Google and the End of Science

Bringing it all back Hume

  • alert
  • submit to reddit

Internet Security Threat Report 2014

WiReD magazine's editor-in-chief Chris Anderson has just seen the end for scientific theories. And it is called Google.

This remarkable revelation was triggered by Google's research director Peter Norvig. Speaking at O'Reilly's Emerging Technology Conference in March. Norvig claimed: "All models are wrong, and increasingly you can succeed without them" - a reference to Google's success at linking web-pages with users. Anderson has generalized that idea to science as a whole in a piece titled The End of Theory: The Data Deluge Makes the Scientific Method Obsolete:

"This is a world where massive amounts of data and applied mathematics replace every other tool that might be brought to bear. Out with every theory of human behavior, from linguistics to sociology. Forget taxonomy, ontology, and psychology. Who knows why people do what they do? The point is they do it, and we can track and measure it with unprecedented fidelity. With enough data, the numbers speak for themselves."

Anderson contends that the same applies for all science - its models are inherently of limited value. Either they are wrong, for example they "caricature... a more complex underlying reality" (quantum mechanics), or we don't know how to prove them experimentally (string theories about the universe), or they raise more questions than they answer (epigenetics in biology).

Yet increasing computing power, both in hardware and statistical analysis algorithms, can still bring forth useful correlations, and new interesting discoveries. Anderson cites Craig Venter's DNA sequencing: having done with sequencing individuals, "in 2005 he started sequencing the air. In the process, he discovered thousands of previously unknown species of bacteria and other life-forms."

"The opportunity is great", he adds, because "correlation supersedes causation, and science can advance even without coherent models, unified theories, or really any mechanistic explanation at all."

Over at Ars Technica, John Timmer evinces shock: "I can't possibly imagine how he comes to that conclusion."

He objects: "Correlations are a way of catching a scientist's attention, but the models and mechanisms that explain them are how we make the predictions that not only advance science, but generate practical applications."

The advancement of science is not itself at issue, but the actual examples Timmer counters with do not seem to convince even Timmer himself.

The royal road would be to demonstrate that models are crucial to science, which would be grounds for thinking that they are logically necessary. Timmer takes the short cut on pragmatic grounds: models have utility, regardless of their truth or falsity. Models, so to speak, make the scientific establishment go around.

"Would Anderson be willing to help test a drug that was based on a poorly understood correlation pulled out of a datamine?" Timmer challenges, apparently unembarrassed to be seen in flagrante putting an ad hominem argument. Of course not, which is why we test on guinea pigs. (And why should Anderson be first?)

But if anything, this is a reason Anderson could use. With sufficiently good correlations, it might finally be possible to spare guinea pigs, chimpanzees, or rats trial by laboratory testing.

The irony here is that eudemonic theories of ethics, which is to say the good and right thing to do is to create happiness, such as hedonism (for me) or utilitarianism (for all of us), are philosophically shakier than statistical inference. Anderson's contention is that technology is changing and, in comparison with the continued daily rising of the sun, the outlook for models and mechanisms on inductive grounds seems less sunny than that of tomorrow's dawn.

A closer shave with history

From an initial pass over the arguments for and against Anderson's "end of theory" claim, it seems that several theories about the justification of science might also have to be added to his hit-list. This is what makes Anderson's argument interesting - an analogue perhaps of "the end of history" claim by Francis Fukuyama in his eponymous book.

How could it happen that Occam's Razor, the (ahem) eponymous principle that explanations should not rely on unnecessary entities, has grown so big that it now threatens to sever the hand that once so securely held it - the hand of scientific practice?

Before addressing that, we should be aware of a slippery complexity - semantics. It is not only Google that "washes" meanings, as The Register's Andrew Orlowski noticed.

The term "model" at one time connoted a physical representation, in scientific context and ordinary contexts, for example, of an atom. It seems now to be used in science to cover a wider range of things: not only the virtual representation of physical models (computer modelling and simulation), but any explanatory matrix where two concepts are mediated by other concepts. Pushed this far, it can be difficult to draw the line between a model and an explanation. And between hypothesis, conjecture, theory, and mechanism. Hold the thought as you read on.

Internet Security Threat Report 2014

Next page: Back home with Hume

More from The Register

next story
Renewable energy 'simply WON'T WORK': Top Google engineers
Windmills, solar, tidal - all a 'false hope', say Stanford PhDs
The next big thing in medical science: POO TRANSPLANTS
Your brother's gonna die, kid, unless we can give him your, well ...
SEX BEAST SEALS may be egging each other on to ATTACK PENGUINS
Boffin: 'I think the behaviour is increasing in frequency'
Post-pub nosh neckfiller: The MIGHTY Scotch egg
Off to the boozer? This delicacy might help mitigate the effects
I'M SO SORRY, sobs Rosetta Brit boffin in 'sexist' sexy shirt storm
'He is just being himself' says proud mum of larger-than-life physicist
NASA launches new climate model at SC14
75 days of supercomputing later ...
Britain's HUMAN DNA-strewing Moon mission rakes in £200k
3 days, and Kickstarter moves lander 37% nearer takeoff
prev story


Why cloud backup?
Combining the latest advancements in disk-based backup with secure, integrated, cloud technologies offer organizations fast and assured recovery of their critical enterprise data.
Forging a new future with identity relationship management
Learn about ForgeRock's next generation IRM platform and how it is designed to empower CEOS's and enterprises to engage with consumers.
How to determine if cloud backup is right for your servers
Two key factors, technical feasibility and TCO economics, that backup and IT operations managers should consider when assessing cloud backup.
High Performance for All
While HPC is not new, it has traditionally been seen as a specialist area – is it now geared up to meet more mainstream requirements?
Security and trust: The backbone of doing business over the internet
Explores the current state of website security and the contributions Symantec is making to help organizations protect critical data and build trust with customers.