Are brains analog, or digital?

Hell breaks loose after Cornell claim

  • alert
  • submit to reddit

SANS - Survey on application security programs

We can't turn off Analog

Did this make sense to neuroscientist Dr Bill Softky? Softky studied under Carver Mead at CalTech and recently worked at the Redwood Neuroscience Institute. He also knows a little about how computers work, and was awarded a prestige Microsoft prize for his work on the Intrinsia debugger - even though he's never worked for Microsoft.

"Because there is no accepted answer for how the brain works, people can say anything. The threshold for disproving something is higher than the threshold for saying it, which is a recipe for the accumulation of bullshit," he says.

Go on, Bill, tell us what you really think.

As people throw the computer metaphor around, they appear to be talking about quite differing ideas of what a computer really is, and the conversations seem to fly right past each other. Softky appears to be on safer empirical ground on what we might more usefully call a "circuitry" metaphor.

"We know that with brain neurons, at least 90 per cent of the bandwidth they use is digital. There is a fibre or there is not a fibre, there is a pulse or there is not a pulse; there is a ground truth to what we see," he explains.

"People see analog signals sloshing around the brain in MRI scans all the time, but much of that is from instrumental blur; the individual circuit elements and computations are still digital."

The two views then, would appear to be irreconcilable.

Or maybe not. Dr Softky says he agrees with much of what Professor Spivey values too, as expressed in a Slashdot post, where he defends the value of probablistic algorithms and neural networks, "(programmed on digitial computers)" as useful and informative ways to build simulations of various human mental processes.

So is saying that the brain works like an analog computer a better fit? Analog computing is still capable of extaordinary computation quite beyond the capabilties of a Turing machine, or a digital computer. You can see its appeal.

Today's digital computers are fast and cheap, easily produced, and ubiquitous, but that's the consequence of politics. And so analog computing is economically and philosophically unacceptable now we live in the "digital era", or "information society", or however it's branded this week. Such arguments neglect the fact that today's "economics" is entirely an agent of politics. In other words, what science does is exactly what we deem acceptable. And so, as Stephen Jay Gould pointed out, just as you can't unwind the tape of evolution and be guaranteed an instant and indentical replay, you can't assume that science would trick out just as it did. Science has always been contingent on economics and ethics, things which science can help inform, but ultimately can't decide for itself.

But it gets interesting. The two views are apparently not irreconcilable.

"I like his point about Hidden Markov models, which are practically the only 'perceptual' system of any use at all. Brains definitely work by trying to figure out what's in the world, and to do that they need probability estimates, which means analog values assigned to multiple possibilities at once (i.e. no clear "winner" at first)," says Softky.

(Jeff Hawkins took some convincing on the value of probabilities, but is now a convert.)

Softky continues -

"But it's a long way from saying that the brain represents probabilities, to somehow saying that's all it does, or saying the circuite elements do that."

"By analogy: speedsheets, tax software, cell phones, and music players all clearly represent analog values too. Does that mean they're 'analog computers'?" he asks.

"Virtually every electronic device that deals with analog information uses digital elements to do so (except for old-fashioned radios and tape players)."

"The outside world is analog in many respects, so brains have to reflect that. But efficient and modular processing elements are digital, so brains ought to take advantage of that too."

Then again, perhaps it's time the computer metaphor was dropped altogether from what may still exist as the "hard sciences", and certainly from the social sciences. Most of today's premiers obliging fire up the computer metaphor when they're stuck for ideas.

Spivey's description of a computer fails to apply to analog hardware, but it doesn't really apply to vector or superscalar processors either. And are we describing the 'chip', or the 'system'? If we can't agree to agree on what a computer looks like, we can't really begin to guess which kind of computer we most resemble. (And at this point you wonder if grown-ups haven't got better things to do than debate such things. What kind of carpet are we? Or car key? Who cares?)

So comparing the brain to a computer isn't fair to humans, and it isn't even fair to computers. You have to allow science to be science, and the metaphors to live their own life. Neither party can't make what David Letterman might call an "assgrab", in order to claim the authority of the other - and recent history is full of such presumptions - without anticipating some pushback. Both camps have, of late, settled into a state of snarling mistrust.

What Spivey's excercise seems to prove is that you can't attempt to "unite" the two camps unless one's own work is very firmly on terra firma to begin with. And when we see Spivey's students moosing around with a mouse, we can see it isn't. ®

Bootnote: But if you want to hear more on the wonders, and limitations of analog - just mail us, and we'll do our best to oblige. ®

Related stories

Scientists see women's brains switch off during sex
Software engineers - the ultimate brain scientists?
Design patterns for a Black Box Brain?
Swiss neurologists to model the brain
Hoax paper fools cybernetic boffins

Top three mobile application threats

More from The Register

next story
Red-faced LOHAN team 'fesses up in blown SPEARS fuse fiasco
Standing in the corner, big pointy 'D' hats
KILLER SPONGES menacing California coastline
Surfers are safe, crustaceans less so
LOHAN's Punch and Judy show relaunches Thursday
Weather looking good for second pop at test flights
Discovery time for 200m WONDER MATERIALS shaved from 4 MILLENNIA... to 4 years
Alloy, Alloy: Boffins in speed-classification breakthrough
Curiosity finds not-very-Australian-shaped rock on Mars
File under 'messianic pastries' and move on, people
Elon Musk's LEAKY THRUSTER gas stalls Space Station supply run
Helium seeps from Falcon 9 first stage, delays new legs for NASA robonaut
Top Secret US payload launched into space successfully
Clandestine NRO spacecraft sets off on its unknown mission
New FEMTO-MOON sighted BIRTHING from Saturn's RING
Icy 'Peggy' looks to be leaving the outer rings
Melting permafrost switches to nasty, high-gear methane release
Result? 'Way more carbon being released into the atmosphere as methane'
prev story


Designing a defence for mobile apps
In this whitepaper learn the various considerations for defending mobile applications; from the mobile application architecture itself to the myriad testing technologies needed to properly assess mobile applications risk.
3 Big data security analytics techniques
Applying these Big Data security analytics techniques can help you make your business safer by detecting attacks early, before significant damage is done.
Five 3D headsets to be won!
We were so impressed by the Durovis Dive headset we’ve asked the company to give some away to Reg readers.
The benefits of software based PBX
Why you should break free from your proprietary PBX and how to leverage your existing server hardware.
Securing web applications made simple and scalable
In this whitepaper learn how automated security testing can provide a simple and scalable way to protect your web applications.