Feeds

Cognitive computing: IBM unveils software for its brain-like SyNAPSE chips

From the firm that brought you the 'cat brain chip'

Next gen security for virtualised datacentres

IBM has unveiled a new computing language to help boffins better program the mad scientist brain-like silicon chips first envisaged in the first three phases of what is known at The Reg as DARPA's "mad cat brain chip" project.

Prototypes of the chips with the cool cognitive computing architectures have already been unveiled to the world, but Big blue says its new programming model will take the complexity of the project to a new level.

According to a Big Blue press release, the technology can mimic the "function, low power, and compact volume of the brain".

It could be used to create next-generation intelligent sensor networks which are capable of tricks like perception, action and cognition - the sort of things humans take for granted. This could allow for human-like collection and analysis of masses of big data, akin to that performed by our human eyes and brains while we navigate the world.

"Architectures and programs are closely intertwined and a new architecture necessitates a new programming paradigm," said Dr Dharmendra Modha, principal investigator and senior manager at IBM Research. "While complementing today's computers, this will bring forth a fundamentally new technological capability in terms of programming and applying emerging learning systems."

IBM's new programming model is "dramatically different" from traditional stored-program architectures, say the boffins.

Big Blue's latest wheeze dispenses with the sequential operation inherent in von Neumann architectures, replacing it with a "distributed, highly interconnected, asynchronous, parallel, large-scale cognitive computing architectures".

Modern computers are superb number-crunchers, but often find it difficult to cope with "real-time processing of the noisy, analog, voluminous, Big Data produced by the world around us," IBM's researchers continued.

The brain, which operates relatively slowly in calculations and offers a low precision, is nonetheless excellent at sifting through staggering amounts of data, before recognising and acting upon patterns.

Impressively, while interpreting the real world, the brain “uses the same amount of power as a 20-watt light bulb.”

IBM first unveiled a chip based on the brain in 2011. The IBM team created a chip that is designed to chew on streams telemetry and rewire itself, much as your brain does as it learns, as it learns about the world from that telemetry.

So far IBM's effort to build a brain has cost about $53m, with $2m in new funding recently pledged by the US Defense Advanced Research Projects Agency (DARPA), the US Army's battalion of crazed weapon inventors.

The project is known as Systems of Neuromorphic Adaptive Plastic Scalable Electronics (SyNAPSE).

In the long term, IBM wants to build a system with ten billion neurons and hundred trillion synapses, which uses just one kilowatt of power and occupies less than two litres in volume.

In a statement, IBM said:

Systems built from these chips could bring the real-time capture and analysis of various types of data closer to the point of collection. They would not only gather symbolic data, which is fixed text or digital information, but also gather sub-symbolic data, which is sensory based and whose values change continuously. This raw data reflects activity in the world of every kind ranging from commerce, social, logistics, location, movement, and environmental conditions.

The technology could be used to copy the human eye, IBM claimed, each of our working eyes deals with more than a terabyte of information each day. Big Blue suggested its chips could build visors for blind people which capture information and then allow blind people to navigate around the world, following instructions conveyed over headphones.

The initial phase of the project simulated the cortex of a cat brain on an IBM BlueGene massively parallel supercomputer with 147,456 cores and 144TB of memory developing the basic synaptic circuits for the brain chip. ®

Build a business case: developing custom apps

More from The Register

next story
Why has the web gone to hell? Market chaos and HUMAN NATURE
Tim Berners-Lee isn't happy, but we should be
Microsoft boots 1,500 dodgy apps from the Windows Store
DEVELOPERS! DEVELOPERS! DEVELOPERS! Naughty, misleading developers!
Mozilla's 'Tiles' ads debut in new Firefox nightlies
You can try turning them off and on again
'Stop dissing Google or quit': OK, I quit, says Code Club co-founder
And now a message from our sponsors: 'STFU or else'
Apple promises to lift Curse of the Drained iPhone 5 Battery
Have you tried turning it off and...? Never mind, here's a replacement
Uber, Lyft and cutting corners: The true face of the Sharing Economy
Casual labour and tired ideas = not really web-tastic
Linux turns 23 and Linus Torvalds celebrates as only he can
No, not with swearing, but by controlling the release cycle
prev story

Whitepapers

Gartner critical capabilities for enterprise endpoint backup
Learn why inSync received the highest overall rating from Druva and is the top choice for the mobile workforce.
Implementing global e-invoicing with guaranteed legal certainty
Explaining the role local tax compliance plays in successful supply chain management and e-business and how leading global brands are addressing this.
Rethinking backup and recovery in the modern data center
Combining intelligence, operational analytics, and automation to enable efficient, data-driven IT organizations using the HP ABR approach.
Consolidation: The Foundation for IT Business Transformation
In this whitepaper learn how effective consolidation of IT and business resources can enable multiple, meaningful business benefits.
Next gen security for virtualised datacentres
Legacy security solutions are inefficient due to the architectural differences between physical and virtual environments.