For decades, scientists have fantasized about creating robots with brain-like intelligence. This year, researchers tempted by that dream made great progress on achieving what has been called the holy grail of computing.
Today, a wide variety of efforts are aimed at creating intelligent computers that can progressively learn and make smarter decisions. Millions of dollars this year were poured in efforts to create "silicon brains," or neuromorphic chips that mimic brain-like functionality to make computers smarter.
The new chips could give eyes and ears to smart robots, which will be able to drive, identify objects, or even point out rotten fruit. This chip technology could let humans get mind control over machines, mobile devices anticipate actions by users and wearable devices like Google Glass to diagnose diseases. In the long run, neural chip implants could boost mental, visual, and cognitive capabilities of humans.
Scientists are looking to create advanced computers with these neural chips, which replicate the brain's circuitry and can retain information and make decisions based on patterns discovered through probabilities and associations. Projects funded by the U.S. government, the European Union, and private organizations are attempting to re-create the manner in which the brain's neurons and synapses work by redesigning the memory, computation, and communication features of traditional circuitry.
The brain has 100 billion interconnected neurons, nerve cells that process and transmit information via electrical and chemical signals. These neurons compute in parallel and communicate via trillions of connections, which are the synapses. Connections among neurons in the neural network are either strengthened or pruned as the brain learns more. Today's processors are wired and regulate voltage differently than the brain's neural network, but researchers are keen on exploiting the parallelism of the brain which, among other things, also reduces power requirements.
Researchers hope neural chips will accomplish cognitive tasks and respond to a wide range of stimuli. Computers can already see and hear; robots have already been built to respond to sensory input. Within five years, computers could get smell and taste, and this sort of sensory information could be fed to chips for processing.
To be sure, most of the chip-development efforts are in early experimental phases. Brains of small insects and worms have been simulated on prototype neural chips, but human brains operate on a different scale. While it could be decades until chips simulate the human brain, the groundwork is bring laid by new models of computing that are now being established.
Among other things, new data-processing techniques are needed that allow more information to be fed to computers, researchers said. Helping this effort, the physical limits of manufacturing techniques for the chips that power today's computers could fall within a decade, opening the door for new computing designs and chip architectures, said Robert Colwell, director of the microsystems technology office at DARPA (Defense Advanced Research Projects Agency), in a speech earlier this year.
Currently, computers don't have the capacity to learn from past experiences. Instead, they rely on pre-programmed code to make decisions. On the other hand, brain cells do not require programming, are high tolerance, can regenerate, and can draw conclusions that computers are not able to reach, said Karlheinz Meier, professor and chair of experimental physics at the University of Heidelberg.
Traditional computers won't go away, meanwhile, as some activities don't require intelligent processing, said Meier, who is also co-director of the European Union-funded Human Brain Project.
"You will always do your text processing and email," Meier said.
But like the brain, neural chips will excel at certain things, like cutting through "noisy" data to make intelligent decisions, said Nabil Imam, a computer scientist and researcher at Cornell University.
The neuromorphic chips will complement, not replace, other processors in a computer, Imam said.
Chips modeled after the human brain have electronic neurons that can dynamically rewire the connections among them, blast information at each other, and forage for relevant data?a process more power efficient than throwing lots of data to CPUs and other coprocessors like GPUs. IBM's Watson supercomputer made history when it beat participants at the game of Jeopardy, but it threw lots of data at processors to find answers.
"Our brains were wired to do certain things very well like pattern recognition. Computers can't do that. These processors have a different class of applications," Imam said.
Imam is involved in the development of neuromorphic chips as part of the multiphase Synapse (Systems of Neuromorphic Adaptive Plastic Scalable Electronics) project funded by DARPA. The Synapse project, initiated in 2008, involves IBM, Hewlett-Packard, Cornell, Stanford University and other universities.
The first tangible results for Synapse came in early 2011, when IBM demonstrated a prototype chip with 256 digital neurons running at slow speeds of 10MHz. The chip was able to demonstrate navigation and pattern recognition abilities.
One chip core had 262,144 programmable synapses, while another core had 65,536 learning synapses. The connections between digital neurons got stronger depending on the number of signals sent. If an electronic spike from one neuron affects the voltage of another neuron, the two are synaptically connected. In chips, spiking neurons communicate with other neurons when triggers, such as certain values, are reached.
Address: 5636 Lemon Ave.
Dallas TX 75209
Phone: +1 214 5203694