SAN FRANCISCO Computers, like humans, can learn. But when Google tries to fill in your search box based only on a few keystrokes, or your iPhone predicts words as you type a text message, its only a narrow mimicry of what the human brain can do.
The challenge in training a computer to behave like a human brain is technological and physiological, testing the limits of computer and brain science. But researchers from IBM Corp. say theyve made a key step toward combining the two worlds.
The company announced last week that it has built two prototype chips that it says process data more like how humans digest information than the chips that now power PCs and supercomputers.
A milestoneThe chips represent a significant milestone in a six-year-long project that has involved 100 researchers and some $41 million in funding from the governments Defense Advanced Research Projects Agency, or DARPA. Thats the Pentagon arm that focuses on long-term research and previously brought the world the Internet. IBM also has committed an undisclosed amount of money.
The prototypes offer further evidence of the growing importance of parallel processing, or computers doing multiple tasks simultaneously. That is important for rendering graphics and crunching large amounts of data.
The uses of the IBM chips are prosaic, such as steering a simulated car through a maze, or playing Pong. It may be a decade or longer before the chips make their way out of the lab and into actual products.
But whats important is not what the chips are doing, but how theyre doing it, said Giulio Tononi, a professor of psychiatry at the University of Wisconsin at Madison who worked with IBM on the project.
Ability to adaptThe chips ability to adapt to types of information that it wasnt specifically programmed to expect is a key feature.
Theres a lot of work to do still, but the most important thing is usually the first step, Tononi said in an interview. And this is not one step, its a few steps.
Technologists have long imagined computers that learn like humans. Your iPhone or Googles servers can be programmed to predict certain behavior based on past events. But the techniques being explored by IBM and other companies and university research labs around cognitive computing could lead to chips that are better able to adapt to unexpected information.
IBMs interest in the chips lies in their ability to potentially help process real-world signals such as temperature or sound or motion and make sense of them for computers.
IBM, which is based in Armonk, N.Y., is a leader in a movement to link physical infrastructure, such as power plants or traffic lights, and information technology, such as servers and software that help regulate their functions. Such projects can be made more efficient with tools to monitor the myriad analog signals present in those environments.
Dharmendra Modha, project leader for IBM Research, said the new chips have parts that behave like digital neurons and synapses that make them different from other chips. Each core, or processing engine, has computing, communication and memory functions.
You have to throw out virtually everything we know about how these chips are designed, he said.
The key, key, key difference really is the memory and the processor are very closely brought together. Theres a massive, massive amount of parallelism.
The project is part of the same research that led to IBMs announcement in 2009 that it had simulated a cats cerebral cortex, the thinking part of the brain, using a massive supercomputer.
Using progressively bigger supercomputers, IBM had previously simulated 40 percent of a mouses brain in 2006, a rats full brain in 2007, and 1 percent of a humans cerebral cortex in 2009.