Brainy computers

ZURICH • Ever since the American computer scientist John McCarthy coined the term "Artificial Intelligence" in 1955, the public has imagined a future of sentient computers and robots that think and act like humans. But while such a future may indeed arrive, it remains, for the moment, a distant prospect.

And yet the foreseeable frontier of computing is no less exciting. We have entered what we at IBM call the Cognitive Era. Breakthroughs in computing are enhancing our ability to make sense of large bodies of data, providing guidance in some of the world's most important decisions, and potentially revolutionising entire industries.

The term "cognitive computing" refers to systems that, rather than being explicitly programmed, are built to learn from their experiences. By extracting useful information from unstructured data, these systems accelerate the information age, helping their users with a broad range of tasks, from identifying unique market opportunities to discovering new treatments for diseases to crafting creative solutions for cities, companies and communities.

The Cognitive Era marks the next stage in the application of science to understand nature and advance human prosperity. Its beginning dates to early 2011, when the cognitive computing system Watson beat two human champions on the game show Jeopardy!. Since then, Watson has gone on to do much more, demonstrating how cognitive computing can use big data to tackle some of the most difficult systemic issues facing humanity.

Broadly, cognitive systems offer five core capabilities. First, they create deeper human engagement, using data about an individual to create more fully human interactions. Second, they scale and elevate expertise, learning from experts in various fields and making that know-how available to broad populations. Third, they provide products, such as those connected to the Internet of Things, with the ability to sense the world around them and to learn about their users.

Fourth, they allow their operators to make sense of large amounts of data, helping to manage workflows, provide context, and allow for continuous learning, better forecasting and improved operational effectiveness. Finally, and perhaps most importantly, they allow their users to perceive patterns and opportunities that would be impossible to discover through traditional means.

Cognitive systems are inspired by the human brain, an organ that still has much to teach us. With systems growing in size and complexity, traditional computer architecture seems to be reaching its limits, as power consumption soars and the transmission delay between components becomes increasingly burdensome. Indeed, when it comes to energy efficiency - measured in terms of the number of computations per energy unit on "unstructured" data - the human brain performs roughly 10,000 times better than the best man-made machines.

Today, computers consume about 10 per cent of the world's electricity output, according to Mr Mark Mills, chief executive officer of the Digital Power Group. In order to benefit fully from the Cognitive Era, we will have to be able to harness huge amounts of information; during the next 15 years, the amount of "digitally accessible" data is expected to grow by a factor of more than 1,000. Performing the calculations necessary for using such a large amount of data will not be possible without huge strides in improving energy efficiency.

Matching the performance and efficiency of the human brain will likely require us to mimic some of its structures. Rather than attempt to squeeze energy-intensive performance out of ever-larger chips, we can arrange computer components in a dense 3D matrix similar to a human brain, maximising not performance, but energy efficiency.

Arranging computer chips in a 3D environment puts the various elements of the computer closer to one another. This not only reduces the time they take to communicate, but also improves energy efficiency by a factor of as much as 5,000 - potentially providing computers with efficiency close to that of a biological brain. Already, a much denser computer built from available mobile technology and hot water cooling allows for 10 times higher efficiency than a conventional system.

But man-made computers are so inefficient not only because they need to power the chips, but also because they need energy to run the air conditioners that remove the heat generated by the processors. The human brain has a lesson to teach here too. Just as the brain uses sugar and blood to cool and provide energy to its various regions, a 3D computer could use coolant fluid to deliver energy to the chips. In addition to dissipating heat, the fluid could be used to power an electrochemical system providing power to the processors. This, in turn, would enable further increases in packaging density - and thus efficiency.

By adopting some of the characteristics of the human brain, computers have the potential to become far more compact, efficient and powerful. And this, in turn, will allow us to take full advantage of cognitive computing - providing our real brains with new sources of support, stimulus and inspiration.

PROJECT SYNDICATE

• The writer is a scientist at IBM Research - Zurich.

Join ST's Telegram channel and get the latest breaking news delivered to you.

A version of this article appeared in the print edition of The Sunday Times on September 18, 2016, with the headline Brainy computers. Subscribe