Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Physical Address
304 North Cardinal St.
Dorchester Center, MA 02124
Two researchers who developed tools for understanding the neural networks that underpin today’s boom in artificial intelligence (AI) have won the 2024 Nobel Prize in Physics.
John Hopfield at Princeton University in New Jersey, and Geoffrey Hinton at the University of Toronto, Canada, share the 11 million Swedish kronor (US$1 million) prize, announced by the Royal Swedish Academy of Sciences in Stockholm on 8 October.
Both used tools from physics to come up with methods that power artificial neural networks, which exploit brain-inspired, layered structures to learn abstract concepts. Their discoveries “form the building blocks of machine learning, that can aid humans in making faster and more-reliable decisions”, said Nobel committee chair Ellen Moons, a physicist at Karlstad University, Sweden, during the announcement. “Artificial neural networks have been used to advance research across physics topics as diverse as particle physics, material science and astrophysics.”
In 1982, Hopfield, a theoretical biologist with a background in physics, came up with a network that described connections between virtual neurons as physical forces1. By storing patterns as a low-energy state of the network, the system could re-create the pattern when prompted with something similar. It became known as associative memory, because the way in which it ‘recalls’ things is similar to the brain trying to remember a word or concept based on related information.
Hinton, a computer scientist, used principles from statistical physics, which collectively describes systems that have too many parts to track individually, to further develop the ‘Hopfield network’. By building probabilities into a layered version of the network, he created a tool that could recognize and classify images, or generate new examples of the type it was trained on2.
Computer science: The learning machines
These processes differed from earlier kinds of computation, in that the networks were able to learn from examples, including from complex data. This would have been challenging for conventional software reliant on step-by-step calculations.
The networks are “grossly idealized models that are as different from real biological neural networks as apples are from planets”, Hinton wrote in Nature Neuroscience in 2000. But they proved useful and have been built upon widely. Neural networks that mimic human learning form the basis of many state-of-the-art AI tools, from large language models (LLMs) to machine-learning algorithms capable of analysing large swathes of data, including the protein-structure-prediction model AlphaFold.
Speaking by telephone at the announcement, Hinton said that learning he had won the Nobel was “a bolt from the blue”. “I’m flabbergasted, I had no idea this would happen,” he said. He added that advances in machine learning “will have a huge influence, it will be comparable with the industrial revolution. But instead of exceeding people in physical strength, it’s going to exceed people in intellectual ability”.
In recent years, Hinton has become of the loudest voices calling for safeguards to be placed around AI. He says he became convinced last year that digital computation had become better than the human brain, thanks to its ability to share learning from multiple copies of an algorithm, running in parallel. “Up until that point, I’d spent 50 years thinking that if we could only make it more like the brain, it will be better,” he said on 31 May, speaking virtually to the AI for Good Global Summit in Geneva, Switzerland. “It made me think [these systems are] going to become more intelligent than us sooner than I thought.”
Hinton also won the Alan Turing Award in 2018 — sometimes described as the ‘Nobel of computer science’. Hopfield, too, has won several other prestigious physics awards, including the Dirac Medal.
“[Hopfield’s] motivation was really physics, and he invented this model of physics to understand certain phases of matter,” says Karl Jansen, a physicist at the German Electron Synchrotron laboratory (DESY) in Zeuthen, who describes the work as “groundbreaking”. After decades of developments, neural networks have become an important tool in analyzing data from physics experiments and in understanding the types of phase transitions that Hopfield had set out to study, Jansen adds.
How to win a Nobel prize: what kind of scientist scoops medals?
Lenka Zdeborová, a specialist in the statistical physics of computation at the Swiss Federal Institute of Technology in Lausanne (EPFL), says that she was pleasantly surprised that the Nobel Committee recognized the importance of physics ideas for understanding complex systems. “This is a very generic idea, whether it’s molecules or people in society.”
In the last five years, the Abel Prize and Fields Medals have also celebrated the cross-fertilization between mathematics, physics and computer science, particularly contributions to statistical physics.
Both winners “have brought incredibly important ideas from physics into AI”, says Yoshua Bengio, a computer scientist who shared the 2018 Turing Award with Hinton and fellow neural-network pioneer Yann LeCun. Hinton’s seminal work and infectious enthusiasm made him a great role model for Bengio and other early proponents of neural networks. “It inspired me incredibly when I was just a university student,” says Bengio, who is director of the Montreal Institute for Learning Algorithms in Canada. Many computer scientists regarded the neural network as fruitless for decades, says Bengio — a major turning point was when Hinton and others used it to win a major image-recognition competition in 2012.
Biology has also benefited from these artificial models of the brain. May-Britt Moser, a neuroscientist at the Norwegian University of Science and Technology in Trondheim and a winner of the 2014 Nobel Prize in Physiology or Medicine, says she was “so happy” when she saw the winners announced. Versions of Hopfield’s network models have been useful to neuroscientists, she says, when investigating how neurons work together in memory and navigation. His model, which describes memories as low points of a surface, helps researchers to visualise how certain thoughts or anxieties can get fixed and retrieved in the brain, she adds. “I love to use this as a metaphor to talk to people when they are stuck.”
Ten computer codes that transformed science
Today, neuroscience relies on network theories and machine learning tools, which stemmed from Hopfield and Hinton’s work, to understand and process data on thousands of cells simultaneously, says Moser. “It’s like a fuel for understanding things that we couldn’t even dream of when we started in this field”.
“The use of machine-learning tools is having immeasurable impact on data analysis and our potential understanding of how brain circuits may compute,” says Eve Marder, a neuroscientist at Brandeis University in Waltham, Massachusetts. “But these impacts are dwarfed by the many impacts that machine learning and artificial intelligence are having in every aspect of our daily lives.”