Researchers at Google DeepMind tried to teach an AI system to have that same sense of “intuitive physics” by training a model that learns how things move by focusing on objects in videos instead of individual pixels. They trained the model on hundreds of thousands of videos to learn how an object behaves. If babies are surprised by something like a ball suddenly flying out of the window, the theory goes, it is because the object is moving in a way that violates the baby’s understanding of physics. The researchers at Google DeepMind managed to get their AI system, too, to show “surprise” when an object moved differently from the way it had learned that objects move.
Yann LeCun, a Turing Prize winner and Meta’s chief AI scientist, has argued that teaching AI systems to observe like children might be the way forward to more intelligent systems. He says humans have a simulation of the world, or a “world model,” in our brains, allowing us to know intuitively that the world is three-dimensional and that objects don’t actually disappear when they go out of view. It lets us predict where a bouncing ball or a speeding bike will be in a few seconds’ time. He’s busy building entirely new architectures for AI that take inspiration from how humans learn. We covered his big bet for the future of AI here.
The AI systems of today excel at narrow tasks, such as playing chess or generating text that sounds like something written by a human. But compared with the human brain—the most powerful machine we know of—these systems are brittle. They lack the sort of common sense that would allow them to operate seamlessly in a messy world, do more sophisticated reasoning, and be more helpful to humans. Studying how babies learn could help us unlock those abilities.
