yego.me
💡 Stop wasting time. Read Youtube instead of watch. Download Chrome Extension

How to get better at video games, according to babies - Brian Christian


3m read
·Nov 8, 2024

In 2013, a group of researchers at DeepMind in London had set their sights on a grand challenge. They wanted to create an AI system that could beat, not just a single Atari game, but every Atari game. They developed a system they called Deep Q Networks, or DQN, and less than two years later, it was superhuman. DQN was getting scores 13 times better than professional human games testers at “Breakout,” 17 times better at “Boxing,” and 25 times better at “Video Pinball.”

But there was one notable, and glaring, exception. When playing “Montezuma’s Revenge,” DQN couldn’t score a single point, even after playing for weeks. What was it that made this particular game so vexingly difficult for AI? And what would it take to solve it? Spoiler alert: babies. We’ll come back to that in a minute.

Playing Atari games with AI involves what’s called reinforcement learning, where the system is designed to maximize some kind of numerical rewards. In this case, those rewards were simply the game's points. This underlying goal drives the system to learn which buttons to press and when to press them to get the most points. Some systems use model-based approaches, where they have a model of the environment that they can use to predict what will happen next once they take a certain action. DQN, however, is model free. Instead of explicitly modeling its environment, it just learns to predict, based on the images on screen, how many future points it can expect to earn by pressing different buttons.

For instance, “if the ball is here and I move left, more points, but if I move right, no more points.” But learning these connections requires a lot of trial and error. The DQN system would start by mashing buttons randomly, and then slowly piece together which buttons to mash when in order to maximize its score. But in playing “Montezuma’s Revenge,” this approach of random button-mashing fell flat on its face. A player would have to perform this entire sequence just to score their first points at the very end. A mistake? Game over. So how could DQN even know it was on the right track?

This is where babies come in. In studies, infants consistently look longer at pictures they haven’t seen before than ones they have. There just seems to be something intrinsically rewarding about novelty. This behavior has been essential in understanding the infant mind. It also turned out to be the secret to beating “Montezuma’s Revenge.” The DeepMind researchers worked out an ingenious way to plug this preference for novelty into reinforcement learning.

They made it so that unusual or new images appearing on the screen were every bit as rewarding as real in-game points. Suddenly, DQN was behaving totally differently from before. It wanted to explore the room it was in, to grab the key and escape through the locked door— not because it was worth 100 points, but for the same reason we would: to see what was on the other side.

With this new drive, DQN not only managed to grab that first key— it explored all the way through 15 of the temple’s 24 chambers. But emphasizing novelty-based rewards can sometimes create more problems than it solves. A novelty-seeking system that’s played a game too long will eventually lose motivation. If it’s seen it all before, why go anywhere?

Alternately, if it encounters, say, a television, it will freeze. The constant novel images are essentially paralyzing. The ideas and inspiration here go in both directions. AI researchers stuck on a practical problem, like how to get DQN to beat a difficult game, are turning increasingly to experts in human intelligence for ideas.

At the same time, AI is giving us new insights into the ways we get stuck and unstuck: into boredom, depression, and addiction, along with curiosity, creativity, and play.

More Articles

View All
I Waterproofed Myself With Aerogel!
I feel confident. “You’re confident that I am NOT gonna be damaged? Not permanently? Okay, let me back up for a moment. I want to talk about the properties of aerogel, the world’s lightest solid. What I’m gonna do is I’m gonna lean in so it’s coming in t…
In Cambodia, a City of Towering Temples in the Forest | National Geographic
Deep in the forests of Cambodia, Siem Reap Province, an ancient stone city soars skyward. This is the sprawling complex of Angkor Archaeological Park. The site is located in the northwestern region of the country and is only four miles from the city of S…
AK-47 vs Prince Rupert's Drop (at 223,000 FPS) - Smarter Every Day 170
Hey, it’s me Destin, welcome back to Smarter Every Day! I’ve been waiting on a sunny day to do this. You remember the last video I fired a .38 special versus a Prince Rupert’s drop, and the Prince Rupert’s drop won. Well, we’re going to fix that today. W…
Perimeter word problem (skating rink) | Math | 3rd grade | Khan Academy
Gus plans to install a handrail around a skating rink. The rink forms a 40 meter by 20 meter rectangle. How many meters of handrail does Gus need? So here’s what we know about this skating rink: it’s a 40 meter by 20 meter rectangle. So let’s draw the sk…
Comparing P value to significance level for test involving difference of proportions | Khan Academy
A veterinarian is studying a certain disease that seems to be affecting male cats more than female cats. They obtain a random sample of records from 500 cats. They find 24 of the 259 male cats have the disease, while 14 of 241 female cats have the disease…
Tom Preston Werner at Startup School 2012
Hi everyone! It’s awesome to be back here. Was here in 2010, two years ago. Lots changed since then. I’m actually gonna put this on the ground. This is my timer. You see, part of being a founder of a company is solving your own problems. So, I was thinki…