Skip to Content
Artificial intelligence

DeepMind’s AI has used teamwork to beat humans at a first-person shooter

Capture the flag game
Capture the flag game
Capture the flag gameDeepMind

Deep-learning algorithms have already mastered games like Starcraft to beat humans, and now they have shown they can team up to beat us too.

The news: In a paper published in Science yesterday, DeepMind showed how it had let AI programs loose in a modified version of the 3D first-person video game Quake III Arena. The team used an algorithm called “For the Win,” which trains a host of agents in parallel using reinforcement learning, the technique that lets AI learn which tactics work and which do not (and that famously enabled DeepMind’s AI to win at Go). This time, AI agents were trained on around 450,000 games of Capture the Flag, the classic game that involves snatching a flag from your opponent’s base while protecting your own.

Each agent could only see a first-person view of the maze-like structure, just as a human player would. The AI agents were mixed up in teams with 40 human players and randomly matched in games—both as opponents and as teammates. To make it even harder, the maps were procedurally generated, meaning no two were the same.

How to win: The teams of AI agents were consistently better than the other pairs and developed teamwork strategies to help them win, including following teammates to outnumber opponents at key moments and waiting near the enemy base to grab a new flag when it appeared. There’s a new video of the agents in action here.

There’s no (A)I in team: The work (which was first published on the arXiv pre-press site before peer review last year) is interesting because it’s hard to get AI to cooperate: cooperation involves so many variables, and all the AI agents are learning independently. There’s the prospect that something like this could help robots operate in the real world more effectively, with each other and with humans.

However, we must be careful not to extrapolate too much. The game was very narrowly defined, and it’s likely the same system couldn’t just transfer to another scenario—never mind real life. In any case, the AI agents were not really collaborating (at least not in the way that humans do, by communicating), Georgia Tech’s Mark Riedl told the New York Times.

For more on the world of AI, sign up here to our weekly AI newsletter, The Algorithm.

Deep Dive

Artificial intelligence

chasm concept
chasm concept

Artificial intelligence is creating a new colonial world order

An MIT Technology Review series investigates how AI is enriching a powerful few by dispossessing communities that have been dispossessed before.

open sourcing language models concept
open sourcing language models concept

Meta has built a massive new language AI—and it’s giving it away for free

Facebook’s parent company is inviting researchers to pore over and pick apart the flaws in its version of GPT-3

spaceman on a horse generated by DALL-E
spaceman on a horse generated by DALL-E

This horse-riding astronaut is a milestone in AI’s journey to make sense of the world

OpenAI’s latest picture-making AI is amazing—but raises questions about what we mean by intelligence.

labor exploitation concept
labor exploitation concept

How the AI industry profits from catastrophe

As the demand for data labeling exploded, an economic catastrophe turned Venezuela into ground zero for a new model of labor exploitation.

Stay connected

Illustration by Rose WongIllustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.