Researchers at the University of Cambridge built a game mod for Civilization that allows people to grow—and tame—a human-beating superintelligence.
The game: Shahar Avin from the university’s Centre for the Study of Existential Risk says the game allows players to build AI R&D capacity in their virtual world. Success is a smarter-than-human AI. Simultaneously, rogue AI risk builds and is mitigated only through safety research.
The goal: “We wanted to test our understanding of the concepts relating to existential risk from artificial superintelligence by deploying them in a restricted yet rich world simulator, [and also test how] players respond … when given the opportunity” to make AI decisions themselves, Avin told me.
Fear or acceptance: The project is also an outreach tool, taking “serious issues to a wider audience.” Avin is “aware” that it could could perpetuate fears about AI. He argues, though, that players “see the risk, but also see the range of tools at their disposal to mitigate it.”
DeepMind’s cofounder: Generative AI is just a phase. What’s next is interactive AI.
“This is a profound moment in the history of technology,” says Mustafa Suleyman.
AI hype is built on high test scores. Those tests are flawed.
With hopes and fears about the technology running wild, it's time to agree on what it can and can't do.
You need to talk to your kid about AI. Here are 6 things you should say.
As children start back at school this week, it’s not just ChatGPT you need to be thinking about.
AI language models are rife with different political biases
New research explains you’ll get more right- or left-wing answers, depending on which AI model you ask.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.