A Game of Civilization May Help People Understand AI’s Existential Threat
Researchers at the University of Cambridge built a game mod for Civilization that allows people to grow—and tame—a human-beating superintelligence.
The game: Shahar Avin from the university’s Centre for the Study of Existential Risk says the game allows players to build AI R&D capacity in their virtual world. Success is a smarter-than-human AI. Simultaneously, rogue AI risk builds and is mitigated only through safety research.
The goal: “We wanted to test our understanding of the concepts relating to existential risk from artificial superintelligence by deploying them in a restricted yet rich world simulator, [and also test how] players respond … when given the opportunity” to make AI decisions themselves, Avin told me.
Fear or acceptance: The project is also an outreach tool, taking “serious issues to a wider audience.” Avin is “aware” that it could could perpetuate fears about AI. He argues, though, that players “see the risk, but also see the range of tools at their disposal to mitigate it.”
Deep Dive
Artificial intelligence
The inside story of how ChatGPT was built from the people who made it
Exclusive conversations that take us behind the scenes of a cultural phenomenon.
AI is dreaming up drugs that no one has ever seen. Now we’ve got to see if they work.
AI automation throughout the drug development pipeline is opening up the possibility of faster, cheaper pharmaceuticals.
GPT-4 is bigger and better than ChatGPT—but OpenAI won’t say why
We got a first look at the much-anticipated big new language model from OpenAI. But this time how it works is even more deeply under wraps.
The original startup behind Stable Diffusion has launched a generative AI for video
Runway’s new model, called Gen-1, can change the visual style of existing videos and movies.
Stay connected
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.