Skip to Content
Artificial intelligence

AI arms control may not be possible, warns Henry Kissinger

March 1, 2019

Henry Kissinger, former US secretary of state and a controversial giant of American foreign policy, believes it may be a lot harder to control the development of AI weapons than nuclear ones.

Warning shots: Speaking at an event at MIT yesterday, Kissinger warned that new developments in AI will bring all sorts of dangers. In particular, he worries that the AI weapons could be harder to control than nukes because development of the technology will happen in secret: “With AI, the other side’s ignorance is one of your best weapons—sharing will be much more difficult.”

Peace movement: He isn’t the only person worried about an AI arms race. A significant number of academics, industry researchers, and tech luminaries have backed a campaign to ban the use of autonomous weapons.

Arms dealing: The issue is challenging, however. As security researcher Paul Scharre writes in a recent book, Army of None, autonomy has been creeping into weapons systems for decades. It isn’t always easy to draw a line around autonomous systems, and the technology is alluring because it can be used to make weapons more reliable.

Dumbing down: Kissinger has been something of an AI naysayer lately. After learning about AlphaGo, he wrote an article in the Atlantic warning that the technology could alter the nature of human knowledge and discovery in ways that ultimately harm humanity (“How the Enlightenment Ends”).

“By mastering certain competencies more rapidly and definitively than humans, [AI] could over time diminish human competence and the human condition itself as it turns it into data,” Kissinger wrote. “Philosophically, intellectually—in every way—human society is unprepared for the rise of artificial intelligence.”

Deep Dive

Artificial intelligence

Why Meta’s latest large language model survived only three days online

Galactica was supposed to help scientists. Instead, it mindlessly spat out biased and incorrect nonsense.

A bot that watched 70,000 hours of Minecraft could unlock AI’s next big thing

Online videos are a vast and untapped source of training data—and OpenAI says it has a new way to use it.

Responsible AI has a burnout problem

Companies say they want ethical AI. But those working in the field say that ambition comes at their expense.

Biotech labs are using AI inspired by DALL-E to invent new drugs

Two groups have announced powerful new generative models that can design new proteins on demand not seen in nature.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.