The news: A deepfake of the president of India’s ruling Bharatiya Janata Party (BJP), Manoj Tiwari, went viral on WhatsApp in the country earlier this month, ahead of legislative assembly elections in Delhi, according to Vice. It’s the first time a political party anywhere has used a deepfake for campaigning purposes. In the original video Tiwari speaks in English, criticizing his political opponent Arvind Kejriwal and encouraging voters to vote for the BJP. The second video has been manipulated using deepfake technology so his mouth moves convincingly as he speaks in Haryanvi, the Hindi dialect spoken by the target voters for the BJP.
The purpose: The BJP has partnered with political communications firm The Ideaz Factory to create deepfakes that let it target voters across the over 20 different languages used in India. The party told Vice that the Tiwari deepfake reached approximately 15 million people in 5,800 WhatsApp groups.
Causing alarm: This isn’t the first time deepfakes have popped up during a political campaign. For example, last December, researchers made a fake video of the two candidates in the UK’s general election endorsing each other. It wasn’t supposed to sway the vote, however—merely to raise awareness about deepfake technology. This case in India seems to be the first time deepfakes have been used for a political campaign. The big risk is that we reach a point where people can no longer trust what they see or hear. In that scenario, a video wouldn’t even need to be digitally altered for people to denounce it as fake. It’s not hard to imagine the corrosive impact that would have on an already fragile political landscape.
Sign up here to our daily newsletter The Download to get your dose of the latest must-read news from the world of emerging tech.
A Roomba recorded a woman on the toilet. How did screenshots end up on Facebook?
Robot vacuum companies say your images are safe, but a sprawling global supply chain for data from our devices creates risk.
The viral AI avatar app Lensa undressed me—without my consent
My avatars were cartoonishly pornified, while my male colleagues got to be astronauts, explorers, and inventors.
Roomba testers feel misled after intimate images ended up on Facebook
An MIT Technology Review investigation recently revealed how images of a minor and a tester on the toilet ended up on social media. iRobot said it had consent to collect this kind of data from inside homes—but participants say otherwise.
How to spot AI-generated text
The internet is increasingly awash with text written by AI software. We need new tools to detect it.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.