Here are five ways deepfake-busting technology could backfire

Deepfakes have become a symbol for the end of truth and, to some, a potential tool to swing elections. (Never mind that most deepfakes are still fake porn.) Everyone from the US government to tech giants to startups is trying to develop deepfake-busting technology. But a new report out today from Witness, a nonprofit that studies synthetic media, points out how these tools could go wrong.
The techniques: Manipulated video is not a new issue, and there are plenty of social problems that even the best deepfake detector can’t fix. (For example, knowing that a video has been edited doesn’t automatically answer the question of whether it should be taken down. What if it’s satire?) That hasn’t prevented companies like Amber Video, Truepic, and eWitness from developing “verified-at-capture” or “controlled-capture” technologies. These use a variety of techniques to sign, geotag, and time-stamp an image or video when it’s created. In theory, this makes it easier to tell if the media has been tampered with.
What’s the problem? The Witness report lays out 14 different ways that these technologies could actually end up being harmful. Some of the key ones:
—The tools being built could be used to surveil people
—Technical restraints could stop these tools from working in places where they’re most needed (and those using old hardware could be left behind)
—Jailbroken devices won’t be able to capture verifiable material
—Companies could delete the data or not let individuals control it
—Requiring more verification for media in court could make the legal process longer and more expensive
So what can be done? There’s no easy solution to these problems, says Witness program director Sam Gregory. The companies building these technologies must address these questions and think about the people who are most likely to be harmed, he adds. It is also possible to build synthetic media tools themselves in a more ethical way. Technology expert Aviv Ovadya, for instance, has ideas for how to make responsible deepfake tools. Companies can do their best to vet which clients are allowed to use their tools and explicitly penalize those who violate their norms. Synthetic media of all kinds are going to become more common. It’ll take a lot of different tactics to keep us all safe.
Deep Dive
Humans and technology
Why embracing complexity is the real challenge in software today
In the midst of industry discussions about productivity and automation, it’s all too easy to overlook the importance of properly reckoning with complexity.
Turning medical data into actionable knowledge
Technology can transform patient care by bringing together data from a variety of sources
Enabling enterprise growth with data intelligence
It's becoming more critical for organizations to organize data and put data infrastructure at the forefront of their data strategy, says Bharti Patel, SVP of product engineering at Hitachi Vantara.
AI gains momentum in core manufacturing services functions
More use cases means customers and employees experience AI’s ability to automate tasks, prioritize work, and empower the user.
Stay connected
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.