Skip to Content

Facebook Will Try to Outsource a Fix for Its Fake-News Problem

The social network is clearly not comfortable deciding whether content in users’ news feeds is true.
December 16, 2016

Facebook has announced a new initiative to help its users distinguish real news from fake, with a tool that neatly sidesteps any requirement for it to act as an arbiter of truth.

The social network will use a number of traits about content on its site, such as how often a post is deleted and whether comments on a post contain arguments over veracity, to weed out fake items. Suspect posts will be presented via a dashboard to fact-checking organizations Snopes, Politifact, ABC News, and FactCheck.org for analysis. If at least two of them contest facts within an article, users will see it tagged as “Disputed by 3rd Party Fact Checkers.”

That neatly absolves the social network from what is, admittedly, the incredibly tough call of whether a news item is genuine. Mark Zuckerberg has made no secret of the fact that he would rather not have Facebook making decisions about what is true and what is false. While he’s promised that more will be done, he’s also said that the issues are “complex, both technically and philosophically.”

Not that Facebook is entirely removed from the process. First, it will tweak its news feeds to suppress stories flagged as untrustworthy by the consortium of truth-seekers. And the Verge notes that employees will have to work out which posts are personal and which are purporting to be news. Facebook says that its employees won’t have to pass judgment on content, though it remains to be seen how easy it is to neatly divide between personal updates and news.

Regardless of how easy it is, it seems we need these kinds of tools. Following the furor surrounding fake news in the wake of the election, a new Pew Research Center study says that 64 percent of U.S. adults say fabricated news stories cause confusion about current affairs. What’s more, 23 percent say that they have shared a made-up news story.  

Maybe next time a similar survey is carried out, those numbers will be smaller.

(Read more: Facebook, Pew Research Center, “Regardless of Its Influence on the Election, Facebook Needs to Change,” “Facebook’s Fake-News Ad Ban Is Not Enough,” “Facebook’s Content Blocking Sends Some Very Mixed Messages”)

Keep Reading

Most Popular

The Steiner tree problem:  Connect a set of points with line segments of minimum total length.
The Steiner tree problem:  Connect a set of points with line segments of minimum total length.

The 50-year-old problem that eludes theoretical computer science

A solution to P vs NP could unlock countless computational problems—or keep them forever out of reach.

section of Rima Sharp captured by the LRO
section of Rima Sharp captured by the LRO

The moon didn’t die as early as we thought

Samples from China’s lunar lander could change everything we know about the moon’s volcanic record.

conceptual illustration of a heart with an arrow going in on one side and a cursor coming out on the other
conceptual illustration of a heart with an arrow going in on one side and a cursor coming out on the other

Forget dating apps: Here’s how the net’s newest matchmakers help you find love

Fed up with apps, people looking for romance are finding inspiration on Twitter, TikTok—and even email newsletters.

ASML machine
ASML machine

Inside the machine that saved Moore’s Law

The Dutch firm ASML spent $9 billion and 17 years developing a way to keep making denser computer chips.

Stay connected

Illustration by Rose WongIllustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.