Skip to Content
Uncategorized

Can 10,000 Humans Clean Up YouTube?

December 5, 2017

The algorithms aren’t working. YouTube, like Facebook, comes under consistent fire for objectionable content—from extremism to child abuse. Tech leaders promise that artificial intelligence will help solve the problem by automatically identifying offensive content before anyone can see it. But YouTube’s best attempts so far spot only 75 percent of such clips before a user reports them.

Clearly, AI is not yet enough. That message appears to be being reiterated by YouTube’s CEO, Susan Wojcicki. In a new blog post, she explains that the video site is swelling its moderation team to include 10,000 staff across Google to help battle objectionable content.

What will they do? To a large extent, more of the same. “Since June, our trust and safety teams have manually reviewed nearly 2 million videos for violent extremist content,” she writes. “We are also taking aggressive action on comments, launching new comment moderation tools and in some cases shutting down comments altogether.”

But Wojcicki also points out that all the work has another use: training AIs. “Human judgment is critical to making contextualized decisions on content,” she writes, adding that by collecting data about how its moderators work, YouTube will be able to “train ... machine-learning technology to identify similar videos in the future.”

If this all sounds familiar, well, that’s because Facebook added 3,000 extra content policers itself earlier this year. At the time, we argued that extra people alone, without robust and reliable AI, would be unlikely to make much of a dent in offensive content—because there’s so damn much of it to sift through.

In YouTube’s case, with 300 hours of footage uploaded every minute, it seems equally unlikely to succeed—at least until its algorithms have learned all they need from the meatspace moderators.

Deep Dive

Uncategorized

Embracing CX in the metaverse

More than just meeting customers where they are, the metaverse offers opportunities to transform customer experience.

Identity protection is key to metaverse innovation

As immersive experiences in the metaverse become more sophisticated, so does the threat landscape.

The modern enterprise imaging and data value chain

For both patients and providers, intelligent, interoperable, and open workflow solutions will make all the difference.

Scientists have created synthetic mouse embryos with developed brains

The stem-cell-derived embryos could shed new light on the earliest stages of human pregnancy.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.