The social-media giant is working on hardware that can analyze and filter live video.
Background: After Facebook rolled out its live video feature in 2016, the company was criticized for a rash of suicides streamed to audiences on the platform. In response, the company created AI tools to spot dangerous behavior and increased the number of reviewers, so that it took less than 10 minutes to remove footage after it was posted.
Improvement: Investing in chips with AI software that can recognize self-harm, sexual acts, or other activities Facebook wants to ban would reduce the need for human moderators to watch suspect videos.
Why it matters: Mark Zuckerberg has big plans for these kinds of custom-built systems. By designing and making its own hardware, Facebook could not only improve its platform but save a lot of money by reducing reliance on chip manufacturers like Nvidia and Intel.
Why Meta’s latest large language model survived only three days online
Galactica was supposed to help scientists. Instead, it mindlessly spat out biased and incorrect nonsense.
DeepMind’s game-playing AI has beaten a 50-year-old record in computer science
The new version of AlphaZero discovered a faster way to do matrix multiplication, a core problem in computing that affects thousands of everyday computer tasks.
A bot that watched 70,000 hours of Minecraft could unlock AI’s next big thing
Online videos are a vast and untapped source of training data—and OpenAI says it has a new way to use it.
The White House just unveiled a new AI Bill of Rights
It's the first big step to hold AI to account.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.