Skip to Content
Artificial intelligence

Put Humans at the Center of AI

At Stanford and Google, Fei-Fei Li is leading the development of artificial intelligence—and working to diversify the field.
October 9, 2017
Blake Farrington

As the director of Stanford’s AI Lab and now as a chief scientist of Google Cloud, Fei-Fei Li is helping to spur the AI revolution. But it’s a revolution that needs to include more people. She spoke with MIT Technology Review senior editor Will Knight about why everyone benefits if we emphasize the human side of the technology.

Why did you join Google?

Researching cutting-edge AI is very satisfying and rewarding, but we’re seeing this great awakening, a great moment in history. For me it’s very important to think about AI’s impact in the world, and one of the most important missions is to democratize this technology. The cloud is this gigantic computing vehicle that delivers computing services to every single industry.

What have you learned so far?

We need to be much more human-centered. If you look at where we are in AI, I would say it’s the great triumph of pattern recognition. It is very task-focused, it lacks contextual awareness, and it lacks the kind of flexible learning that humans have. We also want to make technology that makes humans’ lives better, our world safer, our lives more productive and better. All this requires a layer of human-level communication and collaboration.

How can we make AI more human-centered?

There’s a great phrase, written in the ’70s: “the definition of today’s AI is a machine that can make a perfect chess move while the room is on fire.” It really speaks to the limitations of AI. In the next wave of AI research, if we want to make more helpful and useful machines, we’ve got to bring back the contextual understanding. We’ve got to bring knowledge abstraction and reasoning. These are all the most important steps.

At Stanford you created Visual Genome, a database of images that are extensively labeled so they can be used for AI systems. Is this interplay of vision and language necessary for the next leap forward?

Absolutely. Vision is a cornerstone of intelligence, and language understanding is a cornerstone of intelligence. What makes humans unique is that evolution gave us the most incredible and sophisticated vision system, motor system, and language system, and they all work together. Visual Genome is exactly the kind of project that’s pushing the boundaries of language understanding and visual understanding. And eventually we’re going to connect with the world of robotics as well.

You’ve talked about the need to have more women involved in AI. Why?

More jobs will be related to artificial intelligence, so we need a huge workforce, and we need a more inclusive base. That’s an economic argument. There are also tons of studies that have shown that when a diverse group of workers come together, the solutions they find in their work are more innovative and more creative. That drives innovation. But it’s also moral and ethical.

When you are making a technology this pervasive and this important for humanity, you want it to carry the values of the entire humanity, and serve the needs of the entire humanity. If the developers of this technology do not represent all walks of life, it is very likely that this will be a biased technology. I say this as a technologist, a researcher, and a mother. And we need to be speaking about this clearly and loudly.

Deep Dive

Artificial intelligence

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

OpenAI teases an amazing new generative video model called Sora

The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.

Google’s Gemini is now in everything. Here’s how you can try it out.

Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.

Google DeepMind’s new generative model makes Super Mario–like games from scratch

Genie learns how to control games by watching hours and hours of video. It could help train next-gen robots too.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.