AI has a culturally biased world view that Google has a plan to change
Google has launched an Inclusive Images Competition, an effort to expand the cultural fluency of image-recognition software. The task for entrants: reduce the bias in a computer vision system trained on a culturally biased image data set.
The context: Machines need to be trained on massive amounts of image data in order to recognize objects. Recent leaps in image recognition have coincided with the release of large, publicly available data sets, including ImageNet and Open Images.
The problem: The most popular data sets, however, are US- and Western-centric—simply because those Western images dominated the internet when the data sets were compiled. As a consequence, systems trained on them often fail to precisely describe scenes from other cultures and locales. Take wedding photos as an example. A standard image-recognition system, trained on open-source data sets, can recognize a bride in a white dress, reflecting the classic Western tradition. But it will fail to recognize a bride in a sari from an Indian ceremony.
The challenge: One way to mitigate this issue is to build more diverse and representative image data sets. While Google is pursuing this approach, the company also believes in advancing another way: by tweaking the machine-learning algorithms themselves to be more inclusive when learning from imperfect data.
The results: Hosted in partnership with the Neural Information Processing Systems (NeurIPS) conference, one of the largest annual gatherings for AI research, the competition received submissions from over 100 participants. Google Brain researcher Pallavi Baljekar noted at a conference talk on Sunday, December 2, that the first-year competition winners were able to make small steps toward more inclusive systems. But only one of the top five approaches successfully recognized an Indian bride. It’s clear that more work needs to be done.
Deep Dive
Artificial intelligence
Large language models can do jaw-dropping things. But nobody knows exactly why.
And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.
OpenAI teases an amazing new generative video model called Sora
The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.
Google’s Gemini is now in everything. Here’s how you can try it out.
Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.
Google DeepMind’s new generative model makes Super Mario–like games from scratch
Genie learns how to control games by watching hours and hours of video. It could help train next-gen robots too.
Stay connected
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.