Google has launched an Inclusive Images Competition, an effort to expand the cultural fluency of image-recognition software. The task for entrants: reduce the bias in a computer vision system trained on a culturally biased image data set.
The context: Machines need to be trained on massive amounts of image data in order to recognize objects. Recent leaps in image recognition have coincided with the release of large, publicly available data sets, including ImageNet and Open Images.
The problem: The most popular data sets, however, are US- and Western-centric—simply because those Western images dominated the internet when the data sets were compiled. As a consequence, systems trained on them often fail to precisely describe scenes from other cultures and locales. Take wedding photos as an example. A standard image-recognition system, trained on open-source data sets, can recognize a bride in a white dress, reflecting the classic Western tradition. But it will fail to recognize a bride in a sari from an Indian ceremony.
The challenge: One way to mitigate this issue is to build more diverse and representative image data sets. While Google is pursuing this approach, the company also believes in advancing another way: by tweaking the machine-learning algorithms themselves to be more inclusive when learning from imperfect data.
The results: Hosted in partnership with the Neural Information Processing Systems (NeurIPS) conference, one of the largest annual gatherings for AI research, the competition received submissions from over 100 participants. Google Brain researcher Pallavi Baljekar noted at a conference talk on Sunday, December 2, that the first-year competition winners were able to make small steps toward more inclusive systems. But only one of the top five approaches successfully recognized an Indian bride. It’s clear that more work needs to be done.
Why Meta’s latest large language model survived only three days online
Galactica was supposed to help scientists. Instead, it mindlessly spat out biased and incorrect nonsense.
A bot that watched 70,000 hours of Minecraft could unlock AI’s next big thing
Online videos are a vast and untapped source of training data—and OpenAI says it has a new way to use it.
Google’s new AI can hear a snippet of song—and then keep on playing
The technique, called AudioLM, generates naturalistic sounds without the need for human annotation.
Responsible AI has a burnout problem
Companies say they want ethical AI. But those working in the field say that ambition comes at their expense.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.