The presidential campaign made clear that chauvinist attitudes toward women remain stubbornly fixed in some parts of society. It turns out we’re inadvertently teaching artificial-intelligence systems to be sexist, too.
New research shows that subtle gender bias is entrenched in the data sets used to teach language skills to AI programs. As these systems become more capable and widespread, their sexist point of view could have negative consequences—in job searches, for instance.
The problem results from the way machines are being taught to read and talk. Computer scientists are feeding them huge quantities of written or spoken language, and letting them draw connections between words and phrases.
The resulting data sets, known as word embeddings, are widely used to train AI systems that handle language—including chatbots, translation systems, image-captioning programs, and recommendation algorithms. Word embeddings represent the relationships between words as mathematical values. This makes it possible for a machine to perceive semantic connections between, say, “king” and “queen” and understand that the relationship between the two words is similar to that between “man” and “woman.” But researchers from Boston University and Microsoft Research New England also found that the data sets considered the word “programmer” closer to the word “man” than “woman,” and that the most similar word for “woman” is “homemaker.”
James Zou, an assistant professor at Stanford University who conducted the research while at Microsoft, says this could have a range of unintended consequences. “We are still trying to understand the full impact that comes from the many AI systems using these biased embeddings,” Zou says.
Zou and colleagues have conducted some simple experiments that show how this gender bias might manifest itself. When they wrote a program designed to read Web pages and rank their relevance, they found the system would rank information about female programmers as less relevant than that about their male counterparts.
The researchers also developed a way to remove gender bias from embeddings by adjusting the mathematical relationship between gender-neutral words like “programmer” and gendered words such as “man” and “woman.”
But not everyone believes gender bias should be eliminated from the data sets. Arvind Narayanan, an assistant professor of computer science at Princeton, has also analyzed word embedding and found gender, racial, and other prejudices. But Narayanan cautions against removing bias automatically, arguing that it could skew a computer’s representation of the real world and make it less adept at making predictions or analyzing data.
“We should think of these not as a bug but a feature,” Narayanan says. “It really depends on the application. What constitutes a terrible bias or prejudice in one application might actually end up being exactly the meaning you want to get out of the data in another application.”
Several word embedding data sets exist, including Word2Vec, created by researchers at Google, and GloVe, developed at Stanford University. Google declined to comment on research showing gender bias in Word2Vec, but the company is clearly conscious of the challenge. A recent blog post describes a technical approach to removing bias from decision-making AI systems without affecting their usefulness.
Biased AI systems could exacerbate the unfairness that already exists, says Barbara Grosz, a professor at Harvard University. “When you are in a society that is evolving in certain ways, then you are actually trying to change the future to be not like the past,” says Grosz, who cowrote a report called AI 100, a project from Stanford University aimed at understanding the potential dangers of AI (see “AI Wants to Be Your Bro, Not Your Foe”). “And to the extent that we rely on algorithms that do that kind of predicting,” Grosz says, “there’s an ethical question about whether we’re inhibiting the very evolution that we want.”
Grosz concedes that there may be situations when it doesn’t make sense to remove bias from a data set. “It’s not that you can avoid all these kinds of bias, but we need to be mindful in our design, and we need to be mindful about what we claim about our programs and their results,” Grosz adds. “For many of these ethical questions, there isn’t a single right answer.”
A Roomba recorded a woman on the toilet. How did screenshots end up on Facebook?
Robot vacuum companies say your images are safe, but a sprawling global supply chain for data from our devices creates risk.
The viral AI avatar app Lensa undressed me—without my consent
My avatars were cartoonishly pornified, while my male colleagues got to be astronauts, explorers, and inventors.
Roomba testers feel misled after intimate images ended up on Facebook
An MIT Technology Review investigation recently revealed how images of a minor and a tester on the toilet ended up on social media. iRobot said it had consent to collect this kind of data from inside homes—but participants say otherwise.
How to spot AI-generated text
The internet is increasingly awash with text written by AI software. We need new tools to detect it.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.