“Spoon” is to “woman” as “tennis racket” is to “man.” At least, that’s according to AI algorithms trained on two of the more common collections of thousands of images that are usually used by researchers to help machines understand the real world.
Wired reports that a team of researchers from the University of Washington studied gender predictions made by computer vision algorithms. What’s particularly interesting is that biases present in image sets—such as the fact that women were 33 percent more likely to appear in a photograph related to cooking in one data batch it studied—are amplified in the connections that the AI’s neural network makes. So, trained on that data set, an AI was 68 percent more likely to predict a woman was cooking, and did so even when an image was clearly of a balding man in a kitchen.
It’s by no means the first time that an AI has been observed to pick up gender biases from training data. Last year, we reported that researchers from Boston University and Microsoft Research found that an AI trained on archives of text learned to associate the word “programmer” with the word “man,” and “homemaker” with the word “woman.”
But it’s especially troubling that biases inherent in data sets may end up being amplified, rather than merely replicated, by the AIs that are trained on them. Sometimes it might cause offense, if, for instance, an AI is being used to target advertising based on images you upload to a social network. But in other applications—such as, say, the controversial practice of predicting criminality from a person’s face—baked-in prejudice could be downright harmful.