Mathematician Cathy O’Neil is offering businesses a chance to test their algorithms for fairness.
Opening the black box: As artificial-intelligence systems get more advanced, the logic paths they follow can be difficult or even impossible to understand, creating a so-called “black box.” As these algorithms come to control increasingly important parts of our lives, like whether we get a job or a loan, it is crucial to understand their biases and decisions.
The solution: O’Neil, who wrote the book Weapons of Math Destruction, has started O’Neil Risk Consulting and Algorithmic Auditing to perform third-party audits on algorithms. It examines everything from the people who programmed the software to the training data to the output, flagging any bias in the process.
Why would businesses choose to do this? Companies aren’t knocking down her door yet (she has only six clients). But they should be: not only is it in society’s best interest, it’s also good marketing. Getting a your algorithm certified for fairness can prove to your customers that your service is equitable, effective, and trustworthy.
How conservative Facebook groups are changing what books children read in school
Parents are gathering online to review books and lobby schools to ban them, often on the basis of sexual content.
Why can’t tech fix its gender problem?
A new generation of tech activists, organizers, and whistleblowers, most of whom are female, non-white, gender-diverse, or queer, may finally bring change.
How the idea of a “transgender contagion” went viral—and caused untold harm
A single paper on the notion that gender dysphoria can spread among young people helped galvanize an anti-trans movement.
The world is moving closer to a new cold war fought with authoritarian tech
At the Shanghai Cooperation Organization summit, Iran, Turkey, and Myanmar promised tighter trade relationships with Russia and China.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.