Google is launching an effort that it hopes may turn humans and intelligent machines into productive work buddies.
Through a project called People + AI Research, or PAIR, Google will release tools designed to help make the inner workings of AI systems more transparent. The company is also launching several research initiatives aimed at finding new ways for humans and AI systems to collaborate effectively.
A great deal has been made recently of the potential for artificial intelligence to eliminate jobs and replace humans, but in truth the technology may often serve as a powerful new kind of tool capable of automating only part of a person’s work.
Figuring out how to get humans and AI algorithms to collaborate effectively could have big economic significance, and it could shape the way the workforce is educated. It might also play a role in tempering negative reactions to the growing role of automation in many settings (see “Is Technology Destroying Jobs?” and “Who Will Own the Robots?”).
Barbara Grosz, a professor at Harvard University who has long argued that computer scientists should design AI systems to complement rather than replace people, says this approach is needed because AI is still so limited in what it can do. Grosz adds that human and machine capabilities can add up to more than the sum of their parts. “AI systems, like all computers, need to be developed for the people who are going to use them,” she says.
Some sense of how AI algorithms may collaborate with humans can be seen in games. Chess or Go players can team up with computer programs to perform at an elevated level. This requires a new set of skills and a new approach to each game.
“The past few years have seen rapid advances in machine learning, with dramatic improvements in technical performance,” the researchers wrote in a blog post announcing the initiative. “But we believe AI can go much further—and be more useful to all of us—if we build systems with people in mind at the start of the process.”
PAIR is being led by Fernanda Viégas and Martin Wattenberg, researchers who specialize in developing visualizations that make complex information more comprehensible. Viégas and Wattenberg previously developed a series of tools to elucidate the behavior of complex and abstract machine-learning models. The opacity of such models is a growing concern for those hoping to deploy them in a wider range of settings (see “The Dark Secret at the Heart of AI”).
The PAIR project is today releasing two tools for visualizing the kind of large data sets used to train a machine-learning model to make useful predictions. These visualizations can help a data scientist identify anomalies in training data.
Grosz says it is especially challenging to build AI systems that work well with humans, partly because those systems may be complex and opaque, but also because interacting with a human in an intelligent way is a major challenge. “In every situation you need to be able to model a person’s mental state,” she says.
This new data poisoning tool lets artists fight back against generative AI
The tool, called Nightshade, messes up training data in ways that could cause serious damage to image-generating AI models.
Rogue superintelligence and merging with machines: Inside the mind of OpenAI’s chief scientist
An exclusive conversation with Ilya Sutskever on his fears for the future of AI and why they’ve made him change the focus of his life’s work.
Driving companywide efficiencies with AI
Advanced AI and ML capabilities revolutionize how administrative and operations tasks are done.
Generative AI deployment: Strategies for smooth scaling
Our global poll examines key decision points for putting AI to use in the enterprise.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.