Inside a modest-looking office building in Tokyo lives an unusually clever industrial robot made by the Japanese company Fanuc. Give the robot a task, like picking widgets out of one box and putting them into another container, and it will spend the night figuring out how to do it. Come morning, the machine should have mastered the job as well as if it had been programmed by an expert.
Industrial robots are capable of extreme precision and speed, but they normally need to be programmed very carefully in order to do something like grasp an object. This is difficult and time-consuming, and it means that such robots can usually work only in tightly controlled environments.
Fanuc’s robot uses a technique known as deep reinforcement learning to train itself, over time, how to learn a new task. It tries picking up objects while capturing video footage of the process. Each time it succeeds or fails, it remembers how the object looked, knowledge that is used to refine a deep learning model, or a large neural network, that controls its action. Deep learning has proved to be a powerful approach in pattern recognition over the past few years.
“After eight hours or so it gets to 90 percent accuracy or above, which is almost the same as if an expert were to program it,” explains Shohei Hido, chief research officer at Preferred Networks, a Tokyo-based company specializing in machine learning. “It works overnight; the next morning it is tuned.”
Robotics researchers are testing reinforcement learning as a way to simplify and speed up the programming of robots that do factory work. Earlier this month, Google published details of its own research on using reinforcement learning to teach robots how to grasp objects.
The Fanuc robot was programmed by Preferred Networks. Fanuc, the world’s largest maker of industrial robots, invested $7.3 million in Preferred Networks in August last year. The companies demonstrated the learning robot at the International Robot Exhibition in Tokyo last December.
One of the big potential benefits of the learning approach, Hido says, is that it can be accelerated if several robots work in parallel and then share what they have learned. So eight robots working together for one hour can perform the same learning as one machine going for eight hours. “Our project is oriented to distributed learning,” Hido says. “You can imagine hundreds of factory robots sharing information.”
This form of distributed learning, sometimes called “cloud robotics,” is shaping up to be a big trend both in research and industry (see “10 Breakthrough Technologies 2016: Robots That Teach Each Other”).
“Fanuc is well placed to think about this,” says Ken Goldberg, a professor of robotics at the University of California, Berkeley, because it installs so many machines in factories around the world. He adds that cloud robotics will most likely reshape the way robots are used in the coming years.
Goldberg and colleagues (including several researchers at Google) are in fact taking this a step further by teaching robots how certain movements may be used to grasp not just specific objects but certain shapes. A paper on this work will appear at the IEEE International Conference on Robotics and Automation in May.
However, Goldberg notes, applying machine learning to robotics is challenging because controlling behavior is more complex than, say, recognizing objects in images. “Deep learning has made enormous progress in pattern recognition,” Goldberg says. “The challenge with robotics is that you’re doing something beyond that. You need to be able to generate the appropriate actions for a huge range of inputs.”
Fanuc may not be the only company developing robots that use machine learning. In 2014, the Swiss robot maker ABB invested in another AI startup called Vicarious. The fruits of that investment have yet to appear, however.
A Roomba recorded a woman on the toilet. How did screenshots end up on Facebook?
Robot vacuum companies say your images are safe, but a sprawling global supply chain for data from our devices creates risk.
The viral AI avatar app Lensa undressed me—without my consent
My avatars were cartoonishly pornified, while my male colleagues got to be astronauts, explorers, and inventors.
Roomba testers feel misled after intimate images ended up on Facebook
An MIT Technology Review investigation recently revealed how images of a minor and a tester on the toilet ended up on social media. iRobot said it had consent to collect this kind of data from inside homes—but participants say otherwise.
How to spot AI-generated text
The internet is increasingly awash with text written by AI software. We need new tools to detect it.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.