Skip to Content
Artificial intelligence

A robot dog has learned to run faster with machine learning

January 16, 2019

Reinforcement learning has helped a four-legged bot move a bit like a real animal, without having to be taught how to make each step.

The news: Roboticists want their creations to mimic animals because animals invariably move in the most energy-efficient way. But the eerily lifelike movement of robots like Boston Dynamics’ Spotmini is usually coded by hand. Now researchers have combined simulation with a technique called reinforcement learning to teach a dog-like robot called “ANYmal” to run faster and recover from falls. Crucially, it did so without any manual intervention. 

Train at super-speed: The simulation let the team run training sessions 1,000 faster than real time on more than 2,000 ANYmals simultaneously. After the simulated training was transferred to a real robot, it was able to exceed its previous top speed by 25% and flip over after falls, the EFH Zurich team explains in a new paper published in Science Robotics today. It’s still pretty limited (as you’ll see above), but it’s a step (ho ho) in the right direction.

Uses: Besides keeping Jeff Bezos company? To be honest, there aren’t that many compelling practical uses at the moment, but the researchers say these sorts of four-legged robots could one day be used to inspect underground tunnels or carry heavy loads on construction sites.

Deep Dive

Artificial intelligence

The inside story of how ChatGPT was built from the people who made it

Exclusive conversations that take us behind the scenes of a cultural phenomenon.

AI is dreaming up drugs that no one has ever seen. Now we’ve got to see if they work.

AI automation throughout the drug development pipeline is opening up the possibility of faster, cheaper pharmaceuticals.

GPT-4 is bigger and better than ChatGPT—but OpenAI won’t say why

We got a first look at the much-anticipated big new language model from OpenAI. But this time how it works is even more deeply under wraps.

The original startup behind Stable Diffusion has launched a generative AI for video

Runway’s new model, called Gen-1, can change the visual style of existing videos and movies.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.