Skip to Content

Buckle up for the Vehicular Zombie Apocalypse

Autonomous technology is being developed at a remarkable rate. This could raise new fears about cars malfunctioning.
March 11, 2013

For self-driving cars, the road to commercialization could contain a few potholes.

Last week Japanese auto-maker Subaru announced that it would recall 50,000 “zombie cars” because several models (Imprezas, Outbacks, and Crosstours) were found to start themselves without human intervention (perhaps before driving off in search of fresh human brains, the report doesn’t say).  

The story caught my eye because I’ve been working on an article about autonomous driving. While these cars aren’t self-driving, and the problem isn’t particularly serious (unless the affected car is left running in a confined space), the incident illustrates how novel forms of automation can bring with them surprising—and potentially alarming—new ways for cars to go wrong. The problem with these Subarus reportedly concerns their remote starting system. If the owner drops the car’s key fob, this apparently can send a message to the vehicle’s remote starter that causes the engine to switch on and off for up to 15 minutes.

The case is important because even more advanced autonomous technology is being developed at a remarkable rate, with sophisticated features already showing up in many commercial vehicles: cruise-control that tracks the speed of other cars, automated parallel parking, and so on. While carmakers are introducing the technology in a careful and responsible way, such technology could theoretically malfunction in novel ways.

Even if the technology works perfectly, public uncertainty over more automation could be a problem. You may recall the incidents of “sudden unintended acceleration” that led to a massive recall of several models of Toyotas between 2009 and 2011. Although many incidents turned out to be due to problems with floor mats interferring with accelerator pedals, the fact that the car’s acceleration system was controlled electronically triggered lingering speculation that the cars electronics were somehow going bananas.

In other words, as cars increasingly become capable of driving themselves, carmakers will need to find ways to reassure drivers that they don’t need to worry about a zombie invasion.

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

OpenAI teases an amazing new generative video model called Sora

The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.

Google’s Gemini is now in everything. Here’s how you can try it out.

Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.

This baby with a head camera helped teach an AI how kids learn language

A neural network trained on the experiences of a single young child managed to learn one of the core components of language: how to match words to the objects they represent.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.