Give Your Dashboard the Finger
An experimental gestural interface for cars lets drivers control any part of their dashboard without taking their hands off the wheel.
Christian Müller, a researcher at the German Research Center for Artificial Intelligence (DFKI) in Saarbrücken, who co-developed the new system, says the idea is to enable drivers to adjust everything from the volume of the car stereo to the climate-control settings while keeping their hands on the wheel and eyes on the road.
The prototype interface uses several sensors to detect the movement of a driver’s right index finger as it disrupts an electric field. It is based on the same principle as the theremin, a musical instrument that is played without being touched. Electromagnetic sensors located in the dashboard detect finger movements, providing the driver is holding the wheel in the recommended ten-to-two position and is driving straight. By detecting the different shapes the driver’s finger draws in the air, the system can detect and interpret a wide range of commands, says Müller.
Müller tested it with DFKI colleague Christoph Endres and Tim Schwartz at the University of Saarland, also in Germany. The researchers used a car simulator. “In our prototype we glued these antennae on the dashboard right behind the steering wheel,” he says.
Six people were asked to try the system, dubbed Geremin. It was able with an accuracy of 86 percent to distinguish 10 different gestures, including moving the finger up or down, left or right, or tracing out circles, triangles, and squares. The work will be presented next week at the International Conference on Intelligent User Interfaces in Palo Alto, California.
Some cars do already have buttons and controls built into the steering wheel or attached to the steering column, says Müller, but buttons are more limiting than gestures. “While you can certainly have designated buttons for a few selected functions or applications on the steering wheel, the number of gestures can be extended,” he says.
The system should be much cheaper than installing cameras to monitor drivers’ movements, as some car manufacturers are now doing. The cost of each sensor is about 50 cents.
“It’s an interesting idea. It could be useful,” says Paul Green, a research professor of the Driver Interface Group at the University of Michigan’s Transport Research Institute. But while it may help keep a driver’s hands on the wheel, recognition errors could prove to be a different kind of distraction, he says. “Also, if each manufacturer has a different set of gestures, then you have a real problem,” he says.
The German researchers hope to extend the gesture set significantly. “We will combine this with speech recognition in order to allow people to dictate text messages in the car,” says Müller.
Nonetheless, there are skeptics. Andrew Howard, head of road safety at the U.K.’s Automobile Association, says the prospect of enabling drivers to text is a “frightening” one. “We shouldn’t be encouraging people to do anything other than driving while driving,” he says.
Dictating texts is “not something we would want to encourage,” says Green, adding that “I think we’re going to find the throughput would be faster using speech-to-text.”
Geoffrey Hinton tells us why he’s now scared of the tech he helped build
“I have suddenly switched my views on whether these things are going to be more intelligent than us.”
ChatGPT is going to change education, not destroy it
The narrative around cheating students doesn’t tell the whole story. Meet the teachers who think generative AI could actually make learning better.
Meet the people who use Notion to plan their whole lives
The workplace tool’s appeal extends far beyond organizing work projects. Many users find it’s just as useful for managing their free time.
Learning to code isn’t enough
Historically, learn-to-code efforts have provided opportunities for the few, but new efforts are aiming to be inclusive.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.