Skip to Content
77 Mass Ave

Cartography on the Fly

Wearable mapping device could help first responders navigate buildings
January 2, 2013

MIT researchers have developed a tool that they hope will be useful for emergency responders: a wearable sensor system that automatically creates a digital map of the wearer’s environment.

The prototype sensor includes a stripped-down camera mounted above a laser rangefinder.

In experiments, a graduate student wearing the system wandered the MIT halls, and the sensors wirelessly relayed data to a laptop in a distant conference room. Observers in the conference room tracked the student’s progress on a map that sprang into being as he moved.

“The operational scenario that was envisioned for this was a hazmat situation where people are suited up with the full suit, and they go in and explore an environment,” says Maurice Fallon, a research scientist in MIT’s Computer Science and Artificial Intelligence Laboratory.

One of the system’s sensors is a laser rangefinder, which sweeps light in a 270° arc and measures the time it takes to return. Automatic mapping systems mounted on robots have used rangefinders to build very accurate maps, but a walking human jostles the rangefinder more than a rolling robot does. Similarly, sensors in a robot’s wheels can provide data about its physical orientation and the distances it covers, which is missing with humans. And because emergency workers might have to move between floors of a building, the new system also has to recognize changes in altitude.

In addition to the rangefinder, the researchers equipped their sensor platform with accelerometers, gyroscopes, and a camera. The gyroscopes can infer when the rangefinder is tilted—information the mapping algorithms can use in interpreting its readings—and the accelerometers provide information about the wearer’s velocity and about changes in altitude.

Should there be any discrepancies in the data from the sensors, the camera serves as the adjudicator. Every few meters, the camera takes a snapshot of its surroundings, and software extracts about 200 visual features from the image—patterns, contours, and inferred three-dimensional shapes. During different passes through the same region, the readings from the other sensors may diverge slightly, but image data can confirm that the region has already been mapped.

The system prototype consists of a handful of devices attached to a sheet of hard plastic about the size of an iPad, which is worn on the chest like a backward backpack. The only sensor that can’t be made much smaller is the rangefinder, so in principle, the whole system could be shrunk to about the size of a coffee mug.

Keep Reading

Most Popular

A Roomba recorded a woman on the toilet. How did screenshots end up on Facebook?

Robot vacuum companies say your images are safe, but a sprawling global supply chain for data from our devices creates risk.

A startup says it’s begun releasing particles into the atmosphere, in an effort to tweak the climate

Make Sunsets is already attempting to earn revenue for geoengineering, a move likely to provoke widespread criticism.

10 Breakthrough Technologies 2023

Every year, we pick the 10 technologies that matter the most right now. We look for advances that will have a big impact on our lives and break down why they matter.

These exclusive satellite images show that Saudi Arabia’s sci-fi megacity is well underway

Weirdly, any recent work on The Line doesn’t show up on Google Maps. But we got the images anyway.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.