Skip to Content
Artificial intelligence

A Camera That Sees like the Human Eye

IBM’s brainlike computer architecture paves the way for a new kind of artificial vision.

The retina is an enormously powerful tool. It sorts through massive amounts of data while operating on only a fraction of the power that a conventional digital camera and computer would require to do the same task.

Dynamic vision: The camera’s strength is in capturing movement, like the milk drops seen here.

Now, engineers at a company called iniLabs in Switzerland are applying lessons from biology in an effort to build a more efficient digital camera inspired by the human retina.

Like the individual neurons in our eyes, the new camera—named the Dynamic Vision Sensor (DVS)—responds only to changes in a given scene. This approach eliminates large swaths of redundant data and could be useful for many fields, including surveillance, robotics, and microscopy.

“Your eye and my eye are digital cameras too. [They’re] just a different kind of digital camera,” says Tobi Delbruck, the chief scientific officer at iniLabs. “We had machine vision that was as good as possible with existing architecture and hardware. But compared to biology, machine vision is pathetically poor.”

An ordinary camera will take in everything it sees, storing the information to be processed later. This uses up a lot of power and a lot of space. Neurons in the eye, however, fire only when they sense a change—such as when a particular part of a scene gets brighter or dimmer. The DVS mimics that selectivity, transmitting information only in response to a shift in the scene. That takes less power and leaves less information to be processed.

Artificial retina: The Dynamic Vision Sensor (DVS) responds only to changes in the scene, eliminating large swaths of irrelevant data.

This feature could be especially useful for recording scenes that do not change often. For example, when sleep researchers videotape their subjects, they are later forced to comb through hours of uneventful footage. With a sensor like the DVS, important, active portions of the data are automatically highlighted.

The pixels in the DVS also mimic the way an individual eye neuron will calibrate itself to a particular location: that cell and those responsible for another area will respond to incoming data in slightly different ways, so one neuron might be very sensitive to input while another takes more stimulation to fire. Similarly, each pixel of the DVS adjusts its own exposure. This allows the DVS to handle uneven lighting conditions, though it also requires enormous pixels that are 10 times the size of those in a modern cell-phone camera.

The DVS is built to work with IBM’s new TrueNorth computer architecture (see “IBM Researchers Show Blueprints for Brainlike Computing”). TrueNorth is a programming approach that mimics biology—information is stored, processed, and shared in a network of “neuromorphic” computer chips, inspired by the neural networks in the brain.

“What we’re talking about—the cameras sending information when something changes—is actually a very central theme to how the brain works, or at least how neuroscientists think it works,” says Nabil Imam, a computer scientist at Cornell University, who is part of the Cornell team that helped IBM develop its neuromorphic chips. “We’re capturing brain features at a high level.”

By combining their camera with TrueNorth architecture, Delbruck and his team believe, they will achieve a device that’s better at handling dynamic, real-time problems.

The DVS is available for purchase for about $2,700 and has been used in several research projects, including one that recorded traffic and another that involved tracking particles in a fluid. The team plans to continue improving the device. The next goals are to add color sensitivity and to enlarge the camera’s retina from its current resolution of 240x180.

Deep Dive

Artificial intelligence

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

Google DeepMind’s new generative model makes Super Mario–like games from scratch

Genie learns how to control games by watching hours and hours of video. It could help train next-gen robots too.

What’s next for generative video

OpenAI's Sora has raised the bar for AI moviemaking. Here are four things to bear in mind as we wrap our heads around what's coming.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.