Select your localized edition:

Close ×

More Ways to Connect

Discover one of our 28 local entrepreneurial communities »

Be the first to know as we launch in new countries and markets around the globe.

Interested in bringing MIT Technology Review to your local market?

MIT Technology ReviewMIT Technology Review - logo

 

Unsupported browser: Your browser does not meet modern web standards. See how it scores »

{ action.text }

Moving image: The image on the right was reconstructed from an fMRI image captured while the subject watched the video clip on the left.

Scientists are a step closer to constructing a digital version of the human visual system. Researchers at the University of California, Berkeley, have developed an algorithm that can be applied to functional magnetic resonance imaging (fMRI) imagery to show a moving image a person is seeing.

Neuroscientists have been using fMRI to study the human visual system for years, which involves measuring changes in blood oxygen levels in the brain. This works fine for studying how we see static images, but it falls short when it comes to moving imagery. Individual neuronal activity occurs over a much faster time scale, so a few years ago the researchers behind the current study set out to devise a computer model to measure this instead. The study shows that this new approach is not only successful but remarkably accurate.

The study, which appears in Current Biology this week, marks the first time that anyone has used brain imaging to determine what moving images a person is seeing. It could help researchers model the human visual system on a computer, and it raises the tantalizing prospect of one day being able to use the model to reconstruct other types of dynamic imagery, such as dreams and memories.

The researchers involved in the study watched hours of movie previews while lying in an fMRI machine. Next they painstakingly deconstructed the data so that they had a specific activation pattern for each second of footage. They ran that data through several different filters to infer what was happening at the neuronal level. “Once you do this, you have a complete model that links the plumbing of the blood flow that you do see with fMRI to the neuronal activity that you don’t see,” says Jack Gallant, who coauthored the study with colleague Shinji Nishimoto.

Next, the researchers compiled a library of 18 million YouTube video clips, chosen at random, to test their model objectively. Previous studies have shown that fMRI can be used to determine static images a subject is looking at, but the new computer model offered the possibility of reconstructing images that had direction of movement as well as shape. “No one has tried to model dynamic vision with this level of detail before,” says Jim Haxby, a neuroimaging expert at Dartmouth College who was not involved in the study.

The researchers used the YouTube library to simulate what would happen on the fMRI images when they watched a new set of movie trailers. The results of the simulations and fMRI scans were close to identical. “Usually you only get that kind of accuracy in physics, not neuroscience,” says Benjamin Singer, an fMRI researcher at Princeton University who was not involved with the study. “It’s a tour de force that brings together decades of work.”

There are two main caveats to the study. The researchers used fMRI data from only one area of the visual system—the V1 area, also known as the primary visual cortex. And the models were customized to each subject. Trying to design a model that would work for everyone would have been too difficult, says Gallant, although he suspects a more generalized model could be developed in the future.

The ultimate goal of this research is to create a computational version of the human brain that “sees” the world as we do. The study also demonstrates an unexpected use for an existing technology. “Everyone always thought it was impossible to recover dynamic brain activity with fMRI,” says Gallant.

5 comments. Share your thoughts »

Credit: Shinji Nishimoto, An T. Vu, Thomas Naselaris, Yuval Benjamini, Bin Yu & Jack L. Gallant

Tagged: Biomedicine, brain, neuroscience, brain imaging, visual system

Reprints and Permissions | Send feedback to the editor

From the Archives

Close

Introducing MIT Technology Review Insider.

Already a Magazine subscriber?

You're automatically an Insider. It's easy to activate or upgrade your account.

Activate Your Account

Become an Insider

It's the new way to subscribe. Get even more of the tech news, research, and discoveries you crave.

Sign Up

Learn More

Find out why MIT Technology Review Insider is for you and explore your options.

Show Me