Skip to Content

Scientists Induce Out-of-Body Experience

Are better video games to follow?
August 24, 2007

Using virtual reality, scientists in Switzerland and Sweden tricked the human brain into believing that the body was at a different location, mimicking an out-of-body experience. The technique could aid in the development of virtual-reality games in which the players actually feel as though they are inside the game. The results were published today in the journal Science.

Often interpreted as a spiritual experience, people have reported out-of-body sensations linked with epilepsy, drug use, and traumatic events like surgery or car accidents. The new experiments support the idea that these experiences are likely induced by a mismatch in sensory information somewhere in the brain.

An article in the Guardian explains the experiment:

In [one] experiment participants wore goggles containing a video screen for each eye. Each screen was fed images from a separate camera behind the participant and, because the two images were combined by the brain into a single image, they saw a 3D image of their own back.

Dr. Ehrsson then moved a plastic rod towards a location just below the cameras while the participant’s real chest was simultaneously touched in the corresponding position. The participants reported feeling that they were located back where the cameras had been placed, watching a body that belonged to someone else.

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

OpenAI teases an amazing new generative video model called Sora

The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.

Google’s Gemini is now in everything. Here’s how you can try it out.

Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.

This baby with a head camera helped teach an AI how kids learn language

A neural network trained on the experiences of a single young child managed to learn one of the core components of language: how to match words to the objects they represent.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.