Skip to Content

Meeting the Future

Internet: Tele-immersion makes virtual conferencing more real
September 1, 2000

Video teleconferencing is often touted for its potential to promote better communications and curb expensive travel. Yet technology for such virtual face-to-face meetings has yet to catch on as a routine business tool. Among its perceived failings: the inability of participants to make eye contact (due to camera placement limitations), the need for a dedicated “dry room” away from office or lab floors, and a lack of shared workspace for collaborative brainstorming.

Given the expected development of far greater bandwidth than is available with current data lines, what is the next step toward more realistic virtual meetings? The answer is known as tele-immersion, a conceptual hybrid of virtual reality and Star Trek’s Holodeck. One of the principal applications areas for Internet2 (a research project involving 170 academic institutions and 50 corporations to develop tomorrow’s faster Internet), tele-immersion visually replicates, in real time and in three dimensions, slabs of space surrounding remote participants in a cybermeeting. The result is a shared, simulated environment that makes it appear as if everyone is in the same room.

In May, researchers at the University of North Carolina (UNC) at Chapel Hill, the University of Pennsylvania in Philadelphia, and Advanced Network and Services in Armonk, N.Y., demonstrated for the first time the building blocks of this meeting room of the future. A participant sees two projected “windows” of life-sized colleagues, each hundreds of miles away from the site and each other. Lean forward and the foreground figure and background bookshelf in a window shift slightly in three dimensions, as if right there.

The 3-D view and precise position tracking, however, come at a cost; awkward goggles and a silvery head tracker perched on a user’s head are required. Less evident is an array of seven standard video cameras and two special ones that capture distance information by reading light patterns subliminally projected into each participant’s environment.

In addition to the need for bulky and costly equipment, the demo suffers from video glitches (just like videoconferencing). Still, it provides a glimpse of what lies beyond today’s videoconferencing. And it provides vindication for the more than two years of research by a collaboration of computer scientists from UNC, Penn, Brown University, Columbia University, the University of Southern California, Carnegie Mellon University and the University of Illinois. “It’s a significant accomplishment,” says Jaron Lanier, chief scientist for the project. “We demonstrated viewpoint-independent real-time scene sensing and reconstruction. And we got rid of the dry room.’”

By the end of the decade, when next-generation broadband is in place, immersive conferencing could be rigged from any office or lab. Still, there are challenges. The researchers want to better integrate the real and virtual worlds and provide overlapping workspaces for shared whiteboards and 3-D modeling. In the demo, the edges of the virtual cubicles do not conjoin; ultimately tele-immersion is meant to be seamless. Lanier, who helped invent and popularize virtual reality in the 1980s and 1990s, says the scientists are also working on autostereo screens (for 3-D without glasses) and advances in haptics (for full-hand tactile simulations).

If it all works, you’ll finally be able to really reach out and touch someone.

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

How scientists traced a mysterious covid case back to six toilets

When wastewater surveillance turns into a hunt for a single infected individual, the ethics get tricky.

The problem with plug-in hybrids? Their drivers.

Plug-in hybrids are often sold as a transition to EVs, but new data from Europe shows we’re still underestimating the emissions they produce.

Google DeepMind’s new generative model makes Super Mario–like games from scratch

Genie learns how to control games by watching hours and hours of video. It could help train next-gen robots too.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.