Skip to Content

CES 2014: Intel’s 3-D Camera Heads to Laptops and Tablets

Controlled with gestures, these depth-sensing devices will appear in the second half of 2014.
January 6, 2014

A combined 2-D and 3-D camera from Intel will be built into laptops and tablets from a range of manufacturers, the company announced at the International Consumer Electronics Show (CES) on Monday. The camera allows a device to be controlled with arm, hand, and finger gestures, and is also intended to allow software to capture and understand the world around it, including people’s facial expressions.

kinect laptops
Depth sensing: This Dell laptop has a 3-D sensing camera embedded above its screen, allowing forms of gesture control previously found only in games consoles.

“We see and touch in 3-D, so why do we have to use computers in 2-D?” asked Mooly Eden, general manager of Intel’s Perceptual Computing Group, before unveiling the new depth-sensing technology, called the Intel RealSense 3-D camera.

He said that Intel intends the technology to become ubiquitous, similar to how conventional webcams have become a standard feature of even the cheapest PCs. “This is something that we will be able to embed in many, many systems to drive the cost down,” said Eden, brandishing a slim naked RealSense circuit board, approximately five inches long and as thin as two quarters.

Intel showed seven different laptops and tablets from Dell, Lenovo, and Asus with the integrated depth camera. Such devices are slated to hit the market in the second half of 2014.

Eden introduced several demonstrations of different applications for the new camera, including gaming, photography, and 3-D scanning. In one demonstration a person waved his hand in front of a Windows computer to move the cursor around, closing his hand to grab a virtual page, moving his hand to scroll the page, and tapping in space to perform the equivalent of a mouse click.

In another, the depth sensing enhanced the tablet’s camera app. While composing a photo, a few taps on the screen were enough to isolate the background and foreground, then apply color filters.

 A similar function was demonstrated in a special version of Microsoft’s Skype software: the depth-camera isolated a video caller from his surroundings and set him against various backgrounds.

Intel further announced a partnership with 3-D Systems, which makes 3-D printers, to develop software so the depth-sensing cameras can scan objects and then edit or print a digital double. Several gesture-controlled games were also shown.

Eden said that enabling computers to sense in 3-D should eventually allow software to better understand human behavior: “The computer will be able to track with the 3-D camera exactly my facial expression and be able to respond appropriately.” Eden said that ability could be combined with voice-operated virtual assistant software so people could speak to their devices more naturally.

Like the depth-sensing technology for Microsoft’s Kinect gaming system, Intel’s has two cameras inside – a conventional camera, and one that senses infrared – as well as an infrared light. (The device can infer depth by detecting infrared light that has bounced back from a scene.)

But Intel’s technology will be much smaller than Microsoft’s, enabling what will be the first generation of laptops and tablets with depth-sensing technology built in. Microsoft has for several years offered a version of its Kinect camera aimed at desktop computers (see “Microsoft’s Plan to Bring About an Era of Gesture Control”), and other companies have launched depth-camera accessories (see “Depth Sensing Cameras Head to Mobile Devices”), but they have had minimal success and were not compact enough to be integrated into PCs or tablets.

PrimeSense, which developed hardware used in Microsoft’s Kinect, previously said it was working on a compact version of its technology (see “PC Makers Bet on Gaze, Gesture, Touch and Voice”). However, it hasn’t publically shown one small enough to be embedded into devices. PrimeSense was recently acquired by Apple, which is notoriously secretive about technology under development.

Keep Reading

Most Popular

Geoffrey Hinton tells us why he’s now scared of the tech he helped build

“I have suddenly switched my views on whether these things are going to be more intelligent than us.”

Meet the people who use Notion to plan their whole lives

The workplace tool’s appeal extends far beyond organizing work projects. Many users find it’s just as useful for managing their free time.

Learning to code isn’t enough

Historically, learn-to-code efforts have provided opportunities for the few, but new efforts are aiming to be inclusive.

Deep learning pioneer Geoffrey Hinton has quit Google

Hinton will be speaking at EmTech Digital on Wednesday.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.