Skip to Content
Artificial intelligence

Amazon’s Echo Look Rates Your Outfits and Slurps Up Revealing Data

The company’s latest smart assistant features a camera to help you choose what to wear—and photograph the inside of your house while it’s at it.

Alexa just got eyes.

The e-tailer’s latest addition to its cadre of smart assistants is called the Echo Look. Much like the Echos before it, you can ask it to play music, read news headlines, or recite your forthcoming schedule. But unlike its brethren, the $200 device also sports cameras that will snap photographs or videos of you when you ask it to. The idea: to act as a kind of smart mirror for the fashion conscious. It even allows you to use use depth perception to blur the background and, in Amazon’s words, “make sure your outfits pop.”

You can use it to see how you actually look from the side or behind, or maybe watch a short video clip to … er, see what your outfit looks like on the dance floor, we guess? There’s also the option to use Amazon’s Style Check feature—already available for Prime members in the Amazon app—which allows you to offer up two images of different outfits to Amazon’s fashion AI and get advice on which looks best.

How that will work if you’re a true fashionista pushing the envelope of style or a dullard whose largest clothing choice is between a pale blue or off-white button down? We’re not so sure. But what is certain is that the device is part of Amazon’s continued push into the fashion market—which some analysts reckon could make it America’s top clothing retailer this year.

The device is fun, and only $20 more expensive than the camera-free Echo speaker, so it will, like its siblings, undoubtedly prove popular. But it’s interesting to think about what the addition of eyes to Alexa’s sensory gamut will mean for Amazon. Beyond hacking concerns raised by placing a connected camera in the location where you dress—and, presumably, undress—there’s the bigger question of what the company itself does with your data.

We asked Amazon, and it confirmed that the images and video gathered by Echo Look will be stored on the company's cloud. That’s also where the AI processing of images will be performed, too. That’s much the same as the way audio is handled by every other device in the Alexa range, with recordings started and stored away to servers every time the assistant hears its name.

That’s great news for Amazon. As our own Tom Simonite has reported, the huge quantities of data supplied by people making voice commands to their Echo are enabling the firm to make impressive breakthroughs in what voice assistants can do. Adding a camera means Amazon will be able to collect huge troves of visual data that it can analyze to determine your tastes in color and style, make recommendations, and then learn from how you respond.

But there are more insights lurking inside those shots, too—about your home decor and whatever else happens to be in frame when you capture selfies. It’s now easy to identify objects in images, and doing so will be fair game for Amazon.

By this point, of course, if you're an Amazon Prime customer, the company has already learned a great deal about your preferences based solely on your shopping habits. To be sure, the Echo Look represents another level of giving up personal data. But maybe it's worth it if you look good doing it?

(Read more: “Alexa Gives Amazon a Powerful Data Advantage,” “In 2016, AI Home Assistants Won Our Hearts,” “10 Breakthrough Technologies: Deep Learning”)

Deep Dive

Artificial intelligence

conceptual illustration showing various women's faces being scanned
conceptual illustration showing various women's faces being scanned

A horrifying new AI app swaps women into porn videos with a click

Deepfake researchers have long feared the day this would arrive.

Conceptual illustration of a therapy session
Conceptual illustration of a therapy session

The therapists using AI to make therapy better

Researchers are learning more about how therapy works by examining the language therapists use with clients. It could lead to more people getting better, and staying better.

a Chichuahua standing on a Great Dane
a Chichuahua standing on a Great Dane

DeepMind says its new language model can beat others 25 times its size

RETRO uses an external memory to look up passages of text on the fly, avoiding some of the costs of training a vast neural network

THE BLOB, 1958, promotional artwork
THE BLOB, 1958, promotional artwork

2021 was the year of monster AI models

GPT-3, OpenAI’s program to mimic human language,  kicked off a new trend in artificial intelligence for bigger and bigger models. How large will they get, and at what cost?

Stay connected

Illustration by Rose WongIllustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.