Skip to Content
Artificial intelligence

IBM Can Run an Experimental AI in Memory, Not on Processors

October 24, 2017

Don’t throw out your CPUs just yet, but there may be a new way to run your neural networks.

In the regular world of computing—whether you’re running exotic deep-learning algorithms or just using Excel—calculations are usually performed on a processor while data is passed back and forth to the memory. That works perfectly well, but some researchers have argued that performing calculations in memory itself would save time and energy that is usually used to move data around.

And that’s exactly the concept that a team from IBM Research in Zurich has now applied to some AI algorithms. The team has used a grid of one million memory devices, pictured above, which are all based on a phase-change material called germanium antimony telluride. The alloy’s special trick is that, when it’s hit by an electrical pulse, its state can be changed—from amorphous, like glass, to crystalline, like metal, or vice versa.

By varying the size and duration of the electric pulses, it’s possible to change the amount by which that crystallization changes. And that, in turn, can be used to represent a number of different states, not just regular 0s and 1s, which can be used to perform calculations rather than just store data. By using that quirk and enough chunks of memory, the IBM researchers have shown that they can perform machine-learning tasks like finding correlations in unknown data streams. The work is published in Nature Communications.

This is, admittedly, a small, niche, lab-based study. But the team reckons it could, if scaled up, create computing systems that perform some AI tasks 200 times faster than regular devices. Even if it can achieve just a fraction of that boost, in-memory AI may have a future.

Deep Dive

Artificial intelligence

Geoffrey Hinton tells us why he’s now scared of the tech he helped build

“I have suddenly switched my views on whether these things are going to be more intelligent than us.”

Deep learning pioneer Geoffrey Hinton has quit Google

Hinton will be speaking at EmTech Digital on Wednesday.

The future of generative AI is niche, not generalized

ChatGPT has sparked speculation about artificial general intelligence. But the next real phase of AI will be in specific domains and contexts.

Welcome to the new surreal. How AI-generated video is changing film.

Exclusive: Watch the world premiere of the AI-generated short film The Frost.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at with a list of newsletters you’d like to receive.