Dedicated machine-learning hardware could help Google fight off rivals in an increasingly competitive cloud AI market.
Backstory: Last year, Google announced it had designed a new chip, called a tensor processing unit (TPU), built to crunch the math AI uses. At the time, it ran the chips itself and allowed just a select group of researchers to make use of them.
What's new: The New York Times reports that Google will allow other companies to make use of the hardware via the cloud. “We are trying to reach as many people as we can as quickly as we can,” Zak Stone, leader of Google’s TPU team, told the newspaper.
Why it matters: Putting AI in the cloud is big business. Google, Amazon, and Microsoft all provide AI software on their cloud servers, and China is joining the race, too. By offering dedicated hardware for AI grunt work, Google will hope to gain a competitive edge over the others.
A chip design that changes everything: 10 Breakthrough Technologies 2023
Computer chip designs are expensive and hard to license. That’s all about to change thanks to the popular open standard known as RISC-V.
Modern data architectures fuel innovation
More diverse data estates require a new strategy—and the infrastructure to support it.
Chinese chips will keep powering your everyday life
The war over advanced semiconductor technology continues, but China will likely take a more important role in manufacturing legacy chips for common devices.
The computer scientist who hunts for costly bugs in crypto code
Programming errors on the blockchain can mean $100 million lost in the blink of an eye. Ronghui Gu and his company CertiK are trying to help.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.