Multicore Patents
Many microchips today incorporate multiple processing units, known as cores, to boost performance (see “Designing for Mobility”). These cores share system resources such as memory. Widely cited U.S. patent 5,617,537, awarded to Japanese communications giant NTT in 1997, arbitrates between cores as they access system memory, so that one core does not, for example, accidentally overwrite information that another core has placed in memory.
A patent map created by IPVision, based in Cambridge, MA, shows a key invention by Japanese communications giant NTT. Awarded in 1997, the patent is important to the growing trend toward so-called multicore chips in computers; these chips incorporate multiple processing units, known as cores, which share system resources, such as memory. The advantage of a multicore system is that, in theory, a pair of two-gigahertz cores, for instance, could work together to do the job of one four-gigahertz core but with a lot less power–an important consideration for mobile devices. In practice, it can be difficult to coordinate the cores so that they work together seamlessly. This widely cited patent describes a way to handle a big piece of the coordination puzzle: how to arbitrate between cores as they access system memory, so that one core does not accidently overwrite information that another core has placed in memory.
Credit: IPVision
Keep Reading
Most Popular
Large language models can do jaw-dropping things. But nobody knows exactly why.
And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.
OpenAI teases an amazing new generative video model called Sora
The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.
Google’s Gemini is now in everything. Here’s how you can try it out.
Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.
This baby with a head camera helped teach an AI how kids learn language
A neural network trained on the experiences of a single young child managed to learn one of the core components of language: how to match words to the objects they represent.
Stay connected
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.