Skip to Content

Seven Must-Read Stories (Week Ending January 3, 2015)

Another chance to catch the most interesting, and important, articles from the previous week on MIT Technology Review.
  1. 2014 in Biomedicine: Rewriting DNA, Decoding the Brain, and a GMO Paradox
    From genetically modified foods to gene therapy, 2014 was a big year for rewriting biology.
  2. Could Passenger Planes Be Tracked More Closely?
    Several technologies allow aircraft to be tracked over the ocean via satellite, but most solutions are costly.
  3. New Form of Memory Could Advance Brain-Inspired Computers
    A new kind of computer memory could help make more capable computer chips that function more like biological brains, say IBM researchers.
  4. 2014 in Computing: Breakthroughs in Artificial Intelligence
    The past year saw progress in developing hardware and software capable of human feats of intelligence.
  5. How Tesla Boosted Its Roadster’s Range by 50 Percent
    An upgrade to the Roadster extends its range and shows how far battery technology has come.
  6. The Top Technology Failures of 2014
    What do the latest technologies to flop, fizzle, and flame out tell us about innovation?
  7. Device Squeezes Cells to Get Drugs In
    A new way to get materials into cells might clear the way for powerful treatments for diseases like cancer and HIV.
  8. <

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

OpenAI teases an amazing new generative video model called Sora

The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.

Google’s Gemini is now in everything. Here’s how you can try it out.

Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.

This baby with a head camera helped teach an AI how kids learn language

A neural network trained on the experiences of a single young child managed to learn one of the core components of language: how to match words to the objects they represent.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.