Skip to Content

3-D View of the Brain

New software for the operating room.
October 15, 2007

The 3-D rendering of a brain tumor and associated brain fibers at left, made by researchers at Thomas ­Jefferson University in Philadelphia, is the product of new software that integrates data from multiple imaging technologies to provide much clearer images. The image shows a tumor (blue mass) and its position relative to brain fibers (threadlike objects) that are affected by the tumor and vital to brain function. The software lets surgeons see which fibers are engulfed by the tumor (dashed lines) and which are outside the tumor (solid lines). Colors indicate the depth of the engulfed fibers or the exterior fibers’ distance from the tumor’s surface; for example, pink and red dashed lines represent deeply engulfed fibers. Pilot studies have demonstrated the software’s usefulness in neurosurgical planning; the researchers expect that with fine-tuning, the technology could be in operating rooms within a year. The images use data from conventional magnetic reso­nance imaging (MRI), which gives details on ­anatomy; functional MRI, which provides information on neural activity; and diffusion tensor imaging, which provides images of fibers connecting different brain areas.

Multimedia

  • See images produced by the new software.

Keep Reading

Most Popular

Large language models can do jaw-dropping things. But nobody knows exactly why.

And that's a problem. Figuring it out is one of the biggest scientific puzzles of our time and a crucial step towards controlling more powerful future models.

OpenAI teases an amazing new generative video model called Sora

The firm is sharing Sora with a small group of safety testers but the rest of us will have to wait to learn more.

Google’s Gemini is now in everything. Here’s how you can try it out.

Gmail, Docs, and more will now come with Gemini baked in. But Europeans will have to wait before they can download the app.

This baby with a head camera helped teach an AI how kids learn language

A neural network trained on the experiences of a single young child managed to learn one of the core components of language: how to match words to the objects they represent.

Stay connected

Illustration by Rose Wong

Get the latest updates from
MIT Technology Review

Discover special offers, top stories, upcoming events, and more.

Thank you for submitting your email!

Explore more newsletters

It looks like something went wrong.

We’re having trouble saving your preferences. Try refreshing this page and updating them one more time. If you continue to get this message, reach out to us at customer-service@technologyreview.com with a list of newsletters you’d like to receive.