Select your localized edition:

Close ×

More Ways to Connect

Discover one of our 28 local entrepreneurial communities »

Be the first to know as we launch in new countries and markets around the globe.

Interested in bringing MIT Technology Review to your local market?

MIT Technology ReviewMIT Technology Review - logo

 

Unsupported browser: Your browser does not meet modern web standards. See how it scores »

When you’re sick, it’s obvious that you and your doctor should work, if possible, to help you get better. Nobody would argue against a treatment that restores normal function to a sick or disabled individual. But the consequences of going further than that–going beyond “normal”–are not commonly studied, nor endorsed by many in medicine. Indeed, in any medical procedure, there is risk. If you are already normal, then conventional wisdom dictates that that’s enough. “Do no harm,” the old aphorism says–we should focus on altering the body and mind only when the risk of the alteration is justified, preferably by the hope of solving a deficit of vastly greater magnitude.

Science has endorsed something of a parallel attitude in its pursuit of biological and biomedical research. Namely, in biology, many key insights have emerged via study of the absence of the phenomenon of interest. For example, throughout the 20th century, many insights about the brain arose from the knocking out of specific genes, or the ablation of specific neurons in animals, or the examination of human patients who have suffered the loss of brain regions from conditions such as war or medically mandated surgery. In this way, we learned that patients without their hippocampi become unable to form new memories; humans and experimental animals with prefrontal-cortex damage make bad decisions and lack impulse control; subjects without dopamine-producing neurons exhibit symptoms of Parkinson’s disease. These studies are good at demonstrating the necessity of a specific neural circuit, or brain region, to the normal state. The idea that biomedical science is supposed to bring us up to normal is embedded, to a degree, in the very structure of the experiments we commonly do in the laboratory.

There is nothing wrong with this line of thought. This angle of research is fully consistent with our medical goal. It has brought us many of the triumphs of the last century, and it continues to yield insights into the vast range of diseases that plague us throughout life. When researchers leave this line of thought, though, they point toward the possibility that going beyond normal may change us in new and unprecedented ways, improving our lives in ways that are hard to even imagine. One fascinating line of research over the past few decades has revealed that life span, which long appeared to be solidly set in stone, can be lengthened through pharmacological and genetic manipulations–at least in worms, yeast, and, most recently, mice. This work may someday (quite possibly soon) lead to drugs that can extend human life span. Or note that this past summer, double amputee Oscar Pistorius won second place in a race against able-bodied runners, racing with his prosthetic carbon-fiber legs. Now he is hoping to take on the Olympics, which has led international athletic bodies to worry that augmented humans may be better at running than normal ones are. As a final example, cognitive-augmentation drugs such as modafinil, which enhances alertness even after long hours of wakefulness, are becoming widespread. (Broadly interpreted, good ole coffee might be considered the original neurotechnology, having augmented attention, alertness, and memory in tired humans for millennia.)

It’s arguably time for a discipline to emerge around the idea of human augmentation. At the MIT Media Lab, we are beginning to search for principles that govern the use of technology to augment human abilities–that make the idea of normal obsolete. As a codirector of the Center for Human Augmentation, I lead a lab, the Neuroengineering and Neuromedia Lab, that is developing devices that will hopefully eventually allow us to enhance memory, creativity, and happiness in humans. One interesting observation that has emerged is that it’s much easier to know when something is gone than it is to characterize it in its intact state. For example, it’s far easier to demonstrate that an animal can form no new memories than it is to characterize the trajectory that memories take as they are learned, consolidated, and forgotten throughout the lifetime of that animal. And whereas many measures of depression and sadness have been defined, a coherent description of happiness remains elusive. How can you augment something if you can’t define it? One of the first things we are doing is developing better, measurable definitions of such phenomena. Another issue is that radically new tools are needed to augment the mind. We are developing new kinds of neural stimulators, for example, that enable highly targeted manipulations of the brain. Some of our inventions, like the ability to turn specific sets of neurons on and off with brief pulses of blue and yellow light, may be used chiefly in animals for the next few years, but I think they will find many compelling uses in humans in the years to come, as their power becomes manifest through the efforts of a great many neuroscientists and engineers.

Cognitive augmentation will require new technologies.

One argument in favor of going for optimality, and forgetting about normal, is that it’s becoming harder and harder to know what is normal. For example, it’s been demonstrated that two-thirds of all people have at least one copy of a DNA sequence that makes them more likely to become depressed after a stressful life event. The rest of all people, a minority of one-third, are more resilient to stress than the other two-thirds are. Thus, it could be argued that becoming depressed in response to stress is the normal state. As a neuroengineer, I think it’s easier just to develop neurotechnologies that will enable us to make people as happy and intelligent as possible, and perhaps to even go farther: taking on the questions that philosophy struggles with, such as how to find meaning in one’s life. (More on that last point in a future post.)

What is a problem, and what is a feature of the human condition? They are not necessarily distinct. But that doesn’t mean we shouldn’t continue to find better ways to make life better. In that way, we’ll hopefully move, in the century to come, from “Do no harm” to “Do good.”

Cite as: Boyden, E. S. “In Pursuit of Human Augmentation.” Ed Boyden’s Blog. Technology Review. 9/17/07. (http://www.technologyreview.com/blog/boyden/21839/).

16 comments. Share your thoughts »

Tagged: Biomedicine, brain, neuroscience, memory, depression, neuroengineering, cognition, stress

Reprints and Permissions | Send feedback to the editor

From the Archives

Close

Introducing MIT Technology Review Insider.

Already a Magazine subscriber?

You're automatically an Insider. It's easy to activate or upgrade your account.

Activate Your Account

Become an Insider

It's the new way to subscribe. Get even more of the tech news, research, and discoveries you crave.

Sign Up

Learn More

Find out why MIT Technology Review Insider is for you and explore your options.

Show Me