Brain-machine interfaces have made it possible for monkeys and some humans to control robotic limbs using just their thoughts. But ideally, a person using an artificial limb or other device would not only be able to control the device, but also feel what it’s touching.
A new study from the lab of Miguel Nicolelis at Duke University Medical Center takes a first step toward such an interface. In a paper published today in Nature, his team reports that monkeys can learn to operate a virtual-reality hand that incorporates tactile feedback.
Nicolelis says that brain-machine interfaces will only be clinically useful if they use bidirectional signals, with both sensory feedback from the device and motor commands from the user. “It’s not enough to just provide motion,” he says. “You need to sense what you’re touching.”
As a first experiment, monkeys used a joystick to control a virtual “avatar” (a monkey arm and hand) on a computer screen, and were encouraged to use the avatar to grab objects on the screen. The virtual objects had textures, and this was conveyed using stimulation through microwave arrays implanted in a part of the brain’s cortex responsible for sensing touch. The monkeys learned to hold the avatar’s hand over objects with a particular texture—conveyed by the frequency of stimulation—in order to be rewarded with food.
In another experiment, the monkeys received the same tactile feedback but controlled the virtual hand using just their thoughts, via microwire arrays implanted in the motor cortex. Although their performance on the task was less accurate, the monkeys improved over time.
Nicolelis says the successful use of a “brain-machine-brain interface” demonstrates that the processes of sensing and responding to tactile sensations can be combined. “We are decoding motor intentions and tactile messages simultaneously,” he says. “That’s never been done before.” Although the stimulation the monkeys receive is artificial, he says, they seem to learn to associate it with tactile information.
The next step is to incorporate the sense of touch into real prosthetics, using pressure sensors that will generate similar tactile feedback about real-world objects. Nicolelis says his group hopes to build a simulator that would test this approach in humans, then incorporate touch sensation in prosthetics it’s creating for people with reduced mobility.
NitishThakor, a biomedical engineer at Johns Hopkins University, says that adding sensory information “is absolutely the next logical step” in brain-machine interface design. Thakor says the experiment not only demonstrates the feasibility of adding touch, but shows that the monkeys can learn a task using these coupled signals. The caveat, he adds, is that textures in the real world are much more complex, as are body movements, and “whether this is scalable remains to be seen.”
DeepMind’s cofounder: Generative AI is just a phase. What’s next is interactive AI.
“This is a profound moment in the history of technology,” says Mustafa Suleyman.
What to know about this autumn’s covid vaccines
New variants will pose a challenge, but early signs suggest the shots will still boost antibody responses.
Human-plus-AI solutions mitigate security threats
With the right human oversight, emerging technologies like artificial intelligence can help keep business and customer data secure
Next slide, please: A brief history of the corporate presentation
From million-dollar slide shows to Steve Jobs’s introduction of the iPhone, a bit of show business never hurt plain old business.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.