Brain-machine interfaces have made it possible for monkeys and some humans to control robotic limbs using just their thoughts. But ideally, a person using an artificial limb or other device would not only be able to control the device, but also feel what it’s touching.
A new study from the lab of Miguel Nicolelis at Duke University Medical Center takes a first step toward such an interface. In a paper published today in Nature, his team reports that monkeys can learn to operate a virtual-reality hand that incorporates tactile feedback.
Nicolelis says that brain-machine interfaces will only be clinically useful if they use bidirectional signals, with both sensory feedback from the device and motor commands from the user. “It’s not enough to just provide motion,” he says. “You need to sense what you’re touching.”
As a first experiment, monkeys used a joystick to control a virtual “avatar” (a monkey arm and hand) on a computer screen, and were encouraged to use the avatar to grab objects on the screen. The virtual objects had textures, and this was conveyed using stimulation through microwave arrays implanted in a part of the brain’s cortex responsible for sensing touch. The monkeys learned to hold the avatar’s hand over objects with a particular texture—conveyed by the frequency of stimulation—in order to be rewarded with food.
In another experiment, the monkeys received the same tactile feedback but controlled the virtual hand using just their thoughts, via microwire arrays implanted in the motor cortex. Although their performance on the task was less accurate, the monkeys improved over time.
Nicolelis says the successful use of a “brain-machine-brain interface” demonstrates that the processes of sensing and responding to tactile sensations can be combined. “We are decoding motor intentions and tactile messages simultaneously,” he says. “That’s never been done before.” Although the stimulation the monkeys receive is artificial, he says, they seem to learn to associate it with tactile information.
The next step is to incorporate the sense of touch into real prosthetics, using pressure sensors that will generate similar tactile feedback about real-world objects. Nicolelis says his group hopes to build a simulator that would test this approach in humans, then incorporate touch sensation in prosthetics it’s creating for people with reduced mobility.
NitishThakor, a biomedical engineer at Johns Hopkins University, says that adding sensory information “is absolutely the next logical step” in brain-machine interface design. Thakor says the experiment not only demonstrates the feasibility of adding touch, but shows that the monkeys can learn a task using these coupled signals. The caveat, he adds, is that textures in the real world are much more complex, as are body movements, and “whether this is scalable remains to be seen.”
Toronto wants to kill the smart city forever
The city wants to get right what Sidewalk Labs got so wrong.
Saudi Arabia plans to spend $1 billion a year discovering treatments to slow aging
The oil kingdom fears that its population is aging at an accelerated rate and hopes to test drugs to reverse the problem. First up might be the diabetes drug metformin.
Yann LeCun has a bold new vision for the future of AI
One of the godfathers of deep learning pulls together old ideas to sketch out a fresh path for AI, but raises as many questions as he answers.
The dark secret behind those cute AI-generated animal images
Google Brain has revealed its own image-making AI, called Imagen. But don't expect to see anything that isn't wholesome.
Get the latest updates from
MIT Technology Review
Discover special offers, top stories, upcoming events, and more.