Thoughts Bring Prostheses to Life

Summary: Using virtual reality technology, researchers found the faulty alignment of prosthetics resulted in a measurable loss in performance for users.

Source: RUB

The more natural the position of an arm prosthesis, the more precise the control – this is the result of a Bochum study on brain-computer interfaces in the therapy of severely paralysed patients.

Invasive brain-computer interfaces aim to improve the quality of life of severely paralysed patients. Movement intentions are read out in the brain, and this information is used to control robotic limbs.

A research team at the Knappschaftskrankenhaus Bochum Langendreer, University Clinic of Ruhr-Universität Bochum, has examined which errors can occur during communication between the brain and the robotic prosthesis and which of them are particularly significant. With the aid of a virtual reality model, the researchers found that a faulty alignment of the prosthesis, the so-called end effector, results in a measurable loss of performance.

The Bochum-based researchers headed by Dr. Christian Klaes from the Department of Neurosurgery published the results in the journal “Scientific Reports”.

Three main sources of error

Brain-computer interfaces can enable severely paralysed patients to move a prosthesis. In the invasive method, a measuring device implanted in the brain translates the signals from the nerve cells into control signals for the end effector, for example a robotic arm prosthesis. The Bochum-based researchers started with the assumption that three main factors have a negative impact on the control of the end-effector: the decoding error, the feedback error and the alignment error.

The decoding error describes the difference between the patient’s real intention to move and the intention to move decoded from the brain signals by the decoder. Alignment error occurs when the end effector of the brain-computer interface is incorrectly positioned relative to the participant’s natural arm.

The feedback error of the brain-computer interface system arises from a lack of somatosensory feedback, i.e. the lack of feedback from the robot arm regarding touch. The Bochum team used a virtual reality model to analyse the misalignment and feedback errors – independently of the decoding error and also independently of each other.

One-to-one translation of movement intentions into movements

“Healthy study participants without sensorimotor disorders slipped into the role of patients with motor dysfunctions in virtual reality,” explains Robin Lienkämper, lead author of the study. “Our model thus provides a one-to-one translation of movement intentions into end-effector movements, comparable to that of a patient using a faultless decoder.”

In virtual reality, the participants were given the task of drawing shapes with a pencil – a square, a circle, a star, a spiral and an asymmetrical shape. This corresponds to a frequently set task in experiments on brain-computer interfaces, which can be used to evaluate and compare motor performance under different conditions.

This shows a man in a VR gaming system
Using Virtual Reality, the researchers analysed different errors that can occur during the control of prostheses via a brain-computer interface. Credit: Susanne Troll

The controller was perceived as a pen by the test participants in the experiment. The researchers achieved the desired feedback effect by having the test person sit at a real table while drawing in the virtual world and the controller touching table surface. There were two groups to control the effect: one group received indirect haptic feedback, the other did not. This means that for the second group, the physical table was removed while the table remained visible in virtual reality.

Ideally, the robotic arm is incorporated in the body schema

Using the collated data, the research team showed that the lack of indirect haptic feedback alone had a minor impact, but amplified the effect of the misalignment. Based on the results, the researchers also suggested that a naturally positioned prosthesis could significantly improve the performance of patients with invasive brain-computer interfaces. They also hypothesised that anchoring the robotic arm to the patient’s own body awareness would have a positive effect and improve motor performance. Ideally, a patient using a brain-computer interface would incorporate the robotic arm into their own body schema.

The researchers concluded their study by highlighting the importance of developing end effectors that enable better incorporation and more natural positioning. According to them, solutions such as exoskeletons or functional muscle stimulation should be considered. “The future of research now lies in using engineering to bring the scientific results to the patient,” says Robin Lienkämper. He expects that, thanks to the commitment of the industry, applications suitable for everyday use will already be available in five to ten years.

Funding:

The study was funded by the Collaborative Research Centres 874 (SFB 874) of the German Research Foundation and the Emmy Noether Programme. The SFB 874 “Integration and Representation of Sensory Processes” has existed at Ruhr-Universität Bochum since 2010. The researchers examined how sensory signals generate neuronal maps, and how this leads to complex behaviour and memory formation.

The Emmy Noether Programme offers highly qualified early career researchers the opportunity to qualify for a university professorship by leading a junior research group.

About this neurotech research news

Source: RUB
Contact: Christian Klaes – RUB
Image: The image is credited to Susanne Troll

Original Research: Open access.
Quantifying the alignment error and the effect of incomplete somatosensory feedback on motor performance in a virtual brain–computer interface setup” by Robin Lienkämper, Susanne Dyck, Muhammad Saif ur Rehman, Marita Metzler, Omair Ali, Christian Klaes. Scientific Reports


Abstract

Quantifying the alignment error and the effect of incomplete somatosensory feedback on motor performance in a virtual brain–computer interface setup

Invasive brain–computer-interfaces (BCIs) aim to improve severely paralyzed patient’s (e.g. tetraplegics) quality of life by using decoded movement intentions to let them interact with robotic limbs.

We argue that the performance in controlling an end-effector using a BCI depends on three major factors: decoding error, missing somatosensory feedback and alignment error caused by translation and/or rotation of the end-effector relative to the real or perceived body.

Using a virtual reality (VR) model of an ideal BCI decoder with healthy participants, we found that a significant performance loss might be attributed solely to the alignment error. We used a shape-drawing task to investigate and quantify the effects of robot arm misalignment on motor performance independent from the other error sources.

We found that a 90° rotation of the robot arm relative to the participant leads to the worst performance, while we did not find a significant difference between a 45° rotation and no rotation. Additionally, we compared a group of subjects with indirect haptic feedback with a group without indirect haptic feedback to investigate the feedback-error.

In the group without feedback, we found a significant difference in performance only when no rotation was applied to the robot arm, supporting that a form of haptic feedback is another important factor to be considered in BCI control.

Join our Newsletter
I agree to have my personal information transferred to AWeber for Neuroscience Newsletter ( more information )
Sign up to receive our recent neuroscience headlines and summaries sent to your email once a day, totally free.
We hate spam and only use your email to contact you about newsletters. You can cancel your subscription any time.