Summary: Researchers call for regulations and ethical guidelines to help protect personal privacy and autonomy for those who use neurotechnologies.
Source: Columbia University.
The convergence of artificial intelligence and brain-computer interfaces may soon restore sight to the blind, allow the paralyzed to move robotic limbs and cure any number of brain and nervous system disorders.
But without regulation, this flurry of innovation spells trouble for humanity, warns a team of researchers led by Columbia University neuroscientist Rafael Yuste and University of Washington bioethicist Sara Goering. In a new essay in Nature, Yuste and Goering join more than two dozen physicians, ethicists, neuroscientists, and computer scientists, in calling for ethical guidelines to cover the evolving use of computer hardware and software to enhance or restore human capabilities.
“We just want to ensure that this new technology which is so exciting, and which could revolutionize our lives, is used for the good of mankind,” said Yuste, director of Columbia’s Neurotechnology Center and a member of the Data Science Institute.
Long the stuff of science fiction, the melding of computers with the human mind to augment or restore brain function is moving closer to reality. The authors estimate that the for-profit brain implant industry is now worth $100 million, led by Bryan Johnson’s startup Kernel, and Elon Musk’s Neuralink. Under President Obama’s BRAIN Initiative alone, The U.S. government has spent another $500 million since 2013, they write.
As these investments bear fruit, the authors see four main threats: the loss of individual privacy, identity and autonomy, and the potential for social inequalities to widen, as corporations, governments, and hackers gain added power to exploit and manipulate people.
To protect privacy, the authors recommend that individuals be required to opt in, as organ donors do, to sharing their brain data from their devices, and that the sale and commercial use of personal data be strictly regulated.
To protect autonomy and identity, the authors recommend that an international convention be created to define what actions would be prohibited, and to educate people about the possible effects on mood, personality and sense of self.
Finally, to address the potential for a brain-enhancement arms race pitting people with super-human intelligence and endurance against everyone else, they suggest creating culture-specific commissions to establish norms and regulations. They also recommend that military use of brain technologies be controlled, much as chemical and biological weapons are already under the Geneva Protocol.
In an earlier essay in the journal Cell, Yuste and Goering laid out similar arguments for integrating ethics into brain technologies, citing the 1970s Belmont Report which set ethical principles and guidelines for research involving human subjects.
Source: Kim Martineau – Columbia University
Publisher: Organized by NeuroscienceNews.com.
Image Source: NeuroscienceNews.com image is in the public domain.
Original Research: Neuroscience Essay “Four ethical priorities for neurotechnologies and AI” by Rafael Yuste, Sara Goering, Blaise Agüera y Arcas, Guoqiang Bi, Jose M. Carmena, Adrian Carter, Joseph J. Fins, Phoebe Friesen, Jack Gallant, Jane E. Huggins, Judy Illes, Philipp Kellmeyer, Eran Klein, Adam Marblestone, Christine Mitchell, Erik Parens, Michelle Pham, Alan Rubel, Norihiro Sadato, Laura Specker Sullivan, Mina Teicher, David Wasserman, Anna Wexler, Meredith Whittaker & Jonathan Wolpaw in Nature. Published online November 8 2017
Four ethical priorities for neurotechnologies and AI
Consider the following scenario. A paralysed man participates in a clinical trial of a brain–computer interface (BCI). A computer connected to a chip in his brain is trained to interpret the neural activity resulting from his mental rehearsals of an action. The computer generates commands that move a robotic arm. One day, the man feels frustrated with the experimental team. Later, his robotic hand crushes a cup after taking it from one of the research assistants, and hurts the assistant. Apologizing for what he says must have been a malfunction of the device, he wonders whether his frustration with the team played a part.
“Four ethical priorities for neurotechnologies and AI” by Rafael Yuste, Sara Goering, Blaise Agüera y Arcas, Guoqiang Bi, Jose M. Carmena, Adrian Carter, Joseph J. Fins, Phoebe Friesen, Jack Gallant, Jane E. Huggins, Judy Illes, Philipp Kellmeyer, Eran Klein, Adam Marblestone, Christine Mitchell, Erik Parens, Michelle Pham, Alan Rubel, Norihiro Sadato, Laura Specker Sullivan, Mina Teicher, David Wasserman, Anna Wexler, Meredith Whittaker & Jonathan Wolpaw in Nature. Published online November 8 2017