A turning point for Rafael Yuste, a neuroscientist at New York's Columbia University, came when his lab discovered it could activate a few neurons in a mouse's visual cortex and make it hallucinate.
Researchers use electroencephalography on a person to measure brain activity. Photo: AFP
The mouse had been trained to lick at a water spout every time it saw two vertical bars, and researchers were able to prompt it to drink even with no bars in sight, said Yuste, whose team published a study on the experiment in 2019.
"We could make the animal see something it didn't see, as if it were a puppet," he told the Thomson Reuters Foundation in a phone interview. "If we can do this today with an animal, we can do it tomorrow with a human for sure."
Yuste is part of a group of scientists and lawmakers, stretching from Switzerland to Chile, who are working to rein in the potential abuses of neuroscience by companies from tech giants to wearable startups.
Following his team's discovery, he launched the NeuroRights Initiative, which advocates five "neuro-rights" to protect how a person's brain data is accessed and used, including a right to mental privacy and free will.
"Right now, it's the wild west," Yuste said.
In Chile, senate member Guido Girardi is pushing to translate those principles into law, with a bill that would give legal protection to a suit of neuro-rights, and a complementary reform to the country's constitution.
In March, the National Commission for Scientific and Technological Research began debating Girardi's proposal, which got unanimous support from parliament in December 2020. His office hopes the bill will be adopted later in the year.
"If this technology is industrialized without the proper regulations and rules, it will threaten fundamental human autonomy," he said in a phone interview.
'Science-fiction scenarios'
Advances in brain science like those made by Yuste's team have made it possible to penetrate the brain using sensors and implants, and access some degree of neural activity.
The US Food and Drug Administration has approved deep brain stimulation procedures - implanting electrodes in the brain - to treat a range of disorders from Parkinson's disease to epilepsy. And major tech firms, from Facebook to Tesla, are working on "computer-brain" interfaces to allow consumers to control devices with their thoughts, while some smaller companies sell wearable devices to monitor brain activity.
But warnings of "science-fiction scenarios" of for-profit mind control are overblown for a line of research that is still so young, said Karen Rommelfanger, director of the neuroethics program at Emory University in Atlanta.
"Yes, the science will get better, not worse," she said. "But exactly how it develops is up in the air."
Ienca at ETH Zurich said major ethical issues could arise if the commercial neurotech data devices are widely shared and analyzed without proper safeguards, he said.
"We already have digital biomarkers that can indicate if someone is predisposed to developing dementia. Let's say [that] data is shared with a prospective employer, you could face discrimination on the job market," he said.
In 2018, Ienca published a review of six commercially available "neuromonitoring" headsets in the journal Nature Biotechnology.
He found that the electroencephalography (EEG) data gathered by the devices as they measure electrical activity in the brain could be leaked online, sold to third parties, or subjected to uses that consumers did not consent to. This also concerns Adam Molnar, cofounder of the neurotech start-up Neurable, which is developing headphones that measure EEG to help users track brain activity and emotions like burnout.
When Neurable launches a new device, he said, it pledges not to sell user data and only uses collected data to improve its own products.
"We want to be the good guys," he said, adding that he hopes the move will help set the tone for other neurotech firms.
Data harvestingRommelfanger at Emory is wary of moving too quickly to regulate brain tech, which she said could stifle innovation. She recommends direct engagement with startups working on commercial devices, encouraging them to develop privacy-conscious and ethically minded products.
Girardi favors strict regulation.
"If we allow for all this brain data to be taken, who knows what the consequences will be? We'll have algorithms deciding what it means to be 'happy,'" Girardi said.
Tim Brown, a neuroethics expert at the University of Washington, said the data currently being collected is not powerful enough to do that.
"A lot of that brain data is basically noise," he said. But, he noted, scientists are working on algorithms to decode and analyze the data gleaned from EEG and functional magnetic resonance imaging scans, hoping to build computer models that can interpret an individual's mental state. He predicted the same dynamics present in the social media or search industry - where companies offer a service for free in return for permission to harvest user data - will likely surface in neurotech.
That could lead to serious consequences for privacy in the coming years, Brown warned, with companies linking users' social media behavior to their brain images in real time to craft ads or other messages.
He also worries about how neurotechnology might exacerbate existing patterns of discrimination and racism. In his research, he has warned of the possibility of "mandatory neurointerventions," when institutions like schools or prisons might deploy neurotechnology to assess mental states.