Summary: A new study reveals current machine learning algorithms may not be reliable in identifying brain regions associated with processing specific syllables. Researchers report machine learning may be effective at decoding mental activity from neuroimaging data, but is not quite so effective at decoding specific information processing mechanisms in the brain.
Source: University of Geneva.
For about the last ten years, researchers have been using artificial intelligence techniques called machine learning to decode human brain activity. Applied to neuroimaging data, these algorithms can reconstitute what we see, hear, and even what we think. For example, they show that words with similar meanings are grouped together in zones in different parts of our brain. However, by recording brain activity during a simple task—whether one hears BA or DA—neuroscientists from the University of Geneva (UNIGE), Switzerland, and the Ecole normale supérieure (ENS) in Paris now show that the brain does not necessarily use the regions of the brain identified by machine learning to perform a task. Above all, these regions reflect the mental associations related to this task. While machine learning is thus effective for decoding mental activity, it is not necessarily effective for understanding the specific information processing mechanisms in the brain. The results are available in the PNAS journal.
Modern neuroscientific data techniques have recently highlighted how the brain spatially organises the portrayal of word sounds, which researchers were able to precisely map by region of activity. UNIGE neuroscientists thus asked how these spatial maps are used by the brain itself when it performs specific tasks. “We have used all the available human neuroimagery techniques to try to answer this question”, says Anne-Lise Giraud, a professor at the Department of Basic Neurosciences of the UNIGE Faculty of Medicine.
A focal region for selecting information
UNIGE neuroscientists had about fifty people listen to a continuum of syllables ranging from BA to DA. The central phonemes were very ambiguous and it was difficult to distinguish between the two options. They then used a functional MRI and magnetoencephalography to see how the brain behaves when the acoustic stimulus is very clear, or, on the contrary, when it is ambiguous and requires an active mental representation of the phoneme and its interpretation by the brain. “We have observed that regardless of how difficult it is to classify the syllable that was heard, between BA and DA, the decision always engages a small region of the posterior superior temporal lobe”, notes Anne-Lise Giraud.
Neuroscientists then double-checked their results on a patient with an injury in the specific region of the posterior superior temporal lobe used to distinguish between BA and DA. “And indeed, although the patient did not appear to have symptoms, he was no longer able to distinguish between the BA and DA phonemes … this confirms that this small region is important in processing this type of phoneme information”, adds Sophie Bouton, a researcher from Anne-Lise Giraud’s team.
The “false positives” of machine learning decoding
But is the information on the identity of the syllable just locally present, as the experiment of these Genevan scientists has shown, or is it present more generally in our brain, as suggested by the maps produced via machine learning? To answer this question, the neuroscientists reproduced the BA / DA task with people who have electrodes directly implanted in their brains for medical reasons. This technique can collect very focal neural activity. A univariate analysis made it possible to see which region of the brain was solicited during the task, electrode by electrode, contact by contact. Solely the contacts in the posterior superior temporal lobe were active, thus confirming the results of the Geneva study.
However, when a machine-learning algorithm was applied to all of the data, thus making a multivariate decoding of data possible, positive results were observed in the entire temporal lobe, and even beyond it. “Learning algorithms are intelligent but ignorant”, specifies Anne-Lise Giraud. “They are very sensitive and use all of the information in the signals. However, they do not allow us to know whether this information was used to perform the task, or if it reflects the consequences of this task—in other words, spreading information in our brain”, continues Valérian Chambon, researcher at the Departement d’études cognitives at the ENS. The mapped regions outside of the posterior superior temporal lobe are thus false positives, in a way. These regions retain information on the decision that the subject makes (BA or DA), but aren’t solicited to perform this task.
This research offers a better understanding of how our brain portrays syllables and, by showing the limits of artificial intelligence in certain research contexts, fosters welcome reflection on how to interpret data produced by machine learning algorithms.
[cbtabs][cbtab title=”MLA”]University of Geneva “BA or DA? Decoding Syllables to Show the Limits of Artificial Intelligence.” NeuroscienceNews. NeuroscienceNews, 31 January 2018. <https://neurosciencenews.com/artificial-intelligence-syllables-8407/>.[/cbtab][cbtab title=”APA”]University of Geneva (2018, January 31). BA or DA? Decoding Syllables to Show the Limits of Artificial Intelligence. NeuroscienceNews. Retrieved January 31, 2018 from https://neurosciencenews.com/artificial-intelligence-syllables-8407/[/cbtab][cbtab title=”Chicago”]University of Geneva “BA or DA? Decoding Syllables to Show the Limits of Artificial Intelligence.” https://neurosciencenews.com/artificial-intelligence-syllables-8407/ (accessed January 31, 2018).[/cbtab][/cbtabs]
Focal versus distributed temporal cortex activity for speech sound category assignment
Percepts and words can be decoded from distributed neural activity measures. However, the existence of widespread representations might conflict with the more classical notions of hierarchical processing and efficient coding, which are especially relevant in speech processing. Using fMRI and magnetoencephalography during syllable identification, we show that sensory and decisional activity colocalize to a restricted part of the posterior superior temporal gyrus (pSTG). Next, using intracortical recordings, we demonstrate that early and focal neural activity in this region distinguishes correct from incorrect decisions and can be machine-decoded to classify syllables. Crucially, significant machine decoding was possible from neuronal activity sampled across different regions of the temporal and frontal lobes, despite weak or absent sensory or decision-related responses. These findings show that speech-sound categorization relies on an efficient readout of focal pSTG neural activity, while more distributed activity patterns, although classifiable by machine learning, instead reflect collateral processes of sensory perception and decision.