Summary: Study reveals how neural populations in different people synchronize during social interaction and tasks of cooperation.
Humans are social creatures. But what leads to them being this way?
To fully understand how the brain gives rise to social behaviors, we need to investigate it during social encounters. Moreover, we need to analyze not only the internal operations of one brain during social activities but also the dynamic interplay between multiple brains engaged in the same activity.
This emerging research field is referred to as “second-person neuroscience” and employs hyperscanning (the simultaneous recording of the activity of multiple brains) as the signature technique.
The earliest hyperscanning studies used functional near-infrared spectroscopy to analyze brain synchronization between participants engaged in experimentally controlled motor tasks. These tasks, however, were not very natural, and the brain synchronization emerging from them may simply have been the result of simultaneous motor action.
In contrast, recent hyperscanning studies employ more natural and free settings for the shared tasks, allowing for nonstructured communication and creative objectives. While this produces more interesting results, a greater effort is needed to extract specific social behaviors from the video footages and link them to particular brain synchronization patterns. In turn, this limits analysis to short-term social interactions, which represent a tiny subset of possible human social behaviors.
Now, a research team led by Yasuyo Minagawa of Keio University, Japan, has worked out an elegant solution to this problem. In their paper published in Neurophotonics, the team presented a novel approach in which they used computer vision techniques to automatically extract social behaviors (or “events”) during a natural two-participant experimental task.
Next, they used an event-related generalized linear model to determine which social behaviors could be linked to particular types of neuronal synchrony, both between brains and within each participant’s brain.
Each pair of participants (39 pairs in total) engaged in a natural, cooperative, and creative task: the design and furnishing of a digital room in a computer game. They were allowed to communicate freely to create a room that satisfied both.
The participants also completed the same task alone as the researchers sought to compare between-brain synchronizations (BBSs) and within-brain synchronizations (WBSs) during the individual and cooperative tasks.
The social behavior that the team focused on during the tasks was eye gaze, that is, whether the participants directed their gaze at the other’s face. They automatically extracted this behavior from the video footage using an open-source software, which made the data analysis easier.
One of the most intriguing findings of the study was that, during cooperative play, there was a strong BBS among the superior and middle temporal regions and specific parts of the prefrontal cortex in the right hemisphere, but little WBS in comparison.
Moreover, the BBS synchronization was strongest when one of the participants raised their gaze to look at the other. Interestingly, the situation reversed during individual play, showing increased WBS within the same regions.
According to Minagawa, these results agree with the idea that our brains work as a “two-in-one system” during certain social interactions.
“Neuron populations within one brain were activated simultaneously with similar neuron populations in the other brain when the participants cooperated to complete the task, as if the two brains functioned together as a single system for creative problem-solving,” she explains.
“These phenomena are consistent with the notion of a ‘we-mode,’ in which interacting agents share their minds in a collective fashion and facilitate interaction by accelerating access to the other’s cognition.”
Overall, this study provides evidence hinting at the remarkable capability of the human brain to understand and synchronize with others’ when the situation calls for it. Minagawa has high hopes for the experimental strategy used and envisions further research with other types of social cues and behaviors in various types of tasks.
“We could apply our method to more detailed social behaviors in future analyses, such as facial expressions and verbal communication. Our analytical approach can provide insights and avenues for future research in interactive social neuroscience,” she concludes. Of course, new techniques that can effectively record and process more complex social behaviors will be needed.
While there is certainly much work left to do, ingenious applications of technologies like computer vision and statistical models can enable a better understanding of how our brains (co)operate during social activities.
About this neuroscience research news
Author: Daneet Steffens
Contact: Daneet Steffens – SPIE
Image: The image is credited to Xu et al
Original Research: Open access.
“Two-in-one system and behavior-specific brain synchrony during goal-free cooperative creation: an analytical approach combining automated behavioral classification and the event-related generalized linear model” by Xu et al. Neurophotonics
Two-in-one system and behavior-specific brain synchrony during goal-free cooperative creation: an analytical approach combining automated behavioral classification and the event-related generalized linear model
In hyperscanning studies of natural social interactions, behavioral coding is usually necessary to extract brain synchronizations specific to a particular behavior. The more natural the task is, the heavier the coding effort is. We propose an analytical approach to resolve this dilemma, providing insights and avenues for future work in interactive social neuroscience.
The objective is to solve the laborious coding problem for naturalistic hyperscanning by proposing a convenient analytical approach and to uncover brain synchronization mechanisms related to human cooperative behavior when the ultimate goal is highly free and creative.
This functional near-infrared spectroscopy hyperscanning study challenged a cooperative goal-free creative game in which dyads can communicate freely without time constraints and developed an analytical approach that combines automated behavior classification (computer vision) with a generalized linear model (GLM) in an event-related manner. Thirty-nine dyads participated in this study.
Conventional wavelet-transformed coherence (WTC) analysis showed that joint play induced robust between-brain synchronization (BBS) among the hub-like superior and middle temporal regions and the frontopolar and dorsomedial/dorsolateral prefrontal cortex (PFC) in the right hemisphere, in contrast to sparse within-brain synchronization (WBS). Contrarily, similar regions within a single brain showed strong WBS with similar connection patterns during independent play. These findings indicate a two-in-one system for performing creative problem-solving tasks. Further, WTC-GLM analysis combined with computer vision successfully extracted BBS, which was specific to the events when one of the participants raised his/her face to the other. This brain-to-brain synchrony between the right dorsolateral PFC and the right temporo-parietal junction suggests joint functioning of these areas when mentalization is necessary under situations with restricted social signals.
Our proposed analytical approach combining computer vision and WTC-GLM can be applied to extract inter-brain synchrony associated with social behaviors of interest.