‘Virtual Partner’ Elicits Emotional Responses From a Human

Summary: Researchers have created a ‘virtual partner’ that is able to elicit emotional response from humans in real time.

Source: Florida Atlantic University.

“How does it ‘feel’ to interact behaviorally with a machine?” To answer that question, scientists created a virtual partner that can elicit emotional responses from its human partner in real-time.

Can machines think? That’s what renowned mathematician Alan Turing sought to understand back in the 1950s when he created an imitation game to find out if a human interrogator could tell a human from a machine based solely on conversation deprived of physical cues. The Turing test was introduced to determine a machine’s ability to show intelligent behavior that is equivalent to or even indistinguishable from that of a human. Turing mainly cared about whether machines could match up to humans’ intellectual capacities.

But there is more to being human than intellectual prowess, so researchers from the Center for Complex Systems and Brain Sciences (CCSBS) in the Charles E. Schmidt College of Science at Florida Atlantic University set out to answer the question: “How does it ‘feel’ to interact behaviorally with a machine?”

They created the equivalent of an “emotional” Turing test, and developed a virtual partner that is able to elicit emotional responses from its human partner while the pair engages in behavioral coordination in real-time.

Results of the study, titled “Enhanced Emotional Responses during Social Coordination with a Virtual Partner,” are recently published in the International Journal of Psychophysiology . The researchers designed the virtual partner so that its behavior is governed by mathematical models of human-to-human interactions in a way that enables humans to interact with the mathematical description of their social selves.

“Our study shows that humans exhibited greater emotional arousal when they thought the virtual partner was a human and not a machine, even though in all cases, it was a machine that they were interacting with,” said Mengsen Zhang, lead author and a Ph.D. student in FAU’s CCSBS. “Maybe we can think of intelligence in terms of coordinated motion within and between brains.”

The virtual partner is a key part of a paradigm developed at FAU called the Human Dynamic Clamp – a state-of-the-art human machine interface technology that allows humans to interact with a computational model that behaves very much like humans themselves. In simple experiments, the model – on receiving input from human movement – drives an image of a moving hand which is displayed on a video screen. To complete the reciprocal coupling, the subject sees and coordinates with the moving image as if it were a real person observed through a video circuit. This social “surrogate” can be precisely tuned and controlled – both by the experimenter and by the input from the human subject.

“The behaviors that gave rise to that distinctive emotional arousal were simple finger movements, not events like facial expressions for example, known to convey emotion,” said Emmanuelle Tognoli, Ph.D., co-author and associate research professor in FAU’s CCSBS. “So the findings are rather startling at first.”

Tognoli is quick to point out that that it is not so much about how fanciful the partner appears or how emotionally prone it is, since usually, fingers have little in the way of tears or laughter. Instead, it is a matter of how well the virtual partner relates its behavior to the human – its competence for social coordination written in its mathematical equations.

The mathematical models that govern the virtual partner’s behavior are grounded in four decades of empirical and theoretical research at FAU led by co-author J.A. Scott Kelso, Ph.D., the Glenwood and Martha Creech Eminent Chair in Science, and founder of FAU’s CCSBS.

Kelso stresses that the key idea behind the Human Dynamic Clamp is the symmetry between the human and the machine, the fact that they are governed by the same laws of coordination dynamics.

“In reality, humans’ interactions with their milieu, including other human beings, are continuous and reciprocal,” said Kelso. “By putting time and reciprocity back in the investigation of emotion and social interaction, the Human Dynamic Clamp affords the opportunity to explore parameter ranges and perturbations that are out of reach of traditional experimental designs. It is a step forward for investigations aimed at understanding complex social behavior.”

Photo of a woman holding a robot's hand.
The virtual partner is a key part of a paradigm developed at FAU called the Human Dynamic Clamp – a state-of-the-art human machine interface technology that allows humans to interact with a computational model that behaves very much like humans themselves. NeuroscienceNews.com image is adapted from the Florida Atlantic University press release.

The study shows that behavioral interaction and emotion are continuously feeding from each other, so that coordination of movement could make useful contribution to the rehabilitation of diseases. Movement coordination disorders are often found in patients with schizophrenia and autism spectrum disorders, who also suffer from social and emotional dysfunctions.

“Artificial Intelligence has been grounded in an algorithmic approach of human cognition. We are now bringing the social and emotional dimensions to the table as well,” said Guillaume Dumas, Ph.D., co-author, and a former post-doctoral member of FAU’s CCSBS who is currently with the Institut Pasteur in Paris.

The researchers anticipate that the virtual partner will soon be developed into the prototype of a cooperative machine that can be used for therapeutic purposes. This type of application might benefit many patients afflicted with social and emotional disorders.

“FAU has nurtured the Center for Complex Systems and Brain Sciences for 30 years, and this work is led by one of our outstanding doctoral students who is advancing our understanding of the orders and disorders that take place in our society and in our brains,” said Daniel C. Flynn, Ph.D., FAU’s vice president for research.

[divider]About this psychology research article[/divider]

Funding: The study was supported by the National Institute of Mental Health’s (NIMH) “Affect, Social Behavior and Social Cognition Program.”

Source: Gisele Gloustain – Florida Atlantic University
Image Source: This NeuroscienceNews.com image is adapted from the Florida Atlantic University press release.
Original Research: Abstract for “Enhanced emotional responses during social coordination with a virtual partner” by Mengsen Zhang, Guillaume Dumas, J.A. Scott Kelso, and Emmanuelle Tognoli in International Journal of Psychophysiology. Published online April 16 2016 doi:10.1016/j.ijpsycho.2016.04.001

[divider]Cite This NeuroscienceNews.com Article[/divider]

[cbtabs][cbtab title=”MLA”]Florida Atlantic University. “‘Virtual Partner’ Elicits Emotional Responses from a Human.” NeuroscienceNews. NeuroscienceNews, 17 May 2016.
<https://neurosciencenews.com/emotion-virtual-partner-psychology-4236/>.[/cbtab][cbtab title=”APA”]Florida Atlantic University. (2016, May 17). ‘Virtual Partner’ Elicits Emotional Responses from a Human. NeuroscienceNews. Retrieved May 17, 2016 from https://neurosciencenews.com/emotion-virtual-partner-psychology-4236/[/cbtab][cbtab title=”Chicago”]Florida Atlantic University. “‘Virtual Partner’ Elicits Emotional Responses from a Human.” NeuroscienceNews.
https://neurosciencenews.com/emotion-virtual-partner-psychology-4236/ (accessed May 17, 2016).[/cbtab][/cbtabs]


Abstract

Enhanced emotional responses during social coordination with a virtual partner

Emotion and motion, though seldom studied in tandem, are complementary aspects of social experience. This study investigates variations in emotional responses during movement coordination between a human and a Virtual Partner (VP), an agent whose virtual finger movements are driven by the Haken-Kelso-Bunz (HKB) equations of Coordination Dynamics. Twenty-one subjects were instructed to coordinate finger movements with the VP in either inphase or antiphase patterns. By adjusting model parameters, we manipulated the ‘intention’ of VP as cooperative or competitive with the human’s instructed goal. Skin potential responses (SPR) were recorded to quantify the intensity of emotional response. At the end of each trial, subjects rated the VP’s intention and whether they thought their partner was another human being or a machine. We found greater emotional responses when subjects reported that their partner was human and when coordination was stable. That emotional responses are strongly influenced by dynamic features of the VP’s behavior, has implications for mental health, brain disorders and the design of socially cooperative machines.

“Enhanced emotional responses during social coordination with a virtual partner” by Mengsen Zhang, Guillaume Dumas, J.A. Scott Kelso, and Emmanuelle Tognoli in International Journal of Psychophysiology. Published online April 16 2016 doi:10.1016/j.ijpsycho.2016.04.001

[divider]Feel free to share this Neuroscience News.[/divider]

Join our Newsletter
I agree to have my personal information transferred to AWeber for Neuroscience Newsletter ( more information )
Sign up to receive the latest neuroscience headlines and summaries sent to your email daily from NeuroscienceNews.com
We hate spam and only use your email to contact you about newsletters. We do not sell email addresses. You can cancel your subscription any time.