Scientists develop ‘virtual partner’ that can elicit emotional responses from a human partner in real-time
Can machines think? That’s what renowned mathematician Alan Turing sought to understand back in the 1950s when he created an imitation game to find out if a human interrogator could tell a human from a machine based solely on conversation deprived of physical cues. The Turing test was introduced to determine a machine’s ability to show intelligent behavior that is equivalent to or even indistinguishable from that of a human. Turing mainly cared about whether machines could match up to humans’ intellectual capacities.
But there is more to being human than intellectual prowess, so researchers from the Center for Complex Systems and Brain Sciences (CCSBS) in the Charles E. Schmidt College of Science at Florida Atlantic University set out to answer the question: “How does it ‘feel’ to interact behaviorally with a machine?”
They created the equivalent of an “emotional” Turing test, and developed a virtual partner that is able to elicit emotional responses from its human partner while the pair engages in behavioral coordination in real-time. Results of the study, titled “Enhanced Emotional Responses during Social Coordination with a Virtual Partner,” are recently published in the International Journal of Psychophysiology. The researchers designed the virtual partner so that its behavior is governed by mathematical models of human-to-human interactions in a way that enables humans to interact with the mathematical description of their social selves.
“Our study shows that humans exhibited greater emotional arousal when they thought the virtual partner was a human and not a machine, even though in all cases, it was a machine that they were interacting with,” said Mengsen Zhang, lead author and a Ph.D. student in FAU’s CCSBS. “Maybe we can think of intelligence in terms of coordinated motion within and between brains.”
The virtual partner is a key part of a paradigm developed at FAU called the Human Dynamic Clamp – a state-of-the-art human machine interface technology that allows humans to interact with a computational model that behaves very much like humans themselves. In simple experiments, the model – on receiving input from human movement – drives an image of a moving hand which is displayed on a video screen. To complete the reciprocal coupling, the subject sees and coordinates with the moving image as if it were a real person observed through a video circuit. This social “surrogate” can be precisely tuned and controlled – both by the experimenter and by the input from the human subject.
“The behaviors that gave rise to that distinctive emotional arousal were simple finger movements, not events like facial expressions for example, known to convey emotion,” said Emmanuelle Tognoli, Ph.D., co-author and associate research professor in FAU’s CCSBS. “So the findings are rather startling at first.”
Tognoli is quick to point out that that it is not so much about how fanciful the partner appears or how emotionally prone it is, since usually, fingers have little in the way of tears or laughter. Instead, it is a matter of how well the virtual partner relates its behavior to the human – its competence for social coordination written in its mathematical equations.
The mathematical models that govern the virtual partner’s behavior are grounded in four decades of empirical and theoretical research at FAU led by co-author J.A. Scott Kelso, Ph.D., the Glenwood and Martha Creech Eminent Chair in Science, and founder of FAU’s CCSBS.
Kelso stresses that the key idea behind the Human Dynamic Clamp is the symmetry between the human and the machine, the fact that they are governed by the same laws of coordination dynamics.
“In reality, humans’ interactions with their milieu, including other human beings, are continuous and reciprocal,” said Kelso. “By putting time and reciprocity back in the investigation of emotion and social interaction, the Human Dynamic Clamp affords the opportunity to explore parameter ranges and perturbations that are out of reach of traditional experimental designs. It is a step forward for investigations aimed at understanding complex social behavior.”
The study shows that behavioral interaction and emotion are continuously feeding from each other, so that coordination of movement could make useful contribution to the rehabilitation of diseases. Movement coordination disorders are often found in patients with schizophrenia and autism spectrum disorders, who also suffer from social and emotional dysfunctions.
“Artificial Intelligence has been grounded in an algorithmic approach of human cognition. We are now bringing the social and emotional dimensions to the table as well,” said Guillaume Dumas, Ph.D., co-author, and a former post-doctoral member of FAU’s CCSBS who is currently with the Institut Pasteur in Paris.
The researchers anticipate that the virtual partner will soon be developed into the prototype of a cooperative machine that can be used for therapeutic purposes. This type of application might benefit many patients afflicted with social and emotional disorders.
“FAU has nurtured the Center for Complex Systems and Brain Sciences for 30 years, and this work is led by one of our outstanding doctoral students who is advancing our understanding of the orders and disorders that take place in our society and in our brains,” said Daniel C. Flynn, Ph.D., FAU’s vice president for research.The study was supported by the National Institute of Mental Health’s (NIMH) “Affect, Social Behavior and Social Cognition Program.”
Article: Enhanced emotional responses during social coordination with a virtual partner, Mengsen Zhang, Guillaume Dumas, J.A. Scott Kelso, Emmanuelle Tognoli, International Journal of Psychophysiology, doi:10.1016/j.ijpsycho.2016.04.001, published online 16 April 2016.