Home > Research > Publications & Outputs > Smiling in the Face and Voice of Avatars and Ro...
View graph of relations

Smiling in the Face and Voice of Avatars and Robots: Evidence for a 'Smiling McGurk Effect'.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published
Close
<mark>Journal publication date</mark>31/05/2024
<mark>Journal</mark>IEEE Transactions on Affective Computing
Issue number2
Volume15
Number of pages12
Pages (from-to)393-404
Publication StatusPublished
Early online date10/10/22
<mark>Original language</mark>English

Abstract

Multisensory integration influences emotional perception, as the McGurk effect demonstrates for the communication between humans. Human physiology implicitly links the production of visual features with other modes like the audio channel: Face muscles responsible for a smiling face also stretch the vocal cords that result in a characteristic smiling voice. For artificial agents capable of multimodal expression, this linkage is modeled explicitly. In our studies, we observe the influence of visual and audio channels on the perception of the agents’ emotional expression. We created videos of virtual characters and social robots either with matching or mismatching emotional expressions in the audio and visual channels. In two online studies, we measured the agents’ perceived valence and arousal. Our results consistently lend support to the ‘emotional McGurk effect’ hypothesis, according to which face transmits valence information, and voice transmits arousal. When dealing with dynamic virtual characters, visual information is enough to convey both valence and arousal, and thus audio expressivity need not be congruent. When dealing with robots with fixed facial expressions, however, both visual and audio information need to be present to convey the intended expression.