面部表情
价(化学)
唤醒
心理学
感知
认知心理学
语音识别
视听
情感配价
情感表达
计算机科学
沟通
多媒体
认知
社会心理学
神经科学
物理
量子力学
作者
Ilaria Torre,Simon Holk,Elmira Yadollahi,Iolanda Leite,Rachel McDonnell,Naomi Harte
出处
期刊:IEEE Transactions on Affective Computing
[Institute of Electrical and Electronics Engineers]
日期:2022-10-10
卷期号:15 (2): 393-404
被引量:8
标识
DOI:10.1109/taffc.2022.3213269
摘要
Multisensory integration influences emotional perception, as the McGurk effect demonstrates for the communication between humans. Human physiology implicitly links the production of visual features with other modes like the audio channel: Face muscles responsible for a smiling face also stretch the vocal cords that result in a characteristic smiling voice. For artificial agents capable of multimodal expression, this linkage is modeled explicitly. In our studies, we observe the influence of visual and audio channels on the perception of the agents' emotional expression. We created videos of virtual characters and social robots either with matching or mismatching emotional expressions in the audio and visual channels. In two online studies, we measured the agents' perceived valence and arousal. Our results consistently lend support to the 'emotional McGurk effect' hypothesis, according to which face transmits valence information, and voice transmits arousal. When dealing with dynamic virtual characters, visual information is enough to convey both valence and arousal, and thus audio expressivity need not be congruent. When dealing with robots with fixed facial expressions, however, both visual and audio information need to be present to convey the intended expression.
科研通智能强力驱动
Strongly Powered by AbleSci AI