iCub
动作(物理)
易读性
凝视
非语言交际
感知
认知心理学
预测(人工智能)
心理学
人机交互
运动(物理)
人机交互
手势
计算机科学
人际交往
眼动
阅读(过程)
机器人
仿人机器人
沟通
人工智能
政治学
法学
神经科学
视觉艺术
量子力学
艺术
物理
作者
Nuno Ferreira Duarte,Mirko Raković,Jovica Tasevski,Moreno I. Coco,Aude Billard,José Santos-Victor
出处
期刊:IEEE robotics and automation letters
日期:2018-10-01
卷期号:3 (4): 4132-4139
被引量:56
标识
DOI:10.1109/lra.2018.2861569
摘要
Humans have the fascinating capacity of processing nonverbal visual cues to understand and anticipate the actions of other humans. This “intention reading” ability is underpinned by shared motor repertoires and action models, which we use to interpret the intentions of others as if they were our own. We investigate how different cues contribute to the legibility of human actions during interpersonal interactions. Our first contribution is a publicly available dataset with recordings of human body motion and eye gaze, acquired in an experimental scenario with an actor interacting with three subjects. From these data, we conducted a human study to analyze the importance of different nonverbal cues for action perception. As our second contribution, we used motion/gaze recordings to build a computational model describing the interaction between two persons. As a third contribution, we embedded this model in the controller of an iCub humanoid robot and conducted a second human study, in the same scenario with the robot as an actor, to validate the model's “intention reading” capability. Our results show that it is possible to model (nonverbal) signals exchanged by humans during interaction, and how to incorporate such a mechanism in robotic systems with the twin goal of being able to “read” human action intentionsand acting in a way that is legible by humans.
科研通智能强力驱动
Strongly Powered by AbleSci AI