计算机科学
运动捕捉
姿势
人工智能
基本事实
计算机视觉
凝视
RGB颜色模型
人机交互
匹配移动
增强现实
钥匙(锁)
多边形网格
运动(物理)
水准点(测量)
模式
计算机图形学(图像)
社会科学
计算机安全
大地测量学
社会学
地理
作者
Siwei Zhang,Qianli Ma,Yan Zhang,Qian Zhi-yin,Taein Kwon,Marc Pollefeys,Federica Bogo,Siyu Tang
标识
DOI:10.1007/978-3-031-20068-7_11
摘要
Understanding social interactions from egocentric views is crucial for many applications, ranging from assistive robotics to AR/VR. Key to reasoning about interactions is to understand the body pose and motion of the interaction partner from the egocentric view. However, research in this area is severely hindered by the lack of datasets. Existing datasets are limited in terms of either size, capture/annotation modalities, ground-truth quality, or interaction diversity. We fill this gap by proposing EgoBody, a novel large-scale dataset for human pose, shape and motion estimation from egocentric views, during interactions in complex 3D scenes. We employ Microsoft HoloLens2 headsets to record rich egocentric data streams (including RGB, depth, eye gaze, head and hand tracking). To obtain accurate 3D ground truth, we calibrate the headset with a multi-Kinect rig and fit expressive SMPL-X body meshes to multi-view RGB-D frames, reconstructing 3D human shapes and poses relative to the scene, over time. We collect 125 sequences, spanning diverse interaction scenarios, and propose the first benchmark for 3D full-body pose and shape estimation of the interaction partner from egocentric views. We extensively evaluate state-of-the-art methods, highlight their limitations in the egocentric scenario, and address such limitations leveraging our high-quality annotations. Data and code are available at https://sanweiliti.github.io/egobody/egobody.html .
科研通智能强力驱动
Strongly Powered by AbleSci AI