适应(眼睛)
动画
面子(社会学概念)
计算机人脸动画
风格(视觉艺术)
计算机科学
多媒体
心理学
人机交互
计算机动画
艺术
视觉艺术
计算机图形学(图像)
语言学
哲学
神经科学
作者
Xukun Zhou,Fengxin Li,Ziqiao Peng,Kejian Wu,Jun He,Biao Qin,Zhaoxin Fan,Hongyan Liu
出处
期刊:Cornell University - arXiv
日期:2024-08-18
标识
DOI:10.48550/arxiv.2408.09357
摘要
Audio-driven 3D face animation is increasingly vital in live streaming and augmented reality applications. While remarkable progress has been observed, most existing approaches are designed for specific individuals with predefined speaking styles, thus neglecting the adaptability to varied speaking styles. To address this limitation, this paper introduces MetaFace, a novel methodology meticulously crafted for speaking style adaptation. Grounded in the novel concept of meta-learning, MetaFace is composed of several key components: the Robust Meta Initialization Stage (RMIS) for fundamental speaking style adaptation, the Dynamic Relation Mining Neural Process (DRMN) for forging connections between observed and unobserved speaking styles, and the Low-rank Matrix Memory Reduction Approach to enhance the efficiency of model optimization as well as learning style details. Leveraging these novel designs, MetaFace not only significantly outperforms robust existing baselines but also establishes a new state-of-the-art, as substantiated by our experimental results.
科研通智能强力驱动
Strongly Powered by AbleSci AI