计算机科学
动作识别
骨架(计算机编程)
图形
卷积(计算机科学)
灵活性(工程)
人工智能
模式识别(心理学)
算法
理论计算机科学
人工神经网络
数学
统计
程序设计语言
班级(哲学)
作者
Haodong Duan,Jiaqi Wang,Kai Chen,Dahua Lin
出处
期刊:Cornell University - arXiv
日期:2022-01-01
被引量:24
标识
DOI:10.48550/arxiv.2210.05895
摘要
Graph convolution networks (GCN) have been widely used in skeleton-based action recognition. We note that existing GCN-based approaches primarily rely on prescribed graphical structures (ie., a manually defined topology of skeleton joints), which limits their flexibility to capture complicated correlations between joints. To move beyond this limitation, we propose a new framework for skeleton-based action recognition, namely Dynamic Group Spatio-Temporal GCN (DG-STGCN). It consists of two modules, DG-GCN and DG-TCN, respectively, for spatial and temporal modeling. In particular, DG-GCN uses learned affinity matrices to capture dynamic graphical structures instead of relying on a prescribed one, while DG-TCN performs group-wise temporal convolutions with varying receptive fields and incorporates a dynamic joint-skeleton fusion module for adaptive multi-level temporal modeling. On a wide range of benchmarks, including NTURGB+D, Kinetics-Skeleton, BABEL, and Toyota SmartHome, DG-STGCN consistently outperforms state-of-the-art methods, often by a notable margin.
科研通智能强力驱动
Strongly Powered by AbleSci AI