计算机科学
人工智能
模式识别(心理学)
图形
特征提取
编码器
RGB颜色模型
计算机视觉
理论计算机科学
操作系统
作者
Jialin Gao,Tong He,Xi Zhou,Shiming Ge
出处
期刊:IEEE Signal Processing Letters
[Institute of Electrical and Electronics Engineers]
日期:2021-01-01
卷期号:28: 2058-2062
被引量:22
标识
DOI:10.1109/lsp.2021.3116513
摘要
Graph Convolutional Networks have been successfully applied in skeleton-based action recognition. The key is fully exploring the spatial-temporal context. This letter proposes a Focusing-Diffusion Graph Convolutional Network (FDGCN) to address this issue. Each skeleton frame is first decomposed into two opposite-direction graphs for subsequent focusing and diffusion processes. Next, the focusing process generates a spatial-level representation for each frame individually by an attention module. This representation is regarded as a supernode to aggregate the feature from each joint node in each frame for spatial context extraction. After generating supernodes for the entire sequence, a transformer encoder layer is proposed to capture the temporal context further. Finally, these supernodes pass the embedded spatial-temporal context back to the spatial joints through the diffusion graph in the diffusing process. Extensive experiments on the NTU RGB+D and Skeleton-Kinetics benchmarks demonstrate the effectiveness of our approach.
科研通智能强力驱动
Strongly Powered by AbleSci AI