计算机科学
动作识别
RGB颜色模型
人工智能
卷积神经网络
骨架(计算机编程)
图形
块(置换群论)
模式识别(心理学)
光学(聚焦)
理论计算机科学
光学
物理
数学
程序设计语言
班级(哲学)
几何学
作者
Bingkun Gao,Le Dong,Hongbo Bi,Yunze Bi
标识
DOI:10.1007/s10489-021-02723-6
摘要
Graph convolutional networks (GCN) have received more and more attention in skeleton-based action recognition. Many existing GCN models pay more attention to spatial information and ignore temporal information, but the completion of actions must be accompanied by changes in temporal information. Besides, the channel, spatial, and temporal dimensions often contain redundant information. In this paper, we design a temporal graph convolutional network (FTGCN) module which can concentrate more temporal information and properly balance them for each action. In order to better integrate channel, spatial and temporal information, we propose a unified attention model of the channel, spatial and temporal (CSTA). A basic block containing these two novelties is called FTC-GCN. Extensive experiments on two large-scale datasets, compared with 17 methods on NTU-RGB+D and 8 methods on Kinetics-Skeleton, show that for skeleton-based human action recognition, our method achieves the best performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI