计算机科学
人工智能
RGB颜色模型
任务(项目管理)
机器学习
比例(比率)
深度学习
可视化
代表(政治)
动作识别
班级(哲学)
动作(物理)
模式识别(心理学)
任务分析
人工神经网络
地理
地图学
政治
量子力学
物理
经济
管理
法学
政治学
作者
Amir Shahroudy,Jun Liu,Tian-Tsong Ng,Gang Wang
标识
DOI:10.1109/cvpr.2016.115
摘要
Recent approaches in depth-based human activity analysis achieved outstanding performance and proved the effectiveness of 3D representation for classification of action classes. Currently available depth-based and RGB+Dbased action recognition benchmarks have a number of limitations, including the lack of training samples, distinct class labels, camera views and variety of subjects. In this paper we introduce a large-scale dataset for RGB+D human action recognition with more than 56 thousand video samples and 4 million frames, collected from 40 distinct subjects. Our dataset contains 60 different action classes including daily, mutual, and health-related actions. In addition, we propose a new recurrent neural network structure to model the long-term temporal correlation of the features for each body part, and utilize them for better action classification. Experimental results show the advantages of applying deep learning methods over state-of-the-art handcrafted features on the suggested cross-subject and cross-view evaluation criteria for our dataset. The introduction of this large scale dataset will enable the community to apply, develop and adapt various data-hungry learning techniques for the task of depth-based and RGB+D-based human activity analysis.
科研通智能强力驱动
Strongly Powered by AbleSci AI