计算机科学
卷积神经网络
活动识别
可穿戴计算机
特征(语言学)
人工智能
传感器融合
模态(人机交互)
可靠性(半导体)
深度学习
模式识别(心理学)
机器学习
嵌入式系统
量子力学
语言学
物理
哲学
功率(物理)
作者
Md. Milon Islam,Sheikh Nooruddin,Fakhri Karray,Ghulam Muhammad
标识
DOI:10.1016/j.inffus.2023.01.015
摘要
Human Activity Recognition (HAR) has become a crucial element for smart healthcare applications due to the fast adoption of wearable sensors and mobile technologies. Most of the existing human activity recognition frameworks deal with a single modality of data that degrades the reliability and recognition accuracy of the system for heterogeneous data sources. In this article, we propose a multi-level feature fusion technique for multimodal human activity recognition using multi-head Convolutional Neural Network (CNN) with Convolution Block Attention Module (CBAM) to process the visual data and Convolutional Long Short Term Memory (ConvLSTM) for dealing with the time-sensitive multi-source sensor information. The architecture is developed to be able to analyze and retrieve channel and spatial dimension features through the use of three branches of CNN along with CBAM for visual information. The ConvLSTM network is designed to capture temporal features from the multiple sensors’ time-series data for efficient activity recognition. An open-access multimodal HAR dataset named UP-Fall detection dataset is utilized in experiments and evaluations to measure the performance of the developed fusion architecture. Finally, we deployed an Internet of Things (IoT) system to test the proposed fusion network in real-world smart healthcare application scenarios. The findings from the experimental results reveal that the developed multimodal HAR framework surpasses the existing state-of-the-art methods in terms of multiple performance metrics.
科研通智能强力驱动
Strongly Powered by AbleSci AI