卷积神经网络
手势
传感器融合
计算机科学
有线手套
计算机视觉
手势识别
深度学习
机器人学
模式识别(心理学)
人工智能
机器人
作者
Ming Wang,Zheng Yan,Ting Wang,Pingqiang Cai,Siyu Gao,Yi Zeng,Changjin Wan,Hong Wang,Liang Pan,Jiancan Yu,Shaowu Pan,Ke He,Jie Lü,Xiaodong Chen
标识
DOI:10.1038/s41928-020-0422-z
摘要
Gesture recognition using machine-learning methods is valuable in the development of advanced cybernetics, robotics and healthcare systems, and typically relies on images or videos. To improve recognition accuracy, such visual data can be combined with data from other sensors, but this approach, which is termed data fusion, is limited by the quality of the sensor data and the incompatibility of the datasets. Here, we report a bioinspired data fusion architecture that can perform human gesture recognition by integrating visual data with somatosensory data from skin-like stretchable strain sensors made from single-walled carbon nanotubes. The learning architecture uses a convolutional neural network for visual processing and then implements a sparse neural network for sensor data fusion and recognition at the feature level. Our approach can achieve a recognition accuracy of 100% and maintain recognition accuracy in non-ideal conditions where images are noisy and under- or over-exposed. We also show that our architecture can be used for robot navigation via hand gestures, with an error of 1.7% under normal illumination and 3.3% in the dark. A bioinspired machine-learning architecture can combine visual data with data from stretchable strain sensors to achieve human gesture recognition with high accuracy in complex environments.
科研通智能强力驱动
Strongly Powered by AbleSci AI