计算机科学
依赖关系(UML)
人工智能
主题(文档)
语音识别
脑-机接口
自然语言处理
模式识别(心理学)
机器学习
脑电图
心理学
神经科学
图书馆学
作者
Hao Sun,Yi Ding,Jianzhu Bao,Ke Qin,Chengxuan Tong,Jing Jin,Cuntai Guan
标识
DOI:10.1016/j.neunet.2024.106470
摘要
Brain-computer interfaces (BCIs) built based on motor imagery paradigm have found extensive utilization in motor rehabilitation and the control of assistive applications. However, traditional MI-BCI systems often exhibit suboptimal classification performance and require significant time for new users to collect subject-specific training data. This limitation diminishes the user-friendliness of BCIs and presents significant challenges in developing effective subject-independent models. In response to these challenges, we propose a novel subject-independent framework for learning temporal dependency for motor imagery BCIs by Contrastive Learning and Self-attention (CLS). In CLS model, we incorporate self-attention mechanism and supervised contrastive learning into a deep neural network to extract important information from electroencephalography (EEG) signals as features. We evaluate the CLS model using two large public datasets encompassing numerous subjects in a subject-independent experiment condition. The results demonstrate that CLS outperforms six baseline algorithms, achieving a mean classification accuracy improvement of 1.3 % and 4.71 % than the best algorithm on the Giga dataset and OpenBMI dataset, respectively. Our findings demonstrate that CLS can effectively learn invariant discriminative features from training data obtained from non-target subjects, thus showcasing its potential for building models for new users without the need for calibration.
科研通智能强力驱动
Strongly Powered by AbleSci AI