计算机科学
脑电图
解码方法
人工智能
变压器
模式识别(心理学)
特征(语言学)
语音识别
工程类
电气工程
神经科学
电信
电压
心理学
语言学
哲学
作者
Yonghao Song,Xueyu Jia,Lie Yang,Longhan Xie
出处
期刊:Cornell University - arXiv
日期:2021-01-01
被引量:70
标识
DOI:10.48550/arxiv.2106.11170
摘要
At present, people usually use some methods based on convolutional neural networks (CNNs) for Electroencephalograph (EEG) decoding. However, CNNs have limitations in perceiving global dependencies, which is not adequate for common EEG paradigms with a strong overall relationship. Regarding this issue, we propose a novel EEG decoding method that mainly relies on the attention mechanism. The EEG data is firstly preprocessed and spatially filtered. And then, we apply attention transforming on the feature-channel dimension so that the model can enhance more relevant spatial features. The most crucial step is to slice the data in the time dimension for attention transforming, and finally obtain a highly distinguishable representation. At this time, global averaging pooling and a simple fully-connected layer are used to classify different categories of EEG data. Experiments on two public datasets indicate that the strategy of attention transforming effectively utilizes spatial and temporal features. And we have reached the level of the state-of-the-art in multi-classification of EEG, with fewer parameters. As far as we know, it is the first time that a detailed and complete method based on the transformer idea has been proposed in this field. It has good potential to promote the practicality of brain-computer interface (BCI). The source code can be found at: \textit{https://github.com/anranknight/EEG-Transformer}.
科研通智能强力驱动
Strongly Powered by AbleSci AI