计算机科学
脑电图
情绪识别
语音识别
卷积神经网络
唤醒
人工智能
深度学习
情绪分类
特征提取
情态动词
音频信号
保险丝(电气)
模式识别(心理学)
人工神经网络
心理学
语音编码
工程类
神经科学
化学
高分子化学
电气工程
精神科
作者
Zhongjie Li,Gaoyan Zhang,Jianwu Dang,Longbiao Wang,Jianguo Wei
标识
DOI:10.1109/ijcnn52387.2021.9533663
摘要
Automatic recognition of human emotional states has attracted many researchers' attention in Human-Computer Interactions and emotional brain-computer interface recently. However, the accuracy of emotion recognition is not satisfying. Considering the advantage of information supplement based on deep learning of multi-modal signals related to emotion, this study proposed a novel emotion recognition architecture to fuse emotional features from brain electroencephalography (EEG) signal and the corresponding audio signal in emotion recognition on DEAP dataset. We used convolutional neural network (CNN) to extract EEG features and bidirectional long short term memory (BiLSTM) neural networks to extract audio features. After that, we combine the multi-modal features into a deep learning architecture to recognize arousal and valence levels. Results showed an improved accuracy compared with previous studies that merely used the EEG signals in both arousal level and valence level, which suggests the effectiveness of our proposed multi-modal fused emotion recognition model. In future work, multi-modal data from nature interaction scenes will be collected and inputted into this architecture to further validate the effectiveness of the method.
科研通智能强力驱动
Strongly Powered by AbleSci AI