计算机科学
情态动词
情绪识别
融合
人工智能
传感器融合
语音识别
模式识别(心理学)
语言学
化学
哲学
高分子化学
作者
Lina Li,Wenjie Deng,Shengli Liao,Qiang Xue,Yuying Rong,Ying Yang,Shixuan Liu,Yumei Zhang
摘要
Emotion recognition plays a vital role in human-computer interaction. However, traditional approaches relying on manual features extraction can get high accuracy results but limited generalization; furthermore results of emotion recognition using a single modality is unreliable. To address these challenges, a multi-modal emotion recognition model called STM-Net, which leverages the spatial and temporal information from electroencephalography (EEG) and eye movement two modalities is proposed based on convolutional neural networks (CNNs) and long short-term memory networks (LSTMs). A CNN-LSTM based model is designed to learn the spatial-temporal features of emotions in EEG signals. For eye-tracking signals, a corresponding CNN-based model is designed for feature extraction. The features from both modalities are fused and fed into a fully connected network for classification, then comprehensive and accurate emotion recognition results are obtained. Experimental results on the SEED-IV multimodal dataset demonstrate the effectiveness of the proposed approach, with accuracy results reaching up to 99.6%, much higher than other similar multimodal emotion recognition models.
科研通智能强力驱动
Strongly Powered by AbleSci AI