强化学习
计算机科学
情绪识别
钢筋
人工智能
语音识别
人机交互
心理学
社会心理学
作者
Chengwen Zhang,Yuhao Zhang,Bo Cheng
标识
DOI:10.1109/icassp48485.2024.10446459
摘要
Multimodal Emotion Recognition in Conversation (ERC) has gained significant attention due to its wide-ranging applications in diverse areas. However, most previous approaches focused on modeling context at the semantic level, neglecting the context of dependency information at the emotional level. In this paper, we proposed a novel Reinforcement Learning framework for the multimodal EMOtion recognition task (RL-EMO), which combines a Multi-modal Graph Convolution Network (MMGCN) [1] module with a novel Reinforcement Learning (RL) module to model context at both the semantic and emotional levels respectively. The RL-EMO approach was evaluated on two widely used multi-modal datasets, IEMOCAP and MELD, and the results show that RL-EMO outperforms several baseline models, achieving significant improvements in F1-score. We release the code at https://github.com/zyh9929/RL-EMO.
科研通智能强力驱动
Strongly Powered by AbleSci AI