背景(考古学)
计算机科学
表达式(计算机科学)
情感表达
认知心理学
可转让性
模式治疗法
人机交互
心理学
心理治疗师
机器学习
古生物学
罗伊特
生物
程序设计语言
作者
C Liu,Zheyong Xie,Sirui Zhao,Jin Zhou,Tong Xu,Minglei Li,Enhong Chen
标识
DOI:10.1145/3652583.3658104
摘要
Recent advancements in Large Language Models~(LLMs) have greatly enhanced the generation capabilities of dialogue systems. However, progress on emotional expression during dialogues might be still limited, especially when capturing and processing the multimodal cues for emotional expression. Therefore, it is urgent to fully adapt the multimodal understanding ability and transferability of LLMs to enhance the emotional-oriented multimodal processing capabilities. To that end, in this paper, we propose a novel Emotion-Guided Multimodal Dialogue model based on LLM, termed ELMD. Specifically, to enhance the emotional expression ability of LLMs, our ELMD customizes an emotional retrieval module, which mainly provides appropriate response demonstration for LLM in understanding emotional context. Subsequently, a two-stage training strategy is proposed, founded on previous demonstration support, to support uncovering nuanced emotions behind multimodal information and constructing natural responses. Comprehensive experiments demonstrate the effectiveness and superiority of ELMD.
科研通智能强力驱动
Strongly Powered by AbleSci AI