亲爱的研友该休息了!由于当前在线用户较少,发布求助请尽量完整地填写文献信息,科研通机器人24小时在线,伴您度过漫漫科研夜!身体可是革命的本钱,早点休息,好梦!

TDFNet: Transformer-Based Deep-Scale Fusion Network for Multimodal Emotion Recognition

计算机科学 深度学习 人工智能 变压器 多模式学习 情感计算 情绪识别 深信不疑网络 特征学习 机器学习 工程类 电气工程 电压
作者
Zhengdao Zhao,Yuhua Wang,guang ze shen,Yuezhu Xu,Jiayuan Zhang
出处
期刊:IEEE/ACM transactions on audio, speech, and language processing [Institute of Electrical and Electronics Engineers]
卷期号:31: 3771-3782 被引量:26
标识
DOI:10.1109/taslp.2023.3316458
摘要

As deep learning technology research continues to progress, artificial intelligence technology is gradually empowering various fields. To achieve a more natural human-computer interaction experience, how to accurately recognize emotional state of speech interactions has become a new research hotspot. Sequence modeling methods based on deep learning techniques have promoted the development of emotion recognition, but the mainstream methods still suffer from insufficient multimodal information interaction, difficulty in learning emotion-related features, and low recognition accuracy. In this paper, we propose a transformer-based deep-scale fusion network (TDFNet) for multimodal emotion recognition, solving the aforementioned problems. The multimodal embedding (ME) module in TDFNet uses pretrained models to alleviate the data scarcity problem by providing a priori knowledge of multimodal information to the model with the help of a large amount of unlabeled data. In addition, a mutual transformer (MT) module is introduced to learn multimodal emotional commonality and speaker-related emotional features to improve contextual emotional semantic understanding. In addition, we design a novel emotion feature learning method named the deep-scale transformer (DST), which further improves emotion recognition by aligning multimodal features and learning multiscale emotion features through GRUs with shared weights. To comparatively evaluate the performance of TDFNet, experiments are conducted with the IEMOCAP corpus under three reasonable data splitting strategies. The experimental results show that TDFNet achieves 82.08% WA and 82.57% UA in RA data splitting, which leads to 1.78% WA and 1.17% UA improvements over the previous state-of-the-art method, respectively. Benefiting from the attentively aligned mutual correlations and fine-grained emotion-related features, TDFNet successfully achieves significant improvements in multimodal emotion recognition.
最长约 10秒,即可获得该文献文件

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
传奇3应助甜美尔风采纳,获得10
3秒前
29秒前
31秒前
甜美尔风发布了新的文献求助10
34秒前
anne发布了新的文献求助10
34秒前
康康XY完成签到 ,获得积分10
36秒前
传奇3应助shinn采纳,获得10
55秒前
威武的晋鹏完成签到,获得积分10
55秒前
肖战战完成签到 ,获得积分10
59秒前
Owen应助威武的晋鹏采纳,获得30
59秒前
1分钟前
1分钟前
1分钟前
anne发布了新的文献求助10
1分钟前
1分钟前
冷静难破发布了新的文献求助10
1分钟前
王誉霖发布了新的文献求助10
1分钟前
1分钟前
shinn发布了新的文献求助10
1分钟前
一粟完成签到 ,获得积分10
1分钟前
shinn发布了新的文献求助10
1分钟前
1分钟前
zqq完成签到,获得积分0
1分钟前
1分钟前
newplayer发布了新的文献求助60
1分钟前
1分钟前
Genetrix应助科研通管家采纳,获得10
1分钟前
滕皓轩完成签到 ,获得积分10
1分钟前
yhgz完成签到,获得积分10
1分钟前
研友_8yN60L完成签到,获得积分10
1分钟前
1分钟前
王誉霖完成签到,获得积分10
1分钟前
小马甲应助andrew12399采纳,获得10
1分钟前
zcq完成签到 ,获得积分10
1分钟前
JamesPei应助huangsi采纳,获得10
1分钟前
王撑撑发布了新的文献求助10
1分钟前
枝头树上的布谷鸟完成签到 ,获得积分10
2分钟前
2分钟前
SARS发布了新的文献求助10
2分钟前
andrew12399完成签到,获得积分10
2分钟前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Kinesiophobia : a new view of chronic pain behavior 2000
Research for Social Workers 1000
Kinesiophobia : a new view of chronic pain behavior 600
Signals, Systems, and Signal Processing 510
Discrete-Time Signals and Systems 510
Psychology and Work Today 500
热门求助领域 (近24小时)
化学 材料科学 生物 医学 工程类 计算机科学 有机化学 物理 生物化学 纳米技术 复合材料 内科学 化学工程 人工智能 催化作用 遗传学 数学 基因 量子力学 物理化学
热门帖子
关注 科研通微信公众号,转发送积分 5893356
求助须知:如何正确求助?哪些是违规求助? 6682592
关于积分的说明 15724435
捐赠科研通 5015012
什么是DOI,文献DOI怎么找? 2701122
邀请新用户注册赠送积分活动 1646893
关于科研通互助平台的介绍 1597471