学习迁移
计算机科学
人工智能
遗忘
人工神经网络
领域(数学分析)
元学习(计算机科学)
深度学习
领域(数学)
机器学习
语音识别
模式识别(心理学)
任务(项目管理)
数学
心理学
数学分析
经济
认知心理学
管理
纯数学
作者
Zhentao Liu,Bao-Han Wu,Meng-Ting Han,Weihua Cao,Min Wu
标识
DOI:10.1016/j.asoc.2023.110766
摘要
Deep learning often requires large amounts of labeled data to train the model, which is not always readily available in the field of speech emotion recognition (SER). Related research work on SER in few shot conditions has reported problem with overfifitting and domain transfer of training. In this study, a few-shot learning method based on meta-transfer learning with domain adaption (MTLDA) is proposed for SER. It not only effectively reduces the over-fitting phenomenon of deep neural networks (DNN) trained with a small number of samples, but also solves the forgetting problem in meta-learning and the target domain adaptability problem in transfer learning. Experiments on three databases (i.e., CASIA is used for pre-training, Emo-DB and SAVEE are used for few-shot learning) are performed for few-shot learning of SER, from which the WAR is 65.12% and UAR is 64.50% on Emo-DB, and the WAR is 58.84% and UAR is 53.26% on SAVEE.
科研通智能强力驱动
Strongly Powered by AbleSci AI