计算机科学
对抗制
人工智能
机器学习
多任务学习
初始化
语言模型
规范化(社会学)
嵌入
自然语言处理
语音识别
一般化
任务(项目管理)
数学分析
数学
管理
社会学
人类学
程序设计语言
经济
作者
Yaqi Chen,Dan Qu,Wenlin Zhang,Fen Yu,Hao Zhang,Xukui Yang
标识
DOI:10.1109/acait56212.2022.10137854
摘要
Low-resource automatic speech recognition is a chal- lenging task. To solve this issue, multilingual meta-learning learns a better model initialization from many source language tasks, allowing for rapid adaption to the target language. However, due to the lack of limitations on multilingual pre-training, the shared semantic space of different languages is difficult to learn. In this work, we propose an adversarial meta-learning training approach to solve this problem. By using the adversarial auxiliary aim of language identification in the meta-learning algorithm, it will guide the model encoder to generate language-independent embedding features, which can improve model generalization. And we use Wasserstein distance and temporal normalization to optimize our adversarial training, making the training more stable and easier. The approach is evaluated on the IARPA BABEL. The results reveal that our approach only requires half as many meta learning training epochs to attain comparable multilingual pre-training performance. It also outperforms the meta learning in all target languages fine-tuning and achieves comparable performance in small data scales. Specially, it can reduce CER from 71% to 62% with fine-tuning 25% of Vietnamese data. Finally, we show why our approach is superior than others by using t-SNE.
科研通智能强力驱动
Strongly Powered by AbleSci AI