稳健性(进化)
计算机科学
对抗制
再培训
水准点(测量)
人工智能
语言模型
一般化
数据建模
机器学习
数据库
数学分析
地理
化学
国际贸易
业务
大地测量学
基因
生物化学
数学
作者
Meizheng Peng,Jia Xu,Min Peng
标识
DOI:10.1109/icassp49357.2023.10096803
摘要
Slot filling is a major problem in spoken language understanding (SLU) task. However, the current SLU models may experience performance degradation when encountering unfamiliar data in different datasets. Meanwhile, as recent models are becoming more complex, retraining the model in a new application scenario would be too expensive. So we think it is important to study the robustness and generalization capability of SLU models. Then we propose the Natural Adversarial Slot Evaluator (NASE), a benchmark with adversarial SLU data to evaluate the robustness and generalization capability of SLU models on the task of slot filling. Our experiments and analysis reveal that all of the six SLU models have a significant performance degradation on NASE. The further analysis points out that the models rely more on the context of the slots than the slot values themselves to make predictions. In addition, the widespread use of joint learning strategy makes unfamiliar intents also affect the slot filling. Based on our findings, we also propose a simple data augmentation method to improve the robustness of SLU models in slot filling. The F1 Scores improve up to about 30% compared to the original model.
科研通智能强力驱动
Strongly Powered by AbleSci AI