计算机科学
水准点(测量)
弹丸
人工智能
训练集
光学(聚焦)
选择(遗传算法)
命名实体识别
机器学习
培训(气象学)
一次性
自然语言处理
模式识别(心理学)
任务(项目管理)
工程类
化学
物理
大地测量学
有机化学
光学
地理
机械工程
系统工程
气象学
作者
Ganghong Huang,Zhong Ping Jiang,Chen Wang,Qizhu Dai,Rongzhen Li
标识
DOI:10.1007/978-3-031-10989-8_8
摘要
Exploiting unlabeled data is one of the plausible methods to improve few-shot named entity recognition (few-shot NER), where only a small number of labeled examples are given for each entity type. Existing works focus on learning deep NER models with self-training for few-shot NER. Self-training may induce incomplete and noisy labels which do not necessarily improve or even deteriorate the model performance. To address this challenge, we propose a prompt-based self-training framework. In the first stage, we introduce a self-training approach with prompt tuning to improve the model performance. Specially, we explore several label selection strategies in self-training to mitigate error propagation from noisy pseudo-labels. In the second stage, we fine-tune the BERT model over the high confidence pseudo-labels and original labels. We conduct experiments on two benchmark datasets. The results show that our method outperforms existing few-shot NER models by significant margins, demonstrating its effectiveness for the few-shot setting.
科研通智能强力驱动
Strongly Powered by AbleSci AI