计算机科学
命名实体识别
人工智能
平滑的
自然语言处理
水准点(测量)
任务(项目管理)
背景(考古学)
编码(集合论)
构造(python库)
模式识别(心理学)
过程(计算)
实体链接
蒸馏
机器学习
情报检索
化学
程序设计语言
计算机视觉
地理
生物
经济
集合(抽象数据类型)
大地测量学
有机化学
管理
古生物学
知识库
作者
Zepeng Li,Shuo Cao,Minyu Zhai,Nengneng Ding,Zhenwen Zhang,Bin Hu
出处
期刊:Neurocomputing
[Elsevier]
日期:2024-04-05
卷期号:586: 127637-127637
被引量:5
标识
DOI:10.1016/j.neucom.2024.127637
摘要
As an important foundational task in the field of natural language processing, the Chinese named entity recognition (NER) task has received widespread attention in recent years. Self-distillation plays a role in exploring the potential of the knowledge carried by internal parameters in the BERT NER model, but few studies have noticed the impact of different granularity semantic information during the distillation process. In this paper, we propose a multi-level semantic enhancement approach based on self-distillation BERT for Chinese named entity recognition. We first design a feasible data augmentation method to improve the training quality for handling complex entity compositions, then construct a boundary smoothing module to achieve the model's moderate learning on entity boundaries. Besides, we utilize the distillation reweighting method to let the model acquire balanced entity and context knowledge. Experimental results on two Chinese named entity recognition benchmark datasets Weibo and Resume have 72.09% and 96.93% F1 scores, respectively. Compared to three different basic distillation BERT models, our model can also produce better results. The source code is available at https://github.com/lookmedandan/MSE.
科研通智能强力驱动
Strongly Powered by AbleSci AI