计算机科学
人工智能
文字嵌入
词(群论)
自然语言处理
背景(考古学)
水准点(测量)
一般化
语言模型
命名实体识别
人工神经网络
钥匙(锁)
图层(电子)
嵌入
地理
任务(项目管理)
生物
经济
有机化学
计算机安全
大地测量学
管理
数学
化学
语言学
哲学
数学分析
古生物学
作者
Hongzhen Cui,Longhao Zhang,Wen Wu,Yunfeng Peng
标识
DOI:10.1109/ijcnn54540.2023.10191631
摘要
Chinese named entity recognition (CNER) is one of the most fundamental tasks in natural language processing (NLP), and is key to extracting information from unstructured texts. In recent years, advances in neural network models and pretrained word-level information embedding techniques have played a driving role in the development of NLP. In this context, how to make full use of word vectors to extract information has become one of the research emphases. The diversity of Chinese expressions and the irregular expressions of texts lead to poor recognition results. This paper proposes a two-layer BiLSTM network model with linear gating logic to enhance the model's learning effect of word vectors within sentences and word memory. The aim is to solve the problem of gradient disappearance and improve the model's generalization ability and entity recognition. Through experiments, our model proved effective on three Chinese benchmark datasets: MSRA, the People's Daily Corpus (PRF), and Boson. The precision of NER performs best among similar models. In addition, using the lab-constructed medical dataset of Chinese Drugs for the Heart for testing, our model outperforms the existing BiLSTM model. Finally, statistical analysis of the changes in F1 during training demonstrated faster convergence of our model.
科研通智能强力驱动
Strongly Powered by AbleSci AI