二元曲线
计算机科学
命名实体识别
人工智能
卷积神经网络
自然语言处理
特征(语言学)
文字嵌入
背景(考古学)
人工神经网络
词(群论)
语音识别
任务(项目管理)
嵌入
三元曲线
语言学
经济
管理
古生物学
哲学
生物
作者
Kaifang Long,Han Zhao,Zengzhen Shao,Yang Cao,Yanfang Geng,Yintai Sun,Wei Xu,Hui Yu
出处
期刊:ACM Transactions on Asian and Low-Resource Language Information Processing
日期:2023-03-23
卷期号:22 (3): 1-16
摘要
Chinese Named Entity Recognition (NER) is an essential task in natural language processing, and its performance directly impacts the downstream tasks. The main challenges in Chinese NER are the high dependence of named entities on context and the lack of word boundary information. Therefore, how to integrate relevant knowledge into the corresponding entity has become the primary task for Chinese NER. Both the lattice LSTM model and the WC-LSTM model did not make excellent use of contextual information. Additionally, the lattice LSTM model had a complex structure and did not exploit the word information well. To address the preceding problems, we propose a Chinese NER method based on the deep neural network with multiple ways of embedding fusion. First, we use a convolutional neural network to combine the contextual information of the input sequence and apply a self-attention mechanism to integrate lexicon knowledge, compensating for the lack of word boundaries. The word feature, context feature, bigram feature, and bigram context feature are obtained for each character. Second, four different features are used to fuse information at the embedding layer. As a result, four different word embeddings are obtained through cascading. Last, the fused feature information is input to the encoding and decoding layer. Experiments on three datasets show that our model can effectively improve the performance of Chinese NER.
科研通智能强力驱动
Strongly Powered by AbleSci AI