计算机科学
人工智能
卷积神经网络
文字嵌入
嵌入
自然语言处理
词(群论)
F1得分
语义学(计算机科学)
深度学习
精确性和召回率
模式识别(心理学)
数学
几何学
程序设计语言
作者
Chao Yang,Xiaotian Wang,Mengyu Li,Li Ji
标识
DOI:10.1109/iccea58433.2023.10135222
摘要
With respect to short texts with high information content, unstructured and non-standard, a text classification model (BERT-CNN-BiLSTM) based on the fusion of BERT model and BiLSTM network with convolutional neural network is proposed. To improve data processing efficiency and classification precision, word vectors are trained in BERT and used as the embedding layer of the model. The embedding layer is utilized to retain semantic information and the semantic representation of words is enhanced. CNN is applied to extract the local semantics of text. Meanwhile, gated linear unit (GLU) is used to optimise the CNN, and gradient dispersion is reduced. BiLSTM is designed to acquire contextual information about the text. Text classification is better implemented. The experimental results show that better results are obtained by BERT training data as word vectors. The BERT-CNN-BiLSTM has significantly improved in terms of classification precision, recall and F1 than the CNN, the BERT-CNN, et al. Precision, recall and F1 values are improved by at least 1.44%, 1.66% and 1.69%, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI