计算机科学
变压器
卷积神经网络
特征提取
提取器
编码器
关系抽取
人工智能
模式识别(心理学)
循环神经网络
人工神经网络
机器学习
信息抽取
工程类
电压
工艺工程
电气工程
操作系统
标识
DOI:10.1109/ijcnn52387.2021.9534183
摘要
Many pre-trained models recently achieved successful results in many NLP tasks. However, such models with heavy structure often led to problems such as a great amount of training time or too much computational cost. This paper presents a more light-weighted model based on Transformer and convolutional neural networks. The proposed model is able to capture important local information of different length by multi-head convolutional layer and effectively extracts long term dependency by position-enriched self-attention layer. These two structures together formed our Transformer with Local-feature Extractor (TLE) encoder. Comparing with previous CNN, RNN, and attention-based baseline models, our model achieves the best performance on multiple relation extraction datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI