计算机科学
变压器
推论
人工智能
语言模型
可扩展性
适应性
自然语言处理
答疑
机器学习
催交
深度学习
工程类
数据库
生物
电气工程
电压
系统工程
生态学
作者
Rania Kora,Ammar Mohammed
标识
DOI:10.1109/miucc58832.2023.10278387
摘要
The rapid progress in deep learning has propelled transformer-based models to the forefront, establishing them as leading solutions for a multiple NLP tasks. These tasks span a wide spectrum, encompassing text classification activities like sentiment analysis, question answering, natural language inference, and news classification. Transformers offer numerous noteworthy benefits compared to alternative language modeling strategies. They possess a remarkable capacity for parallel processing, enabling them to handle multiple segments of a sequence concurrently, thereby significantly expediting both training and inference. Furthermore, transformers excel in capturing extensive contextual relationships within text, thereby enhancing their comprehension of the broader context and resulting in the generation of more logically coherent textual content. Additionally, transformers exhibit heightened adaptability and scalability, simplifying their utilization across various tasks. Hence, this review paper aims to analyze in-depth the most popular transformer techniques specifically applied to text classification tasks: the various BERT models. We also provide an overview of more than 109 research efforts that used various BERT models in text classification, determining the methodology used. In addition, we describe the three variants of the transformer model with an illustration of the most common models for each variant and the appropriate NLP tasks for each variant.
科研通智能强力驱动
Strongly Powered by AbleSci AI