变压器
计算机科学
氙气
加速
现场可编程门阵列
编码器
计算机体系结构
建筑
嵌入式系统
并行计算
计算机硬件
电气工程
工程类
操作系统
艺术
视觉艺术
电压
作者
Γεώργιος Τζάνος,Christoforos Kachris,Dimitrios Soudris
标识
DOI:10.1109/pacet56979.2022.9976354
摘要
Natural Language Processing (NLP) allows program computers to process and analyze large amounts of natural language data. In the last few years, NLP has shown tremendous growth, with many organizations presenting models such as BERT (Bidirectional Encoder Representations from Transformers), GPT2 (Generative Pre-trained Transformer 2), GPT3 (Generative Pre-trained Transformer 3), etc. The cornerstone of these models and the reason for the growth of the NLP is mainly due to the Transformer networks. However, very few architectures have been presented, for the acceleration of the Transformer networks using FPGAs. In this paper, we propose a novel architecture for Transformer Networks optimized on FPGAs. The performance evaluation on a Xilinx Alveo U200 FPGA achieved up to 80.5x speed-up over a single-core CPU and up to 2.3x speedup over a 40-thread Xeon CPU running BERT model.
科研通智能强力驱动
Strongly Powered by AbleSci AI