Softmax函数
计算
变压器
计算机科学
数学
算法
离散数学
算术
人工神经网络
人工智能
电气工程
工程类
电压
作者
Zhengyu Mei,Hongxi Dong,Yuxuan Wang,Hongbing Pan
出处
期刊:IEEE Transactions on Circuits and Systems Ii-express Briefs
[Institute of Electrical and Electronics Engineers]
日期:2023-04-10
卷期号:70 (9): 3594-3598
被引量:1
标识
DOI:10.1109/tcsii.2023.3265710
摘要
With the popularity of Transformer neural networks, it is inevitable for hardware accelerators to perform nonlinear computation mainly based on the softmax operation. However, a better compromise between the algorithm performance and hardware overhead is always a constant challenge. Hence, this brief advances a tiny and efficient architecture named TEA-S to implement the softmax function with the universal approximation scheme based on Piecewise Linear Approximation Computation (PLAC). With the first co-optimization of calculation and memory, TEA-S can better achieve the design goals of the tiny area and high efficiency. The implementation results show that the peak efficiency of processing 8-bit quantized data will be 487.51 Gps/(mm $^{{2}}{\cdot }$ mW) with the tiny area of $3052.43~{\mu }{\mathrm {m}}^{2}$ at the frequency of 0.5 GHz under 90-nm CMOS technology. Moreover, TEA-S can offer the universal solution to any lengths of input sequences, providing negligible accuracy loss in Transformers compared to the quantized baselines.
科研通智能强力驱动
Strongly Powered by AbleSci AI