XNOR门
现场可编程门阵列
计算机科学
查阅表格
人工神经网络
推论
冗余(工程)
计算机工程
深层神经网络
修剪
计算机体系结构
嵌入式系统
计算机硬件
逻辑门
算法
人工智能
与非门
操作系统
生物
程序设计语言
农学
作者
Erwei Wang,James J. Davis,Peter Y. K. Cheung,George A. Constantinides
出处
期刊:Cornell University - arXiv
日期:2019-01-01
被引量:6
标识
DOI:10.48550/arxiv.1904.00938
摘要
Research has shown that deep neural networks contain significant redundancy, and that high classification accuracies can be achieved even when weights and activations are quantised down to binary values. Network binarisation on FPGAs greatly increases area efficiency by replacing resource-hungry multipliers with lightweight XNOR gates. However, an FPGA's fundamental building block, the K-LUT, is capable of implementing far more than an XNOR: it can perform any K-input Boolean operation. Inspired by this observation, we propose LUTNet, an end-to-end hardware-software framework for the construction of area-efficient FPGA-based neural network accelerators using the native LUTs as inference operators. We demonstrate that the exploitation of LUT flexibility allows for far heavier pruning than possible in prior works, resulting in significant area savings while achieving comparable accuracy. Against the state-of-the-art binarised neural network implementation, we achieve twice the area efficiency for several standard network models when inferencing popular datasets. We also demonstrate that even greater energy efficiency improvements are obtainable.
科研通智能强力驱动
Strongly Powered by AbleSci AI