计算机科学
人工神经网络
推论
计算机体系结构
软件
软件工程
程序设计语言
人工智能
作者
Yu Wang,Xuefei Ning,Shulin Zeng,Yi Cai,Ke Guo,Hanbo Sun,Changcheng Tang,Tianyi Lu,Shuang Liang,Tianchen Zhao
出处
期刊:Chapman and Hall/CRC eBooks
[Informa]
日期:2022-01-12
卷期号:: 55-90
标识
DOI:10.1201/9781003162810-4
摘要
In this chapter, we introduce our efforts in accelerating neural network inference. From the hardware design aspect, we introduce the instructions-set-architecture deep learning accelerator to support all kinds of DNN models with customized ISA and optimized software compiler. And from the algorithm aspect, we introduce several practices we have used: sensitivity-based pruning without hardware model, quantization, iterative pruning with hardware model, and neural architecture search. Take-aways Discusses hardware design: An instructions-set-architecture deep learning accelerator to support all kinds of DNN models with customized ISA and optimized software compile Discusses software practices: Sensitivity-based pruning without hardware model, quantization, iterative pruning with hardware model, neural architecture search.
科研通智能强力驱动
Strongly Powered by AbleSci AI