量化(信号处理)
计算机科学
舍入
人工神经网络
人工智能
蒸馏
机器学习
算法
操作系统
有机化学
化学
作者
Zehua Pei,Xufeng Yao,Wenqian Zhao,Bei Yu
标识
DOI:10.1109/tnnls.2023.3300309
摘要
Quantization is a critical technique employed across various research fields for compressing deep neural networks (DNNs) to facilitate deployment within resource-limited environments. This process necessitates a delicate balance between model size and performance. In this work, we explore knowledge distillation (KD) as a promising approach for improving quantization performance by transferring knowledge from high-precision networks to low-precision counterparts. We specifically investigate feature-level information loss during distillation and emphasize the importance of feature-level network quantization perception. We propose a novel quantization method that combines feature-level distillation and contrastive learning to extract and preserve more valuable information during the quantization process. Furthermore, we utilize the hyperbolic tangent function to estimate gradients with respect to the rounding function, which smoothens the training procedure. Our extensive experimental results demonstrate that the proposed approach achieves competitive model performance with the quantized network compared to its full-precision counterpart, thus validating its efficacy and potential for real-world applications.
科研通智能强力驱动
Strongly Powered by AbleSci AI