系列(地层学)
蒸馏
回归
回归分析
时间序列
计量经济学
计算机科学
统计
数学
机器学习
色谱法
地质学
化学
古生物学
作者
Qing Xu,Keyu Wu,Min Wu,Kezhi Mao,Xiaoli Li,Zhenghua Chen
出处
期刊:IEEE transactions on artificial intelligence
[Institute of Electrical and Electronics Engineers]
日期:2024-06-01
卷期号:5 (6): 3184-3194
标识
DOI:10.1109/tai.2023.3341854
摘要
As one of the most popular and effective methods in model compression, knowledge distillation (KD) attempts to transfer knowledge from single or multiple large-scale networks (i.e., Teachers ) to a compact network (i.e., Student ). For the multi-teacher scenario, existing methods either assign equal or fixed weights for different teacher models during distillation, which can be inefficient as teachers might perform variously or even oppositely on different training samples. To address this issue, we propose a novel reinforced knowledge distillation method with negatively correlated teachers which are generated via negative correlation learning. The negatively correlated teachers would encourage teachers to learn different aspects of data and thus the ensemble of them can be more comprehensive and suitable for multi-teacher KD. Subsequently, a reinforced KD algorithm is proposed to dynamically employ proper teachers for different training instances via dueling Double Deep Q-Network (DDQN). Our proposed method complements the existing KD procedure on teacher generation and selection. Extensive experimental results on two real-world time series regression tasks clearly demonstrate that the proposed approach could achieve superior performance over state-of-the-art methods. The PyTorch implementation of our proposed approach is available at https://github.com/xuqing88/RL-KD-for-time-series-regression .
科研通智能强力驱动
Strongly Powered by AbleSci AI