Robust knowledge distillation based on feature variance against backdoored teacher model

差异(会计) 计算机科学 蒸馏 特征(语言学) 机器学习 人工智能 工艺工程 化学 色谱法 工程类 语言学 会计 哲学 业务
作者
Jinyin Chen,Xiaoming Zhao,Haibin Zheng,Xiao Li,Sheng Xiang,Haifeng Guo
出处
期刊:Applied Soft Computing [Elsevier]
卷期号:: 111907-111907
标识
DOI:10.1016/j.asoc.2024.111907
摘要

Benefiting from large well-trained deep neural networks (DNNs), model compression has captured special attention for computing resource limited equipment, especially edge devices. Knowledge distillation (KD) is one of the widely used compression techniques for edge deployment, by obtaining a lightweight student model from a well-trained teacher model released on public platforms. However, it has been empirically noticed that the backdoor in the teacher model will be transferred to the student model during the process of KD. Although numerous KD methods have been proposed, most of them focus on the distillation of a high-performing student model without robustness consideration. Besides, some research adopts KD techniques as effective backdoor mitigation tools, but they fail to perform model compression at the same time. Consequently, it is still an open problem to well achieve two objectives of robust KD, i.e., student model's performance and backdoor mitigation. To address these issues, we propose RobustKD, a robust knowledge distillation that compresses the model while mitigating backdoor based on feature variance. Specifically, RobustKD distinguishes the previous works in three key aspects: (1) effectiveness - by distilling the feature map of the teacher model after detoxification, the main task performance of the student model is comparable to that of the teacher model; (2) robustness - by reducing the characteristic variance between the teacher model and the student model, it mitigates the backdoor of the student model under backdoored teacher model scenario; (3) generic - RobustKD still has good performance in the face of multiple data models (e.g., WRN 28-4, Pyramid-200) and diverse DNNs (e.g., ResNet50, MobileNet). Comprehensive experiments are conducted on four datasets, six models, two distillation methods, and two backdoor attack methods, compared with four baselines, and the results verified that the proposed method achieves the state-of-the-art performance in both aspects of accuracy and robustness. In addition, RobustKD is still effective when adaptive attacks are considered. The code of RobustKD is open-sourced at https://github.com/Xming-Z/RobustKD.

科研通智能强力驱动
Strongly Powered by AbleSci AI
科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
yaya发布了新的文献求助10
刚刚
1秒前
HelloFM发布了新的文献求助10
2秒前
2秒前
可燃冰发布了新的文献求助10
2秒前
3秒前
你好完成签到,获得积分10
3秒前
整箱发布了新的文献求助10
4秒前
4秒前
4秒前
所所应助任性的山芙采纳,获得10
4秒前
5秒前
XHH1994完成签到,获得积分10
5秒前
5秒前
6秒前
我是老大应助Khalil采纳,获得10
6秒前
6秒前
岑梨愁发布了新的文献求助10
7秒前
8秒前
Akim应助jia_hui1009采纳,获得10
8秒前
柚子叶完成签到 ,获得积分10
9秒前
10秒前
leo7完成签到,获得积分10
10秒前
今后应助纯真的滑板采纳,获得10
10秒前
顾君如发布了新的文献求助10
10秒前
高高友易发布了新的文献求助20
10秒前
领导范儿应助ndndd采纳,获得10
10秒前
54687完成签到,获得积分10
10秒前
小蘑菇应助平安喜乐采纳,获得10
11秒前
11秒前
宏hong发布了新的文献求助10
11秒前
ttyj完成签到,获得积分20
11秒前
静飞完成签到 ,获得积分10
11秒前
cc发布了新的文献求助10
11秒前
科目三应助含糊的水之采纳,获得10
12秒前
12秒前
无醇大师完成签到,获得积分10
13秒前
李荷月完成签到,获得积分10
13秒前
yunkong关注了科研通微信公众号
13秒前
heiehi发布了新的文献求助10
13秒前
高分求助中
(应助此贴封号)【重要!!请各用户(尤其是新用户)详细阅读】【科研通的精品贴汇总】 10000
Kinesiophobia : a new view of chronic pain behavior 3000
Molecular Biology of Cancer: Mechanisms, Targets, and Therapeutics 1100
Signals, Systems, and Signal Processing 510
Discrete-Time Signals and Systems 510
Proceedings of the Fourth International Congress of Nematology, 8-13 June 2002, Tenerife, Spain 500
Le genre Cuphophyllus (Donk) st. nov 500
热门求助领域 (近24小时)
化学 材料科学 生物 医学 工程类 计算机科学 有机化学 物理 生物化学 纳米技术 复合材料 内科学 化学工程 人工智能 催化作用 遗传学 数学 基因 量子力学 物理化学
热门帖子
关注 科研通微信公众号,转发送积分 5939618
求助须知:如何正确求助?哪些是违规求助? 7050600
关于积分的说明 15879571
捐赠科研通 5069751
什么是DOI,文献DOI怎么找? 2726815
邀请新用户注册赠送积分活动 1685394
关于科研通互助平台的介绍 1612731