对抗制
可解释性
计算机科学
透视图(图形)
人工智能
稳健性(进化)
深层神经网络
深度学习
数据科学
管理科学
机器学习
工程类
生物化学
基因
化学
作者
Sicong Han,Chenhao Lin,Chao Shen,Qian Wang,Xiaohong Guan
出处
期刊:ACM Computing Surveys
[Association for Computing Machinery]
日期:2023-04-28
卷期号:55 (14s): 1-38
被引量:30
摘要
Deep learning technology is increasingly being applied in safety-critical scenarios but has recently been found to be susceptible to imperceptible adversarial perturbations. This raises a serious concern regarding the adversarial robustness of deep neural network (DNN)–based applications. Accordingly, various adversarial attacks and defense approaches have been proposed. However, current studies implement different types of attacks and defenses with certain assumptions. There is still a lack of full theoretical understanding and interpretation of adversarial examples. Instead of reviewing technical progress in adversarial attacks and defenses, this article presents a framework consisting of three perspectives to discuss recent works focusing on theoretically explaining adversarial examples comprehensively. In each perspective, various hypotheses are further categorized and summarized into several subcategories and introduced systematically. To the best of our knowledge, this study is the first to concentrate on surveying existing research on adversarial examples and adversarial robustness from the interpretability perspective. By drawing on the reviewed literature, this survey characterizes current problems and challenges that need to be addressed and highlights potential future research directions to further investigate adversarial examples.
科研通智能强力驱动
Strongly Powered by AbleSci AI