对抗制
计算机科学
分类器(UML)
深层神经网络
人工智能
稳健性(进化)
摄动(天文学)
训练集
上下文图像分类
机器学习
人工神经网络
图像(数学)
生物化学
量子力学
基因
物理
化学
作者
Chaithanya Kumar Mummadi,Thomas Brox,Jan Hendrik Metzen
标识
DOI:10.1109/iccv.2019.00503
摘要
Classifiers such as deep neural networks have been shown to be vulnerable against adversarial perturbations on problems with high-dimensional input space. While adversarial training improves the robustness of image classifiers against such adversarial perturbations, it leaves them sensitive to perturbations on a non-negligible fraction of the inputs. In this work, we show that adversarial training is more effective in preventing universal perturbations, where the same perturbation needs to fool a classifier on many inputs. Moreover, we investigate the trade-off between robustness against universal perturbations and performance on unperturbed data and propose an extension of adversarial training that handles this trade-off more gracefully. We present results for image classification and semantic segmentation to showcase that universal perturbations that fool a model hardened with adversarial training become clearly perceptible and show patterns of the target scene.
科研通智能强力驱动
Strongly Powered by AbleSci AI