稳健性(进化)
人工智能
计算机科学
模式识别(心理学)
上下文图像分类
训练集
连接(主束)
计算机视觉
机器学习
图像(数学)
数学
生物化学
化学
几何学
基因
作者
Ali Karkehabadi,Parisa Oftadeh,Danial Shafaie,Jamshid Hassanpour
标识
DOI:10.1109/icicip60808.2024.10477811
摘要
Although Deep Neural Networks (DNNs) have established a solid place for themselves in different applications, their mysterious inner working impedes their usage in sensitive applications. Interpretability-based methods try to overcome this issue by providing explanations for the models. Saliency Guided Training (SGT) is such method that directs the model's focus toward the most relevant features. This technique enhances the clarity of saliency maps, aiding in a better understanding of the model's decision-making. This research investigates the robustness SGT algorithm against adversarial attacks. Although saliency-guided training promises enhanced interpretability for a reliable application of DNNs, our investigation shows that this method increases the model's vulnerability against adversarial attacks. This study underscores the pressing necessity for researchers to achieve an equilibrium between clarity of interpretation and defense against adversarial interventions. Also, the outcome shows a need for attention when deploying saliency-based DNNs in different applications. We employ diverse architectures such as a conventional CNN, ResNet-18, and the Tiny Transformer on popular datasets such as MNIST, CIFAR-10, CIFAR-100, and Caltech101 to substantiate our conclusion.
科研通智能强力驱动
Strongly Powered by AbleSci AI