计算机科学
对抗制
稳健性(进化)
深层神经网络
人工智能
边距(机器学习)
机器学习
水准点(测量)
困境
词(群论)
人工神经网络
基因
地理
化学
哲学
认识论
生物化学
语言学
大地测量学
作者
Jiahui Shi,Linjing Li,Daniel Zeng
标识
DOI:10.1016/j.neucom.2022.09.032
摘要
Attacks with adversarial examples can tremendously worsen the performance of deep neural networks (DNNs). Hence, defending against such adversarial attacks is crucial for nearly all DNN-based applications. Adversarial training is an effective and extensively adopted approach for increasing the robustness of DNNs in which benign examples and their adversarial counterparts are considered together in the training stage. However, this may result in a decrease in accuracy on benign examples because it does not account for the inter-class distance of benign examples. To overcome the aforementioned dilemma, we devise a novel defense approach named adversarial supervised contrastive learning (ASCL), which combines adversarial training with supervised contrastive learning to enhance the robustness of DNN-based models while maintaining their clean accuracy. We validate the effectiveness of the proposed ASCL approach in the scenario of defending against word substitution attacks by means of extensive experiments on benchmark tasks and datasets. The experimental results show that ASCL reduces the attack success rate to 20% while maintaining the accuracy for clean inputs within a 2% margin.
科研通智能强力驱动
Strongly Powered by AbleSci AI