对抗制
计算机科学
初始化
推论
人工智能
嵌入
黑匣子
可微函数
机器学习
理论计算机科学
数学
数学分析
程序设计语言
作者
Zhaorong Liu,Xi Xiong,Yuanyuan Li,Yuanyuan Wang,Jiazhong Lu,Shuai Zhang,Fei Xiong
标识
DOI:10.1016/j.neunet.2024.106461
摘要
Hard-label black-box textual adversarial attacks present a highly challenging task due to the discrete and non-differentiable nature of text data and the lack of direct access to the model's predictions. Research in this issue is still in its early stages, and the performance and efficiency of existing methods has potential for improvement. For instance, exchange-based and gradient-based attacks may become trapped in local optima and require excessive queries, hindering the generation of adversarial examples with high semantic similarity and low perturbation under limited query conditions. To address these issues, we propose a novel framework called HyGloadAttack (adversarial Attacks via Hybrid optimization and Global random initialization) for crafting high-quality adversarial examples. HyGloadAttack utilizes a perturbation matrix in the word embedding space to find nearby adversarial examples after global initialization and selects synonyms that maximize similarity while maintaining adversarial properties. Furthermore, we introduce a gradient-based quick search method to accelerate the search process of optimization. Extensive experiments on five datasets of text classification and natural language inference, as well as two real APIs, demonstrate the significant superiority of our proposed HyGloadAttack method over state-of-the-art baseline methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI