论文标题
提高黑盒对抗攻击的查询效率
Improving Query Efficiency of Black-box Adversarial Attack
论文作者
论文摘要
深度神经网络(DNN)在各种任务上表现出了出色的性能,但是当攻击者可以访问目标模型时,它们具有对抗性示例的风险,可以轻松生成。由于已经通过在线服务部署了许多机器学习模型,这些模型仅提供无法访问的模型(例如Google Cloud Vision API2)的查询输出,因此Black-Box对抗攻击(不访问目标模型)在实践中是至关重要的安全问题,而不是白盒。但是,现有的基于查询的黑盒对抗攻击通常需要过多的模型查询以保持高攻击成功率。因此,为了提高查询效率,我们借助于以神经过程为特征的图像结构信息来探讨围绕良性输入的对抗性示例的分布,并在本文中提出了基于神经过程的黑盒对抗性攻击(NP-攻击)。广泛的实验表明,NP攻击可以大大降低黑框设置下的查询计数。
Deep neural networks (DNNs) have demonstrated excellent performance on various tasks, however they are under the risk of adversarial examples that can be easily generated when the target model is accessible to an attacker (white-box setting). As plenty of machine learning models have been deployed via online services that only provide query outputs from inaccessible models (e.g. Google Cloud Vision API2), black-box adversarial attacks (inaccessible target model) are of critical security concerns in practice rather than white-box ones. However, existing query-based black-box adversarial attacks often require excessive model queries to maintain a high attack success rate. Therefore, in order to improve query efficiency, we explore the distribution of adversarial examples around benign inputs with the help of image structure information characterized by a Neural Process, and propose a Neural Process based black-box adversarial attack (NP-Attack) in this paper. Extensive experiments show that NP-Attack could greatly decrease the query counts under the black-box setting.