论文标题

学习为多攻击鲁棒性产生噪音

Learning to Generate Noise for Multi-Attack Robustness

论文作者

Madaan, Divyam, Shin, Jinwoo, Hwang, Sung Ju

论文摘要

对抗性学习已成为绕过现有方法对抗扰动的敏感性的成功技术之一。但是,大多数现有的防御方法都是为了防御一种对抗性扰动而定制的(例如$ \ ell_ \ infty $ -ATTACK)。在安全至关重要的应用中,这使这些方法无关紧要,因为攻击者可以采用不同的对手来欺骗系统。此外,对多种扰动的培训同时可以显着增加培训期间的计算开销。为了应对这些挑战,我们提出了一个新颖的元学习框架,该框架明确学习以产生噪声,以提高模型对多种攻击的鲁棒性。它的关键组件是元噪声发生器(MNG),它输出最佳噪声以随机处理给定样品,从而有助于降低各种对抗性扰动的误差。通过利用MNG生成的样品,我们通过在多个扰动中执行标签一致性来训练模型。我们验证了我们计划在各种数据集上训练的模型的鲁棒性,并不利于各种扰动,这表明它以边缘计算成本大大优于多个扰动的基准。

Adversarial learning has emerged as one of the successful techniques to circumvent the susceptibility of existing methods against adversarial perturbations. However, the majority of existing defense methods are tailored to defend against a single category of adversarial perturbation (e.g. $\ell_\infty$-attack). In safety-critical applications, this makes these methods extraneous as the attacker can adopt diverse adversaries to deceive the system. Moreover, training on multiple perturbations simultaneously significantly increases the computational overhead during training. To address these challenges, we propose a novel meta-learning framework that explicitly learns to generate noise to improve the model's robustness against multiple types of attacks. Its key component is Meta Noise Generator (MNG) that outputs optimal noise to stochastically perturb a given sample, such that it helps lower the error on diverse adversarial perturbations. By utilizing samples generated by MNG, we train a model by enforcing the label consistency across multiple perturbations. We validate the robustness of models trained by our scheme on various datasets and against a wide variety of perturbations, demonstrating that it significantly outperforms the baselines across multiple perturbations with a marginal computational cost.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源