论文标题

对抗性的愚弄超越“翻转标签”

Adversarial Fooling Beyond "Flipping the Label"

论文作者

Mopuri, Konda Reddy, Shaj, Vaisakh, Babu, R. Venkatesh

论文摘要

CNN的最新进展已在各种CV/AI应用中取得了显着的成就。尽管在许多关键任务中,CNN在人类上表现出或比人类表现更好,但它们很容易受到对抗性攻击的影响。这些攻击在现实生活中可能是危险的。尽管近年来提出了许多对抗性攻击,但没有适当的方法来量化这些攻击的有效性。截至今天,仅使用愚弄率来衡量模型的敏感性或对抗攻击的有效性。欺骗率只是考虑了标签翻转的标签,不考虑这种翻转的成本,例如,在某些部署中,在两种狗之间翻转可能不像将狗类别与车辆混淆。因此,量化模型脆弱性的度量也应捕获翻转的严重性。在这项工作中,我们首先提出了现有评估的缺点,并提出了新颖的指标来捕捉愚蠢的各个方面。此外,我们第一次对一组不同的CNN体​​系结构进行了几次重要的对抗性攻击。我们认为,提出的分析带来了有关当前对抗性攻击和CNN模型的宝贵见解。

Recent advancements in CNNs have shown remarkable achievements in various CV/AI applications. Though CNNs show near human or better than human performance in many critical tasks, they are quite vulnerable to adversarial attacks. These attacks are potentially dangerous in real-life deployments. Though there have been many adversarial attacks proposed in recent years, there is no proper way of quantifying the effectiveness of these attacks. As of today, mere fooling rate is used for measuring the susceptibility of the models, or the effectiveness of adversarial attacks. Fooling rate just considers label flipping and does not consider the cost of such flipping, for instance, in some deployments, flipping between two species of dogs may not be as severe as confusing a dog category with that of a vehicle. Therefore, the metric to quantify the vulnerability of the models should capture the severity of the flipping as well. In this work we first bring out the drawbacks of the existing evaluation and propose novel metrics to capture various aspects of the fooling. Further, for the first time, we present a comprehensive analysis of several important adversarial attacks over a set of distinct CNN architectures. We believe that the presented analysis brings valuable insights about the current adversarial attacks and the CNN models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源