论文标题

通过负相关合奏捍卫对抗示例

Defending Adversarial Examples by Negative Correlation Ensemble

论文作者

Luo, Wenjian, Zhang, Hongwei, Kong, Linghao, Chen, Zhijian, Tang, Ke

论文摘要

DNN中的安全问题,例如对抗性示例,引起了很多关注。对抗性示例是指能够通过引入精心设计的扰动来诱导DNN完全返回预测的示例。显然,对抗性的例子给深度学习的发展带来了很大的安全风险。最近,提出了一些针对对抗性例子的防御方法,但是,我们认为这些方法的表现仍然有限。在本文中,我们提出了一种名为负相关集成(NCEN)的新的集合防御方法,该方法通过在整体中引入每个成员的梯度方向和梯度大小来实现令人信服的结果,并同时降低了它们中对抗性例子的转移性。已经进行了广泛的实验,结果表明NCEN可以有效地改善合奏的对抗性鲁棒性。

The security issues in DNNs, such as adversarial examples, have attracted much attention. Adversarial examples refer to the examples which are capable to induce the DNNs return completely predictions by introducing carefully designed perturbations. Obviously, adversarial examples bring great security risks to the development of deep learning. Recently, Some defense approaches against adversarial examples have been proposed, however, in our opinion, the performance of these approaches are still limited. In this paper, we propose a new ensemble defense approach named the Negative Correlation Ensemble (NCEn), which achieves compelling results by introducing gradient directions and gradient magnitudes of each member in the ensemble negatively correlated and at the same time, reducing the transferability of adversarial examples among them. Extensive experiments have been conducted, and the results demonstrate that NCEn can improve the adversarial robustness of ensembles effectively.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源