论文标题
SAT:通过基于课程的损失平滑改善对抗性训练
SAT: Improving Adversarial Training via Curriculum-Based Loss Smoothing
论文作者
论文摘要
对抗训练(AT)已成为训练强大网络的流行选择。但是,它倾向于大力牺牲清洁准确性,以促进鲁棒性,并遭受巨大的概括错误。为了解决这些问题,我们提出了平稳的对抗训练(SAT),并在我们对Hessian损失特征的分析的指导下进行了指导。我们发现,课程学习是一种强调开始“轻松”并逐渐加剧培训的“难度”的计划,可以使对抗性损失格局平滑,以进行适当选择的难度度量。我们提出了在对抗环境中课程学习的一般公式,并提出了基于最大Hessian特征值(H-SAT)和SoftMax概率(P-SA)的两个难度指标。我们证明,SAT甚至可以为大型扰动规范而稳定网络培训,并允许网络以更好的清洁精度与稳健性权衡曲线相比,与AT相比。与AT(交易和其他基线)相比,这会导致清洁准确性和鲁棒性的显着提高。为了强调一些结果,我们的最佳模型将正常和鲁棒精度提高了6%和1%的CIFAR-100,而AT分别提高了6%和1%。在Imagenette(Imagenette)的一个十级子集中,我们的模型的表现分别以正常和稳健的精度优于23%和3%。
Adversarial training (AT) has become a popular choice for training robust networks. However, it tends to sacrifice clean accuracy heavily in favor of robustness and suffers from a large generalization error. To address these concerns, we propose Smooth Adversarial Training (SAT), guided by our analysis on the eigenspectrum of the loss Hessian. We find that curriculum learning, a scheme that emphasizes on starting "easy" and gradually ramping up on the "difficulty" of training, smooths the adversarial loss landscape for a suitably chosen difficulty metric. We present a general formulation for curriculum learning in the adversarial setting and propose two difficulty metrics based on the maximal Hessian eigenvalue (H-SAT) and the softmax probability (P-SA). We demonstrate that SAT stabilizes network training even for a large perturbation norm and allows the network to operate at a better clean accuracy versus robustness trade-off curve compared to AT. This leads to a significant improvement in both clean accuracy and robustness compared to AT, TRADES, and other baselines. To highlight a few results, our best model improves normal and robust accuracy by 6% and 1% on CIFAR-100 compared to AT, respectively. On Imagenette, a ten-class subset of ImageNet, our model outperforms AT by 23% and 3% on normal and robust accuracy respectively.