论文标题

您只需要一个简单的微调:通过对抗性微调进行强大的深度学习

A Simple Fine-tuning Is All You Need: Towards Robust Deep Learning Via Adversarial Fine-tuning

论文作者

Jeddi, Ahmadreza, Shafiee, Mohammad Javad, Wong, Alexander

论文摘要

具有预计梯度下降(PGD)的对抗训练(AT)是改善深神经网络鲁棒性的有效方法。但是,PGD AT已显示出两个主要局限性:i)较高的计算成本,ii)在训练过程中极度拟合,从而导致模型概括减少。尽管已经对诸如模型能力和训练数据量表等因素的影响进行了广泛的研究,但在每个网络优化对对抗性鲁棒性的每个网络优化中,很少关注非常重要的参数的影响:学习率。特别是,我们假设在对抗训练期间的有效学习率调度可以显着减少过度拟合的问题,以至于甚至不需要从头开始对手训练模型,而可以简单地对对手进行预训练的模型。在这一假设的推动下,我们提出了一种基于$ \ textit {慢起步,快速衰减} $学习率调度策略的简单但非常有效的对抗性微调方法,该方法不仅显着降低了所需的计算成本,而且还大大提高了深度神经网络的准确性和鲁棒性。实验结果表明,提出的对抗性微调方法的表现优于CIFAR-10,CIFAR-100和IMAGENET数据集的最新方法,同时将计算成本降低了8-10 $ \ times $。此外,提出的对抗性微调方法的一个非常重要的好处是,它使得能够提高任何预训练的深度神经网络的鲁棒性而无需从头开始训练模型,从而在研究文献中尚未证明这些模型以前尚未证明这一点。

Adversarial Training (AT) with Projected Gradient Descent (PGD) is an effective approach for improving the robustness of the deep neural networks. However, PGD AT has been shown to suffer from two main limitations: i) high computational cost, and ii) extreme overfitting during training that leads to reduction in model generalization. While the effect of factors such as model capacity and scale of training data on adversarial robustness have been extensively studied, little attention has been paid to the effect of a very important parameter in every network optimization on adversarial robustness: the learning rate. In particular, we hypothesize that effective learning rate scheduling during adversarial training can significantly reduce the overfitting issue, to a degree where one does not even need to adversarially train a model from scratch but can instead simply adversarially fine-tune a pre-trained model. Motivated by this hypothesis, we propose a simple yet very effective adversarial fine-tuning approach based on a $\textit{slow start, fast decay}$ learning rate scheduling strategy which not only significantly decreases computational cost required, but also greatly improves the accuracy and robustness of a deep neural network. Experimental results show that the proposed adversarial fine-tuning approach outperforms the state-of-the-art methods on CIFAR-10, CIFAR-100 and ImageNet datasets in both test accuracy and the robustness, while reducing the computational cost by 8-10$\times$. Furthermore, a very important benefit of the proposed adversarial fine-tuning approach is that it enables the ability to improve the robustness of any pre-trained deep neural network without needing to train the model from scratch, which to the best of the authors' knowledge has not been previously demonstrated in research literature.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源