论文标题
更广泛的神经网络真的有助于对抗性鲁棒性吗?
Do Wider Neural Networks Really Help Adversarial Robustness?
论文作者
论文摘要
对抗训练是针对对抗性例子的强大防御类型。先前的经验结果表明,对抗训练需要更广泛的网络才能更好地表现。但是,仍然难以捉摸的神经网络宽度如何影响模型的鲁棒性。在本文中,我们仔细检查了网络宽度与模型鲁棒性之间的关系。具体而言,我们表明模型鲁棒性与自然准确性和扰动稳定性之间的权衡密切相关,该稳定性由可靠的正则化参数$λ$控制。凭借相同的$λ$,更宽的网络可以实现更好的自然精度,但扰动稳定性较差,从而导致整体模型的鲁棒性可能更差。为了了解这种现象的起源,我们将扰动稳定性与网络的本地Lipschitzness联系起来。通过利用有关神经切线内核的最新结果,我们从理论上表明,更广泛的网络往往具有较差的扰动稳定性。我们的分析表明:1)在小型网络上首先调整$λ$的常见策略,然后直接将其用于广泛的模型训练可能会导致模型稳健性恶化; 2)需要正确扩大$λ$,以完全释放更广泛模型的稳健性。最后,我们提出了一种新的宽度调整正规化方法(WAR)方法,该方法可自适应地扩大$λ$在宽模型上,并大大节省了调谐时间。
Adversarial training is a powerful type of defense against adversarial examples. Previous empirical results suggest that adversarial training requires wider networks for better performances. However, it remains elusive how neural network width affects model robustness. In this paper, we carefully examine the relationship between network width and model robustness. Specifically, we show that the model robustness is closely related to the tradeoff between natural accuracy and perturbation stability, which is controlled by the robust regularization parameter $λ$. With the same $λ$, wider networks can achieve better natural accuracy but worse perturbation stability, leading to a potentially worse overall model robustness. To understand the origin of this phenomenon, we further relate the perturbation stability with the network's local Lipschitzness. By leveraging recent results on neural tangent kernels, we theoretically show that wider networks tend to have worse perturbation stability. Our analyses suggest that: 1) the common strategy of first fine-tuning $λ$ on small networks and then directly use it for wide model training could lead to deteriorated model robustness; 2) one needs to properly enlarge $λ$ to unleash the robustness potential of wider models fully. Finally, we propose a new Width Adjusted Regularization (WAR) method that adaptively enlarges $λ$ on wide models and significantly saves the tuning time.