论文标题
CNN denoiser和非本地过滤器之间的神经切线链接
The Neural Tangent Link Between CNN Denoisers and Non-Local Filters
论文作者
论文摘要
卷积神经网络(CNN)现在是解决计算成像问题的良好工具。基于CNN的现代算法在各种图像恢复问题中获得最先进的性能。此外,最近已经表明,尽管过度参数过度参数化,但接受单个损坏图像训练的网络仍然可以执行以及训练有素的网络。我们通过其神经切线内核(NTK)和众所周知的非本地滤波技术(例如非本地均值或BM3D)引入了此类网络之间的正式联系。与给定网络体系结构相关的过滤功能可以以封闭形式获得,而无需训练网络,其完全以网络权重的随机初始化为特征。尽管NTK理论准确地预测了与使用标准梯度下降训练的网络相关的过滤器,但我们的分析表明,它不足以解释使用流行的Adam优化器训练的网络的行为。后者在隐藏层中实现了更大的重量,在训练过程中适应了非本地过滤功能。我们通过广泛的图像剥夺实验评估我们的发现。
Convolutional Neural Networks (CNNs) are now a well-established tool for solving computational imaging problems. Modern CNN-based algorithms obtain state-of-the-art performance in diverse image restoration problems. Furthermore, it has been recently shown that, despite being highly overparameterized, networks trained with a single corrupted image can still perform as well as fully trained networks. We introduce a formal link between such networks through their neural tangent kernel (NTK), and well-known non-local filtering techniques, such as non-local means or BM3D. The filtering function associated with a given network architecture can be obtained in closed form without need to train the network, being fully characterized by the random initialization of the network weights. While the NTK theory accurately predicts the filter associated with networks trained using standard gradient descent, our analysis shows that it falls short to explain the behaviour of networks trained using the popular Adam optimizer. The latter achieves a larger change of weights in hidden layers, adapting the non-local filtering function during training. We evaluate our findings via extensive image denoising experiments.