论文标题
与低级本地连通性重新访问空间不变性
Revisiting Spatial Invariance with Low-Rank Local Connectivity
论文作者
论文摘要
卷积神经网络是深度学习中最成功的架构之一,这一成功至少部分归因于空间不变性作为归纳偏见的功效。本地连接的层与卷积层仅在缺乏空间不变性方面不同,通常在实践中表现较差。但是,这些观察结果仍然使某种程度的空间不变性放松可能会产生比卷积或局部连通性更好的感应偏见的可能性。为了检验这一假设,我们设计了一种以受控方式放松网络层的空间不变性的方法。我们创建了一个\ textit {low-rank}本地连接的层,在该图层中,在每个位置上应用的滤波器库被构造为具有空间变化的组合权重的滤波器库基集的线性组合。通过改变基本过滤器库的数量,我们可以控制空间不变性的放松程度。在小型卷积网络的实验中,我们发现放松的空间不变性提高了MNIST,CIFAR-10和CEELBA数据集的卷积和本地连接层的分类精度,因此表明空间不变性可能是过度限制的先验。
Convolutional neural networks are among the most successful architectures in deep learning with this success at least partially attributable to the efficacy of spatial invariance as an inductive bias. Locally connected layers, which differ from convolutional layers only in their lack of spatial invariance, usually perform poorly in practice. However, these observations still leave open the possibility that some degree of relaxation of spatial invariance may yield a better inductive bias than either convolution or local connectivity. To test this hypothesis, we design a method to relax the spatial invariance of a network layer in a controlled manner; we create a \textit{low-rank} locally connected layer, where the filter bank applied at each position is constructed as a linear combination of basis set of filter banks with spatially varying combining weights. By varying the number of basis filter banks, we can control the degree of relaxation of spatial invariance. In experiments with small convolutional networks, we find that relaxing spatial invariance improves classification accuracy over both convolution and locally connected layers across MNIST, CIFAR-10, and CelebA datasets, thus suggesting that spatial invariance may be an overly restrictive prior.