论文标题

修剪的神经网络令人惊讶地模块化

Pruned Neural Networks are Surprisingly Modular

论文作者

Filan, Daniel, Hod, Shlomi, Wild, Cody, Critch, Andrew, Russell, Stuart

论文摘要

神经网络的学习重量通常被认为缺乏可仔的内部结构。为了辨别这些权重的结构,我们引入了多层感知器(MLP)的模块化概念,并研究了在小图像数据集中训练的MLP的模块化结构。我们的模块化概念来自图群集文献:“模块”是一组具有强大内部连接性但外部连接较弱的神经元。我们发现,训练和重量修剪会产生比随机初始化的MLP更模块化的,并且通常比重量相同(稀疏)分布的随机MLP明显更模块化。有趣的是,经过辍学训练时,它们会更加模块化。我们还介绍了不同模块对性能的重要性以及模块如何相互依赖的探索性分析。当存在这种结构时,了解神经网络的模块化结构将使他们的内部运作更加可解释工程师。请注意,本文已被“神经网络中的可突出性”,ARXIV:2103.03386和“量化深神经网络中的本地专业化”,ARXIV:2110.08058!

The learned weights of a neural network are often considered devoid of scrutable internal structure. To discern structure in these weights, we introduce a measurable notion of modularity for multi-layer perceptrons (MLPs), and investigate the modular structure of MLPs trained on datasets of small images. Our notion of modularity comes from the graph clustering literature: a "module" is a set of neurons with strong internal connectivity but weak external connectivity. We find that training and weight pruning produces MLPs that are more modular than randomly initialized ones, and often significantly more modular than random MLPs with the same (sparse) distribution of weights. Interestingly, they are much more modular when trained with dropout. We also present exploratory analyses of the importance of different modules for performance and how modules depend on each other. Understanding the modular structure of neural networks, when such structure exists, will hopefully render their inner workings more interpretable to engineers. Note that this paper has been superceded by "Clusterability in Neural Networks", arxiv:2103.03386 and "Quantifying Local Specialization in Deep Neural Networks", arxiv:2110.08058!

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源