论文标题

大多数激活功能都可以赢得彩票,而无需过度深度

Most Activation Functions Can Win the Lottery Without Excessive Depth

论文作者

Burkholz, Rebekka

论文摘要

强有力的彩票假说突出了通过修剪训练深神经网络的潜力,这激发了人们对神经网络如何代表功能的有趣实用和理论见解。对于具有RELU激活功能的网络,已证明具有深度$ L $的目标网络可以通过随机初始化的神经网络的子网近似,该网络的子网具有将目标深度$ 2L $两倍,并且通过对数因子更宽。我们表明,深度$ L+1 $网络就足够了。该结果表明,我们可以期望在现实的,常用的深度上找到彩票,而只需要对数过度参数化。我们的新型构造方法适用于大量的激活功能,不仅限于RES。

The strong lottery ticket hypothesis has highlighted the potential for training deep neural networks by pruning, which has inspired interesting practical and theoretical insights into how neural networks can represent functions. For networks with ReLU activation functions, it has been proven that a target network with depth $L$ can be approximated by the subnetwork of a randomly initialized neural network that has double the target's depth $2L$ and is wider by a logarithmic factor. We show that a depth $L+1$ network is sufficient. This result indicates that we can expect to find lottery tickets at realistic, commonly used depths while only requiring logarithmic overparametrization. Our novel construction approach applies to a large class of activation functions and is not limited to ReLUs.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源