论文标题

渠道均衡网络,用于学习深层表示

Channel Equilibrium Networks for Learning Deep Representation

论文作者

Shao, Wenqi, Tang, Shitao, Pan, Xingang, Tan, Ping, Wang, Xiaogang, Luo, Ping

论文摘要

卷积神经网络(CNN)通常是通过堆叠多个构件来构建的,每个构件都包含一个归一化层,例如批处理归一化(BN)和一个矫正的线性函数,例如Relu。但是,这项工作表明,归一化和整流线性函数的组合导致抑制通道,这些通道的幅度很小,对学习的特征表示几乎没有贡献,从而阻碍了CNN的普遍性能力。与以前的艺术只是简单地删除了被抑制的通道,我们建议在设计新颖的神经构建块(CE)平衡(CE)块的训练期间“唤醒”,这使同一层的通道能够同样贡献学习的表示。我们表明,CE能够从经验和理论上预防抑制通道。 CE有一些吸引人的好处。 (1)可以将其集成到许多高级CNN架构中,例如Resnet和Mobilenet,表现优于其原始网络。 (2)CE与Nash平衡有着有趣的联系,Nash平衡是非合作游戏的众所周知的解决方案。 (3)广泛的实验表明,CE可以在ImageNet和Coco等各种具有挑战性的基准上实现最先进的性能。

Convolutional Neural Networks (CNNs) are typically constructed by stacking multiple building blocks, each of which contains a normalization layer such as batch normalization (BN) and a rectified linear function such as ReLU. However, this work shows that the combination of normalization and rectified linear function leads to inhibited channels, which have small magnitude and contribute little to the learned feature representation, impeding the generalization ability of CNNs. Unlike prior arts that simply removed the inhibited channels, we propose to "wake them up" during training by designing a novel neural building block, termed Channel Equilibrium (CE) block, which enables channels at the same layer to contribute equally to the learned representation. We show that CE is able to prevent inhibited channels both empirically and theoretically. CE has several appealing benefits. (1) It can be integrated into many advanced CNN architectures such as ResNet and MobileNet, outperforming their original networks. (2) CE has an interesting connection with the Nash Equilibrium, a well-known solution of a non-cooperative game. (3) Extensive experiments show that CE achieves state-of-the-art performance on various challenging benchmarks such as ImageNet and COCO.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源