论文标题
TRP:训练有素的等级修剪有效的深神经网络
TRP: Trained Rank Pruning for Efficient Deep Neural Networks
论文作者
论文摘要
为了在移动电话等边缘设备上启用DNN,由于其坚实的理论原理和有效的实现,因此广泛采用了低级近似。以前的几项作品试图通过低排放分解直接近似预贴模型。但是,参数的小近似误差可能会因大型预测损失而荡漾。结果,性能通常会大大下降,需要进行微调的精致努力才能恢复准确性。显然,将低级近似值与训练分开不是最佳的。与以前的作品不同,本文将低级近似值和正则化整合到培训过程中。我们提出了训练有素的等级修剪(TRP),该排名在低级近似和训练之间交替。 TRP保持原始网络的能力,同时在训练过程中施加低级约束。通过随机下梯度下降优化的核正则化来进一步促进TRP中的低等级。 TRP训练的网络固有地具有低级别的结构,并且与可忽略不计的性能损失近似,从而消除了低级分解后的微调过程。该方法在CIFAR-10和Imagenet上进行了全面评估,使用低秩近似值优于先前的压缩方法。
To enable DNNs on edge devices like mobile phones, low-rank approximation has been widely adopted because of its solid theoretical rationale and efficient implementations. Several previous works attempted to directly approximate a pretrained model by low-rank decomposition; however, small approximation errors in parameters can ripple over a large prediction loss. As a result, performance usually drops significantly and a sophisticated effort on fine-tuning is required to recover accuracy. Apparently, it is not optimal to separate low-rank approximation from training. Unlike previous works, this paper integrates low rank approximation and regularization into the training process. We propose Trained Rank Pruning (TRP), which alternates between low rank approximation and training. TRP maintains the capacity of the original network while imposing low-rank constraints during training. A nuclear regularization optimized by stochastic sub-gradient descent is utilized to further promote low rank in TRP. The TRP trained network inherently has a low-rank structure, and is approximated with negligible performance loss, thus eliminating the fine-tuning process after low rank decomposition. The proposed method is comprehensively evaluated on CIFAR-10 and ImageNet, outperforming previous compression methods using low rank approximation.