论文标题

通过自我知识蒸馏扩展标签平滑正则化

Extending Label Smoothing Regularization with Self-Knowledge Distillation

论文作者

Wang, Ji-Yue, Zhang, Pei, Pang, Wen-feng, Li, Jie

论文摘要

受标签平滑正则化(LSR)和知识蒸馏(KD)之间的牢固相关性的启发,我们提出了一种算法LSRKD,通过将LSR方法扩展到KD状态并应用较高的温度,以训练训练提升。然后,我们通过教师校正(TC)方法改进LSRKD,该方法在统一分配教师中手动设置了较大的班级比例。为了进一步提高LSRKD的性能,我们开发了一种名为Memory-Replay知识蒸馏(MRKD)的自我鉴定方法,该方法提供了一位知识渊博的老师,以取代LSRKD中的统一分布。 MRKD方法惩罚了当前模型的输出分布与培训轨迹的副本之间的KD损失。通过防止模型学习远离其历史产出分布空间,MRKD可以稳定学习并找到更强大的最低限度。我们的实验表明,LSRKD可以一致地提高LSR性能,尤其是在LSR无效的几个深神经网络上。此外,MRKD可以显着改善单个模型训练。实验结果证实,TC可以帮助LSRKD和MRKD提高训练,尤其是在网络上失败的网络。总体而言,LSRKD,MRKD及其TC变体与LSR方法相当或胜过LSR方法,这表明这些KD方法的广泛适用性。

Inspired by the strong correlation between the Label Smoothing Regularization(LSR) and Knowledge distillation(KD), we propose an algorithm LsrKD for training boost by extending the LSR method to the KD regime and applying a softer temperature. Then we improve the LsrKD by a Teacher Correction(TC) method, which manually sets a constant larger proportion for the right class in the uniform distribution teacher. To further improve the performance of LsrKD, we develop a self-distillation method named Memory-replay Knowledge Distillation (MrKD) that provides a knowledgeable teacher to replace the uniform distribution one in LsrKD. The MrKD method penalizes the KD loss between the current model's output distributions and its copies' on the training trajectory. By preventing the model learning so far from its historical output distribution space, MrKD can stabilize the learning and find a more robust minimum. Our experiments show that LsrKD can improve LSR performance consistently at no cost, especially on several deep neural networks where LSR is ineffectual. Also, MrKD can significantly improve single model training. The experiment results confirm that the TC can help LsrKD and MrKD to boost training, especially on the networks they are failed. Overall, LsrKD, MrKD, and their TC variants are comparable to or outperform the LSR method, suggesting the broad applicability of these KD methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源