论文标题
朝着基于亲和力的知识蒸馏的统一观点
Towards a Unified View of Affinity-Based Knowledge Distillation
论文作者
论文摘要
人工神经网络之间的知识转移已成为深度学习的重要主题。在开放的问题中,需要保留哪种知识以进行转移以及如何有效地实现。最近的几项工作显示了使用基于关系的知识的蒸馏方法的良好性能。这些算法极具吸引力,因为它们基于简单的样本间相似性。然而,在这种情况下,适当的亲和力和使用它的使用远远不够。在本文中,通过将知识蒸馏显式地模块化为三个组成部分的框架,即,正常化和损失,我们对这些算法进行了统一的处理,并研究了许多模块的未开发组合。通过此框架,我们对图像分类进行了许多蒸馏目标的广泛评估,并获得了一些有效的设计选择的有用见解,同时证明尽管具有简单性,但基于关系的知识蒸馏如何实现与艺术的可比性能。
Knowledge transfer between artificial neural networks has become an important topic in deep learning. Among the open questions are what kind of knowledge needs to be preserved for the transfer, and how it can be effectively achieved. Several recent work have shown good performance of distillation methods using relation-based knowledge. These algorithms are extremely attractive in that they are based on simple inter-sample similarities. Nevertheless, a proper metric of affinity and use of it in this context is far from well understood. In this paper, by explicitly modularising knowledge distillation into a framework of three components, i.e. affinity, normalisation, and loss, we give a unified treatment of these algorithms as well as study a number of unexplored combinations of the modules. With this framework we perform extensive evaluations of numerous distillation objectives for image classification, and obtain a few useful insights for effective design choices while demonstrating how relation-based knowledge distillation could achieve comparable performance to the state of the art in spite of the simplicity.