论文标题
通过知识蒸馏,无监督的多目标域适应
Unsupervised Multi-Target Domain Adaptation Through Knowledge Distillation
论文作者
论文摘要
无监督的域适应性(UDA)试图减轻域转移的问题,从目标域W.R.T.未标记的数据分布之间。来自源域的数据标记。尽管单一目标UDA方案在文献中进行了很好的研究,但多目标域的适应性(MTDA)仍然在很大程度上没有探索,尽管其实际上很重要,例如,在多相机视频配置应用中。 MTDA问题可以通过调整每个目标域的一个专业模型来解决,尽管在许多现实世界中,该解决方案太昂贵了。已经提出了混合MTDA的多个目标,但该解决方案可能导致模型特异性和准确性降低。在本文中,我们提出了一种新型的无监督MTDA方法来训练可以很好地跨越多个目标域的CNN。我们的多老师MTDA(MT-MTDA)方法依靠多教老师知识蒸馏(KD)来迭代地提取目标领域知识,从多个教师到普通学生。 KD过程是以渐进的方式进行的,每个老师都会培训学生如何为特定目标执行UDA,而不是直接学习域名适应的功能。最后,MT-MTDA并没有结合每个老师的知识,而是在教师之间交替提炼知识,从而在学习适应学生时保留了每个目标(老师)的特殊性。将MT-MTDA与几种具有挑战性的UDA基准的最新方法进行了比较,并且经验结果表明,我们提出的模型可以提供跨多个目标域的高度准确性。我们的代码可在以下网址找到:https://github.com/liviaets/mt-mtda
Unsupervised domain adaptation (UDA) seeks to alleviate the problem of domain shift between the distribution of unlabeled data from the target domain w.r.t. labeled data from the source domain. While the single-target UDA scenario is well studied in the literature, Multi-Target Domain Adaptation (MTDA) remains largely unexplored despite its practical importance, e.g., in multi-camera video-surveillance applications. The MTDA problem can be addressed by adapting one specialized model per target domain, although this solution is too costly in many real-world applications. Blending multiple targets for MTDA has been proposed, yet this solution may lead to a reduction in model specificity and accuracy. In this paper, we propose a novel unsupervised MTDA approach to train a CNN that can generalize well across multiple target domains. Our Multi-Teacher MTDA (MT-MTDA) method relies on multi-teacher knowledge distillation (KD) to iteratively distill target domain knowledge from multiple teachers to a common student. The KD process is performed in a progressive manner, where the student is trained by each teacher on how to perform UDA for a specific target, instead of directly learning domain adapted features. Finally, instead of combining the knowledge from each teacher, MT-MTDA alternates between teachers that distill knowledge, thereby preserving the specificity of each target (teacher) when learning to adapt to the student. MT-MTDA is compared against state-of-the-art methods on several challenging UDA benchmarks, and empirical results show that our proposed model can provide a considerably higher level of accuracy across multiple target domains. Our code is available at: https://github.com/LIVIAETS/MT-MTDA