论文标题

先验知识指导了无监督的领域适应

Prior Knowledge Guided Unsupervised Domain Adaptation

论文作者

Sun, Tao, Lu, Cheng, Ling, Haibin

论文摘要

目标域中的标签放弃使无监督的域适应性(UDA)成为许多现实世界应用中有吸引力的技术,尽管它也带来了巨大的挑战,因为没有标记目标数据的模型适应变得更加困难。在本文中,我们通过从目标领域的先验知识中寻求赔偿来解决这个问题,这在实践中通常(部分)可用于人类专业知识。这导致了一个新颖而实用的环境,除了训练数据外,还提供有关目标类别分布的一些先验知识。我们将设置称为知识引导的无监督域适应性(KUDA)。特别是,我们考虑了有关目标域中类别分布的两种特定类型的先验知识:描述单个类概率的下层和上限的Unary Bound,以及描述了两个类别概率之间关系的二进制关系。我们提出了一个使用此类先验知识来完善模型生成的伪标签的一般整流模块。该模块被配制为从先验知识和光滑的正常化程序中得出的零一个编程问题。它可以很容易地插入基于自我训练的UDA方法中,我们将其与两种最先进的方法结合使用,即射击和用餐。四个基准的经验结果证实,整流模块显然改善了伪标签的质量,这反过来又有益于自我训练阶段。在先验知识的指导下,两种方法的性能都大大提高。我们希望我们的工作能够激发进一步的调查,以整合UDA的先验知识。代码可在https://github.com/tsun/kuda上找到。

The waive of labels in the target domain makes Unsupervised Domain Adaptation (UDA) an attractive technique in many real-world applications, though it also brings great challenges as model adaptation becomes harder without labeled target data. In this paper, we address this issue by seeking compensation from target domain prior knowledge, which is often (partially) available in practice, e.g., from human expertise. This leads to a novel yet practical setting where in addition to the training data, some prior knowledge about the target class distribution are available. We term the setting as Knowledge-guided Unsupervised Domain Adaptation (KUDA). In particular, we consider two specific types of prior knowledge about the class distribution in the target domain: Unary Bound that describes the lower and upper bounds of individual class probabilities, and Binary Relationship that describes the relations between two class probabilities. We propose a general rectification module that uses such prior knowledge to refine model generated pseudo labels. The module is formulated as a Zero-One Programming problem derived from the prior knowledge and a smooth regularizer. It can be easily plugged into self-training based UDA methods, and we combine it with two state-of-the-art methods, SHOT and DINE. Empirical results on four benchmarks confirm that the rectification module clearly improves the quality of pseudo labels, which in turn benefits the self-training stage. With the guidance from prior knowledge, the performances of both methods are substantially boosted. We expect our work to inspire further investigations in integrating prior knowledge in UDA. Code is available at https://github.com/tsun/KUDA.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源