论文标题

领域概括 - 因果观点

Domain Generalization -- A Causal Perspective

论文作者

Sheth, Paras, Moraffah, Raha, Candan, K. Selçuk, Raglin, Adrienne, Liu, Huan

论文摘要

机器学习模型依靠各种假设来获得高精度。这些模型的初步假设之一是独立且相同的分布,这表明火车和测试数据是从相同分布中采样的。但是,由于分布的变化,这种假设很少在现实世界中。结果,依靠这一假设的模型表现出较差的概括能力。近年来,已经做出了专门的努力来提高这些模型的概括能力,共同称为 - \ textit {域概括方法}。这些方法背后的主要思想是识别在不同分布中保持不变的稳定特征或机制。许多概括方法采用因果理论来描述不变性,因为因果关系和不变性是密不可分的。但是,当前的调查涉及非常高级的因果感知领域的概括方法。此外,我们认为可以根据如何利用因果关系在该方法以及使用模型管道的哪一部分中对方法进行分类。为此,我们将因果领域的泛化方法分为三类,即(i)通过因果数据增强方法不变性,这些方法在数据预处理阶段中应用,(ii)通过因果代表学习方法不变性在代表性学习阶段和(iii II II II III阶段)的范围内进行启发的方法。此外,该调查还包括对域泛化方法的基准数据集和代码存储库的深入见解。我们通过对未来方向的见解和讨论来结束调查。

Machine learning models rely on various assumptions to attain high accuracy. One of the preliminary assumptions of these models is the independent and identical distribution, which suggests that the train and test data are sampled from the same distribution. However, this assumption seldom holds in the real world due to distribution shifts. As a result models that rely on this assumption exhibit poor generalization capabilities. Over the recent years, dedicated efforts have been made to improve the generalization capabilities of these models collectively known as -- \textit{domain generalization methods}. The primary idea behind these methods is to identify stable features or mechanisms that remain invariant across the different distributions. Many generalization approaches employ causal theories to describe invariance since causality and invariance are inextricably intertwined. However, current surveys deal with the causality-aware domain generalization methods on a very high-level. Furthermore, we argue that it is possible to categorize the methods based on how causality is leveraged in that method and in which part of the model pipeline is it used. To this end, we categorize the causal domain generalization methods into three categories, namely, (i) Invariance via Causal Data Augmentation methods which are applied during the data pre-processing stage, (ii) Invariance via Causal representation learning methods that are utilized during the representation learning stage, and (iii) Invariance via Transferring Causal mechanisms methods that are applied during the classification stage of the pipeline. Furthermore, this survey includes in-depth insights into benchmark datasets and code repositories for domain generalization methods. We conclude the survey with insights and discussions on future directions.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源