论文标题
具有性能保证的无增强图对比度学习
Augmentation-Free Graph Contrastive Learning with Performance Guarantee
论文作者
论文摘要
图形对比学习(GCL)是图形结构数据的最具代表性和普遍的自我监督学习方法。尽管取得了显着的成功,但现有的GCL方法在很大程度上依赖于增强计划,以了解各种增强视图中不变的表示。在这项工作中,我们通过通过光谱理论的镜头来检查增强技术对图数据的影响,在GCL中重新审视这种惯例。我们发现,图表保留了低频组件,并扰动该图的中和高频组件,这有助于GCL算法在同质图上的成功,但由于异性数据的高频偏好。在此激励的基础上,我们提出了一种名为AF-GCL的新颖,理论原理和无增强的GCL方法,该方法(1)(1)利用图形神经网络汇总的特征来构建自主性信号而不是增强信号,因此(2)对图同质程度不太敏感。从理论上讲,我们介绍了AF-GCL的性能保证以及了解AF-GCL功效的分析。在14个基准数据集上具有不同程度的异质性实验表明,AF-GCL在同质图上表现出竞争性或更好的性能,并且在异性图上胜过所有现有的最新GCL方法,其计算范围明显较小。
Graph contrastive learning (GCL) is the most representative and prevalent self-supervised learning approach for graph-structured data. Despite its remarkable success, existing GCL methods highly rely on an augmentation scheme to learn the representations invariant across different augmentation views. In this work, we revisit such a convention in GCL through examining the effect of augmentation techniques on graph data via the lens of spectral theory. We found that graph augmentations preserve the low-frequency components and perturb the middle-and high-frequency components of the graph, which contributes to the success of GCL algorithms on homophilic graphs but hinder its application on heterophilic graphs, due to the high-frequency preference of heterophilic data. Motivated by this, we propose a novel, theoretically-principled, and augmentation-free GCL method, named AF-GCL, that (1) leverages the features aggregated by Graph Neural Network to construct the self-supervision signal instead of augmentations and therefore (2) is less sensitive to the graph homophily degree. Theoretically, We present the performance guarantee for AF-GCL as well as an analysis for understanding the efficacy of AF-GCL. Extensive experiments on 14 benchmark datasets with varying degrees of heterophily show that AF-GCL presents competitive or better performance on homophilic graphs and outperforms all existing state-of-the-art GCL methods on heterophilic graphs with significantly less computational overhead.