论文标题

图表回顾性

Graph Attention Retrospective

论文作者

Fountoulakis, Kimon, Levi, Amit, Yang, Shenghao, Baranwal, Aseem, Jagannath, Aukosh

论文摘要

基于图形的学习是机器学习的快速增长的子场,其中具有社交网络,引文网络和生物信息学中的应用。最受欢迎的模型之一是图形注意力网络。引入它们是为了使节点以不均匀的方式从邻居节点的特征汇总信息,而简单的图形卷积则没有区分节点的邻居。在本文中,我们从理论上研究图形注意力网络的行为。对于上下文随机块模型,我们证明了有关图形注意机制的性能的多个结果。在这里,节点特征是从高斯的混合物和从随机块模型的边缘获得的。我们表明,在“简单”制度中,高斯人的平均值之间的距离足够大,图形注意力可以将阶层间的阶层与阶层内边缘区分开。因此,它保持了重要边缘的权重,并显着降低了不重要的边缘的权重。因此,我们表明这意味着完美的节点分类。在“硬”制度中,我们表明,每种注意力机制都无法将阶级内部和阶层边缘区分开。此外,我们表明,即使可以将阶层内边缘与阶层间边缘分开,图形注意力卷积也无法(几乎)完美地对节点进行分类。除了完美的节点分类之外,我们还为图中图中的图表的鲁棒性提供了积极的结果。特别是,我们的鲁棒性结果意味着图形注意力可以比简单的图形卷积和最佳的节点特征线性分类器更好。我们评估了关于合成和现实世界数据的理论结果。

Graph-based learning is a rapidly growing sub-field of machine learning with applications in social networks, citation networks, and bioinformatics. One of the most popular models is graph attention networks. They were introduced to allow a node to aggregate information from features of neighbor nodes in a non-uniform way, in contrast to simple graph convolution which does not distinguish the neighbors of a node. In this paper, we theoretically study the behaviour of graph attention networks. We prove multiple results on the performance of the graph attention mechanism for the problem of node classification for a contextual stochastic block model. Here, the node features are obtained from a mixture of Gaussians and the edges from a stochastic block model. We show that in an "easy" regime, where the distance between the means of the Gaussians is large enough, graph attention is able to distinguish inter-class from intra-class edges. Thus it maintains the weights of important edges and significantly reduces the weights of unimportant edges. Consequently, we show that this implies perfect node classification. In the "hard" regime, we show that every attention mechanism fails to distinguish intra-class from inter-class edges. In addition, we show that graph attention convolution cannot (almost) perfectly classify the nodes even if intra-class edges could be separated from inter-class edges. Beyond perfect node classification, we provide a positive result on graph attention's robustness against structural noise in the graph. In particular, our robustness result implies that graph attention can be strictly better than both the simple graph convolution and the best linear classifier of node features. We evaluate our theoretical results on synthetic and real-world data.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源