论文标题

大型图学习的未耦合培训体系结构

An Uncoupled Training Architecture for Large Graph Learning

论文作者

Yang, Dalong, Chen, Chuan, Zheng, Youhao, Zheng, Zibin, Liao, Shih-wei

论文摘要

图形卷积网络(GCN)已被广泛用于图形学习任务。但是,基于GCN的模型(GCN)是一个固有的耦合训练框架,重复地进行了复杂的相邻聚集,这导致了处理大规模图的灵活性的限制。随着层深度的增加,由于递归邻居的扩张,GCN的计算和记忆成本在爆炸性上增长。为了解决这些问题,我们提出了Node2Grids,这是一个灵活的未耦合培训框架,利用独立的映射数据来获取嵌入。节点2grids不是直接处理耦合节点作为GCN,而是支持一种更有效的方法,而是将耦合的图数据映射到可以将其馈送到有效的卷积神经网络(CNN)中的独立网格样数据中。这种简单但有效的策略可大大节省内存和计算资源,同时与领先的基于GCN的模型获得可比的结果。具体而言,通过通过度对每个节点的影响进行排名,Node2Grids选择了具有中央节点融合信息的最具影响力的一阶和二阶邻居来构建类似网格的数据。为了进一步提高下游任务的效率,采用了一个简单的基于CNN的神经网络来从映射的网格样数据中捕获重要信息。此外,实施了网格级的注意机制,该机制能够暗中指定具有不同影响的相邻节点的不同权重。除了典型的转导和归纳学习任务外,我们还验证了我们在百万尺度图上的框架,以证明所提出的节点2Grids模型的优越性,而不是基于GCN的最先进的方法。

Graph Convolutional Network (GCN) has been widely used in graph learning tasks. However, GCN-based models (GCNs) is an inherently coupled training framework repetitively conducting the complex neighboring aggregation, which leads to the limitation of flexibility in processing large-scale graph. With the depth of layers increases, the computational and memory cost of GCNs grow explosively due to the recursive neighborhood expansion. To tackle these issues, we present Node2Grids, a flexible uncoupled training framework that leverages the independent mapped data for obtaining the embedding. Instead of directly processing the coupled nodes as GCNs, Node2Grids supports a more efficacious method in practice, mapping the coupled graph data into the independent grid-like data which can be fed into the efficient Convolutional Neural Network (CNN). This simple but valid strategy significantly saves memory and computational resource while achieving comparable results with the leading GCN-based models. Specifically, by ranking each node's influence through degree, Node2Grids selects the most influential first-order as well as second-order neighbors with central node fusion information to construct the grid-like data. For further improving the efficiency of downstream tasks, a simple CNN-based neural network is employed to capture the significant information from the mapped grid-like data. Moreover, the grid-level attention mechanism is implemented, which enables implicitly specifying the different weights for neighboring nodes with different influences. In addition to the typical transductive and inductive learning tasks, we also verify our framework on million-scale graphs to demonstrate the superiority of the proposed Node2Grids model against the state-of-the-art GCN-based approaches.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源