论文标题
一个强大的堆叠框架,用于训练具有多面节点功能的深图模型
A Robust Stacking Framework for Training Deep Graph Models with Multifaceted Node Features
论文作者
论文摘要
带有数值节点特征和图形结构的图形神经网络(GNN)作为输入,在具有图形数据的各种监督学习任务上表现出了出色的性能。但是,GNN使用的数值节点特征通常是从大多数真实世界应用中的文本或表格(数字/分类)类型的原始数据中提取的。在大多数标准监督的学习设置中,使用IID(非图形)数据的最佳模型不是简单的神经网络层,因此不容易被合并到GNN中。在这里,我们提出了一个可靠的堆叠框架,该框架将图形感知传播与用于IID数据的任意模型融合在一起,这些模型是在多层中结合和堆叠的。我们的层面框架利用行李和堆叠策略来享受强有力的概括,从而有效地减轻了标签泄漏和过度拟合的方式。在各种具有表格/文本节点特征的图形数据集中,我们的方法相对于表格/文本和图形神经网络模型以及将两者结合的现有最新的混合策略都具有可比性或卓越的性能。
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data. However the numerical node features utilized by GNNs are commonly extracted from raw data which is of text or tabular (numeric/categorical) type in most real-world applications. The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not simple neural network layers and thus are not easily incorporated into a GNN. Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data, which are ensembled and stacked in multiple layers. Our layer-wise framework leverages bagging and stacking strategies to enjoy strong generalization, in a manner which effectively mitigates label leakage and overfitting. Across a variety of graph datasets with tabular/text node features, our method achieves comparable or superior performance relative to both tabular/text and graph neural network models, as well as existing state-of-the-art hybrid strategies that combine the two.