论文标题
攻击图结构时梯度告诉什么
What Does the Gradient Tell When Attacking the Graph Structure
论文作者
论文摘要
最近的研究表明,图形神经网络(GNN)容易受到针对图结构的对抗性攻击。鉴于培训标签,恶意攻击者可以操纵有限数量的边缘,以损害受害者模型的表现。先前的经验研究表明,基于梯度的攻击者倾向于增加边缘而不是删除边缘。在本文中,我们提出了一个理论上的演示,表明攻击者由于GNN的信息传递机制而倾向于增加阶层间边缘,这解释了一些先前的经验观察。通过连接不同的节点,攻击者可以更有效地损坏节点特征,从而使此类攻击更加有利。但是,我们证明了GNN信息传递的固有平滑度倾向于模糊特征空间中的差异性,从而导致在正向过程中丢失了重要信息。为了解决这个问题,我们提出了一个具有多层次传播的新型替代模型,以保留节点差异信息。该模型使未聚集的原始特征和多跳聚合特征的传播同时引入批处理归一化以增强节点表示中的差异,并抵消拓扑聚合产生的平滑度。我们的实验在我们的方法中表现出显着的改善。Furthermore,理论和实验证据都表明,添加一流的边缘构成了易于观察的攻击模式。我们提出了一种创新的攻击损失,以平衡攻击效率和不可识别性,牺牲了一些攻击效率以实现更大的不可识别。我们还提供实验,以验证通过此攻击损失实现的折衷性绩效。
Recent research has revealed that Graph Neural Networks (GNNs) are susceptible to adversarial attacks targeting the graph structure. A malicious attacker can manipulate a limited number of edges, given the training labels, to impair the victim model's performance. Previous empirical studies indicate that gradient-based attackers tend to add edges rather than remove them. In this paper, we present a theoretical demonstration revealing that attackers tend to increase inter-class edges due to the message passing mechanism of GNNs, which explains some previous empirical observations. By connecting dissimilar nodes, attackers can more effectively corrupt node features, making such attacks more advantageous. However, we demonstrate that the inherent smoothness of GNN's message passing tends to blur node dissimilarity in the feature space, leading to the loss of crucial information during the forward process. To address this issue, we propose a novel surrogate model with multi-level propagation that preserves the node dissimilarity information. This model parallelizes the propagation of unaggregated raw features and multi-hop aggregated features, while introducing batch normalization to enhance the dissimilarity in node representations and counteract the smoothness resulting from topological aggregation. Our experiments show significant improvement with our approach.Furthermore, both theoretical and experimental evidence suggest that adding inter-class edges constitutes an easily observable attack pattern. We propose an innovative attack loss that balances attack effectiveness and imperceptibility, sacrificing some attack effectiveness to attain greater imperceptibility. We also provide experiments to validate the compromise performance achieved through this attack loss.