论文标题
与准确的差异学习图表自我监督学习
Graph Self-supervised Learning with Accurate Discrepancy Learning
论文作者
论文摘要
图形神经网络(GNN)的自我监督学习旨在以无监督的方式学习图形的准确表示,以获取它们的可转移表示形式,以实现各种下游任务。预测性学习和对比度学习是图形自学学习的两种最普遍的方法。但是,他们有自己的缺点。尽管预测学习方法可以学习相邻节点和边缘之间的上下文关系,但他们无法学习全局的图形级别相似性。对比学习,尽管它可以学习全局图级相似性,但其目标是最大化两个不同扰动图之间的相似性,可能会导致表示无法区分两个具有不同属性的相似图的表示。为了解决此类限制,我们提出了一个框架,旨在了解原始图和扰动图之间的确切差异,这是基于差异的自我监督学习(D-SLA)。具体而言,我们以不同程度的相似性创建给定图的多个扰动,并训练模型以预测每个图是原始图还是扰动的图。此外,我们进一步旨在使用图表编辑距离准确捕获每个扰动图的差异。我们在各种与图形相关的下游任务上验证了D-SLA,包括分子属性预测,蛋白质功能预测和链接预测任务,我们在这些任务上很大程度上胜过相关的基准。
Self-supervised learning of graph neural networks (GNNs) aims to learn an accurate representation of the graphs in an unsupervised manner, to obtain transferable representations of them for diverse downstream tasks. Predictive learning and contrastive learning are the two most prevalent approaches for graph self-supervised learning. However, they have their own drawbacks. While the predictive learning methods can learn the contextual relationships between neighboring nodes and edges, they cannot learn global graph-level similarities. Contrastive learning, while it can learn global graph-level similarities, its objective to maximize the similarity between two differently perturbed graphs may result in representations that cannot discriminate two similar graphs with different properties. To tackle such limitations, we propose a framework that aims to learn the exact discrepancy between the original and the perturbed graphs, coined as Discrepancy-based Self-supervised LeArning (D-SLA). Specifically, we create multiple perturbations of the given graph with varying degrees of similarity, and train the model to predict whether each graph is the original graph or the perturbed one. Moreover, we further aim to accurately capture the amount of discrepancy for each perturbed graph using the graph edit distance. We validate our D-SLA on various graph-related downstream tasks, including molecular property prediction, protein function prediction, and link prediction tasks, on which ours largely outperforms relevant baselines.