论文标题
基于图的半监督学习的对比和生成图卷积网络
Contrastive and Generative Graph Convolutional Networks for Graph-based Semi-Supervised Learning
论文作者
论文摘要
基于图的半监督学习(SSL)旨在通过图将少数标记数据的标签传输到剩余的大量未标记数据。作为最受欢迎的基于图的SSL方法之一,最近提出的图形卷积网络(GCN)通过将神经网络的声音表现力与图形结构相结合,从而取得了显着的进步。但是,现有的基于图的方法并未直接解决SSL的核心问题,即监督短缺,因此它们的性能仍然非常有限。为了解决这个问题,本文提出了一种基于GCN的新型SSL算法,以通过使用数据相似性和图形结构来丰富监督信号。首先,通过设计半监视的对比损失,可以通过最大化同一数据的不同视图或同一类的数据之间的一致性来生成改进的节点表示形式。因此,丰富的未标记数据和稀缺但有价值的标签数据可以共同提供大量的监督信息,以学习歧视性节点表示,这有助于改善后续的分类结果。其次,通过使用与输入特征相关的图生成损失,将数据特征与输入图拓扑之间的基本确定关系作为SSL的补充监督信号。与其他最先进的方法相比,对各种现实数据集的深入实验结果牢固地验证了我们算法的有效性。
Graph-based Semi-Supervised Learning (SSL) aims to transfer the labels of a handful of labeled data to the remaining massive unlabeled data via a graph. As one of the most popular graph-based SSL approaches, the recently proposed Graph Convolutional Networks (GCNs) have gained remarkable progress by combining the sound expressiveness of neural networks with graph structure. Nevertheless, the existing graph-based methods do not directly address the core problem of SSL, i.e., the shortage of supervision, and thus their performances are still very limited. To accommodate this issue, a novel GCN-based SSL algorithm is presented in this paper to enrich the supervision signals by utilizing both data similarities and graph structure. Firstly, by designing a semi-supervised contrastive loss, improved node representations can be generated via maximizing the agreement between different views of the same data or the data from the same class. Therefore, the rich unlabeled data and the scarce yet valuable labeled data can jointly provide abundant supervision information for learning discriminative node representations, which helps improve the subsequent classification result. Secondly, the underlying determinative relationship between the data features and input graph topology is extracted as supplementary supervision signals for SSL via using a graph generative loss related to the input features. Intensive experimental results on a variety of real-world datasets firmly verify the effectiveness of our algorithm compared with other state-of-the-art methods.