论文标题
图形卷积网络的自我监督培训
Self-supervised Training of Graph Convolutional Networks
论文作者
论文摘要
图形卷积网络(GCN)已成功应用于分析非网格数据,而经典的卷积神经网络(CNN)无法直接使用。 GCN和CNN共享的一个相似性是需要大量标记数据进行网络培训的数据。此外,GCN需要邻接矩阵作为输入来定义这些非网格数据之间的关系,这导致所有数据包括培训,验证和测试数据,通常仅形成一个用于培训的图形结构。此外,邻接矩阵通常是预先定义和固定的,这使得数据增强策略不能在构造的图形结构数据上使用以增加培训数据的量。为了进一步提高在有限的培训数据下的学习能力和模型性能,在本文中,我们提出了两种类型的自我监管的学习策略,以从输入图结构数据本身中利用可用信息。我们提出的自我监督学习策略将在两个代表性的GCN模型上使用三个公共引用网络数据集(Citeseer,Cora和PubMed)进行了检查。实验结果证明了我们提出的策略的概括能力以及可移植性,这可以显着提高GCN的性能,并具有自我监督学习的力量,以改善特征学习。
Graph Convolutional Networks (GCNs) have been successfully applied to analyze non-grid data, where the classical convolutional neural networks (CNNs) cannot be directly used. One similarity shared by GCNs and CNNs is the requirement of massive amount of labeled data for network training. In addition, GCNs need the adjacency matrix as input to define the relationship between those non-grid data, which leads to all of data including training, validation and test data typically forms only one graph structures data for training. Furthermore, the adjacency matrix is usually pre-defined and stationary, which makes the data augmentation strategies cannot be employed on the constructed graph structures data to augment the amount of training data. To further improve the learning capacity and model performance under the limited training data, in this paper, we propose two types of self-supervised learning strategies to exploit available information from the input graph structure data itself. Our proposed self-supervised learning strategies are examined on two representative GCN models with three public citation network datasets - Citeseer, Cora and Pubmed. The experimental results demonstrate the generalization ability as well as the portability of our proposed strategies, which can significantly improve the performance of GCNs with the power of self-supervised learning in improving feature learning.