论文标题

通过变异推断学习大规模的文本属性图

Learning on Large-scale Text-attributed Graphs via Variational Inference

论文作者

Zhao, Jianan, Qu, Meng, Li, Chaozhuo, Yan, Hao, Liu, Qian, Li, Rui, Xie, Xing, Tang, Jian

论文摘要

本文研究了有关文本属性图(标签)的学习,其中每个节点都与文本描述相关联。解决此类问题的理想解决方案是将文本和图形结构信息与大语言模型和图形神经网络(GNN)集成在一起。但是,当图形大大时,由于训练大型语言模型和GNN的高度计算复杂性,问题变得非常具有挑战性。在本文中,我们通过将图形结构和语言学习与各种期望最大化(EM)框架(称为Glem)融合,为大型文本归属图的学习提出了一个有效的解决方案。 Glem并没有同时在大图上训练大型语言模型和GNN,而是建议更新E-Step和M-Step中的两个模块。这样的过程允许分别训练两个模块,同时允许两个模块相互作用并相互增强。对多个数据集的广泛实验证明了所提出方法的效率和有效性。

This paper studies learning on text-attributed graphs (TAGs), where each node is associated with a text description. An ideal solution for such a problem would be integrating both the text and graph structure information with large language models and graph neural networks (GNNs). However, the problem becomes very challenging when graphs are large due to the high computational complexity brought by training large language models and GNNs together. In this paper, we propose an efficient and effective solution to learning on large text-attributed graphs by fusing graph structure and language learning with a variational Expectation-Maximization (EM) framework, called GLEM. Instead of simultaneously training large language models and GNNs on big graphs, GLEM proposes to alternatively update the two modules in the E-step and M-step. Such a procedure allows training the two modules separately while simultaneously allowing the two modules to interact and mutually enhance each other. Extensive experiments on multiple data sets demonstrate the efficiency and effectiveness of the proposed approach.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源