论文标题

关于混合池在基于混合的图表学习中用于语言处理的有效性

On the Effectiveness of Hybrid Pooling in Mixup-Based Graph Learning for Language Processing

论文作者

Dong, Zeming, Hu, Qiang, Zhang, Zhenya, Guo, Yuejun, Cordy, Maxime, Papadakis, Mike, Traon, Yves Le, Zhao, Jianjun

论文摘要

基于图形神经网络(GNN)的图形学习在自然语言和编程语言处理中很受欢迎,尤其是在文本和源代码分类中。通常,GNN是通过合并交替的图层来构建的,该图层学习图形节点特征的转换,以及使用图形池池操作员(例如,最大值)的图池层层,以有效地减少节点的数量,同时保留图形的语义信息。最近,为了增强图形学习任务中的GNN,通过线性混合一对图数据及其标签来产生合成图数据的数据增强技术,已被广泛采用。但是,歧管混合的性能可能会受到图形合并操作员的高度影响,而且没有很多致力于发现这种感情的研究。为了弥合这一差距,我们采取了早期步骤来探讨图形池操作员如何影响基于混合的图形学习的性能。为此,我们通过将歧管混合应用于基于11个图形池操作(9个混合池操作员,2个非Hybrid池操作员)的形式表征来进行全面的经验研究。自然语言数据集(GossipCop,Politifact)和编程语言数据集(JAVA250,Python800)的实验结果表明,与标准的最大最大泵和最先进的图形图形多功能图(GMT)相比,Hybrid合并操作员对歧管混合更有效。

Graph neural network (GNN)-based graph learning has been popular in natural language and programming language processing, particularly in text and source code classification. Typically, GNNs are constructed by incorporating alternating layers which learn transformations of graph node features, along with graph pooling layers that use graph pooling operators (e.g., Max-pooling) to effectively reduce the number of nodes while preserving the semantic information of the graph. Recently, to enhance GNNs in graph learning tasks, Manifold-Mixup, a data augmentation technique that produces synthetic graph data by linearly mixing a pair of graph data and their labels, has been widely adopted. However, the performance of Manifold-Mixup can be highly affected by graph pooling operators, and there have not been many studies that are dedicated to uncovering such affection. To bridge this gap, we take an early step to explore how graph pooling operators affect the performance of Mixup-based graph learning. To that end, we conduct a comprehensive empirical study by applying Manifold-Mixup to a formal characterization of graph pooling based on 11 graph pooling operations (9 hybrid pooling operators, 2 non-hybrid pooling operators). The experimental results on both natural language datasets (Gossipcop, Politifact) and programming language datasets (JAVA250, Python800) demonstrate that hybrid pooling operators are more effective for Manifold-Mixup than the standard Max-pooling and the state-of-the-art graph multiset transformer (GMT) pooling, in terms of producing more accurate and robust GNN models.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源