论文标题
对歧视说否:学习敏感属性信息有限的学习公平图神经网络
Say No to the Discrimination: Learning Fair Graph Neural Networks with Limited Sensitive Attribute Information
论文作者
论文摘要
图形神经网络(GNN)在建模图结构数据方面表现出很大的功能。但是,与其他机器学习模型类似,GNN可能会对受保护的敏感属性(例如肤色和性别)产生偏见。因为对包括GNN在内的机器学习算法进行了训练,以反映培训数据的分布,这些数据通常包含对敏感属性的历史偏见。另外,GNN中的歧视可以通过图形结构和通信机制放大。结果,GNN在诸如犯罪率预测之类的敏感领域中的应用将在很大程度上受到限制。尽管已经对I.I.D数据进行了公平分类的广泛研究,但解决非I.I.D数据歧视问题的方法是相当有限的。此外,在现有作品中很少考虑敏感属性中稀疏注释的实际情况。因此,我们研究具有有限敏感属性信息的学习公平GNN的新颖而重要的问题。提议FAIRGNN通过利用图形结构和有限的敏感信息来消除GNN的偏差,同时维持高节点分类的精度。我们的理论分析表明,在有限的敏感属性的有限的节点下,pabrgnn可以确保在轻度条件下GNN的公平性。对现实世界数据集的广泛实验也证明了pabrgnn在依据和保持高准确性方面的有效性。
Graph neural networks (GNNs) have shown great power in modeling graph structured data. However, similar to other machine learning models, GNNs may make predictions biased on protected sensitive attributes, e.g., skin color and gender. Because machine learning algorithms including GNNs are trained to reflect the distribution of the training data which often contains historical bias towards sensitive attributes. In addition, the discrimination in GNNs can be magnified by graph structures and the message-passing mechanism. As a result, the applications of GNNs in sensitive domains such as crime rate prediction would be largely limited. Though extensive studies of fair classification have been conducted on i.i.d data, methods to address the problem of discrimination on non-i.i.d data are rather limited. Furthermore, the practical scenario of sparse annotations in sensitive attributes is rarely considered in existing works. Therefore, we study the novel and important problem of learning fair GNNs with limited sensitive attribute information. FairGNN is proposed to eliminate the bias of GNNs whilst maintaining high node classification accuracy by leveraging graph structures and limited sensitive information. Our theoretical analysis shows that FairGNN can ensure the fairness of GNNs under mild conditions given limited nodes with known sensitive attributes. Extensive experiments on real-world datasets also demonstrate the effectiveness of FairGNN in debiasing and keeping high accuracy.