论文标题
forml:学习重新加权数据以获得公平性
FORML: Learning to Reweight Data for Fairness
论文作者
论文摘要
对机器学习模型进行了训练,以最大程度地减少单个度量标准的平均损失,因此通常不考虑公平性和鲁棒性。当培训数据不平衡或测试分布不同时,忽略培训中的这种指标可能会使这些模型容易违反公平。这项工作通过元学习(FormL)引入了公平性优化的重新加权,这是一种培训算法,通过共同学习培训样本权重和神经网络参数来平衡公平和鲁棒性与准确性。该方法通过学会通过动态重新重新加权从用户指定的持有集合中学到的数据来代表所需的分布的分布来平衡分布的数据,从而提高了模型公平性。 Forml提高了图像分类任务上的机会公平标准的平等性,减少了损坏的标签的偏见,并通过数据凝结来促进建立更多公平数据集。这些改进是在没有预处理数据或后处理模型输出的情况下实现的,而无需学习额外的加权函数,而无需更改模型体系结构,而在原始预测度量方面保持准确性。
Machine learning models are trained to minimize the mean loss for a single metric, and thus typically do not consider fairness and robustness. Neglecting such metrics in training can make these models prone to fairness violations when training data are imbalanced or test distributions differ. This work introduces Fairness Optimized Reweighting via Meta-Learning (FORML), a training algorithm that balances fairness and robustness with accuracy by jointly learning training sample weights and neural network parameters. The approach increases model fairness by learning to balance the contributions from both over- and under-represented sub-groups through dynamic reweighting of the data learned from a user-specified held-out set representative of the distribution under which fairness is desired. FORML improves equality of opportunity fairness criteria on image classification tasks, reduces bias of corrupted labels, and facilitates building more fair datasets via data condensation. These improvements are achieved without pre-processing data or post-processing model outputs, without learning an additional weighting function, without changing model architecture, and while maintaining accuracy on the original predictive metric.