论文标题

通过概率分类进行回归的公平措施

Fairness Measures for Regression via Probabilistic Classification

论文作者

Steinberg, Daniel, Reid, Alistair, O'Callaghan, Simon

论文摘要

算法公平性涉及表达诸如公平或合理处理之类的概念,作为机器学习算法可以优化的可量化措施。迄今为止,大多数文献中的大多数工作都集中在预测是分类的分类问题上,例如接受或拒绝贷款申请。这部分是因为分类公平措施可以通过比较结果速度来计算,从而导致行为,例如确保将相同的合格男人选择作为合格女性。但是,这些措施在计算上很难推广到定价或分配付款等问题的连续回归设置。难度是由估计条件密度的(例如,系统将超过一定量的概率密度)。对于回归设置,我们通过观察将其分解为受保护属性的不同条件概率的比率来引入独立性,分离和充分性标准的可行近似值。我们介绍和训练机器学习分类器(与预测指标)不同,是一种从数据中估算这些概率的机制。这自然会导致对标准的不可知论,可拖动的近似值,我们通过实验探索。

Algorithmic fairness involves expressing notions such as equity, or reasonable treatment, as quantifiable measures that a machine learning algorithm can optimise. Most work in the literature to date has focused on classification problems where the prediction is categorical, such as accepting or rejecting a loan application. This is in part because classification fairness measures are easily computed by comparing the rates of outcomes, leading to behaviours such as ensuring that the same fraction of eligible men are selected as eligible women. But such measures are computationally difficult to generalise to the continuous regression setting for problems such as pricing, or allocating payments. The difficulty arises from estimating conditional densities (such as the probability density that a system will over-charge by a certain amount). For the regression setting we introduce tractable approximations of the independence, separation and sufficiency criteria by observing that they factorise as ratios of different conditional probabilities of the protected attributes. We introduce and train machine learning classifiers, distinct from the predictor, as a mechanism to estimate these probabilities from the data. This naturally leads to model agnostic, tractable approximations of the criteria, which we explore experimentally.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源