论文标题

在多个利益相关者分布中衡量和签署公平作为绩效

Measuring and signing fairness as performance under multiple stakeholder distributions

论文作者

Lopez-Paz, David, Bouchacourt, Diane, Sagun, Levent, Usunier, Nicolas

论文摘要

随着学习机器对人类生命的决策的影响,分析其公平性能成为重要性的主题。然而,我们衡量学习系统公平性的最佳工具是将数学单线封装的刚性公平指标,为参与预测任务的利益相关者提供了有限的权力,并且当我们敦促过度压力以优化它们时,很容易操纵。为了促进这些问题,我们建议将重点从塑造公平度量指标转变为策划计算这些示例的分布。特别是,我们认为,关于公平性的每一个主张都应立即遵循标语“在哪些例子下公平,并由谁收集?”。通过强调与域概括中文献的联系,我们建议衡量公平性作为系统在多个压力测试下概括的能力 - 具有社会相关性的示例的分布。我们鼓励每个利益相关者策划一个或多个压力测试,其中包含反映其(可能相互矛盾的)利益的例子。该机器通过缺乏或超过预定义的度量值而通过或未能使每个压力测试失败。测试结果涉及所有利益相关者参与有关如何改善学习系统的讨论,并根据上下文和基于可解释的数据提供对公平性的灵活评估。我们为压力测试提供了完整的实施指南,既说明了该框架的好处和缺点,又引入了一个加密计划,以使系统提供商获得一定程度的预测问责制。

As learning machines increase their influence on decisions concerning human lives, analyzing their fairness properties becomes a subject of central importance. Yet, our best tools for measuring the fairness of learning systems are rigid fairness metrics encapsulated as mathematical one-liners, offer limited power to the stakeholders involved in the prediction task, and are easy to manipulate when we exhort excessive pressure to optimize them. To advance these issues, we propose to shift focus from shaping fairness metrics to curating the distributions of examples under which these are computed. In particular, we posit that every claim about fairness should be immediately followed by the tagline "Fair under what examples, and collected by whom?". By highlighting connections to the literature in domain generalization, we propose to measure fairness as the ability of the system to generalize under multiple stress tests -- distributions of examples with social relevance. We encourage each stakeholder to curate one or multiple stress tests containing examples reflecting their (possibly conflicting) interests. The machine passes or fails each stress test by falling short of or exceeding a pre-defined metric value. The test results involve all stakeholders in a discussion about how to improve the learning system, and provide flexible assessments of fairness dependent on context and based on interpretable data. We provide full implementation guidelines for stress testing, illustrate both the benefits and shortcomings of this framework, and introduce a cryptographic scheme to enable a degree of prediction accountability from system providers.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源