论文标题

Fairneuron:通过选择性神经元的对手游戏改善深层神经网络公平性

FairNeuron: Improving Deep Neural Network Fairness with Adversary Games on Selective Neurons

论文作者

Gao, Xuanqi, Zhai, Juan, Ma, Shiqing, Shen, Chao, Chen, Yufei, Wang, Qian

论文摘要

随着深度神经网络(DNN)被整合到越来越多的关键系统中,对社会产生了深远的影响,人们对其道德表现(例如公平性)的关注越来越多。不幸的是,在许多情况下,模型的公平性和准确性是优化的矛盾目标。为了解决这个问题,已经有许多工作试图通过在模型级别使用对抗性游戏来改善模型公平性。这种方法引入了一个对手,该对手除了对主要任务的预测准确性外评估模型的公平性,并执行联合优化以实现平衡结果。在本文中,我们注意到,在进行基于向后传播的训练时,这种矛盾的现象已在单个神经元水平上显示。基于此观察结果,我们提出了一种DNN模型自动修复工具Fairneuron,以减轻公平关注并平衡准确性的权衡,而无需引入其他模型。它致力于从准确性和公平训练目标中检测具有矛盾优化方向的神经元,并通过选择性辍学实现权衡。与最先进的方法相比,我们的方法轻巧,使其可扩展和更有效。我们在3个数据集上的评估表明,Fairneuron可以有效地改善所有模型的公平性,同时保持稳定的实用程序。

With Deep Neural Network (DNN) being integrated into a growing number of critical systems with far-reaching impacts on society, there are increasing concerns on their ethical performance, such as fairness. Unfortunately, model fairness and accuracy in many cases are contradictory goals to optimize. To solve this issue, there has been a number of work trying to improve model fairness by using an adversarial game in model level. This approach introduces an adversary that evaluates the fairness of a model besides its prediction accuracy on the main task, and performs joint-optimization to achieve a balanced result. In this paper, we noticed that when performing backward propagation based training, such contradictory phenomenon has shown on individual neuron level. Based on this observation, we propose FairNeuron, a DNN model automatic repairing tool, to mitigate fairness concerns and balance the accuracy-fairness trade-off without introducing another model. It works on detecting neurons with contradictory optimization directions from accuracy and fairness training goals, and achieving a trade-off by selective dropout. Comparing with state-of-the-art methods, our approach is lightweight, making it scalable and more efficient. Our evaluation on 3 datasets shows that FairNeuron can effectively improve all models' fairness while maintaining a stable utility.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源