论文标题

公平SA:敏感性分析面部识别的公平性

Fair SA: Sensitivity Analysis for Fairness in Face Recognition

论文作者

Joshi, Aparna R., Suau, Xavier, Sivakumar, Nivedha, Zappella, Luca, Apostoloff, Nicholas

论文摘要

随着在高影响域中使用深度学习变得无处不在,评估模型的弹性越来越重要。这样的高影响域是面部识别,现实世界的应用涉及受各种降解影响的图像,例如运动模糊或高暴露。此外,跨性别和种族等不同属性捕获的图像也可以挑战面部识别算法的鲁棒性。尽管传统的摘要统计数据表明,面部识别模型的总体表现持续改善,但这些指标并不能直接衡量模型的鲁棒性或公平性。视觉心理物理学敏感性分析(VPSA)[1]提供了一种方法,通过引入数据中的增量扰动来查明单个失败的原因。但是,扰动可能对亚组的影响有所不同。在本文中,我们提出了一种基于鲁棒性的新公平评估,以扩展VPSA的通用框架的形式。使用此框架,我们可以分析模型对受扰动影响的人群的不同亚组的表现的能力,并通过测量目标鲁棒性来查明亚组的确切故障模式。随着对模型公平性的越来越重视,我们将面部识别作为我们框架的示例应用程序,并建议通过AUC矩阵紧凑地可视化模型的公平分析。我们分析了共同的面部识别模型的性能,并从经验上表明,当图像受到干扰时,某些亚组处于不利影响,从而发现了使用该模型在没有扰动的亚组上的模型性能的趋势。

As the use of deep learning in high impact domains becomes ubiquitous, it is increasingly important to assess the resilience of models. One such high impact domain is that of face recognition, with real world applications involving images affected by various degradations, such as motion blur or high exposure. Moreover, images captured across different attributes, such as gender and race, can also challenge the robustness of a face recognition algorithm. While traditional summary statistics suggest that the aggregate performance of face recognition models has continued to improve, these metrics do not directly measure the robustness or fairness of the models. Visual Psychophysics Sensitivity Analysis (VPSA) [1] provides a way to pinpoint the individual causes of failure by way of introducing incremental perturbations in the data. However, perturbations may affect subgroups differently. In this paper, we propose a new fairness evaluation based on robustness in the form of a generic framework that extends VPSA. With this framework, we can analyze the ability of a model to perform fairly for different subgroups of a population affected by perturbations, and pinpoint the exact failure modes for a subgroup by measuring targeted robustness. With the increasing focus on the fairness of models, we use face recognition as an example application of our framework and propose to compactly visualize the fairness analysis of a model via AUC matrices. We analyze the performance of common face recognition models and empirically show that certain subgroups are at a disadvantage when images are perturbed, thereby uncovering trends that were not visible using the model's performance on subgroups without perturbations.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源