Elsevier

Pattern Recognition Letters

Volume 140, December 2020, Pages 332-338
Pattern Recognition Letters

Post-comparison mitigation of demographic bias in face recognition using fair score normalization

https://doi.org/10.1016/j.patrec.2020.11.007Get rights and content

Highlights

  • Enhance face verification performance for demographics effected by decision bias.

  • Operates on comparison score-level to enable using existing face recognition models.

  • Based on the notation of individual fairness to treat similar face groups similarly.

  • Effectiveness proved on three databases and two face embeddings.

  • Unsupervised enhancement of the overall and intra-class recognition performance.

Abstract

Current face recognition systems achieve high progress on several benchmark tests. Despite this progress, recent works showed that these systems are strongly biased against demographic sub-groups. Consequently, an easily integrable solution is needed to reduce the discriminatory effect of these biased systems. Previous work mainly focused on learning less biased face representations, which comes at the cost of a strongly degraded overall recognition performance. In this work, we propose a novel unsupervised fair score normalization approach that is specifically designed to reduce the effect of bias in face recognition and subsequently lead to a significant overall performance boost. Our hypothesis is built on the notation of individual fairness by designing a normalization approach that leads to treating “similar” individuals “similarly”. Experiments were conducted on three publicly available datasets captured under controlled and in-the-wild circumstances. Results demonstrate that our solution reduces demographic biases, e.g. by up to 82.7% in the case when gender is considered. Moreover, it mitigates the bias more consistently than existing works. In contrast to previous works, our fair normalization approach enhances the overall performance by up to 53.2% at false match rate of 103 and up to 82.9% at a false match rate of 105. Additionally, it is easily integrable into existing recognition systems and not limited to face biometrics.

Introduction

Large-scale face recognition systems are spreading worldwide and are increasingly involved in critical decision-making processes, such as in forensics and law enforcement. Consequently, these systems also have a growing effect on everybody’s daily life. However, current biometric solutions are mainly optimized for maximum recognition accuracy [1] and are heavily biased for certain demographic groups [2], [3], [4], [5], [6], [7]. This means that, for example, specific demographic groups can be falsely identified as black-listed individuals more frequently than other groups. Consequently, there is an increased need that guarantees fairness for biometric solutions [6], [8], [9] to prevent discriminatory decisions.

From a political perspective, there are several regulations to guarantee fairness. Article 7 of the Universal Declaration on Human Rights and Article 14 of the European Convention of Human Rights ensure people the right to non-discrimination. Also the General Data Protection Regulation (GDPR) [10] aims at preventing discriminatory effects (article 71). In spite of these political efforts, several works [2], [3], [4], [5], [6], [7] showed that commercial [6], as well as open-source [2] face recognition systems, are strongly biased towards different demographic groups. Consequently, there is an increased need for fair and unbiased biometric solutions [2], [7].

Recent works mainly focused on learning less-biased face representations [11], [12], [13], [14], [15], [16], [17] for specific demographics. However, this requires computationally expensive template-replacement of the whole database if the recognition system is updated. Moreover, the bias-mitigation often comes at the cost of a strong decrease in recognition performance.

In this work, we propose a novel and unsupervised fair score normalization bias mitigation approach that is easily-integrable. Unlike previous work, increasing fairness also leads to an improved performance of the system in total. Our theoretical motivation is based on the notation of individual fairness [18], resulting in a solution that treats similar individuals similarly and thus, more fairly. The proposed approach clusters samples in the embedding space such that similar identities are categorized without the need for pre-defined demographic classes. For each cluster, an optimal local threshold is computed and used to develop a score normalization approach that ensures a more individual, unbiased, and fair treatment. The experiments are conducted on three publicly available datasets captured under controlled and in-the-wild conditions and on two face embeddings. To justify the concept of our fair normalization approach, we provide a visual illustration that demonstrates (a) the suitability of the notation of individual fairness for face recognition and (b) the need for more individualized treatment of face recognition systems. The results show a higher consistency and efficiency of our unsupervised normalization approach compared to related works. It efficiently mitigates demographic-bias by up to 82.7% while consistently enhancing the total recognition performance by up to 82.9%. The source code for this work is available at the following link.1

Section snippets

Related work

Bias in biometrics was found in several disciplines such as presentation attack detection [19], biometric image quality estimation [20], and the estimation of facial characteristics [21], [22], [23]. In face biometrics, bias might be induced by non-equally distributed classes in training data [14], [16]. This results in face recognition performances that are strongly influenced by demographic attributes [24]. These findings motivated the recent research towards mitigating bias in face

Methodology

The goal of this work is to enhance the fairness of existing face recognition systems in an easily-integrable manner. In this work, we follow the notation of individual fairness [18]. This notation emphasizes that similar individuals should be treated similarly. We transfer this idea to the embedding and score level to propose a novel fair group-based score normalization method, without the need for pre-defined demographic groups. The proposed approach is able to treat all identities more

Experimental setup

Database In order to evaluate the face recognition performance of our approach under controlled and unconstrained conditions, we conducted experiments on the public available Adience [30], ColorFeret [31], and Morph [32] datasets. ColorFeret [31] consists of 14,126 images from 1199 different individuals with different poses under controlled conditions. Furthermore, a variety of face poses, facial expressions, and lighting conditions are included in the dataset. The Adience dataset [30] consists

Visual demonstration of the need for individuality

Since our approach is based on the idea of individual fairness, we first want to visually demonstrate why this notation is suitable for face recognition. Fig. 1 shows an t-SNE visualization of the embedding space for the dataset Adience. The t-SNE algorithm maps the high-dimensional embedding space into a two-dimensional space such that similar samples in the high-dimension space lie closely together in two dimensions. Furthermore, each sample is colored based on the local thresholds computed

Conclusion

Despite the progress achieved by current face recognition systems, recent works showed that biometric systems impose a strong bias against subgroups of the population. Consequently, there is an increased need for solutions that increase the fairness of such systems. Previous works focused on learning bias-mitigated face representations. However, these solutions are often hardly-integrable and degrade the overall recognition performance. In this work, we propose a novel fair score normalization

Declaration of Competing Interest

None.

Acknowledgments

This work was supported by the German Federal Ministry of Education and Research (BMBF) as well as by the Hessen State Ministry for Higher Education, Research and the Arts (HMWK) within the National Research Center for Applied Cybersecurity (ATHENE). Portions of the research in this paper use the FERET database of facial images collected under the FERET program, sponsored by the Counterdrug Technology Development Program Office.

References (39)

  • P. Voigt et al.

    The EU General Data Protection Regulation (GDPR): A Practical Guide

    (2017)
  • S. Gong, X. Liu, A.K. Jain, DebFace: De-biasing face recognition, CoRR...
  • J. Liang et al.

    Additive adversarial learning for unbiased authentication

    The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)

    (2019)
  • M. Wang, W. Deng, Mitigate bias in face recognition using skewness-aware reinforcement learning, CoRR abs/1911.10692...
  • C. Huang, Y. Li, C.C. Loy, X. Tang, Deep imbalanced learning for face recognition and attribute prediction, CoRR...
  • A. Kortylewski et al.

    Analyzing and reducing the damage of dataset bias to face recognition with synthetic data

    The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops

    (2019)
  • X. Yin et al.

    Feature transfer learning for face recognition with under-represented data

    IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2019, Long Beach, CA, USA, June 16–20, 2019

    (2019)
  • C. Dwork et al.

    Fairness through awareness

    Proceedings of the 3rd Innovations in Theoretical Computer Science Conference, ITCS ’12

    (2012)
  • M. Fang, N. Damer, F. Kirchbuchner, A. Kuijper, Demographic bias in presentation attack detection of iris recognition...
  • Cited by (28)

    • Ring-Regularized Cosine Similarity Learning for Fine-Grained Face Verification

      2021, Pattern Recognition Letters
      Citation Excerpt :

      Face recognition is one of the popular research topics in computer vision and pattern recognition and it has been comprehensively studied for decades [1,9,12,13,15,18,20,21,23,27,30,34].

    • Brief Audit of Post-pandemic Biometrics

      2023, Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
    View all citing articles on Scopus
    View full text