Quantifying Bias in a Face Verification System

Machine learning models perform face verification (FV) for a variety of highly consequential applications, such as biometric authentication, face identification, and surveillance. Many state-of-the-art FV systems suffer from unequal performance across demographic groups, which is commonly overlooked...

Description complète

Détails bibliographiques
Auteurs principaux: Frisella, Megan, Khorrami, Pooya, Matterer, Jason, Kratkiewicz, Kendra, Torres-Carrasquillo, Pedro
Autres auteurs: Lincoln Laboratory
Format: Article
Publié: Multidisciplinary Digital Publishing Institute 2022
Accès en ligne:https://hdl.handle.net/1721.1/142034
_version_ 1826203055227404288
author Frisella, Megan
Khorrami, Pooya
Matterer, Jason
Kratkiewicz, Kendra
Torres-Carrasquillo, Pedro
author2 Lincoln Laboratory
author_facet Lincoln Laboratory
Frisella, Megan
Khorrami, Pooya
Matterer, Jason
Kratkiewicz, Kendra
Torres-Carrasquillo, Pedro
author_sort Frisella, Megan
collection MIT
description Machine learning models perform face verification (FV) for a variety of highly consequential applications, such as biometric authentication, face identification, and surveillance. Many state-of-the-art FV systems suffer from unequal performance across demographic groups, which is commonly overlooked by evaluation measures that do not assess population-specific performance. Deployed systems with bias may result in serious harm against individuals or groups who experience underperformance. We explore several fairness definitions and metrics, attempting to quantify bias in Google’s FaceNet model. In addition to statistical fairness metrics, we analyze clustered face embeddings produced by the FV model. We link well-clustered embeddings (well-defined, dense clusters) for a demographic group to biased model performance against that group. We present the intuition that FV systems underperform on protected demographic groups because they are less sensitive to differences between features within those groups, as evidenced by clustered embeddings. We show how this performance discrepancy results from a combination of representation and aggregation bias.
first_indexed 2024-09-23T12:30:40Z
format Article
id mit-1721.1/142034
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T12:30:40Z
publishDate 2022
publisher Multidisciplinary Digital Publishing Institute
record_format dspace
spelling mit-1721.1/1420342023-02-14T19:43:47Z Quantifying Bias in a Face Verification System Frisella, Megan Khorrami, Pooya Matterer, Jason Kratkiewicz, Kendra Torres-Carrasquillo, Pedro Lincoln Laboratory Machine learning models perform face verification (FV) for a variety of highly consequential applications, such as biometric authentication, face identification, and surveillance. Many state-of-the-art FV systems suffer from unequal performance across demographic groups, which is commonly overlooked by evaluation measures that do not assess population-specific performance. Deployed systems with bias may result in serious harm against individuals or groups who experience underperformance. We explore several fairness definitions and metrics, attempting to quantify bias in Google’s FaceNet model. In addition to statistical fairness metrics, we analyze clustered face embeddings produced by the FV model. We link well-clustered embeddings (well-defined, dense clusters) for a demographic group to biased model performance against that group. We present the intuition that FV systems underperform on protected demographic groups because they are less sensitive to differences between features within those groups, as evidenced by clustered embeddings. We show how this performance discrepancy results from a combination of representation and aggregation bias. 2022-04-25T12:36:10Z 2022-04-25T12:36:10Z 2022-04-20 2022-04-21T21:03:49Z Article http://purl.org/eprint/type/JournalArticle https://hdl.handle.net/1721.1/142034 Computer Sciences & Mathematics Forum 3 (1): 6 (2022) PUBLISHER_CC http://dx.doi.org/10.3390/cmsf2022003006 Creative Commons Attribution https://creativecommons.org/licenses/by/4.0/ application/pdf Multidisciplinary Digital Publishing Institute Multidisciplinary Digital Publishing Institute
spellingShingle Frisella, Megan
Khorrami, Pooya
Matterer, Jason
Kratkiewicz, Kendra
Torres-Carrasquillo, Pedro
Quantifying Bias in a Face Verification System
title Quantifying Bias in a Face Verification System
title_full Quantifying Bias in a Face Verification System
title_fullStr Quantifying Bias in a Face Verification System
title_full_unstemmed Quantifying Bias in a Face Verification System
title_short Quantifying Bias in a Face Verification System
title_sort quantifying bias in a face verification system
url https://hdl.handle.net/1721.1/142034
work_keys_str_mv AT frisellamegan quantifyingbiasinafaceverificationsystem
AT khorramipooya quantifyingbiasinafaceverificationsystem
AT mattererjason quantifyingbiasinafaceverificationsystem
AT kratkiewiczkendra quantifyingbiasinafaceverificationsystem
AT torrescarrasquillopedro quantifyingbiasinafaceverificationsystem