Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications

Bias in software systems is a serious threat to human rights: when software makes decisions that allocate resources or opportunities, may disparately impact people based on personal traits (e.g., gender, ethnic group, etc.), systematically (dis)advantaging certain social groups. The cause is very of...

Full description

Bibliographic Details
Main Authors: Mariachiara Mecati, Marco Torchiano, Antonio Vetro, Juan Carlos de Martin
Format: Article
Language:English
Published: IEEE 2023-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10058507/
_version_ 1797861349001265152
author Mariachiara Mecati
Marco Torchiano
Antonio Vetro
Juan Carlos de Martin
author_facet Mariachiara Mecati
Marco Torchiano
Antonio Vetro
Juan Carlos de Martin
author_sort Mariachiara Mecati
collection DOAJ
description Bias in software systems is a serious threat to human rights: when software makes decisions that allocate resources or opportunities, may disparately impact people based on personal traits (e.g., gender, ethnic group, etc.), systematically (dis)advantaging certain social groups. The cause is very often the imbalance of training data, that is, unequal distribution of data between the classes of an attribute. Previous studies showed that lower levels of balance in protected attributes are related to higher levels of unfairness in the output. In this paper we contribute to the current status of knowledge on balance measures as risk indicators of systematic discriminations by studying imbalance on two further aspects: the intersectionality among the classes of protected attributes, and the combination of the target variable with protected attributes. We conduct an empirical study to verify whether: i) it is possible to infer the balance of intersectional attributes from the balance of the primary attributes, ii) measures of balance on intersectional attributes are helpful to detect unfairness in the classification outcome, iii) the computation of balance on the combination of a target variable with protected attributes improves the detection of unfairness. Overall the results reveal positive answers, but not for every combination of balance measure and fairness criterion. For this reason, we recommend selecting the fairness and balance measures that are most suitable to the application context when applying our risk approach to real cases.
first_indexed 2024-04-09T22:00:47Z
format Article
id doaj.art-5618c19cab344636b91236b3bb03141e
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-04-09T22:00:47Z
publishDate 2023-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-5618c19cab344636b91236b3bb03141e2023-03-23T23:00:16ZengIEEEIEEE Access2169-35362023-01-0111269962701110.1109/ACCESS.2023.325237010058507Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair ClassificationsMariachiara Mecati0https://orcid.org/0000-0002-0041-1809Marco Torchiano1https://orcid.org/0000-0001-5328-368XAntonio Vetro2https://orcid.org/0000-0003-2027-3308Juan Carlos de Martin3Department of Control and Computer Engineering, Politecnico di Torino, Turin, ItalyDepartment of Control and Computer Engineering, Politecnico di Torino, Turin, ItalyDepartment of Control and Computer Engineering, Politecnico di Torino, Turin, ItalyDepartment of Control and Computer Engineering, Politecnico di Torino, Turin, ItalyBias in software systems is a serious threat to human rights: when software makes decisions that allocate resources or opportunities, may disparately impact people based on personal traits (e.g., gender, ethnic group, etc.), systematically (dis)advantaging certain social groups. The cause is very often the imbalance of training data, that is, unequal distribution of data between the classes of an attribute. Previous studies showed that lower levels of balance in protected attributes are related to higher levels of unfairness in the output. In this paper we contribute to the current status of knowledge on balance measures as risk indicators of systematic discriminations by studying imbalance on two further aspects: the intersectionality among the classes of protected attributes, and the combination of the target variable with protected attributes. We conduct an empirical study to verify whether: i) it is possible to infer the balance of intersectional attributes from the balance of the primary attributes, ii) measures of balance on intersectional attributes are helpful to detect unfairness in the classification outcome, iii) the computation of balance on the combination of a target variable with protected attributes improves the detection of unfairness. Overall the results reveal positive answers, but not for every combination of balance measure and fairness criterion. For this reason, we recommend selecting the fairness and balance measures that are most suitable to the application context when applying our risk approach to real cases.https://ieeexplore.ieee.org/document/10058507/Data biasdata imbalanceintersectionalityalgorithmic fairnessautomated decision-makingdata ethics
spellingShingle Mariachiara Mecati
Marco Torchiano
Antonio Vetro
Juan Carlos de Martin
Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
IEEE Access
Data bias
data imbalance
intersectionality
algorithmic fairness
automated decision-making
data ethics
title Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
title_full Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
title_fullStr Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
title_full_unstemmed Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
title_short Measuring Imbalance on Intersectional Protected Attributes and on Target Variable to Forecast Unfair Classifications
title_sort measuring imbalance on intersectional protected attributes and on target variable to forecast unfair classifications
topic Data bias
data imbalance
intersectionality
algorithmic fairness
automated decision-making
data ethics
url https://ieeexplore.ieee.org/document/10058507/
work_keys_str_mv AT mariachiaramecati measuringimbalanceonintersectionalprotectedattributesandontargetvariabletoforecastunfairclassifications
AT marcotorchiano measuringimbalanceonintersectionalprotectedattributesandontargetvariabletoforecastunfairclassifications
AT antoniovetro measuringimbalanceonintersectionalprotectedattributesandontargetvariabletoforecastunfairclassifications
AT juancarlosdemartin measuringimbalanceonintersectionalprotectedattributesandontargetvariabletoforecastunfairclassifications