GBMix: Enhancing Fairness by Group-Balanced Mixup

Mixup is a powerful data augmentation strategy that has been shown to improve the generalization and adversarial robustness of machine learning classifiers, particularly in computer vision applications. Despite its simplicity and effectiveness, the impact of Mixup on the fairness of a model has not...

Full description

Bibliographic Details
Main Authors: Sangwoo Hong, Youngseok Yoon, Hyungjun Joo, Jungwoo Lee
Format: Article
Language:English
Published: IEEE 2024-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/10413562/
_version_ 1827355275529551872
author Sangwoo Hong
Youngseok Yoon
Hyungjun Joo
Jungwoo Lee
author_facet Sangwoo Hong
Youngseok Yoon
Hyungjun Joo
Jungwoo Lee
author_sort Sangwoo Hong
collection DOAJ
description Mixup is a powerful data augmentation strategy that has been shown to improve the generalization and adversarial robustness of machine learning classifiers, particularly in computer vision applications. Despite its simplicity and effectiveness, the impact of Mixup on the fairness of a model has not been thoroughly investigated yet. In this paper, we demonstrate that Mixup can perpetuate or even exacerbate bias presented in the training set. We provide insight to understand the reasons behind this behavior and propose GBMix, a group-balanced Mixup strategy to train fair classifiers. It groups the dataset based on their attributes and balances the Mixup ratio between the groups. Through the reorganization and balance of Mixup among groups, GBMix effectively enhances both average and worst-case accuracy concurrently. We empirically show that GBMix effectively mitigates bias in the training set and reduces the performance gap between groups. This effect is observed across a range of datasets and networks, and GBMix outperforms all the state-of-the-art methods.
first_indexed 2024-03-08T04:08:20Z
format Article
id doaj.art-2e77c5614055405392aab5bbf121068f
institution Directory Open Access Journal
issn 2169-3536
language English
last_indexed 2024-03-08T04:08:20Z
publishDate 2024-01-01
publisher IEEE
record_format Article
series IEEE Access
spelling doaj.art-2e77c5614055405392aab5bbf121068f2024-02-09T00:03:24ZengIEEEIEEE Access2169-35362024-01-0112188771888710.1109/ACCESS.2024.335827510413562GBMix: Enhancing Fairness by Group-Balanced MixupSangwoo Hong0https://orcid.org/0000-0002-0270-2781Youngseok Yoon1Hyungjun Joo2Jungwoo Lee3https://orcid.org/0000-0002-6804-980XDepartment of Electrical and Computer Engineering, Communications and Machine Learning Laboratory, Seoul National University, Seoul, South KoreaDepartment of Electrical and Computer Engineering, Communications and Machine Learning Laboratory, Seoul National University, Seoul, South KoreaDepartment of Electrical and Computer Engineering, Communications and Machine Learning Laboratory, Seoul National University, Seoul, South KoreaDepartment of Electrical and Computer Engineering, Communications and Machine Learning Laboratory, Seoul National University, Seoul, South KoreaMixup is a powerful data augmentation strategy that has been shown to improve the generalization and adversarial robustness of machine learning classifiers, particularly in computer vision applications. Despite its simplicity and effectiveness, the impact of Mixup on the fairness of a model has not been thoroughly investigated yet. In this paper, we demonstrate that Mixup can perpetuate or even exacerbate bias presented in the training set. We provide insight to understand the reasons behind this behavior and propose GBMix, a group-balanced Mixup strategy to train fair classifiers. It groups the dataset based on their attributes and balances the Mixup ratio between the groups. Through the reorganization and balance of Mixup among groups, GBMix effectively enhances both average and worst-case accuracy concurrently. We empirically show that GBMix effectively mitigates bias in the training set and reduces the performance gap between groups. This effect is observed across a range of datasets and networks, and GBMix outperforms all the state-of-the-art methods.https://ieeexplore.ieee.org/document/10413562/Mixupfairnessdata augmentationbiasspurious correlation
spellingShingle Sangwoo Hong
Youngseok Yoon
Hyungjun Joo
Jungwoo Lee
GBMix: Enhancing Fairness by Group-Balanced Mixup
IEEE Access
Mixup
fairness
data augmentation
bias
spurious correlation
title GBMix: Enhancing Fairness by Group-Balanced Mixup
title_full GBMix: Enhancing Fairness by Group-Balanced Mixup
title_fullStr GBMix: Enhancing Fairness by Group-Balanced Mixup
title_full_unstemmed GBMix: Enhancing Fairness by Group-Balanced Mixup
title_short GBMix: Enhancing Fairness by Group-Balanced Mixup
title_sort gbmix enhancing fairness by group balanced mixup
topic Mixup
fairness
data augmentation
bias
spurious correlation
url https://ieeexplore.ieee.org/document/10413562/
work_keys_str_mv AT sangwoohong gbmixenhancingfairnessbygroupbalancedmixup
AT youngseokyoon gbmixenhancingfairnessbygroupbalancedmixup
AT hyungjunjoo gbmixenhancingfairnessbygroupbalancedmixup
AT jungwoolee gbmixenhancingfairnessbygroupbalancedmixup