Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review

<h4>Background</h4> While artificial intelligence (AI) offers possibilities of advanced clinical prediction and decision-making in healthcare, models trained on relatively homogeneous datasets, and populations poorly-representative of underlying diversity, limits generalisability and ris...

Full description

Bibliographic Details
Main Authors: Leo Anthony Celi, Jacqueline Cellini, Marie-Laure Charpignon, Edward Christopher Dee, Franck Dernoncourt, Rene Eber, William Greig Mitchell, Lama Moukheiber, Julian Schirmer, Julia Situ, Joseph Paguio, Joel Park, Judy Gichoya Wawira, Seth Yao
Format: Article
Language:English
Published: Public Library of Science (PLoS) 2022-03-01
Series:PLOS Digital Health
Online Access:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931338/?tool=EBI
_version_ 1797703272517074944
author Leo Anthony Celi
Jacqueline Cellini
Marie-Laure Charpignon
Edward Christopher Dee
Franck Dernoncourt
Rene Eber
William Greig Mitchell
Lama Moukheiber
Julian Schirmer
Julia Situ
Joseph Paguio
Joel Park
Judy Gichoya Wawira
Seth Yao
author_facet Leo Anthony Celi
Jacqueline Cellini
Marie-Laure Charpignon
Edward Christopher Dee
Franck Dernoncourt
Rene Eber
William Greig Mitchell
Lama Moukheiber
Julian Schirmer
Julia Situ
Joseph Paguio
Joel Park
Judy Gichoya Wawira
Seth Yao
author_sort Leo Anthony Celi
collection DOAJ
description <h4>Background</h4> While artificial intelligence (AI) offers possibilities of advanced clinical prediction and decision-making in healthcare, models trained on relatively homogeneous datasets, and populations poorly-representative of underlying diversity, limits generalisability and risks biased AI-based decisions. Here, we describe the landscape of AI in clinical medicine to delineate population and data-source disparities. <h4>Methods</h4> We performed a scoping review of clinical papers published in PubMed in 2019 using AI techniques. We assessed differences in dataset country source, clinical specialty, and author nationality, sex, and expertise. A manually tagged subsample of PubMed articles was used to train a model, leveraging transfer-learning techniques (building upon an existing BioBERT model) to predict eligibility for inclusion (original, human, clinical AI literature). Of all eligible articles, database country source and clinical specialty were manually labelled. A BioBERT-based model predicted first/last author expertise. Author nationality was determined using corresponding affiliated institution information using Entrez Direct. And first/last author sex was evaluated using the Gendarize.io API. <h4>Results</h4> Our search yielded 30,576 articles, of which 7,314 (23.9%) were eligible for further analysis. Most databases came from the US (40.8%) and China (13.7%). Radiology was the most represented clinical specialty (40.4%), followed by pathology (9.1%). Authors were primarily from either China (24.0%) or the US (18.4%). First and last authors were predominately data experts (i.e., statisticians) (59.6% and 53.9% respectively) rather than clinicians. And the majority of first/last authors were male (74.1%). <h4>Interpretation</h4> U.S. and Chinese datasets and authors were disproportionately overrepresented in clinical AI, and almost all of the top 10 databases and author nationalities were from high income countries (HICs). AI techniques were most commonly employed for image-rich specialties, and authors were predominantly male, with non-clinical backgrounds. Development of technological infrastructure in data-poor regions, and diligence in external validation and model re-calibration prior to clinical implementation in the short-term, are crucial in ensuring clinical AI is meaningful for broader populations, and to avoid perpetuating global health inequity. Author summary Artificial Intelligence (AI) creates opportunities for accurate, objective and immediate decision support in healthcare with little expert input–especially valuable in resource-poor settings where there is shortage of specialist care. Given that AI poorly generalises to cohorts outside those whose data was used to train and validate the algorithms, populations in data-rich regions stand to benefit substantially more vs data-poor regions, entrenching existing healthcare disparities. Here, we show that more than half of the datasets used for clinical AI originate from either the US or China. In addition, the U.S. and China contribute over 40% of the authors of the publications. While the models may perform on-par/better than clinician decision-making in the well-represented regions, benefits elsewhere are not guaranteed. Further, we show discrepancies in gender and specialty representation–notably that almost three-quarters of the coveted first/senior authorship positions were held by men, and radiology accounted for 40% of all clinical AI manuscripts. We emphasize that building equitable sociodemographic representation in data repositories, in author nationality, gender and expertise, and in clinical specialties is crucial in ameliorating health inequities.
first_indexed 2024-03-12T05:02:10Z
format Article
id doaj.art-cced3c08553244ca8cb277a24cb43975
institution Directory Open Access Journal
issn 2767-3170
language English
last_indexed 2024-03-12T05:02:10Z
publishDate 2022-03-01
publisher Public Library of Science (PLoS)
record_format Article
series PLOS Digital Health
spelling doaj.art-cced3c08553244ca8cb277a24cb439752023-09-03T09:09:46ZengPublic Library of Science (PLoS)PLOS Digital Health2767-31702022-03-0113Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global reviewLeo Anthony CeliJacqueline CelliniMarie-Laure CharpignonEdward Christopher DeeFranck DernoncourtRene EberWilliam Greig MitchellLama MoukheiberJulian SchirmerJulia SituJoseph PaguioJoel ParkJudy Gichoya WawiraSeth Yao<h4>Background</h4> While artificial intelligence (AI) offers possibilities of advanced clinical prediction and decision-making in healthcare, models trained on relatively homogeneous datasets, and populations poorly-representative of underlying diversity, limits generalisability and risks biased AI-based decisions. Here, we describe the landscape of AI in clinical medicine to delineate population and data-source disparities. <h4>Methods</h4> We performed a scoping review of clinical papers published in PubMed in 2019 using AI techniques. We assessed differences in dataset country source, clinical specialty, and author nationality, sex, and expertise. A manually tagged subsample of PubMed articles was used to train a model, leveraging transfer-learning techniques (building upon an existing BioBERT model) to predict eligibility for inclusion (original, human, clinical AI literature). Of all eligible articles, database country source and clinical specialty were manually labelled. A BioBERT-based model predicted first/last author expertise. Author nationality was determined using corresponding affiliated institution information using Entrez Direct. And first/last author sex was evaluated using the Gendarize.io API. <h4>Results</h4> Our search yielded 30,576 articles, of which 7,314 (23.9%) were eligible for further analysis. Most databases came from the US (40.8%) and China (13.7%). Radiology was the most represented clinical specialty (40.4%), followed by pathology (9.1%). Authors were primarily from either China (24.0%) or the US (18.4%). First and last authors were predominately data experts (i.e., statisticians) (59.6% and 53.9% respectively) rather than clinicians. And the majority of first/last authors were male (74.1%). <h4>Interpretation</h4> U.S. and Chinese datasets and authors were disproportionately overrepresented in clinical AI, and almost all of the top 10 databases and author nationalities were from high income countries (HICs). AI techniques were most commonly employed for image-rich specialties, and authors were predominantly male, with non-clinical backgrounds. Development of technological infrastructure in data-poor regions, and diligence in external validation and model re-calibration prior to clinical implementation in the short-term, are crucial in ensuring clinical AI is meaningful for broader populations, and to avoid perpetuating global health inequity. Author summary Artificial Intelligence (AI) creates opportunities for accurate, objective and immediate decision support in healthcare with little expert input–especially valuable in resource-poor settings where there is shortage of specialist care. Given that AI poorly generalises to cohorts outside those whose data was used to train and validate the algorithms, populations in data-rich regions stand to benefit substantially more vs data-poor regions, entrenching existing healthcare disparities. Here, we show that more than half of the datasets used for clinical AI originate from either the US or China. In addition, the U.S. and China contribute over 40% of the authors of the publications. While the models may perform on-par/better than clinician decision-making in the well-represented regions, benefits elsewhere are not guaranteed. Further, we show discrepancies in gender and specialty representation–notably that almost three-quarters of the coveted first/senior authorship positions were held by men, and radiology accounted for 40% of all clinical AI manuscripts. We emphasize that building equitable sociodemographic representation in data repositories, in author nationality, gender and expertise, and in clinical specialties is crucial in ameliorating health inequities.https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931338/?tool=EBI
spellingShingle Leo Anthony Celi
Jacqueline Cellini
Marie-Laure Charpignon
Edward Christopher Dee
Franck Dernoncourt
Rene Eber
William Greig Mitchell
Lama Moukheiber
Julian Schirmer
Julia Situ
Joseph Paguio
Joel Park
Judy Gichoya Wawira
Seth Yao
Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
PLOS Digital Health
title Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
title_full Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
title_fullStr Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
title_full_unstemmed Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
title_short Sources of bias in artificial intelligence that perpetuate healthcare disparities—A global review
title_sort sources of bias in artificial intelligence that perpetuate healthcare disparities a global review
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9931338/?tool=EBI
work_keys_str_mv AT leoanthonyceli sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT jacquelinecellini sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT marielaurecharpignon sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT edwardchristopherdee sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT franckdernoncourt sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT reneeber sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT williamgreigmitchell sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT lamamoukheiber sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT julianschirmer sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT juliasitu sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT josephpaguio sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT joelpark sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT judygichoyawawira sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview
AT sethyao sourcesofbiasinartificialintelligencethatperpetuatehealthcaredisparitiesaglobalreview