Understanding and evaluating harms of AI-generated image captions in political images

The use of AI-generated image captions has been increasing. Scholars of disability studies have long studied accessibility and AI issues concerning technology bias, focusing on image captions and tags. However, less attention has been paid to the individuals and social groups depicted in images and...

Full description

Bibliographic Details
Main Authors: Habiba Sarhan, Simon Hegelich
Format: Article
Language:English
Published: Frontiers Media S.A. 2023-09-01
Series:Frontiers in Political Science
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fpos.2023.1245684/full
_version_ 1827811856657416192
author Habiba Sarhan
Simon Hegelich
author_facet Habiba Sarhan
Simon Hegelich
author_sort Habiba Sarhan
collection DOAJ
description The use of AI-generated image captions has been increasing. Scholars of disability studies have long studied accessibility and AI issues concerning technology bias, focusing on image captions and tags. However, less attention has been paid to the individuals and social groups depicted in images and captioned using AI. Further research is needed to understand the underlying representational harms that could affect these social groups. This paper investigates the potential representational harms to social groups depicted in images. There is a high risk of harming certain social groups, either by stereotypical descriptions or erasing their identities from the caption, which could affect the understandings, beliefs, and attitudes that people hold about these specific groups. For the purpose of this article, 1,000 images with human-annotated captions were collected from news agencies “politics” sections. Microsoft's Azure Cloud Services was used to generate AI-generated captions with the December 2021 public version. The pattern observed from the politically salient images gathered and their captions highlight the tendency of the model used to generate more generic descriptions, which may potentially harm misrepresented social groups. Consequently, a balance between those harms needs to be struck, which is intertwined with the trade-off between generating generic vs. specific descriptions. The decision to generate generic descriptions, being extra cautious not to use stereotypes, erases and demeans excluded and already underrepresented social groups, while the decision to generate specific descriptions stereotypes social groups as well as reifies them. The appropriate trade-off is, therefore, crucial, especially when examining politically salient images.
first_indexed 2024-03-11T23:11:14Z
format Article
id doaj.art-18c65f49ab15477db0ef91f739672bd6
institution Directory Open Access Journal
issn 2673-3145
language English
last_indexed 2024-03-11T23:11:14Z
publishDate 2023-09-01
publisher Frontiers Media S.A.
record_format Article
series Frontiers in Political Science
spelling doaj.art-18c65f49ab15477db0ef91f739672bd62023-09-21T08:19:52ZengFrontiers Media S.A.Frontiers in Political Science2673-31452023-09-01510.3389/fpos.2023.12456841245684Understanding and evaluating harms of AI-generated image captions in political imagesHabiba SarhanSimon HegelichThe use of AI-generated image captions has been increasing. Scholars of disability studies have long studied accessibility and AI issues concerning technology bias, focusing on image captions and tags. However, less attention has been paid to the individuals and social groups depicted in images and captioned using AI. Further research is needed to understand the underlying representational harms that could affect these social groups. This paper investigates the potential representational harms to social groups depicted in images. There is a high risk of harming certain social groups, either by stereotypical descriptions or erasing their identities from the caption, which could affect the understandings, beliefs, and attitudes that people hold about these specific groups. For the purpose of this article, 1,000 images with human-annotated captions were collected from news agencies “politics” sections. Microsoft's Azure Cloud Services was used to generate AI-generated captions with the December 2021 public version. The pattern observed from the politically salient images gathered and their captions highlight the tendency of the model used to generate more generic descriptions, which may potentially harm misrepresented social groups. Consequently, a balance between those harms needs to be struck, which is intertwined with the trade-off between generating generic vs. specific descriptions. The decision to generate generic descriptions, being extra cautious not to use stereotypes, erases and demeans excluded and already underrepresented social groups, while the decision to generate specific descriptions stereotypes social groups as well as reifies them. The appropriate trade-off is, therefore, crucial, especially when examining politically salient images.https://www.frontiersin.org/articles/10.3389/fpos.2023.1245684/fullAI-generated image captionsrepresentational harmsinclusionresponsible AIAI harms frontiers
spellingShingle Habiba Sarhan
Simon Hegelich
Understanding and evaluating harms of AI-generated image captions in political images
Frontiers in Political Science
AI-generated image captions
representational harms
inclusion
responsible AI
AI harms frontiers
title Understanding and evaluating harms of AI-generated image captions in political images
title_full Understanding and evaluating harms of AI-generated image captions in political images
title_fullStr Understanding and evaluating harms of AI-generated image captions in political images
title_full_unstemmed Understanding and evaluating harms of AI-generated image captions in political images
title_short Understanding and evaluating harms of AI-generated image captions in political images
title_sort understanding and evaluating harms of ai generated image captions in political images
topic AI-generated image captions
representational harms
inclusion
responsible AI
AI harms frontiers
url https://www.frontiersin.org/articles/10.3389/fpos.2023.1245684/full
work_keys_str_mv AT habibasarhan understandingandevaluatingharmsofaigeneratedimagecaptionsinpoliticalimages
AT simonhegelich understandingandevaluatingharmsofaigeneratedimagecaptionsinpoliticalimages