Login / Signup

Now you see me, now you don't: an exploration of religious exnomination in DALL-E.

Mark AlfanoEhsan AbedinRitsaart ReimannMarinus FerreiraMarc Cheong
Published in: Ethics and information technology (2024)
Artificial intelligence (AI) systems are increasingly being used not only to classify and analyze but also to generate images and text. As recent work on the content produced by text and image Generative AIs has shown (e.g., Cheong et al., 2024, Acerbi & Stubbersfield, 2023), there is a risk that harms of representation and bias, already documented in prior AI and natural language processing (NLP) algorithms may also be present in generative models. These harms relate to protected categories such as gender, race, age, and religion. There are several kinds of harms of representation to consider in this context, including stereotyping, lack of recognition, denigration, under-representation, and many others (Crawford in Soundings 41:45-55, 2009; in: Barocas et al., SIGCIS Conference, 2017). Whereas the bulk of researchers' attention thus far has been given to stereotyping and denigration, in this study we examine 'exnomination', as conceived by Roland Barthes (1972), of religious groups. Our case study is DALL-E, a tool that generates images from natural language prompts. Using DALL-E mini, we generate images from generic prompts such as "religious person." We then examine whether the generated images are recognizably members of a nominated group. Thus, we assess whether the generated images normalize some religions while neglecting others. We hypothesize that Christianity will be recognizably represented more frequently than other religious groups. Our results partially support this hypothesis but introduce further complexities, which we then explore.
Keyphrases
  • deep learning
  • artificial intelligence
  • convolutional neural network
  • machine learning
  • big data
  • optical coherence tomography
  • autism spectrum disorder
  • mental health
  • smoking cessation
  • working memory