"They only care to show us the wheelchair": Disability Representation in Text-to-Image AI Models

被引:0
|
作者
Mack, Kelly Avery [1 ,2 ]
Qadri, Rida [3 ]
Denton, Remi [4 ]
Kane, Shaun K. [5 ]
Bennett, Cynthia L. [4 ]
机构
[1] Google Res, Seattle, WA 98101 USA
[2] Univ Washington, Paul G Allen Sch Comp Sci, Seattle, WA 98195 USA
[3] Google Res, San Francisco, CA USA
[4] Google Res, New York, NY USA
[5] Google Res, Boulder, CO USA
关键词
disability representation; generative AI; algorithmic harms; human-centered AI; AI harms; text-to-image models;
D O I
10.1145/3613904.3642166
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper reports on disability representation in images output from text-to-image (T2I) generative AI systems. Through eight focus groups with 25 people with disabilities, we found that models repeatedly presented reductive archetypes for different disabilities. Often these representations reflected broader societal stereotypes and biases, which our participants were concerned to see reproduced through T2I. Our participants discussed further challenges with using these models including the current reliance on prompt engineering to reach satisfactorily diverse results. Finally, they offered suggestions for how to improve disability representation with solutions like showing multiple, heterogeneous images for a single prompt and including the prompt with images generated. Our discussion reflects on tensions and tradeoffs we found among the diverse perspectives shared to inform future research on representation-oriented generative AI system evaluation metrics and development processes.
引用
收藏
页数:23
相关论文
共 50 条
  • [31] BAGM: A Backdoor Attack for Manipulating Text-to-Image Generative Models
    Vice, Jordan
    Akhtar, Naveed
    Hartley, Richard
    Mian, Ajmal
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 4865 - 4880
  • [32] Text-to-Image Diffusion Models are Zero-Shot Classifiers
    Clark, Kevin
    Jaini, Priyank
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [33] Bridge-GAN: Interpretable Representation Learning for Text-to-Image Synthesis
    Yuan, Mingkuan
    Peng, Yuxin
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (11) : 4258 - 4268
  • [34] EvilPromptFuzzer: generating inappropriate content based on text-to-image models
    He, Juntao
    Dai, Haoran
    Sui, Runqi
    Yuan, Xuejing
    Liu, Dun
    Feng, Hao
    Liu, Xinyue
    Yang, Wenchuan
    Cui, Baojiang
    Li, Kedan
    [J]. CYBERSECURITY, 2024, 7 (01):
  • [35] A Study on Generating Webtoons Using Multilingual Text-to-Image Models
    Yu, Kyungho
    Kim, Hyoungju
    Kim, Jeongin
    Chun, Chanjun
    Kim, Pankoo
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (12):
  • [36] The Chosen One: Consistent Characters in Text-to-Image Diffusion Models
    Avrahami, Omri
    Hertz, Amir
    Vinker, Yael
    Arar, Moab
    Fruchter, Shlomi
    Fried, Ohad
    Cohen-Or, Daniel
    Lischinski, Dani
    [J]. PROCEEDINGS OF SIGGRAPH 2024 CONFERENCE PAPERS, 2024,
  • [37] Towards Consistent Video Editing with Text-to-Image Diffusion Models
    Zhang, Zicheng
    Li, Bonan
    Nie, Xuecheng
    Han, Congying
    Guo, Tiande
    Liu, Luoqi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [38] Exposing fake images generated by text-to-image diffusion models
    Xu, Qiang
    Wang, Hao
    Meng, Laijin
    Mi, Zhongjie
    Yuan, Jianye
    Yan, Hong
    [J]. PATTERN RECOGNITION LETTERS, 2023, 176 : 76 - 82
  • [39] Partiality and Misconception: Investigating Cultural Representativeness in Text-To-Image Models
    Zhang, Lili
    Liao, Xi
    Yang, Zaijia
    Gao, Baihang
    Wang, Chunjie
    Yang, Qiuling
    Li, Deshun
    [J]. PROCEEDINGS OF THE 2024 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYTEMS (CHI 2024), 2024,
  • [40] Best Prompts for Text-to-Image Models and How to Find Them
    Pavlichenko, Nikita
    Ustalov, Dmitry
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 2067 - 2071