Partiality and Misconception: Investigating Cultural Representativeness in Text-To-Image Models

被引:0
|
作者
Zhang, Lili [1 ]
Liao, Xi [1 ]
Yang, Zaijia [1 ]
Gao, Baihang [1 ]
Wang, Chunjie [1 ]
Yang, Qiuling [2 ]
Li, Deshun [2 ]
机构
[1] Hainan Univ, Haikou, Hainan, Peoples R China
[2] Hainan Univ, China Innovat Platform Acad Hainan Prov, Haikou, Hainan, Peoples R China
基金
中国国家自然科学基金;
关键词
text-to-image generation; cultural representativeness; cultural cluster; bias; stereotype;
D O I
10.1145/3613904.3642877
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text-to-image (T2I) models enable users worldwide to create high-defnition and realistic images through text prompts, where the underrepresentation and potential misinformation of images have raised growing concerns. However, few existing works examine cultural representativeness, especially involving whether the generated content can fairly and accurately refect global cultures. Combining automated and human methods, we investigate this issue in multiple dimensions quantifcationally and conduct a set of evaluations on three prevailing T2I models (DALL-E v2, Stable Difusion v1.5 and v2.1). Introducing attributes of cultural cluster and subject, we provide a fresh interdisciplinary perspective to bias analysis. The benchmark dataset UCOGC is presented, which encompasses authentic images of unique cultural objects from global clusters. Our results reveal that the culture of a disadvantaged country is prone to be neglected, some specifed subjects often present a stereotype or a simple patchwork of elements, and over half of cultural objects are mispresented.
引用
收藏
页数:25
相关论文
共 50 条
  • [41] Expressive Text-to-Image Generation with Rich Text
    Ge, Songwei
    Park, Taesung
    Zhu, Jun-Yan
    Huang, Jia-Bin
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 7511 - 7522
  • [42] From text to mask: Localizing entities using the attention of text-to-image diffusion models
    Xiao, Changming
    Yang, Qi
    Zhou, Feng
    Zhang, Changshui
    [J]. NEUROCOMPUTING, 2024, 610
  • [43] PromptMix: Text-to-image diffusion models enhance the performance of lightweight networks
    Bakhtiarnia, Arian
    Zhang, Qi
    Iosifidis, Alexandros
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [44] Comparative Review of Text-to-Image Generation Techniques Based on Diffusion Models
    Gao, Xinyu
    Du, Fang
    Song, Lijuan
    [J]. Computer Engineering and Applications, 2024, 60 (24) : 44 - 64
  • [45] Point-Cloud Completion with Pretrained Text-to-image Diffusion Models
    Kasten, Yoni
    Rahamim, Ohad
    Chechik, Gal
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [46] MagicFusion: Boosting Text-to-Image Generation Performance by Fusing Diffusion Models
    Zhao, Jing
    Zheng, Heliang
    Wang, Chaoyue
    Lan, Long
    Yang, Wenjing
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 22535 - 22545
  • [47] Open-Vocabulary Panoptic Segmentation with Text-to-Image Diffusion Models
    Xu, Jiarui
    Liu, Sifei
    Vahdat, Arash
    Byeon, Wonmin
    Wang, Xiaolong
    De Meo, Shalini
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 2955 - 2966
  • [48] Designing with words: exploring the integration of text-to-image models in industrial design
    Barros, Mario
    Ai, Qi
    [J]. DIGITAL CREATIVITY, 2024,
  • [49] Video Colorization with Pre-trained Text-to-Image Diffusion Models
    Liu, Hanyuan
    Xie, Minshan
    Xing, Jinbo
    Li, Chengze
    Wong, Tien-Tsin
    [J]. arXiv, 2023,
  • [50] Compositional Text-to-Image Synthesis with Attention Map Control of Diffusion Models
    Wang, Ruichen
    Chen, Zekang
    Chen, Chen
    Ma, Jian
    Lu, Haonan
    Lin, Xiaodong
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 6, 2024, : 5544 - 5552