MarIA and BETO are sexist: evaluating gender bias in large language models for Spanish

被引:3
|
作者
Garrido-Munoz, Ismael [1 ]
Martinez-Santiago, Fernando [1 ]
Montejo-Raez, Arturo [1 ]
机构
[1] Univ Jaen, CEATIC, Campus Las Lagunillas, Jaen 23071, Spain
关键词
Deep learning; Gender bias; Bias evaluation; Language model; BERT; RoBERTa;
D O I
10.1007/s10579-023-09670-3
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
The study of bias in language models is a growing area of work, however, both research and resources are focused on English. In this paper, we make a first approach focusing on gender bias in some freely available Spanish language models trained using popular deep neural networks, like BERT or RoBERTa. Some of these models are known for achieving state-of-the-art results on downstream tasks. These promising results have promoted such models' integration in many real-world applications and production environments, which could be detrimental to people affected for those systems. This work proposes an evaluation framework to identify gender bias in masked language models, with explainability in mind to ease the interpretation of the evaluation results. We have evaluated 20 different models for Spanish, including some of the most popular pretrained ones in the research community. Our findings state that varying levels of gender bias are present across these models.This approach compares the adjectives proposed by the model for a set of templates. We classify the given adjectives into understandable categories and compute two new metrics from model predictions, one based on the internal state (probability) and the other one on the external state (rank). Those metrics are used to reveal biased models according to the given categories and quantify the degree of bias of the models under study.
引用
收藏
页码:1387 / 1417
页数:31
相关论文
共 50 条
  • [1] Evaluating and Mitigating Gender Bias in Generative Large Language Models
    Zhou, H.
    Inkpen, D.
    Kantarci, B.
    INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2024, 19 (06)
  • [2] Gender bias and stereotypes in Large Language Models
    Kotek, Hadas
    Dockum, Rikker
    Sun, David Q.
    PROCEEDINGS OF THE ACM COLLECTIVE INTELLIGENCE CONFERENCE, CI 2023, 2023, : 12 - 24
  • [3] MarIA: Spanish Language Models
    Gutierrez-Fandino, Asier
    Armengol-Estape, Jordi
    Pamies, Marc
    Llop-Palao, Joan
    Silveira-Ocampo, Joaquin
    Carrino, Casimiro Pio
    Armentano-Oller, Carme
    Rodriguez-Penagos, Carlos
    Gonzalez-Agirre, Aitor
    Villegas, Marta
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2022, (68): : 39 - 60
  • [4] Locating and Mitigating Gender Bias in Large Language Models
    Cai, Yuchen
    Cao, Ding
    Guo, Rongxi
    Wen, Yaqin
    Liu, Guiquan
    Chen, Enhong
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14878 : 471 - 482
  • [5] Evaluating the Performance of Large Language Models for Spanish Language in Undergraduate Admissions Exams
    Miranda, Sabino
    Pichardo-Lagunas, Obdulia
    Martinez-Seis, Bella
    Baldi, Pierre
    COMPUTACION Y SISTEMAS, 2023, 27 (04): : 1241 - 1248
  • [6] GenderCARE: A Comprehensive Framework for Assessing and Reducing Gender Bias in Large Language Models
    Tang, Kunsheng
    Zhou, Wenbo
    Zhang, Jie
    Liu, Aishan
    Deng, Gelei
    Li, Shuai
    Qi, Peigui
    Zhang, Weiming
    Zhang, Tianwei
    Yu, Nenghai
    PROCEEDINGS OF THE 2024 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2024, 2024, : 1196 - 1210
  • [7] Evaluating Bias and Fairness in Gender-Neutral Pretrained Vision-and-Language Models
    Cabello, Laura
    Bugliarello, Emanuele
    Brandl, Stephanie
    Elliott, Desmond
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 8465 - 8483
  • [8] CAN LEGAL LANGUAGE BE GENDER-NEUTRAL? SOME THOUGHTS ON (NON)-SEXIST LANGUAGE IN ENGLISH AND SPANISH
    Pano, Ana
    Turci, Monica
    REVISTA GENERAL DE DERECHO PUBLICO COMPARADO, 2008, (02):
  • [9] Evaluating Gender Bias of Pre-trained Language Models in Natural Language Inference by Considering All Labels
    Anantaprayoon, Panatchakorn
    Kaneko, Masahiro
    Okazaki, Naoaki
    arXiv, 2023,
  • [10] Evaluating the Performance of Large Language Models in Predicting Diagnostics for Spanish Clinical Cases in Cardiology
    Delaunay, Julien
    Cusido, Jordi
    APPLIED SCIENCES-BASEL, 2025, 15 (01):