Pipelines for Social Bias Testing of Large Language Models

被引:0
|
作者
Nozza, Debora [1 ]
Bianchi, Federico [1 ]
Hovy, Dirk [1 ]
机构
[1] Bocconi Univ, Via Sarfatti 25, Milan, Italy
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The maturity level of language models is now at a stage in which many companies rely on them to solve various tasks. However, while research has shown how biased and harmful these models are, systematic ways of integrating social bias tests into development pipelines are still lacking. This short paper suggests how to use these verification techniques in development pipelines. We take inspiration from software testing and suggest addressing social bias evaluation as software testing. We hope to open a discussion on the best methodologies to handle social bias testing in language models.
引用
收藏
页码:68 / 74
页数:7
相关论文
共 50 条
  • [31] A critical review of large language models: Sensitivity, bias, and the path toward specialized AI
    Hajikhani, Arash
    Cole, Carolyn
    [J]. QUANTITATIVE SCIENCE STUDIES, 2024, 5 (03): : 736 - 756
  • [32] The Unequal Opportunities of Large Language Models: Revealing Demographic Bias through Job Recommendations
    Salinas, Abel
    Shah, Parth Vipul
    Huang, Yuzhong
    McCormack, Robert
    Morstatter, Fred
    [J]. PROCEEDINGS OF 2023 ACM CONFERENCE ON EQUITY AND ACCESS IN ALGORITHMS, MECHANISMS, AND OPTIMIZATION, EAAMO 2023, 2023,
  • [33] Large Language Models
    Vargas, Diego Collarana
    Katsamanis, Nassos
    [J]. ERCIM NEWS, 2024, (136): : 12 - 13
  • [34] Large Language Models
    Cerf, Vinton G.
    [J]. COMMUNICATIONS OF THE ACM, 2023, 66 (08) : 7 - 7
  • [35] A Natural Bias for Language Generation Models
    Meister, Clara
    Stokowiec, Wojciech
    Pimentel, Tiago
    Yu, Lei
    Rimell, Laura
    Kuncoro, Adhiguna
    [J]. 61ST CONFERENCE OF THE THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 2, 2023, : 243 - 255
  • [36] Building Reactive Large Language Model Pipelines with MOTION
    Shankar, Shreya
    Parameswaran, Aditya G.
    [J]. COMPANION OF THE 2024 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA, SIGMOD-COMPANION 2024, 2024, : 520 - 523
  • [37] Foundation and large language models: fundamentals, challenges, opportunities, and social impacts
    Myers, Devon
    Mohawesh, Rami
    Chellaboina, Venkata Ishwarya
    Sathvik, Anantha Lakshmi
    Venkatesh, Praveen
    Ho, Yi-Hui
    Henshaw, Hanna
    Alhawawreh, Muna
    Berdik, David
    Jararweh, Yaser
    [J]. CLUSTER COMPUTING-THE JOURNAL OF NETWORKS SOFTWARE TOOLS AND APPLICATIONS, 2024, 27 (01): : 1 - 26
  • [38] Foundation and large language models: fundamentals, challenges, opportunities, and social impacts
    Devon Myers
    Rami Mohawesh
    Venkata Ishwarya Chellaboina
    Anantha Lakshmi Sathvik
    Praveen Venkatesh
    Yi-Hui Ho
    Hanna Henshaw
    Muna Alhawawreh
    David Berdik
    Yaser Jararweh
    [J]. Cluster Computing, 2024, 27 : 1 - 26
  • [39] Testing models of belief bias: An experiment
    Coutts, Alexander
    [J]. GAMES AND ECONOMIC BEHAVIOR, 2019, 113 : 549 - 565
  • [40] Pilot study on large language models for risk-of-bias assessments in systematic reviews: A(I) new type of bias?
    Barsby, Joseph
    Hume, Samuel
    Lemmey, Hamish A. L.
    Cutteridge, Joseph
    Lee, Regent
    Bera, Katarzyna D.
    [J]. BMJ EVIDENCE-BASED MEDICINE, 2024,