Gender-tuning: Empowering Fine-tuning for Debiasing Pre-trained Language Models

被引:0
|
作者
Ghanbarzadeh, Somayeh [1 ]
Huang, Yan [1 ]
Palangi, Hamid [2 ]
Moreno, Radames Cruz [2 ]
Khanpour, Hamed [2 ]
机构
[1] Univ North Texas, Denton, TX 76203 USA
[2] Microsoft Res, Redmond, WA USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent studies have revealed that the widely-used Pre-trained Language Models (PLMs) propagate societal biases from the large unmoderated pre-training corpora. Existing solutions require debiasing training processes and datasets for debiasing, which are resource-intensive and costly. Furthermore, these methods hurt the PLMs' performance on downstream tasks. In this study, we propose Gender-tuning, which debiases the PLMs through fine-tuning on downstream tasks' datasets. For this aim, Gender-tuning integrates Masked Language Modeling (MLM) training objectives into fine-tuning's training process. Comprehensive experiments show that Gender-tuning outperforms the state-of-the-art baselines in terms of average gender bias scores in PLMs while improving PLMs' performance on downstream tasks solely using the downstream tasks' dataset. Also, Gender-tuning is a deployable debiasing tool for any PLM that works with original fine-tuning.
引用
收藏
页码:5448 / 5458
页数:11
相关论文
共 50 条
  • [41] Securely Fine-tuning Pre-trained Encoders Against Adversarial Examples
    Zhou, Ziqi
    Li, Minghui
    Liu, Wei
    Hu, Shengshan
    Zhang, Yechao
    Wang, Wei
    Xue, Lulu
    Zhang, Leo Yu
    Yao, Dezhong
    Jin, Hai
    45TH IEEE SYMPOSIUM ON SECURITY AND PRIVACY, SP 2024, 2024, : 3015 - 3033
  • [42] Variational Monte Carlo on a Budget - Fine-tuning pre-trained NeuralWavefunctions
    Scherbela, Michael
    Gerard, Leon
    Grohs, Philipp
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [43] Fine-Tuning Pre-Trained CodeBERT for Code Search in Smart Contract
    JIN Huan
    LI Qinying
    Wuhan University Journal of Natural Sciences, 2023, 28 (03) : 237 - 245
  • [44] Fine-tuning Pre-trained Language Models for Few-shot Intent Detection: Supervised Pre-training and Isotropization
    Zhang, Haode
    Liang, Haowen
    Zhang, Yuwei
    Zhan, Liming
    Wu, Xiao-Ming
    Lu, Xiaolei
    Lam, Albert Y. S.
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 532 - 542
  • [45] An Empirical Study of Parameter-Efficient Fine-Tuning Methods for Pre-trained Code Models
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, : 397 - 408
  • [46] APPT: Boosting Automated Patch Correctness Prediction via Fine-Tuning Pre-Trained Models
    Zhang, Quanjun
    Fang, Chunrong
    Sun, Weisong
    Liu, Yan
    He, Tieke
    Hao, Xiaodong
    Chen, Zhenyu
    IEEE TRANSACTIONS ON SOFTWARE ENGINEERING, 2024, 50 (03) : 474 - 494
  • [47] Few-Sample Named Entity Recognition for Security Vulnerability Reports by Fine-Tuning Pre-trained Language Models
    Yang, Guanqun
    Dineen, Shay
    Lin, Zhipeng
    Liu, Xueqing
    DEPLOYABLE MACHINE LEARNING FOR SECURITY DEFENSE, MLHAT 2021, 2021, 1482 : 55 - 78
  • [48] Enhancing Automated Essay Scoring Performance via Fine-tuning Pre-trained Language Models with Combination of Regression and Ranking
    Yang, Ruosong
    Cao, Jiannong
    Wen, Zhiyuan
    Wu, Youzheng
    He, Xiaodong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1560 - 1569
  • [49] Enhancing Health Mention Classification Through Reexamining Misclassified Samples and Robust Fine-Tuning Pre-Trained Language Models
    Meng, Deyu
    Phuntsho, Tshewang
    Gonsalves, Tad
    IEEE ACCESS, 2024, 12 : 190445 - 190453
  • [50] Fine-tuning Happens in Tiny Subspaces: Exploring Intrinsic Task-specific Subspaces of Pre-trained Language Models
    Zhang, Zhong
    Liu, Bang
    Shao, Junming
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 1701 - 1713