Neutralizing Gender Bias in Word Embeddings with Latent Disentanglement and Counterfactual Generation

被引:0
|
作者
Shin, Seungjae [1 ]
Song, Kyungwoo [1 ]
Jang, JoonHo [1 ]
Kim, Hyemi [1 ]
Joo, Weonyoung [1 ]
Moon, Il-Chul [1 ]
机构
[1] Korea Adv Inst Sci & Technol KAIST, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent research demonstrates that word embeddings, trained on the human-generated corpus, have strong gender biases in embedding spaces, and these biases can result in the discriminative results from the various downstream tasks. Whereas the previous methods project word embeddings into a linear subspace for debiasing, we introduce a Latent Disentanglement method with a siamese auto-encoder structure with an adapted gradient reversal layer. Our structure enables the separation of the semantic latent information and gender latent information of given word into the disjoint latent dimensions. Afterwards, we introduce a Counterfactual Generation to convert the gender information of words, so the original and the modified embeddings can produce a gender-neutralized word embedding after geometric alignment regularization, without loss of semantic information. From the various quantitative and qualitative debiasing experiments, our method shows to be better than existing debiasing methods in debiasing word embeddings. In addition, Our method shows the ability to preserve semantic information during debiasing by minimizing the semantic information losses for extrinsic NLP downstream tasks.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Understanding Gender Bias in Knowledge Base Embeddings
    Du, Yupei
    Zheng, Qi
    Wu, Yuanbin
    Lan, Man
    Yang, Yan
    Ma, Meirong
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 1381 - 1395
  • [32] A World Full of Stereotypes? Further Investigation on Origin and Gender Bias in Multi-Lingual Word Embeddings
    Kurpicz-Briki, Mascha
    Leoni, Tomaso
    [J]. FRONTIERS IN BIG DATA, 2021, 4
  • [33] Word embeddings are biased. But whose bias are they reflecting?
    Petreski, Davor
    Hashim, Ibrahim C.
    [J]. AI & SOCIETY, 2023, 38 (02) : 975 - 982
  • [34] Word embeddings are biased. But whose bias are they reflecting?
    Davor Petreski
    Ibrahim C. Hashim
    [J]. AI & SOCIETY, 2023, 38 : 975 - 982
  • [35] Automated Template Generation based on Word Embeddings
    Manatuica, Maria
    Dascalu, Mihai
    Ruseti, Stefan
    Trausan-Matu, Stefan
    [J]. PROCEEDINGS OF THE 14TH INTERNATIONAL SCIENTIFIC CONFERENCE ELEARNING AND SOFTWARE FOR EDUCATION: ELEARNING CHALLENGES AND NEW HORIZONS, VOL 2, 2018, : 392 - 398
  • [36] Incorporating Latent Meanings of Morphological Compositions to Enhance Word Embeddings
    Xu, Yang
    Liu, Jiawei
    Yang, Wei
    Huang, Liusheng
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL), VOL 1, 2018, : 1232 - 1242
  • [37] A Transparent Framework for Evaluating Unintended Demographic Bias in Word Embeddings
    Sweeney, Chris
    Najafian, Maryam
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 1662 - 1667
  • [38] Debiasing Embeddings for Reduced Gender Bias in Text Classification
    Prost, Flavien
    Thain, Nithum
    Bolukbasi, Tolga
    [J]. GENDER BIAS IN NATURAL LANGUAGE PROCESSING (GEBNLP 2019), 2019, : 69 - 75
  • [39] Towards Detection of Subjective Bias using Contextualized Word Embeddings
    Pant, Kartikey
    Dadu, Tanvi
    Mamidi, Radhika
    [J]. WWW'20: COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2020, 2020, : 75 - 76
  • [40] Latent Semantic Analysis Approach for Document Summarization Based on Word Embeddings
    Al-Sabahi, Kamal
    Zhang Zuping
    Kang, Yang
    [J]. KSII TRANSACTIONS ON INTERNET AND INFORMATION SYSTEMS, 2019, 13 (01): : 254 - 276