Mitigating demographic bias of machine learning models on social media

被引:0
|
作者
Wang, Yanchen [1 ]
Singh, Lisa [1 ]
机构
[1] Georgetown Univ, Washington, DC 20057 USA
基金
美国国家科学基金会;
关键词
algorithmic fairness; social media; demographic inference;
D O I
10.1145/3617694.3623244
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Social media posts have been used to predict different user behaviors and attitudes, including mental health condition, political affiliation, and vaccine hesitancy. Unfortunately, while social media platforms make APIs available for collecting user data, they also make it challenging to collect well structured demographic features about individuals who post on their platforms. This makes it difficult for researchers to assess the fairness of models they develop using these data. Researchers have begun considering approaches for determining fairness of machine learning models built using social media data. In this paper, we consider both the case when the sensitive demographic feature is available to the researcher and when it is not. After framing our specific problem and discussing the challenges, we focus on the scenario when the training data does not explicitly contain a sensitive demographic feature, but instead contains a hidden sensitive feature that can be approximated using a sensitive feature proxy. In this case, we propose an approach for determining whether a sensitive feature proxy exists in the training data and apply a fixing method to reduce the correlation between the sensitive feature proxy and the sensitive feature. To demonstrate our approach, we present two case studies using micro-linked Twitter/X data and show biases resulting from sensitive feature proxies that are present in the training data and are highly correlated to hidden sensitive features. We then show that a standard fixing approach can effectively reduce bias even if the sensitive attribute needs to be inferred by the researcher using existing reliable inference models. This is an important step toward understanding approaches for improving fairness on social media.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Mitigating Bias in Clinical Machine Learning Models
    Julio C. Perez-Downes
    Andrew S. Tseng
    Keith A. McConn
    Sara M. Elattar
    Olayemi Sokumbi
    Ronnie A. Sebro
    Megan A. Allyse
    Bryan J. Dangott
    Rickey E. Carter
    Demilade Adedinsewo
    Current Treatment Options in Cardiovascular Medicine, 2024, 26 : 29 - 45
  • [2] Mitigating Bias in Clinical Machine Learning Models
    Perez-Downes, Julio C.
    Tseng, Andrew S.
    McConn, Keith A.
    Elattar, Sara M.
    Sokumbi, Olayemi
    Sebro, Ronnie A.
    Allyse, Megan A.
    Dangott, Bryan J.
    Carter, Rickey E.
    Adedinsewo, Demilade
    CURRENT TREATMENT OPTIONS IN CARDIOVASCULAR MEDICINE, 2024, 26 (03) : 29 - 45
  • [3] Detecting and Mitigating Bias in Social Media
    Morstatter, Fred
    PROCEEDINGS OF THE 2016 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING ASONAM 2016, 2016, : 1347 - 1348
  • [4] Evaluating and mitigating bias in machine learning models for cardiovascular disease prediction
    Li, Fuchen
    Wu, Patrick
    Ong, Henry H.
    Peterson, Josh F.
    Wei, Wei-qi
    Zhao, Juan
    JOURNAL OF BIOMEDICAL INFORMATICS, 2023, 138
  • [5] Mitigating Racial Bias in Machine Learning
    Kostick-Quenet, Kristin M.
    Cohen, I. Glenn
    Gerke, Sara
    Lo, Bernard
    Antaki, James
    Movahedi, Faezah
    Njah, Hasna
    Schoen, Lauren
    Estep, Jerry E.
    Blumenthal-Barby, J. S.
    JOURNAL OF LAW MEDICINE & ETHICS, 2022, 50 (01): : 92 - 100
  • [6] Mitigating bias in machine learning for medicine
    Vokinger, Kerstin N.
    Feuerriegel, Stefan
    Kesselheim, Aaron S.
    COMMUNICATIONS MEDICINE, 2021, 1 (01):
  • [7] Mitigating bias in machine learning for medicine
    Kerstin N. Vokinger
    Stefan Feuerriegel
    Aaron S. Kesselheim
    Communications Medicine, 1
  • [8] Limitations of mitigating judicial bias with machine learning
    Kristian Lum
    Nature Human Behaviour, 1 (7)
  • [9] Designing Against Bias: Identifying and Mitigating Bias in Machine Learning and AI
    Corliss, David J.
    INTELLIGENT SYSTEMS AND APPLICATIONS, VOL 3, INTELLISYS 2023, 2024, 824 : 411 - 418
  • [10] Deconstructing demographic bias in speech-based machine learning models for digital health
    Yang, Michael
    El-Attar, Abd-Allah
    Chaspari, Theodora
    FRONTIERS IN DIGITAL HEALTH, 2024, 6