Towards Fair and Robust Classification

被引:2
|
作者
Sun, Haipei [1 ]
Wu, Kun [2 ]
Wang, Ting [3 ]
Wang, Wendy Hui [2 ]
机构
[1] Facebook Inc, Seattle, WA 98109 USA
[2] Stevens Inst Technol, Hoboken, NJ 07030 USA
[3] Penn State Univ, University Pk, PA 16802 USA
基金
美国国家科学基金会;
关键词
Algorithmic fairness; adversarial robustness; classification; trustworthy machine learning; ATTACKS;
D O I
10.1109/EuroSP53844.2022.00030
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Robustness and fairness are two equally important issues for machine learning systems. Despite the active research on robustness and fairness of ML recently, these efforts focus on either fairness or robustness, but not both. To bridge this gap, in this paper, we design Fair and Robust Classification (FRoC) models that equip the classification models with both fairness and robustness. Meeting both fairness and robustness constraints is not trivial due to the tension between them. The trade-off between fairness, robustness, and model accuracy also introduces additional challenge. To address these challenges, we design two FRoC methods, namely FROC-PRE that modifies the input data before model training, and FROC-IN that modifies the model with an adversarial objective function to address both fairness and robustness during training. FROC-IN is suitable to the settings where the users (e.g., ML service providers) only have the access to the model but not the original data, while FROC-PRE works for the settings where the users (e.g., data owners) have the access to both data and a surrogate model that may have similar architecture as the target model. Our extensive experiments on real-world datasets demonstrate that both FROC-IN and FROC-PRE can achieve both fairness and robustness with insignificant accuracy loss of the target model.
引用
收藏
页码:356 / 376
页数:21
相关论文
共 50 条
  • [1] Towards Threshold Invariant Fair Classification
    Chen, Mingliang
    Wu, Min
    [J]. CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 560 - 569
  • [2] To be Robust or to be Fair: Towards Fairness in Adversarial Training
    Xu, Han
    Liu, Xiaorui
    Li, Yaxin
    Jain, Anil K.
    Tang, Jiliang
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [3] Fair and Robust Classification Under Sample Selection Bias
    Du, Wei
    Wu, Xintao
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2999 - 3003
  • [4] SURE: Robust, Explainable, and Fair Classification without Sensitive Attributes
    Chakrabarti, Deepayan
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 179 - 189
  • [5] Adversarial Debiasing techniques towards 'fair' skin lesion classification
    Correa-Medero, Ramon L.
    Patel, Bhavik
    Banerjee, Imon
    [J]. 2023 11TH INTERNATIONAL IEEE/EMBS CONFERENCE ON NEURAL ENGINEERING, NER, 2023,
  • [6] Towards Robust Domain Generation Algorithm Classification
    Drichel, Arthur
    Meyer, Marc
    Meyer, Ulrike
    [J]. PROCEEDINGS OF THE 19TH ACM ASIA CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, ACM ASIACCS 2024, 2024, : 2 - 18
  • [7] Towards tamper detection and classification with robust watermarks
    Knowles, H
    Winne, D
    Canagarajah, N
    Bull, D
    [J]. PROCEEDINGS OF THE 2003 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL II: COMMUNICATIONS-MULTIMEDIA SYSTEMS & APPLICATIONS, 2003, : 959 - 962
  • [8] Towards Robust Phoneme Classification With Hybrid Features
    Yousafzai, Jibran
    Cvetkovic, Zoran
    Sollich, Peter
    [J]. 2010 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, 2010, : 1643 - 1647
  • [9] Towards robust classification detection for adversarial examples
    Liu, Huangxiaolie
    Zhang, Dong
    Chen, Huijun
    [J]. INTERNATIONAL CONFERENCE FOR INTERNET TECHNOLOGY AND SECURED TRANSACTIONS (ICITST-2020), 2020, : 23 - 29
  • [10] Ensembles of Bireducts: Towards Robust Classification and Simple Representation
    Slezak, Dominik
    Janusz, Andrzej
    [J]. FUTURE GENERATION INFORMATION TECHNOLOGY, 2011, 7105 : 64 - 77