Interpretability and Fairness in Machine Learning: A Formal Methods Approach

被引:0
|
作者
Ghosh, Bishwamittra [1 ]
机构
[1] Natl Univ Singapore, Sch Comp, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The last decades have witnessed significant progress in machine learning with applications in different safety-critical domains, such as medical, law, education, and transportation. In high-stake domains, machine learning predictions have farreaching consequences on the end-users. With the aim of applying machine learning for societal goods, there have been increasing efforts to regulate machine learning by imposing interpretability, fairness, robustness, privacy, etc. in predictions. Towards responsible and trustworthy machine learning, we propose two research themes in our dissertation research: interpretability and fairness of machine learning classifiers. In particular, we design algorithms to learn interpretable rule-based classifiers, formally verify fairness, and explain the sources of unfairness. Prior approaches to these problems are often limited by scalability, accuracy, or both. To overcome these limitations, we closely integrate automated reasoning and formal methods with fairness and interpretability to develop scalable and accurate solutions.
引用
收藏
页码:7083 / 7084
页数:2
相关论文
共 50 条
  • [1] Accuracy, Fairness, and Interpretability of Machine Learning Criminal Recidivism Models
    Ingram, Eric
    Gursoy, Furkan
    Kakadiaris, Ioannis A.
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING, APPLICATIONS AND TECHNOLOGIES, BDCAT, 2022, : 233 - 241
  • [2] Evaluating Attribution Methods in Machine Learning Interpretability
    Ratul, Qudrat E. Alahy
    Serra, Edoardo
    Cuzzocrea, Alfredo
    2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 5239 - 5245
  • [3] Machine Learning Interpretability: A Survey on Methods and Metrics
    Carvalho, Diogo, V
    Pereira, Eduardo M.
    Cardoso, Jaime S.
    ELECTRONICS, 2019, 8 (08)
  • [4] Wasserstein-based fairness interpretability framework for machine learning models
    Alexey Miroshnikov
    Konstandinos Kotsiopoulos
    Ryan Franks
    Arjun Ravi Kannan
    Machine Learning, 2022, 111 : 3307 - 3357
  • [5] Wasserstein-based fairness interpretability framework for machine learning models
    Miroshnikov, Alexey
    Kotsiopoulos, Konstandinos
    Franks, Ryan
    Kannan, Arjun Ravi
    MACHINE LEARNING, 2022, 111 (09) : 3307 - 3357
  • [6] Explainable AI: A Review of Machine Learning Interpretability Methods
    Linardatos, Pantelis
    Papastefanopoulos, Vasilis
    Kotsiantis, Sotiris
    ENTROPY, 2021, 23 (01) : 1 - 45
  • [7] Formal Reasoning Methods for Explainability in Machine Learning
    Marquez-Silva, Joao
    ELECTRONIC PROCEEDINGS IN THEORETICAL COMPUTER SCIENCE, 2020, (325):
  • [8] Approach to provide interpretability in machine learning models for image classification
    Anja Stadlhofer
    Vitaliy Mezhuyev
    Industrial Artificial Intelligence, 1 (1):
  • [9] Evaluation of compression index of red mud by machine learning interpretability methods
    Yang, Fan
    Zhang, Jieya
    Xie, Mingxing
    Cui, Wenwen
    Dong, Xiaoqiang
    COMPUTERS AND GEOTECHNICS, 2025, 181
  • [10] Interpretability methods of machine learning algorithms with applications in breast cancer diagnosis
    Karatza, P.
    Dalakleidi, K.
    Athanasiou, M.
    Nikita, K. S.
    2021 43RD ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE & BIOLOGY SOCIETY (EMBC), 2021, : 2310 - 2313