Costs and Benefits of Fair Representation Learning

被引:21
|
作者
McNamara, Daniel [1 ,2 ]
Ong, Cheng Soon [1 ,2 ]
Williamson, Robert C. [1 ,2 ]
机构
[1] Australian Natl Univ, Canberra, ACT, Australia
[2] CSIRO Data6l, Canberra, ACT, Australia
关键词
fairness; representation learning; machine learning;
D O I
10.1145/3306618.3317964
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Machine learning algorithms are increasingly used to make or support important decisions about people's lives. This has led to interest in the problem of fair classification, which involves learning to make decisions that are non-discriminatory with respect to a sensitive variable such as race or gender. Several methods have been proposed to solve this problem, including fair representation learning, which cleans the input data used by the algorithm to remove information about the sensitive variable. We show that using fair representation learning as an intermediate step in fair classification incurs a cost compared to directly solving the problem, which we refer to as the cost of mistrust. We show that fair representation learning in fact addresses a different problem, which is of interest when the data user is not trusted to access the sensitive variable. We quantify the benefits of fair representation learning, by showing that any subsequent use of the cleaned data will not be too unfair. The benefits we identify result from restricting the decisions of adversarial data users, while the costs are due to applying those same restrictions to other data users.
引用
收藏
页码:263 / 270
页数:8
相关论文
共 50 条