145
49

Learning Fair Classifiers

Abstract

Automated data-driven decision systems are ubiquitous across a wide variety of online services, from online social networking and e-commerce to e-government. These systems rely on complex learning methods and vast amounts of data to optimize the service functionality, satisfaction of the end user and profitability. However, there is a growing concern that these automated decisions can lead to user discrimination, even in the absence of intent, leading to a lack of fairness, i.e., their outcomes have a disproportionally large adverse impact on particular groups of people sharing one or more sensitive attributes (e.g., race, sex). In this paper, we introduce a flexible mechanism to design fair classifiers in a principled manner, by leveraging a novel intuitive measure of decision boundary (un)fairness. Then, we instantiate this mechanism on two well-known classifiers: logistic regression and support vector machines. Experiments on both synthetic and real-world data show that our mechanism allows for a fine-grained control of the level of fairness, often at a minimal cost in terms of accuracy, and it provides more flexibility than alternatives.

View on arXiv
Comments on this paper