Fairness without Harm: Decoupled Classifiers with Preference Guarantees

Berk Ustun, Yang Liu, David Parkes
Proceedings of the 36th International Conference on Machine Learning, PMLR 97:6373-6382, 2019.

Abstract

In domains such as medicine, it can be acceptable for machine learning models to include sensitive attributes such as gender and ethnicity. In this work, we argue that when there is this kind of treatment disparity, then it should be in the best interest of each group. Drawing on ethical principles such as beneficence ("do the best") and non-maleficence ("do no harm"), we show how to use sensitive attributes to train decoupled classifiers that satisfy preference guarantees. These guarantees ensure the majority of individuals in each group prefer their assigned classifier to (i) a pooled model that ignores group membership (rationality), and (ii) the model assigned to any other group (envy-freeness). We introduce a recursive procedure that adaptively selects group attributes for decoupling, and present formal conditions to ensure preference guarantees in terms of generalization error. We validate the effectiveness of the procedure on real-world datasets, showing that it improves accuracy without violating preference guarantees on test data.

Cite this Paper


BibTeX
@InProceedings{pmlr-v97-ustun19a, title = {Fairness without Harm: Decoupled Classifiers with Preference Guarantees}, author = {Ustun, Berk and Liu, Yang and Parkes, David}, booktitle = {Proceedings of the 36th International Conference on Machine Learning}, pages = {6373--6382}, year = {2019}, editor = {Chaudhuri, Kamalika and Salakhutdinov, Ruslan}, volume = {97}, series = {Proceedings of Machine Learning Research}, month = {09--15 Jun}, publisher = {PMLR}, pdf = {http://proceedings.mlr.press/v97/ustun19a/ustun19a.pdf}, url = {http://proceedings.mlr.press/v97/ustun19a.html}, abstract = {In domains such as medicine, it can be acceptable for machine learning models to include sensitive attributes such as gender and ethnicity. In this work, we argue that when there is this kind of treatment disparity, then it should be in the best interest of each group. Drawing on ethical principles such as beneficence ("do the best") and non-maleficence ("do no harm"), we show how to use sensitive attributes to train decoupled classifiers that satisfy preference guarantees. These guarantees ensure the majority of individuals in each group prefer their assigned classifier to (i) a pooled model that ignores group membership (rationality), and (ii) the model assigned to any other group (envy-freeness). We introduce a recursive procedure that adaptively selects group attributes for decoupling, and present formal conditions to ensure preference guarantees in terms of generalization error. We validate the effectiveness of the procedure on real-world datasets, showing that it improves accuracy without violating preference guarantees on test data.} }
Endnote
%0 Conference Paper %T Fairness without Harm: Decoupled Classifiers with Preference Guarantees %A Berk Ustun %A Yang Liu %A David Parkes %B Proceedings of the 36th International Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2019 %E Kamalika Chaudhuri %E Ruslan Salakhutdinov %F pmlr-v97-ustun19a %I PMLR %P 6373--6382 %U http://proceedings.mlr.press/v97/ustun19a.html %V 97 %X In domains such as medicine, it can be acceptable for machine learning models to include sensitive attributes such as gender and ethnicity. In this work, we argue that when there is this kind of treatment disparity, then it should be in the best interest of each group. Drawing on ethical principles such as beneficence ("do the best") and non-maleficence ("do no harm"), we show how to use sensitive attributes to train decoupled classifiers that satisfy preference guarantees. These guarantees ensure the majority of individuals in each group prefer their assigned classifier to (i) a pooled model that ignores group membership (rationality), and (ii) the model assigned to any other group (envy-freeness). We introduce a recursive procedure that adaptively selects group attributes for decoupling, and present formal conditions to ensure preference guarantees in terms of generalization error. We validate the effectiveness of the procedure on real-world datasets, showing that it improves accuracy without violating preference guarantees on test data.
APA
Ustun, B., Liu, Y. & Parkes, D.. (2019). Fairness without Harm: Decoupled Classifiers with Preference Guarantees. Proceedings of the 36th International Conference on Machine Learning, in Proceedings of Machine Learning Research 97:6373-6382 Available from http://proceedings.mlr.press/v97/ustun19a.html.

Related Material