Fairness Trade-Offs and Partial Debiasing

Francois Buet-Golfouse, Islam Utyagulov
Proceedings of The 14th Asian Conference on Machine Learning, PMLR 189:112-136, 2023.

Abstract

Previous literature has shown that bias mitigating algorithms were sometimes prone to overfitting and had poor out-of-sample generalisation. This paper is first and foremost concerned with establishing a mathematical framework to tackle the specific issue of generalisation. Throughout this work, we consider fairness trade-offs and objectives mixing statistical loss over the whole sample and fairness penalties on categories (which could stem from different values of protected attributes), encompassing partial de-biasing. We do so by adopting two different but complementary viewpoints: first, we consider a PAC-type setup and derive probabilistic upper bounds involving sample-only information; second, we leverage an asymptotic framework to derive a closed-form limiting distribution for the difference between the empirical trade-off and the true trade-off. While these results provide guarantees for learning fairness metrics across categories, they also point out to the key (but asymmetric) role played by class imbalance. To summarise, learning fairness without having access to enough category-level samples is hard, and a simple numerical experiment shows that it can lead to spurious results.

Cite this Paper


BibTeX
@InProceedings{pmlr-v189-buet-golfouse23b, title = {Fairness Trade-Offs and Partial Debiasing}, author = {Buet-Golfouse, Francois and Utyagulov, Islam}, booktitle = {Proceedings of The 14th Asian Conference on Machine Learning}, pages = {112--136}, year = {2023}, editor = {Khan, Emtiyaz and Gonen, Mehmet}, volume = {189}, series = {Proceedings of Machine Learning Research}, month = {12--14 Dec}, publisher = {PMLR}, pdf = {https://proceedings.mlr.press/v189/buet-golfouse23b/buet-golfouse23b.pdf}, url = {https://proceedings.mlr.press/v189/buet-golfouse23b.html}, abstract = {Previous literature has shown that bias mitigating algorithms were sometimes prone to overfitting and had poor out-of-sample generalisation. This paper is first and foremost concerned with establishing a mathematical framework to tackle the specific issue of generalisation. Throughout this work, we consider fairness trade-offs and objectives mixing statistical loss over the whole sample and fairness penalties on categories (which could stem from different values of protected attributes), encompassing partial de-biasing. We do so by adopting two different but complementary viewpoints: first, we consider a PAC-type setup and derive probabilistic upper bounds involving sample-only information; second, we leverage an asymptotic framework to derive a closed-form limiting distribution for the difference between the empirical trade-off and the true trade-off. While these results provide guarantees for learning fairness metrics across categories, they also point out to the key (but asymmetric) role played by class imbalance. To summarise, learning fairness without having access to enough category-level samples is hard, and a simple numerical experiment shows that it can lead to spurious results.} }
Endnote
%0 Conference Paper %T Fairness Trade-Offs and Partial Debiasing %A Francois Buet-Golfouse %A Islam Utyagulov %B Proceedings of The 14th Asian Conference on Machine Learning %C Proceedings of Machine Learning Research %D 2023 %E Emtiyaz Khan %E Mehmet Gonen %F pmlr-v189-buet-golfouse23b %I PMLR %P 112--136 %U https://proceedings.mlr.press/v189/buet-golfouse23b.html %V 189 %X Previous literature has shown that bias mitigating algorithms were sometimes prone to overfitting and had poor out-of-sample generalisation. This paper is first and foremost concerned with establishing a mathematical framework to tackle the specific issue of generalisation. Throughout this work, we consider fairness trade-offs and objectives mixing statistical loss over the whole sample and fairness penalties on categories (which could stem from different values of protected attributes), encompassing partial de-biasing. We do so by adopting two different but complementary viewpoints: first, we consider a PAC-type setup and derive probabilistic upper bounds involving sample-only information; second, we leverage an asymptotic framework to derive a closed-form limiting distribution for the difference between the empirical trade-off and the true trade-off. While these results provide guarantees for learning fairness metrics across categories, they also point out to the key (but asymmetric) role played by class imbalance. To summarise, learning fairness without having access to enough category-level samples is hard, and a simple numerical experiment shows that it can lead to spurious results.
APA
Buet-Golfouse, F. & Utyagulov, I.. (2023). Fairness Trade-Offs and Partial Debiasing. Proceedings of The 14th Asian Conference on Machine Learning, in Proceedings of Machine Learning Research 189:112-136 Available from https://proceedings.mlr.press/v189/buet-golfouse23b.html.

Related Material