Learning Optimal Fair Decision Trees: Trade-offs Between Interpretability, Fairness, and Accuracy

19Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The increasing use of machine learning in high-stakes domains - where people's livelihoods are impacted - creates an urgent need for interpretable, fair, and highly accurate algorithms. With these needs in mind, we propose a mixed integer optimization (MIO) framework for learning optimal classification trees - one of the most interpretable models - that can be augmented with arbitrary fairness constraints. In order to better quantify the "price of interpretability", we also propose a new measure of model interpretability called decision complexity that allows for comparisons across different classes of machine learning models. We benchmark our method against state-of-the-art approaches for fair classification on popular datasets; in doing so, we conduct one of the first comprehensive analyses of the trade-offs between interpretability, fairness, and predictive accuracy. Given a fixed disparity threshold, our method has a price of interpretability of about 4.2 percentage points in terms of out-of-sample accuracy compared to the best performing, complex models. However, our method consistently finds decisions with almost full parity, while other methods rarely do.

Cite

CITATION STYLE

APA

Jo, N., Aghaei, S., Benson, J., Gomez, A., & Vayanos, P. (2023). Learning Optimal Fair Decision Trees: Trade-offs Between Interpretability, Fairness, and Accuracy. In AIES 2023 - Proceedings of the 2023 AAAI/ACM Conference on AI, Ethics, and Society (pp. 181–192). Association for Computing Machinery, Inc. https://doi.org/10.1145/3600211.3604664

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free