Causal Feature Selection for Algorithmic Fairness

31Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The use of machine learning (ML) in high-stakes societal decisions has encouraged the consideration of fairness throughout the ML lifecycle. Although data integration is one of the primary steps to generate high-quality training data, most of the fairness literature ignores this stage. In this work, we consider fairness in the integration component of data management, aiming to identify features that improve prediction without adding any bias to the dataset. We work under the causal fairness paradigm. Without requiring the underlying structural causal model a priori, we propose an approach to identify a sub-collection of features that ensure fairness of the dataset by performing conditional independence tests between different subsets of features. We use group testing to improve the complexity of the approach. We theoretically prove the correctness of the proposed algorithm and show that sublinear conditional independence tests are sufficient to identify these variables. A detailed empirical evaluation is performed on real-world datasets to demonstrate the efficacy and efficiency of our technique.

Cite

CITATION STYLE

APA

Galhotra, S., Shanmugam, K., Sattigeri, P., & Varshney, K. R. (2022). Causal Feature Selection for Algorithmic Fairness. In Proceedings of the ACM SIGMOD International Conference on Management of Data (pp. 276–285). Association for Computing Machinery. https://doi.org/10.1145/3514221.3517909

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free