Fairness in the Eyes of the Data: Certifying Machine-Learning Models

20Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a framework that allows to certify the fairness degree of a model based on an interactive and privacy-preserving test. The framework verifies any trained model, regardless of its training process and architecture. Thus, it allows us to evaluate any deep learning model on multiple fairness definitions empirically. We tackle two scenarios, where either the test data is privately available only to the tester or is publicly known in advance, even to the model creator. We investigate the soundness of the proposed approach using theoretical analysis and present statistical guarantees for the interactive test. Finally, we provide a cryptographic technique to automate fairness testing and certified inference with only black-box access to the model at hand while hiding the participants' sensitive data.

Cite

CITATION STYLE

APA

Segal, S., Adi, Y., Pinkas, B., Baum, C., Ganesh, C., & Keshet, J. (2021). Fairness in the Eyes of the Data: Certifying Machine-Learning Models. In AIES 2021 - Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society (pp. 926–935). Association for Computing Machinery, Inc. https://doi.org/10.1145/3461702.3462554

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free