Strategic classification from revealed preferences

141Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We study an online linear classification problem in which the data is generated by strategic agents who manipulate their features in an effort to change the classification outcome. In rounds, the learner deploys a classifier, then an adversarially chosen agent arrives and possibly manipulates her features to optimally respond to the learner's choice of classifier. The learner has no knowledge of the agents' utility functions or “real” features, which may vary widely across agents. Instead, the learner is only able to observe their “revealed preferences”, i.e., the manipulated feature vectors they provide. For a broad family of agent cost functions, we give a computationally efficient learning algorithm that is able to obtain diminishing “Stackelberg regret” - a form of policy regret that guarantees that the learner is realizing loss nearly as small as that of the best classifier in hindsight, even allowing for the fact that agents would have best-responded differently to the optimal classifier.

Cite

CITATION STYLE

APA

Dong, J., Roth, A., Schutzman, Z., Waggoner, B., & Wu, Z. S. (2018). Strategic classification from revealed preferences. In ACM EC 2018 - Proceedings of the 2018 ACM Conference on Economics and Computation (pp. 55–70). Association for Computing Machinery, Inc. https://doi.org/10.1145/3219166.3219193

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free