Generalization analysis for game-theoretic machine learning

3Citations
Citations of this article
32Readers
Mendeley users who have this article in their library.

Abstract

For Internet applications like sponsored search, cautions need to be taken when using machine learning to optimize their mechanisms (e.g., auction) since self-interested agents in these applications may change their behaviors (and thus the data distribution) in response to the mechanisms. To tackle this problem, a framework called game-theoretic machine learning (GTML) was recently proposed, which first learns a Markov behavior model to characterize agents behaviors, and then learns the optimal mechanism by simulating agents' behavior changes in response to the mechanism. While GTML has demonstrated practical success, its generalization analysis is challenging because the behavior data are non-i.i.d. and dependent on the mechanism. To address this challenge, first, we decompose the generalization error for GTML into the behavior learning error and the mechanism learning error; second, for the behavior learning error, we obtain novel non-asymptotic error bounds for both parametric and non-parametric behavior learning methods; third, for the mechanism learning error, we derive a uniform convergence bound based on a new concept called nested covering number of the mechanism space and the generalization analysis techniques developed for mixing sequences.

Cite

CITATION STYLE

APA

Li, H., Tian, F., Chen, W., Qin, T., Ma, Z. M., & Liu, T. Y. (2015). Generalization analysis for game-theoretic machine learning. In Proceedings of the National Conference on Artificial Intelligence (Vol. 3, pp. 2089–2095). AI Access Foundation. https://doi.org/10.1609/aaai.v29i1.9436

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free