DeepFP for Finding Nash Equilibrium in Continuous Action Spaces

8Citations
Citations of this article
11Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Finding Nash equilibrium in continuous action spaces is a challenging problem and has applications in domains such as protecting geographic areas from potential attackers. We present DeepFP, an approximate extension of fictitious play in continuous action spaces. DeepFP represents players’ approximate best responses via generative neural networks which are highly expressive implicit density approximators. It additionally uses a game-model network which approximates the players’ expected payoffs given their actions, and trains the networks end-to-end in a model-based learning regime. Further, DeepFP allows using domain-specific oracles if available and can hence exploit techniques such as mathematical programming to compute best responses for structured games. We demonstrate stable convergence to Nash equilibrium on several classic games and also apply DeepFP to a large forest security domain with a novel defender best response oracle. We show that DeepFP learns strategies robust to adversarial exploitation and scales well with growing number of players’ resources.

Cite

CITATION STYLE

APA

Kamra, N., Gupta, U., Wang, K., Fang, F., Liu, Y., & Tambe, M. (2019). DeepFP for Finding Nash Equilibrium in Continuous Action Spaces. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11836 LNCS, pp. 238–258). Springer. https://doi.org/10.1007/978-3-030-32430-8_15

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free