Policy learning for continuous space security games using neural networks

27Citations
Citations of this article
40Readers
Mendeley users who have this article in their library.

Abstract

A wealth of algorithms centered around (integer) linear programming have been proposed to compute equilibrium strategies in security games with discrete states and actions. However, in practice many domains possess continuous state and action spaces. In this paper, we consider a continuous space security game model with infinite-size action sets for players and present a novel deep learning based approach to extend the existing toolkit for solving security games. Specifically, we present (i) OptGradFP, a novel and general algorithm that searches for the optimal defender strategy in a parameterized continuous search space, and can also be used to learn policies over multiple game states simultaneously; (ii) OptGradFP-NN, a convolutional neural network based implementation of OptGradFP for continuous space security games. We demonstrate the potential to predict good defender strategies via experiments and analysis of OptGradFP and OptGradFP-NN on discrete and continuous game settings.

Cite

CITATION STYLE

APA

Kamra, N., Gupta, U., Fang, F., Liu, Y., & Tambe, M. (2018). Policy learning for continuous space security games using neural networks. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 1103–1112). AAAI press. https://doi.org/10.1609/aaai.v32i1.11446

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free