Deep neural network approximated dynamic programming for combinatorial optimization

23Citations
Citations of this article
30Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we propose a general framework for combining deep neural networks (DNNs) with dynamic programming to solve combinatorial optimization problems. For problems that can be broken into smaller subproblems and solved by dynamic programming, we train a set of neural networks to replace value or policy functions at each decision step. Two variants of the neural network approximated dynamic programming (NDP) methods are proposed; in the value-based NDP method, the networks learn to estimate the value of each choice at the corresponding step, while in the policy-based NDP method the DNNs only estimate the best decision at each step. The training procedure of the NDP starts from the smallest problem size and a new DNN for the next size is trained to cooperate with previous DNNs. After all the DNNs are trained, the networks are fine-tuned together to further improve overall performance. We test NDP on the linear sum assignment problem, the traveling salesman problem and the talent scheduling problem. Experimental results show that NDP can achieve considerable computation time reduction on hard problems with reasonable performance loss. In general, NDP can be applied to reducible combinatorial optimization problems for the purpose of computation time reduction.

Cite

CITATION STYLE

APA

Xu, S., Panwar, S. S., Kodialam, M., & Lakshman, T. V. (2020). Deep neural network approximated dynamic programming for combinatorial optimization. In AAAI 2020 - 34th AAAI Conference on Artificial Intelligence (pp. 1684–1691). AAAI press. https://doi.org/10.1609/aaai.v34i02.5531

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free