Multiple policy value Monte Carlo tree search

4Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.

Abstract

Many of the strongest game playing programs use a combination of Monte Carlo tree search (MCTS) and deep neural networks (DNN), where the DNNs are used as policy or value evaluators. Given a limited budget, such as online playing or during the self-play phase of AlphaZero (AZ) training, a balance needs to be reached between accurate state estimation and more MCTS simulations, both of which are critical for a strong game playing agent. Typically, larger DNNs are better at generalization and accurate evaluation, while smaller DNNs are less costly, and therefore can lead to more MCTS simulations and bigger search trees with the same budget. This paper introduces a new method called the multiple policy value MCTS (MPV-MCTS), which combines multiple policy value neural networks (PV-NNs) of various sizes to retain advantages of each network, where two PV-NNs fS and fL are used in this paper. We show through experiments on the game NoGo that a combined fS and fL MPV-MCTS outperforms single PV-NN with policy value MCTS, called PV-MCTS. Additionally, MPV-MCTS also outperforms PV-MCTS for AZ training.

Cite

CITATION STYLE

APA

Lan, L. C., Li, W., Wei, T. H., & Wu, I. C. (2019). Multiple policy value Monte Carlo tree search. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2019-August, pp. 4704–4710). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2019/653

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free