Adding expert knowledge and exploration in Monte-Carlo tree search

37Citations
Citations of this article
47Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a new exploration term, more efficient than classical UCT-like exploration terms. It combines efficiently expert rules, patterns extracted from datasets, All-Moves-As-First values, and classical online values. As this improved bandit formula does not solve several important situations (semeais, nakade) in computer Go, we present three other important improvements which are central in the recent progress of our program MoGo. We show an expert-based improvement of Monte-Carlo simulations for nakade situations; we also emphasize some limitations of this modification. We show a technique which preserves diversity in the Monte-Carlo simulation, which greatly improves the results in 19x19. Whereas the UCB-based exploration term is not efficient in MoGo, we show a new exploration term which is highly efficient in MoGo. MoGo recently won a game with handicap 7 against a 9Dan Pro player, Zhou JunXun, winner of the LG Cup 2007, and a game with handicap 6 against a 1Dan pro player, Li-Chen Chien. © 2010 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Chaslot, G., Fiter, C., Hoock, J. B., Rimmel, A., & Teytaud, O. (2010). Adding expert knowledge and exploration in Monte-Carlo tree search. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6048 LNCS, pp. 1–13). https://doi.org/10.1007/978-3-642-12993-3_1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free