Bandit based Monte-Carlo planning

1.9kCitations
Citations of this article
1.2kReaders
Mendeley users who have this article in their library.

This article is free to access.

Abstract

For large state-space Markovian Decision Problems Monte-Carlo planning is one of the few viable approaches to find near-optimal solutions. In this paper we introduce a new algorithm, UCT, that applies bandit ideas to guide Monte-Carlo planning. In finite-horizon or discounted MDPs the algorithm is shown to be consistent and finite sample bounds are derived on the estimation error due to sampling. Experimental results show that in several domains, UCT is significantly more efficient than its alternatives. © Springer-Verlag Berlin Heidelberg 2006.

Cite

CITATION STYLE

APA

Kocsis, L., & Szepesvári, C. (2006). Bandit based Monte-Carlo planning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 4212 LNAI, pp. 282–293). Springer Verlag. https://doi.org/10.1007/11871842_29

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free