Trial-Based Dynamic Programming for Multi-Agent Planning

2Citations
Citations of this article
20Readers
Mendeley users who have this article in their library.

Abstract

Trial-based approaches offer an efficient way to solve single-agent MDPs and POMDPs. These approaches allow agents to focus their computations on regions of the environment they encounter during the trials, leading to significant computational savings. We present a novel trial-based dynamic programming (TBDP) algorithm for DEC-POMDPs that extends these benefits to multi-agent settings. The algorithm uses trial-based methods for both belief generation and policy evaluation. Policy improvement is implemented efficiently using linear programming and a sub-policy reuse technique that helps bound the amount of memory. The results show that TBDP can produce significant value improvements and is much faster than the best existing planning algorithms.

Cite

CITATION STYLE

APA

Wu, F., Zilberstein, S., & Chen, X. (2010). Trial-Based Dynamic Programming for Multi-Agent Planning. In Proceedings of the 24th AAAI Conference on Artificial Intelligence, AAAI 2010 (pp. 908–914). AAAI Press. https://doi.org/10.1609/aaai.v24i1.7616

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free