Gaussian processes in reinforcement learning

ISSN: 10495258
121Citations
Citations of this article
361Readers
Mendeley users who have this article in their library.

Abstract

We exploit some useful properties of Gaussian process (GP) regression models for reinforcement learning in continuous state spaces and discrete time. We demonstrate how the GP model allows evaluation of the value function in closed form. The resulting policy iteration algorithm is demonstrated on a simple problem with a two dimensional state space. Further, we speculate that the intrinsic ability of GP models to characterise distributions of functions would allow the method to capture entire distributions over future values instead of merely their expectation, which has traditionally been the focus of much of reinforcement learning.

Cite

CITATION STYLE

APA

Rasmussen, C. E., & Kuss, M. (2004). Gaussian processes in reinforcement learning. In Advances in Neural Information Processing Systems. Neural information processing systems foundation.

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free