Discounted reinforcement learning does not scale

1Citations
Citations of this article
21Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Currently popular reinforcement learning methods are based on estimating value functions that indicate the long-term value of each problem state. In many domains, such as those traditionally studied in AI planning research, the size of state spaces precludes the individual storage of state value estimates. Consequently, most practical implementations of reinforcement learning methods have stored value functions using generalizing function approximators, with mixed results. We analyze the effects of approximation error on the performance of goal-based tasks, revealing potentially severe scaling difficulties. Empirical evidence is presented that suggests when difficulties are likely to occur and explains some of the widely differing results reported in the literature.

Cite

CITATION STYLE

APA

Mcdonald, M. A. F., & Hingston, P. (1997). Discounted reinforcement learning does not scale. Computational Intelligence, 13(1), 126–143. https://doi.org/10.1111/0824-7935.00035

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free