SVD reduction in continuos environment reinforcement learning

6Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Reinforcement learning methods, surviving the control difficulties of the unknown environment, are gaining more and more popularity recently in the autonomous robotics community. One of the possible difficulties of the reinforcement learning applications in complex situations is the huge size of the state-value- or action-value-function representation [2]. The case of continuous environment (continuous valued) reinforcement learning could be even complicated, as the state-value- or action-value-functions are turning into continuous functions. In this paper we suggest a way for tackling these difficulties by the application of SVD (Singular Value Decomposition) methods [3], [4], [15], [26]. © Springer-Verlag 2001.

Cite

CITATION STYLE

APA

Kovács, S. (2001). SVD reduction in continuos environment reinforcement learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 2206 LNCS, pp. 719–738). Springer Verlag. https://doi.org/10.1007/3-540-45493-4_71

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free