Towards generating simulated walking motion using position based deep reinforcement learning

3Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Much of robotics research aims to develop control solutions that exploit the machine's dynamics in order to achieve an extraordinarily agile behaviour [1]. This, however, is limited by the use of traditional model-based control techniques such as model predictive control and quadratic programming. These solutions are often based on simplified mechanical models which result in mechanically constrained and inefficient behaviour, thereby limiting the agility of the robotic system in development [2]. Treating the control of robotic systems as a reinforcement learning (RL) problem enables the use of model-free algorithms that attempt to learn a policy which maximizes the expected future (discounted) reward without inferring the effects of an executed action on the environment.

Cite

CITATION STYLE

APA

Jones, W., Gangapurwala, S., Havoutis, I., & Yoshida, K. (2019). Towards generating simulated walking motion using position based deep reinforcement learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11650 LNAI, pp. 467–470). Springer Verlag. https://doi.org/10.1007/978-3-030-25332-5_42

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free