Gaussian processes non-linear inverse reinforcement learning

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The authors analyse a Bayesian framework for posing and solving inverse reinforcement learning (IRL) problems that arise in decision-making and optimisation settings. The authors propose a non-parametric Bayesian model using Gaussian process (GP) and preference graphs, which offer an effective and computationally efficient method for ill-posed IRL problems in large or infinite state space. This approach only requires a finite number of demonstrations that is much less than the amount required for approximating the feature expectation or value functions in previous IRL methods. The proposed learning framework is expressive as it relies on a Bayesian structure in which assumptions are explicit and changeable. It is also robust in that it formalises on convex optimisation, which retains the promise of computationally manageable implementations for practical problems. To deal with more realistic IRL problems where the dynamics is also unknown, the GP model can be easily combined with the method to learn the dynamics at the same time. Experimental results prove the superiority of the authors method to current prevailing IRL algorithms with fewer numbers of demonstrations in both discrete and continuous state space.

Cite

CITATION STYLE

APA

Qiao, Q., & Lin, X. (2021). Gaussian processes non-linear inverse reinforcement learning. IET Cyber-Systems and Robotics, 3(2), 150–163. https://doi.org/10.1049/csy2.12017

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free