Imitation learning refers to the problem where an agent learns a policy that mimics the demonstration provided by the expert, without any information on the cost function of the environment. Classical approaches to imitation learning usually rely on a restrictive class of cost functions that best explains the expert's demonstration, exemplified by linear functions of pre-defined features on states and actions. We show that the kernelization of a classical algorithm naturally reduces the imitation learning to a distribution learning problem, where the imitation policy tries to match the state-action visitation distribution of the expert. Closely related to our approach is the recent work on leveraging generative adversarial networks (GANs) for imitation learning, but our reduction to distribution learning is much simpler, robust to scarce expert demonstration, and sample efficient. We demonstrate the effectiveness of our approach on a wide range of high-dimensional control tasks.
CITATION STYLE
Kim, K. E., & Park, H. S. (2018). Imitation learning via kernel mean embedding. In 32nd AAAI Conference on Artificial Intelligence, AAAI 2018 (pp. 3415–3422). AAAI press. https://doi.org/10.1609/aaai.v32i1.11720
Mendeley helps you to discover research relevant for your work.