This chapter provides an overview of the most popular methods of inverse reinforcement learning (IRL) and imitation learning (IL). These methods solve the problem of optimal control in a data-driven way, similarly to reinforcement learning, however with the critical difference that now rewards are not observed. The problem is rather to learn the reward function from the observed behavior of an agent. As behavioral data without rewards is widely available, the problem of learning from such data is certainly very interesting. This chapter provides a moderate-level technical description of the most promising IRL methods, equips the reader with sufficient knowledge to understand and follow the current literature on IRL, and presents examples that use simple simulated environments to evaluate how these methods perform when the "ground-truth" rewards are known. We then present use cases for IRL in quantitative finance which include applications in trading strategy identification, sentiment-based trading, option pricing, inference of portfolio investors, and market modeling.
CITATION STYLE
Dixon, M. F., Halperin, I., & Bilokon, P. (2020). Inverse Reinforcement Learning and Imitation Learning. In Machine Learning in Finance (pp. 419–517). Springer International Publishing. https://doi.org/10.1007/978-3-030-41068-1_11
Mendeley helps you to discover research relevant for your work.