Abstract
Reinforcement Learning (RL) provides effective results with an agent learning from a standalone reward function. However, it presents unique challenges with large amounts of environment states and action spaces, as well as in the determination of rewards. Imitation Learning (IL) offers a promising solution for those challenges using a teacher. In IL, the learning process can take advantage of human-sourced assistance and/or control over the agent and environment. A human teacher and an agent learner are considered in this study. The teacher takes part in the agent’s training towards dealing with the environment, tackling a specific objective, and achieving a predefined goal. This paper proposes a novel approach combining IL with different types of RL methods, namely, state-action-reward-state-action (SARSA) and Asynchronous Advantage Actor-Critic Agents (A3C), to overcome the problems of both stand-alone systems. How to effectively leverage the teacher’s feedback-be it direct binary or indirect detailed-for the agent learner to learn sequential decisionmaking policies is addressed. The results of this study on various OpenAI-Gym environments show that this algorithmic method can be incorporated with different combinations, and significantly decreases both human endeavors and tedious exploration process.
Author supplied keywords
Cite
CITATION STYLE
Navidi, N., & Landry, R. (2021). New approach in human-ai interaction by reinforcement-imitation learning. Applied Sciences (Switzerland), 11(7). https://doi.org/10.3390/app11073068
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.