A Hybrid PAC Reinforcement Learning Algorithm for Human-Robot Interaction

1Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper offers a new hybrid probably approximately correct (PAC) reinforcement learning (RL) algorithm for Markov decision processes (MDPs) that intelligently maintains favorable features of both model-based and model-free methodologies. The designed algorithm, referred to as the Dyna-Delayed Q-learning (DDQ) algorithm, combines model-free Delayed Q-learning and model-based R-max algorithms while outperforming both in most cases. The paper includes a PAC analysis of the DDQ algorithm and a derivation of its sample complexity. Numerical results are provided to support the claim regarding the new algorithm’s sample efficiency compared to its parents as well as the best known PAC model-free and model-based algorithms in application. A real-world experimental implementation of DDQ in the context of pediatric motor rehabilitation facilitated by infant-robot interaction highlights the potential benefits of the reported method.

Cite

CITATION STYLE

APA

Zehfroosh, A., & Tanner, H. G. (2022). A Hybrid PAC Reinforcement Learning Algorithm for Human-Robot Interaction. Frontiers in Robotics and AI, 9. https://doi.org/10.3389/frobt.2022.797213

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free