Applying and Verifying an Explainability Method Based on Policy Graphs in the Context of Reinforcement Learning

6Citations
Citations of this article
12Readers
Mendeley users who have this article in their library.

Abstract

The advancement on explainability techniques is quite relevant in the field of Reinforcement Learning (RL) and its applications can be beneficial for the development of intelligent agents that are understandable by humans and are able cooperate with them. When dealing with Deep RL some approaches already exist in the literature, but a common problem is that it can be tricky to define whether the explanations generated for an agent really reflect the behaviour of the trained agent. In this work we will apply an approach for explainability based on the creation of a Policy Graph (PG) that represents the agent's behaviour. Our main contribution is a way to measure the similarity between the explanations and the agent's behaviour, by building another agent that follows a policy based on the explainability method and comparing the behaviour of both agents.

Cite

CITATION STYLE

APA

Climent, A., Gnatyshak, D., & Alvarez-Napagao, S. (2021). Applying and Verifying an Explainability Method Based on Policy Graphs in the Context of Reinforcement Learning. In Frontiers in Artificial Intelligence and Applications (Vol. 339, pp. 455–464). IOS Press BV. https://doi.org/10.3233/FAIA210166

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free