Explaining Deep Q-Learning Experience Replay with SHapley Additive exPlanations

3Citations
Citations of this article
38Readers
Mendeley users who have this article in their library.

Abstract

Reinforcement Learning (RL) has shown promise in optimizing complex control and decision-making processes but Deep Reinforcement Learning (DRL) lacks interpretability, limiting its adoption in regulated sectors like manufacturing, finance, and healthcare. Difficulties arise from DRL’s opaque decision-making, hindering efficiency and resource use, this issue is amplified with every advancement. While many seek to move from Experience Replay to A3C, the latter demands more resources. Despite efforts to improve Experience Replay selection strategies, there is a tendency to keep the capacity high. We investigate training a Deep Convolutional Q-learning agent across 20 Atari games intentionally reducing Experience Replay capacity from (Formula presented.) to (Formula presented.). We find that a reduction from (Formula presented.) to (Formula presented.) doesn’t significantly affect rewards, offering a practical path to resource-efficient DRL. To illuminate agent decisions and align them with game mechanics, we employ a novel method: visualizing Experience Replay via Deep SHAP Explainer. This approach fosters comprehension and transparent, interpretable explanations, though any capacity reduction must be cautious to avoid overfitting. Our study demonstrates the feasibility of reducing Experience Replay and advocates for transparent, interpretable decision explanations using the Deep SHAP Explainer to promote enhancing resource efficiency in Experience Replay.

References Powered by Scopus

Human-level control through deep reinforcement learning

22573Citations
N/AReaders
Get full text

Use of Ranks in One-Criterion Variance Analysis

9111Citations
N/AReaders
Get full text

Learning to Predict by the Methods of Temporal Differences

3722Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Recent Applications of Explainable AI (XAI): A Systematic Literature Review

3Citations
N/AReaders
Get full text

A Comparative Analysis of SHAP, LIME, ANCHORS, and DICE for Interpreting a Dense Neural Network in Credit Card Fraud Detection

2Citations
N/AReaders
Get full text

Analyzing deep reinforcement learning model decisions with Shapley additive explanations for counter drone operations

0Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Sullivan, R. S., & Longo, L. (2023). Explaining Deep Q-Learning Experience Replay with SHapley Additive exPlanations. Machine Learning and Knowledge Extraction, 5(4), 1433–1455. https://doi.org/10.3390/make5040072

Readers' Seniority

Tooltip

Researcher 22

81%

PhD / Post grad / Masters / Doc 4

15%

Professor / Associate Prof. 1

4%

Readers' Discipline

Tooltip

Computer Science 23

85%

Business, Management and Accounting 3

11%

Engineering 1

4%

Article Metrics

Tooltip
Mentions
Blog Mentions: 1
News Mentions: 1
Social Media
Shares, Likes & Comments: 1

Save time finding and organizing research with Mendeley

Sign up for free