A probabilistic argumentation framework for reinforcement learning agents: Towards a mentalistic approach to agent profiles

11Citations
Citations of this article
35Readers
Mendeley users who have this article in their library.
Get full text

Abstract

A bounded-reasoning agent may face two dimensions of uncertainty: firstly, the uncertainty arising from partial information and conflicting reasons, and secondly, the uncertainty arising from the stochastic nature of its actions and the environment. This paper attempts to address both dimensions within a single unified framework, by bringing together probabilistic argumentation and reinforcement learning. We show how a probabilistic rule-based argumentation framework can capture Markov decision processes and reinforcement learning agents; and how the framework allows us to characterise agents and their argument-based motivations from both a logic-based perspective and a probabilistic perspective. We advocate and illustrate the use of our approach to capture models of agency and norms, and argue that, in addition to providing a novel method for investigating agent types, the unified framework offers a sound basis for taking a mentalistic approach to agent profiles.

Cite

CITATION STYLE

APA

Riveret, R., Gao, Y., Governatori, G., Rotolo, A., Pitt, J., & Sartor, G. (2019). A probabilistic argumentation framework for reinforcement learning agents: Towards a mentalistic approach to agent profiles. Autonomous Agents and Multi-Agent Systems, 33(1–2), 216–274. https://doi.org/10.1007/s10458-019-09404-2

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free