Expectation optimization with probabilistic guarantees in POMDPs with discounted-sum objectives

3Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.

Abstract

Partially-observable Markov decision processes (POMDPs) with discounted-sum payoff are a standard framework to model a wide range of problems related to decision making under uncertainty. Traditionally, the goal has been to obtain policies that optimize the expectation of the discounted-sum payoff. A key drawback of the expectation measure is that even low probability events with extreme payoff can significantly affect the expectation, and thus the obtained policies are not necessarily risk-averse. An alternate approach is to optimize the probability that the payoff is above a certain threshold, which allows obtaining risk-averse policies, but ignores optimization of the expectation. We consider the expectation optimization with probabilistic guarantee (EOPG) problem, where the goal is to optimize the expectation ensuring that the payoff is above a given threshold with at least a specified probability. We present several results on the EOPG problem, including the first algorithm to solve it.

Cite

CITATION STYLE

APA

Chatterjee, K., Elgyütt, A., Novotný, P., & Rouillé, O. (2018). Expectation optimization with probabilistic guarantees in POMDPs with discounted-sum objectives. In IJCAI International Joint Conference on Artificial Intelligence (Vol. 2018-July, pp. 4692–4699). International Joint Conferences on Artificial Intelligence. https://doi.org/10.24963/ijcai.2018/652

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free