Contractivity of Bellman operator in risk averse dynamic programming with infinite horizon

0Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The paper deals with a risk averse dynamic programming problem with infinite horizon. First, the required assumptions are formulated to have the problem well defined. Then the Bellman equation is derived, which may be also seen as a standalone reinforcement learning problem. The fact that the Bellman operator is contraction is proved, guaranteeing convergence of various solution algorithms used for dynamic programming as well as reinforcement learning problems, which we demonstrate on the value iteration and the policy iteration algorithms.

Cite

CITATION STYLE

APA

Kopa, M., & Šmíd, M. (2023). Contractivity of Bellman operator in risk averse dynamic programming with infinite horizon. Operations Research Letters, 51(2), 133–136. https://doi.org/10.1016/j.orl.2023.01.008

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free