Many reinforcement learning (RL) algorithms approximate an optimal value function. Once the function is known, it is easy to determine an optimal policy. For most real-world applications, however, the value function is too complex to be represented by lookup tables, making it necessary to use function approximators such as neural networks. In this case, convergence to the optimal value function is no longer guaranteed and it becomes important to know to which extent performance diminishes when one uses approximate value functions instead of optimal ones. This problem has recently been discussed in the context of expectation-based Markov decision problems. Our analysis generalizes this work to minimax-based Markov decision problems, yields new results for expectation-based tasks, and shows how minimax-based and expectation-based Markov decision problems relate.
CITATION STYLE
Heger, M. (1996). Loss from imperfect value functions in expectation-based and minimax-based tasks. Machine Learning, 22(1–3), 197–225. https://doi.org/10.1007/BF00114728
Mendeley helps you to discover research relevant for your work.