Aversion to Option Loss in a Restless Bandit Task

5Citations
Citations of this article
14Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In everyday life, people need to make choices without full information about the environment, which poses an explore-exploit dilemma in which one must balance the need to learn about the world and the need to obtain rewards from it. The explore-exploit dilemma is often studied using the multi-armed restless bandit task, in which people repeatedly select from multiple options, and human behaviour is modelled as a form of reinforcement learning via Kalman filters. Inspired by work in the judgment and decision-making literature, we present two experiments using multi-armed bandit tasks in both static and dynamic environments, in situations where options can become unviable and vanish if they are not pursued. A Kalman filter model using Thompson sampling provides an excellent account of human learning in a standard restless bandit task, but there are systematic departures in the vanishing bandit task. We explore the nature of this loss aversion signal and consider theoretical explanations for the results.

Cite

CITATION STYLE

APA

Navarro, D. J., Tran, P., & Baz, N. (2018). Aversion to Option Loss in a Restless Bandit Task. Computational Brain and Behavior, 1(2), 151–164. https://doi.org/10.1007/s42113-018-0010-8

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free