Reference-point centering and range-adaptation enhance human reinforcement learning at the cost of irrational preferences

50Citations
Citations of this article
117Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

In economics and perceptual decision-making contextual effects are well documented, where decision weights are adjusted as a function of the distribution of stimuli. Yet, in reinforcement learning literature whether and how contextual information pertaining to decision states is integrated in learning algorithms has received comparably little attention. Here, we investigate reinforcement learning behavior and its computational substrates in a task where we orthogonally manipulate outcome valence and magnitude, resulting in systematic variations in state-values. Model comparison indicates that subjects’ behavior is best accounted for by an algorithm which includes both reference point-dependence and range-adaptation—two crucial features of state-dependent valuation. In addition, we find that state-dependent outcome valuation progressively emerges, is favored by increasing outcome information and correlated with explicit understanding of the task structure. Finally, our data clearly show that, while being locally adaptive (for instance in negative valence and small magnitude contexts), state-dependent valuation comes at the cost of seemingly irrational choices, when options are extrapolated out from their original contexts.

References Powered by Scopus

Technical Note: Q-Learning

11576Citations
N/AReaders
Get full text

A neural substrate of prediction and reward

6653Citations
N/AReaders
Get full text

Choices, values, and frames

4793Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Resource-rational decision making

75Citations
N/AReaders
Get full text

Modelling Stock Markets by Multi-agent Reinforcement Learning

46Citations
N/AReaders
Get full text

Contextual influence on confidence judgments in human reinforcement learning

46Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Bavard, S., Lebreton, M., Khamassi, M., Coricelli, G., & Palminteri, S. (2018). Reference-point centering and range-adaptation enhance human reinforcement learning at the cost of irrational preferences. Nature Communications, 9(1). https://doi.org/10.1038/s41467-018-06781-2

Readers over time

‘18‘19‘20‘21‘22‘23‘24‘2507142128

Readers' Seniority

Tooltip

PhD / Post grad / Masters / Doc 66

77%

Researcher 16

19%

Professor / Associate Prof. 3

3%

Lecturer / Post doc 1

1%

Readers' Discipline

Tooltip

Psychology 36

47%

Neuroscience 31

40%

Engineering 6

8%

Computer Science 4

5%

Save time finding and organizing research with Mendeley

Sign up for free
0