Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice

Metha, Jeremy A. and Brian, Maddison L. and Oberrauch, Sara and Barnes, Samuel A. and Featherby, Travis J. and Bossaerts, Peter and Murawski, Carsten and Hoyer, Daniel and Jacobson, Laura H. (2020) Separating Probability and Reversal Learning in a Novel Probabilistic Reversal Learning Task for Mice. Frontiers in Behavioral Neuroscience, 13. ISSN 1662-5153

[thumbnail of pubmed-zip/versions/1/package-entries/fnbeh-13-00270/fnbeh-13-00270.pdf] Text
pubmed-zip/versions/1/package-entries/fnbeh-13-00270/fnbeh-13-00270.pdf - Published Version

Download (3MB)

Abstract

The exploration/exploitation tradeoff – pursuing a known reward vs. sampling from lesser known options in the hope of finding a better payoff – is a fundamental aspect of learning and decision making. In humans, this has been studied using multi-armed bandit tasks. The same processes have also been studied using simplified probabilistic reversal learning (PRL) tasks with binary choices. Our investigations suggest that protocols previously used to explore PRL in mice may prove beyond their cognitive capacities, with animals performing at a no-better-than-chance level. We sought a novel probabilistic learning task to improve behavioral responding in mice, whilst allowing the investigation of the exploration/exploitation tradeoff in decision making. To achieve this, we developed a two-lever operant chamber task with levers corresponding to different probabilities (high/low) of receiving a saccharin reward, reversing the reward contingencies associated with levers once animals reached a threshold of 80% responding at the high rewarding lever. We found that, unlike in existing PRL tasks, mice are able to learn and behave near optimally with 80% high/20% low reward probabilities. Altering the reward contingencies towards equality showed that some mice displayed preference for the high rewarding lever with probabilities as close as 60% high/40% low. Additionally, we show that animal choice behavior can be effectively modelled using reinforcement learning (RL) models incorporating learning rates for positive and negative prediction error, a perseveration parameter, and a noise parameter. This new decision task, coupled with RL analyses, advances access to investigate the neuroscience of the exploration/exploitation tradeoff in decision making.

Item Type: Article
Subjects: STM Library Press > Biological Science
Depositing User: Unnamed user with email support@stmlibrarypress.com
Date Deposited: 16 Jan 2023 09:20
Last Modified: 01 Jul 2024 07:28
URI: http://journal.scienceopenlibraries.com/id/eprint/189

Actions (login required)

View Item
View Item