Multiple model-based reinforcement learning explains dopamine neuronal activity.

Neural Netw

ATR Computational Neuroscience Labs, 2-2-2 Hikaridai, Keihanna Science City, Kyoto 619-0288, Japan.

Published: August 2007

A number of computational models have explained the behavior of dopamine neurons in terms of temporal difference learning. However, earlier models cannot account for recent results of conditioning experiments; specifically, the behavior of dopamine neurons in case of variation of the interval between a cue stimulus and a reward has not been satisfyingly accounted for. We address this problem by using a modular architecture, in which each module consists of a reward predictor and a value estimator. A "responsibility signal", computed from the accuracy of the predictions of the reward predictors, is used to weight the contributions and learning of the value estimators. This multiple-model architecture gives an accurate account of the behavior of dopamine neurons in two specific experiments: when the reward is delivered earlier than expected, and when the stimulus-reward interval varies uniformly over a fixed range.

Download full-text PDF

Source
http://dx.doi.org/10.1016/j.neunet.2007.04.028DOI Listing

Publication Analysis

Top Keywords

behavior dopamine
12
dopamine neurons
12
multiple model-based
4
model-based reinforcement
4
reinforcement learning
4
learning explains
4
dopamine
4
explains dopamine
4
dopamine neuronal
4
neuronal activity
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!