Evaluating the TD model of classical conditioning.

Learn Behav

Princeton Neuroscience Institute and Department of Mechanical & Aerospace Engineering, Princeton University, 3-N-12 Green Hall, Princeton, NJ 08542, USA.

Published: September 2012

The temporal-difference (TD) algorithm from reinforcement learning provides a simple method for incrementally learning predictions of upcoming events. Applied to classical conditioning, TD models suppose that animals learn a real-time prediction of the unconditioned stimulus (US) on the basis of all available conditioned stimuli (CSs). In the TD model, similar to other error-correction models, learning is driven by prediction errors--the difference between the change in US prediction and the actual US. With the TD model, however, learning occurs continuously from moment to moment and is not artificially constrained to occur in trials. Accordingly, a key feature of any TD model is the assumption about the representation of a CS on a moment-to-moment basis. Here, we evaluate the performance of the TD model with a heretofore unexplored range of classical conditioning tasks. To do so, we consider three stimulus representations that vary in their degree of temporal generalization and evaluate how the representation influences the performance of the TD model on these conditioning tasks.

Download full-text PDF

Source
http://dx.doi.org/10.3758/s13420-012-0082-6DOI Listing

Publication Analysis

Top Keywords

classical conditioning
12
performance model
8
conditioning tasks
8
model
5
evaluating model
4
model classical
4
conditioning
4
conditioning temporal-difference
4
temporal-difference algorithm
4
algorithm reinforcement
4

Similar Publications

Want AI Summaries of new PubMed Abstracts delivered to your In-box?

Enter search terms and have AI summaries delivered each week - change queries or unsubscribe any time!