Content area

Abstract

The temporal-difference (TD) algorithm from reinforcement learning provides a simple method for incrementally learning predictions of upcoming events. Applied to classical conditioning, TD models suppose that animals learn a real-time prediction of the unconditioned stimulus (US) on the basis of all available conditioned stimuli (CSs). In the TD model, similar to other error-correction models, learning is driven by prediction errors-the difference between the change in US prediction and the actual US. With the TD model, however, learning occurs continuously from moment to moment and is not artificially constrained to occur in trials. Accordingly, a key feature of any TD model is the assumption about the representation of a CS on a moment-to-moment basis. Here, we evaluate the performance of the TD model with a heretofore unexplored range of classical conditioning tasks. To do so, we consider three stimulus representations that vary in their degree of temporal generalization and evaluate how the representation influences the performance of the TD model on these conditioning tasks. [PUBLICATION ABSTRACT]

Details

Title
Evaluating the TD model of classical conditioning
Author
Ludvig, Elliot A; Sutton, Richard S; Kehoe, E James
Pages
305-319
Publication year
2012
Publication date
Sep 2012
Publisher
Springer Nature B.V.
ISSN
15434494
e-ISSN
15434508
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
1470425373
Copyright
Copyright Springer Science & Business Media Sep 2012