Ontology highlight
ABSTRACT:
SUBMITTER: Glascher J
PROVIDER: S-EPMC2895323 | biostudies-other | 2010 May
REPOSITORIES: biostudies-other
Gläscher Jan J Daw Nathaniel N Dayan Peter P O'Doherty John P JP
Neuron 20100501 4
Reinforcement learning (RL) uses sequential experience with situations ("states") and outcomes to assess actions. Whereas model-free RL uses this experience directly, in the form of a reward prediction error (RPE), model-based RL uses it indirectly, building a model of the state transition and outcome structure of the environment, and evaluating actions by searching this model. A state prediction error (SPE) plays a central role, reporting discrepancies between the current model and the observed ...[more]