Multiple associative structures created by reinforcement and incidental statistical learning mechanisms

Nat Commun. 2019 Oct 23;10(1):4835. doi: 10.1038/s41467-019-12557-z.

Abstract

Learning the structure of the world can be driven by reinforcement but also occurs incidentally through experience. Reinforcement learning theory has provided insight into how prediction errors drive updates in beliefs but less attention has been paid to the knowledge resulting from such learning. Here we contrast associative structures formed through reinforcement and experience of task statistics. BOLD neuroimaging in human volunteers demonstrates rigid representations of rewarded sequences in temporal pole and posterior orbito-frontal cortex, which are constructed backwards from reward. By contrast, medial prefrontal cortex and a hippocampal-amygdala border region carry reward-related knowledge but also flexible statistical knowledge of the currently relevant task model. Intriguingly, ventral striatum encodes prediction error responses but not the full RL- or statistically derived task knowledge. In summary, representations of task knowledge are derived via multiple learning processes operating at different time scales that are associated with partially overlapping and partially specialized anatomical regions.

Publication types

  • Research Support, Non-U.S. Gov't

MeSH terms

  • Adult
  • Amygdala / diagnostic imaging
  • Amygdala / physiology
  • Association Learning / physiology*
  • Brain / diagnostic imaging*
  • Brain / physiology
  • Female
  • Functional Neuroimaging
  • Hippocampus / diagnostic imaging
  • Hippocampus / physiology
  • Humans
  • Learning / physiology
  • Magnetic Resonance Imaging
  • Male
  • Prefrontal Cortex / diagnostic imaging
  • Prefrontal Cortex / physiology
  • Reinforcement, Psychology*
  • Temporal Lobe / diagnostic imaging
  • Temporal Lobe / physiology
  • Ventral Striatum / diagnostic imaging
  • Ventral Striatum / physiology
  • Young Adult