Abstract
Reinforcement learning models have been productively applied to identify neural correlates of the value of social information. However, by operationalizing social information as a lean, reward-predictive cue, this literature underestimates the richness of human social learning: Humans readily go beyond action-outcome mappings and can draw flexible inferences from a single observation. We argue that computational models of social learning need minds, that is, a generative model of how others’ unobservable mental states cause their observable actions. Recent advances in inferential social learning suggest that even young children learn from others by using an intuitive, generative model of other minds. Bridging developmental, Bayesian, and reinforcement learning perspectives can enrich our understanding of the neural bases of distinctively human social learning.
Original language | English (US) |
---|---|
Pages (from-to) | 110-115 |
Number of pages | 6 |
Journal | Current Opinion in Behavioral Sciences |
Volume | 38 |
DOIs | |
State | Published - Apr 2021 |
Externally published | Yes |
All Science Journal Classification (ASJC) codes
- Psychiatry and Mental health
- Cognitive Neuroscience
- Behavioral Neuroscience