Modeling rapid language learning by distilling Bayesian priors into artificial neural networks

Research output: Contribution to journalArticlepeer-review

6 Scopus citations

Abstract

Humans can learn languages from remarkably little experience. Developing computational models that explain this ability has been a major challenge in cognitive science. Existing approaches have been successful at explaining how humans generalize rapidly in controlled settings but are usually too restrictive to tractably handle naturalistic data. We show that learning from limited naturalistic data is possible with an approach that bridges the divide between two popular modeling traditions: Bayesian models and neural networks. This approach distills a Bayesian model’s inductive biases—the factors that guide generalization—into a neural network that has flexible representations. Like a Bayesian model, the resulting system can learn formal linguistic patterns from limited data. Like a neural network, it can also learn aspects of English syntax from naturally-occurring sentences. Thus, this model provides a single system that can learn rapidly and can handle naturalistic data.

Original languageEnglish (US)
Article number4676
JournalNature communications
Volume16
Issue number1
DOIs
StatePublished - Dec 2025

All Science Journal Classification (ASJC) codes

  • General Chemistry
  • General Biochemistry, Genetics and Molecular Biology
  • General Physics and Astronomy

Fingerprint

Dive into the research topics of 'Modeling rapid language learning by distilling Bayesian priors into artificial neural networks'. Together they form a unique fingerprint.

Cite this