Adaptor grammars: A framework for specifying compositional nonparametric Bayesian models

Mark Johnson, Thomas L. Griffiths, Sharon Goldwater

Research output: Chapter in Book/Report/Conference proceedingConference contribution

163 Scopus citations

Abstract

This paper introduces adaptor grammars, a class of probabilistic models of language that generalize probabilistic context-free grammars (PCFGs). Adaptor grammars augment the probabilistic rules of PCFGs with "adaptors" that can induce dependencies among successive uses. With a particular choice of adaptor, based on the Pitman-Yor process, nonparametric Bayesian models of language using Dirichlet processes and hierarchical Dirichlet processes can be written as simple grammars. We present a general-purpose inference algorithm for adaptor grammars, making it easy to define and use such models, and illustrate how several existing nonparametric Bayesian models can be expressed within this framework.

Original languageEnglish (US)
Title of host publicationAdvances in Neural Information Processing Systems 19 - Proceedings of the 2006 Conference
Pages641-648
Number of pages8
StatePublished - 2007
Externally publishedYes
Event20th Annual Conference on Neural Information Processing Systems, NIPS 2006 - Vancouver, BC, Canada
Duration: Dec 4 2006Dec 7 2006

Publication series

NameAdvances in Neural Information Processing Systems
ISSN (Print)1049-5258

Other

Other20th Annual Conference on Neural Information Processing Systems, NIPS 2006
Country/TerritoryCanada
CityVancouver, BC
Period12/4/0612/7/06

All Science Journal Classification (ASJC) codes

  • Computer Networks and Communications
  • Information Systems
  • Signal Processing

Fingerprint

Dive into the research topics of 'Adaptor grammars: A framework for specifying compositional nonparametric Bayesian models'. Together they form a unique fingerprint.

Cite this