TY - GEN
T1 - Constructing a hypothesis space from the Web for large-scale Bayesian word learning
AU - Abbott, Joshua T.
AU - Austerweil, Joseph L.
AU - Griffiths, Thomas L.
N1 - Publisher Copyright:
© CogSci 2012.All rights reserved.
PY - 2012
Y1 - 2012
N2 - The Bayesian generalization framework has been successful in explaining how people generalize a property from a few observed stimuli to novel stimuli, across several different domains. To create a successful Bayesian generalization model, modelers typically specify a hypothesis space and prior probability distribution for each specific domain. However, this raises two problems: the models do not scale beyond the (typically small-scale) domain that they were designed for, and the explanatory power of the models is reduced by their reliance on a hand-coded hypothesis space and prior. To solve these two problems, we propose a method for deriving hypothesis spaces and priors from large online databases. We evaluate our method by constructing a hypothesis space and prior for a Bayesian word learning model from WordNet, a large online database that encodes the semantic relationships between words as a network. After validating our approach by replicating a previous word learning study, we apply the same model to a new experiment featuring three additional taxonomic domains (clothing, containers, and seats). In both experiments, we found that the same automatically constructed hypothesis space explains the complex pattern of generalization behavior, producing accurate predictions across a total of six different domains.
AB - The Bayesian generalization framework has been successful in explaining how people generalize a property from a few observed stimuli to novel stimuli, across several different domains. To create a successful Bayesian generalization model, modelers typically specify a hypothesis space and prior probability distribution for each specific domain. However, this raises two problems: the models do not scale beyond the (typically small-scale) domain that they were designed for, and the explanatory power of the models is reduced by their reliance on a hand-coded hypothesis space and prior. To solve these two problems, we propose a method for deriving hypothesis spaces and priors from large online databases. We evaluate our method by constructing a hypothesis space and prior for a Bayesian word learning model from WordNet, a large online database that encodes the semantic relationships between words as a network. After validating our approach by replicating a previous word learning study, we apply the same model to a new experiment featuring three additional taxonomic domains (clothing, containers, and seats). In both experiments, we found that the same automatically constructed hypothesis space explains the complex pattern of generalization behavior, producing accurate predictions across a total of six different domains.
KW - Bayesian modeling
KW - concept learning
KW - generalization
KW - online databases
KW - word learning
UR - http://www.scopus.com/inward/record.url?scp=84898930325&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84898930325&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:84898930325
T3 - Building Bridges Across Cognitive Sciences Around the World - Proceedings of the 34th Annual Meeting of the Cognitive Science Society, CogSci 2012
SP - 54
EP - 59
BT - Building Bridges Across Cognitive Sciences Around the World - Proceedings of the 34th Annual Meeting of the Cognitive Science Society, CogSci 2012
A2 - Miyake, Naomi
A2 - Peebles, David
A2 - Cooper, Richard P.
PB - The Cognitive Science Society
T2 - 34th Annual Meeting of the Cognitive Science Society: Building Bridges Across Cognitive Sciences Around the World, CogSci 2012
Y2 - 1 August 2012 through 4 August 2012
ER -