Bayesian Word Learning in Multiple Language Environments.

Published

Journal Article

Infant language learners are faced with the difficult inductive problem of determining how new words map to novel or known objects in their environment. Bayesian inference models have been successful at using the sparse information available in natural child-directed speech to build candidate lexicons and infer speakers' referential intentions. We begin by asking how a Bayesian model optimized for monolingual input (the Intentional Model; Frank et al., 2009) generalizes to new monolingual or bilingual corpora and find that, especially in the case of the bilingual input, the model shows a significant decrease in performance. In the next experiment, we propose the ME Model, a modified Bayesian model, which approximates infants' mutual exclusivity bias to support the differential demands of monolingual and bilingual learning situations. The extended model is assessed using the same corpora of real child-directed speech, showing that its performance is more robust against varying input and less dependent than the Intentional Model on optimization of its parsimony parameter. We argue that both monolingual and bilingual demands on word learning are important considerations for a computational model, as they can yield significantly different results than when only one such context is considered.

Full Text

Cited Authors

  • Zinszer, BD; Rolotti, SV; Li, F; Li, P

Published Date

  • May 2018

Published In

Volume / Issue

  • 42 Suppl 2 /

Start / End Page

  • 439 - 462

PubMed ID

  • 29154481

Pubmed Central ID

  • 29154481

Electronic International Standard Serial Number (EISSN)

  • 1551-6709

Digital Object Identifier (DOI)

  • 10.1111/cogs.12567

Language

  • eng

Conference Location

  • United States