Skip to main content
Journal cover image

Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings.

Publication ,  Journal Article
Bulgarelli, F; Bergelson, E
Published in: Behavior research methods
April 2020

The LENA system has revolutionized research on language acquisition, providing both a wearable device to collect day-long recordings of children's environments, and a set of automated outputs that process, identify, and classify speech using proprietary algorithms. This output includes information about input sources (e.g., adult male, electronics). While this system has been tested across a variety of settings, here we delve deeper into validating the accuracy and reliability of LENA's automated diarization, i.e., tags of who is talking. Specifically, we compare LENA's output with a gold standard set of manually generated talker tags from a dataset of 88 day-long recordings, taken from 44 infants at 6 and 7 months, which includes 57,983 utterances. We compare accuracy across a range of classifications from the original Lena Technical Report, alongside a set of analyses examining classification accuracy by utterance type (e.g., declarative, singing). Consistent with previous validations, we find overall high agreement between the human and LENA-generated speaker tags for adult speech in particular, with poorer performance identifying child, overlap, noise, and electronic speech (accuracy range across all measures: 0-92%). We discuss several clear benefits of using this automated system alongside potential caveats based on the error patterns we observe, concluding with implications for research using LENA-generated speaker tags.

Duke Scholars

Altmetric Attention Stats
Dimensions Citation Stats

Published In

Behavior research methods

DOI

EISSN

1554-3528

ISSN

1554-351X

Publication Date

April 2020

Volume

52

Issue

2

Start / End Page

641 / 653

Related Subject Headings

  • Speech
  • Reproducibility of Results
  • Language Development
  • Infant
  • Humans
  • Experimental Psychology
  • Data Collection
  • Automation
  • Algorithms
  • 5204 Cognitive and computational psychology
 

Citation

APA
Chicago
ICMJE
MLA
NLM
Bulgarelli, F., & Bergelson, E. (2020). Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings. Behavior Research Methods, 52(2), 641–653. https://doi.org/10.3758/s13428-019-01265-7
Bulgarelli, Federica, and Elika Bergelson. “Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings.Behavior Research Methods 52, no. 2 (April 2020): 641–53. https://doi.org/10.3758/s13428-019-01265-7.
Bulgarelli F, Bergelson E. Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings. Behavior research methods. 2020 Apr;52(2):641–53.
Bulgarelli, Federica, and Elika Bergelson. “Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings.Behavior Research Methods, vol. 52, no. 2, Apr. 2020, pp. 641–53. Epmc, doi:10.3758/s13428-019-01265-7.
Bulgarelli F, Bergelson E. Look who's talking: A comparison of automated and human-generated speaker tags in naturalistic day-long recordings. Behavior research methods. 2020 Apr;52(2):641–653.
Journal cover image

Published In

Behavior research methods

DOI

EISSN

1554-3528

ISSN

1554-351X

Publication Date

April 2020

Volume

52

Issue

2

Start / End Page

641 / 653

Related Subject Headings

  • Speech
  • Reproducibility of Results
  • Language Development
  • Infant
  • Humans
  • Experimental Psychology
  • Data Collection
  • Automation
  • Algorithms
  • 5204 Cognitive and computational psychology