A causal deep learning framework for classifying phonemes in cochlear implants

Conference Paper

Speech intelligibility in cochlear implant (CI) users degrades considerably in listening environments with reverberation and noise. Previous research in automatic speech recognition (ASR) has shown that phoneme-based speech enhancement algorithms improve ASR system performance in reverberant environments as compared to a global model. However, phoneme-specific speech processing has not yet been implemented in CIs. In this paper, we propose a causal deep learning framework for classifying phonemes using features extracted at the time-frequency resolution of a CI processor. We trained and tested long short-term memory networks to classify phonemes and manner of articulation in anechoic and reverberant conditions. The results showed that CI-inspired features provide slightly higher levels of performance than traditional ASR features. To the best of our knowledge, this study is the first to provide a classification framework with the potential to categorize phonetic units in real-time in a CI.

Full Text

Duke Authors

Cited Authors

  • Chu, K; Collins, L; Mainsah, B

Published Date

  • January 1, 2021

Published In

Volume / Issue

  • 2021-June /

Start / End Page

  • 6498 - 6502

International Standard Serial Number (ISSN)

  • 1520-6149

Digital Object Identifier (DOI)

  • 10.1109/ICASSP39728.2021.9413986

Citation Source

  • Scopus