Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of t...
Main Authors: | , , , |
---|---|
Other Authors: | |
Format: | Article |
Published: |
Frontiers Media SA
2021
|
Online Access: | https://hdl.handle.net/1721.1/129383 |
_version_ | 1811096326143737856 |
---|---|
author | Haro, Stephanie Smalt, Christopher J. Ciccarelli, Gregory A. Quatieri, Thomas F. |
author2 | Lincoln Laboratory |
author_facet | Lincoln Laboratory Haro, Stephanie Smalt, Christopher J. Ciccarelli, Gregory A. Quatieri, Thomas F. |
author_sort | Haro, Stephanie |
collection | MIT |
description | Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of these factors complicates the prescription oftreatment or therapy to improve hearing function. Damage tothe auditory peripherycan be studied in animals; however, this method alone is not enough to understandthe impact of hearing loss on speech perception. Computational auditory models bridgethe gap between animal studies and human speech perception.Perturbations to themodeled auditory systems can permit mechanism-based investigations into observedhuman behavior. In this study, we propose a computational model that accounts forthe complex interactions between different hearing damagemechanisms and simulateshuman speech-in-noise perception. The model performs a digit classification task asa human would, with only acoustic sound pressure as input. Thus, we can use themodel’s performance as a proxy for human performance. This two-stage model consistsof a biophysical cochlear-nerve spike generator followed by a deep neural network(DNN) classifier. We hypothesize that sudden damage to the periphery affects speechperception and that central nervous system adaptation overtime may compensatefor peripheral hearing damage. Our model achieved human-like performance acrosssignal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving50% digit recognition accuracy at−20.7 dB SNR. Results were comparable to eightNH participants on the same task who achieved 50% behavioralperformance at−22dB SNR. We also simulated medial olivocochlear reflex (MOCR)and auditory nervefiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs comparedto higher SNRs. Our simulated performance following ANF loss is consistent withthe hypothesis that cochlear synaptopathy impacts communication in backgroundnoise more so than in quiet. Following the insult of various cochlear degradations, weimplemented extreme and conservative adaptation through the DNN. At the lowest SNRs(<0 dB), both adapted models were unable to fully recover NH performance, even withhundreds of thousands of training samples. This implies a limit on performance recoveryfollowing peripheral damage in our human-inspired DNN architecture. |
first_indexed | 2024-09-23T16:42:02Z |
format | Article |
id | mit-1721.1/129383 |
institution | Massachusetts Institute of Technology |
last_indexed | 2024-09-23T16:42:02Z |
publishDate | 2021 |
publisher | Frontiers Media SA |
record_format | dspace |
spelling | mit-1721.1/1293832022-09-29T20:51:26Z Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance Haro, Stephanie Smalt, Christopher J. Ciccarelli, Gregory A. Quatieri, Thomas F. Lincoln Laboratory Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of these factors complicates the prescription oftreatment or therapy to improve hearing function. Damage tothe auditory peripherycan be studied in animals; however, this method alone is not enough to understandthe impact of hearing loss on speech perception. Computational auditory models bridgethe gap between animal studies and human speech perception.Perturbations to themodeled auditory systems can permit mechanism-based investigations into observedhuman behavior. In this study, we propose a computational model that accounts forthe complex interactions between different hearing damagemechanisms and simulateshuman speech-in-noise perception. The model performs a digit classification task asa human would, with only acoustic sound pressure as input. Thus, we can use themodel’s performance as a proxy for human performance. This two-stage model consistsof a biophysical cochlear-nerve spike generator followed by a deep neural network(DNN) classifier. We hypothesize that sudden damage to the periphery affects speechperception and that central nervous system adaptation overtime may compensatefor peripheral hearing damage. Our model achieved human-like performance acrosssignal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving50% digit recognition accuracy at−20.7 dB SNR. Results were comparable to eightNH participants on the same task who achieved 50% behavioralperformance at−22dB SNR. We also simulated medial olivocochlear reflex (MOCR)and auditory nervefiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs comparedto higher SNRs. Our simulated performance following ANF loss is consistent withthe hypothesis that cochlear synaptopathy impacts communication in backgroundnoise more so than in quiet. Following the insult of various cochlear degradations, weimplemented extreme and conservative adaptation through the DNN. At the lowest SNRs(<0 dB), both adapted models were unable to fully recover NH performance, even withhundreds of thousands of training samples. This implies a limit on performance recoveryfollowing peripheral damage in our human-inspired DNN architecture. United States. Department of Defense. Research and Engineering (Air Force Contract No. FA8702-15-D-0001) National Institutes of Health (U.S.) (T32 Trainee Grant No. 5T32DC000038-27) National Science Foundation (U.S.). Graduate Research Fellowship Program (Grant DGE1745303) 2021-01-12T16:27:45Z 2021-01-12T16:27:45Z 2020-12 2020-07 Article http://purl.org/eprint/type/JournalArticle 2381-2710 https://hdl.handle.net/1721.1/129383 Haro, Stephanie et al. “Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance.” Frontiers in neuroscience, 14 (December 2020): 588448 © 2020 The Author(s) 10.3389/fenrg.2020.585461 Frontiers in neuroscience Creative Commons Attribution 4.0 International license https://creativecommons.org/licenses/by/4.0/ application/pdf Frontiers Media SA Frontiers |
spellingShingle | Haro, Stephanie Smalt, Christopher J. Ciccarelli, Gregory A. Quatieri, Thomas F. Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title | Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title_full | Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title_fullStr | Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title_full_unstemmed | Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title_short | Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance |
title_sort | deep neural network model of hearing impaired speech in noise performance |
url | https://hdl.handle.net/1721.1/129383 |
work_keys_str_mv | AT harostephanie deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance AT smaltchristopherj deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance AT ciccarelligregorya deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance AT quatierithomasf deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance |