Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance

Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of t...

Full description

Bibliographic Details
Main Authors: Haro, Stephanie, Smalt, Christopher J., Ciccarelli, Gregory A., Quatieri, Thomas F.
Other Authors: Lincoln Laboratory
Format: Article
Published: Frontiers Media SA 2021
Online Access:https://hdl.handle.net/1721.1/129383
_version_ 1811096326143737856
author Haro, Stephanie
Smalt, Christopher J.
Ciccarelli, Gregory A.
Quatieri, Thomas F.
author2 Lincoln Laboratory
author_facet Lincoln Laboratory
Haro, Stephanie
Smalt, Christopher J.
Ciccarelli, Gregory A.
Quatieri, Thomas F.
author_sort Haro, Stephanie
collection MIT
description Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of these factors complicates the prescription oftreatment or therapy to improve hearing function. Damage tothe auditory peripherycan be studied in animals; however, this method alone is not enough to understandthe impact of hearing loss on speech perception. Computational auditory models bridgethe gap between animal studies and human speech perception.Perturbations to themodeled auditory systems can permit mechanism-based investigations into observedhuman behavior. In this study, we propose a computational model that accounts forthe complex interactions between different hearing damagemechanisms and simulateshuman speech-in-noise perception. The model performs a digit classification task asa human would, with only acoustic sound pressure as input. Thus, we can use themodel’s performance as a proxy for human performance. This two-stage model consistsof a biophysical cochlear-nerve spike generator followed by a deep neural network(DNN) classifier. We hypothesize that sudden damage to the periphery affects speechperception and that central nervous system adaptation overtime may compensatefor peripheral hearing damage. Our model achieved human-like performance acrosssignal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving50% digit recognition accuracy at−20.7 dB SNR. Results were comparable to eightNH participants on the same task who achieved 50% behavioralperformance at−22dB SNR. We also simulated medial olivocochlear reflex (MOCR)and auditory nervefiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs comparedto higher SNRs. Our simulated performance following ANF loss is consistent withthe hypothesis that cochlear synaptopathy impacts communication in backgroundnoise more so than in quiet. Following the insult of various cochlear degradations, weimplemented extreme and conservative adaptation through the DNN. At the lowest SNRs(<0 dB), both adapted models were unable to fully recover NH performance, even withhundreds of thousands of training samples. This implies a limit on performance recoveryfollowing peripheral damage in our human-inspired DNN architecture.
first_indexed 2024-09-23T16:42:02Z
format Article
id mit-1721.1/129383
institution Massachusetts Institute of Technology
last_indexed 2024-09-23T16:42:02Z
publishDate 2021
publisher Frontiers Media SA
record_format dspace
spelling mit-1721.1/1293832022-09-29T20:51:26Z Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance Haro, Stephanie Smalt, Christopher J. Ciccarelli, Gregory A. Quatieri, Thomas F. Lincoln Laboratory Many individuals struggle to understand speech in listening scenarios that includereverberation and background noise. An individual’s ability to understand speech arisesfrom a combination of peripheral auditory function, central auditory function, and generalcognitive abilities. The interaction of these factors complicates the prescription oftreatment or therapy to improve hearing function. Damage tothe auditory peripherycan be studied in animals; however, this method alone is not enough to understandthe impact of hearing loss on speech perception. Computational auditory models bridgethe gap between animal studies and human speech perception.Perturbations to themodeled auditory systems can permit mechanism-based investigations into observedhuman behavior. In this study, we propose a computational model that accounts forthe complex interactions between different hearing damagemechanisms and simulateshuman speech-in-noise perception. The model performs a digit classification task asa human would, with only acoustic sound pressure as input. Thus, we can use themodel’s performance as a proxy for human performance. This two-stage model consistsof a biophysical cochlear-nerve spike generator followed by a deep neural network(DNN) classifier. We hypothesize that sudden damage to the periphery affects speechperception and that central nervous system adaptation overtime may compensatefor peripheral hearing damage. Our model achieved human-like performance acrosssignal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving50% digit recognition accuracy at−20.7 dB SNR. Results were comparable to eightNH participants on the same task who achieved 50% behavioralperformance at−22dB SNR. We also simulated medial olivocochlear reflex (MOCR)and auditory nervefiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs comparedto higher SNRs. Our simulated performance following ANF loss is consistent withthe hypothesis that cochlear synaptopathy impacts communication in backgroundnoise more so than in quiet. Following the insult of various cochlear degradations, weimplemented extreme and conservative adaptation through the DNN. At the lowest SNRs(<0 dB), both adapted models were unable to fully recover NH performance, even withhundreds of thousands of training samples. This implies a limit on performance recoveryfollowing peripheral damage in our human-inspired DNN architecture. United States. Department of Defense. Research and Engineering (Air Force Contract No. FA8702-15-D-0001) National Institutes of Health (U.S.) (T32 Trainee Grant No. 5T32DC000038-27) National Science Foundation (U.S.). Graduate Research Fellowship Program (Grant DGE1745303) 2021-01-12T16:27:45Z 2021-01-12T16:27:45Z 2020-12 2020-07 Article http://purl.org/eprint/type/JournalArticle 2381-2710 https://hdl.handle.net/1721.1/129383 Haro, Stephanie et al. “Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance.” Frontiers in neuroscience, 14 (December 2020): 588448 © 2020 The Author(s) 10.3389/fenrg.2020.585461 Frontiers in neuroscience Creative Commons Attribution 4.0 International license https://creativecommons.org/licenses/by/4.0/ application/pdf Frontiers Media SA Frontiers
spellingShingle Haro, Stephanie
Smalt, Christopher J.
Ciccarelli, Gregory A.
Quatieri, Thomas F.
Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title_full Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title_fullStr Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title_full_unstemmed Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title_short Deep Neural Network Model of Hearing-Impaired Speech-in-Noise Performance
title_sort deep neural network model of hearing impaired speech in noise performance
url https://hdl.handle.net/1721.1/129383
work_keys_str_mv AT harostephanie deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance
AT smaltchristopherj deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance
AT ciccarelligregorya deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance
AT quatierithomasf deepneuralnetworkmodelofhearingimpairedspeechinnoiseperformance