Interactive spatial speech recognition maps based on simulated speech recognition experiments
In their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required fact...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
EDP Sciences
2022-01-01
|
Series: | Acta Acustica |
Subjects: | |
Online Access: | https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.html |
_version_ | 1797761012579958784 |
---|---|
author | Schädler Marc René |
author_facet | Schädler Marc René |
author_sort | Schädler Marc René |
collection | DOAJ |
description | In their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required factors simultaneously to predict the individual speech recognition performance in complex, that is, e.g. multi-source dynamic, acoustic environments. While such predictions may still not be sufficiently accurate for serious applications, such as, e.g. individual hearing aid fitting, they can already be performed. This raises an interesting question: What could we do if we had a perfect speech intelligibility model? In a first step, means to explore and interpret the predicted outcomes of large numbers of speech recognition experiments would be helpful, and large amounts of data demand an accessible, that is, easily comprehensible, representation. In this contribution, an interactive, that is, user manipulable, representation of speech recognition performance is proposed and investigated by means of a concrete example, which focuses on the listener’s head orientation and the spatial dimensions – in particular width and depth – of an acoustic scene. An exemplary modeling toolchain, that is, a combination of an acoustic model, a hearing device model, and a listener model, was used to generate a data set for demonstration purposes. Using the spatial speech recognition maps to explore this data set demonstrated the suitability of the approach to observe possibly relevant listener behavior. The proposed representation was found to be a suitable target to compare and validate modeling approaches in ecologically relevant contexts, and should help to explore possible applications of future speech recognition models. Ultimately, it may serve as a tool to use validated prediction models in the design of spaces and devices which take speech communication into account. |
first_indexed | 2024-03-12T19:06:51Z |
format | Article |
id | doaj.art-11f46b05b7d74cf3917f78eaeb1fbaab |
institution | Directory Open Access Journal |
issn | 2681-4617 |
language | English |
last_indexed | 2024-03-12T19:06:51Z |
publishDate | 2022-01-01 |
publisher | EDP Sciences |
record_format | Article |
series | Acta Acustica |
spelling | doaj.art-11f46b05b7d74cf3917f78eaeb1fbaab2023-08-02T06:07:02ZengEDP SciencesActa Acustica2681-46172022-01-0163110.1051/aacus/2022028aacus210031Interactive spatial speech recognition maps based on simulated speech recognition experimentsSchädler Marc René0https://orcid.org/0000-0001-6731-7377Medizinische Physik and Cluster of Excellence Hearing4all, Universität OldenburgIn their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required factors simultaneously to predict the individual speech recognition performance in complex, that is, e.g. multi-source dynamic, acoustic environments. While such predictions may still not be sufficiently accurate for serious applications, such as, e.g. individual hearing aid fitting, they can already be performed. This raises an interesting question: What could we do if we had a perfect speech intelligibility model? In a first step, means to explore and interpret the predicted outcomes of large numbers of speech recognition experiments would be helpful, and large amounts of data demand an accessible, that is, easily comprehensible, representation. In this contribution, an interactive, that is, user manipulable, representation of speech recognition performance is proposed and investigated by means of a concrete example, which focuses on the listener’s head orientation and the spatial dimensions – in particular width and depth – of an acoustic scene. An exemplary modeling toolchain, that is, a combination of an acoustic model, a hearing device model, and a listener model, was used to generate a data set for demonstration purposes. Using the spatial speech recognition maps to explore this data set demonstrated the suitability of the approach to observe possibly relevant listener behavior. The proposed representation was found to be a suitable target to compare and validate modeling approaches in ecologically relevant contexts, and should help to explore possible applications of future speech recognition models. Ultimately, it may serve as a tool to use validated prediction models in the design of spaces and devices which take speech communication into account.https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.htmlinteractive speech recognition mapspeech perception modelspeech recognition performancecomplex acoustic scenespeech maskingspeech in noiseimpaired hearinghearing loss compensation |
spellingShingle | Schädler Marc René Interactive spatial speech recognition maps based on simulated speech recognition experiments Acta Acustica interactive speech recognition map speech perception model speech recognition performance complex acoustic scene speech masking speech in noise impaired hearing hearing loss compensation |
title | Interactive spatial speech recognition maps based on simulated speech recognition experiments |
title_full | Interactive spatial speech recognition maps based on simulated speech recognition experiments |
title_fullStr | Interactive spatial speech recognition maps based on simulated speech recognition experiments |
title_full_unstemmed | Interactive spatial speech recognition maps based on simulated speech recognition experiments |
title_short | Interactive spatial speech recognition maps based on simulated speech recognition experiments |
title_sort | interactive spatial speech recognition maps based on simulated speech recognition experiments |
topic | interactive speech recognition map speech perception model speech recognition performance complex acoustic scene speech masking speech in noise impaired hearing hearing loss compensation |
url | https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.html |
work_keys_str_mv | AT schadlermarcrene interactivespatialspeechrecognitionmapsbasedonsimulatedspeechrecognitionexperiments |