Interactive spatial speech recognition maps based on simulated speech recognition experiments

In their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required fact...

Full description

Bibliographic Details
Main Author: Schädler Marc René
Format: Article
Language:English
Published: EDP Sciences 2022-01-01
Series:Acta Acustica
Subjects:
Online Access:https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.html
_version_ 1797761012579958784
author Schädler Marc René
author_facet Schädler Marc René
author_sort Schädler Marc René
collection DOAJ
description In their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required factors simultaneously to predict the individual speech recognition performance in complex, that is, e.g. multi-source dynamic, acoustic environments. While such predictions may still not be sufficiently accurate for serious applications, such as, e.g. individual hearing aid fitting, they can already be performed. This raises an interesting question: What could we do if we had a perfect speech intelligibility model? In a first step, means to explore and interpret the predicted outcomes of large numbers of speech recognition experiments would be helpful, and large amounts of data demand an accessible, that is, easily comprehensible, representation. In this contribution, an interactive, that is, user manipulable, representation of speech recognition performance is proposed and investigated by means of a concrete example, which focuses on the listener’s head orientation and the spatial dimensions – in particular width and depth – of an acoustic scene. An exemplary modeling toolchain, that is, a combination of an acoustic model, a hearing device model, and a listener model, was used to generate a data set for demonstration purposes. Using the spatial speech recognition maps to explore this data set demonstrated the suitability of the approach to observe possibly relevant listener behavior. The proposed representation was found to be a suitable target to compare and validate modeling approaches in ecologically relevant contexts, and should help to explore possible applications of future speech recognition models. Ultimately, it may serve as a tool to use validated prediction models in the design of spaces and devices which take speech communication into account.
first_indexed 2024-03-12T19:06:51Z
format Article
id doaj.art-11f46b05b7d74cf3917f78eaeb1fbaab
institution Directory Open Access Journal
issn 2681-4617
language English
last_indexed 2024-03-12T19:06:51Z
publishDate 2022-01-01
publisher EDP Sciences
record_format Article
series Acta Acustica
spelling doaj.art-11f46b05b7d74cf3917f78eaeb1fbaab2023-08-02T06:07:02ZengEDP SciencesActa Acustica2681-46172022-01-0163110.1051/aacus/2022028aacus210031Interactive spatial speech recognition maps based on simulated speech recognition experimentsSchädler Marc René0https://orcid.org/0000-0001-6731-7377Medizinische Physik and Cluster of Excellence Hearing4all, Universität OldenburgIn their everyday life, the speech recognition performance of human listeners is influenced by diverse factors, such as the acoustic environment, the talker and listener positions, possibly impaired hearing, and optional hearing devices. Prediction models come closer to considering all required factors simultaneously to predict the individual speech recognition performance in complex, that is, e.g. multi-source dynamic, acoustic environments. While such predictions may still not be sufficiently accurate for serious applications, such as, e.g. individual hearing aid fitting, they can already be performed. This raises an interesting question: What could we do if we had a perfect speech intelligibility model? In a first step, means to explore and interpret the predicted outcomes of large numbers of speech recognition experiments would be helpful, and large amounts of data demand an accessible, that is, easily comprehensible, representation. In this contribution, an interactive, that is, user manipulable, representation of speech recognition performance is proposed and investigated by means of a concrete example, which focuses on the listener’s head orientation and the spatial dimensions – in particular width and depth – of an acoustic scene. An exemplary modeling toolchain, that is, a combination of an acoustic model, a hearing device model, and a listener model, was used to generate a data set for demonstration purposes. Using the spatial speech recognition maps to explore this data set demonstrated the suitability of the approach to observe possibly relevant listener behavior. The proposed representation was found to be a suitable target to compare and validate modeling approaches in ecologically relevant contexts, and should help to explore possible applications of future speech recognition models. Ultimately, it may serve as a tool to use validated prediction models in the design of spaces and devices which take speech communication into account.https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.htmlinteractive speech recognition mapspeech perception modelspeech recognition performancecomplex acoustic scenespeech maskingspeech in noiseimpaired hearinghearing loss compensation
spellingShingle Schädler Marc René
Interactive spatial speech recognition maps based on simulated speech recognition experiments
Acta Acustica
interactive speech recognition map
speech perception model
speech recognition performance
complex acoustic scene
speech masking
speech in noise
impaired hearing
hearing loss compensation
title Interactive spatial speech recognition maps based on simulated speech recognition experiments
title_full Interactive spatial speech recognition maps based on simulated speech recognition experiments
title_fullStr Interactive spatial speech recognition maps based on simulated speech recognition experiments
title_full_unstemmed Interactive spatial speech recognition maps based on simulated speech recognition experiments
title_short Interactive spatial speech recognition maps based on simulated speech recognition experiments
title_sort interactive spatial speech recognition maps based on simulated speech recognition experiments
topic interactive speech recognition map
speech perception model
speech recognition performance
complex acoustic scene
speech masking
speech in noise
impaired hearing
hearing loss compensation
url https://acta-acustica.edpsciences.org/articles/aacus/full_html/2022/01/aacus210031/aacus210031.html
work_keys_str_mv AT schadlermarcrene interactivespatialspeechrecognitionmapsbasedonsimulatedspeechrecognitionexperiments