Improving N-Best Rescoring in Under-Resourced Code-Switched Speech Recognition Using Pretraining and Data Augmentation

In this study, we present improvements in N-best rescoring of code-switched speech achieved by n-gram augmentation as well as optimised pretraining of long short-term memory (LSTM) language models with larger corpora of out-of-domain monolingual text. Our investigation specifically considers the imp...

Full description

Bibliographic Details
Main Authors: Joshua Jansen van Vüren, Thomas Niesler
Format: Article
Language:English
Published: MDPI AG 2022-09-01
Series:Languages
Subjects:
Online Access:https://www.mdpi.com/2226-471X/7/3/236