Context-dependent factored language modelsGregor Donaj
, Zdravko Kačič
, 2017, original scientific article
Abstract: The incorporation of grammatical information into speech recognition systems is often used to increase performance in morphologically rich languages. However, this introduces demands for sufficiently large training corpora and proper methods of using the additional information. In this paper, we present a method for building factored language models that use data obtained by morphosyntactic tagging. The models use only relevant factors that help to increase performance and ignore data from other factors, thus also reducing the need for large morphosyntactically tagged training corpora. Which data is relevant is determined at run-time, based on the current text segment being estimated, i.e., the context. We show that using a context-dependent model in a two-pass recognition algorithm, the overall speech recognition accuracy in a Broadcast News application improved by 1.73% relatively, while simpler models using the same data achieved only 0.07% improvement. We also present a more detailed error analysis based on lexical features, comparing first-pass and second-pass results.
Keywords: speech recognition, factored language model, dynamic backoff path, word context, inflectional language, morphosyntactic tags
Published: 26.06.2017; Views: 596; Downloads: 209
Full text (1,17 MB)
This document has many files! More...