Deriving phrase-based language models
2002pp. 41–48
Citations Over TimeTop 16% of 2002 papers
Abstract
Phrase-based language models have grown in popularity since they allow the speech recognition process to make use of more context in recognizing the words. Previous approaches have used perplexity reduction to identify groups of words to be linked into phrases and have used these phrases as the basis for computing the language model probabilities. In this paper, we argue that perplexity reduction is only one of three aspects to be considered in choosing the phrases. We also argue that the chosen phrases should not be the basis for computing the language model probabilities. Rather, the probabilities should be derived from a language model built at the lexical level.
Related Papers
- Combination of Recurrent Neural Networks and Factored Language Models for Code-Switching Language Modeling(2013)
- → Improved topic-dependent language modeling using information retrieval techniques(1999)55 cited
- → Verifying the long-range dependency of RNN language models(2016)2 cited
- → Attention-based Persian Language Modeling(2022)
- → Building Personalized Language Models Through Language Model Interpolation(2023)