Structure and Performance of a Dependency Language Model
1997
Citations Over Time
Ciprian Chelba, David M. Engle, Harry Printz, Frederick Jelinek, Eric Sven Ristad, Víctor Jiménez, Roni Rosenfeld, Sanjeev Khudanpur, Andreas Stolcke, Lidia Mangue, Dekai Wu
Abstract
Abstract : We present a maximum entropy language model that incorporates both syntax and semantics via a dependency grammar. Such a grammar expresses the relations between words by a directed graph. Because the edges of this graph may connect words that are arbitrarily far apart in a sentence, this technique can incorporate the predictive power of words that lie outside of bigram or trigram range. we have built several simple dependency models, as we call them, and tested them in a speech recognition experiment. We report experimental results for these models here, including one that has a small but statistically significant advantage (p .02) over a digram language model.
Related Papers
- → Class phrase models for language modeling(2002)45 cited
- → Speech recognition using a stochastic language model integrating local and global constraints(1994)9 cited
- → Decoupling Word-Pair Distance and Co-occurrence Information for Effective Long History Context Language Modeling(2015)2 cited
- Modeling of term-distance and term-occurrence information for improving n-gram language model performance(2013)
- → Improving language models by using distant information(2007)3 cited