𝔖 Bobbio Scriptorium
✦   LIBER   ✦

Variable-length categoryn-gram language models

✍ Scribed by T.R. Niesler; P.C. Woodland


Publisher
Elsevier Science
Year
1999
Tongue
English
Weight
253 KB
Volume
13
Category
Article
ISSN
0885-2308

No coin nor oath required. For personal study only.

✦ Synopsis


This paper presents a language model based on n-grams of word groups (categories). The length of each n-gram is increased selectively according to an estimate of the resulting improvement in predictive quality. This allows the model size to be controlled while including longer-range dependencies when these benefit performance. The categories are chosen to correspond to part-of-speech classifications in a bid to exploit a priori grammatical information. To account for different grammatical functions, the language model allows words to belong to multiple categories, and implicitly involves a statistical tagging operation which may be used to label new text. Intrinsic generalization by the category-based model leads to good performance with sparse data sets. However word-based n-grams deliver superior average performance as the amount of training material increases. Nevertheless, the category model continues to supply better predictions for word n-tuples not present in the training set. Consequently, a method allowing the two approaches to be combined within a backoff framework is presented. Experiments with the LOB, Switchboard and Wall Street Journal corpora demonstrate that this technique greatly improves language model perplexities for sparse training sets, and offers significantly improved size vs. performance tradeoffs when compared with standard trigram models.


πŸ“œ SIMILAR VOLUMES


Language modeling using stochastic autom
✍ Jianying Hu; William Turin; Michael K. Brown πŸ“‚ Article πŸ“… 1997 πŸ› Elsevier Science 🌐 English βš– 299 KB

It is well known that language models are effective for increasing the accuracy of speech and handwriting recognizers, but large language models are often required to achieve low model perplexity (or entropy) and still have adequate language coverage. We study three efficient methods for variable or

Learning Variable-Length Markov Models o
✍ Aphrodite Galata; Neil Johnson; David Hogg πŸ“‚ Article πŸ“… 2001 πŸ› Elsevier Science 🌐 English βš– 264 KB

In recent years there has been an increased interest in the modeling and recognition of human activities involving highly structured and semantically rich behavior such as dance, aerobics, and sign language. A novel approach for automatically acquiring stochastic models of the high-level structure o

A weighted average n-gram model of natur
✍ P. O'Boyle; M. Owens; F.J. Smith πŸ“‚ Article πŸ“… 1994 πŸ› Elsevier Science 🌐 English βš– 403 KB

A new \(n\)-gram model of natural language designed to aid speech recognition is presented in which the probabilities are calculated as a weighted average of maximum likelihood probabilities obtained from a training corpus. This simple approach produces a model that can be constructed quickly and is

Multiclass composite N-gram language mod
✍ Hirofumi Yamamoto; Yoshinori Sagisaka πŸ“‚ Article πŸ“… 2003 πŸ› John Wiley and Sons 🌐 English βš– 530 KB

## Abstract The authors propose a method to generate a compact, highly reliable language model for speech recognition based on the efficient classification of words. In this method, the connectedness with the words immediately before and after the word is taken to represent separate attributes, and