As online information grows dramatically, search engines such as Google are playing a more and more important role in our lives. Critical to all search engines is the problem of designing an effective retrieval model that can rank documents accurately for a given query. This has been a central resea
Language Modeling for Information Retrieval
β Scribed by John Lafferty, ChengXiang Zhai (auth.), W. Bruce Croft, John Lafferty (eds.)
- Publisher
- Springer Netherlands
- Year
- 2003
- Tongue
- English
- Leaves
- 252
- Series
- The Springer International Series on Information Retrieval 13
- Edition
- 1
- Category
- Library
No coin nor oath required. For personal study only.
β¦ Synopsis
A statisticallanguage model, or more simply a language model, is a probΒ abilistic mechanism for generating text. Such adefinition is general enough to include an endless variety of schemes. However, a distinction should be made between generative models, which can in principle be used to synthesize artificial text, and discriminative techniques to classify text into predefined catΒ egories. The first statisticallanguage modeler was Claude Shannon. In exploring the application of his newly founded theory of information to human language, Shannon considered language as a statistical source, and measured how weH simple n-gram models predicted or, equivalently, compressed natural text. To do this, he estimated the entropy of English through experiments with human subjects, and also estimated the cross-entropy of the n-gram models on natural 1 text. The ability of language models to be quantitatively evaluated in tbis way is one of their important virtues. Of course, estimating the true entropy of language is an elusive goal, aiming at many moving targets, since language is so varied and evolves so quickly. Yet fifty years after Shannon's study, language models remain, by all measures, far from the Shannon entropy liInit in terms of their predictive power. However, tbis has not kept them from being useful for a variety of text processing tasks, and moreover can be viewed as encouragement that there is still great room for improvement in statisticallanguage modeling.
β¦ Table of Contents
Front Matter....Pages i-xiii
Probabilistic Relevance Models Based on Document and Query Generation....Pages 1-10
Relevance Models in Information Retrieval....Pages 11-56
Language Modeling and Relevance....Pages 57-71
Contributions of Language Modeling to the Theory and Practice of Information Retrieval....Pages 73-93
Language Models for Topic Tracking....Pages 95-123
A Probabilistic Approach to Term Translation for Cross-Lingual Retrieval....Pages 125-140
Using Compression-Based Language Models for Text Categorization....Pages 141-165
Applications of Score Distributions in Information Retrieval....Pages 167-188
An Unbiased Generative Model for Setting Dissemination Thresholds....Pages 189-217
Language Modeling Experiments in Non-Extractive Summarization....Pages 219-244
Back Matter....Pages 245-245
β¦ Subjects
Information Storage and Retrieval; Computer Science, general; Artificial Intelligence (incl. Robotics)
π SIMILAR VOLUMES
As online information grows dramatically, search engines such as Google are playing a more and more important role in our lives. Critical to all search engines is the problem of designing an effective retrieval model that can rank documents accurately for a given query. This has been a central resea
ΠΠ· ΡΠ΅ΡΠΈΠΈ Foundations and Trends in Information Retrieval ΠΈΠ·Π΄Π°ΡΠ΅Π»ΡΡΡΠ²Π° NOWPress, 2009, -77 pp.<div class="bb-sep"></div>Statistical language models have recently been successfully applied to many information retrieval problems. A great deal of recent work has shown that statistical language models no
<p>In recent years, there have been several attempts to define a logic for information retrieval (IR). The aim was to provide a rich and uniform representation of information and its semantics with the goal of improving retrieval effectiveness. The basis of a logical model for IR is the assumption t