<p><p>Current language technology is dominated by approaches that either enumerate a large set of rules, or are focused on a large amount of manually labelled data. The creation of both is time-consuming and expensive, which is commonly thought to be the reason why automated natural language underst
Structure Discovery in Natural Language
β Scribed by Chris Biemann (auth.)
- Publisher
- Springer-Verlag Berlin Heidelberg
- Year
- 2012
- Tongue
- English
- Leaves
- 199
- Series
- Theory and Applications of Natural Language Processing
- Edition
- 1
- Category
- Library
No coin nor oath required. For personal study only.
β¦ Synopsis
Current language technology is dominated by approaches that either enumerate a large set of rules, or are focused on a large amount of manually labelled data. The creation of both is time-consuming and expensive, which is commonly thought to be the reason why automated natural language understanding has still not made its way into βreal-lifeβ applications yet.
This book sets an ambitious goal: to shift the development of language processing systems to a much more automated setting than previous works. A new approach is defined: what if computers analysed large samples of language data on their own, identifying structural regularities that perform the necessary abstractions and generalisations in order to better understand language in the process?
After defining the framework of Structure Discovery and shedding light on the nature and the graphic structure of natural language data, several procedures are described that do exactly this: let the computer discover structures without supervision in order to boost the performance of language technology applications. Here, multilingual documents are sorted by language, word classes are identified, and semantic ambiguities are discovered and resolved without using a dictionary or other explicit human input. The book concludes with an outlook on the possibilities implied by this paradigm and sets the methods in perspective to human computer interaction.
The target audience are academics on all levels (undergraduate and graduate students, lecturers and professors) working in the fields of natural language processing and computational linguistics, as well as natural language engineers who are seeking to improve their systems.
β¦ Table of Contents
Front Matter....Pages i-xx
Introduction....Pages 1-17
Graph Models....Pages 19-37
SmallWorlds of Natural Language....Pages 39-71
Graph Clustering....Pages 73-100
Unsupervised Language Separation....Pages 101-111
Unsupervised Part-of-Speech Tagging....Pages 113-144
Word Sense Induction and Disambiguation....Pages 145-155
Conclusion....Pages 157-160
Back Matter....Pages 161-178
β¦ Subjects
Artificial Intelligence (incl. Robotics); Computational Linguistics; Graph Theory
π SIMILAR VOLUMES
<p><p>Current language technology is dominated by approaches that either enumerate a large set of rules, or are focused on a large amount of manually labelled data. The creation of both is time-consuming and expensive, which is commonly thought to be the reason why automated natural language underst
<p><strong>`</strong>Her book deserves a reading by anyone interested in questions of phrase structure.<strong>'</strong><br/><strong>Language, 69</strong><br/><strong>`</strong>Speas' book is a unique and comprehensive presentation of how syntactic structure is projected from lexical items, and how
<p><span>Algebraic Structures in Natural Language</span><span> addresses a central problem in cognitive science concerning the learning procedures through which humans acquire and represent natural language. Until recently algebraic systems have dominated the study of natural language in formal and
In Coherence and Natural Language, Florian Wolf and Edward Gibson specify and evaluate criteria for descriptively adequate data structures for representing discourse coherence. They test the influence of discourse structure on pronoun processing, ev