
Statistical Language Learning
List price: US$25.00
Currently unavailable
We can notify you when this item is back in stock
Add to wishlistAbeBooks may have this title (opens in new window).
Try AbeBooksDescription
Eugene Charniak breaks new ground in artificial intelligence research by presenting statistical language processing from an artificial intelligence point of view in a text for researchers and scientists with a traditional computer science background. New, exacting empirical methods are needed to break the deadlock in such areas of artificial intelligence as robotics, knowledge representation, machine learning, machine translation, and natural language processing (NLP). It is time, Charniak observes, to switch paradigms. This text introduces statistical language processing techniques; word tagging, parsing with probabilistic context free grammars, grammar induction, syntactic disambiguation, semantic wordclasses, word-sense disambiguation; along with the underlying mathematics and chapter exercises. Charniak points out that as a method of attacking NLP problems, the statistical approach has several advantages. It is grounded in real text and therefore promises to produce usable results, and it offers an obvious way to approach learning: "one simply gathers statistics."
show more
show more
Product details
- Paperback | 190 pages
- 150 x 224 x 10mm | 318g
- 01 Sep 1996
- MIT Press Ltd
- Bradford Books
- Massachusetts, United States
- English
- Revised ed.
- 0262531410
- 9780262531412
- 2,124,080
Table of contents
Part 1 The Standard Model: Two Technologies; Morphology and Knowledge of Words; Syntax and Context-Free Grammars; Chart Parsing; Meaning and Semantic Processing; Exercises. Part 2 Statistical Models and the Entropy of English: A Fragment of Probability Theory; Statistical Models; Speech Recognition; Entropy; Markov Chains; Cross Entropy; Cross Entropy as a Model Evaluator; Exercises. Part 3 Hidden Markov Models and Two Applications: Trigram Models of English; Hidden Markov Models; Part-of-Speech Tagging; Exercises. Part 4 Algorithms for Hidden Markov Models: Finding the Most Likely Path; Computing HMM Output Probabilities; HMM Training; Exercises. Part 5 Probabilistic Context-Free Grammars: Probabilistic Grammars; PCFGs and Syntactic Ambiguity; PCFGs and Grammar Induction; PCFGs and Ungrammaticality; PCFGs and Language Modelling; Basic Algorithms for PCFGs; Exercises. Part 6 The Mathematics of PCFGs: Relation of HMMs to PCFGs; Finding Sentence Probabilities for PCFGs; Training PCFGs; Exercises. Part 7 Learning Probabilistic Grammars: Why the Simple Approach Fails; Learning Dependency Grammars; Learning from a Bracketed Corpus; Improving a Partial Grammar; Exercises. Part 8 Syntactic Disambiguation: Simple Methods for Prepositional Phrases; Using Semantic Information; Relative-Clause Attachment; Uniform Use of Lexical/Semantic Information; Exercises. Part 9 Word Classes and Meaning: Clustering; Clustering by Next Word; Clustering with Syntactic Information; Problems with Word Clustering; Exercises. Part 10 Word Senses and Their Disambiguation: Word Senses Using Outside Information; Word Senses Without Outside Information; Meanings and Selectional Restrictions; Discussion; Exercises.
show more
show more
Review quote
"This is a lovely book." -- David Nye
show more
show more
About Eugene Charniak
Eugene Charniak is Professor of Computer Science at Brown University. He is the author of Statistical Language Learning (MIT Press) and other books.
show more
show more