Natural Language Processing
with Python
Steven Bird, Ewan Klein, and Edward Loper
Beijing • Cambridge • Farnham • Köln • S...
Natural Language Processing with Python
by Steven Bird, Ewan Klein, and Edward Loper
Copyright © 2009 Steven Bird, Ewan Kl...
Table of Contents
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ...
3.10 Summary 121
3.11 Further Reading 122
3.12 Exercises 123
4. Writing Structured Programs . . . . . . . . . . . . . . . ...
7.2 Chunking 264
7.3 Developing and Evaluating Chunkers 270
7.4 Recursion in Linguistic Structure 277
7.5 Named Entity Rec...
11.5 Working with Toolbox Data 431
11.6 Describing Language Resources Using OLAC Metadata 435
11.7 Summary 437
11.8 Furthe...
Preface
This is a book about Natural Language Processing. By “natural language” we mean a
language that is used for everyd...
Audience
NLP is important for scientific, economic, social, and cultural reasons. NLP is experi-
encing rapid growth as it...
Note that this book is not a reference work. Its coverage of Python and NLP is selective,
and presented in a tutorial styl...
ways to parse a sentence, recognize its syntactic structure, and construct representa-
tions of meaning (Chapters 8–10). T...
we can use to break a line into its words. To apply a method to an object, we write the
object name, followed by a period,...
NLTK-Data
This contains the linguistic corpora that are analyzed and processed in the book.
NumPy (recommended)
This is a ...
Language processing task NLTK modules Functionality
Linguistic fieldwork nltk.toolbox Manipulate data in SIL Toolbox forma...
A significant fraction of any NLP syllabus deals with algorithms and data structures.
On their own these can be rather dry...
Chapter Arts and Humanities Science and Engineering
Chapter 5, CategorizingandTaggingWords 2–4 2–4
Chapter 6, LearningtoCl...
writing a program that uses several chunks of code from this book does not require
permission. Selling or distributing a C...
The authors provide additional materials for each chapter via the NLTK website at:
http://www.nltk.org/
To comment or ask ...
Figure P-1. Edward Loper, Ewan Klein, and Steven Bird, Stanford, July 2007
xx | Preface
CHAPTER 1
Language Processing and Python
It is easy to get our hands on millions of words of text. What can we do with it,...
Getting Started with Python
One of the friendly things about Python is that it allows you to type directly into the
intera...
>>> 1 +
File "<stdin>", line 1
1 +
^
SyntaxError: invalid syntax
>>>
This produced a syntax error. In Python, it doesn’t m...
tells the interpreter to load some texts for us to explore: from nltk.book import *. This
says “from NLTK’s book module, l...
ght have been rummaged out of this monstrous cabinet there is no telling . But
of Whale - Bones ; for Whales of a monstrou...
Your Turn: Pick another pair of words and compare their usage in two
different texts, using the similar() and common_conte...
>>> text3.generate()
In the beginning of his brother is a hairy man , whose top may reach
unto heaven ; and ye shall sow t...
items of text3 with the command: set(text3). When you do this, many screens of
words will fly past. Now try the following:...
called a function, and we define a short name for our function with the keyword def.
The next example shows how to define ...
a sense of the power and creativity of programming. Don’t worry if you find it a bit
confusing right now.
Later we’ll see ...
Some more lists have been defined for you, one for the opening sentence of each of our
texts, sent2 … sent9. We inspect tw...
Indexing Lists
As we have seen, a text in Python is a list of words, represented using a combination
of brackets and quote...
This practice of counting from zero is initially confusing, but typical of
modern programming languages. You’ll quickly ge...
>>> sent[0] = 'First'
>>> sent[9] = 'Last'
>>> len(sent)
10
>>> sent[1:9] = ['Second', 'Third']
>>> sent
['First', 'Second...
Notice in the previous example that we split the definition of my_sent
over two lines. Python expressions can be split acr...
>>> name = 'Monty'
>>> name[0]
'M'
>>> name[:4]
'Mont'
>>>
We can also perform multiplication and addition with strings:
>...
Frequency Distributions
How can we automatically identify the words of a text that are most informative about
the topic an...
Your Turn: Try the preceding frequency distribution example for your-
self, for text2. Be careful to use the correct paren...
If the frequent words don’t help us, how about the words that occur once only, the so-
called hapaxes? View them by typing...
Let’s return to our task of finding words that characterize a text. Notice that the long
words in text4 reflect its nation...
National Government; United Nations; public money
>>> text8.collocations()
Building collocations list
medium build; social...
differences between authors, genres, or languages. Table 1-2 summarizes the functions
defined in frequency distributions.
...
Operator Relationship
!= Not equal to
> Greater than
>= Greater than or equal to
We can use these to select different word...
>>> sorted([w for w in set(text1) if w.endswith('ableness')])
['comfortableness', 'honourableness', 'immutableness', 'indi...
>>> len(set(text1))
19317
>>> len(set([word.lower() for word in text1]))
17231
>>>
Now that we are not double-counting wor...
An if statement is known as a control structure because it controls whether the code
in the indented block will be run. An...
Ishmael is a titlecase word
. is punctuation
>>>
As you can see, even with this small amount of Python knowledge, you can ...
In this section we describe some language understanding technologies, to give you a
sense of the interesting challenges th...
semantic role labeling—identifying how a noun phrase relates to the verb (as agent,
patient, instrument, and so on).
Gener...
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
Natural language processing with python (2009)
of 491

Natural language processing with python (2009)

Published on: Mar 3, 2016
Source: www.slideshare.net


Transcripts - Natural language processing with python (2009)

  • 1. Natural Language Processing with Python Steven Bird, Ewan Klein, and Edward Loper Beijing • Cambridge • Farnham • Köln • Sebastopol • Taipei • Tokyo
  • 2. Natural Language Processing with Python by Steven Bird, Ewan Klein, and Edward Loper Copyright © 2009 Steven Bird, Ewan Klein, and Edward Loper. All rights reserved. Printed in the United States of America. Published by O’Reilly Media, Inc., 1005 Gravenstein Highway North, Sebastopol, CA 95472. O’Reilly books may be purchased for educational, business, or sales promotional use. Online editions are also available for most titles (http://my.safaribooksonline.com). For more information, contact our corporate/institutional sales department: (800) 998-9938 or corporate@oreilly.com. Editor: Julie Steele Production Editor: Loranah Dimant Copyeditor: Genevieve d’Entremont Proofreader: Loranah Dimant Indexer: Ellen Troutman Zaig Cover Designer: Karen Montgomery Interior Designer: David Futato Illustrator: Robert Romano Printing History: June 2009: First Edition. Nutshell Handbook, the Nutshell Handbook logo, and the O’Reilly logo are registered trademarks of O’Reilly Media, Inc. Natural Language Processing with Python, the image of a right whale, and related trade dress are trademarks of O’Reilly Media, Inc. Many of the designations used by manufacturers and sellers to distinguish their products are claimed as trademarks. Where those designations appear in this book, and O’Reilly Media, Inc. was aware of a trademark claim, the designations have been printed in caps or initial caps. While every precaution has been taken in the preparation of this book, the publisher and authors assume no responsibility for errors or omissions, or for damages resulting from the use of the information con- tained herein. ISBN: 978-0-596-51649-9 [M] 1244726609
  • 3. Table of Contents Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ix 1. Language Processing and Python . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1 1.1 Computing with Language: Texts and Words 1 1.2 A Closer Look at Python: Texts as Lists of Words 10 1.3 Computing with Language: Simple Statistics 16 1.4 Back to Python: Making Decisions and Taking Control 22 1.5 Automatic Natural Language Understanding 27 1.6 Summary 33 1.7 Further Reading 34 1.8 Exercises 35 2. Accessing Text Corpora and Lexical Resources . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39 2.1 Accessing Text Corpora 39 2.2 Conditional Frequency Distributions 52 2.3 More Python: Reusing Code 56 2.4 Lexical Resources 59 2.5 WordNet 67 2.6 Summary 73 2.7 Further Reading 73 2.8 Exercises 74 3. Processing Raw Text . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79 3.1 Accessing Text from the Web and from Disk 80 3.2 Strings: Text Processing at the Lowest Level 87 3.3 Text Processing with Unicode 93 3.4 Regular Expressions for Detecting Word Patterns 97 3.5 Useful Applications of Regular Expressions 102 3.6 Normalizing Text 107 3.7 Regular Expressions for Tokenizing Text 109 3.8 Segmentation 112 3.9 Formatting: From Lists to Strings 116 v
  • 4. 3.10 Summary 121 3.11 Further Reading 122 3.12 Exercises 123 4. Writing Structured Programs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129 4.1 Back to the Basics 130 4.2 Sequences 133 4.3 Questions of Style 138 4.4 Functions: The Foundation of Structured Programming 142 4.5 Doing More with Functions 149 4.6 Program Development 154 4.7 Algorithm Design 160 4.8 A Sample of Python Libraries 167 4.9 Summary 172 4.10 Further Reading 173 4.11 Exercises 173 5. Categorizing and Tagging Words . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179 5.1 Using a Tagger 179 5.2 Tagged Corpora 181 5.3 Mapping Words to Properties Using Python Dictionaries 189 5.4 Automatic Tagging 198 5.5 N-Gram Tagging 202 5.6 Transformation-Based Tagging 208 5.7 How to Determine the Category of a Word 210 5.8 Summary 213 5.9 Further Reading 214 5.10 Exercises 215 6. Learning to Classify Text . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 221 6.1 Supervised Classification 221 6.2 Further Examples of Supervised Classification 233 6.3 Evaluation 237 6.4 Decision Trees 242 6.5 Naive Bayes Classifiers 245 6.6 Maximum Entropy Classifiers 250 6.7 Modeling Linguistic Patterns 254 6.8 Summary 256 6.9 Further Reading 256 6.10 Exercises 257 7. Extracting Information from Text . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 261 7.1 Information Extraction 261 vi | Table of Contents
  • 5. 7.2 Chunking 264 7.3 Developing and Evaluating Chunkers 270 7.4 Recursion in Linguistic Structure 277 7.5 Named Entity Recognition 281 7.6 Relation Extraction 284 7.7 Summary 285 7.8 Further Reading 286 7.9 Exercises 286 8. Analyzing Sentence Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291 8.1 Some Grammatical Dilemmas 292 8.2 What’s the Use of Syntax? 295 8.3 Context-Free Grammar 298 8.4 Parsing with Context-Free Grammar 302 8.5 Dependencies and Dependency Grammar 310 8.6 Grammar Development 315 8.7 Summary 321 8.8 Further Reading 322 8.9 Exercises 322 9. Building Feature-Based Grammars . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327 9.1 Grammatical Features 327 9.2 Processing Feature Structures 337 9.3 Extending a Feature-Based Grammar 344 9.4 Summary 356 9.5 Further Reading 357 9.6 Exercises 358 10. Analyzing the Meaning of Sentences . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361 10.1 Natural Language Understanding 361 10.2 Propositional Logic 368 10.3 First-Order Logic 372 10.4 The Semantics of English Sentences 385 10.5 Discourse Semantics 397 10.6 Summary 402 10.7 Further Reading 403 10.8 Exercises 404 11. Managing Linguistic Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 407 11.1 Corpus Structure: A Case Study 407 11.2 The Life Cycle of a Corpus 412 11.3 Acquiring Data 416 11.4 Working with XML 425 Table of Contents | vii
  • 6. 11.5 Working with Toolbox Data 431 11.6 Describing Language Resources Using OLAC Metadata 435 11.7 Summary 437 11.8 Further Reading 437 11.9 Exercises 438 Afterword: The Language Challenge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 441 Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 449 NLTK Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 459 General Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 463 viii | Table of Contents
  • 7. Preface This is a book about Natural Language Processing. By “natural language” we mean a language that is used for everyday communication by humans; languages such as Eng- lish, Hindi, or Portuguese. In contrast to artificial languages such as programming lan- guages and mathematical notations, natural languages have evolved as they pass from generation to generation, and are hard to pin down with explicit rules. We will take Natural Language Processing—or NLP for short—in a wide sense to cover any kind of computer manipulation of natural language. At one extreme, it could be as simple as counting word frequencies to compare different writing styles. At the other extreme, NLP involves “understanding” complete human utterances, at least to the extent of being able to give useful responses to them. Technologies based on NLP are becoming increasingly widespread. For example, phones and handheld computers support predictive text and handwriting recognition; web search engines give access to information locked up in unstructured text; machine translation allows us to retrieve texts written in Chinese and read them in Spanish. By providing more natural human-machine interfaces, and more sophisticated access to stored information, language processing has come to play a central role in the multi- lingual information society. This book provides a highly accessible introduction to the field of NLP. It can be used for individual study or as the textbook for a course on natural language processing or computational linguistics, or as a supplement to courses in artificial intelligence, text mining, or corpus linguistics. The book is intensely practical, containing hundreds of fully worked examples and graded exercises. The book is based on the Python programming language together with an open source library called the Natural Language Toolkit (NLTK). NLTK includes extensive soft- ware, data, and documentation, all freely downloadable from http://www.nltk.org/. Distributions are provided for Windows, Macintosh, and Unix platforms. We strongly encourage you to download Python and NLTK, and try out the examples and exercises along the way. ix
  • 8. Audience NLP is important for scientific, economic, social, and cultural reasons. NLP is experi- encing rapid growth as its theories and methods are deployed in a variety of new lan- guage technologies. For this reason it is important for a wide range of people to have a working knowledge of NLP. Within industry, this includes people in human-computer interaction, business information analysis, and web software development. Within academia, it includes people in areas from humanities computing and corpus linguistics through to computer science and artificial intelligence. (To many people in academia, NLP is known by the name of “Computational Linguistics.”) This book is intended for a diverse range of people who want to learn how to write programs that analyze written language, regardless of previous programming experience: New to programming? The early chapters of the book are suitable for readers with no prior knowledge of programming, so long as you aren’t afraid to tackle new concepts and develop new computing skills. The book is full of examples that you can copy and try for your- self, together with hundreds of graded exercises. If you need a more general intro- duction to Python, see the list of Python resources at http://docs.python.org/. New to Python? Experienced programmers can quickly learn enough Python using this book to get immersed in natural language processing. All relevant Python features are carefully explained and exemplified, and you will quickly come to appreciate Python’s suit- ability for this application area. The language index will help you locate relevant discussions in the book. Already dreaming in Python? Skim the Python examples and dig into the interesting language analysis material that starts in Chapter 1. You’ll soon be applying your skills to this fascinating domain. Emphasis This book is a practical introduction to NLP. You will learn by example, write real programs, and grasp the value of being able to test an idea through implementation. If you haven’t learned already, this book will teach you programming. Unlike other programming books, we provide extensive illustrations and exercises from NLP. The approach we have taken is also principled, in that we cover the theoretical underpin- nings and don’t shy away from careful linguistic and computational analysis. We have tried to be pragmatic in striking a balance between theory and application, identifying the connections and the tensions. Finally, we recognize that you won’t get through this unless it is also pleasurable, so we have tried to include many applications and ex- amples that are interesting and entertaining, and sometimes whimsical. x | Preface
  • 9. Note that this book is not a reference work. Its coverage of Python and NLP is selective, and presented in a tutorial style. For reference material, please consult the substantial quantity of searchable resources available at http://python.org/ and http://www.nltk .org/. This book is not an advanced computer science text. The content ranges from intro- ductory to intermediate, and is directed at readers who want to learn how to analyze text using Python and the Natural Language Toolkit. To learn about advanced algo- rithms implemented in NLTK, you can examine the Python code linked from http:// www.nltk.org/, and consult the other materials cited in this book. What You Will Learn By digging into the material presented here, you will learn: • How simple programs can help you manipulate and analyze language data, and how to write these programs • How key concepts from NLP and linguistics are used to describe and analyze language • How data structures and algorithms are used in NLP • How language data is stored in standard formats, and how data can be used to evaluate the performance of NLP techniques Depending on your background, and your motivation for being interested in NLP, you will gain different kinds of skills and knowledge from this book, as set out in Table P-1. Table P-1. Skills and knowledge to be gained from reading this book, depending on readers’ goals and background Goals Background in arts and humanities Background in science and engineering Language analysis Manipulating large corpora, exploring linguistic models, and testing empirical claims. Using techniques in data modeling, data mining, and knowledge discovery to analyze natural language. Language technology Buildingrobustsystemstoperformlinguistictasks with technological applications. Usinglinguisticalgorithmsanddatastructuresinrobust language processing software. Organization The early chapters are organized in order of conceptual difficulty, starting with a prac- tical introduction to language processing that shows how to explore interesting bodies of text using tiny Python programs (Chapters 1–3). This is followed by a chapter on structured programming (Chapter 4) that consolidates the programming topics scat- tered across the preceding chapters. After this, the pace picks up, and we move on to a series of chapters covering fundamental topics in language processing: tagging, clas- sification, and information extraction (Chapters 5–7). The next three chapters look at Preface | xi
  • 10. ways to parse a sentence, recognize its syntactic structure, and construct representa- tions of meaning (Chapters 8–10). The final chapter is devoted to linguistic data and how it can be managed effectively (Chapter 11). The book concludes with an After- word, briefly discussing the past and future of the field. Within each chapter, we switch between different styles of presentation. In one style, natural language is the driver. We analyze language, explore linguistic concepts, and use programming examples to support the discussion. We often employ Python con- structsthathavenotbeenintroducedsystematically,soyoucanseetheirpurposebefore delving into the details of how and why they work. This is just like learning idiomatic expressions in a foreign language: you’re able to buy a nice pastry without first having learned the intricacies of question formation. In the other style of presentation, the programming language will be the driver. We’ll analyze programs, explore algorithms, and the linguistic examples will play a supporting role. Each chapter ends with a series of graded exercises, which are useful for consolidating the material. The exercises are graded according to the following scheme: ○ is for easy exercises that involve minor modifications to supplied code samples or other simple activities; ◑ is for intermediate exercises that explore an aspect of the material in more depth, requiring careful analysis and design; ● is for difficult, open-ended tasks that will challenge your understanding of the material and force you to think independently (readers new to programming should skip these). Each chapter has a further reading section and an online “extras” section at http://www .nltk.org/, with pointers to more advanced materials and online resources. Online ver- sions of all the code examples are also available there. Why Python? Python is a simple yet powerful programming language with excellent functionality for processing linguistic data. Python can be downloaded for free from http://www.python .org/. Installers are available for all platforms. Here is a five-line Python program that processes file.txt and prints all the words ending in ing: >>> for line in open("file.txt"): ... for word in line.split(): ... if word.endswith('ing'): ... print word This program illustrates some of the main features of Python. First, whitespace is used to nest lines of code; thus the line starting with if falls inside the scope of the previous line starting with for; this ensures that the ing test is performed for each word. Second, Python is object-oriented; each variable is an entity that has certain defined attributes and methods. For example, the value of the variable line is more than a sequence of characters. It is a string object that has a “method” (or operation) called split() that xii | Preface
  • 11. we can use to break a line into its words. To apply a method to an object, we write the object name, followed by a period, followed by the method name, i.e., line.split(). Third, methods have arguments expressed inside parentheses. For instance, in the ex- ample, word.endswith('ing') had the argument 'ing' to indicate that we wanted words ending with ing and not something else. Finally—and most importantly—Python is highly readable, so much so that it is fairly easy to guess what this program does even if you have never written a program before. We chose Python because it has a shallow learning curve, its syntax and semantics are transparent, and it has good string-handling functionality. As an interpreted language, Python facilitates interactive exploration. As an object-oriented language, Python per- mits data and methods to be encapsulated and re-used easily. As a dynamic language, Python permits attributes to be added to objects on the fly, and permits variables to be typed dynamically, facilitating rapid development. Python comes with an extensive standard library, including components for graphical programming, numerical pro- cessing, and web connectivity. Python is heavily used in industry, scientific research, and education around the world. Python is often praised for the way it facilitates productivity, quality, and main- tainability of software. A collection of Python success stories is posted at http://www .python.org/about/success/. NLTK defines an infrastructure that can be used to build NLP programs in Python. It provides basic classes for representing data relevant to natural language processing; standard interfaces for performing tasks such as part-of-speech tagging, syntactic pars- ing, and text classification; and standard implementations for each task that can be combined to solve complex problems. NLTK comes with extensive documentation. In addition to this book, the website at http://www.nltk.org/ provides API documentation that covers every module, class, and functioninthetoolkit,specifyingparametersandgivingexamplesofusage.Thewebsite also provides many HOWTOs with extensive examples and test cases, intended for users, developers, and instructors. Software Requirements To get the most out of this book, you should install several free software packages. Current download pointers and instructions are available at http://www.nltk.org/. Python The material presented in this book assumes that you are using Python version 2.4 or 2.5. We are committed to porting NLTK to Python 3.0 once the libraries that NLTK depends on have been ported. NLTK The code examples in this book use NLTK version 2.0. Subsequent releases of NLTK will be backward-compatible. Preface | xiii
  • 12. NLTK-Data This contains the linguistic corpora that are analyzed and processed in the book. NumPy (recommended) This is a scientific computing library with support for multidimensional arrays and linear algebra, required for certain probability, tagging, clustering, and classifica- tion tasks. Matplotlib (recommended) This is a 2D plotting library for data visualization, and is used in some of the book’s code samples that produce line graphs and bar charts. NetworkX (optional) This is a library for storing and manipulating network structures consisting of nodes and edges. For visualizing semantic networks, also install the Graphviz library. Prover9 (optional) This is an automated theorem prover for first-order and equational logic, used to support inference in language processing. Natural Language Toolkit (NLTK) NLTK was originally created in 2001 as part of a computational linguistics course in the Department of Computer and Information Science at the University of Pennsylva- nia. Since then it has been developed and expanded with the help of dozens of con- tributors. It has now been adopted in courses in dozens of universities, and serves as the basis of many research projects. Table P-2 lists the most important NLTK modules. Table P-2. Language processing tasks and corresponding NLTK modules with examples of functionality Language processing task NLTK modules Functionality Accessing corpora nltk.corpus Standardized interfaces to corpora and lexicons String processing nltk.tokenize, nltk.stem Tokenizers, sentence tokenizers, stemmers Collocation discovery nltk.collocations t-test, chi-squared, point-wise mutual information Part-of-speech tagging nltk.tag n-gram, backoff, Brill, HMM, TnT Classification nltk.classify, nltk.cluster Decision tree, maximum entropy, naive Bayes, EM, k-means Chunking nltk.chunk Regular expression, n-gram, named entity Parsing nltk.parse Chart, feature-based, unification, probabilistic, dependency Semantic interpretation nltk.sem, nltk.inference Lambda calculus, first-order logic, model checking Evaluation metrics nltk.metrics Precision, recall, agreement coefficients Probability and estimation nltk.probability Frequency distributions, smoothed probability distributions Applications nltk.app, nltk.chat Graphical concordancer, parsers, WordNet browser, chatbots xiv | Preface
  • 13. Language processing task NLTK modules Functionality Linguistic fieldwork nltk.toolbox Manipulate data in SIL Toolbox format NLTK was designed with four primary goals in mind: Simplicity To provide an intuitive framework along with substantial building blocks, giving users a practical knowledge of NLP without getting bogged down in the tedious house-keeping usually associated with processing annotated language data Consistency To provide a uniform framework with consistent interfaces and data structures, and easily guessable method names Extensibility To provide a structure into which new software modules can be easily accommo- dated, including alternative implementations and competing approaches to the same task Modularity To provide components that can be used independently without needing to un- derstand the rest of the toolkit Contrasting with these goals are three non-requirements—potentially useful qualities that we have deliberately avoided. First, while the toolkit provides a wide range of functions, it is not encyclopedic; it is a toolkit, not a system, and it will continue to evolve with the field of NLP. Second, while the toolkit is efficient enough to support meaningful tasks, it is not highly optimized for runtime performance; such optimiza- tions often involve more complex algorithms, or implementations in lower-level pro- gramming languages such as C or C++. This would make the software less readable and more difficult to install. Third, we have tried to avoid clever programming tricks, since we believe that clear implementations are preferable to ingenious yet indecipher- able ones. For Instructors Natural Language Processing is often taught within the confines of a single-semester course at the advanced undergraduate level or postgraduate level. Many instructors have found that it is difficult to cover both the theoretical and practical sides of the subject in such a short span of time. Some courses focus on theory to the exclusion of practical exercises, and deprive students of the challenge and excitement of writing programs to automatically process language. Other courses are simply designed to teach programming for linguists, and do not manage to cover any significant NLP con- tent. NLTK was originally developed to address this problem, making it feasible to cover a substantial amount of theory and practice within a single-semester course, even if students have no prior programming experience. Preface | xv
  • 14. A significant fraction of any NLP syllabus deals with algorithms and data structures. On their own these can be rather dry, but NLTK brings them to life with the help of interactive graphical user interfaces that make it possible to view algorithms step-by- step. Most NLTK components include a demonstration that performs an interesting task without requiring any special input from the user. An effective way to deliver the materials is through interactive presentation of the examples in this book, entering them in a Python session, observing what they do, and modifying them to explore some empirical or theoretical issue. This book contains hundreds of exercises that can be used as the basis for student assignments. The simplest exercises involve modifying a supplied program fragment in a specified way in order to answer a concrete question. At the other end of the spectrum, NLTK provides a flexible framework for graduate-level research projects, with standard implementations of all the basic data structures and algorithms, interfaces to dozens of widely used datasets (corpora), and a flexible and extensible architecture. Additional support for teaching using NLTK is available on the NLTK website. We believe this book is unique in providing a comprehensive framework for students to learn about NLP in the context of learning to program. What sets these materials apart is the tight coupling of the chapters and exercises with NLTK, giving students— even those with no prior programming experience—a practical introduction to NLP. After completing these materials, students will be ready to attempt one of the more advanced textbooks, such as Speech and Language Processing, by Jurafsky and Martin (Prentice Hall, 2008). This book presents programming concepts in an unusual order, beginning with a non- trivial data type—lists of strings—then introducing non-trivial control structures such as comprehensions and conditionals. These idioms permit us to do useful language processing from the start. Once this motivation is in place, we return to a systematic presentation of fundamental concepts such as strings, loops, files, and so forth. In this way, we cover the same ground as more conventional approaches, without expecting readers to be interested in the programming language for its own sake. Two possible course plans are illustrated in Table P-3. The first one presumes an arts/ humanities audience, whereas the second one presumes a science/engineering audi- ence. Other course plans could cover the first five chapters, then devote the remaining time to a single area, such as text classification (Chapters 6 and 7), syntax (Chapters 8 and 9), semantics (Chapter 10), or linguistic data management (Chapter 11). Table P-3. Suggested course plans; approximate number of lectures per chapter Chapter Arts and Humanities Science and Engineering Chapter 1, LanguageProcessingandPython 2–4 2 Chapter 2, AccessingTextCorporaandLexicalResources 2–4 2 Chapter 3, ProcessingRawText 2–4 2 Chapter 4, WritingStructuredPrograms 2–4 1–2 xvi | Preface
  • 15. Chapter Arts and Humanities Science and Engineering Chapter 5, CategorizingandTaggingWords 2–4 2–4 Chapter 6, LearningtoClassifyText 0–2 2–4 Chapter 7, ExtractingInformationfromText 2 2–4 Chapter 8, AnalyzingSentenceStructure 2–4 2–4 Chapter 9, BuildingFeature-BasedGrammars 2–4 1–4 Chapter 10, AnalyzingtheMeaningofSentences 1–2 1–4 Chapter 11, ManagingLinguisticData 1–2 1–4 Total 18–36 18–36 Conventions Used in This Book The following typographical conventions are used in this book: Bold Indicates new terms. Italic Used within paragraphs to refer to linguistic examples, the names of texts, and URLs; also used for filenames and file extensions. Constant width Used for program listings, as well as within paragraphs to refer to program elements such as variable or function names, statements, and keywords; also used for pro- gram names. Constant width italic Shows text that should be replaced with user-supplied values or by values deter- mined by context; also used for metavariables within program code examples. This icon signifies a tip, suggestion, or general note. This icon indicates a warning or caution. Using Code Examples This book is here to help you get your job done. In general, you may use the code in this book in your programs and documentation. You do not need to contact us for permission unless you’re reproducing a significant portion of the code. For example, Preface | xvii
  • 16. writing a program that uses several chunks of code from this book does not require permission. Selling or distributing a CD-ROM of examples from O’Reilly books does require permission. Answering a question by citing this book and quoting example code does not require permission. Incorporating a significant amount of example code from this book into your product’s documentation does require permission. We appreciate, but do not require, attribution. An attribution usually includes the title, author, publisher, and ISBN. For example: “Natural Language Processing with Py- thon, by Steven Bird, Ewan Klein, and Edward Loper. Copyright 2009 Steven Bird, Ewan Klein, and Edward Loper, 978-0-596-51649-9.” If you feel your use of code examples falls outside fair use or the permission given above, feel free to contact us at permissions@oreilly.com. Safari® Books Online When you see a Safari® Books Online icon on the cover of your favorite technology book, that means the book is available online through the O’Reilly Network Safari Bookshelf. Safari offers a solution that’s better than e-books. It’s a virtual library that lets you easily search thousands of top tech books, cut and paste code samples, download chapters, and find quick answers when you need the most accurate, current information. Try it for free at http://my.safaribooksonline.com. How to Contact Us Please address comments and questions concerning this book to the publisher: O’Reilly Media, Inc. 1005 Gravenstein Highway North Sebastopol, CA 95472 800-998-9938 (in the United States or Canada) 707-829-0515 (international or local) 707-829-0104 (fax) We have a web page for this book, where we list errata, examples, and any additional information. You can access this page at: http://www.oreilly.com/catalog/9780596516499 xviii | Preface
  • 17. The authors provide additional materials for each chapter via the NLTK website at: http://www.nltk.org/ To comment or ask technical questions about this book, send email to: bookquestions@oreilly.com For more information about our books, conferences, Resource Centers, and the O’Reilly Network, see our website at: http://www.oreilly.com Acknowledgments The authors are indebted to the following people for feedback on earlier drafts of this book:DougArnold,MichaelaAtterer,GregAumann,KennethBeesley,StevenBethard, Ondrej Bojar, Chris Cieri, Robin Cooper, Grev Corbett, James Curran, Dan Garrette, Jean Mark Gawron, Doug Hellmann, Nitin Indurkhya, Mark Liberman, Peter Ljunglöf, Stefan Müller, Robin Munn, Joel Nothman, Adam Przepiorkowski, Brandon Rhodes, Stuart Robinson, Jussi Salmela, Kyle Schlansker, Rob Speer, and Richard Sproat. We are thankful to many students and colleagues for their comments on the class materials that evolved into these chapters, including participants at NLP and linguistics summer schools in Brazil, India, and the USA. This book would not exist without the members of the nltk-dev developer community, named on the NLTK website, who have given so freely of their time and expertise in building and extending NLTK. We are grateful to the U.S. National Science Foundation, the Linguistic Data Consor- tium, an Edward Clarence Dyason Fellowship, and the Universities of Pennsylvania, Edinburgh, and Melbourne for supporting our work on this book. We thank Julie Steele, Abby Fox, Loranah Dimant, and the rest of the O’Reilly team, for organizing comprehensive reviews of our drafts from people across the NLP and Python communities, for cheerfully customizing O’Reilly’s production tools to accom- modate our needs, and for meticulous copyediting work. Finally, we owe a huge debt of gratitude to our partners, Kay, Mimo, and Jee, for their love, patience, and support over the many years that we worked on this book. We hope that our children—Andrew, Alison, Kirsten, Leonie, and Maaike—catch our enthusi- asm for language and computation from these pages. Royalties Royalties from the sale of this book are being used to support the development of the Natural Language Toolkit. Preface | xix
  • 18. Figure P-1. Edward Loper, Ewan Klein, and Steven Bird, Stanford, July 2007 xx | Preface
  • 19. CHAPTER 1 Language Processing and Python It is easy to get our hands on millions of words of text. What can we do with it, assuming we can write some simple programs? In this chapter, we’ll address the following questions: 1. What can we achieve by combining simple programming techniques with large quantities of text? 2. How can we automatically extract key words and phrases that sum up the style and content of a text? 3. What tools and techniques does the Python programming language provide for such work? 4. What are some of the interesting challenges of natural language processing? This chapter is divided into sections that skip between two quite different styles. In the “computing with language” sections, we will take on some linguistically motivated programming tasks without necessarily explaining how they work. In the “closer look at Python” sections we will systematically review key programming concepts. We’ll flag the two styles in the section titles, but later chapters will mix both styles without being so up-front about it. We hope this style of introduction gives you an authentic taste of what will come later, while covering a range of elementary concepts in linguis- tics and computer science. If you have basic familiarity with both areas, you can skip to Section 1.5; we will repeat any important points in later chapters, and if you miss anything you can easily consult the online reference material at http://www.nltk.org/. If the material is completely new to you, this chapter will raise more questions than it answers, questions that are addressed in the rest of this book. 1.1 Computing with Language: Texts and Words We’re all very familiar with text, since we read and write it every day. Here we will treat text as raw data for the programs we write, programs that manipulate and analyze it in a variety of interesting ways. But before we can do this, we have to get started with the Python interpreter. 1
  • 20. Getting Started with Python One of the friendly things about Python is that it allows you to type directly into the interactive interpreter—the program that will be running your Python programs. You can access the Python interpreter using a simple graphical interface called the In- teractive DeveLopment Environment (IDLE). On a Mac you can find this under Ap- plications→MacPython, and on Windows under All Programs→Python. Under Unix you can run Python from the shell by typing idle (if this is not installed, try typing python). The interpreter will print a blurb about your Python version; simply check that you are running Python 2.4 or 2.5 (here it is 2.5.1): Python 2.5.1 (r251:54863, Apr 15 2008, 22:57:26) [GCC 4.0.1 (Apple Inc. build 5465)] on darwin Type "help", "copyright", "credits" or "license" for more information. >>> If you are unable to run the Python interpreter, you probably don’t have Python installed correctly. Please visit http://python.org/ for detailed in- structions. The >>> prompt indicates that the Python interpreter is now waiting for input. When copying examples from this book, don’t type the “>>>” yourself. Now, let’s begin by using Python as a calculator: >>> 1 + 5 * 2 - 3 8 >>> Once the interpreter has finished calculating the answer and displaying it, the prompt reappears. This means the Python interpreter is waiting for another instruction. Your Turn: Enter a few more expressions of your own. You can use asterisk (*) for multiplication and slash (/) for division, and parentheses for bracketing expressions. Note that division doesn’t always behave as you might expect—it does integer division (with rounding of fractions downwards) when you type 1/3 and “floating-point” (or decimal) divi- sion when you type 1.0/3.0. In order to get the expected behavior of division (standard in Python 3.0), you need to type: from __future__ import division. The preceding examples demonstrate how you can work interactively with the Python interpreter, experimenting with various expressions in the language to see what they do. Now let’s try a non-sensical expression to see how the interpreter handles it: 2 | Chapter 1: Language Processing and Python
  • 21. >>> 1 + File "<stdin>", line 1 1 + ^ SyntaxError: invalid syntax >>> This produced a syntax error. In Python, it doesn’t make sense to end an instruction with a plus sign. The Python interpreter indicates the line where the problem occurred (line 1 of <stdin>, which stands for “standard input”). Now that we can use the Python interpreter, we’re ready to start working with language data. Getting Started with NLTK Before going further you should install NLTK, downloadable for free from http://www .nltk.org/. Follow the instructions there to download the version required for your platform. Once you’ve installed NLTK, start up the Python interpreter as before, and install the data required for the book by typing the following two commands at the Python prompt, then selecting the book collection as shown in Figure 1-1. >>> import nltk >>> nltk.download() Figure 1-1. Downloading the NLTK Book Collection: Browse the available packages using nltk.download(). The Collections tab on the downloader shows how the packages are grouped into sets, and you should select the line labeled book to obtain all data required for the examples and exercises in this book. It consists of about 30 compressed files requiring about 100Mb disk space. The full collection of data (i.e., all in the downloader) is about five times this size (at the time of writing) and continues to expand. Once the data is downloaded to your machine, you can load some of it using the Python interpreter. The first step is to type a special command at the Python prompt, which 1.1 Computing with Language: Texts and Words | 3
  • 22. tells the interpreter to load some texts for us to explore: from nltk.book import *. This says “from NLTK’s book module, load all items.” The book module contains all the data you will need as you read this chapter. After printing a welcome message, it loads the text of several books (this will take a few seconds). Here’s the command again, together with the output that you will see. Take care to get spelling and punctuation right, and remember that you don’t type the >>>. >>> from nltk.book import * *** Introductory Examples for the NLTK Book *** Loading text1, ..., text9 and sent1, ..., sent9 Type the name of the text or sentence to view it. Type: 'texts()' or 'sents()' to list the materials. text1: Moby Dick by Herman Melville 1851 text2: Sense and Sensibility by Jane Austen 1811 text3: The Book of Genesis text4: Inaugural Address Corpus text5: Chat Corpus text6: Monty Python and the Holy Grail text7: Wall Street Journal text8: Personals Corpus text9: The Man Who Was Thursday by G . K . Chesterton 1908 >>> Any time we want to find out about these texts, we just have to enter their names at the Python prompt: >>> text1 <Text: Moby Dick by Herman Melville 1851> >>> text2 <Text: Sense and Sensibility by Jane Austen 1811> >>> Now that we can use the Python interpreter, and have some data to work with, we’re ready to get started. Searching Text There are many ways to examine the context of a text apart from simply reading it. A concordance view shows us every occurrence of a given word, together with some context. Here we look up the word monstrous in Moby Dick by entering text1 followed by a period, then the term concordance, and then placing "monstrous" in parentheses: >>> text1.concordance("monstrous") Building index... Displaying 11 of 11 matches: ong the former , one was of a most monstrous size . ... This came towards us , ON OF THE PSALMS . " Touching that monstrous bulk of the whale or ork we have r ll over with a heathenish array of monstrous clubs and spears . Some were thick d as you gazed , and wondered what monstrous cannibal and savage could ever hav that has survived the flood ; most monstrous and most mountainous ! That Himmal they might scout at Moby Dick as a monstrous fable , or still worse and more de th of Radney .'" CHAPTER 55 Of the monstrous Pictures of Whales . I shall ere l ing Scenes . In connexion with the monstrous pictures of whales , I am strongly ere to enter upon those still more monstrous stories of them which are to be fo 4 | Chapter 1: Language Processing and Python
  • 23. ght have been rummaged out of this monstrous cabinet there is no telling . But of Whale - Bones ; for Whales of a monstrous size are oftentimes cast up dead u >>> Your Turn: Try searching for other words; to save re-typing, you might be able to use up-arrow, Ctrl-up-arrow, or Alt-p to access the previous command and modify the word being searched. You can also try search- es on some of the other texts we have included. For example, search Sense and Sensibility for the word affection, using text2.concord ance("affection"). Search the book of Genesis to find out how long some people lived, using: text3.concordance("lived"). You could look at text4, the Inaugural Address Corpus, to see examples of English going back to 1789, and search for words like nation, terror, god to see how these words have been used differently over time. We’ve also included text5, the NPS Chat Corpus: search this for unconventional words like im, ur, lol. (Note that this corpus is uncensored!) Once you’ve spent a little while examining these texts, we hope you have a new sense of the richness and diversity of language. In the next chapter you will learn how to access a broader range of text, including text in languages other than English. A concordance permits us to see words in context. For example, we saw that mon- strous occurred in contexts such as the ___ pictures and the ___ size. What other words appear in a similar range of contexts? We can find out by appending the term similar to the name of the text in question, then inserting the relevant word in parentheses: >>> text1.similar("monstrous") Building word-context index... subtly impalpable pitiable curious imperial perilous trustworthy abundant untoward singular lamentable few maddens horrible loving lazy mystifying christian exasperate puzzled >>> text2.similar("monstrous") Building word-context index... very exceedingly so heartily a great good amazingly as sweet remarkably extremely vast >>> Observe that we get different results for different texts. Austen uses this word quite differently from Melville; for her, monstrous has positive connotations, and sometimes functions as an intensifier like the word very. The term common_contexts allows us to examine just the contexts that are shared by two or more words, such as monstrous and very. We have to enclose these words by square brackets as well as parentheses, and separate them with a comma: >>> text2.common_contexts(["monstrous", "very"]) be_glad am_glad a_pretty is_pretty a_lucky >>> 1.1 Computing with Language: Texts and Words | 5
  • 24. Your Turn: Pick another pair of words and compare their usage in two different texts, using the similar() and common_contexts() functions. It is one thing to automatically detect that a particular word occurs in a text, and to display some words that appear in the same context. However, we can also determine the location of a word in the text: how many words from the beginning it appears. This positional information can be displayed using a dispersion plot. Each stripe represents an instance of a word, and each row represents the entire text. In Figure 1-2 we see some striking patterns of word usage over the last 220 years (in an artificial text con- structed by joining the texts of the Inaugural Address Corpus end-to-end). You can produce this plot as shown below. You might like to try more words (e.g., liberty, constitution) and different texts. Can you predict the dispersion of a word before you view it? As before, take care to get the quotes, commas, brackets, and parentheses exactly right. >>> text4.dispersion_plot(["citizens", "democracy", "freedom", "duties", "America"]) >>> Important: You need to have Python’s NumPy and Matplotlib pack- ages installed in order to produce the graphical plots used in this book. Please see http://www.nltk.org/ for installation instructions. Now, just for fun, let’s try generating some random text in the various styles we have just seen. To do this, we type the name of the text followed by the term generate. (We need to include the parentheses, but there’s nothing that goes between them.) Figure 1-2. Lexical dispersion plot for words in U.S. Presidential Inaugural Addresses: This can be used to investigate changes in language use over time. 6 | Chapter 1: Language Processing and Python
  • 25. >>> text3.generate() In the beginning of his brother is a hairy man , whose top may reach unto heaven ; and ye shall sow the land of Egypt there was no bread in all that he was taken out of the month , upon the earth . So shall thy wages be ? And they made their father ; and Isaac was old , and kissed him : and Laban with his cattle in the midst of the hands of Esau thy first born , and Phichol the chief butler unto his son Isaac , she >>> Note that the first time you run this command, it is slow because it gathers statistics about word sequences. Each time you run it, you will get different output text. Now try generating random text in the style of an inaugural address or an Internet chat room. Although the text is random, it reuses common words and phrases from the source text and gives us a sense of its style and content. (What is lacking in this randomly generated text?) When generate produces its output, punctuation is split off from the preceding word. While this is not correct formatting for English text, we do it to make clear that words and punctuation are independent of one another. You will learn more about this in Chapter 3. Counting Vocabulary The most obvious fact about texts that emerges from the preceding examples is that they differ in the vocabulary they use. In this section, we will see how to use the com- puter to count the words in a text in a variety of useful ways. As before, you will jump right in and experiment with the Python interpreter, even though you may not have studied Python systematically yet. Test your understanding by modifying the examples, and trying the exercises at the end of the chapter. Let’s begin by finding out the length of a text from start to finish, in terms of the words and punctuation symbols that appear. We use the term len to get the length of some- thing, which we’ll apply here to the book of Genesis: >>> len(text3) 44764 >>> So Genesis has 44,764 words and punctuation symbols, or “tokens.” A token is the technical name for a sequence of characters—such as hairy, his, or :)—that we want to treat as a group. When we count the number of tokens in a text, say, the phrase to be or not to be, we are counting occurrences of these sequences. Thus, in our example phrase there are two occurrences of to, two of be, and one each of or and not. But there are only four distinct vocabulary items in this phrase. How many distinct words does the book of Genesis contain? To work this out in Python, we have to pose the question slightly differently. The vocabulary of a text is just the set of tokens that it uses, since in a set, all duplicates are collapsed together. In Python we can obtain the vocabulary 1.1 Computing with Language: Texts and Words | 7
  • 26. items of text3 with the command: set(text3). When you do this, many screens of words will fly past. Now try the following: >>> sorted(set(text3)) ['!', "'", '(', ')', ',', ',)', '.', '.)', ':', ';', ';)', '?', '?)', 'A', 'Abel', 'Abelmizraim', 'Abidah', 'Abide', 'Abimael', 'Abimelech', 'Abr', 'Abrah', 'Abraham', 'Abram', 'Accad', 'Achbor', 'Adah', ...] >>> len(set(text3)) 2789 >>> By wrapping sorted() around the Python expression set(text3) , we obtain a sorted list of vocabulary items, beginning with various punctuation symbols and continuing with words starting with A. All capitalized words precede lowercase words. We dis- cover the size of the vocabulary indirectly, by asking for the number of items in the set, and again we can use len to obtain this number . Although it has 44,764 tokens, this book has only 2,789 distinct words, or “word types.” A word type is the form or spelling of the word independently of its specific occurrences in a text—that is, the word considered as a unique item of vocabulary. Our count of 2,789 items will include punctuation symbols, so we will generally call these unique items types instead of word types. Now, let’s calculate a measure of the lexical richness of the text. The next example shows us that each word is used 16 times on average (we need to make sure Python uses floating-point division): >>> from __future__ import division >>> len(text3) / len(set(text3)) 16.050197203298673 >>> Next, let’s focus on particular words. We can count how often a word occurs in a text, and compute what percentage of the text is taken up by a specific word: >>> text3.count("smote") 5 >>> 100 * text4.count('a') / len(text4) 1.4643016433938312 >>> Your Turn: How many times does the word lol appear in text5? How much is this as a percentage of the total number of words in this text? You may want to repeat such calculations on several texts, but it is tedious to keep retyping the formula. Instead, you can come up with your own name for a task, like “lexical_diversity” or “percentage”, and associate it with a block of code. Now you only have to type a short name instead of one or more complete lines of Python code, and you can reuse it as often as you like. The block of code that does a task for us is 8 | Chapter 1: Language Processing and Python
  • 27. called a function, and we define a short name for our function with the keyword def. The next example shows how to define two new functions, lexical_diversity() and percentage(): >>> def lexical_diversity(text): ... return len(text) / len(set(text)) ... >>> def percentage(count, total): ... return 100 * count / total ... Caution! The Python interpreter changes the prompt from >>> to ... after en- countering the colon at the end of the first line. The... prompt indicates that Python expects an indented code block to appear next. It is up to you to do the indentation, by typing four spaces or hitting the Tab key. To finish the indented block, just enter a blank line. In the definition of lexical diversity() , we specify a parameter labeled text. This parameter is a “placeholder” for the actual text whose lexical diversity we want to compute, and reoccurs in the block of code that will run when the function is used, in line . Similarly, percentage() is defined to take two parameters, labeled count and total . Once Python knows that lexical_diversity() and percentage() are the names for spe- cific blocks of code, we can go ahead and use these functions: >>> lexical_diversity(text3) 16.050197203298673 >>> lexical_diversity(text5) 7.4200461589185629 >>> percentage(4, 5) 80.0 >>> percentage(text4.count('a'), len(text4)) 1.4643016433938312 >>> To recap, we use or call a function such as lexical_diversity() by typing its name, followed by an open parenthesis, the name of the text, and then a close parenthesis. These parentheses will show up often; their role is to separate the name of a task—such as lexical_diversity()—from the data that the task is to be performed on—such as text3. The data value that we place in the parentheses when we call a function is an argument to the function. You have already encountered several functions in this chapter, such as len(), set(), and sorted(). By convention, we will always add an empty pair of parentheses after a function name, as in len(), just to make clear that what we are talking about is a func- tion rather than some other kind of Python expression. Functions are an important concept in programming, and we only mention them at the outset to give newcomers 1.1 Computing with Language: Texts and Words | 9
  • 28. a sense of the power and creativity of programming. Don’t worry if you find it a bit confusing right now. Later we’ll see how to use functions when tabulating data, as in Table 1-1. Each row of the table will involve the same computation but with different data, and we’ll do this repetitive work using a function. Table 1-1. Lexical diversity of various genres in the Brown Corpus Genre Tokens Types Lexical diversity skill and hobbies 82345 11935 6.9 humor 21695 5017 4.3 fiction: science 14470 3233 4.5 press: reportage 100554 14394 7.0 fiction: romance 70022 8452 8.3 religion 39399 6373 6.2 1.2 A Closer Look at Python: Texts as Lists of Words You’ve seen some important elements of the Python programming language. Let’s take a few moments to review them systematically. Lists What is a text? At one level, it is a sequence of symbols on a page such as this one. At another level, it is a sequence of chapters, made up of a sequence of sections, where each section is a sequence of paragraphs, and so on. However, for our purposes, we will think of a text as nothing more than a sequence of words and punctuation. Here’s how we represent text in Python, in this case the opening sentence of Moby Dick: >>> sent1 = ['Call', 'me', 'Ishmael', '.'] >>> After the prompt we’ve given a name we made up, sent1, followed by the equals sign, and then some quoted words, separated with commas, and surrounded with brackets. This bracketed material is known as a list in Python: it is how we store a text. We can inspect it by typing the name . We can ask for its length . We can even apply our own lexical_diversity() function to it . >>> sent1 ['Call', 'me', 'Ishmael', '.'] >>> len(sent1) 4 >>> lexical_diversity(sent1) 1.0 >>> 10 | Chapter 1: Language Processing and Python
  • 29. Some more lists have been defined for you, one for the opening sentence of each of our texts, sent2 … sent9. We inspect two of them here; you can see the rest for yourself using the Python interpreter (if you get an error saying that sent2 is not defined, you need to first type from nltk.book import *). >>> sent2 ['The', 'family', 'of', 'Dashwood', 'had', 'long', 'been', 'settled', 'in', 'Sussex', '.'] >>> sent3 ['In', 'the', 'beginning', 'God', 'created', 'the', 'heaven', 'and', 'the', 'earth', '.'] >>> Your Turn: Make up a few sentences of your own, by typing a name, equals sign, and a list of words, like this: ex1 = ['Monty', 'Python', 'and', 'the', 'Holy', 'Grail']. Repeat some of the other Python op- erations we saw earlier in Section 1.1, e.g., sorted(ex1), len(set(ex1)), ex1.count('the'). A pleasant surprise is that we can use Python’s addition operator on lists. Adding two lists creates a new list with everything from the first list, followed by everything from the second list: >>> ['Monty', 'Python'] + ['and', 'the', 'Holy', 'Grail'] ['Monty', 'Python', 'and', 'the', 'Holy', 'Grail'] This special use of the addition operation is called concatenation; it combines the lists together into a single list. We can concatenate sen- tences to build up a text. We don’t have to literally type the lists either; we can use short names that refer to pre- defined lists. >>> sent4 + sent1 ['Fellow', '-', 'Citizens', 'of', 'the', 'Senate', 'and', 'of', 'the', 'House', 'of', 'Representatives', ':', 'Call', 'me', 'Ishmael', '.'] >>> What if we want to add a single item to a list? This is known as appending. When we append() to a list, the list itself is updated as a result of the operation. >>> sent1.append("Some") >>> sent1 ['Call', 'me', 'Ishmael', '.', 'Some'] >>> 1.2 A Closer Look at Python: Texts as Lists of Words | 11
  • 30. Indexing Lists As we have seen, a text in Python is a list of words, represented using a combination of brackets and quotes. Just as with an ordinary page of text, we can count up the total number of words in text1 with len(text1), and count the occurrences in a text of a particular word—say, heaven—using text1.count('heaven'). With some patience, we can pick out the 1st, 173rd, or even 14,278th word in a printed text. Analogously, we can identify the elements of a Python list by their order of oc- currence in the list. The number that represents this position is the item’s index. We instruct Python to show us the item that occurs at an index such as 173 in a text by writing the name of the text followed by the index inside square brackets: >>> text4[173] 'awaken' >>> We can do the converse; given a word, find the index of when it first occurs: >>> text4.index('awaken') 173 >>> Indexes are a common way to access the words of a text, or, more generally, the ele- ments of any list. Python permits us to access sublists as well, extracting manageable pieces of language from large texts, a technique known as slicing. >>> text5[16715:16735] ['U86', 'thats', 'why', 'something', 'like', 'gamefly', 'is', 'so', 'good', 'because', 'you', 'can', 'actually', 'play', 'a', 'full', 'game', 'without', 'buying', 'it'] >>> text6[1600:1625] ['We', "'", 're', 'an', 'anarcho', '-', 'syndicalist', 'commune', '.', 'We', 'take', 'it', 'in', 'turns', 'to', 'act', 'as', 'a', 'sort', 'of', 'executive', 'officer', 'for', 'the', 'week'] >>> Indexes have some subtleties, and we’ll explore these with the help of an artificial sentence: >>> sent = ['word1', 'word2', 'word3', 'word4', 'word5', ... 'word6', 'word7', 'word8', 'word9', 'word10'] >>> sent[0] 'word1' >>> sent[9] 'word10' >>> Notice that our indexes start from zero: sent element zero, written sent[0], is the first word, 'word1', whereas sent element 9 is 'word10'. The reason is simple: the moment Python accesses the content of a list from the computer’s memory, it is already at the first element; we have to tell it how many elements forward to go. Thus, zero steps forward leaves it at the first element. 12 | Chapter 1: Language Processing and Python
  • 31. This practice of counting from zero is initially confusing, but typical of modern programming languages. You’ll quickly get the hang of it if you’ve mastered the system of counting centuries where 19XY is a year in the 20th century, or if you live in a country where the floors of a building are numbered from 1, and so walking up n-1 flights of stairs takes you to level n. Now, if we accidentally use an index that is too large, we get an error: >>> sent[10] Traceback (most recent call last): File "<stdin>", line 1, in ? IndexError: list index out of range >>> This time it is not a syntax error, because the program fragment is syntactically correct. Instead, it is a runtime error, and it produces a Traceback message that shows the context of the error, followed by the name of the error, IndexError, and a brief explanation. Let’s take a closer look at slicing, using our artificial sentence again. Here we verify that the slice 5:8 includes sent elements at indexes 5, 6, and 7: >>> sent[5:8] ['word6', 'word7', 'word8'] >>> sent[5] 'word6' >>> sent[6] 'word7' >>> sent[7] 'word8' >>> By convention, m:n means elements m…n-1. As the next example shows, we can omit the first number if the slice begins at the start of the list , and we can omit the second number if the slice goes to the end : >>> sent[:3] ['word1', 'word2', 'word3'] >>> text2[141525:] ['among', 'the', 'merits', 'and', 'the', 'happiness', 'of', 'Elinor', 'and', 'Marianne', ',', 'let', 'it', 'not', 'be', 'ranked', 'as', 'the', 'least', 'considerable', ',', 'that', 'though', 'sisters', ',', 'and', 'living', 'almost', 'within', 'sight', 'of', 'each', 'other', ',', 'they', 'could', 'live', 'without', 'disagreement', 'between', 'themselves', ',', 'or', 'producing', 'coolness', 'between', 'their', 'husbands', '.', 'THE', 'END'] >>> We can modify an element of a list by assigning to one of its index values. In the next example, we put sent[0] on the left of the equals sign . We can also replace an entire slice with new material . A consequence of this last change is that the list only has four elements, and accessing a later value generates an error . 1.2 A Closer Look at Python: Texts as Lists of Words | 13
  • 32. >>> sent[0] = 'First' >>> sent[9] = 'Last' >>> len(sent) 10 >>> sent[1:9] = ['Second', 'Third'] >>> sent ['First', 'Second', 'Third', 'Last'] >>> sent[9] Traceback (most recent call last): File "<stdin>", line 1, in ? IndexError: list index out of range >>> Your Turn: Take a few minutes to define a sentence of your own and modify individual words and groups of words (slices) using the same methods used earlier. Check your understanding by trying the exercises on lists at the end of this chapter. Variables From the start of Section 1.1, you have had access to texts called text1, text2, and so on. It saved a lot of typing to be able to refer to a 250,000-word book with a short name like this! In general, we can make up names for anything we care to calculate. We did this ourselves in the previous sections, e.g., defining a variable sent1, as follows: >>> sent1 = ['Call', 'me', 'Ishmael', '.'] >>> Such lines have the form: variable = expression. Python will evaluate the expression, and save its result to the variable. This process is called assignment. It does not gen- erate any output; you have to type the variable on a line of its own to inspect its contents. The equals sign is slightly misleading, since information is moving from the right side to the left. It might help to think of it as a left-arrow. The name of the variable can be anything you like, e.g., my_sent, sentence, xyzzy. It must start with a letter, and can include numbers and underscores. Here are some examples of variables and assignments: >>> my_sent = ['Bravely', 'bold', 'Sir', 'Robin', ',', 'rode', ... 'forth', 'from', 'Camelot', '.'] >>> noun_phrase = my_sent[1:4] >>> noun_phrase ['bold', 'Sir', 'Robin'] >>> wOrDs = sorted(noun_phrase) >>> wOrDs ['Robin', 'Sir', 'bold'] >>> Remember that capitalized words appear before lowercase words in sorted lists. 14 | Chapter 1: Language Processing and Python
  • 33. Notice in the previous example that we split the definition of my_sent over two lines. Python expressions can be split across multiple lines, so long as this happens within any kind of brackets. Python uses the ... prompt to indicate that more input is expected. It doesn’t matter how much indentation is used in these continuation lines, but some inden- tation usually makes them easier to read. It is good to choose meaningful variable names to remind you—and to help anyone else who reads your Python code—what your code is meant to do. Python does not try to make sense of the names; it blindly follows your instructions, and does not object if you do something confusing, such as one = 'two' or two = 3. The only restriction is that a variable name cannot be any of Python’s reserved words, such as def, if, not, and import. If you use a reserved word, Python will produce a syntax error: >>> not = 'Camelot' File "<stdin>", line 1 not = 'Camelot' ^ SyntaxError: invalid syntax >>> We will often use variables to hold intermediate steps of a computation, especially when this makes the code easier to follow. Thus len(set(text1)) could also be written: >>> vocab = set(text1) >>> vocab_size = len(vocab) >>> vocab_size 19317 >>> Caution! Take care with your choice of names (or identifiers) for Python varia- bles. First, you should start the name with a letter, optionally followed by digits (0 to 9) or letters. Thus, abc23 is fine, but 23abc will cause a syntax error. Names are case-sensitive, which means that myVar and myvar are distinct variables. Variable names cannot contain whitespace, but you can separate words using an underscore, e.g., my_var. Be careful not to insert a hyphen instead of an underscore: my-var is wrong, since Python interprets the - as a minus sign. Strings Some of the methods we used to access the elements of a list also work with individual words, or strings. For example, we can assign a string to a variable , index a string , and slice a string . 1.2 A Closer Look at Python: Texts as Lists of Words | 15
  • 34. >>> name = 'Monty' >>> name[0] 'M' >>> name[:4] 'Mont' >>> We can also perform multiplication and addition with strings: >>> name * 2 'MontyMonty' >>> name + '!' 'Monty!' >>> We can join the words of a list to make a single string, or split a string into a list, as follows: >>> ' '.join(['Monty', 'Python']) 'Monty Python' >>> 'Monty Python'.split() ['Monty', 'Python'] >>> We will come back to the topic of strings in Chapter 3. For the time being, we have two important building blocks—lists and strings—and are ready to get back to some language analysis. 1.3 Computing with Language: Simple Statistics Let’s return to our exploration of the ways we can bring our computational resources to bear on large quantities of text. We began this discussion in Section 1.1, and saw how to search for words in context, how to compile the vocabulary of a text, how to generate random text in the same style, and so on. In this section, we pick up the question of what makes a text distinct, and use automatic methods to find characteristic words and expressions of a text. As in Section 1.1, you can try new features of the Python language by copying them into the interpreter, and you’ll learn about these features systematically in the following section. Before continuing further, you might like to check your understanding of the last sec- tion by predicting the output of the following code. You can use the interpreter to check whether you got it right. If you’re not sure how to do this task, it would be a good idea to review the previous section before continuing further. >>> saying = ['After', 'all', 'is', 'said', 'and', 'done', ... 'more', 'is', 'said', 'than', 'done'] >>> tokens = set(saying) >>> tokens = sorted(tokens) >>> tokens[-2:] what output do you expect here? >>> 16 | Chapter 1: Language Processing and Python
  • 35. Frequency Distributions How can we automatically identify the words of a text that are most informative about the topic and genre of the text? Imagine how you might go about finding the 50 most frequent words of a book. One method would be to keep a tally for each vocabulary item, like that shown in Figure 1-3. The tally would need thousands of rows, and it would be an exceedingly laborious process—so laborious that we would rather assign the task to a machine. Figure 1-3. Counting words appearing in a text (a frequency distribution). The table in Figure 1-3 is known as a frequency distribution , and it tells us the frequency of each vocabulary item in the text. (In general, it could count any kind of observable event.) It is a “distribution” since it tells us how the total number of word tokens in the text are distributed across the vocabulary items. Since we often need frequency distributions in language processing, NLTK provides built-in support for them. Let’s use a FreqDist to find the 50 most frequent words of Moby Dick. Try to work out what is going on here, then read the explanation that follows. >>> fdist1 = FreqDist(text1) >>> fdist1 <FreqDist with 260819 outcomes> >>> vocabulary1 = fdist1.keys() >>> vocabulary1[:50] [',', 'the', '.', 'of', 'and', 'a', 'to', ';', 'in', 'that', "'", '-', 'his', 'it', 'I', 's', 'is', 'he', 'with', 'was', 'as', '"', 'all', 'for', 'this', '!', 'at', 'by', 'but', 'not', '--', 'him', 'from', 'be', 'on', 'so', 'whale', 'one', 'you', 'had', 'have', 'there', 'But', 'or', 'were', 'now', 'which', '?', 'me', 'like'] >>> fdist1['whale'] 906 >>> When we first invoke FreqDist, we pass the name of the text as an argument . We can inspect the total number of words (“outcomes”) that have been counted up — 260,819 in the case of Moby Dick. The expression keys() gives us a list of all the distinct types in the text , and we can look at the first 50 of these by slicing the list . 1.3 Computing with Language: Simple Statistics | 17
  • 36. Your Turn: Try the preceding frequency distribution example for your- self, for text2. Be careful to use the correct parentheses and uppercase letters. If you get an error message NameError: name 'FreqDist' is not defined, you need to start your work with from nltk.book import *. Do any words produced in the last example help us grasp the topic or genre of this text? Only one word, whale, is slightly informative! It occurs over 900 times. The rest of the words tell us nothing about the text; they’re just English “plumbing.” What proportion of the text is taken up with such words? We can generate a cumulative frequency plot for these words, using fdist1.plot(50, cumulative=True), to produce the graph in Figure 1-4. These 50 words account for nearly half the book! Figure 1-4. Cumulative frequency plot for the 50 most frequently used words in Moby Dick, which account for nearly half of the tokens. 18 | Chapter 1: Language Processing and Python
  • 37. If the frequent words don’t help us, how about the words that occur once only, the so- called hapaxes? View them by typing fdist1.hapaxes(). This list contains lexicographer, cetological, contraband, expostulations, and about 9,000 others. It seems that there are too many rare words, and without seeing the context we probably can’t guess what half of the hapaxes mean in any case! Since neither frequent nor infrequent words help, we need to try something else. Fine-Grained Selection of Words Next, let’s look at the long words of a text; perhaps these will be more characteristic and informative. For this we adapt some notation from set theory. We would like to find the words from the vocabulary of the text that are more than 15 characters long. Let’s call this property P, so that P(w) is true if and only if w is more than 15 characters long. Now we can express the words of interest using mathematical set notation as shown in (1a). This means “the set of all w such that w is an element of V (the vocabu- lary) and w has property P.” (1) a. {w | w ∈ V & P(w)} b. [w for w in V if p(w)] The corresponding Python expression is given in (1b). (Note that it produces a list, not a set, which means that duplicates are possible.) Observe how similar the two notations are. Let’s go one more step and write executable Python code: >>> V = set(text1) >>> long_words = [w for w in V if len(w) > 15] >>> sorted(long_words) ['CIRCUMNAVIGATION', 'Physiognomically', 'apprehensiveness', 'cannibalistically', 'characteristically', 'circumnavigating', 'circumnavigation', 'circumnavigations', 'comprehensiveness', 'hermaphroditical', 'indiscriminately', 'indispensableness', 'irresistibleness', 'physiognomically', 'preternaturalness', 'responsibilities', 'simultaneousness', 'subterraneousness', 'supernaturalness', 'superstitiousness', 'uncomfortableness', 'uncompromisedness', 'undiscriminating', 'uninterpenetratingly'] >>> For each word w in the vocabulary V, we check whether len(w) is greater than 15; all other words will be ignored. We will discuss this syntax more carefully later. Your Turn: Try out the previous statements in the Python interpreter, and experiment with changing the text and changing the length condi- tion. Does it make an difference to your results if you change the variable names, e.g., using [word for word in vocab if ...]? 1.3 Computing with Language: Simple Statistics | 19
  • 38. Let’s return to our task of finding words that characterize a text. Notice that the long words in text4 reflect its national focus—constitutionally, transcontinental—whereas those in text5 reflect its informal content: boooooooooooglyyyyyy and yuuuuuuuuuuuummmmmmmmmmmm. Have we succeeded in automatically extract- ing words that typify a text? Well, these very long words are often hapaxes (i.e., unique) and perhaps it would be better to find frequently occurring long words. This seems promising since it eliminates frequent short words (e.g., the) and infrequent long words (e.g., antiphilosophists). Here are all words from the chat corpus that are longer than seven characters, that occur more than seven times: >>> fdist5 = FreqDist(text5) >>> sorted([w for w in set(text5) if len(w) > 7 and fdist5[w] > 7]) ['#14-19teens', '#talkcity_adults', '((((((((((', '........', 'Question', 'actually', 'anything', 'computer', 'cute.-ass', 'everyone', 'football', 'innocent', 'listening', 'remember', 'seriously', 'something', 'together', 'tomorrow', 'watching'] >>> Notice how we have used two conditions: len(w) > 7 ensures that the words are longer than seven letters, and fdist5[w] > 7 ensures that these words occur more than seven times. At last we have managed to automatically identify the frequently occurring con- tent-bearing words of the text. It is a modest but important milestone: a tiny piece of code, processing tens of thousands of words, produces some informative output. Collocations and Bigrams A collocation is a sequence of words that occur together unusually often. Thus red wine is a collocation, whereas the wine is not. A characteristic of collocations is that they are resistant to substitution with words that have similar senses; for example, maroon wine sounds very odd. To get a handle on collocations, we start off by extracting from a text a list of word pairs, also known as bigrams. This is easily accomplished with the function bigrams(): >>> bigrams(['more', 'is', 'said', 'than', 'done']) [('more', 'is'), ('is', 'said'), ('said', 'than'), ('than', 'done')] >>> Here we see that the pair of words than-done is a bigram, and we write it in Python as ('than', 'done'). Now, collocations are essentially just frequent bigrams, except that we want to pay more attention to the cases that involve rare words. In particular, we want to find bigrams that occur more often than we would expect based on the fre- quency of individual words. The collocations() function does this for us (we will see how it works later): >>> text4.collocations() Building collocations list United States; fellow citizens; years ago; Federal Government; General Government; American people; Vice President; Almighty God; Fellow citizens; Chief Magistrate; Chief Justice; God bless; Indian tribes; public debt; foreign nations; political parties; State governments; 20 | Chapter 1: Language Processing and Python
  • 39. National Government; United Nations; public money >>> text8.collocations() Building collocations list medium build; social drinker; quiet nights; long term; age open; financially secure; fun times; similar interests; Age open; poss rship; single mum; permanent relationship; slim build; seeks lady; Late 30s; Photo pls; Vibrant personality; European background; ASIAN LADY; country drives >>> The collocations that emerge are very specific to the genre of the texts. In order to find red wine as a collocation, we would need to process a much larger body of text. Counting Other Things Counting words is useful, but we can count other things too. For example, we can look at the distribution of word lengths in a text, by creating a FreqDist out of a long list of numbers, where each number is the length of the corresponding word in the text: >>> [len(w) for w in text1] [1, 4, 4, 2, 6, 8, 4, 1, 9, 1, 1, 8, 2, 1, 4, 11, 5, 2, 1, 7, 6, 1, 3, 4, 5, 2, ...] >>> fdist = FreqDist([len(w) for w in text1]) >>> fdist <FreqDist with 260819 outcomes> >>> fdist.keys() [3, 1, 4, 2, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 20] >>> We start by deriving a list of the lengths of words in text1 , and the FreqDist then counts the number of times each of these occurs . The result is a distribution containing a quarter of a million items, each of which is a number corresponding to a word token in the text. But there are only 20 distinct items being counted, the numbers 1 through 20, because there are only 20 different word lengths. I.e., there are words consisting of just 1 character, 2 characters, ..., 20 characters, but none with 21 or more characters. One might wonder how frequent the different lengths of words are (e.g., how many words of length 4 appear in the text, are there more words of length 5 than length 4, etc.). We can do this as follows: >>> fdist.items() [(3, 50223), (1, 47933), (4, 42345), (2, 38513), (5, 26597), (6, 17111), (7, 14399), (8, 9966), (9, 6428), (10, 3528), (11, 1873), (12, 1053), (13, 567), (14, 177), (15, 70), (16, 22), (17, 12), (18, 1), (20, 1)] >>> fdist.max() 3 >>> fdist[3] 50223 >>> fdist.freq(3) 0.19255882431878046 >>> From this we see that the most frequent word length is 3, and that words of length 3 account for roughly 50,000 (or 20%) of the words making up the book. Although we will not pursue it here, further analysis of word length might help us understand 1.3 Computing with Language: Simple Statistics | 21
  • 40. differences between authors, genres, or languages. Table 1-2 summarizes the functions defined in frequency distributions. Table 1-2. Functions defined for NLTK’s frequency distributions Example Description fdist = FreqDist(samples) Create a frequency distribution containing the given samples fdist.inc(sample) Increment the count for this sample fdist['monstrous'] Count of the number of times a given sample occurred fdist.freq('monstrous') Frequency of a given sample fdist.N() Total number of samples fdist.keys() The samples sorted in order of decreasing frequency for sample in fdist: Iterate over the samples, in order of decreasing frequency fdist.max() Sample with the greatest count fdist.tabulate() Tabulate the frequency distribution fdist.plot() Graphical plot of the frequency distribution fdist.plot(cumulative=True) Cumulative plot of the frequency distribution fdist1 < fdist2 Test if samples in fdist1 occur less frequently than in fdist2 Our discussion of frequency distributions has introduced some important Python con- cepts, and we will look at them systematically in Section 1.4. 1.4 Back to Python: Making Decisions and Taking Control So far, our little programs have had some interesting qualities: the ability to work with language, and the potential to save human effort through automation. A key feature of programming is the ability of machines to make decisions on our behalf, executing instructions when certain conditions are met, or repeatedly looping through text data until some condition is satisfied. This feature is known as control, and is the focus of this section. Conditionals Python supports a wide range of operators, such as < and >=, for testing the relationship between values. The full set of these relational operators are shown in Table 1-3. Table 1-3. Numerical comparison operators Operator Relationship < Less than <= Less than or equal to == Equal to (note this is two “=”signs, not one) 22 | Chapter 1: Language Processing and Python
  • 41. Operator Relationship != Not equal to > Greater than >= Greater than or equal to We can use these to select different words from a sentence of news text. Here are some examples—notice only the operator is changed from one line to the next. They all use sent7, the first sentence from text7 (Wall Street Journal). As before, if you get an error saying that sent7 is undefined, you need to first type: from nltk.book import *. >>> sent7 ['Pierre', 'Vinken', ',', '61', 'years', 'old', ',', 'will', 'join', 'the', 'board', 'as', 'a', 'nonexecutive', 'director', 'Nov.', '29', '.'] >>> [w for w in sent7 if len(w) < 4] [',', '61', 'old', ',', 'the', 'as', 'a', '29', '.'] >>> [w for w in sent7 if len(w) <= 4] [',', '61', 'old', ',', 'will', 'join', 'the', 'as', 'a', 'Nov.', '29', '.'] >>> [w for w in sent7 if len(w) == 4] ['will', 'join', 'Nov.'] >>> [w for w in sent7 if len(w) != 4] ['Pierre', 'Vinken', ',', '61', 'years', 'old', ',', 'the', 'board', 'as', 'a', 'nonexecutive', 'director', '29', '.'] >>> There is a common pattern to all of these examples: [w for w in text if condition], where condition is a Python “test” that yields either true or false. In the cases shown in the previous code example, the condition is always a numerical comparison. How- ever,wecanalsotestvariouspropertiesofwords,usingthefunctionslistedinTable1-4. Table 1-4. Some word comparison operators Function Meaning s.startswith(t) Test if s starts with t s.endswith(t) Test if s ends with t t in s Test if t is contained inside s s.islower() Test if all cased characters in s are lowercase s.isupper() Test if all cased characters in s are uppercase s.isalpha() Test if all characters in s are alphabetic s.isalnum() Test if all characters in s are alphanumeric s.isdigit() Test if all characters in s are digits s.istitle() Test if s is titlecased (all words in s have initial capitals) Here are some examples of these operators being used to select words from our texts: words ending with -ableness; words containing gnt; words having an initial capital; and words consisting entirely of digits. 1.4 Back to Python: Making Decisions and Taking Control | 23
  • 42. >>> sorted([w for w in set(text1) if w.endswith('ableness')]) ['comfortableness', 'honourableness', 'immutableness', 'indispensableness', ...] >>> sorted([term for term in set(text4) if 'gnt' in term]) ['Sovereignty', 'sovereignties', 'sovereignty'] >>> sorted([item for item in set(text6) if item.istitle()]) ['A', 'Aaaaaaaaah', 'Aaaaaaaah', 'Aaaaaah', 'Aaaah', 'Aaaaugh', 'Aaagh', ...] >>> sorted([item for item in set(sent7) if item.isdigit()]) ['29', '61'] >>> We can also create more complex conditions. If c is a condition, then not c is also a condition. If we have two conditions c1 and c2, then we can combine them to form a new condition using conjunction and disjunction: c1 and c2, c1 or c2. YourTurn: Run the following examples and try to explain what is going on in each one. Next, try to make up some conditions of your own. >>> sorted([w for w in set(text7) if '-' in w and 'index' in w]) >>> sorted([wd for wd in set(text3) if wd.istitle() and len(wd) > 10]) >>> sorted([w for w in set(sent7) if not w.islower()]) >>> sorted([t for t in set(text2) if 'cie' in t or 'cei' in t]) Operating on Every Element In Section 1.3, we saw some examples of counting items other than words. Let’s take a closer look at the notation we used: >>> [len(w) for w in text1] [1, 4, 4, 2, 6, 8, 4, 1, 9, 1, 1, 8, 2, 1, 4, 11, 5, 2, 1, 7, 6, 1, 3, 4, 5, 2, ...] >>> [w.upper() for w in text1] ['[', 'MOBY', 'DICK', 'BY', 'HERMAN', 'MELVILLE', '1851', ']', 'ETYMOLOGY', '.', ...] >>> These expressions have the form [f(w) for ...] or [w.f() for ...], where f is a function that operates on a word to compute its length, or to convert it to uppercase. For now, you don’t need to understand the difference between the notations f(w) and w.f(). Instead, simply learn this Python idiom which performs the same operation on every element of a list. In the preceding examples, it goes through each word in text1, assigning each one in turn to the variable w and performing the specified oper- ation on the variable. The notation just described is called a “list comprehension.” This is our first example of a Python idiom, a fixed notation that we use habitually without bothering to analyze each time. Mastering such idioms is an important part of becoming a fluent Python programmer. Let’s return to the question of vocabulary size, and apply the same idiom here: >>> len(text1) 260819 24 | Chapter 1: Language Processing and Python
  • 43. >>> len(set(text1)) 19317 >>> len(set([word.lower() for word in text1])) 17231 >>> Now that we are not double-counting words like This and this, which differ only in capitalization, we’ve wiped 2,000 off the vocabulary count! We can go a step further and eliminate numbers and punctuation from the vocabulary count by filtering out any non-alphabetic items: >>> len(set([word.lower() for word in text1 if word.isalpha()])) 16948 >>> This example is slightly complicated: it lowercases all the purely alphabetic items. Per- haps it would have been simpler just to count the lowercase-only items, but this gives the wrong answer (why?). Don’t worry if you don’t feel confident with list comprehensions yet, since you’ll see many more examples along with explanations in the following chapters. Nested Code Blocks Most programming languages permit us to execute a block of code when a conditional expression, or if statement, is satisfied. We already saw examples of conditional tests in code like [w for w in sent7 if len(w) < 4]. In the following program, we have created a variable called word containing the string value 'cat'. The if statement checks whether the test len(word) < 5 is true. It is, so the body of the if statement is invoked and the print statement is executed, displaying a message to the user. Remember to indent the print statement by typing four spaces. >>> word = 'cat' >>> if len(word) < 5: ... print 'word length is less than 5' ... word length is less than 5 >>> When we use the Python interpreter we have to add an extra blank line in order for it to detect that the nested block is complete. If we change the conditional test to len(word) >= 5, to check that the length of word is greater than or equal to 5, then the test will no longer be true. This time, the body of the if statement will not be executed, and no message is shown to the user: >>> if len(word) >= 5: ... print 'word length is greater than or equal to 5' ... >>> 1.4 Back to Python: Making Decisions and Taking Control | 25
  • 44. An if statement is known as a control structure because it controls whether the code in the indented block will be run. Another control structure is the for loop. Try the following, and remember to include the colon and the four spaces: >>> for word in ['Call', 'me', 'Ishmael', '.']: ... print word ... Call me Ishmael . >>> This is called a loop because Python executes the code in circular fashion. It starts by performing the assignment word = 'Call', effectively using the word variable to name the first item of the list. Then, it displays the value of word to the user. Next, it goes back to the for statement, and performs the assignment word = 'me' before displaying this new value to the user, and so on. It continues in this fashion until every item of the list has been processed. Looping with Conditions Now we can combine the if and for statements. We will loop over every item of the list, and print the item only if it ends with the letter l. We’ll pick another name for the variable to demonstrate that Python doesn’t try to make sense of variable names. >>> sent1 = ['Call', 'me', 'Ishmael', '.'] >>> for xyzzy in sent1: ... if xyzzy.endswith('l'): ... print xyzzy ... Call Ishmael >>> You will notice that if and for statements have a colon at the end of the line, before the indentation begins. In fact, all Python control structures end with a colon. The colon indicates that the current statement relates to the indented block that follows. We can also specify an action to be taken if the condition of the if statement is not met. Here we see the elif (else if) statement, and the else statement. Notice that these also have colons before the indented code. >>> for token in sent1: ... if token.islower(): ... print token, 'is a lowercase word' ... elif token.istitle(): ... print token, 'is a titlecase word' ... else: ... print token, 'is punctuation' ... Call is a titlecase word me is a lowercase word 26 | Chapter 1: Language Processing and Python
  • 45. Ishmael is a titlecase word . is punctuation >>> As you can see, even with this small amount of Python knowledge, you can start to build multiline Python programs. It’s important to develop such programs in pieces, testing that each piece does what you expect before combining them into a program. This is why the Python interactive interpreter is so invaluable, and why you should get comfortable using it. Finally, let’s combine the idioms we’ve been exploring. First, we create a list of cie and cei words, then we loop over each item and print it. Notice the comma at the end of the print statement, which tells Python to produce its output on a single line. >>> tricky = sorted([w for w in set(text2) if 'cie' in w or 'cei' in w]) >>> for word in tricky: ... print word, ancient ceiling conceit conceited conceive conscience conscientious conscientiously deceitful deceive ... >>> 1.5 Automatic Natural Language Understanding We have been exploring language bottom-up, with the help of texts and the Python programming language. However, we’re also interested in exploiting our knowledge of language and computation by building useful language technologies. We’ll take the opportunity now to step back from the nitty-gritty of code in order to paint a bigger picture of natural language processing. At a purely practical level, we all need help to navigate the universe of information locked up in text on the Web. Search engines have been crucial to the growth and popularity of the Web, but have some shortcomings. It takes skill, knowledge, and some luck, to extract answers to such questions as: What tourist sites can I visit between Philadelphia and Pittsburgh on a limited budget? What do experts say about digital SLR cameras? What predictions about the steel market were made by credible commentators in the past week? Getting a computer to answer them automatically involves a range of language processing tasks, including information extraction, inference, and summari- zation, and would need to be carried out on a scale and with a level of robustness that is still beyond our current capabilities. On a more philosophical level, a long-standing challenge within artificial intelligence has been to build intelligent machines, and a major part of intelligent behavior is un- derstanding language. For many years this goal has been seen as too difficult. However, as NLP technologies become more mature, and robust methods for analyzing unre- stricted text become more widespread, the prospect of natural language understanding has re-emerged as a plausible goal. 1.5 Automatic Natural Language Understanding | 27
  • 46. In this section we describe some language understanding technologies, to give you a sense of the interesting challenges that are waiting for you. Word Sense Disambiguation In word sense disambiguation we want to work out which sense of a word was in- tended in a given context. Consider the ambiguous words serve and dish: (2) a. serve: help with food or drink; hold an office; put ball into play b. dish: plate; course of a meal; communications device In a sentence containing the phrase: he served the dish, you can detect that both serve and dish are being used with their food meanings. It’s unlikely that the topic of discus- sion shifted from sports to crockery in the space of three words. This would force you to invent bizarre images, like a tennis pro taking out his frustrations on a china tea-set laid out beside the court. In other words, we automatically disambiguate words using context, exploiting the simple fact that nearby words have closely related meanings. As another example of this contextual effect, consider the word by, which has several meanings, for example, the book by Chesterton (agentive—Chesterton was the author of the book); the cup by the stove (locative—the stove is where the cup is); and submit by Friday (temporal—Friday is the time of the submitting). Observe in (3) that the meaning of the italicized word helps us interpret the meaning of by. (3) a. The lost children were found by the searchers (agentive) b. The lost children were found by the mountain (locative) c. The lost children were found by the afternoon (temporal) Pronoun Resolution A deeper kind of language understanding is to work out “who did what to whom,” i.e., to detect the subjects and objects of verbs. You learned to do this in elementary school, but it’s harder than you might think. In the sentence the thieves stole the paintings, it is easy to tell who performed the stealing action. Consider three possible following sen- tences in (4), and try to determine what was sold, caught, and found (one case is ambiguous). (4) a. The thieves stole the paintings. They were subsequently sold. b. The thieves stole the paintings. They were subsequently caught. c. The thieves stole the paintings. They were subsequently found. Answering this question involves finding the antecedent of the pronoun they, either thieves or paintings. Computational techniques for tackling this problem include ana- phora resolution—identifying what a pronoun or noun phrase refers to—and 28 | Chapter 1: Language Processing and Python
  • 47. semantic role labeling—identifying how a noun phrase relates to the verb (as agent, patient, instrument, and so on). Generating Language Output If we can automatically solve such problems of language understanding, we will be able to move on to tasks that involve generating language output, such as question answering and machine translation. In the first case, a machine should be able to answer a user’s questions relating to collection of texts: (5) a. Text: ... The thieves stole the paintings. They were subsequently sold. ... b. Human: Who or what was sold? c. Machine: The paintings. The machine’s answer demonstrates that it has correctly worked out that they refers to paintings and not to thieves. In the second case, the machine should be able to translate the text into another language, accurately conveying the meaning of the original text. In translating the example text into French, we are forced to choose the gender of the pronoun in the second sentence: ils (masculine) if the thieves are sold, and elles (fem- inine) if the paintings are sold. Correct translation actually depends on correct under- standing of the pronoun. (6) a. The thieves stole the paintings. They were subsequently found. b. Les voleurs ont volé les peintures. Ils ont été trouvés plus tard. (the thieves) c. Les voleurs ont volé les peintures. Elles ont été trouvées plus tard. (the paintings) In all of these examples, working out the sense of a word, the subject of a verb, and the antecedent of a pronoun are steps in establishing the meaning of a sentence, things we would expect a language understanding system to be able to do. Machine Translation For a long time now, machine translation (MT) has been the holy grail of language understanding, ultimately seeking to provide high-quality, idiomatic translation be- tween any pair of languages. Its roots go back to the early days of the Cold War, when the promise of automatic translation led to substantial government sponsorship, and with it, the genesis of NLP itself. Today, practical translation systems exist for particular pairs of languages, and some are integrated into web search engines. However, these systems have some serious shortcomings. We can explore them with the help of NLTK’s “babelizer” (which is automatically loaded when you import this chapter’s materials using from nltk.book import *). This program submits a sentence for translation into a specified language, 1.5 Automatic Natural Language Understanding | 29

Related Documents