Word and phrase coding

Word2Phrase in Apache Spark

This is the implementation of word2phrase (see section 4 Learning Phrases of http://arxiv.org/pdf/1310.4546.pdf).

The estimator will take a dataframe as a training set, and produce a model that can be used with the transformer pipeline.

Below is an example using the testSentences.scala file:

First we create our training dataset; it’s a dataframe where the occurrences «new york» and «test drive» appears frequently. (The sentences make no sense as they are randomly generated words. See below for the full dataframe.)

You can copy/paste this (be sure to include the full dataframe included in the repo) into your spark shell to test it.

First run spark-shell and download the word2phrase mvn package:

spark-shell --packages com.reputation.spark.word2phrase.1.0.1

Import the algorithm and create the dataframe:

import org.apache.spark.ml.feature.Word2Phrase

val wordDataFrame = sqlContext.createDataFrame(Seq(
(0, "new york test drive cool york how always learn media new york ."),
(1, "online york new york learn to media cool time ."),
(2, "media play how cool times play ."),
(3, "code to to code york to loaded times media ."),
(4, "play awesome to york ."),
.
.
.
(1099, "work please ideone how awesome times ."),
(1100, "play how play awesome to new york york awesome use new york work please loaded always like ."),
(1101, "learn like I media online new york ."),
(1102, "media follow learn code code there to york times ."),
(1103, "cool use play work please york cool new york how follow ."),
(1104, "awesome how loaded media use us cool new york online code judge ideone like ."),
(1105, "judge media times time ideone new york new york time us fun ."),
(1106, "new york to time there media time fun there new like media time time ."),
(1107, "awesome to new times learn cool code play how to work please to learn to ."),
(1108, "there work please online new york how to play play judge how always work please ."),
(1109, "fun ideone to play loaded like how ."),
(1110, "fun york test drive awesome play times ideone new us media like follow .")
)).toDF("label", "inputWords")

We set the input and output column names and create the model (the estimator step, represented by the fit(wordDataFrame) function).

val t = new Word2Phrase().setInputCol("inputWords").setOutputCol("out")

val model = t.fit(wordDataFrame)

We then use this model to transform our original dataframe sentences and view the results. Unfortunately you can’t see the entire row in the spark-shell, but in the out column it’s clear that all instances of «new york» and «test drive» have been transformed into «new_york» and «test_drive».

val bi_gram_data = model.transform(wordDataFrame)

bi_gram_data.show()
//this is the final result
+-----+--------------------+--------------------+
|label|          inputWords|                 out|
+-----+--------------------+--------------------+
|    0|new york test dri...|new_york test_dri...|
|    1|online york new y...|online york new_y...|
|    2|media play how co...|media play how co...|
|    3|code to to code y...|code to to code y...|
|    4|play awesome to y...|play awesome to y...|
|    5|   like I I always .|   like I I always .|
|    6|how to there lear...|how to there lear...|
|    7|judge time us pla...|judge time us pla...|
|    8|judge test drive ...|judge test_drive ...|
|    9|judge follow fun ...|judge follow fun ...|
|   10|how I follow ideo...|how I follow ideo...|
|   11|use use learn I t...|use use learn I t...|
|   12|us new york alway...|us new_york alway...|
|   13|there always how ...|there always how ...|
|   14|always time media...|always time media...|
|   15|how test drive to...|how test_drive to...|
|   16|cool us online ti...|cool us online ti...|
|   17|follow time aweso...|follow time aweso...|
|   18|us york test driv...|us york test_driv...|
|   19|use fun new york ...|use fun new_york ...|
+-----+--------------------+--------------------+
only showing top 20 rows

See the blog post for more details on word2phrase:

http://tech.reputation.com/meaningless-words-to-useful-phrases-in-spark-word2phrase/ ‎

Ekaterina Chernyak, Dmitry Ilvovsky

Содержание

  • 1 Part of speech (POS)
  • 2 POS ambiguation
  • 3 POS taggers
    • 3.1 NLTK POS default tagger
  • 4 Exercise 3.1 Genre comparison
  • 5 Key word and phrase extraction
  • 6 Supervised methods for key word and phrase extraction
  • 7 Unsupervised methods for key word and phrase extraction from a single text
    • 7.1 Exercise 3.2
  • 8 Bigram association measures
    • 8.1 Pointwise Mutual Information [Manning, Shuetze, 1999]
    • 8.2 T-Score [Manning, Shuetze, 1999]
    • 8.3 Chi-squared [Manning, Shuetze, 1999]
    • 8.4 Bigram association measures in NLTK
      • 8.4.1 NLTK BigramCollocationFinder
  • 9 TextRank: using graph centrality measures for key word and phrase extraction [Mihalcea, Tarau, 2004]
    • 9.1 PageRank [Brin, Page, 1998]
    • 9.2 HITS [Kleinberg, 1999]
    • 9.3 Exercise 3.3
  • 10 Unsupervised methods for key word and phrase selection from a text in a collection
    • 10.1 Term frequency [Luhn, 1957]
    • 10.2 Inverse document frequency [Spaerck Jones, 1972]
  • 11 Variants of TF and IDF weights
  • 12 TF-IDF in NLTK
    • 12.1 NLTK TextCollection class
    • 12.2 Exercise 3.4
  • 13 TF-IDF alternatives
    • 13.1 Exercise 3.5
  • 14 Using TF-IDF to measure text similarity

Part of speech (POS)

Part of speech [Manning, Shuetze, 1999]

Words of a language are grouped into classes which show similar syntactic behavior. These word classes are called parts of speech (POS). Three
important parts of speech are noun, verb, and adjective. The major types of morphological process are in ection, derivation, and compounding.

There are around 9 POS according to different schools:

  • Nouns (NN, NP), pronouns (PN, PRP), adjectives (JJ): number, gender, case
  • Adjective (JJ): comparative, superlative, short form
  • Verbs (VB): subject number, subject person, tense, aspect, modality, participles, voice
  • Adverbs (RB), prepositions (IN), conjunctions (, CS), articles (AT)

and particles (RP): nothing

POS ambiguation

Ship (noun or verb?)

  • a luxury cruise ship
  • Both products are due to ship at the beginning of June
  • A new engine was shipped over from the US
  • The port is closed to all shipping

Contest (noun or verb?)

  • Stone decided to hold a contest to see who could write the best song.
  • She plans to contest a seat in Congress next year.

POS taggers

  • Corpus- or dictionary-based VS rule-based
  • Ngram-based taggers:
    • unigram tagging: assign the most frequent tag
    • ngram tagging: look at the context of n previous words (requires a lot of training data)
  • Trade-off between the accuracy and the coverage: combine different taggers

NLTK POS default tagger

In[1]: from nltk.tag import pos tag

In[2]: print pos tag(['ship'])

Out[1]: [('ship', 'NN')]

In[3]: print pos tag(['shipping'])

Out[2]: [('shipping', 'VBG')]

See [1] for more details on
learning taggers.

Exercise 3.1 Genre comparison

Text genre [Santini, Sharoff, 2009]

The concept of genre is hard to agree upon. Many interpretations have been proposed since Aristotles Poetics without reaching any definite
conclusions about the inventory or even principles for classifying documents into genres. The lack of an agreed definition of what genre is causes the problem of the loose boundaries between the term genre» with other neighbouring terms, such as «register», «domain», «topic», and «style».

Exercise 3.1

Input: Two texts of different genre (for example, Wikipedia article and
blog post)
Output: rank all of POS tags for both texts

How can you describe the difference between two genres?

Key word and phrase extraction

There are many definitions of key word and phrase. Thus there are many methods for their extraction:

  • supervised VS unsupervised
  • frequency-based VS more complex
  • from individual text VS from text collection
  • word (unigram) VS bigram VS ngram
  • term VS named entity VS collocation
  • sequential words VS using window

Supervised methods for key word and phrase extraction

I am a word. Am I a key word? Let us build a classifier.

  • Am I in the beginning or in the end of the sentence?
  • Am I capitalized?
  • How many times do I occur?
  • Am I used in Wikipedia as a title of a category or an article?
  • Am I a term?
  • Am I a NE?
  • etc.

But we need a collection of marked up texts!

Unsupervised methods for key word and phrase extraction from a single text

  • POS patterns
  • Association measures: PMI, T-Score, LLR
  • Graph methods: TextRank [Mihalcea, Tarau, 2004]
  • Syntactic patterns

Exercise 3.2

Input: sif1.txt (or your own text)

Key word: top n1 NN

Key phrase: top n2 phrases, that satisfy the following patterns: JJ + NN, NN + NN, NN + IN + NN

Output: list of key words and phrases

Hint: use nltk.ngrams to get ngrams.

Bigram association measures

Pointwise Mutual Information [Manning, Shuetze, 1999]

PMI measures the reduction of uncertainty about the occurrence of one
word when we are told about the occurrence of the other one.

L3 p1.jpg

T-Score [Manning, Shuetze, 1999]

T-Score is a statistical t-test applied to finding collocations. The t-test
looks at the difference between the observed and expected means, scaled
by the variance of the data. The T-score is most useful as a method for
ranking collocations. The level of significance itself is less useful.

L3 p2.jpg

L3 p3.jpg

But Chi-squared has many other interesting applications.

Chi-squared [Manning, Shuetze, 1999]

In general, for the problem of finding collocation, the difference between the T-score and the Chi-squared does not seem to be large.

Bigram association measures in NLTK

NLTK BigramCollocationFinder

In[1]: from nltk.collocations import *

In[2]: bigram measures = nltk.collocations.BigramAssocMeasures()

In[3]: finder = BigramCollocationFinder.from words(tokens)

In[4]: finder.apply freq filter(3)

In[5]: for i in finder.nbest(bigram measures.pmi, 20):

...

Bigram measures:

  • bigram measures.pmi
  • bigram measures.student_t
  • bigram measures.chi_sq
  • igram measures.likelihood_ratio

See [2] for more more bigram association measures.

TextRank: using graph centrality measures for key word and phrase extraction [Mihalcea, Tarau, 2004]

  1. Add words as vertices in the graph.
  2. Identify relations that connect words:
    1. consequent words;
    2. words inside (left or right) the window (2-5 words); and use these relations to draw edges between vertices in the graph. Edges can be directed or undirected, weighted or unweighted.
  3. Iterate the graph-based ranking algorithm until convergence (for example, PageRank).
  4. Sort vertices based on their final score. Use the values attached to each vertex for ranking/selection decisions.
  5. If two adjacent words are selected as potential keywords by TextRank, collapse them into one single key phrase.

See original paper: [3]

Compatibility of systems of linear constraints over the set of natural numbers. Criteria of compatibility of a system of linear Diophantine equations, strict inequations, and nonstrict inequations are considered. Upper bounds for components of a minimal set of solutions and algorithms of construction of minimal generating sets of solutions for all types of systems are given. These criteria and the corresponding algorithms for constructing a minimal supporting set of solutions can be used in solving all the considered types systems and systems of mixed types.

TextRank

G = (V, E) — a graph, V — vertices, E — edges
In(Vi) — the set of vertices that point to it
Out(Vi) — the set of vertices that Vi points to
Graph centrality measures:

L3 p5.jpg

HITS [Kleinberg, 1999]

L3 p6.jpg

Key words and phrases, assigned by TextRank using PageRank centrality measure: linear constraints; linear diophantine equations; natural numbers; nonstrict inequations; strict inequations; upper bounds

Exercise 3.3

Input: sif.txt (or your own text)
Output: key words and phrases computed by PageRank (using PR(Vi),HITSA(Vi) and HITSH(Vi) as centrality measures)

Hint: use NetworkX for PageRank and HITS [4]

Unsupervised methods for key word and phrase selection from a text in a collection

The problem: given a collection of texts find those words and phrases (terms) that occur in this text significant frequently than in other texts.

Term frequency [Luhn, 1957]

The weight of a term that occurs in a document is simply proportional to the term frequency.

Inverse document frequency [Spaerck Jones, 1972]

The specificity of a term can be quantified as an inverse function of the number of documents in which it occurs.

tfidf (term, text, collection) = tf (term, document) × idf (term, collection)

Variants of TF and IDF weights

L3 p7.jpg

TF-IDF in NLTK

NLTK TextCollection class

In[1]: from nltk.text import TextCollection

In[2]: collection = [WhitespaceTokenizer().tokenize(text) for text in collection]

In[3]: corpus = TextCollection(collection)

In[4]: for i in collection[0]: print i, corpus.tf idf(i,collection[0])

Exercise 3.4

Input: sif2.txt (or your own collection of texts)

Output: list of key words and phrases according to TF-IDF for one text

Hint: use sorted(mylist,key=lambda l:l[1], reverse=True) to sort mylist in descending order based on the second parameter

TF-IDF alternatives

Exercise 3.5

Write MI and Chi-squared as an alternative for TF-IDF for measuring significance of a term in a text in a collection.
Check your ideas in Sebastiani[5], 2001

Using TF-IDF to measure text similarity

L3p8.jpg

I would start with a jagged array of character pointers like this

char *phrase1[] = { "Good", "morning", "Mr.", "", "", "you", "have", "", "$", "in", "your", "account.", NULL };
char *phrase2[] = { "Member", "", "", "has", "chosen", "", "from", "the", "list.", NULL };
char **phrases[] = { phrase1, phrase2, NULL };

In each phrase, there are

  • words that must be matched exactly, e.g. "morning"
  • empty strings "" that mark the location of dynamic items
  • NULL pointers that mark the end of the phrase

When using the array, phrases[p][i] is the i’th word in phrase p,
and phrases[p][i][0] is the first character in the i’th word in phrase p.

Hence, the code to check for a dynamic item is

if ( phrases[p][i][0] == '' )
    // this is a dynamic item

To check for the end of the phrase

if ( phrases[p][i] == NULL )
    // this is the end of the phrase

Otherwise, to compare the word

if ( strcmp( phrases[p][i], word ) == 0 )
    // the word matches 

Word and phrase tables are key inputs to machine translations, but costly to
produce. New unsupervised learning methods represent words and phrases in a
high-dimensional vector space, and these monolingual embeddings have been shown
to encode syntactic and semantic relationships between language elements. The
information captured by these embeddings can be exploited for bilingual
translation by learning a transformation matrix that allows matching relative
positions across two monolingual vector spaces. This method aims to identify
high-quality candidates for word and phrase translation more cost-effectively
from unlabeled data.
This paper expands the scope of previous attempts of bilingual translation to
four languages (English, German, Spanish, and French). It shows how to process
the source data, train a neural network to learn the high-dimensional
embeddings for individual languages and expands the framework for testing their
quality beyond the English language. Furthermore, it shows how to learn
bilingual transformation matrices and obtain candidates for word and phrase
translation, and assess their quality.


PDF


Abstract

Code


Tasks


Datasets





 

Add Datasets

introduced or used in this paper

Results from the Paper



 
Submit

results from this paper

to get state-of-the-art GitHub badges and help the
community compare results to other papers.

Methods


No methods listed for this paper. Add

relevant methods here

Training a Word2Vec model with phrases is very similar to training a Word2Vec model with single words. The difference: you would need to add a layer of intelligence in processing your text data to pre-discover phrases. In this tutorial, you will learn how to create embeddings with phrases without explicitly specifying the number of words that should make-up a phrase (i.e. the n-gram size). This means that you could have phrases with 2 words, 3 words and in some rare cases even 4 or 5.

At a high level, the steps would include:

  • Step 1:  Discovering common phrases in your corpora
  • Step 2: Tagging your corpora with phrases
  • Step 3: Training a Word2Vec model with the newly found phrases

Step 1: Discovering common phrases in your corpora

The first step towards generating embeddings for phrases is recognizing groups of words that make up a phrase. There are many ways to recognize phrases. One way is to use a linguistic heavy approach called “chunking” to detect phrases. NLTK for example, has a chunk capability that you could use.

For this task, I will show you how you can use a text data mining approach with Spark, where you leverage the volume and evidence from your corpora for phrase detection. I like this approach because it’s lightweight, speedy and scales to the amount of data that you need to process.

So here’s how it works. At a high level, the entire corpora of text is segmented using a set of delimiter tokens. This can be special characters, stop words and other terms that can indicate phrase boundary. I specifically used some special characters and a very basic set of English stop words.

Stop words are excellent for splitting text into a set of phrases as they usually consist of connector and filler words used to connect ideas, details, or clauses together in order to make one clear, detailed sentence. You can get creative and use a more complete stop word list or you can even over-simplify this list to make it a minimal stop word list.

The code below shows you how you can use both special characters and stop words to break text into a set of candidate phrases. Check the phrase-at-scale repo for the full source code.

def generate_candidate_phrases(text, stopwords):
    """ generate phrases using phrase boundary markers """

    # generate approximate phrases with punctation
    coarse_candidates = char_splitter.split(text.lower())

    candidate_phrases = []

    for coarse_phrase
            in coarse_candidates:

        words = re.split("\s+", coarse_phrase)
        previous_stop = False

        # examine each word to determine if it is a phrase boundary marker or part of a phrase or lone ranger
        for w in words:

            if w in stopwords and not previous_stop:
                # phrase boundary encountered, so put a hard indicator
                candidate_phrases.append(";")
                previous_stop = True
            elif w not in stopwords and len(w) > 3:
                # keep adding words to list until a phrase boundary is detected
                candidate_phrases.append(w.strip())
                previous_stop = False

    # get a list of candidate phrases without boundary demarcation
    phrases = re.split(";+", ' '.join(candidate_phrases))

    return phrases

In the code above, we are first splitting text into coarse-grained units using some special characters like comma, period and semi-colon. This is then followed by more fine-grained boundary detection using stop words. When you repeat this process for all documents or sentences in your corpora, you will end up with a huge set of phrases. You can then surface the top phrases using frequency counts and other measures such as Pointwise Mutual Information which can measure strength of association between words in your phrase. For the phrase embedding task, we naturally have to use lots and lots of data, so frequency counts alone would suffice for this task. In some other tasks, I have combined frequency counts with Pointwise Mutual Information to get a better measure phrase quality.

To ensure scalability, I really like using Spark since you can leverage its built-in multi-threading capability on a single machine or use multiple machines to get more CPU power if you really have massive amounts of data to process. The code below shows you the PySpark method that reads your text files, cleans it up, generates candidate phrases, counts frequency of the phrases and filters it down to a set of phrases that satisfy a minimum frequency count. On a 450 MB dataset, run locally, this takes about a minute to discover top phrases and 7 minutes to annotate the entire text corpora with phrases. You can follow instructions in the phrase-at-scale repo to use this PySpark code to discover phrases for your data.

def generate_and_tag_phrases(text_rdd,min_phrase_count=50):
    """Find top phrases, tag corpora with those top phrases"""

    # load stop words for phrase boundary marking
    logging.info ("Loading stop words...")
    stopwords = load_stop_words ()

    # get top phrases with counts > min_phrase_count
    logging.info ("Generating and collecting top phrases...")
    top_phrases_rdd = 
        text_rdd.map(lambda txt: remove_special_characters(txt))
        .map(lambda txt: generate_candidate_phrases(txt, stopwords)) 
        .flatMap(lambda phrases: phrase_to_counts(phrases)) 
        .reduceByKey(add) 
        .sortBy(lambda phrases: phrases[1], ascending=False) 
        .filter(lambda phrases: phrases[1] >= min_phrase_count) 
        .sortBy(lambda phrases: phrases[0], ascending=True) 
        .map(lambda phrases: (phrases[0], phrases[0].replace(" ", "_")))

    shortlisted_phrases = top_phrases_rdd.collectAsMap()
    logging.info("Done with phrase generation...")

    # write phrases to file which you can use down the road to tag your text
    logging.info("Saving top phrases to {0}".format(phrases_file))
    with open(os.path.join(abspath, phrases_file), "w") as f:
        for phrase in shortlisted_phrases:
            f.write(phrase)
            f.write("n")

    # tag corpora and save as new corpora
    logging.info("Tagging corpora with phrases...this will take a while")
    tagged_text_rdd = text_rdd.map(
            lambda txt: tag_data(
                txt,
                shortlisted_phrases))

    return tagged_text_rdd

Here is a tiny snapshot of phrases found using the code above on a restaurant review dataset.

Step 2: Tagging your corpora with phrases

There are two ways you can mark certain words as phrases in your corpora. One approach is to pre-annotate your entire corpora and generate a new “annotated corpora”. The other way is to annotate your sentences or documents during the pre-processing phase prior to learning the embeddings. It’s much cleaner to have a separate layer for annotation which does not interfere with the training phase. Otherwise, it will be harder to gauge if your model is slow due to training or annotation.

In annotating your corpora, all you need to do is to somehow join the words that make-up a phrase. For this task, I just use an underscore to join the individual words. So, “…ate fried chicken and onion rings…” would become “…ate fried_chicken and onion_rings…”

Step 3: Training a Phrase2Vec model using Word2Vec

Once you have phrases explicitly tagged in your corpora the training phase is quite similar to any Word2Vec model with Gensim or any other library. You can follow my Word2Vec Gensim Tutorial for a full example on how to train and use Word2Vec.

Example Usage of Phrase Embeddings

The examples below show you the power of phrase embeddings when used to find similar concepts.  These are concepts from the restaurant domain, trained on 450 MB worth of restaurant reviews using Gensim.

Similar and related unigrams, bigrams and trigrams

Notice below that we are able to capture highly related concepts that are unigrams, bigrams and higher order n-grams.

Most similar to 'green_curry':
--------------------------------------
('panang_curry', 0.8900948762893677)
('yellow_curry', 0.884008526802063)
('panang', 0.8525004386901855)
('drunken_noodles', 0.850254237651825)
('basil_chicken', 0.8400430679321289)
('coconut_soup', 0.8296557664871216)
('massaman_curry', 0.827597975730896)
('pineapple_fried_rice', 0.8266736268997192)


Most similar to 'singapore_noodles':
--------------------------------------
('shrimp_fried_rice', 0.7932361960411072)
('drunken_noodles', 0.7914629578590393)
('house_fried_rice', 0.7901676297187805)
('mongolian_beef', 0.7796567678451538)
('crab_rangoons', 0.773795485496521)
('basil_chicken', 0.7726351022720337)
('crispy_beef', 0.7671589255332947)
('steamed_dumplings', 0.7614079117774963)


Most similar to 'chicken_tikka_masala':
--------------------------------------
('korma', 0.8702514171600342)
('butter_chicken', 0.8668922781944275)
('tikka_masala', 0.8444720506668091)
('garlic_naan', 0.8395442962646484)
('lamb_vindaloo', 0.8390569686889648)
('palak_paneer', 0.826908528804779)
('chicken_biryani', 0.8210495114326477)
('saag_paneer', 0.8197864294052124)


Most similar to 'breakfast_burrito':
--------------------------------------
('huevos_rancheros', 0.8463341593742371)
('huevos', 0.789624035358429)
('chilaquiles', 0.7711247801780701)
('breakfast_sandwich', 0.7659544944763184)
('rancheros', 0.7541004419326782)
('omelet', 0.7512155175209045)
('scramble', 0.7490915060043335)
('omlet', 0.747859001159668)

Most similar to 'little_salty':
--------------------------------------
('little_bland', 0.745500385761261)
('little_spicy', 0.7443351149559021)
('little_oily', 0.7373550534248352)
('little_overcooked', 0.7355216145515442)
('kinda_bland', 0.7207454442977905)
('slightly_overcooked', 0.712611973285675)
('little_greasy', 0.6943882703781128)
('cooked_nicely', 0.6860566139221191)


Most similar to 'celiac_disease':
--------------------------------------
('celiac', 0.8376057744026184)
('intolerance', 0.7442486882209778)
('gluten_allergy', 0.7399739027023315)
('celiacs', 0.7183824181556702)
('intolerant', 0.6730632781982422)
('gluten_free', 0.6726624965667725)
('food_allergies', 0.6587174534797668)
('gluten', 0.6406026482582092)

Similar concepts expressed differently

Here you will see that similar concepts that are expressed differently can also be captured.

Most similar to 'reasonably_priced':
--------------------------------------
('fairly_priced', 0.8588327169418335)
('affordable', 0.7922118306159973)
('inexpensive', 0.7702735066413879)
('decently_priced', 0.7376087307929993)
('reasonable_priced', 0.7328246831893921)
('priced_reasonably', 0.6946456432342529)
('priced_right', 0.6871092915534973)
('moderately_priced', 0.6844340562820435)


Most similar to 'highly_recommend':
--------------------------------------
('definitely_recommend', 0.9155156016349792)
('strongly_recommend', 0.86533123254776)
('absolutely_recommend', 0.8545517325401306)
('totally_recommend', 0.8534528017044067)
('recommend', 0.8257364630699158)
('certainly_recommend', 0.785507082939148)
('highly_reccomend', 0.7751532196998596)
('highly_recommended', 0.7553941607475281)

Summary

In summary, to generate embeddings of phrases, you would need to add a layer for phrase discovery before training a Word2Vec model. If you have lots of data, a text data mining approach has the benefit of being lightweight and scalable, without compromising on quality. In addition, you wouldn’t have to specify a phrase size in advance or be limited by a specific vocabulary. A linguistic heavy approach gives you a lot more specificity in terms of parts of speech and the types of phrases (e.g. noun phrase vs. verb phrase) that you are dealing with. If you really need that information, then you can consider a chunking approach over a text mining approach.

Resources

Here are some resources that might come handy to you:

  • Source code for discovering phrases at scale
  • Gensim Word2Vec Tutorial – Full Working Example
  • Understanding stop words
  • Constructing a domain specific stop word list

Back in elementary school you learnt the difference between nouns, verbs,
adjectives, and adverbs. These «word classes» are not just
the idle invention of grammarians, but are useful categories for many
language processing tasks. As we will see, they arise from simple analysis
of the distribution of words in text. The goal of this chapter is to
answer the following questions:

  1. What are lexical categories and how are they used in natural language processing?
  2. What is a good Python data structure for storing words and their categories?
  3. How can we automatically tag each word of a text with its word class?

Along the way, we’ll cover some fundamental techniques in NLP, including
sequence labeling, n-gram models, backoff, and evaluation. These techniques
are useful in many areas, and tagging gives us a simple context in which
to present them. We will also see how tagging is the second step in the typical
NLP pipeline, following tokenization.

The process of classifying words into their parts of speech and
labeling them accordingly is known as part-of-speech tagging,
POS-tagging, or simply tagging. Parts of speech
are also known as word classes or lexical categories.
The collection of tags
used for a particular task is known as a tagset. Our emphasis
in this chapter is on exploiting tags, and tagging text automatically.

1   Using a Tagger

A part-of-speech tagger, or POS-tagger, processes a sequence of words, and attaches a
part of speech tag to each word (don’t forget to import nltk):

 
>>> text = word_tokenize("And now for something completely different")
>>> nltk.pos_tag(text)
[('And', 'CC'), ('now', 'RB'), ('for', 'IN'), ('something', 'NN'),
('completely', 'RB'), ('different', 'JJ')]

Here we see that and is CC, a coordinating conjunction;
now and completely are RB, or adverbs;
for is IN, a preposition;
something is NN, a noun; and
different is JJ, an adjective.

Note

NLTK provides documentation for each tag, which can be queried using
the tag, e.g. nltk.help.upenn_tagset(‘RB’), or a regular
expression, e.g. nltk.help.upenn_tagset(‘NN.*’).
Some corpora have README files with tagset documentation,
see nltk.corpus.???.readme(), substituting in the name
of the corpus.

Let’s look at another example, this time including some homonyms:

 
>>> text = word_tokenize("They refuse to permit us to obtain the refuse permit")
>>> nltk.pos_tag(text)
[('They', 'PRP'), ('refuse', 'VBP'), ('to', 'TO'), ('permit', 'VB'), ('us', 'PRP'),
('to', 'TO'), ('obtain', 'VB'), ('the', 'DT'), ('refuse', 'NN'), ('permit', 'NN')]

Notice that refuse and permit both appear as a
present tense verb (VBP) and a noun (NN).
E.g. refUSE is a verb meaning «deny,» while REFuse is
a noun meaning «trash» (i.e. they are not homophones).
Thus, we need to know which word is being used in order to pronounce
the text correctly. (For this reason,
text-to-speech systems usually perform POS-tagging.)

Note

Your Turn:
Many words, like ski and race, can be used as nouns
or verbs with no difference in pronunciation. Can you think of
others? Hint: think of a commonplace object and try to put
the word to before it to see if it can also be a verb, or
think of an action and try to put the before it to see if
it can also be a noun. Now make up a sentence with both uses
of this word, and run the POS-tagger on this sentence.

Lexical categories like «noun» and part-of-speech tags like NN seem to have
their uses, but the details will be obscure to many readers. You might wonder what
justification there is for introducing this extra level of information.
Many of these categories arise from superficial analysis the distribution
of words in text. Consider the following analysis involving
woman (a noun), bought (a verb),
over (a preposition), and the (a determiner).
The text.similar() method takes a word w, finds all contexts
w1w w2,
then finds all words w’ that appear in the same context,
i.e. w1w’w2.

 
>>> text = nltk.Text(word.lower() for word in nltk.corpus.brown.words())
>>> text.similar('woman')
Building word-context index...
man day time year car moment world family house boy child country job
state girl place war way case question
>>> text.similar('bought')
made done put said found had seen given left heard been brought got
set was called felt in that told
>>> text.similar('over')
in on to of and for with from at by that into as up out down through
about all is
>>> text.similar('the')
a his this their its her an that our any all one these my in your no
some other and

Observe that searching for woman finds nouns;
searching for bought mostly finds verbs;
searching for over generally finds prepositions;
searching for the finds several determiners.
A tagger can correctly identify the tags on these words
in the context of a sentence, e.g. The woman bought over $150,000
worth of clothes
.

A tagger can also model our knowledge of unknown words,
e.g. we can guess that scrobbling is probably a verb,
with the root scrobble,
and likely to occur in contexts like he was scrobbling.

2   Tagged Corpora

2.1   Representing Tagged Tokens

By convention in NLTK, a tagged token is represented using a
tuple consisting of the token and the tag.
We can create one of these special tuples from the standard string
representation of a tagged token, using the function str2tuple():

 
>>> tagged_token = nltk.tag.str2tuple('fly/NN')
>>> tagged_token
('fly', 'NN')
>>> tagged_token[0]
'fly'
>>> tagged_token[1]
'NN'

We can construct a list of tagged tokens directly from a string. The first
step is to tokenize the string
to access the individual word/tag strings, and then to convert
each of these into a tuple (using str2tuple()).

 
>>> sent = '''
... The/AT grand/JJ jury/NN commented/VBD on/IN a/AT number/NN of/IN
... other/AP topics/NNS ,/, AMONG/IN them/PPO the/AT Atlanta/NP and/CC
... Fulton/NP-tl County/NN-tl purchasing/VBG departments/NNS which/WDT it/PPS
... said/VBD ``/`` ARE/BER well/QL operated/VBN and/CC follow/VB generally/RB
... accepted/VBN practices/NNS which/WDT inure/VB to/IN the/AT best/JJT
... interest/NN of/IN both/ABX governments/NNS ''/'' ./.
... '''
>>> [nltk.tag.str2tuple(t) for t in sent.split()]
[('The', 'AT'), ('grand', 'JJ'), ('jury', 'NN'), ('commented', 'VBD'),
('on', 'IN'), ('a', 'AT'), ('number', 'NN'), ... ('.', '.')]

2.2   Reading Tagged Corpora

Several of the corpora included with NLTK have been tagged for
their part-of-speech. Here’s an example of what you might see if you
opened a file from the Brown Corpus with a text editor:

The/at Fulton/np-tl County/nn-tl Grand/jj-tl Jury/nn-tl
said/vbd Friday/nr an/at investigation/nn of/in Atlanta’s/np$
recent/jj primary/nn election/nn produced/vbd / no/at
evidence/nn »/» that/cs any/dti irregularities/nns took/vbd
place/nn ./.

Other corpora use a variety of formats for storing part-of-speech tags.
NLTK’s corpus readers provide a uniform interface so that you
don’t have to be concerned with the different file formats.
In contrast with the file fragment shown above,
the corpus reader for the Brown Corpus represents the data as shown below.
Note that part-of-speech tags have been converted to uppercase, since this has
become standard practice since the Brown Corpus was published.

 
>>> nltk.corpus.brown.tagged_words()
[('The', 'AT'), ('Fulton', 'NP-TL'), ...]
>>> nltk.corpus.brown.tagged_words(tagset='universal')
[('The', 'DET'), ('Fulton', 'NOUN'), ...]

Whenever a corpus contains tagged text, the NLTK corpus interface
will have a tagged_words() method.
Here are some more examples, again using the output format
illustrated for the Brown Corpus:

 
>>> print(nltk.corpus.nps_chat.tagged_words())
[('now', 'RB'), ('im', 'PRP'), ('left', 'VBD'), ...]
>>> nltk.corpus.conll2000.tagged_words()
[('Confidence', 'NN'), ('in', 'IN'), ('the', 'DT'), ...]
>>> nltk.corpus.treebank.tagged_words()
[('Pierre', 'NNP'), ('Vinken', 'NNP'), (',', ','), ...]

Not all corpora employ the same set of tags; see the
tagset help functionality and the readme() methods
mentioned above for documentation.
Initially we want to avoid the complications of these tagsets,
so we use a built-in mapping to the «Universal Tagset»:

 
>>> nltk.corpus.brown.tagged_words(tagset='universal')
[('The', 'DET'), ('Fulton', 'NOUN'), ...]
>>> nltk.corpus.treebank.tagged_words(tagset='universal')
[('Pierre', 'NOUN'), ('Vinken', 'NOUN'), (',', '.'), ...]

Tagged corpora for several other languages are distributed with NLTK,
including Chinese, Hindi, Portuguese, Spanish, Dutch and Catalan.
These usually contain non-ASCII text,
and Python always displays this in hexadecimal when printing a larger structure
such as a list.

 
>>> nltk.corpus.sinica_treebank.tagged_words()
[('ä', 'Neu'), ('åæ', 'Nad'), ('åç', 'Nba'), ...]
>>> nltk.corpus.indian.tagged_words()
[('মহিষের', 'NN'), ('সন্তান', 'NN'), (':', 'SYM'), ...]
>>> nltk.corpus.mac_morpho.tagged_words()
[('Jersei', 'N'), ('atinge', 'V'), ('mxe9dia', 'N'), ...]
>>> nltk.corpus.conll2002.tagged_words()
[('Sao', 'NC'), ('Paulo', 'VMI'), ('(', 'Fpa'), ...]
>>> nltk.corpus.cess_cat.tagged_words()
[('El', 'da0ms0'), ('Tribunal_Suprem', 'np0000o'), ...]

If your environment is set up correctly, with appropriate editors and fonts,
you should be able to display individual strings in a human-readable way.
For example, 2.1 shows data accessed using
nltk.corpus.indian.

../images/tag-indian.png

Figure 2.1: POS-Tagged Data from Four Indian Languages: Bangla, Hindi, Marathi, and Telugu

If the corpus is also segmented into sentences, it will have
a tagged_sents() method that divides up the tagged words into
sentences rather than presenting them as one big list.
This will be useful when we come to developing automatic taggers,
as they are trained and tested on lists of sentences, not words.

2.4   Nouns

Nouns generally refer to people, places, things, or concepts, e.g.:
woman, Scotland, book, intelligence. Nouns can appear after
determiners and adjectives, and can be the subject or object of the
verb, as shown in 2.2.

Table 2.2:

Syntactic Patterns involving some Nouns

Word After a determiner Subject of the verb
woman the woman who I saw yesterday … the woman sat down
Scotland the Scotland I remember as a child … Scotland has five million people
book the book I bought yesterday … this book recounts the colonization of Australia
intelligence the intelligence displayed by the child … Mary’s intelligence impressed her teachers

The simplified noun tags are N for common nouns like book,
and NP for proper nouns like Scotland.

Let’s inspect some tagged text to see what parts of speech occur before a noun,
with the most frequent ones first. To begin with, we construct a list
of bigrams whose members are themselves word-tag pairs such as
((‘The’, ‘DET’), (‘Fulton’, ‘NP’)) and ((‘Fulton’, ‘NP’), (‘County’, ‘N’)).
Then we construct a FreqDist from the tag parts of the bigrams.

 
>>> word_tag_pairs = nltk.bigrams(brown_news_tagged)
>>> noun_preceders = [a[1] for (a, b) in word_tag_pairs if b[1] == 'NOUN']
>>> fdist = nltk.FreqDist(noun_preceders)
>>> [tag for (tag, _) in fdist.most_common()]
['NOUN', 'DET', 'ADJ', 'ADP', '.', 'VERB', 'CONJ', 'NUM', 'ADV', 'PRT', 'PRON', 'X']

This confirms our assertion that nouns occur after determiners and
adjectives, including numeral adjectives (tagged as NUM).

2.5   Verbs

Verbs are words that describe events and actions, e.g. fall,
eat in 2.3.
In the context of a sentence, verbs typically express a relation
involving the referents of one or more noun phrases.

Table 2.3:

Syntactic Patterns involving some Verbs

Word Simple With modifiers and adjuncts (italicized)
fall Rome fell Dot com stocks suddenly fell like a stone
eat Mice eat cheese John ate the pizza with gusto

What are the most common verbs in news text? Let’s sort all the verbs by frequency:

 
>>> wsj = nltk.corpus.treebank.tagged_words(tagset='universal')
>>> word_tag_fd = nltk.FreqDist(wsj)
>>> [wt[0] for (wt, _) in word_tag_fd.most_common() if wt[1] == 'VERB']
['is', 'said', 'are', 'was', 'be', 'has', 'have', 'will', 'says', 'would',
 'were', 'had', 'been', 'could', "'s", 'can', 'do', 'say', 'make', 'may',
 'did', 'rose', 'made', 'does', 'expected', 'buy', 'take', 'get', 'might',
 'sell', 'added', 'sold', 'help', 'including', 'should', 'reported', ...]

Note that the items being counted in the frequency distribution are word-tag pairs.
Since words and tags are paired, we can treat the word as a condition and the tag
as an event, and initialize a conditional frequency distribution with a list of
condition-event pairs. This lets us see a frequency-ordered list of tags given a word:

 
>>> cfd1 = nltk.ConditionalFreqDist(wsj)
>>> cfd1['yield'].most_common()
[('VERB', 28), ('NOUN', 20)]
>>> cfd1['cut'].most_common()
[('VERB', 25), ('NOUN', 3)]

We can reverse the order of the pairs, so that the tags are the conditions, and the
words are the events. Now we can see likely words for a given tag. We
will do this for the WSJ tagset rather than the universal tagset:

 
>>> wsj = nltk.corpus.treebank.tagged_words()
>>> cfd2 = nltk.ConditionalFreqDist((tag, word) for (word, tag) in wsj)
>>> list(cfd2['VBN'])
['been', 'expected', 'made', 'compared', 'based', 'priced', 'used', 'sold',
'named', 'designed', 'held', 'fined', 'taken', 'paid', 'traded', 'said', ...]

To clarify the distinction between VBD (past tense) and VBN
(past participle), let’s find words which can be both VBD and
VBN, and see some surrounding text:

 
>>> [w for w in cfd1.conditions() if 'VBD' in cfd1[w] and 'VBN' in cfd1[w]]
['Asked', 'accelerated', 'accepted', 'accused', 'acquired', 'added', 'adopted', ...]
>>> idx1 = wsj.index(('kicked', 'VBD'))
>>> wsj[idx1-4:idx1+1]
[('While', 'IN'), ('program', 'NN'), ('trades', 'NNS'), ('swiftly', 'RB'),
 ('kicked', 'VBD')]
>>> idx2 = wsj.index(('kicked', 'VBN'))
>>> wsj[idx2-4:idx2+1]
[('head', 'NN'), ('of', 'IN'), ('state', 'NN'), ('has', 'VBZ'), ('kicked', 'VBN')]

In this case, we see that the past participle of kicked is preceded by a form of
the auxiliary verb have. Is this generally true?

Note

Your Turn:
Given the list of past participles produced by
list(cfd2[‘VN’]), try to collect a list of all the word-tag
pairs that immediately precede items in that list.

2.6   Adjectives and Adverbs

Two other important word classes are adjectives and adverbs.
Adjectives describe nouns, and can be used as modifiers
(e.g. large in the large pizza), or in predicates (e.g. the
pizza is large
). English adjectives can have internal structure
(e.g. fall+ing in the falling
stocks
). Adverbs modify verbs to specify the time, manner, place or
direction of the event described by the verb (e.g. quickly in
the stocks fell quickly). Adverbs may also modify adjectives
(e.g. really in Mary’s teacher was really nice).

English has several categories of closed class words in addition to
prepositions, such as articles (also often called determiners)
(e.g., the, a), modals (e.g., should,
may), and personal pronouns (e.g., she, they).
Each dictionary and grammar classifies these words differently.

Note

Your Turn:
If you are uncertain about some of these parts of speech, study them using
nltk.app.concordance(), or watch some of the Schoolhouse Rock!
grammar videos available at YouTube, or consult the Further Reading
section at the end of this chapter.

2.8   Exploring Tagged Corpora

Let’s briefly return to the kinds of exploration of corpora we saw in previous chapters,
this time exploiting POS tags.

Suppose we’re studying the word often and want to see how it is used
in text. We could ask to see the words that follow often

 
>>> brown_learned_text = brown.words(categories='learned')
>>> sorted(set(b for (a, b) in nltk.bigrams(brown_learned_text) if a == 'often'))
[',', '.', 'accomplished', 'analytically', 'appear', 'apt', 'associated', 'assuming',
'became', 'become', 'been', 'began', 'call', 'called', 'carefully', 'chose', ...]

However, it’s probably more instructive to use the tagged_words()
method to look at the part-of-speech tag of the following words:

 
>>> brown_lrnd_tagged = brown.tagged_words(categories='learned', tagset='universal')
>>> tags = [b[1] for (a, b) in nltk.bigrams(brown_lrnd_tagged) if a[0] == 'often']
>>> fd = nltk.FreqDist(tags)
>>> fd.tabulate()
 PRT  ADV  ADP    . VERB  ADJ
   2    8    7    4   37    6

Notice that the most high-frequency parts of speech following often are verbs.
Nouns never appear in this position (in this particular corpus).

Next, let’s look at some larger context, and find words involving
particular sequences of tags and words (in this case «<Verb> to <Verb>»).
In code-three-word-phrase we consider each three-word window in the sentence [1],
and check if they meet our criterion [2]. If the tags
match, we print the corresponding words [3].

Finally, let’s look for words that are highly ambiguous as to their part of speech tag.
Understanding why such words are tagged as they are in each context can help us clarify
the distinctions between the tags.

 
>>> brown_news_tagged = brown.tagged_words(categories='news', tagset='universal')
>>> data = nltk.ConditionalFreqDist((word.lower(), tag)
...                                 for (word, tag) in brown_news_tagged)
>>> for word in sorted(data.conditions()):
...     if len(data[word]) > 3:
...         tags = [tag for (tag, _) in data[word].most_common()]
...         print(word, ' '.join(tags))
...
best ADJ ADV NP V
better ADJ ADV V DET
close ADV ADJ V N
cut V N VN VD
even ADV DET ADJ V
grant NP N V -
hit V VD VN N
lay ADJ V NP VD
left VD ADJ N VN
like CNJ V ADJ P -
near P ADV ADJ DET
open ADJ V N ADV
past N ADJ DET P
present ADJ ADV V N
read V VN VD NP
right ADJ N DET ADV
second NUM ADV DET N
set VN V VD N -
that CNJ V WH DET

Note

Your Turn:
Open the POS concordance tool nltk.app.concordance() and load the complete
Brown Corpus (simplified tagset). Now pick some of the above words and see how the tag
of the word correlates with the context of the word.
E.g. search for near to see all forms mixed together, near/ADJ to see it used
as an adjective, near N to see just those cases where a noun follows, and so forth.
For a larger set of examples, modify the supplied code so that it lists words having
three distinct tags.

3   Mapping Words to Properties Using Python Dictionaries

As we have seen, a tagged word of the form (word, tag) is
an association between a word and a part-of-speech tag.
Once we start doing part-of-speech tagging, we will be creating
programs that assign a tag to a word, the tag which is most
likely in a given context. We can think of this process as
mapping from words to tags. The most natural way to
store mappings in Python uses the so-called dictionary data type
(also known as an associative array or hash array
in other programming languages).
In this section we look at dictionaries and see how they can
represent a variety of language information, including
parts of speech.

3.1   Indexing Lists vs Dictionaries

A text, as we have seen, is treated in Python as a list of words.
An important property of lists is that we can «look up» a particular
item by giving its index, e.g. text1[100]. Notice how we specify
a number, and get back a word. We can think of a list as a simple
kind of table, as shown in 3.1.

../images/maps01.png

Figure 3.1: List Look-up: we access the contents of a Python list with the help of an integer index.

Contrast this situation with frequency distributions (3),
where we specify a word, and get back a number, e.g. fdist[‘monstrous’], which
tells us the number of times a given word has occurred in a text. Look-up using words is
familiar to anyone who has used a dictionary. Some more examples are shown in
3.2.

../images/maps02.png

Figure 3.2: Dictionary Look-up: we access the entry of a dictionary using a key
such as someone’s name, a web domain, or an English word;
other names for dictionary are map, hashmap, hash, and associative array.

In the case of a phonebook, we look up an entry using a name,
and get back a number. When we type a domain name in a web browser,
the computer looks this up to get back an IP address. A word
frequency table allows us to look up a word and find its frequency in
a text collection. In all these cases, we are mapping from names to
numbers, rather than the other way around as with a list.
In general, we would like to be able to map between
arbitrary types of information. 3.1 lists a variety
of linguistic objects, along with what they map.

Table 3.1:

Linguistic Objects as Mappings from Keys to Values

Linguistic Object Maps From Maps To
Document Index Word List of pages (where word is found)
Thesaurus Word sense List of synonyms
Dictionary Headword Entry (part-of-speech, sense definitions, etymology)
Comparative Wordlist Gloss term Cognates (list of words, one per language)
Morph Analyzer Surface form Morphological analysis (list of component morphemes)

Most often, we are mapping from a «word» to some structured object.
For example, a document index maps from a word (which we can represent
as a string), to a list of pages (represented as a list of integers).
In this section, we will see how to represent such mappings in Python.

3.2   Dictionaries in Python

Python provides a dictionary data type that can be used for
mapping between arbitrary types. It is like a conventional dictionary,
in that it gives you an efficient way to look things up. However,
as we see from 3.1, it has a much wider range of uses.

To illustrate, we define pos to be an empty dictionary and then add four
entries to it, specifying the part-of-speech of some words. We add
entries to a dictionary using the familiar square bracket notation:

So, for example, [1] says that
the part-of-speech of colorless is adjective, or more
specifically, that the key ‘colorless’
is assigned the value ‘ADJ’ in dictionary pos.
When we inspect the value of pos [2] we see
a set of key-value pairs. Once we have populated the dictionary
in this way, we can employ the keys to retrieve values:

 
>>> pos['ideas']
'N'
>>> pos['colorless']
'ADJ'

Of course, we might accidentally use a key that hasn’t been assigned a value.

 
>>> pos['green']
Traceback (most recent call last):
  File "<stdin>", line 1, in ?
KeyError: 'green'

This raises an important question. Unlike lists and strings, where we
can use len() to work out which integers will be legal indexes,
how do we work out the legal keys for a dictionary? If the dictionary
is not too big, we can simply inspect its contents by evaluating the
variable pos. As we saw above (line [2]), this gives
us the key-value pairs. Notice that they are not in the same order they
were originally entered; this is because dictionaries are not sequences
but mappings (cf. 3.2), and the keys are not inherently
ordered.

Alternatively, to just find the keys, we can convert the
dictionary to a list [1] — or use
the dictionary in a context where a list is expected,
as the parameter of sorted() [2],
or in a for loop [3].

Note

When you type list(pos) you might see a different order
to the one shown above. If you want to see the keys in order, just sort them.

As well as iterating over all keys
in the dictionary with a for loop, we can use the for loop
as we did for printing lists:

 
>>> for word in sorted(pos):
...     print(word + ":", pos[word])
...
colorless: ADJ
furiously: ADV
sleep: V
ideas: N

Finally, the dictionary methods keys(), values() and
items() allow us to access the keys, values, and key-value pairs as separate lists.
We can even sort tuples [1], which orders them according to their first element
(and if the first elements are the same, it uses their second elements).

 
>>> list(pos.keys())
['colorless', 'furiously', 'sleep', 'ideas']
>>> list(pos.values())
['ADJ', 'ADV', 'V', 'N']
>>> list(pos.items())
[('colorless', 'ADJ'), ('furiously', 'ADV'), ('sleep', 'V'), ('ideas', 'N')]
>>> for key, val in sorted(pos.items()): [1]
...     print(key + ":", val)
...
colorless: ADJ
furiously: ADV
ideas: N
sleep: V

We want to be sure that when we look something up in a dictionary, we
only get one value for each key. Now
suppose we try to use a dictionary to store the fact that the
word sleep can be used as both a verb and a noun:

 
>>> pos['sleep'] = 'V'
>>> pos['sleep']
'V'
>>> pos['sleep'] = 'N'
>>> pos['sleep']
'N'

Initially, pos[‘sleep’] is given the value ‘V’. But this is
immediately overwritten with the new value ‘N’.
In other words, there can only be one entry in the dictionary for ‘sleep’.
However, there is a way of storing multiple values in
that entry: we use a list value,
e.g. pos[‘sleep’] = [‘N’, ‘V’]. In fact, this is what we
saw in 4 for the CMU Pronouncing Dictionary,
which stores multiple pronunciations for a single word.

3.3   Defining Dictionaries

We can use the same key-value pair format to create a dictionary. There’s
a couple of ways to do this, and we will normally use the first:

 
>>> pos = {'colorless': 'ADJ', 'ideas': 'N', 'sleep': 'V', 'furiously': 'ADV'}
>>> pos = dict(colorless='ADJ', ideas='N', sleep='V', furiously='ADV')

Note that dictionary keys must be immutable types, such as strings and tuples.
If we try to define a dictionary using a mutable key, we get a TypeError:

 
>>> pos = {['ideas', 'blogs', 'adventures']: 'N'}
Traceback (most recent call last):
  File "<stdin>", line 1, in <module>
TypeError: list objects are unhashable

3.4   Default Dictionaries

If we try to access a key that is not in a dictionary, we get an error.
However, its often useful if a dictionary can automatically create
an entry for this new key and give it a default value, such as zero or
the empty list. For this reason, a special kind of dictionary
called a defaultdict is available.
In order to use it, we have to supply a parameter which can be used to
create the default value, e.g. int, float, str, list, dict,
tuple.

 
>>> from collections import defaultdict
>>> frequency = defaultdict(int)
>>> frequency['colorless'] = 4
>>> frequency['ideas']
0
>>> pos = defaultdict(list)
>>> pos['sleep'] = ['NOUN', 'VERB']
>>> pos['ideas']
[]

Note

These default values are actually functions that convert other
objects to the specified type (e.g. int(«2»), list(«2»)).
When they are called with no parameter — int(), list()
— they return 0 and [] respectively.

The above examples specified the default value of a dictionary entry to
be the default value of a particular data type. However, we can specify
any default value we like, simply by providing the name of a function
that can be called with no arguments to create the required value.
Let’s return to our part-of-speech example, and create a dictionary
whose default value for any entry is ‘N’ [1].
When we access a non-existent entry [2],
it is automatically added to the dictionary [3].

Note

The above example used a lambda expression, introduced in
4.4. This lambda expression specifies no
parameters, so we call it using parentheses with no arguments.
Thus, the definitions of f and g below are equivalent:

 
>>> f = lambda: 'NOUN'
>>> f()
'NOUN'
>>> def g():
...     return 'NOUN'
>>> g()
'NOUN'

Let’s see how default dictionaries could be used in a more substantial
language processing task.
Many language processing tasks — including tagging — struggle to correctly process
the hapaxes of a text. They can perform better with a fixed vocabulary and a
guarantee that no new words will appear. We can preprocess a text to replace
low-frequency words with a special «out of vocabulary» token UNK, with
the help of a default dictionary. (Can you work out how to do this without
reading on?)

We need to create a default dictionary that maps each word to its replacement.
The most frequent n words will be mapped to themselves.
Everything else will be mapped to UNK.

 
>>> alice = nltk.corpus.gutenberg.words('carroll-alice.txt')
>>> vocab = nltk.FreqDist(alice)
>>> v1000 = [word for (word, _) in vocab.most_common(1000)]
>>> mapping = defaultdict(lambda: 'UNK')
>>> for v in v1000:
...     mapping[v] = v
...
>>> alice2 = [mapping[v] for v in alice]
>>> alice2[:100]
['UNK', 'Alice', "'", 's', 'UNK', 'in', 'UNK', 'by', 'UNK', 'UNK', 'UNK',
'UNK', 'CHAPTER', 'I', '.', 'UNK', 'the', 'Rabbit', '-', 'UNK', 'Alice',
'was', 'beginning', 'to', 'get', 'very', 'tired', 'of', 'sitting', 'by',
'her', 'sister', 'on', 'the', 'UNK', ',', 'and', 'of', 'having', 'nothing',
'to', 'do', ':', 'once', 'or', 'twice', 'she', 'had', 'UNK', 'into', 'the',
'book', 'her', 'sister', 'was', 'UNK', ',', 'but', 'it', 'had', 'no',
'pictures', 'or', 'UNK', 'in', 'it', ',', "'", 'and', 'what', 'is', 'the',
'use', 'of', 'a', 'book', ",'", 'thought', 'Alice', "'", 'without',
'pictures', 'or', 'conversation', "?'" ...]
>>> len(set(alice2))
1001

3.5   Incrementally Updating a Dictionary

We can employ dictionaries to count occurrences, emulating the method
for tallying words shown in fig-tally.
We begin by initializing an empty defaultdict, then process each
part-of-speech tag in the text. If the tag hasn’t been seen before,
it will have a zero count by default. Each time we encounter a tag,
we increment its count using the += operator.

 
>>> from collections import defaultdict
>>> counts = defaultdict(int)
>>> from nltk.corpus import brown
>>> for (word, tag) in brown.tagged_words(categories='news', tagset='universal'):
...     counts[tag] += 1
...
>>> counts['NOUN']
30640
>>> sorted(counts)
['ADJ', 'PRT', 'ADV', 'X', 'CONJ', 'PRON', 'VERB', '.', 'NUM', 'NOUN', 'ADP', 'DET']

>>> from operator import itemgetter
>>> sorted(counts.items(), key=itemgetter(1), reverse=True)
[('NOUN', 30640), ('VERB', 14399), ('ADP', 12355), ('.', 11928), ...]
>>> [t for t, c in sorted(counts.items(), key=itemgetter(1), reverse=True)]
['NOUN', 'VERB', 'ADP', '.', 'DET', 'ADJ', 'ADV', 'CONJ', 'PRON', 'PRT', 'NUM', 'X']

Example 3.3 (code_dictionary.py): Figure 3.3: Incrementally Updating a Dictionary, and Sorting by Value

The listing in 3.3 illustrates an important idiom for
sorting a dictionary by its values, to show words in decreasing
order of frequency. The first parameter of sorted() is the items
to sort, a list of tuples consisting of a POS tag and a frequency.
The second parameter specifies the sort key using a function itemgetter().
In general, itemgetter(n) returns a function that can be called on
some other sequence object to obtain the nth element, e.g.:

 
>>> pair = ('NP', 8336)
>>> pair[1]
8336
>>> itemgetter(1)(pair)
8336

The last parameter of sorted() specifies that the items should be returned
in reverse order, i.e. decreasing values of frequency.

There’s a second useful programming idiom at the beginning of
3.3, where we initialize a defaultdict and then use a
for loop to update its values. Here’s a schematic version:

>>> my_dictionary = defaultdict(

function to create default value

)

my_dictionary[

item_key

]

is updated with information about item

Here’s another instance of this pattern, where we index words according to their last two letters:

 
>>> last_letters = defaultdict(list)
>>> words = nltk.corpus.words.words('en')
>>> for word in words:
...     key = word[-2:]
...     last_letters[key].append(word)
...
>>> last_letters['ly']
['abactinally', 'abandonedly', 'abasedly', 'abashedly', 'abashlessly', 'abbreviately',
'abdominally', 'abhorrently', 'abidingly', 'abiogenetically', 'abiologically', ...]
>>> last_letters['zy']
['blazy', 'bleezy', 'blowzy', 'boozy', 'breezy', 'bronzy', 'buzzy', 'Chazy', ...]

The following example uses the same pattern to create an anagram dictionary.
(You might experiment with the third line to get an idea of why this program works.)

 
>>> anagrams = defaultdict(list)
>>> for word in words:
...     key = ''.join(sorted(word))
...     anagrams[key].append(word)
...
>>> anagrams['aeilnrt']
['entrail', 'latrine', 'ratline', 'reliant', 'retinal', 'trenail']

Since accumulating words like this is such a common task,
NLTK provides a more convenient way of creating a defaultdict(list),
in the form of nltk.Index().

 
>>> anagrams = nltk.Index((''.join(sorted(w)), w) for w in words)
>>> anagrams['aeilnrt']
['entrail', 'latrine', 'ratline', 'reliant', 'retinal', 'trenail']

Note

nltk.Index is a defaultdict(list) with extra support for
initialization. Similarly,
nltk.FreqDist is essentially a defaultdict(int) with extra
support for initialization (along with sorting and plotting methods).

3.6   Complex Keys and Values

We can use default dictionaries with complex keys and values.
Let’s study the range of possible tags for a word, given the
word itself, and the tag of the previous word. We will see
how this information can be used by a POS tagger.

This example uses a dictionary whose default value for an entry
is a dictionary (whose default value is int(), i.e. zero).
Notice how we iterated over the bigrams of the tagged
corpus, processing a pair of word-tag pairs for each iteration [1].
Each time through the loop we updated our pos dictionary’s
entry for (t1, w2), a tag and its following word [2].
When we look up an item in pos we must specify a compound key [3],
and we get back a dictionary object.
A POS tagger could use such information to decide that the
word right, when preceded by a determiner, should be tagged as ADJ.

3.7   Inverting a Dictionary

Dictionaries support efficient lookup, so long as you want to get the value for
any key. If d is a dictionary and k is a key, we type d[k] and
immediately obtain the value. Finding a key given a value is slower and more
cumbersome:

 
>>> counts = defaultdict(int)
>>> for word in nltk.corpus.gutenberg.words('milton-paradise.txt'):
...     counts[word] += 1
...
>>> [key for (key, value) in counts.items() if value == 32]
['brought', 'Him', 'virtue', 'Against', 'There', 'thine', 'King', 'mortal',
'every', 'been']

If we expect to do this kind of «reverse lookup» often, it helps to construct
a dictionary that maps values to keys. In the case that no two keys have
the same value, this is an easy thing to do. We just get all the key-value
pairs in the dictionary, and create a new dictionary of value-key
pairs. The next example also illustrates another way of initializing a
dictionary pos with key-value pairs.

 
>>> pos = {'colorless': 'ADJ', 'ideas': 'N', 'sleep': 'V', 'furiously': 'ADV'}
>>> pos2 = dict((value, key) for (key, value) in pos.items())
>>> pos2['N']
'ideas'

Let’s first make our part-of-speech dictionary a bit more realistic
and add some more words to pos using the dictionary update() method, to
create the situation where multiple keys have the same value. Then the
technique just shown for reverse lookup will no longer work (why
not?). Instead, we have to use append() to accumulate the words
for each part-of-speech, as follows:

 
>>> pos.update({'cats': 'N', 'scratch': 'V', 'peacefully': 'ADV', 'old': 'ADJ'})
>>> pos2 = defaultdict(list)
>>> for key, value in pos.items():
...     pos2[value].append(key)
...
>>> pos2['ADV']
['peacefully', 'furiously']

Now we have inverted the pos dictionary, and can look up any part-of-speech and find
all words having that part-of-speech. We can do the same thing even
more simply using NLTK’s support for indexing as follows:

 
>>> pos2 = nltk.Index((value, key) for (key, value) in pos.items())
>>> pos2['ADV']
['peacefully', 'furiously']

A summary of Python’s dictionary methods is given in 3.2.

Table 3.2:

Python’s Dictionary Methods: A summary of commonly-used methods and idioms
involving dictionaries.

Example Description
d = {} create an empty dictionary and assign it to d
d[key] = value assign a value to a given dictionary key
d.keys() the list of keys of the dictionary
list(d) the list of keys of the dictionary
sorted(d) the keys of the dictionary, sorted
key in d test whether a particular key is in the dictionary
for key in d iterate over the keys of the dictionary
d.values() the list of values in the dictionary
dict([(k1,v1), (k2,v2), …]) create a dictionary from a list of key-value pairs
d1.update(d2) add all items from d2 to d1
defaultdict(int) a dictionary whose default value is zero

4   Automatic Tagging

In the rest of this chapter we will explore various ways to automatically
add part-of-speech tags to text. We will see that the tag of a word depends
on the word and its context within a sentence. For this reason, we will
be working with data at the level of (tagged) sentences rather than words.
We’ll begin by loading the data we will be using.

 
>>> from nltk.corpus import brown
>>> brown_tagged_sents = brown.tagged_sents(categories='news')
>>> brown_sents = brown.sents(categories='news')

4.1   The Default Tagger

The simplest possible tagger assigns the same tag to each token. This
may seem to be a rather banal step, but it establishes an important
baseline for tagger performance. In order to get the best result, we
tag each word with the most likely tag. Let’s find out which tag is
most likely (now using the unsimplified tagset):

 
>>> tags = [tag for (word, tag) in brown.tagged_words(categories='news')]
>>> nltk.FreqDist(tags).max()
'NN'

Now we can create a tagger that tags everything as NN.

 
>>> raw = 'I do not like green eggs and ham, I do not like them Sam I am!'
>>> tokens = nltk.word_tokenize(raw)
>>> default_tagger = nltk.DefaultTagger('NN')
>>> default_tagger.tag(tokens)
[('I', 'NN'), ('do', 'NN'), ('not', 'NN'), ('like', 'NN'), ('green', 'NN'),
('eggs', 'NN'), ('and', 'NN'), ('ham', 'NN'), (',', 'NN'), ('I', 'NN'),
('do', 'NN'), ('not', 'NN'), ('like', 'NN'), ('them', 'NN'), ('Sam', 'NN'),
('I', 'NN'), ('am', 'NN'), ('!', 'NN')]

Unsurprisingly, this method performs rather poorly.
On a typical corpus, it will tag only about an eighth of the tokens correctly,
as we see below:

 
>>> default_tagger.evaluate(brown_tagged_sents)
0.13089484257215028

Default taggers assign their tag to every single word, even words that
have never been encountered before. As it happens, once we have processed
several thousand words of English text, most new words will be nouns.
As we will see, this means that default taggers can help to improve the
robustness of a language processing system. We will return to them
shortly.

4.2   The Regular Expression Tagger

The regular expression tagger assigns tags to tokens on the basis of
matching patterns. For instance, we might guess that any word ending
in ed is the past participle of a verb, and any word ending with
‘s is a possessive noun. We can express these as a list of
regular expressions:

 
>>> patterns = [
...     (r'.*ing$', 'VBG'),                
...     (r'.*ed$', 'VBD'),                 
...     (r'.*es$', 'VBZ'),                 
...     (r'.*ould$', 'MD'),                
...     (r'.*'s$', 'NN$'),                
...     (r'.*s$', 'NNS'),                  
...     (r'^-?[0-9]+(.[0-9]+)?$', 'CD'),  
...     (r'.*', 'NN')                      
... ]

Note that these are processed in order, and the first one that matches is applied.
Now we can set up a tagger and use it to tag a sentence. Now its right about a fifth
of the time.

 
>>> regexp_tagger = nltk.RegexpTagger(patterns)
>>> regexp_tagger.tag(brown_sents[3])
[('``', 'NN'), ('Only', 'NN'), ('a', 'NN'), ('relative', 'NN'), ('handful', 'NN'),
('of', 'NN'), ('such', 'NN'), ('reports', 'NNS'), ('was', 'NNS'), ('received', 'VBD'),
("''", 'NN'), (',', 'NN'), ('the', 'NN'), ('jury', 'NN'), ('said', 'NN'), (',', 'NN'),
('``', 'NN'), ('considering', 'VBG'), ('the', 'NN'), ('widespread', 'NN'), ...]
>>> regexp_tagger.evaluate(brown_tagged_sents)
0.20326391789486245

The final regular expression «.*» is a catch-all that tags everything as a noun.
This is equivalent to the default tagger (only much less efficient).
Instead of re-specifying this as part of the regular expression tagger,
is there a way to combine this tagger with the default tagger? We
will see how to do this shortly.

Note

Your Turn:
See if you can come up with patterns to improve the performance of the above
regular expression tagger. (Note that 1
describes a way to partially automate such work.)

4.3   The Lookup Tagger

A lot of high-frequency words do not have the NN tag.
Let’s find the hundred most frequent words and store their most likely tag.
We can then use this information as the model for a «lookup tagger»
(an NLTK UnigramTagger):

 
>>> fd = nltk.FreqDist(brown.words(categories='news'))
>>> cfd = nltk.ConditionalFreqDist(brown.tagged_words(categories='news'))
>>> most_freq_words = fd.most_common(100)
>>> likely_tags = dict((word, cfd[word].max()) for (word, _) in most_freq_words)
>>> baseline_tagger = nltk.UnigramTagger(model=likely_tags)
>>> baseline_tagger.evaluate(brown_tagged_sents)
0.45578495136941344

It should come as no surprise by now that simply
knowing the tags for the 100 most frequent words enables us to tag a large fraction of
tokens correctly (nearly half in fact).
Let’s see what it does on some untagged input text:

 
>>> sent = brown.sents(categories='news')[3]
>>> baseline_tagger.tag(sent)
[('``', '``'), ('Only', None), ('a', 'AT'), ('relative', None),
('handful', None), ('of', 'IN'), ('such', None), ('reports', None),
('was', 'BEDZ'), ('received', None), ("''", "''"), (',', ','),
('the', 'AT'), ('jury', None), ('said', 'VBD'), (',', ','),
('``', '``'), ('considering', None), ('the', 'AT'), ('widespread', None),
('interest', None), ('in', 'IN'), ('the', 'AT'), ('election', None),
(',', ','), ('the', 'AT'), ('number', None), ('of', 'IN'),
('voters', None), ('and', 'CC'), ('the', 'AT'), ('size', None),
('of', 'IN'), ('this', 'DT'), ('city', None), ("''", "''"), ('.', '.')]

Many words have been assigned a tag of None,
because they were not among the 100 most frequent words.
In these cases we would like to assign the default tag of NN.
In other words, we want to use the lookup table first,
and if it is unable to assign a tag, then use the default tagger,
a process known as backoff (5).
We do this by specifying one tagger as a parameter to the other,
as shown below. Now the lookup tagger will only store word-tag pairs
for words other than nouns, and whenever it cannot assign a tag to a
word it will invoke the default tagger.

 
>>> baseline_tagger = nltk.UnigramTagger(model=likely_tags,
...                                      backoff=nltk.DefaultTagger('NN'))

Let’s put all this together and write a program to create and
evaluate lookup taggers having a range of sizes, in 4.1.

 
def performance(cfd, wordlist):
    lt = dict((word, cfd[word].max()) for word in wordlist)
    baseline_tagger = nltk.UnigramTagger(model=lt, backoff=nltk.DefaultTagger('NN'))
    return baseline_tagger.evaluate(brown.tagged_sents(categories='news'))

def display():
    import pylab
    word_freqs = nltk.FreqDist(brown.words(categories='news')).most_common()
    words_by_freq = [w for (w, _) in word_freqs]
    cfd = nltk.ConditionalFreqDist(brown.tagged_words(categories='news'))
    sizes = 2 ** pylab.arange(15)
    perfs = [performance(cfd, words_by_freq[:size]) for size in sizes]
    pylab.plot(sizes, perfs, '-bo')
    pylab.title('Lookup Tagger Performance with Varying Model Size')
    pylab.xlabel('Model Size')
    pylab.ylabel('Performance')
    pylab.show()

Example 4.1 (code_baseline_tagger.py): Figure 4.1: Lookup Tagger Performance with Varying Model Size

../images/tag-lookup.png

Figure 4.2: Lookup Tagger

Observe that performance initially increases rapidly as the model size grows, eventually
reaching a plateau, when large increases in model size yield little improvement
in performance. (This example used the pylab plotting package, discussed
in 4.8.)

5   N-Gram Tagging

5.1   Unigram Tagging

Unigram taggers are based on a simple statistical algorithm:
for each token, assign the tag that is most likely for
that particular token. For example, it will assign the tag JJ to any
occurrence of the word frequent, since frequent is used as an
adjective (e.g. a frequent word) more often than it is used as a
verb (e.g. I frequent this cafe).
A unigram tagger behaves just like a lookup tagger (4),
except there is a more convenient technique for setting it up,
called training. In the following code sample,
we train a unigram tagger, use it to tag a sentence, then evaluate:

 
>>> from nltk.corpus import brown
>>> brown_tagged_sents = brown.tagged_sents(categories='news')
>>> brown_sents = brown.sents(categories='news')
>>> unigram_tagger = nltk.UnigramTagger(brown_tagged_sents)
>>> unigram_tagger.tag(brown_sents[2007])
[('Various', 'JJ'), ('of', 'IN'), ('the', 'AT'), ('apartments', 'NNS'),
('are', 'BER'), ('of', 'IN'), ('the', 'AT'), ('terrace', 'NN'), ('type', 'NN'),
(',', ','), ('being', 'BEG'), ('on', 'IN'), ('the', 'AT'), ('ground', 'NN'),
('floor', 'NN'), ('so', 'QL'), ('that', 'CS'), ('entrance', 'NN'), ('is', 'BEZ'),
('direct', 'JJ'), ('.', '.')]
>>> unigram_tagger.evaluate(brown_tagged_sents)
0.9349006503968017

We train a UnigramTagger by specifying tagged sentence data as
a parameter when we initialize the tagger. The training process involves
inspecting the tag of each word and storing the most likely tag for any word
in a dictionary, stored inside the tagger.

5.2   Separating the Training and Testing Data

Now that we are training a tagger on some data, we must be careful not to test it on the
same data, as we did in the above example. A tagger that simply memorized its training data
and made no attempt to construct a general model would get a perfect score, but would also
be useless for tagging new text. Instead, we should split the data, training on 90% and
testing on the remaining 10%:

 
>>> size = int(len(brown_tagged_sents) * 0.9)
>>> size
4160
>>> train_sents = brown_tagged_sents[:size]
>>> test_sents = brown_tagged_sents[size:]
>>> unigram_tagger = nltk.UnigramTagger(train_sents)
>>> unigram_tagger.evaluate(test_sents)
0.811721...

Although the score is worse, we now have a better picture of the usefulness of
this tagger, i.e. its performance on previously unseen text.

5.3   General N-Gram Tagging

When we perform a language processing task based on unigrams, we are using
one item of context. In the case of tagging, we only consider the current
token, in isolation from any larger context. Given such a model, the best
we can do is tag each word with its a priori most likely tag.
This means we would tag a word such as wind with the same tag,
regardless of whether it appears in the context the wind or
to wind.

An n-gram tagger is a generalization of a unigram tagger whose context is
the current word together with the part-of-speech tags of the
n-1 preceding tokens, as shown in 5.1. The tag to be
chosen, tn, is circled, and the context is shaded
in grey. In the example of an n-gram tagger shown in 5.1,
we have n=3; that is, we consider the tags of the two preceding words in addition
to the current word. An n-gram tagger
picks the tag that is most likely in the given context.

../images/tag-context.png

Figure 5.1: Tagger Context

Note

A 1-gram tagger is another term for a unigram tagger: i.e.,
the context used to tag a token is just the text of the token itself.
2-gram taggers are also called bigram taggers, and 3-gram taggers
are called trigram taggers.

The NgramTagger class uses a tagged training corpus to determine which
part-of-speech tag is most likely for each context. Here we see
a special case of an n-gram tagger, namely a bigram tagger.
First we train it, then use it to tag untagged sentences:

 
>>> bigram_tagger = nltk.BigramTagger(train_sents)
>>> bigram_tagger.tag(brown_sents[2007])
[('Various', 'JJ'), ('of', 'IN'), ('the', 'AT'), ('apartments', 'NNS'),
('are', 'BER'), ('of', 'IN'), ('the', 'AT'), ('terrace', 'NN'),
('type', 'NN'), (',', ','), ('being', 'BEG'), ('on', 'IN'), ('the', 'AT'),
('ground', 'NN'), ('floor', 'NN'), ('so', 'CS'), ('that', 'CS'),
('entrance', 'NN'), ('is', 'BEZ'), ('direct', 'JJ'), ('.', '.')]
>>> unseen_sent = brown_sents[4203]
>>> bigram_tagger.tag(unseen_sent)
[('The', 'AT'), ('population', 'NN'), ('of', 'IN'), ('the', 'AT'), ('Congo', 'NP'),
('is', 'BEZ'), ('13.5', None), ('million', None), (',', None), ('divided', None),
('into', None), ('at', None), ('least', None), ('seven', None), ('major', None),
('``', None), ('culture', None), ('clusters', None), ("''", None), ('and', None),
('innumerable', None), ('tribes', None), ('speaking', None), ('400', None),
('separate', None), ('dialects', None), ('.', None)]

Notice that the bigram tagger manages to tag every word in a sentence it saw during
training, but does badly on an unseen sentence. As soon as it encounters a new word
(i.e., 13.5), it is unable to assign a tag. It cannot tag the following word
(i.e., million) even if it was seen during training, simply because it never
saw it during training with a None tag on the previous word. Consequently, the
tagger fails to tag the rest of the sentence. Its overall accuracy score is very low:

 
>>> bigram_tagger.evaluate(test_sents)
0.102063...

As n gets larger, the specificity of the contexts increases,
as does the chance that the data we wish to tag contains contexts that
were not present in the training data. This is known as the sparse
data
problem, and is quite pervasive in NLP. As a consequence, there is a
trade-off between the accuracy and the coverage of our results (and
this is related to the precision/recall trade-off in information
retrieval).

Caution!

n-gram taggers should not consider context that crosses a
sentence boundary. Accordingly, NLTK taggers are designed to work
with lists of sentences, where each sentence is a list of words. At
the start of a sentence, tn-1 and preceding
tags are set to None.

5.4   Combining Taggers

One way to address the trade-off between accuracy and coverage is to
use the more accurate algorithms when we can, but to fall back on
algorithms with wider coverage when necessary. For example, we could
combine the results of a bigram tagger, a unigram tagger, and
a default tagger, as follows:

  1. Try tagging the token with the bigram tagger.
  2. If the bigram tagger is unable to find a tag for the token, try
    the unigram tagger.
  3. If the unigram tagger is also unable to find a tag, use a default tagger.

Most NLTK taggers permit a backoff-tagger to be specified.
The backoff-tagger may itself have a backoff tagger:

 
>>> t0 = nltk.DefaultTagger('NN')
>>> t1 = nltk.UnigramTagger(train_sents, backoff=t0)
>>> t2 = nltk.BigramTagger(train_sents, backoff=t1)
>>> t2.evaluate(test_sents)
0.844513...

Note

Your Turn:
Extend the above example by defining a TrigramTagger called
t3, which backs off to t2.

Note that we specify the backoff tagger when the tagger is
initialized so that training can take advantage of the backoff tagger.
Thus, if the bigram tagger would assign the same tag
as its unigram backoff tagger in a certain context,
the bigram tagger discards the training instance.
This keeps the bigram tagger model as small as possible. We can
further specify that a tagger needs to see more than one instance of a
context in order to retain it, e.g. nltk.BigramTagger(sents, cutoff=2, backoff=t1)
will discard contexts that have only been seen once or twice.

5.5   Tagging Unknown Words

Our approach to tagging unknown words still uses backoff to a regular-expression tagger
or a default tagger. These are unable to make use of context. Thus, if our tagger
encountered the word blog, not seen during training, it would assign it the same tag,
regardless of whether this word appeared in the context the blog or to blog.
How can we do better with these unknown words, or out-of-vocabulary items?

A useful method to tag unknown words based on context is to limit the vocabulary
of a tagger to the most frequent n words, and to replace every other word
with a special word UNK using the method shown in 3.
During training, a unigram tagger will probably learn that UNK is usually a noun.
However, the n-gram taggers will detect contexts in which it has some other tag.
For example, if the preceding word is to (tagged TO), then UNK
will probably be tagged as a verb.

5.6   Storing Taggers

Training a tagger on a large corpus may take a significant time. Instead of training a tagger
every time we need one, it is convenient to save a trained tagger in a file for later re-use.
Let’s save our tagger t2 to a file t2.pkl.

 
>>> from pickle import dump
>>> output = open('t2.pkl', 'wb')
>>> dump(t2, output, -1)
>>> output.close()

Now, in a separate Python process, we can load our saved tagger.

 
>>> from pickle import load
>>> input = open('t2.pkl', 'rb')
>>> tagger = load(input)
>>> input.close()

Now let’s check that it can be used for tagging.

 
>>> text = """The board's action shows what free enterprise
...     is up against in our complex maze of regulatory laws ."""
>>> tokens = text.split()
>>> tagger.tag(tokens)
[('The', 'AT'), ("board's", 'NN$'), ('action', 'NN'), ('shows', 'NNS'),
('what', 'WDT'), ('free', 'JJ'), ('enterprise', 'NN'), ('is', 'BEZ'),
('up', 'RP'), ('against', 'IN'), ('in', 'IN'), ('our', 'PP$'), ('complex', 'JJ'),
('maze', 'NN'), ('of', 'IN'), ('regulatory', 'NN'), ('laws', 'NNS'), ('.', '.')]

5.7   Performance Limitations

What is the upper limit to the performance of an n-gram tagger?
Consider the case of a trigram tagger. How many cases of part-of-speech ambiguity does it
encounter? We can determine the answer to this question empirically:

 
>>> cfd = nltk.ConditionalFreqDist(
...            ((x[1], y[1], z[0]), z[1])
...            for sent in brown_tagged_sents
...            for x, y, z in nltk.trigrams(sent))
>>> ambiguous_contexts = [c for c in cfd.conditions() if len(cfd[c]) > 1]
>>> sum(cfd[c].N() for c in ambiguous_contexts) / cfd.N()
0.049297702068029296

Thus, one out of twenty trigrams is ambiguous [EXAMPLES]. Given the
current word and the previous two tags, in 5% of cases there is more than one tag
that could be legitimately assigned to the current word according to
the training data. Assuming we always pick the most likely tag in
such ambiguous contexts, we can derive a lower bound on
the performance of a trigram tagger.

Another way to investigate the performance of a tagger is to study
its mistakes. Some tags may be harder than others to assign, and
it might be possible to treat them specially by pre- or post-processing
the data. A convenient way to look at tagging errors is the
confusion matrix. It charts expected tags (the gold standard)
against actual tags generated by a tagger:

 
>>> test_tags = [tag for sent in brown.sents(categories='editorial')
...                  for (word, tag) in t2.tag(sent)]
>>> gold_tags = [tag for (word, tag) in brown.tagged_words(categories='editorial')]
>>> print(nltk.ConfusionMatrix(gold_tags, test_tags))           

Based on such analysis we may decide to modify the tagset. Perhaps
a distinction between tags that is difficult to make can be dropped,
since it is not important in the context of some larger processing task.

Another way to analyze the performance bound on a tagger comes from
the less than 100% agreement between human annotators. [MORE]

In general, observe that the tagging process collapses distinctions:
e.g. lexical identity is usually lost when all personal pronouns are
tagged PRP. At the same time, the tagging process introduces
new distinctions and removes ambiguities: e.g. deal tagged as VB or NN.
This characteristic of collapsing certain distinctions and introducing new
distinctions is an important feature of tagging which
facilitates classification and prediction.
When we introduce finer distinctions in a tagset, an n-gram tagger gets
more detailed information about the left-context when it is deciding
what tag to assign to a particular word.
However, the tagger simultaneously has to do more work to classify the
current token, simply because there are more tags to choose from.
Conversely, with fewer distinctions (as with the simplified tagset),
the tagger has less information about context, and it has a smaller
range of choices in classifying the current token.

We have seen that ambiguity in the training data leads to an upper limit
in tagger performance. Sometimes more context will resolve the
ambiguity. In other cases however, as noted by (Church, Young, & Bloothooft, 1996), the
ambiguity can only be resolved with reference to syntax, or to world
knowledge. Despite these imperfections, part-of-speech tagging has
played a central role in the rise of statistical approaches to natural
language processing. In the early 1990s, the surprising accuracy of
statistical taggers was a striking demonstration that it was possible
to solve one small part of the language understanding problem, namely
part-of-speech disambiguation, without reference to deeper sources of
linguistic knowledge. Can this idea be pushed further? In 7.,
we shall see that it can.

6   Transformation-Based Tagging

A potential issue with n-gram taggers is the size of their n-gram
table (or language model). If tagging is to be employed in a variety
of language technologies deployed on mobile computing devices, it is
important to strike a balance between model size and tagger
performance. An n-gram tagger with backoff may store trigram and
bigram tables, large sparse arrays which may have hundreds of millions
of entries.

A second issue concerns context. The only information an n-gram
tagger considers from prior context is tags, even though words
themselves might be a useful source of information. It is simply
impractical for n-gram models to be conditioned on the identities of
words in the context. In this section we examine Brill tagging,
an inductive tagging method which performs very well using models
that are only a tiny fraction of the size of n-gram taggers.

Brill tagging is a kind of transformation-based learning, named
after its inventor. The
general idea is very simple: guess the tag of each word, then go back
and fix the mistakes. In this way, a Brill tagger successively
transforms a bad tagging of a text into a better one. As with n-gram
tagging, this is a supervised learning method, since we need
annotated training data to figure out whether the tagger’s guess is a
mistake or not. However, unlike n-gram tagging, it does
not count observations but compiles a list of transformational
correction rules.

The process of Brill tagging is usually explained by analogy with
painting. Suppose we were painting a tree, with all its details of
boughs, branches, twigs and leaves, against a uniform sky-blue
background. Instead of painting the tree first then trying to paint
blue in the gaps, it is simpler to paint the whole canvas blue, then
«correct» the tree section by over-painting the blue background. In
the same fashion we might paint the trunk a uniform brown before going
back to over-paint further details with even finer brushes. Brill
tagging uses the same idea: begin with broad brush strokes then fix up
the details, with successively finer changes. Let’s look at an
example involving the following sentence:

(1) The President said he will ask Congress to increase grants to states
for vocational rehabilitation

We will examine the operation of two rules:
(a) Replace NN with VB when the previous word is TO;
(b) Replace TO with IN when the next tag is NNS.
6.1
illustrates this process, first tagging with the unigram tagger, then
applying the rules to fix the errors.

Table 6.1:

Steps in Brill Tagging

Phrase to increase grants to states for vocational rehabilitation
Unigram TO NN NNS TO NNS IN JJ NN
Rule 1   VB            
Rule 2       IN        
Output TO VB NNS IN NNS IN JJ NN
Gold TO VB NNS IN NNS IN JJ NN

In this table we see two rules. All such rules are generated from a
template of the following form: «replace T1 with
T2 in the context C«. Typical contexts are the
identity or the tag of the preceding or following word, or the
appearance of a specific tag within 2-3 words of the current word. During
its training phase, the tagger guesses values for T1,
T2 and C, to create thousands of candidate rules.
Each rule is scored according to its net benefit: the
number of incorrect tags that it corrects, less the number of correct
tags it incorrectly modifies.

Brill taggers have another interesting property: the rules are
linguistically interpretable. Compare this with the n-gram taggers,
which employ a potentially massive table of n-grams. We cannot learn
much from direct inspection of such a table, in comparison to the
rules learned by the Brill tagger.
6.1 demonstrates NLTK’s Brill tagger.

 
>>> from nltk.tbl import demo as brill_demo
>>> brill_demo.demo()
Training Brill tagger on 80 sentences...
Finding initial useful rules...
    Found 6555 useful rules.

           B      |
   S   F   r   O  |        Score = Fixed - Broken
   c   i   o   t  |  R     Fixed = num tags changed incorrect -> correct
   o   x   k   h  |  u     Broken = num tags changed correct -> incorrect
   r   e   e   e  |  l     Other = num tags changed incorrect -> incorrect
   e   d   n   r  |  e
------------------+-------------------------------------------------------
  12  13   1   4  | NN -> VB if the tag of the preceding word is 'TO'
   8   9   1  23  | NN -> VBD if the tag of the following word is 'DT'
   8   8   0   9  | NN -> VBD if the tag of the preceding word is 'NNS'
   6   9   3  16  | NN -> NNP if the tag of words i-2...i-1 is '-NONE-'
   5   8   3   6  | NN -> NNP if the tag of the following word is 'NNP'
   5   6   1   0  | NN -> NNP if the text of words i-2...i-1 is 'like'
   5   5   0   3  | NN -> VBN if the text of the following word is '*-1'
   ...
>>> print(open("errors.out").read())
             left context |    word/test->gold     | right context
--------------------------+------------------------+--------------------------
                          |      Then/NN->RB       | ,/, in/IN the/DT guests/N
, in/IN the/DT guests/NNS |       '/VBD->POS       | honor/NN ,/, the/DT speed
'/POS honor/NN ,/, the/DT |    speedway/JJ->NN     | hauled/VBD out/RP four/CD
NN ,/, the/DT speedway/NN |     hauled/NN->VBD     | out/RP four/CD drivers/NN
DT speedway/NN hauled/VBD |      out/NNP->RP       | four/CD drivers/NNS ,/, c
dway/NN hauled/VBD out/RP |      four/NNP->CD      | drivers/NNS ,/, crews/NNS
hauled/VBD out/RP four/CD |    drivers/NNP->NNS    | ,/, crews/NNS and/CC even
P four/CD drivers/NNS ,/, |     crews/NN->NNS      | and/CC even/RB the/DT off
NNS and/CC even/RB the/DT |    official/NNP->JJ    | Indianapolis/NNP 500/CD a
                          |     After/VBD->IN      | the/DT race/NN ,/, Fortun
ter/IN the/DT race/NN ,/, |    Fortune/IN->NNP     | 500/CD executives/NNS dro
s/NNS drooled/VBD like/IN |  schoolboys/NNP->NNS   | over/IN the/DT cars/NNS a
olboys/NNS over/IN the/DT |      cars/NN->NNS      | and/CC drivers/NNS ./.

Example 6.1 (code_brill_demo.py): Figure 6.1: Brill Tagger Demonstration: the tagger has a collection of
templates of the form X -> Y if the preceding word is Z;
the variables in these templates are instantiated to particular
words and tags to create «rules»; the score for a rule is the
number of broken examples it corrects minus the number of
correct cases it breaks; apart from training a tagger, the
demonstration displays residual errors.

7   How to Determine the Category of a Word

Now that we have examined word classes in detail, we turn to a more
basic question: how do we decide what category a word belongs to in
the first place? In general, linguists use morphological, syntactic,
and semantic clues to determine the category of a word.

7.1   Morphological Clues

The internal structure of a word may give useful clues as to the
word’s category. For example, -ness is a suffix
that combines with an adjective to produce a noun, e.g.
happyhappiness, illillness. So
if we encounter a word that ends in -ness, this is very likely
to be a noun. Similarly, -ment is a suffix that combines
with some verbs to produce a noun, e.g.
governgovernment and establishestablishment.

English verbs can also be morphologically complex. For instance, the
present participle of a verb ends in -ing, and expresses
the idea of ongoing, incomplete action (e.g. falling, eating).
The -ing suffix also appears on nouns derived from verbs, e.g. the
falling of the leaves
(this is known as the gerund).

7.2   Syntactic Clues

Another source of information is the typical contexts in which a word can
occur. For example, assume that we have already determined the
category of nouns. Then we might say that a syntactic criterion for an
adjective in English is that it can occur immediately before a noun,
or immediately following the words be or very. According
to these tests, near should be categorized as an adjective:

(2)
b. The end is (very) near.

7.3   Semantic Clues

Finally, the meaning of a word is a useful clue as to its lexical
category. For example, the best-known definition of a noun is
semantic: «the name of a person, place or thing». Within modern linguistics,
semantic criteria for word classes are treated with suspicion, mainly
because they are hard to formalize. Nevertheless, semantic criteria
underpin many of our intuitions about word classes, and enable us to
make a good guess about the categorization of words in languages that
we are unfamiliar with. For example, if all we know about the Dutch word
verjaardag is that it means the same as the English word
birthday, then we can guess that verjaardag is a noun in
Dutch. However, some care is needed: although we might translate zij
is vandaag jarig
as it’s her birthday today, the word
jarig is in fact an adjective in Dutch, and has no exact
equivalent in English.

7.4   New Words

All languages acquire new lexical items. A list of words recently
added to the Oxford Dictionary of English includes cyberslacker,
fatoush, blamestorm, SARS, cantopop, bupkis, noughties, muggle
, and
robata. Notice that all these new words are nouns, and this is
reflected in calling nouns an open class. By contrast, prepositions
are regarded as a closed class. That is, there is a limited set of
words belonging to the class (e.g., above, along, at, below, beside,
between, during, for, from, in, near, on, outside, over, past,
through, towards, under, up, with
), and membership of the set only
changes very gradually over time.

8   Summary

  • Words can be grouped into classes, such as nouns, verbs, adjectives, and adverbs.
    These classes are known as lexical categories or parts of speech.
    Parts of speech are assigned short labels, or tags, such as NN, VB,
  • The process of automatically assigning parts of speech to words in text
    is called part-of-speech tagging, POS tagging, or just tagging.
  • Automatic tagging is an important step in the NLP pipeline,
    and is useful in a variety of situations including:
    predicting the behavior of previously unseen words,
    analyzing word usage in corpora, and text-to-speech systems.
  • Some linguistic corpora, such as the Brown Corpus, have been POS tagged.
  • A variety of tagging methods are possible, e.g.
    default tagger, regular expression tagger, unigram tagger and n-gram taggers.
    These can be combined using a technique known as backoff.
  • Taggers can be trained and evaluated using tagged corpora.
  • Backoff is a method for combining models: when a more specialized
    model (such as a bigram tagger) cannot assign a tag in a given
    context, we backoff to a more general model (such as a unigram tagger).
  • Part-of-speech tagging is an important, early example of a sequence
    classification task in NLP: a classification decision at any one point
    in the sequence makes use of words and tags in the local context.
  • A dictionary is used to map between arbitrary types of information,
    such as a string and a number: freq[‘cat’] = 12. We create
    dictionaries using the brace notation: pos = {},
    pos = {‘furiously’: ‘adv’, ‘ideas’: ‘n’, ‘colorless’: ‘adj’}.
  • N-gram taggers can be defined for large values of n, but once
    n is larger than 3 we usually encounter the sparse data problem;
    even with a large quantity of training data we only see a tiny
    fraction of possible contexts.
  • Transformation-based tagging involves learning a series
    of repair rules of the form «change tag s to tag
    t in context c«, where each rule
    fixes mistakes and possibly introduces a (smaller) number
    of errors.

9   Further Reading

Extra materials for this chapter are posted at http://nltk.org/, including links to freely
available resources on the web.
For more examples of tagging with NLTK, please see the
Tagging HOWTO at http://nltk.org/howto.
Chapters 4 and 5 of (Jurafsky & Martin, 2008) contain more advanced
material on n-grams and part-of-speech tagging.
The «Universal Tagset» is described by (Petrov, Das, & McDonald, 2012).
Other approaches to tagging involve machine learning methods (chap-data-intensive).
In 7. we will see a generalization of tagging called chunking in which a
contiguous sequence of words is assigned a single tag.

For tagset documentation, see
nltk.help.upenn_tagset() and nltk.help.brown_tagset().
Lexical categories are introduced in linguistics textbooks, including those
listed in 1..

There are many other kinds of tagging.
Words can be tagged with directives to a speech synthesizer,
indicating which words should be emphasized. Words can be tagged with sense
numbers, indicating which sense of the word was used. Words can also
be tagged with morphological features.
Examples of each of these kinds of tags are shown below.
For space reasons, we only show the tag for a single
word. Note also that the first two examples use XML-style
tags, where elements in angle brackets enclose the word that is
tagged.

  1. Speech Synthesis Markup Language (W3C SSML):
    That is a <emphasis>big</emphasis> car!
  2. SemCor: Brown Corpus tagged with WordNet senses:
    Space in any <wf pos=«NN» lemma=«form» wnsn=«4»>form</wf>
    is completely measured by the three dimensions.

    (Wordnet form/nn sense 4: «shape, form, configuration,
    contour, conformation»)
  3. Morphological tagging, from the Turin University Italian Treebank:
    E‘ italiano , come progetto e realizzazione , il
    primo (PRIMO ADJ ORDIN M SING) porto turistico dell’ Albania .

Note that tagging is also performed at higher levels. Here is an example
of dialogue act tagging, from the NPS Chat Corpus (Forsyth & Martell, 2007) included with
NLTK. Each turn of the dialogue is categorized as to its communicative
function:

Statement  User117 Dude..., I wanted some of that
ynQuestion User120 m I missing something?
Bye        User117 I'm gonna go fix food, I'll be back later.
System     User122 JOIN
System     User2   slaps User122 around a bit with a large trout.
Statement  User121 18/m pm me if u tryin to chat

10   Exercises


  1. Search the web for «spoof newspaper headlines», to find such gems as:
    British Left Waffles on Falkland Islands, and
    Juvenile Court to Try Shooting Defendant.
    Manually tag these headlines to see if knowledge of the part-of-speech
    tags removes the ambiguity.

  2. Working with someone else, take turns to pick a word that can be
    either a noun or a verb (e.g. contest); the opponent has to
    predict which one is likely to be the most frequent in the Brown corpus; check the
    opponent’s prediction, and tally the score over several turns.

  3. Tokenize and tag the following sentence:
    They wind back the clock, while we chase after the wind.
    What different pronunciations and parts of speech are involved?
  4. ☼ Review the mappings in 3.1. Discuss any other
    examples of mappings you can think of. What type of information do they map
    from and to?
  5. ☼ Using the Python interpreter in interactive mode, experiment with
    the dictionary examples in this chapter. Create a dictionary d, and add
    some entries. What happens if you try to access a non-existent
    entry, e.g. d[‘xyz’]?
  6. ☼ Try deleting an element from a dictionary d, using the syntax
    del d[‘abc’]. Check that the item was deleted.
  7. ☼ Create two dictionaries, d1 and d2, and add some entries to
    each. Now issue the command d1.update(d2). What did this do?
    What might it be useful for?
  8. ☼ Create a dictionary e, to represent a single lexical entry
    for some word of your choice.
    Define keys like headword, part-of-speech, sense, and
    example, and assign them suitable values.
  9. ☼ Satisfy yourself that there are
    restrictions on the distribution of go and went, in the
    sense that they cannot be freely interchanged in the kinds of contexts
    illustrated in (3d) in 7.

  10. Train a unigram tagger and run it on some new text.
    Observe that some words are not assigned a tag. Why not?

  11. Learn about the affix tagger (type help(nltk.AffixTagger)).
    Train an affix tagger and run it on some new text.
    Experiment with different settings for the affix length
    and the minimum word length. Discuss your findings.

  12. Train a bigram tagger with no backoff tagger, and run it on some of the training
    data. Next, run it on some new data.
    What happens to the performance of the tagger? Why?
  13. ☼ We can use a dictionary to specify the values to be
    substituted into a formatting string. Read Python’s library
    documentation for formatting strings
    http://docs.python.org/lib/typesseq-strings.html
    and use this method to display today’s date in two
    different formats.
  14. ◑ Use sorted() and set() to get a sorted list of tags used in the Brown
    corpus, removing duplicates.
  15. ◑ Write programs to process the Brown Corpus and find answers to the following
    questions:

    1. Which nouns are more common in their plural form, rather than their singular
      form? (Only consider regular plurals, formed with the -s suffix.)
    2. Which word has the greatest number of distinct tags. What are they, and
      what do they represent?
    3. List tags in order of decreasing frequency. What do the 20 most frequent tags represent?
    4. Which tags are nouns most commonly found after? What do these tags represent?
  16. ◑ Explore the following issues that arise in connection with the lookup tagger:
    1. What happens to the tagger performance for the various
      model sizes when a backoff tagger is omitted?
    2. Consider the curve in 4.2; suggest a
      good size for a lookup tagger that balances memory and performance.
      Can you come up with scenarios where it would be preferable to
      minimize memory usage, or to maximize performance with no regard for memory usage?
  17. ◑ What is the upper limit of performance for a lookup tagger,
    assuming no limit to the size of its table? (Hint: write a program
    to work out what percentage of tokens of a word are assigned
    the most likely tag for that word, on average.)
  18. ◑ Generate some statistics for tagged data to answer the following questions:
    1. What proportion of word types are always assigned the same part-of-speech tag?
    2. How many words are ambiguous, in the sense that they appear with at least two tags?
    3. What percentage of word tokens in the Brown Corpus involve
      these ambiguous words?
  19. ◑ The evaluate() method works out how accurately
    the tagger performs on this text. For example, if the supplied tagged text
    was [(‘the’, ‘DT’), (‘dog’, ‘NN’)] and the tagger produced the output
    [(‘the’, ‘NN’), (‘dog’, ‘NN’)], then the score would be 0.5.
    Let’s try to figure out how the evaluation method works:

    1. A tagger t takes a list of words as input, and produces a list of tagged words
      as output. However, t.evaluate() is given correctly tagged text as its only parameter.
      What must it do with this input before performing the tagging?
    2. Once the tagger has created newly tagged text, how might the evaluate() method
      go about comparing it with the original tagged text and computing the accuracy score?
    3. Now examine the source code to see how the method is implemented. Inspect
      nltk.tag.api.__file__ to discover the location of the source code,
      and open this file using an editor (be sure to use the api.py file and
      not the compiled api.pyc binary file).
  20. ◑ Write code to search the Brown Corpus for particular words and phrases
    according to tags, to answer the following questions:

    1. Produce an alphabetically sorted list of the distinct words tagged as MD.
    2. Identify words that can be plural nouns or third person singular verbs
      (e.g. deals, flies).
    3. Identify three-word prepositional phrases of the form IN + DET + NN
      (eg. in the lab).
    4. What is the ratio of masculine to feminine pronouns?
  21. ◑ In 3.1 we saw a table involving frequency counts for
    the verbs adore, love, like, prefer and
    preceding qualifiers absolutely and definitely.
    Investigate the full range of adverbs that appear before these four verbs.

  22. We defined the regexp_tagger that can be used
    as a fall-back tagger for unknown words. This tagger only checks for
    cardinal numbers. By testing for particular prefix or suffix strings,
    it should be possible to guess other tags. For example,
    we could tag any word that ends with -s as a plural noun.
    Define a regular expression tagger (using RegexpTagger())
    that tests for at least five other patterns in the spelling of words.
    (Use inline documentation to explain the rules.)

  23. Consider the regular expression tagger developed in the exercises in
    the previous section. Evaluate the tagger using its accuracy() method,
    and try to come up with ways to improve its performance. Discuss your findings.
    How does objective evaluation help in the development process?

  24. How serious is the sparse data problem? Investigate the
    performance of n-gram taggers as n increases from 1 to 6.
    Tabulate the accuracy score. Estimate the training data required
    for these taggers, assuming a vocabulary size of
    105 and a tagset size of 102.
  25. ◑ Obtain some tagged data for another language, and train and
    evaluate a variety of taggers on it. If the language is
    morphologically complex, or if there are any orthographic clues
    (e.g. capitalization) to word classes, consider developing a
    regular expression tagger for it (ordered after the unigram
    tagger, and before the default tagger). How does the accuracy of
    your tagger(s) compare with the same taggers run on English data?
    Discuss any issues you encounter in applying these methods to the language.
  26. ◑ 4.1 plotted a curve showing
    change in the performance of a lookup tagger as the model size was increased.
    Plot the performance curve for a unigram tagger, as the amount of training
    data is varied.

  27. Inspect the confusion matrix for the bigram tagger t2 defined in 5,
    and identify one or more sets of tags to collapse. Define a dictionary to do
    the mapping, and evaluate the tagger on the simplified data.

  28. Experiment with taggers using the simplified tagset (or make one of your
    own by discarding all but the first character of each tag name).
    Such a tagger has fewer distinctions to make, but much less
    information on which to base its work. Discuss your findings.

  29. Recall the example of a bigram tagger which encountered a word it hadn’t
    seen during training, and tagged the rest of the sentence as None.
    It is possible for a bigram tagger to fail part way through a sentence
    even if it contains no unseen words (even if the sentence was used during
    training). In what circumstance can this happen? Can you write a program
    to find some examples of this?

  30. Preprocess the Brown News data by replacing low frequency words with UNK,
    but leaving the tags untouched. Now train and evaluate a bigram tagger
    on this data. How much does this help? What is the contribution of the unigram
    tagger and default tagger now?

  31. Modify the program in 4.1 to use a logarithmic scale on
    the x-axis, by replacing pylab.plot() with pylab.semilogx().
    What do you notice about the shape of the resulting plot? Does the gradient
    tell you anything?

  32. Consult the documentation for the Brill tagger demo function,
    using help(nltk.tag.brill.demo).
    Experiment with the tagger by setting different values for the parameters.
    Is there any trade-off between training time (corpus size) and performance?
  33. ◑ Write code that builds a dictionary of dictionaries of sets.
    Use it to store the set of POS tags that can follow a given word having
    a given POS tag, i.e. wordi → tagi
    → tagi+1.
  34. ★ There are 264 distinct words in the Brown Corpus having exactly
    three possible tags.

    1. Print a table with the integers 1..10 in one column, and the
      number of distinct words in the corpus having 1..10 distinct tags
      in the other column.
    2. For the word with the greatest number of distinct tags, print
      out sentences from the corpus containing the word, one for each
      possible tag.
  35. ★ Write a program to classify contexts involving the word must according
    to the tag of the following word. Can this be used to discriminate between the
    epistemic and deontic uses of must?

  36. Create a regular expression tagger and various unigram and n-gram taggers,
    incorporating backoff, and train them on part of the Brown corpus.

    1. Create three different combinations of the taggers. Test the
      accuracy of each combined tagger. Which combination works best?
    2. Try varying the size of the training corpus. How does it affect
      your results?

  37. Our approach for tagging an unknown word has been to consider the letters of the word
    (using RegexpTagger()), or to ignore the word altogether and tag
    it as a noun (using nltk.DefaultTagger()). These methods will not do well for texts having
    new words that are not nouns.
    Consider the sentence I like to blog on Kim’s blog. If blog is a new
    word, then looking at the previous tag (TO versus NP$) would probably be helpful.
    I.e. we need a default tagger that is sensitive to the preceding tag.

    1. Create a new kind of unigram tagger that looks at the tag of the previous word,
      and ignores the current word. (The best way to do this is to modify the source
      code for UnigramTagger(), which presumes knowledge of object-oriented
      programming in Python.)
    2. Add this tagger to the sequence of backoff taggers (including ordinary trigram
      and bigram taggers that look at words), right before the usual default tagger.
    3. Evaluate the contribution of this new unigram tagger.

  38. Consider the code in 5 which
    determines the upper bound for accuracy of a trigram tagger.
    Review Abney’s discussion concerning the impossibility of
    exact tagging (Church, Young, & Bloothooft, 1996). Explain why correct tagging of
    these examples requires access to other kinds of information than
    just words and tags. How might you estimate the scale of this problem?

  39. Use some of the estimation techniques in nltk.probability,
    such as Lidstone or Laplace estimation, to develop a statistical
    tagger that does a better job than n-gram backoff taggers in cases where
    contexts encountered during testing were not seen during training.

  40. Inspect the diagnostic files created by the Brill tagger rules.out and
    errors.out. Obtain the demonstration code by accessing the source code
    (at http://www.nltk.org/code)
    and create your own version of the Brill tagger.
    Delete some of the rule templates, based on what you learned from inspecting rules.out.
    Add some new rule templates which employ contexts that might help to
    correct the errors you saw in errors.out.

  41. Develop an n-gram backoff tagger that permits «anti-n-grams» such as
    [«the», «the»] to be specified when a tagger is initialized.
    An anti-ngram is assigned a count of zero and is used to prevent
    backoff for this n-gram (e.g. to avoid
    estimating P(the | the) as just P(the)).

  42. Investigate three different ways to define the split between training and
    testing data when developing a tagger using the Brown Corpus:
    genre (category), source (fileid), and sentence.
    Compare their relative performance and discuss which method
    is the most legitimate. (You might use n-fold cross validation,
    discussed in 3, to improve the accuracy of the evaluations.)

  43. Develop your own NgramTagger class that inherits from NLTK’s class,
    and which encapsulates the method of collapsing the vocabulary of
    the tagged training and testing data that was described in
    this chapter. Make sure that the unigram and default backoff taggers
    have access to the full vocabulary.

About this document…

UPDATED FOR NLTK 3.0.
This is a chapter from Natural Language Processing with Python,
by Steven Bird, Ewan Klein and Edward Loper,
Copyright © 2019 the authors.
It is distributed with the Natural Language Toolkit [http://nltk.org/],
Version 3.0, under the terms of the
Creative Commons Attribution-Noncommercial-No Derivative Works 3.0 United States License
[http://creativecommons.org/licenses/by-nc-nd/3.0/us/].

This document was built on
Wed 4 Sep 2019 11:40:48 ACST

Like this post? Please share to your friends:
  • Word and spanish accents
  • Word and pdf viewer
  • Word and spaces count
  • Word and pdf reader
  • Word and sound records