18
Unsupervised Models for Named Entity Classifcation Michael Collins Yoram Singer AT&T Labs, 1999

Unsupervised Models for Named Entity Classifcation

Embed Size (px)

DESCRIPTION

Unsupervised Models for Named Entity Classifcation. Michael Collins Yoram Singer AT&T Labs, 1999. The Task. Tag phrases with “ person ”, “ organization ” or “ location ”. For example, R alph Grishman , of NYU , sure is swell. WHY?. Labeled data. Unlabeled data. Spelling Rules. - PowerPoint PPT Presentation

Citation preview

Page 1: Unsupervised Models for Named Entity Classifcation

Unsupervised Models for Named Entity Classifcation

Michael Collins

Yoram Singer

AT&T Labs, 1999

Page 2: Unsupervised Models for Named Entity Classifcation

The Task

• Tag phrases with “person”, “organization” or “location”.

For example,

Ralph Grishman, of NYU, sure is swell.

Page 3: Unsupervised Models for Named Entity Classifcation

WHY?

Unlabeled data

Labeled data

Page 4: Unsupervised Models for Named Entity Classifcation

Spelling Rules

• The approach uses two kinds of rules– Spelling

• Simple look up to see “Honduras” is a location!

• Look for words in string, like “Mr.”

Page 5: Unsupervised Models for Named Entity Classifcation

Contextual Rules

– Contextual• Words surrounding the string

– A rule that any proper name modified by an appositive whose head is “president” is a person.

Page 6: Unsupervised Models for Named Entity Classifcation

Two Categories of Rules

• The key to the method is redundancy in the two kind of rules.

…says Mr. Cooper, a vice president of…

spelling contextual

Unlabeled data gives us these hints!

contextualspelling

Page 7: Unsupervised Models for Named Entity Classifcation

The Experiment

• 970,000 New York Times sentences were parsed.

• Sequences of NNP and NNPS were then extracted as named entity examples if they met one of two critereon.

Page 8: Unsupervised Models for Named Entity Classifcation

Kinds of Noun Phrases

1. There was an appositive modifier to the NP, whose head is a singular noun (tagged NN).

– …says Maury Cooper, a vice president…

2. The NP is a compliment to a preposition which is the head of a PP. This PP modifies another NP whose head is a singular noun.

– … fraud related to work on a federally funded sewage plant in Georgia.

Page 9: Unsupervised Models for Named Entity Classifcation

(spelling, context) pairs created

• …says Maury Cooper, a vice president…– (Maury Cooper, president)

• … fraud related to work on a federally funded sewage plant in Georgia.– (Georgia, plant_in)

Page 10: Unsupervised Models for Named Entity Classifcation

Rules

• Set of rules– Full-string=x (full-string=Maury Cooper)– Contains(x) (contains(Maury))– Allcap1 IBM– Allcap2 N.Y.– Nonalpha=x A.T.&T. (nonalpha=..&.)– Context = x (context = president)– Context-type = x appos or prep

Page 11: Unsupervised Models for Named Entity Classifcation

SEED RULES

• Full-string = New York

• Full-string = California

• Full-string = U.S.

• Contains(Mr.)

• Contains(Incorporated)

• Full-string=Microsoft

• Full-string=I.B.M.

Page 12: Unsupervised Models for Named Entity Classifcation

The Algorithm

1. Initialize: Set the spelling decision list equal to the set of seed rules.

2. Label the training set using these rules.

3. Use these to get contextual rules. (x = feature, y = label)

4. Label set using contextual rules, and use to get sp. rules.

5. Set spelling rules to seed plus the new rules.

6. If less than threshold new rules, go to 2 and add 15 more.

7. When finished, label the training data with the combined spelling/contextual decision list, then induce a final decision list from the labeled examples where all rules are added to the decision list.

kxCount

yxCountyx

)(

),(maxarg,

Page 13: Unsupervised Models for Named Entity Classifcation

Example

• (IBM, company)– …IBM, the company that makes…

• (General Electric, company) – ..General Electric, a leading company in the area,…

• (General Electric, employer )– … joined General Electric, the biggest employer…

• (NYU, employer)– NYU, the employer of the famous Ralph Grishman,…

Page 14: Unsupervised Models for Named Entity Classifcation

The Power

I.B.M.

Mr.

Two classifiers both give labels on 49.2% of unlabeled examples

Agree on 99.25% of them!

Page 15: Unsupervised Models for Named Entity Classifcation

Evaluation

• 88,962 (spelling, context) pairs.– 971,746 sentences

• 1,000 randomly extracted to be test set.• Location, person, organization, noise• 186, 289, 402, 123• Took out 38 temporal noise.

• Clean Accuracy: Nc/ 962 • Noise Accuracy: Nc/(962-85)

Page 16: Unsupervised Models for Named Entity Classifcation

Results

Algorithm Clean Accuracy Noise Accuracy

Baseline 45.8% 41.8%

EM 83.1% 75.8%

Yarowsky 95 81.3% 74.1%

Yarowsky Cautious 91.2% 83.2%

DL-CoTrain 91.3% 83.3%

CoBoost 91.1% 83.1%

Page 17: Unsupervised Models for Named Entity Classifcation

QUESTIONS

Page 18: Unsupervised Models for Named Entity Classifcation

Thank you!

• www.lightrail.com/• www.cnnfn.com/• pbskids.org/• www.szilagyi.us• www.dflt.org