14
Why Not Grab a Free Lunch? Mining Large Corpora for Parallel Sentences to Improve Translation Modeling Ferhan Ture and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

Ferhan Ture and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

  • Upload
    abel

  • View
    45

  • Download
    0

Embed Size (px)

DESCRIPTION

Why Not Grab a Free Lunch? Mining Large Corpora for Parallel Sentences to Improve Translation Modeling. Ferhan Ture and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012. Extracting Bilingual Text. Problem: Mine bitext from comparable corpora - PowerPoint PPT Presentation

Citation preview

Page 1: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

Why Not Grab a Free Lunch? Mining Large Corpora for

Parallel Sentences to Improve Translation Modeling

Ferhan Ture and Jimmy LinUniversity of Maryland, College Park

NAACL-HLT’12June 6, 2012

Page 2: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

2

Problem: Mine bitext from comparable corpora

Application: Improve quality of MT models

Approach:Phase 1

Identify similar document pairs from comparable corporaPhase 2

1. Generate candidate sentence pairs2. Classify each candidate as ‘parallel’ or ‘not

parallel’

Extracting Bilingual Text

Page 3: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

3

Extracting Bilingual Text

Signature Generation

Candidate Generatio

n

2-stepClassifi

er

docvectorsF

docvectorsE

source-languagecollection Ftarget-languagecollection E

Phase 2candidate

sentence pairs

aligned bilingual sentencepairs

cross-lingualdocument pairs

Signature Generation

“No Free Lunch: Brute Force vs.

Locality-Sensitive Hashing for Cross-lingual Pairwise

Similarity”. Ture et al. SIGIR 2011.

Phase 11.5m German Wikipedia articles3.5m English Wikipedia articles

64mGerman-Englisharticle pairs

Page 4: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

4

Challenge: 64m document pairs hundreds of billions sentence pairs

Solution: 2-step classification approach

1. a simple classifier efficiently filters out irrelevant pairs

2. a complex classifier effectively classifies remaining pairs

Extracting Bilingual Text

Page 5: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

5

Extracting bitext from web pages (Resnik&Smith’03), news stories (Munteanu&Marcu’05), and Wikipedia articles (Smith et al’10).• no heuristics on document/time structure (i.e.,

generalizable)• scalable implementation

Recent Google paper with similar motivation (Uszkoreit et al’10)• far less computational resources• control “efficiency vs effectiveness”• not simply “more data is better”• significant results with much less data

Related Work

Page 6: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

6

• cosine similarity of the two sentences s1 and s2

where u1 and u2 are vector representations of s1 and s2

• sentence length ratio: the ratio of lengths of the two sentences

• word translation ratio: ratio of words in s1 that have translations in s2, (only consider translations with at least 0.01 probability)

Bitext ClassifierFeatures

Page 7: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

7

• Maximum entropy classifier (OpenNLP-MaxEnt)• Europarl v6 German-English corpus

Trained on1000 parallel, 5000 non-parallel (sampled from all possible)

Tested on 1000 parallel, 999000 non-parallel (all possible)

Bitext ClassifierEvaluation

comparable with Smith et al’10

4 times faster

good out-of-domain performance

Page 8: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

8

MapReduce

• Easy-to-understand programming model for designing scalable and distributed algorithms

• Experiments on Hadoop cluster- 12 nodes, each with 2 quad-core 2.2GHz Intel processors, 24

GB RAM

Page 9: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

9

Bitext Extraction Algorithm

sentence

detection+tf-idf

cross-lingualdocument pairs (ne,nf)

({se},{ve})({sf},{vf})

{ve,vf)}{(ve,vf

)’}simple

classification

complexclassificat

ionbitext

S1

bitext S2

source (ne , de)

target (nf , df)

sentences andsent. vectors

cartesian product

X

MAP < ne , de > ↦ < (ne , nf) , ({se}’,{ve}’) >

REDUCE < (ne , nf) , ({se}’,{ve}’,{sf}’,{vf}’) > ↦ <(ne , nf) , (se , sf)>

({se}’,{ve}’)({sf}’,{vf}’)

candidategeneration2.4 hours

shuffle&sort1.25 hours

simple classification4.13 hours

complex classification0.52 hours

Page 10: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

10

Evaluation on MTTrain with GIZA++, Hiero-style SCFGTune with MIRA on WMT10 development set (2525 sentences)Decode with cdec (2489 sentences) using 5-gram English LM (SRILM) Baseline system

all standard cdec features21.37 BLEU on test set5th out of 9 WMT10 teams with comparable resultsbest teams use novel techniques to exploit specific

aspects strong and competitive baseline

Page 11: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

Candidate generation64 million German-English article pairs from phase 1400 billion candidate sentence pairs214 billion after (# terms ≥ 3 and sentence length ≥

5)132 billion after (1/2 < sentence length ratio < 2)

End-to-End Experiments

WMT10 train simple > 0.98

WMT10 train simple > 0.986

complex > 0.60

0 3.1 8.1 16.9data size(in millions)

complex > 0.65

simple > 0.992

5.3

random sampling

11

2-step

1-step

Page 12: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

12

turning point when the benefits of more data exceeds the extra noise introduced

random > S2when low-scoringsentence pairs may be helpful inMT

S2>S1 consiste

ntly

2.39 BLEU improvementover baseline

Baseline = 21.37

Evaluation on MT

Page 13: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

13

Conclusions• Built approach to extract parallel

sentences from freely available resources– 5m sentence pairs highest BLEU in WMT’10– data-driven > task-specific engineeringWhy not grab a free lunch?

• We plan to extend to more language pairs and share our findings with the community

• All of our code and data is freely available

Page 14: Ferhan Ture  and Jimmy Lin University of Maryland, College Park NAACL-HLT’12 June 6, 2012

Thank you!Code: ivory.ccData: www.github.com/ferhanture/WikiBitext