22
Vienna, September 22, 2005 CLEF 2005 Workshop Overview of the Multilingual Question Answering Track Alessandro Vallin, Danilo Giampiccolo, Bernardo Magnini CELCT - ITC-irst (Trento, Italy)

Overview of the Multilingual Question Answering Track

  • Upload
    kiral

  • View
    52

  • Download
    0

Embed Size (px)

DESCRIPTION

Overview of the Multilingual Question Answering Track Alessandro Vallin, Danilo Giampiccolo, Bernardo Magnini. CELCT - ITC-irst (Trento, Italy). Outline. Multilingual Question Answering Task Participants Results Approaches Conclusions and Future Directions. Question Answering. - PowerPoint PPT Presentation

Citation preview

Page 1: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Overview of the Multilingual Question

Answering Track

Alessandro Vallin, Danilo Giampiccolo, Bernardo Magnini

CELCT - ITC-irst(Trento, Italy)

Page 2: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Outline

Multilingual Question Answering Task Participants Results Approaches Conclusions and Future Directions

Page 3: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Open Domain Question AnsweringQuestion Answering

Domain specific Domain-independent

Structured data Free text

WebFixed set

of collectionsSingle

document

QA at CLEF: Focus on multilingual question answering- monolingual tasks in other languages than English- cross-language QA: questions in a source language, answers in a target language

Page 4: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• Answers may be found in languages different from the language of the question.

• Interest in QA systems for languages other than English.

• Force the QA community to design real multilingual systems.

• Check/improve the portability of the technologies implemented in current English QA systems.

Multilingual Question Answering

Page 5: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Cross-Language QA Quanto è alto il Mont Ventoux?

(How tall is Mont Ventoux?)

“Le Mont Ventoux, impérial avec ses 1909 mètres et sa tour blanche telle un étendard, règne de toutes …”

1909 metri

English corpusItalian corpus Spanish corpusFrench corpus

Page 6: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

QA at CLEF

• TREC QA style– Prevalence of Factoid questions– Exact answers + document id

• Use the CLEF corpora (news, 1994 -1995)

• Return the answer in the language of the text collection in which it has been found (i.e. no translation of the answer)

• QA-CLEF started as a pilot in 2003

Page 7: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Nine groups coordinated the QA track:

- CELCT / ITC-irst (A. Vallin, D. Giampiccolo): Italian

- DFKI (G. Erbach, B. Sacalenau): German

- ELDA/ELRA (C. Ayache): French

- Linguateca (D. Santos): Portuguese

- UNED (A. Penas): Spanish

- U. Amsterdam (M. De Rijke): Dutch

- U. Limerick (R. Sutcliff): English

- Bulgarian Academy of Sciences (P. Osenova): Bulgarian

- U. Helsinki (l. Aunimo): Finnish

- University of Indonesia: Indonesian as source

QA-CLEF-05: Organization

Page 8: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

QA-CLEF-05: TaskQuestions: 200 open domain questions; three kinds:

• Factoid (ca. 50%): Which Arab country produces the most oil?

• Definition (ca. 25%): Who is Josef Paul Kleihues?

• Temporally restricted (ca. 15%): by period, event and date

• NIL (ca. 10%):

• Answers: exact answer in the target language

• Document collections: open domain news corpora (on average 230MB)

• Evaluation: • Each answer: Right, Wrong, ineXact, Unsupported

• Two runs for participant

Page 9: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• By period: ex. Which liberal politician was Italy's Health Minister from 1989 to 1993?

• By event: ex. Who was the Norwegian Prime Minister when the referendum on Norway's possible accession to the EU was held?

• By date: ex. Which country did Jimmy Carter visit in 1994 at Bill Clinton's request?

QA-CLEF-05: Temporally Restricted

Questions

Page 10: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• Main evaluation measure was accuracy (fraction of Right responses).

• Whenever possible, a Confidence-Weighted Score was calculated:

• Beside CWS, two new measures were introduced, namely K1 and r value, to take into account both accuracy and confidence

1

QQ

i=1

number of correct responses in first i ranksi

CWS =

QA-CLEF-05: Evaluation

Page 11: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

documentcollections

translationEN => 10 languages

systems’answers

100 monolingual Q&A pairs with EN translation

IT

FR

NL

ES

900 Q&A pairs in 1 language + EN

selection ofadditional

80 + 20 questions

Multinine-05XML collection

of 900 Q&Ain 10 languages

extraction ofplain text test sets

experiments(1 week window)

manualassessment

Exercise

evaluation(2 p/d for 1 run)

QA-CLEF-05: Task Organization

Page 12: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

QA-CLEF-05: MultiNine<q answer_type="ORGANIZATION" category="D" cnt="0001" restriction="NONE">

<language original="TRUE" val="BG">  <question group="BTB">Кой е Джулио Андреоти?</question>   <answer docid="SEGA20020315-052" n="1">бившия италиански премиер и пожизнен сенатор</answer>   </language>

<language original="FALSE" val="DE">  <question group="DFKI">Wer ist Giulio Andreotti?</question>   </language>

<language original="FALSE" val="EN">  <question group="BTB">Who is Giulio Andreotti?</question>   <answer docid="GH950303-000082" n="1">former Italian Prime

Minister</answer>   </language>

<language original="FALSE" val="ES">  <question group="UNED">¿Quién es Giulio Andreotti?</question>

  <answer docid="EFE19951018-12131" n="1">ex primer ministro italiano</answer>   <answer docid="EFE19950921-13683" n="2">senador vitalicio</answer>   </language>

<language original="FALSE" val="FI">  <question group="HELS">Kuka on Giulio Andreotti?</question>   <answer docid="aamu19950927.sgml" n="1">Italian entinen pääministeri</answer>   </language>

<language original="FALSE" val="FR">  <question group="ELDA">Qui est Giulio Andreotti ?</question>   </language>

<language original="FALSE" val="IN">  <question group="UIN">Siapakah Giulio Andreotti itu?</question>   </language>

<language original="FALSE" val="IT">  <question group="CELCT">Chi è Giulio Andreotti?</question>   </language>

<language original="FALSE" val="NL">  <question group="UAM">Wie is Giulio Andreotti?</question>   <answer docid="AD19950128-0066" n="1">Italiaanse oud-premier</answer>     </q>

Page 13: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

8 Monolingual and 73 bilingual tasks

BG DE EN ES FI FR IN IT NL PT

BG

DE

EN

ES

FI

FR

IT

NL

PT

QA-CLEF-05: Potential Tasks

S

T

Page 14: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

America Europe Asia Australia TOTALsubmitted

runs

TREC-1999 13 3 3 1 20 46

TREC-2000 14 7 6 - 27 75

TREC-2001 19 8 8 - 35 67

TREC-2002 16 10 6 - 32 67

TREC-2003 13 8 4 - 25 54

TREC-2004 ? ? ? ? 28 63

CLEF 2003 3 5 - - 8 17

CLEF 2004 1 17 - - 18 48

CLEF 2005 1 22 1 - 24 67

QA-CLEF-05: Participants

Page 15: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

8 Monolingual and 15 bilingual tasks; (6 + 13 in 2004)Tasks / participant = 1.8; (1.6 in 2004)Comparability (tasks > 2 part. / tasks) = 0.2 (0.2 in 2004)

BG DE EN ES FI FR IN IT NL PT

BG 2 1

DE 2 1

EN 1 1 1 1 2 1 1

ES 2 7 1

FI 1

FR 1 7 1 1

IT 3

NL 2

PT 1 3

QA-CLEF-05: Activated Tasks

S

T

Page 16: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

34.01

23.5

32.5

24.5 28

45.5

28.64

43.65

65

56.5

48.5

44

60

37.18

27.5

43.5

25.5

42

23

64

27.5

49.5

64.5

36

58.5 61.5

73.5

26.5

80.5

46.5

68

77.5

0

10

20

30

40

50

60

70

80

90

Bulgarian

Germ

an

English

Spanish

Finnish

French

Italian

Dutch

Portuguese

Best2004

Combination2004

Best2005

Combination2005

QA-CLEF-05: Results

Page 17: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

67

83

70

41,535

45,5

35

64,5

39,5

23 22 21,4

29

17

23,7

14,7

29,36

18,48

0

10

20

30

40

50

60

70

80

90

TR

EC

-01

TR

EC

-02

TR

EC

-03

TR

EC

-04

Mono

Bilingual

Mono

Bilingual

Mono

Bilingual

Best

Average

CLEF03

CLEF04

CLEF05

QA-CLEF-05: Results

Page 18: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• Linguistic processors and resources are used by most of the systems.

• POS-tagging, Named Entities Recognition, WordNet, Gazzetters, partial parsing (chunking).

• Deep parsing is adopted by many systems

• Semantics (logical representation) is used by few systems

• Answer patterns:

• superficial patterns (regular expressions)

• deep (dependency trees): pre-processing the document collection, matching dependency trees, off-line answer patter retrieval.

QA-CLEF-05: Approaches (1)

Page 19: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• Few system use some form of “semantic” indexing based on syntactic information or named entities

•Few systems consult the Web at run-time

• to find answers in specialized portals

• to validate a candidate answer

• Architecture of QA modules: e.g. XML based

• Cross-languages

• commercial translators, word by word translation

• keyword translation

QA-CLEF-05: Approaches (2)

Page 20: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

• Increasing interest in multilingual QA

• More participants (+ 33%)

• Two new languages as target (Bulgarian and Finnish) and one as source (Indonesian)

• More activated tasks (44, they were 29 in 2004, + 51%)

• Monolingual is the prevalent interest (61%)

• Comparability among tasks is 20% (as in 2004)

• Interesting results: 6 systems above 40% accuracy

• Future. Under discussion:

• Add new languages (Romanian is a candidate)

• Introducing pilots

QA-CLEF-05: Conclusions

Page 21: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

PartecipantsAcronym Name Country

Alicante U. Alicante - Comp.Sci + Spain

INAOEIst.Nac.Astrofisica, Optica,

ElectronicaMexico

LINGUATECA-N Linguateca-Sintef Norway

TALP U.Politecnica Catalunya Spain

MIRACLE Daedalus & Madrid Univs Spain

Jaen U.Jaen - Intell.Systems Spain

DEPOK U.Indonesia - Comp.Sci Indonesia

DIRO U.Montreal Canada

Hildesheim U.Hildesheim - Inf.Sci Germany

DFKI Center Art. Intelligence Germany

Helsinki-2 U.Helsinki - Comp.Sci Finland

U.Glasgow U.Glasgow - IR UK

XLDB U.Lisbon Portugal

Hagen U.Hagen - Comp.Sci Germany

ILPS U.Amsterdam - Informatics Netherlands

ITC-irst IRST-Trento Italy

NICTA U.Melbourne - Comp.Sci Australia

Metacarta Metacarta Inc. USA

UNED UNED-LSI Spain

Priberam Priberam Informatica Portugal

SYNAPSE SYNAPSE Developpment France

Acronym Name CountryIRIT Inst.Informatique,Toulouse France

U.Evora U.Evora - Informatics Portugal

BUAP U.Autonomous Puebla Mexico

U.Sheffield U.Sheffield - Inf.Studies UK

U.Valencia U.Valencia - AI Spain

LINA Lab Informatique-NLP France

U.Groningen U.Groningen - Letters Netherlands

SICS Swedish Inst.Comp.Sci Sweden

Budapest U. Budapest U. Tech.&Econom Hungary

LIC2M CEA France

CLAWS MS group-U.Amsterdam Netherlands

Hummingbird Hummingbird Core Tech. Canada

CINDI U.Concordia - Comp.Sci Canada

REINA U.Salamanca - Translation Spain

Wolverhampton U.Wolverhampton - CL UK

BulTreeBank Bulgarian Acad.Sciences Bulgaria

LIR LIMSI-LIR France

U.Limerick U.Limerick Ireland

IMS Stuttgart U. Stuttgart Germany

Page 22: Overview of the  Multilingual Question Answering Track

Vienna, September 22, 2005

CLEF 2005 Workshop

Evaluation Exercise – CL Approaches

QuestionAnalysis / keyword extraction

INPUT (source language)

CandidateDocumentSelection

DocumentCollection

DocumentCollection

Preprocessing

PreprocessedDocuments

CandidateDocumentAnalysis

AnswerExtraction

OUTPUT (target language)

question translationinto target language

translation ofretrieved data

U. AmsterdamU. EdinburghU. Neuchatel

Bulg. Ac. of SciencesITC-IrstU. LimerickU. HelsinkiDFKILIMSI-CNRS