19
IRF Report 2010-00002 Date: 2010-07-23 CLEF-IP 2009: Evaluation Summary Part Two Florina Piroi, Giovanna Roda, Veronika Zenz ABSTRACT: This document presents a summary of the evaluation activities that were carried out for CLEF-IP 2009 track, with respect to 12 specific topics. The data from the 70 submitted runs was preprocessed and grouped into bundles according to experiment size and task. For each experiment we computed precision and recall at various levels, and map measures.

CLEF-IP 2009: Evaluation Summary Part Two - TU Wienclef-ip/pubs/CLEF-IP-2009-IRF-TR-2010-00002.pdf · CLEF-IP 2009: Evaluation Summary – Part Two Florina Piroi, ... EN, DE, FR S(4),M(4),

Embed Size (px)

Citation preview

IRF Report 2010-00002

Date: 2010-07-23

CLEF-IP 2009: Evaluation Summary –

Part Two

Florina Piroi, Giovanna Roda, Veronika Zenz

ABSTRACT: This document presents a summary of the evaluation activities that

were carried out for CLEF-IP 2009 track, with respect to 12 specific topics. The data

from the 70 submitted runs was preprocessed and grouped into bundles according to

experiment size and task. For each experiment we computed precision and recall at

various levels, and map measures.

Clef–Ip 2009: Evaluation Summary – Part Two

Florina [email protected]

Giovanna [email protected]

Veronika [email protected]

July 23, 2010

Abstract

This document presents a summary of the evaluation activities that were carried out forClef–Ip 2009 track, with respect to 12 specific topics. The data from the 70 submittedruns was preprocessed and grouped into bundles according to experiment size and task. Foreach experiment we computed precision and recall at various levels, and map measures.

1 Further Evaluations at Clef–Ip 2009

In this report we present the results of run evaluations for a set of 12 topics. The 70 runs used forthese evaluations were submitted by fourteen different teams corresponding to fifteen differentparticipating institutions (see Table 1).

For each experiment we computed 9 standard IR measures:; Precision, Precision@5, Precision@10, Precision@100; Recall, Recall@5, Recall@10, Recall@100; MAPThe values shown in this evaluation report are computed with trec eval 1.

Table 1: List of active participants and runs submitted

ID Institution Tasks Sizes RunsTUD Tech. Univ. Darmstadt, Dept. of CS,

Ubiquitous Knowledge Processing LabDE Main, EN,

DE, FRS(4),M(4),L(4), XL(4)

16

UniNE Univ. Neuchatel - Computer Science CH Main S(7), XL(1) 8uscom Santiago de Compostela Univ. - Dept.

Electronica y ComputacionES Main S(8) 8

UTASICS University of Tampere - Info Studies& Interactive Media and Swedish In-stitute of Computer Science

FISE

Main XL(8) 8

clefip-ug Glasgow Univ. - IR Group Keith UK Main M(4),XL(1) 5clefip-unige

Geneva Univ. - Centre Universitaired’Informatique

CH Main XL(5) 5

cwi Centrum Wiskunde & Informatica -Interactive Information Access

NL Main M(1),XL(4) 4

hcuge Geneva Univ. Hospitals - Service ofMedical Informatics

CH Main, EN,DE, FR

M(3),XL(1) 4

humb Humboldt Univ. - Dept. of GermanLanguage and Linguistics

DE Main, EN,DE, FR

XL(4) 4

clefip-dcu Dublin City Univ. - School ofComputing

IR Main XL(3) 3

clefip-run Radboud Univ. Nijmegen - Centrefor Language Studies & SpeechTechnologies

NL Main, EN S(2) 2

Continued on next page

1http://trec.nist.gov/trec eval

1

Evaluation Summary - Part Two 2

Table 1: List of active participants and runs submitted

ID Institution Tasks Sizes RunsHildesheim Hildesheim Univ. - Inst. of Informa-

tion Systems & Nat. Lang. ProcessingDE Main S(1) 1

NLEL Technical Univ. Valencia - NaturalLangugage Engineering

ES Main S(1) 1

UAIC AI. I. Cuza University of Iasi - NaturalLanguage Processing

RO EN S(1) 1

2 Selection of Topic Patents, Data Preprocessing

For this round of evaluations we have been collaborating with volunteering experts in intelectualproperty. Each patent expert aggreed to assess retrieval results for up to four patents in theirown area of expertise. Each expertise area was communicated to us as an Ipc code. Therefore,the twelve selected patents (see Table 2) are not randomly extracted from the Clef–Ip topicsets (Small, Medium, Large, eXtra Large), but selected depending on their Ipc code, with apreference for those with a higher number of recorded citations.

Table 2: Topics manually assessed by patent experts

ID Patent IPC codes Citations In Topic Set1 EP1421078 A61K, C07D, A61P 12 S2 EP1278338 H04L, G05B, H04Q 5 S3 EP1133104 G06F, H04L, H04N, H04B, H04M 8 S4 EP1236569 B41N, C08L, B41C, C08K, G03F 3 S5 EP1333887 A61K, A61P 4 S6 EP1314753 B32B, C08K, E04B, C08J, C09J 3 S7 EP1130071 C08L, C08F, C09J 12 M8 EP1255138 G02B 4 M9 EP1297752 A23D, C11B, A21D, A23L, C11C 3 Xl10 EP1175873 A61C 3 Xl11 EP1149535 A23F 3 Xl12 EP1333726 A23F, A23L 5 Xl

Before computing retrieval efficiency measures, we filtered the submitted experiments tocontain only the results for the 12 topics. Then, using a similar process to the one described in[1], we’ve created three evaluation bundles: one for experiments that contain retrieve results fortopics 1 to 6; a second one for experiments with results for topics 1 to 8; and a third one forthe experiments with retrieval results for all the 12 topics. Not only for convenience, we use thesame S, M, and XL bundle names as in the first evaluation summary [1].2 The list of (filtered)experiments in the resulting bundles are the same as in [1], Section 2, and we do not reproducethem here.

Section 4 shows the evaluation results grouped by S, M, and XL bundles of runs.

3 Assessments

3.1 Manual Relevance Assessments

To record the assessments done by the volunteering patent experts, we have used a customisedversion of the web plartform system used by Trec ’Legal Track’ since 20073. Upon login, thesystem allowed an IP expert to select the patent topic to assess, to inspect each retrieved patent,and mark its relevancy. However, only a few patent experts used the provided systems. Most of

2The ‘S’, ‘M’, ‘XL’ bundle names also reflect the fact that topics 1 to 6 are to be found in the Small topic set,topics 1 to 8 in the Medium topic set, and topics 1 to 12 in the eXtra Large topic set.

3The web platform has been originally desigend by Ian Soboroff, at NIST.

Evaluation Summary - Part Two 3

them preferred using their own system for doing assessments. Therefore, we have provided themwith text files of the documents to assess, and they sent us back the results of their assessments.

In average, each expert had to assess 264 results per topic, which correspond to a pooling upto rank 20 of the experimental data. The original relevance levels an assessor could assign were:

Figure 1: Assessment system after login.

‘unsure’, ‘not relevant’, ‘A’ (background information), ‘X’ (invalidating one or more claims), or’Y’ (combined with other references, invalidating one ore more claims). As, according to the Ipexperts, a decision wether a document is ‘X’ or ‘Y’-relevant for a given topic is time consuming,we have later aggreed that they would use only ‘not relevant’ and ‘A’ (or ‘relevant’) for theassessments.

Evaluation Summary - Part Two 4

Figure 2: Assessing one topic patent.

4 Measurements

In this section we present the values of the measures mentioned in Section 1 for all run filesgrouped by task and size.

4.1 Main Task

Table 3: Measures for bundle S, task Main

run id P R MAPclefip-dcu baseline 0.0092 0.5957 0.0615clefip-dcu Filtered2 0.0085 0.579 0.0729clefip-dcu Filtered 0.0082 0.5594 0.0501

clefip-run ClaimsBOW 0.0217 0.1971 0.0174clefip-ug bm25medstandard 0. 0. 0.

clefip-ug infdocfreqBM25EnglishTerms 0. 0. 0.clefip-ug infdocfreqBM25FrenchTermsNew 0. 0. 0.

clefip-ug infdocfreqBM25GermanTerms 0. 0. 0.clefip-ug infdocfreqCosEnglishTerms 0.0075 0.383 0.0193

clefip-unige RUN1 0.0075 0.5028 0.0934clefip-unige RUN2 0.0082 0.5646 0.0936clefip-unige RUN3 0.0075 0.5028 0.0934clefip-unige RUN4 0.0077 0.5123 0.0413clefip-unige RUN5 0.0045 0.351 0.0395

cwi bm25 0.0072 0.5104 0.0208cwi boolean 0.0059 0.3108 0.0031

cwi categorybm25 0.0058 0.381 0.0066cwi category 0.0062 0.4744 0.0197hcuge BiTeM 0.008 0.5204 0.0319

Hildesheim MethodAnew 0. 0. 0.humb 1 0.0079 0.4025 0.1631

NLEL MethodA 0.0058 0.37 0.0574TUD 800noTitle 0.0088 0.5699 0.0466

UAIC MethodAnew 0.0002 0.0128 0.0026UniNE strat1 0.005 0.3655 0.0346UniNE strat2 0.0073 0.5533 0.0779UniNE strat3 0.0098 0.5681 0.075UniNE strat4 0.0089 0.585 0.0714UniNE strat5 0.0065 0.4544 0.0197UniNE strat6 0.0096 0.6411 0.062

run id P R MAPContinued on next page

Evaluation Summary - Part Two 5

Table 3: Measures for bundle S, task Main, continued

run id P R MAPUniNE strat7 0.0097 0.6514 0.0199UniNE strat8 0.0082 0.5595 0.0189uscom BM25a 0.0084 0.5402 0.1165uscom BM25at 0.0088 0.586 0.1222uscom BM25b 0.0096 0.6012 0.0416uscom BM25bt 0.0099 0.6039 0.0503uscom BM25c 0.0094 0.5883 0.032uscom BM25ct 0.0096 0.5899 0.0372uscom BM25d 0.0092 0.5755 0.0315uscom BM25dt 0.0096 0.5899 0.0371

UTASICS abs-des-ratf-GT 0.0093 0.6541 0.0758UTASICS abs-des-ratf 0.009 0.6118 0.0447

UTASICS abs-tit-cla-ratf-GT-ipcr 0.0087 0.6746 0.0642UTASICS abs-tit-cla-ratf-ipcr 0.0087 0.641 0.0701

UTASICS all-ratf-GT-ipcr 0.0093 0.675 0.0953UTASICS all-ratf-ipcr 0.0092 0.6621 0.0853

UTASICS all-tf-idf-GT-ipcr 0.0082 0.6172 0.1102UTASICS all-tf-idf-ipcr 0.0087 0.6353 0.0981

run id P R MAP

Table 4: Measures for bundle M, task Main

run id P R MAPclefip-dcu baseline 0.0081 0.5975 0.055clefip-dcu Filtered2 0.0081 0.5997 0.0672clefip-dcu Filtered 0.0081 0.5924 0.0484

clefip-ug bm25medstandard 0.0009 0.0257 0.0002clefip-ug infdocfreqBM25EnglishTerms 0. 0. 0.clefip-ug infdocfreqBM25FrenchTerms 0. 0. 0.

clefip-ug infdocfreqBM25GermanTerms 0. 0. 0.clefip-ug infdocfreqCosEnglishTerms 0.0097 0.4932 0.0323

clefip-unige RUN1 0.0079 0.5573 0.0998clefip-unige RUN2 0.0086 0.6109 0.0934clefip-unige RUN3 0.0079 0.5573 0.0998clefip-unige RUN4 0.0076 0.5533 0.0563clefip-unige RUN5 0.0058 0.4471 0.0462

cwi bm25 0.0065 0.5298 0.0471cwi boolean 0.0186 0.3201 0.0036

cwi categorybm25 0.0067 0.4659 0.0164cwi category 0.0058 0.5029 0.0164hcuge BiTeM 0.0086 0.5668 0.0348

humb 1 0.0099 0.4784 0.149TUD 800noTitle 0.0095 0.626 0.0574

UniNE strat8 0.0088 0.6108 0.0261UTASICS abs-des-ratf-GT 0.0101 0.6965 0.071

UTASICS abs-des-ratf 0.01 0.6684 0.0677UTASICS abs-tit-cla-ratf-GT-ipcr 0.0093 0.7008 0.0539

UTASICS abs-tit-cla-ratf-ipcr 0.0093 0.6756 0.0583UTASICS all-ratf-GT-ipcr 0.0106 0.7268 0.0851

UTASICS all-ratf-ipcr 0.0105 0.7172 0.0717UTASICS all-tf-idf-GT-ipcr 0.0098 0.6835 0.0954

UTASICS all-tf-idf-ipcr 0.0099 0.6897 0.0792

run id P R MAP

Evaluation Summary - Part Two 6

Table 5: Measures for bundle XL, task Main

run id P R MAPclefip-dcu baseline 0.0112 0.6599 0.084clefip-dcu Filtered2 0.0118 0.625 0.0935clefip-dcu Filtered 0.0118 0.6201 0.0759

clefip-ug infdocfreqCosEnglishTerms 0.0131 0.5452 0.0614clefip-unige RUN1 0.0102 0.5857 0.1156clefip-unige RUN2 0.0103 0.6013 0.1077clefip-unige RUN3 0.0102 0.5857 0.1157clefip-unige RUN4 0.0107 0.617 0.0837clefip-unige RUN5 0.0077 0.4688 0.0481

cwi bm25 0.0098 0.6063 0.0736cwi boolean 0.0528 0.3659 0.0258cwi category 0.0131 0.5396 0.0407hcuge BiTeM 0.0267 0.5788 0.0748

humb 1 0.0316 0.5089 0.1664TUD 800noTitle 0.0129 0.6486 0.0685

UniNE strat8 0.0088 0.5376 0.0337UTASICS abs-des-ratf-GT 0.0124 0.7142 0.098

UTASICS abs-des-ratf 0.0124 0.703 0.1034UTASICS abs-tit-cla-ratf-GT-ipcr 0.0124 0.7642 0.0956

UTASICS abs-tit-cla-ratf-ipcr 0.0121 0.7292 0.0874UTASICS all-ratf-GT-ipcr 0.0136 0.783 0.1245

UTASICS all-ratf-ipcr 0.0131 0.7616 0.099UTASICS all-tf-idf-GT-ipcr 0.0126 0.7378 0.122

UTASICS all-tf-idf-ipcr 0.0124 0.733 0.105

run id P R MAP

Evaluation Summary - Part Two 7

Tab

le6:

Measuresforbundle

S,task

Main

run

idP

P5

P10

P100

RR5

R10

R100

MAP

clefi

p-dcu

baseline

0.009

20.1

0.13

330.04

830.59

570.026

90.07

060.289

0.0615

clefi

p-dcu

Filtered2

0.008

50.13

330.13

330.04

170.57

90.057

0.10

230.2697

0.0729

clefi

p-dcu

Filtered

0.008

20.1

0.1

0.03

830.55

940.032

50.06

230.2471

0.0501

clefi

p-runClaim

sBOW

0.021

70.

0.03

330.02

170.19

710.

0.03

130.1971

0.0174

clefi

p-ugbm25

med

stan

dard

0.0.

0.0.

0.0.

0.

0.0.

clefi

p-uginfdocfreqBM25E

nglishTerms

0.0.

0.0.

0.0.

0.

0.0.

clefi

p-uginfdocfreqBM25F

renchTermsN

ew0.

0.0.

0.0.

0.

0.

0.0.

clefi

p-uginfdocfreqBM25

GermanTerms

0.0.

0.0.

0.0.

0.

0.0.

clefi

p-uginfdocfreqCosEnglishTerms

0.007

50.06

670.05

0.02

170.38

30.022

70.02

690.1316

0.0193

clefi

p-unigeRUN1

0.007

50.2

0.11

670.03

670.50

280.088

30.10

110.2759

0.0934

clefi

p-unigeRUN2

0.008

20.16

670.13

330.03

670.56

460.069

80.11

090.2872

0.0936

clefi

p-unigeRUN3

0.007

50.2

0.11

670.03

670.50

280.088

30.10

110.2759

0.0934

clefi

p-unigeRUN4

0.007

70.1

0.08

330.03

170.51

230.026

90.04

950.246

0.0413

clefi

p-unigeRUN5

0.004

50.1

0.06

670.01

670.35

10.049

90.06

270.1625

0.0395

cwibm25

0.007

20.06

670.05

0.01

830.51

040.008

30.02

120.0944

0.0208

cwiboolean

0.005

90.

0.0.00

670.31

080.

0.

0.0423

0.0031

cwicatego

rybm25

0.005

80.

0.01

670.01

0.38

10.

0.00

420.048

0.0066

cwicategory

0.006

20.03

330.05

0.02

50.47

440.004

20.02

120.1569

0.0197

hcu

geBiTeM

0.008

0.03

330.06

670.03

330.52

040.009

80.04

230.2698

0.0319

Hildesheim

MethodAnew

0.0.

0.0.

0.0.

0.

0.0.

humb1

0.007

90.23

330.15

0.03

330.40

250.161

20.18

910.3177

0.1631

NLELMethodA

0.005

80.1

0.13

330.03

170.37

0.029

80.09

660.2114

0.0574

TUD

800

noT

itle

0.008

80.1

0.06

670.04

170.56

990.021

20.02

530.2755

0.0466

UAIC

MethodAnew

0.000

20.03

330.01

670.00

170.01

280.012

80.01

280.0128

0.0026

UniN

Estrat1

0.005

0.06

670.06

670.01

830.36

550.028

30.05

10.1201

0.0346

UniN

Estrat2

0.007

30.13

330.08

330.03

50.55

330.068

40.07

820.3001

0.0779

UniN

Estrat3

0.009

80.1

0.08

330.04

170.56

810.041

20.06

380.3336

0.075

UniN

Estrat4

0.008

90.06

670.08

330.03

50.58

50.037

0.06

680.3084

0.0714

UniN

Estrat5

0.006

50.

0.03

330.03

170.45

440.

0.02

830.248

0.0197

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Con

tinued

onnextpage

Evaluation Summary - Part Two 8

Table

6:Measuresforbundle

S,task

Main,continued

run

idP

P5

P10

P100

RR5

R10

R100

MAP

UniN

Estrat6

0.009

60.13

330.06

670.03

330.64

110.051

0.05

10.2353

0.062

UniN

Estrat7

0.009

70.

0.03

330.02

830.65

140.

0.01

70.1738

0.0199

UniN

Estrat8

0.008

20.

0.05

0.02

330.55

950.

0.04

650.1758

0.0189

uscom

BM25

a0.008

40.16

670.13

330.03

330.54

020.075

50.12

540.3076

0.1165

uscom

BM25

at0.008

80.16

670.15

0.04

0.58

60.075

50.13

820.3386

0.1222

uscom

BM25b

0.009

60.1

0.05

0.03

330.60

120.026

80.02

680.2902

0.0416

uscom

BM25b

t0.009

90.1

0.08

330.03

330.60

390.029

80.05

240.2932

0.0503

uscom

BM25c

0.009

40.06

670.03

330.03

0.58

830.014

0.01

40.2732

0.032

uscom

BM25ct

0.009

60.03

330.06

670.03

170.58

990.004

20.03

960.2803

0.0372

uscom

BM25d

0.009

20.06

670.05

0.02

830.57

550.014

0.03

250.2603

0.0315

uscom

BM25d

t0.009

60.06

670.08

330.03

170.58

990.017

0.05

810.2803

0.0371

UTASIC

Sabs-des-ratf-GT

0.009

30.1

0.1

0.03

50.65

410.041

20.08

230.3084

0.0758

UTASIC

Sabs-des-ratf

0.009

0.03

330.03

330.03

330.61

180.018

50.03

130.3042

0.0447

UTASIC

Sab

s-tit-cla-ratf-G

T-ipcr

0.008

70.13

330.11

670.04

0.67

460.059

70.09

220.2765

0.0642

UTASIC

Sabs-tit-cla-ratf-ipcr

0.008

70.2

0.13

330.03

830.64

10.073

60.11

60.2958

0.0701

UTASIC

Sall-ratf-G

T-ipcr

0.009

30.23

330.15

0.04

170.67

50.092

20.11

480.3798

0.0953

UTASIC

Sall-ratf-ipcr

0.009

20.23

330.16

670.04

50.66

210.092

20.11

890.3846

0.0853

UTASIC

Sall-tf-idf-GT-ipcr

0.008

20.23

330.13

330.04

0.61

720.100

80.11

060.2863

0.1102

UTASIC

Sall-tf-idf-ipcr

0.008

70.23

330.13

330.04

670.63

530.100

80.11

060.3745

0.0981

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Table

7:Measuresforbundle

M,task

Main

run

idP

P5

P10

P100

RR5

R10

R100

MAP

clefi

p-dcu

baseline

0.00

810.07

50.1

0.04

370.59

750.020

10.053

0.3528

0.055

clefi

p-dcu

Filtered2

0.00

810.1

0.11

250.03

880.59

970.042

70.080

40.3383

0.0672

clefi

p-dcu

Filtered

0.00

810.07

50.07

50.03

620.59

240.024

40.046

70.3213

0.0484

clefi

p-ugbm25m

edstandard

0.00

090.

0.0.

0.02

570.

0.

0.0.0002

clefi

p-uginfdocfreqBM25

EnglishTerms

0.

0.0.

0.0.

0.0.

0.0.

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Continued

onnextpage

Evaluation Summary - Part Two 9

Tab

le7:Measuresforbundle

M,task

Main,continued

run

idP

P5

P10

P100

RR5

R10

R100

MAP

clefi

p-uginfdocfreqBM25F

renchTerms

0.

0.0.

0.0.

0.0.

0.0.

clefi

p-uginfdocfreqBM25G

erman

Terms

0.

0.0.

0.0.

0.0.

0.0.

clefi

p-uginfdocfreqCosE

nglishTerms

0.00

970.07

50.07

50.03

120.49

320.020

70.031

20.2188

0.0323

clefi

p-unigeRUN1

0.00

790.2

0.13

750.04

370.55

730.111

60.128

60.3687

0.0998

clefi

p-unigeRUN2

0.00

860.15

0.11

250.03

750.61

090.094

0.124

90.3588

0.0934

clefi

p-unigeRUN3

0.00

790.2

0.13

750.04

370.55

730.111

60.128

60.3687

0.0998

clefi

p-unigeRUN4

0.00

760.1

0.08

750.03

250.55

330.061

80.120

50.3242

0.0563

clefi

p-unigeRUN5

0.00

580.12

50.1

0.02

630.44

710.044

70.061

70.2003

0.0462

cwibm25

0.00

650.1

0.07

50.01

750.52

980.089

60.102

90.1578

0.0471

cwiboolean

0.01

860.02

50.01

250.00

620.32

010.003

70.003

70.0354

0.0036

cwicategorybm25

0.00

670.02

50.03

750.01

750.46

590.003

70.048

50.1414

0.0164

cwicatego

ry0.00

580.02

50.03

750.02

130.50

290.003

10.015

90.163

0.0164

hcu

geBiTeM

0.00

860.02

50.05

0.03

620.56

680.007

40.031

70.3494

0.0348

humb1

0.00

990.22

50.16

250.04

250.47

840.128

20.156

60.3658

0.149

TUD

800n

oTitle

0.00

950.12

50.1

0.04

880.62

60.023

20.033

70.372

0.0574

UniN

Estrat8

0.00

880.02

50.05

0.03

120.61

080.003

70.038

50.2862

0.0261

UTASIC

Sab

s-des-ratf-G

T0.01

010.1

0.12

50.04

250.69

650.034

60.076

50.3171

0.071

UTASIC

Sab

s-des-ratf

0.01

0.1

0.08

750.04

630.66

840.024

90.041

90.4046

0.0677

UTASIC

Sabs-tit-cla-ratf-G

T-ipcr

0.00

930.12

50.1

0.03

630.70

080.048

40.072

80.2257

0.0539

UTASIC

Sab

s-tit-cla-ratf-ipcr

0.00

930.17

50.11

250.03

50.67

560.058

90.090

60.2402

0.0583

UTASIC

Sall-ratf-G

T-ipcr

0.01

060.2

0.13

750.04

630.72

680.072

80.093

40.367

0.0851

UTASIC

Sall-ratf-ipcr

0.01

050.17

50.13

750.04

0.71

720.069

10.092

90.3449

0.0717

UTASIC

Sall-tf-idf-GT-ipcr

0.00

980.2

0.11

250.04

750.68

350.079

30.086

60.2662

0.0954

UTASIC

Sall-tf-idf-ipcr

0.00

990.17

50.1

0.04

50.68

970.075

60.083

0.3103

0.0792

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Evaluation Summary - Part Two 10

Tab

le8:

Measuresforbundle

XL,task

Main

run

idP

P5

P10

P100

RR5

R10

R100

MAP

clefi

p-dcu

baseline

0.011

20.08

330.14

170.06

750.65

990.021

80.066

40.4171

0.084

clefi

p-dcu

Filtered2

0.011

80.11

670.15

830.05

830.62

50.039

80.087

60.3797

0.0935

clefi

p-dcu

Filtered

0.011

80.1

0.12

50.05

670.62

010.027

60.062

20.3684

0.0759

clefi

p-uginfdocfreqCosE

nglishTerms

0.013

10.13

330.11

670.04

830.54

520.030

10.053

50.2673

0.0614

clefi

p-unigeRUN1

0.010

20.26

670.17

50.05

50.58

570.109

90.127

70.3619

0.1156

clefi

p-unigeRUN2

0.010

30.21

670.15

0.04

750.60

130.095

0.122

0.3381

0.1077

clefi

p-unigeRUN3

0.010

20.26

670.17

50.05

580.58

570.109

90.127

70.3652

0.1157

clefi

p-unigeRUN4

0.010

70.11

670.16

670.05

330.61

70.052

70.131

0.3557

0.0837

clefi

p-unigeRUN5

0.007

70.11

670.10

830.03

420.46

880.037

20.060

10.2076

0.0481

cwibm25

0.009

80.11

670.08

330.04

750.60

630.076

40.088

40.2642

0.0736

cwiboolean

0.052

80.03

330.04

170.03

170.36

590.006

60.018

20.1451

0.0258

cwicatego

ry0.013

10.08

330.07

50.03

920.53

960.016

40.033

20.2223

0.0407

hcu

geBiTeM

0.026

70.13

330.12

50.05

50.57

880.034

90.065

60.3712

0.0748

humb1

0.031

60.28

330.23

330.06

170.50

890.124

90.172

60.4129

0.1664

TUD

800

noT

itle

0.012

90.15

0.11

670.06

420.64

860.028

80.042

90.3951

0.0685

UniN

Estrat8

0.008

80.08

330.07

50.03

250.53

760.018

20.044

60.2393

0.0337

UTASIC

Sabs-des-ratf-GT

0.012

40.13

330.15

0.06

670.71

420.037

60.077

60.3852

0.098

UTASIC

Sab

s-des-ratf

0.012

40.18

330.12

50.07

080.70

30.047

0.061

30.4508

0.1034

UTASIC

Sabs-tit-cla-ratf-G

T-ipcr

0.012

40.2

0.16

670.05

920.76

420.059

90.093

80.32

0.0956

UTASIC

Sab

s-tit-cla-ratf-ipcr

0.012

10.18

330.14

170.05

670.72

920.057

70.093

60.3179

0.0874

UTASIC

Sall-ratf-G

T-ipcr

0.013

60.26

670.19

170.07

170.78

30.075

20.107

40.4331

0.1245

UTASIC

Sall-ratf-ipcr

0.013

10.16

670.17

50.06

0.76

160.061

50.107

30.379

0.099

UTASIC

Sall-tf-idf-GT-ipcr

0.012

60.23

330.15

830.07

080.73

780.073

50.090

60.36

0.122

UTASIC

Sall-tf-idf-ipcr

0.012

40.2

0.14

170.06

580.73

30.073

0.091

40.3716

0.105

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Evaluation Summary - Part Two 11

4.2 Language Tasks

Table 9: Measures for bundle S, task DE

run id P R MAPhcuge BiTeM 0.0073 0.4622 0.0242

humb 1 0.0073 0.3859 0.0373TUD 800noTitle 0.0085 0.5991 0.0516

run id P R MAP

Table 10: Measures for bundle M, task DE

run id P R MAPhcuge BiTeM 0.0073 0.4937 0.0231

humb 1 0.0094 0.4659 0.0597TUD 800noTitle 0.009 0.6405 0.0596

run id P R MAP

Table 11: Measures for bundle XL, task DE

run id P R MAPhumb 1 0.0315 0.5095 0.0782

TUD 800noTitle 0.0134 0.6489 0.0753

run id P R MAP

Evaluation Summary - Part Two 12

Tab

le12

:Measuresforbundle

S,task

DE

run

idP

P5

P10

P100

RR5

R10

R100

MAP

hcu

geBiTeM

0.00

730.033

30.06

670.02

670.46

220.00

980.065

40.20

480.0242

humb1

0.00

730.066

70.08

330.01

330.38

590.03

130.115

30.20

460.0373

TUD

800n

oTitle

0.00

850.1

0.11

670.03

830.59

910.02

120.043

50.26

710.0516

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Tab

le13

:Measuresforbundle

M,task

DE

run

idP

P5

P10

P100

RR5

R10

R100

MAP

hcu

geBiTeM

0.00

730.02

50.05

0.02

370.49

370.00

740.049

10.27

860.02

31humb1

0.00

940.12

50.12

50.02

750.46

590.03

450.104

90.28

090.05

97TUD

800n

oTitle

0.00

90.15

0.13

750.04

50.64

050.02

690.047

30.32

410.05

96

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Table

14:Measuresforbundle

XL,task

DE

run

idP

P5

P10

P100

RR5

R10

R100

MAP

humb1

0.03

150.166

70.13

330.04

080.50

950.05

210.108

30.31

030.0782

TUD

800n

oTitle

0.01

340.166

70.15

830.06

170.64

890.03

130.064

50.36

340.0753

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Evaluation Summary - Part Two 13

Table 15: Measures for bundle S, task EN

run id P R MAPclefip-run ClaimsBOW 0.0217 0.1971 0.0174

hcuge BiTeM 0.0069 0.4396 0.0272humb 1 0.0079 0.4025 0.1414

TUD 800noTitle 0.0087 0.5427 0.0407UAIC MethodAnew 0.0002 0.0128 0.0026

run id P R MAP

Table 16: Measures for bundle M, task EN

run id P R MAPhcuge BiTeM 0.0078 0.5062 0.0292

humb 1 0.0099 0.4784 0.1137TUD 800noTitle 0.009 0.5945 0.0429

run id P R MAP

Table 17: Measures for bundle XL, task EN

run id P R MAPhumb 1 0.0319 0.5208 0.1319

TUD 800noTitle 0.0127 0.6244 0.0595

run id P R MAP

Evaluation Summary - Part Two 14

Tab

le18

:Measuresforbundle

S,task

EN

run

idP

P5

P10

P100

RR5

R10

R100

MAP

clefi

p-runClaim

sBOW

0.02

170.

0.03

330.02

170.19

710.

0.03

130.197

10.0174

hcu

geBiTeM

0.00

690.

0.01

670.02

830.43

960.

0.00

420.246

50.0272

humb1

0.00

790.2

0.11

670.02

830.40

250.14

260.14

680.271

30.1414

TUD

800n

oTitle

0.00

870.06

670.08

330.03

50.54

270.01

70.08

660.219

20.0407

UAIC

MethodAnew

0.00

020.03

330.01

670.00

170.01

280.01

280.01

280.012

80.0026

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Table

19:Measuresforbundle

M,task

EN

run

idP

P5

P10

P100

RR5

R10

R100

MAP

hcu

geBiTeM

0.00

780.

0.01

250.03

380.50

620.

0.003

10.33

560.02

92humb1

0.00

990.15

0.08

750.03

120.47

840.10

70.110

10.34

680.11

37TUD

800n

oTitle

0.00

90.07

50.07

50.03

880.59

450.01

640.068

60.31

510.04

29

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Tab

le20

:Measuresforbundle

XL,task

EN

run

idP

P5

P10

P100

RR5

R10

R100

MAP

humb1

0.03

190.216

70.14

170.05

50.52

080.10

150.113

0.40

340.1319

TUD

800n

oTitle

0.01

270.1

0.10

830.05

920.62

440.02

730.075

50.35

680.0595

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Evaluation Summary - Part Two 15

Table 21: Measures for bundle S, task FR

run id P R MAPhcuge BiTeM 0.0056 0.4018 0.0425

humb 1 0.0064 0.365 0.067TUD 800noTitle 0.0083 0.5488 0.0296

run id P R MAP

Table 22: Measures for bundle M, task FR

run id P R MAPhcuge BiTeM 0.0072 0.4704 0.038

humb 1 0.0087 0.4503 0.0551TUD 800noTitle 0.0082 0.5844 0.0264

run id P R MAP

Table 23: Measures for bundle XL, task FR

run id P R MAPhumb 1 0.0311 0.499 0.0786

TUD 800noTitle 0.0134 0.6209 0.0458

run id P R MAP

Evaluation Summary - Part Two 16

Table

24:Measuresforbundle

S,task

FR

run

idP

P5

P10

P100

RR5

R10

R100

MAP

hcu

geBiTeM

0.00

560.033

30.03

330.02

50.40

180.04

170.045

80.27

780.0425

humb1

0.00

640.133

30.08

330.01

330.36

50.07

370.115

30.20

460.067

TUD

800n

oTitle

0.00

830.066

70.06

670.03

330.54

880.02

270.059

30.21

810.0296

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Tab

le25

:Measuresforbundle

M,task

FR

run

idP

P5

P10

P100

RR5

R10

R100

MAP

hcu

geBiTeM

0.00

720.02

50.02

50.02

870.47

040.03

120.034

40.35

170.038

humb1

0.00

870.1

0.06

250.01

880.45

030.05

530.086

50.21

720.05

51TUD

800n

oTitle

0.00

820.05

0.05

0.03

250.58

440.01

70.044

50.26

160.02

64

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Table

26:Measuresforbundle

XL,task

FR

run

idP

P5

P10

P100

RR5

R10

R100

MAP

humb1

0.03

110.15

0.12

50.03

250.49

90.06

570.114

20.25

890.07

86TUD

800n

oTitle

0.01

340.08

330.07

50.04

830.62

090.02

150.046

0.30

20.04

58

run

idP

P5

P10

P100

RR5

R10

R100

MAP

Evaluation Summary - Part Two 17

References

[1] Florina Piroi, Giovanna Roda, and Veronika Zenz. CLEF-IP 2009 Evaluation Summary. July2009.

Evaluation Summary - Part Two 18

List of Tables

1 List of active participants and runs submitted . . . . . . . . . . . . . . . . . . . . 11 List of active participants and runs submitted . . . . . . . . . . . . . . . . . . . . 22 Topics manually assessed by patent experts . . . . . . . . . . . . . . . . . . . . . 23 Bundle S, task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44 Bundle M, task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 55 Bundle XL, task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66 Bundle S , task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77 Bundle M , task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88 Bundle XL , task Main . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109 Bundle S, task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1110 Bundle M, task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1111 Bundle XL, task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1112 Bundle S , task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1213 Bundle M , task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1214 Bundle XL , task DE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1215 Bundle S, task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1316 Bundle M, task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1317 Bundle XL, task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1318 Bundle S , task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1419 Bundle M , task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1420 Bundle XL , task EN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1421 Bundle S, task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1522 Bundle M, task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1523 Bundle XL, task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1524 Bundle S , task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1625 Bundle M , task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1626 Bundle XL , task FR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16