53
Deep Semantic Role Labeling: What works and what’s next Luheng He , Kenton Lee , Mike Lewis and Luke Zettlemoyer †* Paul G. Allen School of Computer Science & Engineering, Univ. of Washington, Facebook AI Research * Allen Institute for Artificial Intelligence 1

Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Deep Semantic Role Labeling:What works and what’s next

Luheng He†, Kenton Lee†, Mike Lewis ‡ and Luke Zettlemoyer†*

† Paul G. Allen School of Computer Science & Engineering, Univ. of Washington, ‡ Facebook AI Research

* Allen Institute for Artificial Intelligence

1

Page 2: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Semantic Role Labeling (SRL)

2

Question Answering Information Extraction Machine Translation

Applications

predicate argument

role label

who what when where why …

Page 3: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

My mug broke into pieces.

The robot broke my mug with a wrench.

Semantic Role Labeling (SRL) - Example

3

Page 4: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

My mug broke into pieces.

The robot broke my mug with a wrench.

Semantic Role Labeling (SRL) - Example

3

v obj

subj v

thing broken

thing broken

breaker instrument

pieces (final state)

Page 5: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

My mug broke into pieces.

The robot broke my mug with a wrench.

Semantic Role Labeling (SRL) - Example

3

v obj

Frame: break.01role description

ARG0 breakerARG1 thing brokenARG2 instrumentARG3 pieces

… …

subj v

thing broken

thing broken

breaker instrument

pieces (final state)

ARG0 ARG1 ARG2

ARG3ARG1

Page 6: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

4

The Proposition Bank (PropBank)

Core roles: Verb-specific roles (ARG0-

ARG5) defined in frame files

Frame: break.01

role descriptionARG0 breakerARG1 thing broken

ARG2 instrument

ARG3 pieces

ARG4 broken away from what?

Frame: buy.01

role descriptionARG0 buyerARG1 thing bough

ARG2 seller

ARG3 price paid

ARG4 benefactive

role descriptionTMP temporalLOC locationMNR mannerDIR directionCAU causePRP purpose…

Adjunct roles: (ARGM-) shared

across verbs

4

Paul Kingsbury and Martha Palmer.From Treebank to PropBank. 2002

Page 7: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

4

The Proposition Bank (PropBank)

Core roles: Verb-specific roles (ARG0-

ARG5) defined in frame files

Frame: break.01

role descriptionARG0 breakerARG1 thing broken

ARG2 instrument

ARG3 pieces

ARG4 broken away from what?

Frame: buy.01

role descriptionARG0 buyerARG1 thing bough

ARG2 seller

ARG3 price paid

ARG4 benefactive

role descriptionTMP temporalLOC locationMNR mannerDIR directionCAU causePRP purpose…

Adjunct roles: (ARGM-) shared

across verbs

Annotated on top of the Penn Treebank Syntax

PropBank Annotation Guidelines, Bonial et al., 2010

4

Paul Kingsbury and Martha Palmer.From Treebank to PropBank. 2002

Page 8: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

5

SRL Systems

syntactic features

candidate argument spans

labeled arguments

prediction

labeling

ILP/DP

sentence, predicate

argument id.

Pipeline Systems

Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et al., 2015

Page 9: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

5

SRL Systems

syntactic features

candidate argument spans

labeled arguments

prediction

labeling

ILP/DP

sentence, predicate

argument id.

Pipeline Systems

Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et al., 2015

sentence, predicate

BIO sequence

prediction

Deep BiLSTM + CRF layer

Viterbi

context window features

End-to-end Systems

Collobert et al., 2011 Zhou and Xu, 2015 Wang et. al, 2015

Page 10: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

5

SRL Systems

syntactic features

candidate argument spans

labeled arguments

prediction

labeling

ILP/DP

sentence, predicate

argument id.

Pipeline Systems

Deep BiLSTM

Hard constraints

BIO sequence

prediction

sentence, predicate

*This work

Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et al., 2015

sentence, predicate

BIO sequence

prediction

Deep BiLSTM + CRF layer

Viterbi

context window features

End-to-end Systems

Collobert et al., 2011 Zhou and Xu, 2015 Wang et. al, 2015

Page 11: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

6

the cats love hats[ ] [ ] [V] [ ]

B-ARG0 0.4I-ARG0 0.05B-ARG1 0.5I-ARG1 0.03

… …

B-ARG0 0.1I-ARG0 0.5B-ARG1 0.1I-ARG1 0.2

… …

B-ARG0 0.001I-ARG0 0.001B-ARG1 0.001

… …B-V 0.95

B-ARG0 0.1I-ARG0 0.1B-ARG1 0.7I-ARG1 0.2

… …

Page 12: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

6

the cats love hats[ ] [ ] [V] [ ]

B-ARG0 0.4I-ARG0 0.05B-ARG1 0.5I-ARG1 0.03

… …

B-ARG0 0.1I-ARG0 0.5B-ARG1 0.1I-ARG1 0.2

… …

B-ARG0 0.001I-ARG0 0.001B-ARG1 0.001

… …B-V 0.95

B-ARG0 0.1I-ARG0 0.1B-ARG1 0.7I-ARG1 0.2

… …

(1) Deep BiLSTM tagger

Page 13: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

6

the cats love hats[ ] [ ] [V] [ ]

B-ARG0 0.4I-ARG0 0.05B-ARG1 0.5I-ARG1 0.03

… …

B-ARG0 0.1I-ARG0 0.5B-ARG1 0.1I-ARG1 0.2

… …

B-ARG0 0.001I-ARG0 0.001B-ARG1 0.001

… …B-V 0.95

B-ARG0 0.1I-ARG0 0.1B-ARG1 0.7I-ARG1 0.2

… …

(1) Deep BiLSTM tagger

(2) Highway connections

Page 14: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

6

the cats love hats[ ] [ ] [V] [ ]

B-ARG0 0.4I-ARG0 0.05B-ARG1 0.5I-ARG1 0.03

… …

B-ARG0 0.1I-ARG0 0.5B-ARG1 0.1I-ARG1 0.2

… …

B-ARG0 0.001I-ARG0 0.001B-ARG1 0.001

… …B-V 0.95

B-ARG0 0.1I-ARG0 0.1B-ARG1 0.7I-ARG1 0.2

… …

(1) Deep BiLSTM tagger

(2) Highway connections

(3) Viterbi decoding with

hard constraints

Page 15: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

7

input from the previous layer

recurrent input

from the prev. timestep

output to the next layer

References: Deep Residual Networks, Kaiming He, ICML 2016 Tutorial

Training Very Deep Networks, Srivastava et al., 2015

Non-linearity

ht

ht�1 F(ct�1,ht�1,xt)

xt

Model - Highway Connections

Page 16: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

7

input from the previous layer

recurrent input

from the prev. timestep

output to the next layer

References: Deep Residual Networks, Kaiming He, ICML 2016 Tutorial

Training Very Deep Networks, Srivastava et al., 2015

Non-linearity

ht shortcut

ht�1 F(ct�1,ht�1,xt)

xt

xt

Model - Highway Connections

Page 17: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

7

input from the previous layer

recurrent input

from the prev. timestep

output to the next layer

References: Deep Residual Networks, Kaiming He, ICML 2016 Tutorial

Training Very Deep Networks, Srivastava et al., 2015

Non-linearity

ht shortcut

ht�1 F(ct�1,ht�1,xt)

gated highway network:

rt � ht + (1� rt) � xt

rt = �(f(ht�1,xt))

xt

xtnew output:

Model - Highway Connections

Page 18: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Softmax

BiLSTM layers …

8

B-ARG1 I-ARG0 B-V B-ARG1Greedy Output

argmax

the cats love hats[ ] [ ] [V] [ ]

Model - Viterbi Decoding with Hard Constraints

Page 19: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Softmax

BiLSTM layers …

8

BIO inconsistency: invalid transition

B-ARG1 I-ARG0 B-V B-ARG1Greedy Output

argmax

the cats love hats[ ] [ ] [V] [ ]

Model - Viterbi Decoding with Hard Constraints

Page 20: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Softmax

BiLSTM layers …

8

BIO inconsistency: invalid transition

s(B-ARG0 ! I-ARG0) =0

s(B-ARG1 ! I-ARG0) =� inf

· · ·

Heuristic transition scores

B-ARG1 I-ARG0 B-V B-ARG1Greedy Output

argmax

the cats love hats[ ] [ ] [V] [ ]

Model - Viterbi Decoding with Hard Constraints

Page 21: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Softmax

BiLSTM layers …

8

B-ARG0 0.4I-ARG0 0.05B-ARG1 0.5I-ARG1 0.03

… …O 0.01

B-ARG0 0.1I-ARG0 0.5B-ARG1 0.1I-ARG1 0.2

… …O 0.05

B-ARG0 0.001I-ARG0 0.001B-ARG1 0.001I-ARG1 0.002

… …B-V 0.95

B-ARG0 0.1I-ARG0 0.1B-ARG1 0.7I-ARG1 0.2

… …O 0.05

s(B-ARG0 ! I-ARG0) =0

s(B-ARG1 ! I-ARG0) =� inf

· · ·

Heuristic transition scores

Viterbi decoding

the cats love hats[ ] [ ] [V] [ ]

Model - Viterbi Decoding with Hard Constraints

Page 22: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Other Implementation Details …

9

• 8 layer BiLSTMs with 300D hidden layers.

• 100D GloVe embeddings, updated during training.

• Orthonormal initialization for LSTM weight matrices (Saxe et al., 2013)

• 0.1 variational dropout between layers (Gal and Ghahramani, 2016)

• Trained for 500 epochs.

Page 23: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

10

CoNLL-2005(PropBank)

CoNLL-2012(OntoNotes)

Size 40k sentences 140k sentences

Domains• WSJ / newswire • Brown (test-only)

• telephone conversations • newswire • newsgroups • broadcast news • broadcast conversation • weblogs

Annotated predicates Verbs Added some nominal

predicates

Datasets CoNLL 2005 Results

CoNLL 2012 (OntoNotes) Results Ablations

Page 24: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

11

F1

60

65

70

75

80

85

90

Ours* Ours Zhou FitzGerald* Täckström Toutanova* Punyakanok*

WSJ Test Brown (out-domain) Test *:Ensemble models

2017 2015 2015 20152017 2008 2008

Datasets CoNLL 2005 Results CoNLL 2012 (OntoNotes) Results Ablations

Page 25: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

11

F1

60

65

70

75

80

85

90

Ours* Ours Zhou FitzGerald* Täckström Toutanova* Punyakanok*

WSJ Test Brown (out-domain) Test

79.480.379.980.382.883.184.6

*:Ensemble models

2017 2015 2015 20152017 2008 2008

Datasets CoNLL 2005 Results CoNLL 2012 (OntoNotes) Results Ablations

Page 26: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

11

F1

60

65

70

75

80

85

90

Ours* Ours Zhou FitzGerald* Täckström Toutanova* Punyakanok*

WSJ Test Brown (out-domain) Test

67.868.871.372.2

69.472.173.6

79.480.379.980.382.883.184.6

*:Ensemble models

2017 2015 2015 20152017 2008 2008

Datasets CoNLL 2005 Results CoNLL 2012 (OntoNotes) Results Ablations

Page 27: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

11

F1

60

65

70

75

80

85

90

Ours* Ours Zhou FitzGerald* Täckström Toutanova* Punyakanok*

WSJ Test Brown (out-domain) Test

67.868.871.372.2

69.472.173.6

79.480.379.980.382.883.184.6

Pipeline modelsBiLSTM models

*:Ensemble models

2017 2015 2015 20152017 2008 2008

Datasets CoNLL 2005 Results CoNLL 2012 (OntoNotes) Results Ablations

Page 28: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

12

F1

60

65

70

75

80

85

90

Ours* Ours Zhou FitzGerald* Täckström Pradhan

77.579.480.281.581.783.4

CoNLL 2012 Test

Pipeline modelsBiLSTM models

*:Ensemble models

2017 2015 2015 20152017 2013

Datasets CoNLL 2005 Results CoNLL 2012 (OntoNotes) Results Ablations

Page 29: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

F1 o

n D

ev. S

et

60

65

70

75

80

85

Num. Epochs

1 50 100 150 200 250 300 350 400 450 500

Full model No highway No orthonormal init. No dropout

(single model, on CoNLL05 Dev)

13

Without dropout, model overfits at ~300 epochs.

Without orthonormal initialization, the deep model learns very slowly

Datasets CoNLL 2005 Results

CoNLL 2012 (OntoNotes) Results Ablations

Page 30: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

What can we learn from the results?

14

F1

60657075808590

Ours* Ours Zhou FitzGerald* Täckström Pradhan

77.579.480.281.581.783.4

1. What’s in the remaining 17%? When does the model still struggle?

Page 31: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

What can we learn from the results?

14

F1

60657075808590

Ours* Ours Zhou FitzGerald* Täckström Pradhan

77.579.480.281.581.783.4

1. What’s in the remaining 17%? When does the model still struggle?

2. BiLSTM-based models are very accurate even without syntax. But can we conclude syntax is no longer useful in SRL?

Page 32: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

15

Question (1): When does the model make mistakes?

Analysis— Error breakdown with oracle transformation — E.g. tease apart labeling errors and boundary errors — Link the error types to known linguistic phenomena, e.g. prepositional phrase (PP) attachment

Page 33: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Oracle Transformations (Error Breakdown)

16

[We] fly to NYC tomorrow.ARG0ARG1

Fix Label:

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Oracle Transformations

Page 34: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Oracle Transformations (Error Breakdown)

16

[We] fly to NYC tomorrow.ARG0ARG1

Fix Label:

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Oracle Transformations

Labeling error 29%

Page 35: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Oracle Transformations (Error Breakdown)

16

I eat [pasta with delight].ARG1

[pasta] [with delight]ARG1 ARGM-MNR

[We] fly to NYC tomorrow.ARG0ARG1

Fix Label:

Split/Merge span:

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Oracle Transformations

Labeling error 29%

I eat [pasta] [with broccoli].

ARG1[pasta with broccoli]

ARG1 ARGM-MNR

Page 36: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Oracle Transformations (Error Breakdown)

16

I eat [pasta with delight].ARG1

[pasta] [with delight]ARG1 ARGM-MNR

[We] fly to NYC tomorrow.ARG0ARG1

Fix Label:

Split/Merge span:

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Oracle Transformations

Labeling error 29%

Attachment error 25%

I eat [pasta] [with broccoli].

ARG1[pasta with broccoli]

ARG1 ARGM-MNR

Page 37: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

17

Confusion matrix for labeling errors

(column normalized)

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Page 38: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

17

Confusion matrix for labeling errors

(column normalized)

• ARG2 is often confused with certain adjuncts (DIR, LOC, MNR), why?

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Page 39: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Predicate: cut

        Arg0-PAG: intentional cutter        Arg1-PPT: thing cut        Arg2-DIR: medium, source        Arg3-MNR: instrument, unintentional cutter        Arg4-GOL: beneficiary

17

Confusion matrix for labeling errors

(column normalized)

Predicate: move

Arg0-PAG: mover        Arg1-PPT: moved        Arg2-GOL: destination        Arg3-VSP: aspect, domain in which arg1 moving

Predicate: strike

        Arg0-PAG: Agent         Arg1-PPT: Theme(-Creation)         Arg2-MNR: Instrument

• ARG2 is often confused with certain adjuncts (DIR, LOC, MNR), why?

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Page 40: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

Predicate: cut

        Arg0-PAG: intentional cutter        Arg1-PPT: thing cut        Arg2-DIR: medium, source        Arg3-MNR: instrument, unintentional cutter        Arg4-GOL: beneficiary

17

Confusion matrix for labeling errors

(column normalized)

Predicate: move

Arg0-PAG: mover        Arg1-PPT: moved        Arg2-GOL: destination        Arg3-VSP: aspect, domain in which arg1 moving

Predicate: strike

        Arg0-PAG: Agent         Arg1-PPT: Theme(-Creation)         Arg2-MNR: Instrument

• ARG2 is often confused with certain adjuncts (DIR, LOC, MNR), why?

• Argument-adjunct distinctions are difficult even for expert annotators!

Error Breakdown Labeling Errors PP Attachment Can Syntax Still Help?

Page 41: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Page 42: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Wrong PP attachment (attach high)

Correct PP attachment (attach low)

Arg2 (PP)Arg1 (NP)

Arg1 (NP)

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Page 43: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Wrong PP attachment (attach high)

Correct PP attachment (attach low)

Arg2 (PP)Arg1 (NP)

Arg1 (NP)

Wrong SRL spans

Correct SRL spans

merge

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Page 44: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Wrong PP attachment (attach high)

Correct PP attachment (attach low)

Arg2 (PP)Arg1 (NP)

Arg1 (NP)

Wrong SRL spans

Correct SRL spans

merge

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Attachment mistakes: 25%.

Categorize the Y spans in : [XY]—>[X][Y] and

[X][Y]—>[XY] operations by gold syntactic labels

Page 45: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Wrong PP attachment (attach high)

Correct PP attachment (attach low)

Arg2 (PP)Arg1 (NP)

Arg1 (NP)

Wrong SRL spans

Correct SRL spans

merge

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Attachment mistakes: 25%.

Categorize the Y spans in : [XY]—>[X][Y] and

[X][Y]—>[XY] operations by gold syntactic labels

Page 46: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

18

Sumimoto financed the acquisition from Sears

Wrong PP attachment (attach high)

Correct PP attachment (attach low)

Arg2 (PP)Arg1 (NP)

Arg1 (NP)

Wrong SRL spans

Correct SRL spans

merge

Error Breakdown Labeling Errors PP Attachment Can Syntax Still

Help?

Attachment mistakes: 25%.

Categorize the Y spans in : [XY]—>[X][Y] and

[X][Y]—>[XY] operations by gold syntactic labels

Takeaway— Traditionally hard tasks, such as argument-adjunct distinction and PP attachment decisions are still challenging! — Use external information/PropBank frame inventory.

Page 47: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

19

Question (2): Can syntax still help SRL?

Recap— PropBank SRL is annotated on top of the PTB syntax. — More than 98% of the gold SRL spans are syntactic constituents. Analysis— At decoding time, make predicted argument spans agree with given syntactic structure (unlabeled). — See if SRL performance increases.

Page 48: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

20

[The cats] love [hats and the dogs] love bananas. ARG0 ARG1

[The cats][hats and the dogs]

2 Syntax Tree

62 Syntax Tree

Error Breakdown

Labeling Errors

PP Attachment

Long-range Dependencies Can Syntax Still Help?

Constrained Decoding with Syntax

Penalize sequence score

Page 49: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

20

[The cats] love [hats and the dogs] love bananas. ARG0 ARG1

[The cats][hats and the dogs]

Sequence score:

2 Syntax Tree

62 Syntax Tree

Error Breakdown

Labeling Errors

PP Attachment

Long-range Dependencies Can Syntax Still Help?

Constrained Decoding with Syntax

tX

i=1

log p(tagt | sentence)� C ⇥X

span

1(span 62 Syntax Tree)

Penalize sequence score

Num. arguments disagree w\ syntax

Penalty strength

Page 50: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

20

• Constraints are not locally decomposable. • A* search (Lewis and Steedman 2014) for a sequence with highest score.

[The cats] love [hats and the dogs] love bananas. ARG0 ARG1

[The cats][hats and the dogs]

Sequence score:

2 Syntax Tree

62 Syntax Tree

Error Breakdown

Labeling Errors

PP Attachment

Long-range Dependencies Can Syntax Still Help?

Constrained Decoding with Syntax

tX

i=1

log p(tagt | sentence)� C ⇥X

span

1(span 62 Syntax Tree)

Penalize sequence score

Num. arguments disagree w\ syntax

Penalty strength

Page 51: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

21

Gold: Penn Treebank constituents. Choe: Parsing as language modeling, Choe and Charniak, 2016 (SOTA) Charniak: A maximum-entropy-inspired parser, Charniak, 2000

Error Breakdown

Labeling Errors

PP Attachment

Long-range Dependencies Can Syntax Still Help?

Syntax Decoding Results

Page 52: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

21

Gold: Penn Treebank constituents. Choe: Parsing as language modeling, Choe and Charniak, 2016 (SOTA) Charniak: A maximum-entropy-inspired parser, Charniak, 2000

Error Breakdown

Labeling Errors

PP Attachment

Long-range Dependencies Can Syntax Still Help?

Syntax Decoding Results

Takeaway— Modest gain when using accurate syntax. — More improvement: Joint training, use syntactic labels, etc.

Page 53: Deep Semantic Role Labeling: What works and what’s next · ILP/DP sentence, predicate argument id. Pipeline Systems Punyakanok et al., 2008 Täckström et al., 2015 FitzGerald et

22

Thank You!

• New state-of-the-art deep network for end-to-end SRL.

• Code and models are publicly available at: https://github.com/luheng/deep_srl

• In-depth error analysis indicating where the models work well and where they still struggle.

• Syntax-based experiments pointing towards directions for future improvements.