13
Semantic Neural Machine Translation using AMR Linfeng Song 1 , Daniel Gildea 1 , Yue Zhang 2 , Zhiguo Wang 3 and Jinsong Su 4 1 Department of Computer Science, University of Rochester, Rochester, NY 14627 2 School of Engineering, Westlake University, China 3 IBM T.J. Watson Research Center, Yorktown Heights, NY 10598 4 Xiamen University, Xiamen, China 1 {lsong10,gildea}@cs.rochester.edu 2 [email protected] 3 [email protected] 4 [email protected] Abstract It is intuitive that semantic representa- tions can be useful for machine transla- tion, mainly because they can help in en- forcing meaning preservation and handling data sparsity (many sentences correspond to one meaning) of machine translation mod- els. On the other hand, little work has been done on leveraging semantics for neural ma- chine translation (NMT). In this work, we study the usefulness of AMR (short for ab- stract meaning representation) on NMT. Ex- periments on a standard English-to-German dataset show that incorporating AMR as additional knowledge can significantly im- prove a strong attention-based sequence-to- sequence neural translation model. 1 Introduction It is intuitive that semantic representations ought to be relevant to machine translation, given that the task is to produce a target language sentence with the same meaning as the source language in- put. Semantic representations formed the core of the earliest symbolic machine translation systems, and have been applied to statistical but non-neural systems as well. Leveraging syntax for neural machine trans- lation (NMT) has been an active research topic (Stahlberg et al., 2016; Aharoni and Goldberg, 2017; Li et al., 2017; Chen et al., 2017; Bast- ings et al., 2017; Wu et al., 2017; Chen et al., 2018). On the other hand, exploring semantics for NMT has so far received relatively little attention. Recently, Marcheggiani et al. (2018) exploited semantic role labeling (SRL) for NMT, show- ing that the predicate-argument information from SRL can improve the performance of an attention- based sequence-to-sequence model by alleviating the “argument switching” problem, 1 one frequent 1 flipping arguments corresponding to different roles A0 (a) John gave his beautiful wife a nice present . A1 A2 John give-01 person have-rel-rol-91 present nice :ARG0 :ARG2 wife :ARG0-of :ARG1 :mod :ARG1 :ARG2 beautiful :mod (b) person name :name :op1 Figure 1: (a) A sentence with semantic roles annota- tions, (b) the corresponding AMR graph of that sen- tence. and severe issue faced by NMT systems (Isabelle et al., 2017). Figure 1 (a) shows one example of semantic role information, which only captures the relations between a predicate (gave) and its ar- guments (John, wife and present). Other important information, such as the relation between John and wife, can not be incorporated. In this paper, we explore the usefulness of ab- stract meaning representation (AMR) (Banarescu et al., 2013) as a semantic representation for NMT. AMR is a semantic formalism that encodes the meaning of a sentence as a rooted, directed graph. Figure 1 (b) shows an AMR graph, in which the nodes (such as give-01 and John) represent the concepts, and edges (such as :ARG0 and :ARG1) represent the relations between concepts they con- nect. Comparing with semantic roles, AMRs cap- ture more relations, such as the relation between John and wife (represented by the subgraph within dotted lines). In addition, AMRs directly capture entity relations and abstract away inflections and function words. As a result, they can serve as a source of knowledge for machine translation that is orthogonal to the textual input. Furthermore, arXiv:1902.07282v1 [cs.CL] 19 Feb 2019

Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

  • Upload
    others

  • View
    16

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

Semantic Neural Machine Translation using AMR

Linfeng Song1, Daniel Gildea1, Yue Zhang2, Zhiguo Wang3 and Jinsong Su4

1Department of Computer Science, University of Rochester, Rochester, NY 146272School of Engineering, Westlake University, China

3IBM T.J. Watson Research Center, Yorktown Heights, NY 105984Xiamen University, Xiamen, China

1{lsong10,gildea}@cs.rochester.edu [email protected]@gmail.com [email protected]

Abstract

It is intuitive that semantic representa-tions can be useful for machine transla-tion, mainly because they can help in en-forcing meaning preservation and handlingdata sparsity (many sentences correspond toone meaning) of machine translation mod-els. On the other hand, little work has beendone on leveraging semantics for neural ma-chine translation (NMT). In this work, westudy the usefulness of AMR (short for ab-stract meaning representation) on NMT. Ex-periments on a standard English-to-Germandataset show that incorporating AMR asadditional knowledge can significantly im-prove a strong attention-based sequence-to-sequence neural translation model.

1 Introduction

It is intuitive that semantic representations oughtto be relevant to machine translation, given thatthe task is to produce a target language sentencewith the same meaning as the source language in-put. Semantic representations formed the core ofthe earliest symbolic machine translation systems,and have been applied to statistical but non-neuralsystems as well.

Leveraging syntax for neural machine trans-lation (NMT) has been an active research topic(Stahlberg et al., 2016; Aharoni and Goldberg,2017; Li et al., 2017; Chen et al., 2017; Bast-ings et al., 2017; Wu et al., 2017; Chen et al.,2018). On the other hand, exploring semantics forNMT has so far received relatively little attention.Recently, Marcheggiani et al. (2018) exploitedsemantic role labeling (SRL) for NMT, show-ing that the predicate-argument information fromSRL can improve the performance of an attention-based sequence-to-sequence model by alleviatingthe “argument switching” problem,1 one frequent

1flipping arguments corresponding to different roles

A0(a)

John gave his beautiful wife a nice present .

A1A2

John

give-01

person

have-rel-rol-91

present

nice

:ARG0

:ARG2

wife

:ARG0-of

:ARG1

:mod

:ARG1

:ARG2

beautiful:mod

(b)

person

name

:name

:op1

Figure 1: (a) A sentence with semantic roles annota-tions, (b) the corresponding AMR graph of that sen-tence.

and severe issue faced by NMT systems (Isabelleet al., 2017). Figure 1 (a) shows one exampleof semantic role information, which only capturesthe relations between a predicate (gave) and its ar-guments (John, wife and present). Other importantinformation, such as the relation between John andwife, can not be incorporated.

In this paper, we explore the usefulness of ab-stract meaning representation (AMR) (Banarescuet al., 2013) as a semantic representation for NMT.AMR is a semantic formalism that encodes themeaning of a sentence as a rooted, directed graph.Figure 1 (b) shows an AMR graph, in which thenodes (such as give-01 and John) represent theconcepts, and edges (such as :ARG0 and :ARG1)represent the relations between concepts they con-nect. Comparing with semantic roles, AMRs cap-ture more relations, such as the relation betweenJohn and wife (represented by the subgraph withindotted lines). In addition, AMRs directly captureentity relations and abstract away inflections andfunction words. As a result, they can serve as asource of knowledge for machine translation thatis orthogonal to the textual input. Furthermore,

arX

iv:1

902.

0728

2v1

[cs

.CL

] 1

9 Fe

b 20

19

Page 2: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

structural information from AMR graphs can helpreduce data sparsity, when training data is not suf-ficient for large-scale training.

Recent advances in AMR parsing keep push-ing the boundary of state-of-the-art performance(Flanigan et al., 2014; Artzi et al., 2015; Pustet al., 2015; Peng et al., 2015; Flanigan et al.,2016; Buys and Blunsom, 2017; Konstas et al.,2017; Wang and Xue, 2017; Lyu and Titov, 2018;Peng et al., 2018; Groschwitz et al., 2018; Guoand Lu, 2018), and have made it possible forautomatically-generated AMRs to benefit down-stream tasks, such as question answering (Mitraand Baral, 2015), summarization (Takase et al.,2016), and event detection (Li et al., 2015a). How-ever, to our knowledge, no existing work has ex-ploited AMR for enhancing NMT.

We fill in this gap, taking an attention-basedsequence-to-sequence system as our baseline,which is similar to Bahdanau et al. (2015). Toleverage knowledge within an AMR graph, weadopt a graph recurrent network (GRN) (Songet al., 2018; Zhang et al., 2018) as the AMR en-coder. In particular, a full AMR graph is consid-ered as a single state, with nodes in the graph be-ing its substates. State transitions are performedon the graph recurrently, allowing substates to ex-change information through edges. At each re-current step, each node advances its current stateby receiving information from the current states ofits adjacent nodes. Thus, with increasing numbersof recurrent steps, each word receives informationfrom a larger context. Figure 3 shows the recur-rent transition, where each node works simultane-ously. Compared with other methods for encod-ing AMRs (Konstas et al., 2017), GRN keeps theoriginal graph structure, and thus no informationis lost (Song et al., 2018). For the decoding stage,two separate attention mechanisms are adopted inthe AMR encoder and sequential encoder, respec-tively.

Experiments on WMT16 English-German data(4.17M) show that adopting AMR significantlyimproves a strong attention-based sequence-to-sequence baseline (25.5 vs 23.7 BLEU). Whentrained with small-scale (226K) data, the improve-ment increases (19.2 vs 16.0 BLEU), which showsthat the structural information from AMR can alle-viate data sparsity when training data are not suffi-cient. To our knowledge, we are the first to inves-tigate AMR for NMT.

Our code and parallel data (training/dev/test)with automatically parsed AMRs are availableat https://github.com/freesunshine0316/semantic-nmt.

2 Related work

Most previous work on exploring semantics forstatistical machine translation (SMT) studies theusefulness of predicate-argument structure fromsemantic role labeling (Wong and Mooney, 2006;Wu and Fung, 2009; Liu and Gildea, 2010; Bakeret al., 2012). Jones et al. (2012) first convertProlog expressions into graphical meaning rep-resentations, leveraging synchronous hyperedgereplacement grammar to parse the input graphswhile generating the outputs. Their graphicalmeaning representation is different from AMR un-der a strict definition, and their experimental dataare limited to 880 sentences. We are the first toinvestigate AMR on a large-scale machine trans-lation task.

Recently, Marcheggiani et al. (2018) investigatesemantic role labeling (SRL) on neural machinetranslation (NMT). The predicate-argument struc-tures are encoded via graph convolutional network(GCN) layers (Kipf and Welling, 2017), whichare laid on top of regular BiRNN or CNN layers.Our work is in line with exploring semantic in-formation, but different in exploiting AMR ratherthan SRL for NMT. In addition, we leverage agraph recurrent network (GRN) (Song et al., 2018;Zhang et al., 2018) for modeling AMRs ratherthan GCN, which is formally consistent with theRNN sentence encoder. Since there is no one-to-one correspondence between AMR nodes andsource words, we adopt a doubly-attentive LSTMdecoder, which is another major difference fromMarcheggiani et al. (2018).

GRNs have recently been used to model graphstructures in NLP tasks. In particular, Zhang et al.(2018) use a GRN model to represent raw sen-tences by building a graph structure of neigh-boring words and a sentence-level node, showingthat the encoder outperforms BiLSTMs and Trans-former (Vaswani et al., 2017) on classification andsequence labeling tasks; Song et al. (2018) builda GRN for encoding AMR graphs for text gener-ation, showing that the representation is superiorcompared to BiLSTM on serialized AMR. We ex-tend Song et al. (2018) by investigating the useful-ness of AMR for neural machine translation. To

Page 3: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

our knowledge, we are the first to use GRN formachine translation.

In addition to GRNs and GCNs, there have beenother graph neural networks, such as graph gatedneural network (GGNN) (Li et al., 2015b; Becket al., 2018). Since our main concern is to em-pirically investigate the effectiveness of AMR forNMT, we leave it to future work to compare GCN,GGNN, and GRN for our task.

3 Baseline: attention-based BiLSTM

We take the attention-based sequence-to-sequencemodel of Bahdanau et al. (2015) as the baseline,but use LSTM cells (Hochreiter and Schmidhuber,1997) instead of GRU cells (Cho et al., 2014).

3.1 BiLSTM encoderThe encoder is a bi-directional LSTM on thesource side. Given a sentence, two sequences ofstates [

←−h 1,←−h 2, . . . ,

←−hN ] and [

−→h 1,−→h 2, . . .

−→hN ]

are generated for representing the input word se-quence x1, x2, . . . , xN in the right-to-left and left-to-right directions, respectively, where for eachword xi,

←−h i = LSTM(

←−h i+1, exi)

−→h i = LSTM(

−→h i−1, exi)

exi is the embedding of word xi.

3.2 Attention-based decoderThe decoder yields a word sequence in the tar-get language y1, y2, . . . , yM by calculating a se-quence of hidden states s1, s2 . . . , sM recurrently.We use an attention-based LSTM decoder (Bah-danau et al., 2015), where the attention memory(H) is the concatenation of the attention vectorsamong all source words. Each attention vector hi

is the concatenation of the encoder states of an in-put token in both directions (

←−h i and

−→h i):

hi = [←−h i;−→h i]

H = [h1;h2; . . . ;hN ]

N is the number of source words.While generating the m-th word, the decoder

considers four factors: (1) the attention memoryH; (2) the previous hidden state of the LSTMmodel sm−1; (3) the embedding of the currentinput (previously generated word) eym ; and (4)the previous context vector ζm−1 from attentionmemory H . When m = 1, we initialize ζ0 as a

zero vector, set ey1 to the embedding of sentencestart token “<s>”, and calculate s0 from the laststep of the encoder states via a dense layer:

s0 =W 1[←−h 0;−→hN ] + b1

whereW 1 and b1 are model parameters.For each decoding stepm, the decoder feeds the

concatenation of the embedding of the current in-put eym and the previous context vector ζm−1 intothe LSTM model to update its hidden state:

sm = LSTM(sm−1, [eym ; ζm−1])

Then the attention probability αm,i on the atten-tion vector hi ∈ H for the current decode step iscalculated as:

εm,i = vᵀ2 tanh(W hhi +W ssm + b2)

αm,i =exp(εm,i)∑Nj=1 exp(εm,j)

W h, W s, v2 and b2 are model parameters. Thenew context vector ζm is calculated via

ζm =

N∑i=1

αm,ihi

The output probability distribution over the targetvocabulary at the current state is calculated by:

P vocab = softmax(V 3[sm, ζm] + b3), (1)

where V 3 and b3 are learnable parameters.

4 Incorporating AMR

Figure 2 shows the overall architecture of ourmodel, which adopts a BiLSTM (bottom left)and our graph recurrent network (GRN)2 (bottomright) for encoding the source sentence and AMR,respectively. An attention-based LSTM decoder isused to generate the output sequence in the targetlanguage, with attention models over both the se-quential encoder and the graph encoder. The atten-tion memory for the graph encoder is from the laststep of the graph state transition process, which isshown in Figure 3.

2We show the advantage of our graph encoder by compar-ing with another popular way for encoding AMRs in Section6.3.

Page 4: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

John wants to go

... ...

möchte gehen

möchte

Figure 2: Overall architecture of our model.

4.1 Encoding AMR with GRN

Figure 3 shows the overall structure of our graphrecurrent network for encoding AMR graphs,which follows Song et al. (2018). Formally, givenan AMR graph G = (V ,E), we use a hiddenstate vector aj to represent each node vj ∈ V .The state of the graph can thus be represented as:

g = {aj}|vj∈V

In order to capture non-local interaction betweennodes, information exchange between nodes isexecuted through a sequence of state transitions,leading to a sequence of states g0, g1, . . . , gT ,where gt = {ajt}|vj∈V , and T is the number ofstate transitions, which is a hyperparameter. Theinitial state g0 consists of a set of initial node statesaj0 = a0, where a0 is a vector of all zeros.

A recurrent neural network is used to model thestate transition process. In particular, the transitionfrom gt−1 to gt consists of a hidden state transi-tion for each node (such as from ajt−1 to ajt ), asshown in Figure 3. At each state transition stept, our model conducts direct communication be-tween a node and all nodes that are directly con-nected to the node. To avoid gradient diminishingor bursting, LSTM (Hochreiter and Schmidhuber,1997) is adopted, where a cell cjt is taken to recordmemory for ajt . We use an input gate ijt , an outputgate ojt and a forget gate f j

t to control informationflow from the inputs and to the output ajt .

The inputs include representations of edges thatare connected to vj , where vj can be either thesource or the target of the edge. We define eachedge as a triple (i, j, l), where i and j are indicesof the source and target nodes, respectively, and lis the edge label. xl

i,j is the representation of edge(i, j, l), detailed in Section 4.1.1. The inputs forvj are grouped into incoming and outgoing edges,

Time

Figure 3: Architecture of the graph recurrent network.

before being summed up:

φj =∑

(i,j,l)∈Ein(j)

xli,j

φj =∑

(j,k,l)∈Eout(j)

xlj,k

where Ein(j) and Eout(j) are the sets of incom-ing and outgoing edges of vj , respectively.

In addition to edge inputs, our model also takesthe hidden states of the incoming and outgoingneighbors of each node during a state transition.Taking vj as an example, the states of its incom-ing and outgoing neighbors are summed up beforebeing passed to the cell and gate nodes:

ψj =∑

(i,j,l)∈Ein(j)

ait−1

ψj =∑

(j,k,l)∈Eout(j)

akt−1,

Based on the above definitions of φj , φj , ψj andψj , the state transition from gt−1 to gt, as repre-sented by ajt , can be defined as:

ijt = σ(W iφj + W iφj +U iψj + U iψj + bi)

ojt = σ(W oφj + W oφj +Uoψj + Uoψj + bo)

f jt = σ(W fφj + W f φj +Ufψj + Uf ψj + bf )

ujt = σ(W uφj + W uφj +Uuψj + Uuψj + bu)

cjt = fjt � c

jt−1 + i

jt � u

jt

ajt = o

jt � tanh(cjt ),

where ijt , ojt and f j

t are the input, output and for-get gates mentioned earlier. W x, W x, Ux, Ux,bx, where x ∈ {i, o, f, u}, are model parameters.

With this state transition mechanism, informa-tion of each node is propagated to all its neighbor-ing nodes after each step. So after several transi-

Page 5: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

tion steps, each node state contains the informa-tion of a large context, including its ancestors, de-scendants, and siblings. For the worst case wherethe input graph is a chain of nodes, the maximumnumber of steps necessary for information fromone arbitrary node to reach another is equal to thesize of the graph. We experiment with differentnumbers of transition steps to study the effective-ness of global encoding.

4.1.1 Input RepresentationThe edges of an AMR graph contain labels, whichrepresent relations between the nodes they con-nect, and are thus important for modeling thegraphs. The representation for each edge (i, j, l)is defined as:

xli,j =W 4

([el; evi ]

)+ b4,

where el and ei are the embeddings of edge labell and source node vi, and W 4 and b4 are modelparameters.

4.2 Incorporating AMR information with adoubly-attentive decoder

There is no one-to-one correspondence betweenAMR nodes and source words. To incorporate ad-ditional knowledge from an AMR graph, an ex-ternal attention model is adopted over the base-line model. In particular, the attention mem-ory from the AMR graph is the last graph stategT = {ajT }|vj∈V . In addition, the contextual vec-tor based on the graph state is calculated as:

εm,i = vᵀ2 tanh(W aa

iT + W ssm + b2)

αm,i =exp(εm,i)∑Nj=1 exp(εm,j)

W a, W s, v2 and b2 are model parameters.The new context vector ζm is calculated via∑N

i=1 αm,iaiT . Finally, ζm is incorporated into

the calculation of the output probability distribu-tion over the target vocabulary (previously definedin Equation 1):

P vocab = softmax(V 3[sm, ζm, ζm] + b3) (2)

5 Training

Given a set of training instances {(X(1),Y (1)),(X(2),Y (2)), . . . }, we train our models using thecross-entropy loss over each gold-standard target

Dataset #Sent. #Tok. (EN) #Tok. (DE)NC-v11 226K 6.4M 7.3MFull 4.17M 109M 118MNews2013 3000 84.7K 95.6KNews2016 2999 88.1K 98.8K

Table 1: Statistics of the dataset. Numbers of tokensare after BPE processing.

Dataset EN-ori EN AMR DENC-v11 79.8K 8.4K 36.6K 8.3KFull 874K 19.3K 403K 19.1K

Table 2: Sizes of vocabularies. EN-ori represents orig-inal English sentences without BPE.

sequence Y (j) = y(j)1 , y

(j)2 , . . . , y

(j)M :

l = −M∑

m=1

log p(y(j)m |y(j)m−1, . . . , y

(j)1 ,X(j);θ)

X(j) represents the inputs for the jth instance,which is a source sentence for our baseline, ora source sentence paired with an automaticallyparsed AMR graph for our model. θ representsthe model parameters.

6 Experiments

We empirically investigate the effectiveness ofAMR for English-to-German translation.

6.1 SetupData We use the WMT163 English-to-

German dataset, which contains around 4.5million sentence pairs for training. In addition, weuse a subset of the full dataset (News Commentaryv11 (NC-v11), containing around 243 thousandsentence pairs) for development and additionalexperiments. For all experiments, we use new-stest2013 and newstest2016 as the developmentand test sets, respectively.

To preprocess the data, the tokenizer fromMoses4 is used to tokenize both the English andGerman sides. The training sentence pairs whereeither side is longer than 50 words are filtered outafter tokenization. To deal with rare and com-pound words, byte-pair encoding (BPE)5 (Sen-nrich et al., 2016) is applied to both sides. Inparticular, 8000 and 16000 BPE merges are used

3http://www.statmt.org/wmt16/translation-task.html4http://www.statmt.org/moses/5https://github.com/rsennrich/subword-nmt

Page 6: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

on the News Commentary v11 subset and thefull training set, respectively. On the other hand,JAMR6 (Flanigan et al., 2016) is adopted to parsethe English sentences into AMRs before BPE isapplied. The statistics of the training data and vo-cabularies after preprocessing are shown in Table1 and 2, respectively. For the experiments with thefull training set, we used the top 40K of the AMRvocabulary, which covers more than 99.6% of thetraining set.

For our dependency-based and SRL-basedbaselines (which will be introduced in Baselinesystems), we choose Stanford CoreNLP (Man-ning et al., 2014) and IBM SIRE to generate de-pendency trees and semantic roles, respectively.Since both dependency trees and semantic rolesare based on the original English sentences with-out BPE, we used the top 100K frequent Englishwords, which cover roughly 99.0% of the trainingset.

Hyperparameters We use the Adam opti-mizer (Kingma and Ba, 2014) with a learning rateof 0.0005. The batch size is set to 128. Betweenlayers, we apply dropout with a probability of0.2. The best model is picked based on the cross-entropy loss on the development set. For modelhyperparameters, we set the graph state transitionnumber to 10 according to development experi-ments. Each node takes information from at most6 neighbors. BLEU (Papineni et al., 2002), TER(Snover et al., 2006) and Meteor (Denkowski andLavie, 2014) are used as the metrics on cased andtokenized results.

For experiments with the NC-v11 subset, bothword embedding and hidden vector sizes are setto 500, and the models are trained for at most 30epochs. For experiments with full training set, theword embedding and hidden state sizes are set to800, and our models are trained for at most 10epochs. For all systems, the word embeddings arerandomly initialized and updated during training.

Baseline systems We compare our modelwith the following systems. Seq2seq repre-sents our attention-based LSTM baseline (§3), andDual2seq is our model, which takes both a se-quential and a graph encoder and adopts a doubly-attentive decoder (§4). To show the merit of AMR,we further contrast our model with the followingbaselines, all of which adopt the same doubly-

6https://github.com/jflanigan/jamr

12.5

13

13.5

14

14.5

15

15.5

16

16.5

1 5 10 12

BLE

U

Number of state transitions

Dual2seq Dual2seq (self) Seq2seq

Figure 4: DEV BLEU scores against transition steps forthe graph encoders. The state transition is not applica-ble to Seq2seq, so we draw a dashed line to representits performance.

attentive framework with a BiLSTM for encod-ing BPE-segmented source sentences: Dual2seq-LinAMR uses another BiLSTM for encoding lin-earized AMRs. Dual2seq-Dep and Dual2seq-SRLadopt our graph recurrent network to encode origi-nal source sentences with dependency and seman-tic role annotations, respectively. The three base-lines are useful for contrasting different methodsof encoding AMRs and for comparing AMRs withother popular structural information for NMT.

We also compare with Transformer (Vaswaniet al., 2017) and OpenNMT (Klein et al., 2017),trained on the same dataset and with the same setof hyperparameters as our systems. In particu-lar, we compare with Transformer-tf, one popu-lar implementation7 of Transformer based on Ten-sorFlow, and we choose OpenNMT-tf, an officialrelease8 of OpenNMT implemented with Tensor-Flow. For a fair comparison, OpenNMT-tf has1 layer for both the encoder and the decoder,and Transformer-tf has the default configuration(N=6), but with parameters being shared amongdifferent blocks.

6.2 Development experiments

Figure 4 shows the system performances as a func-tion of the number of graph state transitions onthe development set. Dual2seq (self) representsour dual-attentive model, but its graph encoder en-codes the source sentence, which is treated as achain graph, instead of an AMR graph. Comparedwith Dual2seq, Dual2seq (self) has the same num-ber of parameters, but without semantic informa-tion from AMR. Due to hardware limitations, we

7https://github.com/Kyubyong/transformer8https://github.com/OpenNMT/OpenNMT-tf

Page 7: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

SystemNC-V11 FULL

BLEU TER↓ Meteor BLEU TER↓ MeteorOpenNMT-tf 15.1 0.6902 0.3040 24.3 0.5567 0.4225Transformer-tf 17.1 0.6647 0.3578 25.1 0.5537 0.4344Seq2seq 16.0 0.6695 0.3379 23.7 0.5590 0.4258Dual2seq-LinAMR 17.3 0.6530 0.3612 24.0 0.5643 0.4246Duel2seq-SRL 17.2 0.6591 0.3644 23.8 0.5626 0.4223Dual2seq-Dep 17.8 0.6516 0.3673 25.0 0.5538 0.4328Dual2seq 19.2* 0.6305 0.3840 25.5* 0.5480 0.4376

Table 3: TEST performance. NC-v11 represents training only with the NC-v11 data, while Full means using thefull training data. * represents significant (Koehn, 2004) result (p < 0.01) over Seq2seq. ↓ indicates the lower thebetter.

do not perform an exhaustive search by evaluat-ing every possible state transition number, but onlytransition numbers of 1, 5, 10 and 12.

Our Dual2seq shows consistent performanceimprovement by increasing the transition numberboth from 1 to 5 (roughly +1.3 BLEU points) andfrom 5 to 10 (roughly 0.2 BLEU points). Theformer shows greater improvement than the latter,showing that the performance starts to convergeafter 5 transition steps. Further increasing transi-tion steps from 10 to 12 gives a slight performancedrop. We set the number of state transition stepsto 10 for all experiments according to these obser-vations.

On the other hand, Dual2seq (self) shows onlysmall improvements by increasing the state tran-sition number, and it does not perform better thanSeq2seq. Both results show that the performancegains of Dual2seq are not due to an increasednumber of parameters.

6.3 Main results

Table 3 shows the TEST BLEU, TER and Me-teor scores of all systems trained on the small-scale News Commentary v11 subset or the large-scale full set. Dual2seq is consistently better thanthe other systems under all three metrics, show-ing the effectiveness of the semantic informationprovided by AMR. Especially, Dual2seq is betterthan both OpenNMT-tf and Transformer-tf . Therecurrent graph state transition of Dual2seq is sim-ilar to Transformer in that it iteratively incorpo-rates global information. The improvement ofDual2seq over Transformer-tf undoubtedly comesfrom the use of AMRs, which provide comple-mentary information to the textual inputs of thesource language.

In terms of BLEU score, Dual2seq is signifi-

cantly better than Seq2seq in both settings, whichshows the effectiveness of incorporating AMRinformation. In particular, the improvement ismuch larger under the small-scale setting (+3.2BLEU) than that under the large-scale setting(+1.7 BLEU). This is an evidence that structuraland coarse-grained semantic information encodedin AMRs can be more helpful when training dataare limited.

When trained on the NC-v11 subset, the gapbetween Seq2seq and Dual2seq under Meteor(around 5 points) is greater than that under BLEU(around 3 points). Since Meteor gives partialcredit to outputs that are synonyms to the refer-ence or share identical stems, one possible ex-planation is that the structural information withinAMRs helps to better translate the concepts fromthe source language, which may be synonyms orparonyms of reference words.

As shown in the second group of Table 3,we further compare our model with other meth-ods of leveraging syntactic or semantic informa-tion. Dual2seq-LinAMR shows much worse per-formance than our model and only slightly out-performs the Seq2seq baseline. Both results showthat simply taking advantage of the AMR conceptswithout their relations does not help very much.One reason may be that AMR concepts, such asJohn and Mary, also appear in the textual input,and thus are also encoded by the other (sequen-tial) encoder.9 The gap between Dual2seq andDual2seq-LinAMR comes from modeling the rela-tions between concepts, which can be helpful fordeciding target word order by enhancing the rela-tions in source sentences. We conclude that prop-erly encoding AMRs is necessary to make them

9AMRs can contain multi-word concepts, such as NewYork City, but they are in the textual input.

Page 8: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

useful.Encoding dependency trees instead of AMRs,

Dual2seq-Dep shows a larger performance gapwith our model (17.8 vs 19.2) on small-scale train-ing data than on large-scale training data (25.0vs 25.5). It is likely because AMRs are moreuseful on alleviating data sparsity than depen-dency trees, since words are lemmatized into uni-fied concepts when parsing sentences into AMRs.For modeling long-range dependencies, AMRshave one crucial advantage over dependency treesby modeling concept-concept relations more di-rectly. It is because AMRs drop function words,thus the distances between concepts are generallycloser in AMRs than in dependency trees. Finally,Dual2seq-SRL is less effective than our model, be-cause the annotations labeled by SRL are a subsetof AMRs.

We outperform Marcheggiani et al. (2018) onthe same datasets, although our systems vary in anumber of respects. When trained on the NC-v11data, they show BLEU scores of 14.9 only withtheir BiLSTM baseline, 16.1 using additional de-pendency information, 15.6 using additional se-mantic roles and 15.8 taking both as additionalknowledge. Using Full as the training data, thescores become 23.3, 23.9, 24.5 and 24.9, respec-tively. In addition to the different semantic rep-resentation being used (AMR vs SRL), Marcheg-giani et al. (2018) laid graph convolutional net-work (GCN) (Kipf and Welling, 2017) layers ontop of a bidirectional LSTM (BiLSTM) layer, andthen concatenated layer outputs as the attentionmemory. GCN layers encode the semantic role in-formation, while BiLSTM layers encode the inputsentence in the source language, and the concate-nated hidden states of both layers contain infor-mation from both semantic role and source sen-tence. For incorporating AMR, since there is noone-to-one word-to-node correspondence betweena sentence and the corresponding AMR graph,we adopt separate attention models. Our BLEUscores are higher than theirs, but we cannot con-clude that the advantage primarily comes fromAMR.

6.4 Analysis

Influence of AMR parsing accuracy To an-alyze the influence of AMR parsing on our modelperformance, we further evaluate on a test setwhere the gold AMRs for the English side are

AMR Anno. BLEUAutomatic 16.8Gold 17.5*

Table 4: BLEU scores of Dual2seq on the little princedata, when gold or automatic AMRs are available.

<=10 11-20 21-30 31+20.0

21.0

22.0

23.0

24.0

25.0

26.0

Seq2seq

Dual2seq

Dual2seq-Dep

Figure 5: Test BLEU score of various sentence lengths

available. In particular, we choose the LittlePrince corpus, which contains 1562 sentenceswith gold AMR annotations.10 Since there areno parallel German sentences, we take a German-version Little Prince novel, and then perform man-ual sentence alignment. Taking the whole LittlePrince corpus as the test set, we measure the in-fluence of AMR parsing accuracy by evaluatingon the test set when gold or automatically-parsedAMRs are available. The automatic AMRs aregenerated by parsing the English sentences withJAMR.

Table 4 shows the BLEU scores of ourDual2seq model taking gold or automatic AMRsas inputs. Not listed in Table 4, Seq2seq achievesa BLEU score of 15.6, which is 1.2 BLEU pointslower than using automatic AMR information.The improvement from automatic AMR to goldAMR (+0.7 BLEU) is significant, which showsthat the translation quality of our model can befurther improved with an increase of AMR pars-ing accuracy. However, the BLEU score with goldAMR does not indicate the potentially best perfor-mance that our model can achieve. The primaryreason is that even though the test set is coupledwith gold AMRs, the training set is not. Trainedwith automatic AMRs, our model can learn to se-lectively trust the AMR structure. An additionalreason is the domain difference: the Little Princedata are in the literary domain while our trainingdata are in the news domain. There can be a furtherperformance gain if the accuracy of the automaticAMRs on the training set is improved.

10https://amr.isi.edu/download.html

Page 9: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

AMR: (s2 / say-01 :ARG0 (p3 / person :ARG1-of (h / have-rel-role-91 :ARG0 (p / person :ARG1-of (m2 / meet-03 :ARG0 (t / they) :ARG2 15) :mod (m / mutual)) :ARG2 (f / friend)) :name (n2 / name :op1 ”Carla” :op2”Hairston”)) :ARG1 (a / and :op1 (p2 / person :name (n / name :op1 ”Lamb”))) :ARG2 (s / she) :time 20)Src: Carla Hairston said she was 15 and Lamb was 20 when they met through mutual friends .Ref: Carla Hairston sagte , sie war 15 und Lamm war 20 , als sie sich durch gemeinsame Freunde trafen .Dual2seq: Carla Hairston sagte , sie war 15 und Lamm war 20 , als sie sich durch gegenseitige Freunde trafen .Seq2seq: Carla Hirston sagte , sie sei 15 und Lamb 20 , als sie durch gegenseitige Freunde trafen .AMR: (s / say-01 :ARG0 (m / media :ARG1-of (l / local-02)) :ARG1 (c2 / come-01 :ARG1 (v / vehicle :mod (p/ police)) :manner (c3 / constant) :path (a / across :op1 (r / refugee :mod (n2 / new))) :time (s2 / since :op1 (t3 /then)) :topic (t / thing :name (n / name :op1 (c / Croatian) :op2 (t2 / Tavarnik)))))Src: Since then , according to local media , police vehicles are constantly coming across new refugees in CroatianTavarnik .Ref: Laut lokalen Medien treffen seitdem im kroatischen Tovarnik standig Polizeifahrzeuge mit neuenFluchtlingen ein .Dual2seq: Seither kommen die Polizeifahrzeuge nach den ortlichen Medien standig uber neue Fluchtlinge inKroatische Tavarnik .Seq2seq: Seitdem sind die Polizeiautos nach den lokalen Medien standig neue Fluchtlinge in Kroatien Tavarnik .AMR: (b2 / breed-01 :ARG0 (p2 / person :ARG0-of (h / have-org-role-91 :ARG2 (s3 / scientist))) :ARG1 (w2 /worm) :ARG2 (s2 / system :ARG1-of (c / control-01 :ARG0 (b / burst-01 :ARG1 (w / wave :mod (s / sound))):ARG1-of (p / possible-01)) :ARG1-of (n / nervous-01) :mod (m / modify-01 :ARG1 (g / genetics))))Src: Scientists have bred worms with genetically modified nervous systems that can be controlled by bursts ofsound waves .Ref: Wissenschaftler haben Wurmer mit genetisch veranderten Nervensystemen gezuchtet , die von Ausbruchenvon Schallwellen gesteuert werden konnen .Dual2seq: Die Wissenschaftler haben die Wurmer mit genetisch veranderten Nervensystemen gezuchtet , diedurch Verbrennungen von Schallwellen kontrolliert werden konnen .Seq2seq: Wissenschaftler haben sich mit genetisch modifiziertem Nervensystem gezuchtet , die durch Verbren-nungen von Klangwellen gesteuert werden konnen .

Figure 6: Sample system outputs

Performance based on sentence length Wehypothesize that AMRs should be more bene-ficial for longer sentences: those are likely tocontain long-distance dependencies (such as dis-course information and predicate-argument struc-tures) which may not be adequately captured bylinear chain RNNs but are directly encoded inAMRs. To test this, we partition the test datainto four buckets by length and calculate BLEUfor each of them. Figure 5 shows the perfor-mances of our model along with Dual2seq-Depand Seq2seq. Our model outperforms the Seq2seqbaseline rather uniformly across all buckets, ex-cept for the first one where they are roughly equal.This may be surprising. On the one hand, Seq2seqfails to capture some dependencies for medium-length instances; on the other hand AMR parsesare more noisy for longer sentences, which pre-vents us from obtaining extra improvements withAMRs.

Dependency trees have been proved useful incapturing long-range dependencies. Figure 5shows that AMRs are comparatively better thandependency trees, especially on medium-length

(21-30) sentences. The reason may be that theAMRs of medium-length sentences are muchmore accurate than longer sentences, thus are bet-ter at capturing the relations between concepts. Onthe other hand, even though dependency trees aremore accurate than AMRs, they still fail to repre-sent relations for long sentences. It is likely be-cause relations for longer sentences are more dif-ficult to detect. Another possible reason is thatdependency trees do not incorporate coreferences,which AMRs consider.

Human evaluation We further study thetranslation quality of predicate-argument struc-tures by conducting a human evaluation on 100instances from the testset. In the evaluation, trans-lations of both Dual2seq and Seq2seq, togetherwith the source English sentence, the German ref-erence, and an AMR are provided to a German-speaking annotator to decide which translationbetter captures the predicate-argument structuresin the source sentence. To avoid annotation bias,translation results of both models are swapped forsome instances, and the German annotator does

Page 10: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

not know which model each translation belongsto. The annotator either selects a “winner” ormakes a “tie” decision, meaning that both resultsare equally good.

Out of the 100 instances, Dual2seq wins on46, Seq2seq wins on 23, and there is a tie onthe remaining 31. Dual2seq wins on almost halfof the instances, about twice as often as Seq2seqwins, indicating that AMRs help in translating thepredicate-argument structures on the source side.

Case study The outputs of the baseline sys-tem (Seq2seq) and our final system (Dual2seq)are shown in Figure 6. In the first sentence, theAMR-based Dual2seq system correctly producesthe reflexive pronoun sich as an argument of theverb trafen (meet), despite the distance betweenthe words in the system output, and despite the factthat the equivalent English words each other donot appear in the system output. This is facilitatedby the argument structure in the AMR analysis.

In the second sentence, the AMR-basedDual2seq system produces an overly literal trans-lation for the English phrasal verb come across.The Seq2seq translation, however, incorrectlystates that the police vehicles are refugees. Thedifficulty for the Seq2seq probably derives in partfrom the fact that are and coming are separated bythe word constantly in the input, while the mainpredicate is clear in the AMR representation.

In the third sentence, the Dual2seq system cor-rectly translates the object of breed as worms,while the Seq2seq translation incorrectly statesthat the scientists breed themselves. Here the diffi-culty is likely the distance between the object andthe verb in the German output, which causes theSeq2seq system to lose track of the correct inputposition to translate.

7 Conclusion

We showed that AMRs can improve neural ma-chine translation. In particular, the structural se-mantic information from AMRs can be comple-mentary to the source textual input by introducinga higher level of information abstraction. A graphrecurrent network (GRN) is leveraged to encodeAMR graphs without breaking the original graphstructure, and a sequential LSTM is used to encodethe source input. The decoder is a doubly-attentiveLSTM, taking the encoding results of both thegraph encoder and the sequential encoder as atten-tion memories. Experiments on a standard bench-

mark showed that AMRs are helpful regardless ofthe sentence length, and are more effective thanother more popular choices, such as dependencytrees and semantic roles.

Acknowledgments

We would like to thank the action editor andthe anonymous reviewers for their insightful com-ments. We also thank Kai Song from Alibaba forsuggestions on large-scale training, Parker Rileyfor comments on the draft, and Rochester’s CIRCfor computational resources.

References

Roee Aharoni and Yoav Goldberg. 2017. To-wards string-to-tree neural machine translation.In Proceedings of the 55th Annual Meeting ofthe Association for Computational Linguistics(ACL-17), pages 132–140.

Yoav Artzi, Kenton Lee, and Luke Zettlemoyer.2015. Broad-coverage CCG semantic parsingwith AMR. In Conference on Empirical Meth-ods in Natural Language Processing (EMNLP-15), pages 1699–1710.

Dzmitry Bahdanau, Kyunghyun Cho, and YoshuaBengio. 2015. Neural machine translation byjointly learning to align and translate. In In-ternational Conference on Learning Represen-tations (ICLR).

Kathryn Baker, Michael Bloodgood, Bonnie JDorr, Chris Callison-Burch, Nathaniel W Fi-lardo, Christine Piatko, Lori Levin, and ScottMiller. 2012. Modality and negation in SIMTuse of modality and negation in semantically-informed syntactic MT. Computational Lin-guistics, 38(2):411–438.

Laura Banarescu, Claire Bonial, Shu Cai,Madalina Georgescu, Kira Griffitt, Ulf Herm-jakob, Kevin Knight, Philipp Koehn, MarthaPalmer, and Nathan Schneider. 2013. Abstractmeaning representation for sembanking. InProceedings of the 7th Linguistic AnnotationWorkshop and Interoperability with Discourse,pages 178–186.

Joost Bastings, Ivan Titov, Wilker Aziz, DiegoMarcheggiani, and Khalil Simaan. 2017. Graphconvolutional encoders for syntax-aware neural

Page 11: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

machine translation. In Conference on Empir-ical Methods in Natural Language Processing(EMNLP-17), pages 1957–1967.

Daniel Beck, Gholamreza Haffari, and TrevorCohn. 2018. Graph-to-sequence learning usinggated graph neural networks. In Proceedingsof the 56th Annual Meeting of the Associationfor Computational Linguistics (ACL-18), pages273–283.

Jan Buys and Phil Blunsom. 2017. Robust incre-mental neural semantic graph parsing. In Pro-ceedings of the 55th Annual Meeting of the As-sociation for Computational Linguistics (ACL-17), pages 1215–1226.

Huadong Chen, Shujian Huang, David Chiang,and Jiajun Chen. 2017. Improved neural ma-chine translation with a syntax-aware encoderand decoder. In Proceedings of the 55th AnnualMeeting of the Association for ComputationalLinguistics (ACL-17), pages 1936–1945.

Kehai Chen, Rui Wang, Masao Utiyama, Ei-ichiro Sumita, and Tiejun Zhao. 2018. Syntax-directed attention for neural machine transla-tion. In Proceedings of the National Conferenceon Artificial Intelligence (AAAI-18).

Kyunghyun Cho, Bart van Merrienboer, CaglarGulcehre, Dzmitry Bahdanau, Fethi Bougares,Holger Schwenk, and Yoshua Bengio. 2014.Learning phrase representations using RNNencoder–decoder for statistical machine trans-lation. In Conference on Empirical Methodsin Natural Language Processing (EMNLP-14),pages 1724–1734.

Michael Denkowski and Alon Lavie. 2014. Me-teor universal: Language specific translationevaluation for any target language. In Proceed-ings of the Ninth Workshop on Statistical Ma-chine Translation, pages 376–380.

Jeffrey Flanigan, Chris Dyer, Noah A. Smith, andJaime Carbonell. 2016. CMU at SemEval-2016Task 8: Graph-based AMR parsing with infi-nite ramp loss. In Proceedings of the 10th In-ternational Workshop on Semantic Evaluation(SemEval-2016), pages 1202–1206.

Jeffrey Flanigan, Sam Thomson, Jaime Carbonell,Chris Dyer, and Noah A. Smith. 2014. A

discriminative graph-based parser for the ab-stract meaning representation. In Proceedingsof the 52nd Annual Meeting of the Associationfor Computational Linguistics (ACL-14), pages1426–1436.

Jonas Groschwitz, Matthias Lindemann, MeaghanFowlie, Mark Johnson, and Alexander Koller.2018. AMR dependency parsing with a typedsemantic algebra. In Proceedings of the 56thAnnual Meeting of the Association for Com-putational Linguistics (ACL-18), pages 1831–1841.

Zhijiang Guo and Wei Lu. 2018. Better transition-based AMR parsing with a refined search space.In Proceedings of the 56th Annual Meeting ofthe Association for Computational Linguistics(ACL-18), pages 1712–1722.

Sepp Hochreiter and Jurgen Schmidhuber. 1997.Long short-term memory. Neural computation,9(8):1735–1780.

Pierre Isabelle, Colin Cherry, and George Foster.2017. A challenge set approach to evaluatingmachine translation. In Conference on Empir-ical Methods in Natural Language Processing(EMNLP-17), pages 2486–2496.

Bevan Jones, Jacob Andreas, Daniel Bauer,Karl Moritz Hermann, and Kevin Knight. 2012.Semantics-based machine translation with hy-peredge replacement grammars. In Proceed-ings of the International Conference on Compu-tational Linguistics (COLING-12), pages 1359–1376.

Diederik Kingma and Jimmy Ba. 2014. Adam:A method for stochastic optimization. arXivpreprint arXiv:1412.6980.

Thomas N. Kipf and Max Welling. 2017. Semi-supervised classification with graph convolu-tional networks. In International Conference onLearning Representations (ICLR).

Guillaume Klein, Yoon Kim, Yuntian Deng, JeanSenellart, and Alexander M. Rush. 2017. Open-NMT: Open-Source Toolkit for Neural MachineTranslation. arXiv preprint arXiv:1701.02810.

Philipp Koehn. 2004. Statistical significance testsfor machine translation evaluation. In Con-ference on Empirical Methods in Natural Lan-

Page 12: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

guage Processing (EMNLP-04), pages 388–395.

Ioannis Konstas, Srinivasan Iyer, Mark Yatskar,Yejin Choi, and Luke Zettlemoyer. 2017. Neu-ral AMR: Sequence-to-sequence models forparsing and generation. In Proceedings ofthe 55th Annual Meeting of the Associationfor Computational Linguistics (ACL-17), pages146–157.

Junhui Li, Deyi Xiong, Zhaopeng Tu, Muhua Zhu,Min Zhang, and Guodong Zhou. 2017. Mod-eling source syntax for neural machine transla-tion. In Proceedings of the 55th Annual Meet-ing of the Association for Computational Lin-guistics (ACL-17), pages 688–697.

Xiang Li, Thien Huu Nguyen, Kai Cao, and RalphGrishman. 2015a. Improving event detectionwith abstract meaning representation. In Pro-ceedings of the First Workshop on ComputingNews Storylines, pages 11–15.

Yujia Li, Daniel Tarlow, Marc Brockschmidt,and Richard Zemel. 2015b. Gated graphsequence neural networks. arXiv preprintarXiv:1511.05493.

Ding Liu and Daniel Gildea. 2010. Semanticrole features for machine translation. In Pro-ceedings of the 23rd International Conferenceon Computational Linguistics (COLING-10),pages 716–724.

Chunchuan Lyu and Ivan Titov. 2018. AMR pars-ing as graph prediction with latent alignment.In Proceedings of the 56th Annual Meeting ofthe Association for Computational Linguistics(ACL-18), pages 397–407.

Christopher D. Manning, Mihai Surdeanu, JohnBauer, Jenny Finkel, Steven J. Bethard, andDavid McClosky. 2014. The Stanford CoreNLPnatural language processing toolkit. In Associ-ation for Computational Linguistics (ACL) Sys-tem Demonstrations, pages 55–60.

Diego Marcheggiani, Joost Bastings, and IvanTitov. 2018. Exploiting semantics in neural ma-chine translation with graph convolutional net-works. In Proceedings of the 2018 Meetingof the North American chapter of the Associ-ation for Computational Linguistics (NAACL-18), pages 486–492.

Arindam Mitra and Chitta Baral. 2015. Loca-tioning a question answering challenge by com-bining statistical methods with inductive rulelearning and reasoning. In Proceedings of theNational Conference on Artificial Intelligence(AAAI-16).

Kishore Papineni, Salim Roukos, Todd Ward, andWei-Jing Zhu. 2002. BLEU: a method for auto-matic evaluation of machine translation. In Pro-ceedings of the 40th Annual Meeting of the As-sociation for Computational Linguistics (ACL-02), pages 311–318.

Xiaochang Peng, Linfeng Song, and DanielGildea. 2015. A synchronous hyperedge re-placement grammar based approach for AMRparsing. In Proceedings of the Nineteenth Con-ference on Computational Natural LanguageLearning, pages 32–41.

Xiaochang Peng, Linfeng Song, Daniel Gildea,and Giorgio Satta. 2018. Sequence-to-sequencemodels for cache transition systems. In Pro-ceedings of the 56th Annual Meeting of the As-sociation for Computational Linguistics (ACL-18), pages 1842–1852.

Michael Pust, Ulf Hermjakob, Kevin Knight,Daniel Marcu, and Jonathan May. 2015. Pars-ing english into abstract meaning representa-tion using syntax-based machine translation.In Conference on Empirical Methods in Natu-ral Language Processing (EMNLP-15), pages1143–1154.

Rico Sennrich, Barry Haddow, and AlexandraBirch. 2016. Neural machine translation ofrare words with subword units. In Proceedingsof the 54th Annual Meeting of the Associationfor Computational Linguistics (ACL-16), pages1715–1725.

Matthew Snover, Bonnie Dorr, Richard Schwartz,Linnea Micciulla, and John Makhoul. 2006. Astudy of translation edit rate with targeted hu-man annotation. In Proceedings of Associationfor Machine Translation in the Americas, pages223–231.

Linfeng Song, Yue Zhang, Zhiguo Wang, andDaniel Gildea. 2018. A graph-to-sequencemodel for AMR-to-text generation. In Proceed-ings of the 56th Annual Meeting of the Associ-

Page 13: Semantic Neural Machine Translation using AMR · 2019-02-21 · Semantic Neural Machine Translation using AMR Linfeng Song 1, Daniel Gildea , Yue Zhang2, Zhiguo Wang3 and Jinsong

ation for Computational Linguistics (ACL-18),pages 1842–1852.

Felix Stahlberg, Eva Hasler, Aurelien Waite, andBill Byrne. 2016. Syntactically guided neuralmachine translation. In Proceedings of the 54thAnnual Meeting of the Association for Compu-tational Linguistics (ACL-16), pages 299–305.

Sho Takase, Jun Suzuki, Naoaki Okazaki, Tsu-tomu Hirao, and Masaaki Nagata. 2016. Neuralheadline generation on abstract meaning repre-sentation. In Conference on Empirical Methodsin Natural Language Processing (EMNLP-16),pages 1054–1059.

Ashish Vaswani, Noam Shazeer, Niki Parmar,Jakob Uszkoreit, Llion Jones, Aidan N Gomez,Łukasz Kaiser, and Illia Polosukhin. 2017. At-tention is all you need. In Advances in Neu-ral Information Processing Systems 30, pages5998–6008.

Chuan Wang and Nianwen Xue. 2017. Getting themost out of AMR parsing. In Conference onEmpirical Methods in Natural Language Pro-cessing (EMNLP-17), pages 1257–1268.

Yuk Wah Wong and Raymond Mooney. 2006.Learning for semantic parsing with statisticalmachine translation. In Proceedings of the2006 Meeting of the North American chapter ofthe Association for Computational Linguistics(NAACL-06), pages 439–446.

Dekai Wu and Pascale Fung. 2009. Semanticroles for SMT: A hybrid two-pass model. InProceedings of the 2009 Meeting of the NorthAmerican chapter of the Association for Com-putational Linguistics (NAACL-09), pages 13–16.

Shuangzhi Wu, Ming Zhou, and DongdongZhang. 2017. Improved neural machine trans-lation with source syntax. In Proceedings of theTwenty-Sixth International Joint Conference onArtificial Intelligence (IJCAI-17), pages 4179–4185.

Yue Zhang, Qi Liu, and Linfeng Song. 2018.Sentence-state LSTM for text representation.In Proceedings of the 56th Annual Meeting ofthe Association for Computational Linguistics(ACL-18), pages 317–327.