115
Computational Psycholinguistics: Day 1 Computational Psycholinguistics, Day 1 Introduction and a classic computational model of sentence production Roger Levy UC San Diego Department of Linguistics LSA Summer Institute 6 July 2007

Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

  • Upload
    others

  • View
    20

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Computational Psycholinguistics, Day 1Introduction and a classic computational

model of sentence production

Roger Levy

UC San DiegoDepartment of Linguistics

LSA Summer Institute6 July 2007

Page 2: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

Computational Psycholinguistics

Psycholinguistics deals with the problem of how humansI acquireI representI comprehendI produce

language.

In this class, we will study these problems from a computationalperspective.

Page 3: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

Class goals

I Overview the literature and major areas in whichcomputational psycholinguistic research is carried out

I Gain experience in reading papers in this fieldI Gain experience in working out the details of a model from

the papersI Critical analysis of models

Page 4: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What is computational modeling? Why do we do it?

I Any phenomenon involving human behavior is so complexthat we cannot hope to formulate a comprehensive theory

I Instead, we devise a model that simplifies thephenomenon to capture some key aspect of it

Page 5: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 6: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 7: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 8: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 9: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 10: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

What might we use a model for?Models can serve any of the following (related) functions:

I Prediction: estimating the behavior/properties of a newstate/datum on the basis of an existing dataset

I Hypothesis testing: as a framework for determiningwhether a given factor has an appreciable influence onsome other variable

I Data simulation: creating artificial data more cheaply andquickly than through empirical data collection

I Summarization: If phenomenon X is complex but relevantto phenomenon Y, it can be most effective to use a simplemodel of X when constructing a model of Y

I Insight: Most generally, a good model can be explored inways that give insight into the phenomenon underconsideration

Page 11: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

Course Contents

I Most of the work we will look at involves probabilisticmodeling of psycholinguistic phenomena

I Probabilistic effects are pervasive in acquisition,representation, comprehension, and production

I . . . but we’ll start off with a paper from another tradition:memory in sentence comprehension

Page 12: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

Feedback from you

Please take a moment to fill out an index card with the followinginfo:

Name (optional)School & Program/DepartmentYear/stage in programComputational Linguistics backgroundPsycholinguistics backgroundProbability/Statistics backgroundOther courses you’re taking at LSA Summer Institute(other side) What do you hope to learn in this class?

Page 13: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Introduction

A bit of history

The formal revolution in linguistic analysis:I Symbolic mathematical models for the description of

natural language sentences Chomsky (1956, 1957)

Its psychological ramifications:I Miller (1956): the human working-memory capacity has

severe limits on how many items can be storedsimultaneously

I Miller and Chomsky (1963): Although human linguisticcompetence requires transformational grammars, humanlinguistic performance might require only simpler (evenfinite-state) grammars

Page 14: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve 1960

A model and an Hypothesis for Language StructurePerhaps the very first paper in computational psycholinguistics

Page 15: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Defining the computational problem

There is an infinite number of sentences in any naturallanguage. How is any given sentence produced?

Page 16: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Basic modeling assumptions

I The grammar by which a sentence is produced consists ofa finite, unordered set of context-free production rules

S→ NP VPNP→ Det N. . .

I An utterance is constructed incrementally—top-down and“left to right”

Page 17: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail

Yngve’s model has the following components:

I A computing register that can store a single symbolI A temporary memory that can store any number of

symbols in a stack (first-in, last-out)I An output

Page 18: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 19: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 20: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 21: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 22: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 23: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 24: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 25: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 26: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

The model in more detail (2)

The model starts with the root symbol S in the computingregister.One execution cycle consists of the following steps:

1. Copy the symbol in the register to the output.2. If the symbol X in the register is a word,

2.1 delete it;2.2 if the temporary memory is not empty, move its top symbol

to the register;2.3 otherwise end.

3. Otherwise: with the register’s symbol on the left-hand side,3.1 choose a production rule X → Yα;3.2 replace the symbol in the register with Y , and put α in the

temporary memory.

Page 27: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 28: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 29: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 30: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 31: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 32: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 33: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 34: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 35: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 36: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 37: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 38: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 39: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 40: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 41: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 42: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 43: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 44: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 45: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 46: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 47: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 48: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 49: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 50: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 51: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 52: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 53: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 54: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Register Temporary MemoryS

NP VPDet N VPthe N VP

N VPdog VPVP

V NPchased NP

NPDet Nthe N

Ncat

*S→ NP VP*NP→ Det N*VP→ VP NP

*Det→ the*N→ dog*N→ cat*V→ chased

S

NP

Det

the

N

dog

VP

V

chased

NP

Det

the

N

cat

Page 55: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents

One type of structure that the model thus far cannot produce isa discontinuous constituent:

S

NP

Pron

it

VP

V

is

Adj

true

Scomp

Comp

that

S

NP

Pron

I

VP

V

lied

Page 56: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 57: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 58: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 59: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 60: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 61: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 62: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 63: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 64: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Discontinuous constituents (2)

To make the model more expressive, allow special rulesaffecting the temporary register:

S→ NP VPNP→ Pron. . . Scomp

When a . . . rule is used, the subsequent symbols go to theright of the first symbol in memory:

Register MemoryS

NP VPPron VP Scomp

S

NP

Pron

VP

Scomp

Page 65: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Model summary

I In modern parlance, Yngve’s model is an incremental,top-down tree generator.

I While deep parts of the tree are being generated,upcoming constituents high in the tree have to be stored inmemory.

I The number of symbols stored in temporary memory atany point in production of a sentence can be said to be thedepth at that point.

Page 66: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Model summary

I In modern parlance, Yngve’s model is an incremental,top-down tree generator.

I While deep parts of the tree are being generated,upcoming constituents high in the tree have to be stored inmemory.

I The number of symbols stored in temporary memory atany point in production of a sentence can be said to be thedepth at that point.

Page 67: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

The Model

Model summary

I In modern parlance, Yngve’s model is an incremental,top-down tree generator.

I While deep parts of the tree are being generated,upcoming constituents high in the tree have to be stored inmemory.

I The number of symbols stored in temporary memory atany point in production of a sentence can be said to be thedepth at that point.

Page 68: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

The hypothesis

I Yngve wanted to identify the temporary memory of hismodel with the human cognitive resource of workingmemory.

I Since working memory is limited, sentences involvingmany items in the temporary memory should be hard toprocess.

Page 69: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

The hypothesis

I Yngve wanted to identify the temporary memory of hismodel with the human cognitive resource of workingmemory.

I Since working memory is limited, sentences involvingmany items in the temporary memory should be hard toprocess.

Page 70: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

The hypothesis (2)

Yngve implicitly proposed two predictions of his theory:

1. The greater the maximum depth of a sentence, the harderit should be to process (produce and/or comprehend).

2. When there are multiple alternative forms that can expressthe same meaning, the alternative with the smallestmaximum depth should tend to be the most felicitous.

3. Natural language grammars should tend to be structuredso as to minimize depth.

Page 71: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

The hypothesis (2)

Yngve implicitly proposed two predictions of his theory:

1. The greater the maximum depth of a sentence, the harderit should be to process (produce and/or comprehend).

2. When there are multiple alternative forms that can expressthe same meaning, the alternative with the smallestmaximum depth should tend to be the most felicitous.

3. Natural language grammars should tend to be structuredso as to minimize depth.

Page 72: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

The hypothesis (2)

Yngve implicitly proposed two predictions of his theory:

1. The greater the maximum depth of a sentence, the harderit should be to process (produce and/or comprehend).

2. When there are multiple alternative forms that can expressthe same meaning, the alternative with the smallestmaximum depth should tend to be the most felicitous.

3. Natural language grammars should tend to be structuredso as to minimize depth.

Page 73: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Center embedding

Using multiple relative clauses in a sentence can makeprocessing difficult:

This is the malt that the rat that the cat that the dogworried killed ate.

It’s not the meaning of the sentence, or the use of relativeclauses, that makes it hard:

This is the malt that was eaten by the rat that waskilled by the cat that was worried by the dog.

What is hard about these sentences?

Page 74: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Center embedding

Using multiple relative clauses in a sentence can makeprocessing difficult:

This is the malt that the rat that the cat that the dogworried killed ate.

It’s not the meaning of the sentence, or the use of relativeclauses, that makes it hard:

This is the malt that was eaten by the rat that waskilled by the cat that was worried by the dog.

What is hard about these sentences?

Page 75: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Center embedding

Using multiple relative clauses in a sentence can makeprocessing difficult:

This is the malt that the rat that the cat that the dogworried killed ate.

It’s not the meaning of the sentence, or the use of relativeclauses, that makes it hard:

This is the malt that was eaten by the rat that waskilled by the cat that was worried by the dog.

What is hard about these sentences?

Page 76: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Center embedding

Using multiple relative clauses in a sentence can makeprocessing difficult:

This is the malt that the rat that the cat that the dogworried killed ate.

It’s not the meaning of the sentence, or the use of relativeclauses, that makes it hard:

This is the malt that was eaten by the rat that waskilled by the cat that was worried by the dog.

What is hard about these sentences?

Page 77: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Center embedding and working memoryS

NP

Dem

this

VP

V

is

NP

Det

the

N

malt

RC

RelPro

that

S

NP

Det

the

N

rat

RC

RelPro

that

S

NP

Det

the

N

cat

RC

RelPro

that

S

NP

Det

the

N

VP

VP

VP

Register Memorythe N VP VP VP

Page 78: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

S

NP

Dem

this

VP

V

is

NP

Det

the

N

malt

RC

RelPro

that

S

VP

V

was

Part

eaten

PP

P

by

NP

Det

the

N

rat

RC

RelPro

that

S

VP

V

was

Part

killed

PP

P

by

NP

Det

the

N

cat

RC

RelPro

that

S

VP

V

was

Part

worried

PP

P

by

NP

Det

the

N

dog

Page 79: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

S

NP

Dem

this

VP

V

is

NP

Det

the

N

malt

RC

RelPro

that

S

VP

V

was

Part

eaten

PP

P

by

NP

Det

the

N

rat

RC

RelPro

that

S

VP

V

was

Part

killed

PP

P

by

NP

Det

the

N

cat

RC

RelPro

that

S

VP

V

was

Part

worried

PP

P

by

NP

Det

the

N

dog

Right-embedding doesn’thave substantial temporarymemory requirements

Page 80: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Only center embedding fills temporary memory

I In center-embedded structures, the model must“remember” which constituents have been started but arenot yet finished

I In right-embedded structures, nothing of the sort needs tobe remembered

I That is, constructions with high maximum depth are hardto process

Page 81: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Only center embedding fills temporary memory

I In center-embedded structures, the model must“remember” which constituents have been started but arenot yet finished

I In right-embedded structures, nothing of the sort needs tobe remembered

I That is, constructions with high maximum depth are hardto process

Page 82: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Only center embedding fills temporary memory

I In center-embedded structures, the model must“remember” which constituents have been started but arenot yet finished

I In right-embedded structures, nothing of the sort needs tobe remembered

I That is, constructions with high maximum depth are hardto process

Page 83: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

A note: temporary memory as “constituentexpectations”

The contents of temporary memory at any point can also bethought of as a set of “constituent expectations”:

S

NP

Dem

this

VP

V

is

NP

Det

the

N

malt

RC

RelPro

that

S

NP

Det

the

N

VP

Page 84: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

A note: temporary memory as “constituentexpectations”

The contents of temporary memory at any point can also bethought of as a set of “constituent expectations”:

S

NP

Dem

this

VP

V

is

NP

Det

the

N

malt

RC

RelPro

that

S

NP

Det

the

N

VP

Register Memorythe N VP

Page 85: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 86: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 87: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 88: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 89: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 90: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

“Heavy Shift”

I Preference to postpone “heavy” constituents until the endof the clause:

Susan gave candy to the boy that visited hisparents in New York last Christmas.? Susan gave the boy that visited his parents inNew York last Christmas candy.

I The medial constituent has an additional memory penaltydue to needing to remember the final constituent

I Hence, the “heavy-medial” variant has the greatestmaximum memory requirement

I That is, when multiple forms are available for the samemeaning, the “shallowest” alternative is preferred

Page 91: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Discontinuous constituents

I Consider the grammar of postmodified superlativeadjectives such as best. . . in the world:

*the best in the world friendthe best friend in the world

I The discontinuous-constituent form used in English is lessdepth-intensive than the continuous-constituent form.

I This type of evidence could be taken as support forYngve’s hypothesis that grammars tend to avoiddepth-intensive constructions.

Page 92: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Discontinuous constituents

I Consider the grammar of postmodified superlativeadjectives such as best. . . in the world:

*the best in the world friendthe best friend in the world

I The discontinuous-constituent form used in English is lessdepth-intensive than the continuous-constituent form.

I This type of evidence could be taken as support forYngve’s hypothesis that grammars tend to avoiddepth-intensive constructions.

Page 93: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Discontinuous constituents

I Consider the grammar of postmodified superlativeadjectives such as best. . . in the world:

*the best in the world friendthe best friend in the world

I The discontinuous-constituent form used in English is lessdepth-intensive than the continuous-constituent form.

I This type of evidence could be taken as support forYngve’s hypothesis that grammars tend to avoiddepth-intensive constructions.

Page 94: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Discontinuous constituents

I Consider the grammar of postmodified superlativeadjectives such as best. . . in the world:

*the best in the world friendthe best friend in the world

I The discontinuous-constituent form used in English is lessdepth-intensive than the continuous-constituent form.

I This type of evidence could be taken as support forYngve’s hypothesis that grammars tend to avoiddepth-intensive constructions.

Page 95: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Discontinuous constituents

I Consider the grammar of postmodified superlativeadjectives such as best. . . in the world:

*the best in the world friendthe best friend in the world

I The discontinuous-constituent form used in English is lessdepth-intensive than the continuous-constituent form.

I This type of evidence could be taken as support forYngve’s hypothesis that grammars tend to avoiddepth-intensive constructions.

Page 96: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Implications of a memory-based theory

I Yngve’s is a memory-based theory where the number ofitems in memory is crucial

I In any such theory, what counts as an item is crucialI This is the problem of the units of representation

Page 97: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Implications of a memory-based theory

I Yngve’s is a memory-based theory where the number ofitems in memory is crucial

I In any such theory, what counts as an item is crucialI This is the problem of the units of representation

Page 98: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Implications of a memory-based theory (2)

Branching structure matters in memory-based theories.Ternary constructions introduce a memory load of 2:

X

O[2]

O[1]

O[0]

Weakly equivalent binary constructions have load of only 1:

X

O[1]

Y

O[1]

O[0]

Page 99: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Implications of a memory-based theory (2)

Branching structure matters in memory-based theories.Ternary constructions introduce a memory load of 2:

X

O[2]

O[1]

O[0]

Weakly equivalent binary constructions have load of only 1:

X

O[1]

Y

O[1]

O[0]

Page 100: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Implications of a memory-based theory (2)

Branching structure matters in memory-based theories.Ternary constructions introduce a memory load of 2:

X

O[2]

O[1]

O[0]

Weakly equivalent binary constructions have load of only 1:

X

O[1]

Y

O[1]

O[0]

Page 101: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Weaknesses for Yngve’s theory

I As an incremental sentence processor, the model“over-commits” in many types of constructions

I Left-branching constructions are pervasive in somelanguages (e.g., Japanese) and not necessarily hard

Page 102: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Overcommittment

Sometimes the model must commit to a structure before anydefinitive evidence for it has appeared:

S

NP

Pron

I

VP

V

eat

S

NP

Pron

I

VP

V

eat

NP

S

NP

Pron

I

VP

V

eat

NP PP

S

NP

Pron

I

VP

V

eat

NP AdvP

Page 103: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Overcommittment

Sometimes the model must commit to a structure before anydefinitive evidence for it has appeared:

S

NP

Pron

I

VP

V

eat

S

NP

Pron

I

VP

V

eat

NP

S

NP

Pron

I

VP

V

eat

NP PP

S

NP

Pron

I

VP

V

eat

NP AdvP

Page 104: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Overcommittment

Sometimes the model must commit to a structure before anydefinitive evidence for it has appeared:

S

NP

Pron

I

VP

V

eat

S

NP

Pron

I

VP

V

eat

NP

S

NP

Pron

I

VP

V

eat

NP PP

S

NP

Pron

I

VP

V

eat

NP AdvP

Page 105: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Overcommittment

Sometimes the model must commit to a structure before anydefinitive evidence for it has appeared:

S

NP

Pron

I

VP

V

eat

S

NP

Pron

I

VP

V

eat

NP

S

NP

Pron

I

VP

V

eat

NP PP

S

NP

Pron

I

VP

V

eat

NP AdvP

Page 106: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Left-embeddingJapanese and many other languages have pervasiveleft-embedding.

S

CC

S

CC

S

NP

PropN

Mary-gaMary

V

kitaarrived

Comp

to

NP

PropN

John-ga

V

ittasaid

Comp

to

V

omou[I] think

Page 107: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Left-embeddingJapanese and many other languages have pervasiveleft-embedding.

S

CC

S

CC

S

NP

PropN

Mary-gaMary

V

kitaarrived

Comp

to

NP

PropN

John-ga

V

ittasaid

Comp

to

V

omou[I] think

Page 108: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 109: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

I The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 110: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

I The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 111: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

I The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 112: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

I The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 113: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

Influence of Yngve on later workI Theories focusing on constituent expectations and

incremental parsing choices/difficulty (Kimball, 1973;Gibson, 1991)

I Theories of center-embedding difficulty have become morenuanced, as our understanding of data have improved(Gibson, 1991; Lewis, 1996):

I The fact that the employee who the manager hired stoleoffice supplies worried the executive.

I # The executive who the fact that the employee stole officesupplies worried hired the manager.

I The mechanism for discontinuous constituents bears aclose relationship to grammars with wrapping mechanisms(Pollard, 1984)

I The empirical phenomenon of “heavy shift” has receivedconsiderable recent attention (Hawkins, 1994; Wasow,2002)

Page 114: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

For Tuesday

Read Jurafsky 1996, focusing on Section 4.

Page 115: Computational Psycholinguistics, Day 1 Introduction and a …idiom.ucsd.edu/~rlevy/lsa308/lecture_1.pdf · 2007-07-06 · Computational Psycholinguistics: Day 1 Introduction Class

Computational Psycholinguistics: Day 1

Yngve 1960

Yngve’s hypothesis

References

Chomsky, N. (1956). Three models for the description of language. IRE Transactions on Information Theory,2(3):113–124.

Chomsky, N. (1957). Syntactic Structures. Mouton.

Gibson, E. (1991). A computational theory of human linguistic processing: memory limitations and processingbreakdown. PhD thesis, Carnegie Mellon.

Hawkins, J. A. (1994). A Performance Theory of Order and Constituency. Cambridge.

Kimball, J. (1973). Seven principles of surface structure parsing in natural language. Cognition, 2:15–47.

Lewis, R. L. (1996). Interference in short-term memory: The magical number two (or three) in sentence processing.Journal of Psycholinguistic Research, 25(1):93–116.

Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processinginformation. Psychological Review, 63:81–97.

Miller, G. A. and Chomsky, N. (1963). Finitary models of language users. In Luce, D. R., Bush, R. R., and Galanter,E., editors, Handbook of Mathematical Psychology, volume II. New York: John Wiley.

Pollard, C. (1984). Generalized Phrase Structure Grammars, Head Grammars, and Natural Languages. PhD thesis,Stanford.

Wasow, T. (2002). Postverbal Behavior. CSLI.