76
Introduction Models Emergent properties Conclusion . Typological consequences of agent interaction Coral Hughto Robert Staubs Joe Pater University of Massachusetts Amherst NECPhon 8 November 15, 2014 Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8 Typological consequences of agent interaction 1

In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

  • Upload
    others

  • View
    19

  • Download
    0

Embed Size (px)

Citation preview

Page 1: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Typological consequences of agent interaction

Coral Hughto Robert Staubs Joe Pater

University of Massachusetts Amherst

NECPhon 8November 15, 2014

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 1

Page 2: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

In standard generative grammar: Grammatical theories areconstructed to generate all and only possible languages.

Some systems are permitted by the theory, others are not. Nodistinction is made within either class.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 2

Page 3: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

In standard generative grammar: Grammatical theories areconstructed to generate all and only possible languages.

Some systems are permitted by the theory, others are not. Nodistinction is made within either class.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 2

Page 4: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Standard goal of learning theories: Show how the systemsgenerated might be learned.

No independent role of learning in typological modeling.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 3

Page 5: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Standard goal of learning theories: Show how the systemsgenerated might be learned.

No independent role of learning in typological modeling.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 3

Page 6: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

We can do better than this: Explain relative frequency based onrelative learnability—combining a learning theory with agrammatical theory (e.g. Heinz 2009, Pater and Moreton 2012,Staubs 2014).

Individual learners acquire particular patterns faster or slower basedon how learning and grammar interact.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 4

Page 7: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

We can do better than this: Explain relative frequency based onrelative learnability—combining a learning theory with agrammatical theory (e.g. Heinz 2009, Pater and Moreton 2012,Staubs 2014).

Individual learners acquire particular patterns faster or slower basedon how learning and grammar interact.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 4

Page 8: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Today we’ll focus on a third model bias.

This bias emerges from interaction between agents, both withinand across generations.

We show consequences particularly for probabilistic models ofgrammar such as Maximum Entropy (Goldwater and Johnson2003).

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 5

Page 9: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Today we’ll focus on a third model bias.

This bias emerges from interaction between agents, both withinand across generations.

We show consequences particularly for probabilistic models ofgrammar such as Maximum Entropy (Goldwater and Johnson2003).

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 5

Page 10: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Today we’ll focus on a third model bias.

This bias emerges from interaction between agents, both withinand across generations.

We show consequences particularly for probabilistic models ofgrammar such as Maximum Entropy (Goldwater and Johnson2003).

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 5

Page 11: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Our focus today: Interaction and transmission tend to reducevariability.

This happens in two fundamentally different network assumptions:iterated and interactive learning.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 6

Page 12: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Our focus today: Interaction and transmission tend to reducevariability.

This happens in two fundamentally different network assumptions:iterated and interactive learning.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 6

Page 13: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

We show that these models show emergent tendencies towards:

1 Categorical outcomes

2 Lexical contrast

3 Avoidance of cumulativity

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 7

Page 14: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Error-driven learning

MaxEnt SGA (perceptron, HG-GLA; Jager 2007, Boersma andPater 2014):

New Weights =Old Weights + η × (Learner Violations − Teacher Violations)

Where η is some assumed learning rate.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 8

Page 15: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Error-driven learning

MaxEnt SGA (perceptron, HG-GLA; Jager 2007, Boersma andPater 2014):

New Weights =Old Weights + η × (Learner Violations − Teacher Violations)

Where η is some assumed learning rate.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 8

Page 16: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Iterated Learning

Iterated learning models present a simplified model of languagechange

These models are based on the observation that language changehappens over time: children’s grammars are not exactly the sameas their parents’

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 9

Page 17: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Agents in this model are arranged in a chain with one learner per“generation”

L1 → L2 → ... → Ln

Each agent in a chain learns its language from the previousgeneration and then teaches it to the next (Kirby and Hurford2002, Griffiths and Kalish 2007)

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 10

Page 18: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Typologically common languages coincide with languages whichare stable (transmitted faithfully) under this learning model

Agents in an iterated learning chain preserve categorical grammarstates better/longer than more variable grammars

This trend towards categoricity emerges through the transmissionof languages between agents, without needing to encode a bias forcategoricity within each agent

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 11

Page 19: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive Learning

Interactive learning models present a simplified model of languagegeneration (Dediu 2009, Pater and Moreton 2012)

A number of agents interact with and learn from each other:

L1 ↔ L2

From these interactions, a shared grammar emerges

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 12

Page 20: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

This model is based on the observation that language change is asocial phenomenon

An individual’s language use continues to change over time, andtheir language use is affected by that of their social network

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 13

Page 21: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Probabilistic typological trends are reflected in the rate at whichthe agents generate particular systems under this model

The shared grammars developed by agents in an interactivelearning model tend to be categorical

These effects are emergent properties of the model, and don’trequire any specifically encoded learning biases

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 14

Page 22: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Iterated or Interactive?

Iterated learning models emphasize the importance of the effect oftransmission of language between generations (from adults tochildren), setting aside the social, interactive aspect of languagelearning

Interactive learning models emphasize the influence of peers onlanguage development, setting aside the influence from adultlanguage users

Both of these models are overly simplified; human languagelearning is probably influenced by both types of interaction

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 15

Page 23: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Categoricity

An interactive learner, starting with equal probabilities oncandidates, will tend toward weights giving categorical outcomes.

Categoricity tableau

*A *B

A -1B -1

Dark lines with gray: means of 100 runs with standard deviations.Learning rate 0.1.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 16

Page 24: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Categoricity

An interactive learner, starting with equal probabilities oncandidates, will tend toward weights giving categorical outcomes.

Categoricity tableau

*A *B

A -1B -1

Dark lines with gray: means of 100 runs with standard deviations.Learning rate 0.1.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 16

Page 25: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Categoricity

An interactive learner, starting with equal probabilities oncandidates, will tend toward weights giving categorical outcomes.

Categoricity tableau

*A *B

A -1B -1

Dark lines with gray: means of 100 runs with standard deviations.Learning rate 0.1.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 16

Page 26: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 1000 2000 3000 4000 5000

0.0

0.2

0.4

0.6

0.8

1.0

Interactive categoricity, zero start weights

Iterations

Pro

babi

lity

diffe

renc

e of

can

dida

tes

Black line: mean, Gray: standard deviation, 100 repetitions, 0.1learning rate

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 17

Page 27: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

The starting distribution is not crucial.

The learners converge on a shared categorical outcome even if theyinitially categorically disagree.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 18

Page 28: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 1000 2000 3000 4000 5000

0.0

0.2

0.4

0.6

0.8

1.0

Interactive categoricity, opposite start weights

Iterations

Pro

babi

lity

diffe

renc

e of

can

dida

tes

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 19

Page 29: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with probabilistic grammars, starting withnon-categorical grammars:

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars, they changeless and less.

3 The system spends most of its time in categorical states.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 20

Page 30: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with probabilistic grammars, starting withnon-categorical grammars:

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars, they changeless and less.

3 The system spends most of its time in categorical states.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 20

Page 31: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with probabilistic grammars, starting withnon-categorical grammars:

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars, they changeless and less.

3 The system spends most of its time in categorical states.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 20

Page 32: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with probabilistic grammars, starting withnon-categorical grammars:

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars, they changeless and less.

3 The system spends most of its time in categorical states.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 20

Page 33: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with constraints (e.g. MaxEnt, Noisy HG;Boersma and Pater 2014):

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars1 They change less and less.

2 The effective change to probability from a change in weightsshrinks.

3 The system spends most of its time in categorical states.

(cf. Wedel 2007 on models where a positive feedback loop createssimilar pressures)

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 21

Page 34: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with constraints (e.g. MaxEnt, Noisy HG;Boersma and Pater 2014):

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars1 They change less and less.2 The effective change to probability from a change in weights

shrinks.

3 The system spends most of its time in categorical states.

(cf. Wedel 2007 on models where a positive feedback loop createssimilar pressures)

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 21

Page 35: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interactive learning with constraints (e.g. MaxEnt, Noisy HG;Boersma and Pater 2014):

1 Errors can push the agents towards either more or lesscategorical states.

2 As the agents drift into categorical grammars1 They change less and less.2 The effective change to probability from a change in weights

shrinks.

3 The system spends most of its time in categorical states.

(cf. Wedel 2007 on models where a positive feedback loop createssimilar pressures)

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 21

Page 36: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 1000 2000 3000 4000 5000

0.0

0.2

0.4

0.6

0.8

1.0

Example run

Iteration

Pro

babi

lity

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 22

Page 37: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 1000 2000 3000 4000 5000

0.0

0.2

0.4

0.6

0.8

1.0

Example run: oscillation

Iteration

Pro

babi

lity

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 23

Page 38: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Part of this pressure (at least) is present in iterated learning as well(see e.g. Dediu 2009, p. 555).

Thus iterated learning can show a pressure for increasingcategoricity over generations.

This requires enough learning to happen in each step in order tomaintain the “emerged” categoricity.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 24

Page 39: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Part of this pressure (at least) is present in iterated learning as well(see e.g. Dediu 2009, p. 555).

Thus iterated learning can show a pressure for increasingcategoricity over generations.

This requires enough learning to happen in each step in order tomaintain the “emerged” categoricity.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 24

Page 40: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Part of this pressure (at least) is present in iterated learning as well(see e.g. Dediu 2009, p. 555).

Thus iterated learning can show a pressure for increasingcategoricity over generations.

This requires enough learning to happen in each step in order tomaintain the “emerged” categoricity.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 24

Page 41: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Terminology reminder

1 iteration: when a datum is exchanged between two agents

2 generation: when a new agent learns from another for anumber of iterations

Thus iterations are relevant to both iterative and interactive.

Generations are not clearly important to interactive learning.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 25

Page 42: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 20 40 60 80 100

0.0

0.2

0.4

0.6

0.8

1.0

Iterative categoricity, 10 iterations

Generations

Pro

babi

lity

diffe

renc

e of

can

dida

tes

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 26

Page 43: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 20 40 60 80 100

0.0

0.2

0.4

0.6

0.8

1.0

Iterative categoricity, 100 iterations

Generations

Pro

babi

lity

diffe

renc

e of

can

dida

tes

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 27

Page 44: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 20 40 60 80 100

0.0

0.2

0.4

0.6

0.8

1.0

Iterative categoricity, 1000 iterations

Generations

Pro

babi

lity

diffe

renc

e of

can

dida

tes

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 28

Page 45: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cumulativity

In a weighted-constraint grammar, constraint violations arecumulative

The optimal candidate is the one whose Harmony score is closestto zero, but the particular combination of constraint weights andviolations doesn’t matter

A candidate which incurs one violation of a constraint with aweight of 6 has the same Harmony score as a candidate whichincurs two violations of a constraint with a weight of 3(1*6 = 2*3 = 6)

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 29

Page 46: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Constraint cumulativity has been cited as a problem forweighted-constraint grammars, as it makes undesirable typologicalpredictions (e.g. Legendre et al. 2006).

If cumulativity effects exist, it seems they might be uncommon.

If either fact is true, we should worry about a model that treatscumulative languages identically with non-cumulative ones.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 30

Page 47: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cumulativity tableaux

3 2X Y H

→A -1 -2B -1 -3

→C -1 -4D -2 -6

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 31

Page 48: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

In an interactive learning model, the agents strongly tend awayfrom cumulative patterns

One reason: many cumulative weightings are intermediate andnon-categorical (Carroll 2012).

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 32

Page 49: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cyan and orange: no cumulativity effect. Black: cumulativity.Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 33

Page 50: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cumulativity

A simulation with two agents beginning with constraint weights atzero, run 1000 times, produced no cumulative patterns

X Y

A -1B -1

C -1D -2

Language Count

A, D 312B, C 688A, C 0

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 34

Page 51: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 500 1000 1500

0.0

0.4

0.8

Avoidance of cumulative pattern, starting at 88% probability

Learning Step

Pro

babi

lity

of [a

]

0 500 1000 1500

0.0

0.4

0.8

Learning Step

Pro

babi

lity

of [c

]

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 35

Page 52: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 1000 2000 3000 4000 5000

0.0

0.4

0.8

Maintaining cumulative pattern, starting at very high probabilities

Learning Step

Pro

babi

lity

of [a

]

0 1000 2000 3000 4000 5000

0.0

0.4

0.8

Learning Step

Pro

babi

lity

of [c

]

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 36

Page 53: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cumulativity

Carroll (2012) analyses the real-world typology of contrastsbetween /s/ and /S/, finding the following distribution oflanguages:

Contrast Type Proportion

Total Neutralization 44.0%Full Contrast 37.0%Complementary Distribution 10.3%Contextual Neutralization 8.2%“Elsewhere” Neutralization 0.5%

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 37

Page 54: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Cumulativity

The “Elsewhere” Neutralization pattern is representable as acumulative pattern in a weighted-constraint grammar, and islargely underrepresented in the typology

Carroll (2012) attempts to account for this skew away fromcumulative patterns through encoding various biases into aMaxEnt learner, but doesn’t find a solution that fits the data aswell as desired

The interactive learning model presented here derives theavoidance of cumulative patterns that Carroll was looking for,without needing to encode specific learning biases

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 38

Page 55: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Contrast

The pressure for categoricity can be extended into a pressure forcontrast.

Now we have agents pronouncing different meanings, not justuninterpretable strings.

We add constraints like M1 → A “Pronounce M1 as A.”

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 39

Page 56: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Contrast

The pressure for categoricity can be extended into a pressure forcontrast.

Now we have agents pronouncing different meanings, not justuninterpretable strings.

We add constraints like M1 → A “Pronounce M1 as A.”

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 39

Page 57: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Contrast

The pressure for categoricity can be extended into a pressure forcontrast.

Now we have agents pronouncing different meanings, not justuninterpretable strings.

We add constraints like M1 → A “Pronounce M1 as A.”

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 39

Page 58: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Contrast tableau

M1 → A M1 → B M2 → A M2 → B

M1A -1B -1

M2A -1B -1

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 40

Page 59: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Meanings are not apparent from surface forms, they must beinferred.

The agents use Robust Interpretive Parsing (RIP; Tesar andSmolensky 2000, Boersma 2003, Jarosz 2013, Boersma and Pater2014):

Agents choose the meaning that they would most likely pronouncewith the observed surface form.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 41

Page 60: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Meanings are not apparent from surface forms, they must beinferred.

The agents use Robust Interpretive Parsing (RIP; Tesar andSmolensky 2000, Boersma 2003, Jarosz 2013, Boersma and Pater2014):

Agents choose the meaning that they would most likely pronouncewith the observed surface form.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 41

Page 61: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Meanings are not apparent from surface forms, they must beinferred.

The agents use Robust Interpretive Parsing (RIP; Tesar andSmolensky 2000, Boersma 2003, Jarosz 2013, Boersma and Pater2014):

Agents choose the meaning that they would most likely pronouncewith the observed surface form.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 41

Page 62: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Interpretation

Teacher Production: M1 → ab

Interpretation: a → M1M2

Learner Production: M2 → ab

Update: Output is not a ⇒ M2 → a ↑, M2 → b ↓

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 42

Page 63: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 200 400 600 800 1000

0.0

0.2

0.4

0.6

0.8

1.0

Emergent contrast, maximizing RIP

Iterations

Pro

babi

lity

diffe

renc

e of

mea

ning

s

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 43

Page 64: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Our assumption that RIP finds the most likely word acceleratescontrast.

Errors in interpretation point to non-categoricalprobabilities—maximizing helps find these.

If we sample instead of maximizing, however, we still get this kindof trend.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 44

Page 65: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 200 400 600 800 1000

0.0

0.2

0.4

0.6

0.8

1.0

Emergent contrast, sampling RIP

Iterations

Pro

babi

lity

diffe

renc

e of

mea

ning

s

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 45

Page 66: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

0 2000 4000 6000 8000 10000

0.0

0.2

0.4

0.6

0.8

1.0

Emergent contrast, sampling RIP

Iterations

Pro

babi

lity

diffe

renc

e of

mea

ning

s

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 46

Page 67: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Similar patterns are found with iterative learning.

Similar pressures for categoricity → similar contrast effects.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 47

Page 68: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Similar patterns are found with iterative learning.

Similar pressures for categoricity → similar contrast effects.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 47

Page 69: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Conclusions

We have shown language learners in a network tend towardsstability with categorical grammars.

This tendency emerges from interaction and transmission:Categorical patterns are those with the most reliability acrossgenerations and interactions.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 48

Page 70: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Conclusions

We have shown language learners in a network tend towardsstability with categorical grammars.

This tendency emerges from interaction and transmission:Categorical patterns are those with the most reliability acrossgenerations and interactions.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 48

Page 71: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

This tendency addresses several possible issues with probabilisticmodels:

Why are languages more categorical than they could be?

How can categorical contrast emerge?

Why are gang effects not (seemingly) ubiquitous?

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 49

Page 72: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

This tendency addresses several possible issues with probabilisticmodels:

Why are languages more categorical than they could be?

How can categorical contrast emerge?

Why are gang effects not (seemingly) ubiquitous?

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 49

Page 73: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

This tendency addresses several possible issues with probabilisticmodels:

Why are languages more categorical than they could be?

How can categorical contrast emerge?

Why are gang effects not (seemingly) ubiquitous?

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 49

Page 74: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

This tendency addresses several possible issues with probabilisticmodels:

Why are languages more categorical than they could be?

How can categorical contrast emerge?

Why are gang effects not (seemingly) ubiquitous?

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 49

Page 75: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

More broadly, this work reinforces the importance of viewinggrammatical models in context:

1 We must consider learning models and their concomitantbiases.

2 We must consider how these learning models interact to formtypological patterns.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 50

Page 76: In standard generative grammar: Grammatical theories are · In standard generative grammar: Grammatical theories are constructed to generate all and only possible languages. Some

Introduction Models Emergent properties Conclusion .

Thank you!

This material is based upon work supported by the NationalScience Foundation under Grant No. S121000000211 to thesecond author, Grants BCS-0813829 and BCS-424077 to theUniversity of Massachusetts, and by the city of Paris under aResearch in Paris fellowship to the third author.

We would also like to thank Lucien Carroll.

Coral Hughto, Robert Staubs, Joe Pater UMass Amherst NECPhon 8

Typological consequences of agent interaction 51