Upload
dominick-stanley
View
215
Download
1
Tags:
Embed Size (px)
Citation preview
2
Outline My own line of research Papers:
Fast Dropout training, ICML, 2013 Distributional Semantics Beyond Words: Supervised Learning
of Analogy and Paraphrase, TACL, 2013.
3
Current Line of Research Conventional approach to a classification problem:
Problems: Never use the label information
Lose the structure in the output Limited to the classes in the training set Hard to leverage unsupervised data
4
Current Line of Research For example take the relation extraction problem:
Conventional Approach: Given sentence s and mentions e1 and e2, find their relation: Output:
“Bill Gates, CEO of Microsoft ….”
Manager
{𝑠 ,𝑒1 ,𝑒2 }→𝑟
𝑟 ∈𝑅={𝑅1 ,... ,𝑅𝑑}
5
Current Line of Research Let’s change the problem a little:
Create a claim about the relation:
“Bill Gates, CEO of Microsoft ….”
R = Manager
Text=“Bill Gates, CEO of Microsoft ….”
Claim=“Bill Gates is manager of Microsoft”
True
6
Current Line of Research Creating data is very easy!
What we do: Use knowledge bases to find entities that are related Find sentences that contain these entities Create claims about the relation inside the original sentence Ask Turker’s to label it
Much easier than extracting labels and labelling
7
Current Line of Research This formulation makes use of the information inherent in
the label This helps us to generalize over the relations that are not
seen in the training data
8
Outline My own line of research Papers:
Fast Dropout training, ICML, 2013 Distributional Semantics Beyond Words: Supervised Learning
of Analogy and Paraphrase, TACL, 2013.
9
Dropout training Proposed by (Hinton et al, 2012)
Each time decide whether to delete one hidden unit with some probability p
10
Dropout training Model averaging effect
Among models, with shared parameters Only a few get trained Much stronger than the known regularizer
What about the input space? Do the same thing!
2H
11
Dropout training Model averaging effect
Among models, with shared parameters Only a few get trained Much stronger than the known regularizer
What about the input space? Do the same thing!
Dropout of 50% of the hidden units and 20% of the input units (Hinton et al, 2012)
2H
12
Outline Can we explicitly show that dropout acts as a regularizer?
Very easy to show for linear regression What about others?
Dropout needs sampling Can be slow!
Can we convert the sampling based update into a deterministic form? Find expected form of updates
13
Linear Regression Reminder:
Consider the standard linear regression
With regularization:
Closed form solution:
Tg w x
2* ( ) ( )arg min T i i
w i
w w x y
2( ) ( ) 2( ) T i ii
i i
L w w x y w
1T Tw X X I X y
~ ( )i iz Bernoulli p i iz p E
(1 )i i iz p p Var
14
Dropout Linear Regression Consider the standard linear regression
LR with dropout:
How to find the parameter?
Tg w x~ ( )i i ix z Bernoulli p
1( ,..., )z mD diag z z
Tzw D x
2( ) ( )( ) T i iz
i
L w w D x y
15
Fast Dropout for Linear Regression We had:
Instead of sampling, minimize the expected loss
Fixed x and y:
2
1
, ~ ( , )m
Tz i i i S S
i
w D x w x z S S N
S
2
1
(1 )m
i i i ii
w x p p
2( ) ( )( ) T i iz
i
L w w D x y
2Tzw D x y
E
1
m
i i ii
w x z
ETzw D x E
1
m
i i ii
w x p
2S T
zw D x Var 2
1
m
i i ii
w x z
Var
16
Fast Dropout for Linear Regression We had:
Instead of sampling minimize the expected loss:
Expected loss:
2
1
, ~ ( , )m
Tz i i i S S
i
w D x w x z S S N
2( ) ( )( ) T i iz
i
L w w D x y
2
2 2
~ ( , )S S
Tz S N
w D x y S y
E E
( )2( ) ( ) 2( ) ( )ii i
S Si
L w L w y E
2 2( ) ( ) 2 ( ), i i jS i i i i
i i j
y c w c x
2 2S Sy
17
Fast Dropout for Linear Regression Expected loss:
Data-dependent regulizer Closed form could be found:
( )2( ) ( ) 2( ) ( )ii i
S Si
L w L w y E
2 2( ) ( ) 2 ( ), i i jS i i i i
i i j
y c w c x
1( )T T Tw X X diag X X X y
18
Some definitions Dropout each input dimension randomly: Probit:
Logistic function / sigmoid :
2 /21( )
2
xtx e dt
1( )
1 zz
e
19
Some definitions useful equalities Useful equalities
We can find the following expectation in closed form:
2
2 2( ) ( ; , ) ( )x N x s dx
s
( ) ( )8
x x
2
2( ) ( ; , ) ( )
8 1x N x s dx
s
2~ ( , )( )
S NS
E
20
Logistic Regression Consider the standard LR
The standard gradient update rule is
For the parameter vector
1( 1| ) ( )
1T
T
w xP Y X x w x
e
( ( ))Tj jw y w x x
log ( ( ))Tw y w x x
21
Dropout on a Logistic Regression Dropout each input dimension randomly:
For the parameter vector
Notation:
~ ( )i i ix z Bernoulli p 1( ,..., )z mD diag z z
log ( ( )) ( ( ))T Tz zw y w x x w y w D x D x
-th dimension of ix i x( ) -th training instancejx j
( ) -th dimension of -th instancejix i j
1 i m 1 j n
22
Fast Dropout training Instead of using we use its expectation: w
; ~ ( )i iavg z z Bernoulli pw w E
2
1
, ~ ( , )m
Tz i i i S S
i
w D x w x z S S N
1 1
m mT
S z i i i i i ii i
w D x w x z w x p
E E
2 22
1 1
(1 )m m
TS z i i i i i i i
i i
w D x w x z w x p p
Var Var
; ~ ( ) ( ( ))i i
Tz z Bernoulli p z zy w D x D x E
23
Fast Dropout training Approx:
By knowing:
How to approximate?
Option 1:
Option 2:
Have closed forms but poor approximations
; ~ ( ) ( ( ))i i
Tavg z z Bernoulli p z zw y w D x D x E
2, ~ ( , )Tz S Sw D x S S N
( ( ))S z zy S D xE E
( ( ))S z zy S D x E E
27
Fast Dropout training Approx:
By knowing:
; ~ ( ) ( ( ))i i
Tavg z z Bernoulli p z zw y w D x D x E
2, ~ ( , )Tz S Sw D x S S N
, ; ~ ( ) ( ( ))i i
Tavg i z z Bernoulli p z i iw y w D x z x E
( 1)i ip z x
| 1 ( ( ))i i
Ti i z z zp x y w D x
E
| 1 ( )i i
Ti i z z zp x y w D x
E
| 1 ( ) ?i i
Tz z zw D x E
| 1 ( ( ))i i
Tz z zy w D x E
28
Fast Dropout training We want to:
Previously:
which could be
found in closed form.
| 1 ( ) ?i i
Tz z zw D x E
2, ~ ( , ), ~ ( )Tz S S i iw D x S S N z Bern p
| 1i iz z
| 1 ( ) ( )i i i
Tz z z S iw D x S E E
S i i i i iw x z w x E
2S i i i i iw x z w x Var
2~ ( , )i ii S SS N T
z i i i i iw D x w x z w z
iS (1 )S i i iw x p 2
iS 2 2( ) (1 )S i i i iw x p p
29
Fast Dropout training We want to:
Previously:
deviates (approximately) from with and
Has closed form!
| 1 ( ) ?i i
Tz z zw D x E
2, ~ ( , )Tz S Sw D x S S N
2( ) (1 )i i i iw x p p
iS S 2
2
2 2
22 ( , )
( )S
S
NS
E
2| 1 ~ ( , )i i
Ti i z i i i i i i S Sz z w D x w x z w z S N
(1 ),i i iw x piS S 2 2 2
iS S
| 1 ( )i i
Tz z zw D x E 2( , )
( )S SN
S
E 2( , )( )
S
S
NS
E