13
Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario Silva, Isabel Trancoso Learning Word Representations from Scarce Data By: Aadil Hayat (13002)

Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

  • Upload
    others

  • View
    1

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario Silva, Isabel Trancoso

Learning Word Representations

from Scarce Data

By: Aadil Hayat (13002)

Page 2: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Learning Word Representations from Scarce and Noisy Data with Embedding Sub-spaces

Introduction Theory Results

Page 3: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Introduction

Unsupervised word embedding for scarce and noisy data

Page 4: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Abstract

A technique to adapt unsupervised word embeddings to specific applications, when only small and noisy labeled datasets are available.

Current methods use pre-trained embeddings to initialize model parameters, and then use the labeled data to tailor them for the intended task.

But this approach is prone to overfitting when the training is performed with scarce and noisy data.

To overcome this issue here the supervised data to find an embedding subspace that fits the task complexity.

All the word representations are adapted through a projection into this task-specific sub-space.

This approach was recently used in the SemEval 2015 Twitter sentiment analysischallenge, attaining state-of-the-art results.

Page 5: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Theory

Page 6: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Unsupervised Structured Skip-Gram

Word embeddings are generally trained by optimizing an objective function that can be measured without annotations.

One popular approach is to estimate the embeddings by maximizing the probability thatthe words within a given window size are predicted correctly.

Here unsupervised structured skip-gram model is used, which is a modification of the skip-gram architecture that has been shown to outperform the original model in syntax based tasks such as POS tagging and dependency parsing.

Structured skip-gram depends on a log-linear model of word prediction. The structured skip-gram models the following probability:

𝒑 𝒘𝒑 = 𝒋 𝒘 = 𝒊) ∝ 𝐞𝐱𝐩 𝑪𝒋𝒑∙ 𝑬 ∙ 𝒘𝒊

𝑤ℎ𝑒𝑟𝑒 𝑤𝑖 ∈ 1,0 𝑣x1 𝑜𝑛𝑒 − ℎ𝑜𝑡 𝑟𝑒𝑝𝑟𝑒𝑠𝑒𝑛𝑡𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑤 = 𝑖 ,𝐸 ∈ 𝑅𝑒x𝑣 𝑖𝑠 𝑡ℎ𝑒 𝑒𝑚𝑏𝑒𝑑𝑑𝑖𝑛𝑔 𝑚𝑎𝑡𝑟𝑖𝑥,

𝐶𝑗𝑝∈ 𝑅vx𝑒 𝑖𝑠 𝑎 𝑠𝑒𝑡 𝑜𝑓 𝑜𝑢𝑡𝑝𝑢𝑡 𝑚𝑎𝑡𝑟𝑖𝑐𝑒𝑠

Page 7: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Adapting Embedding with Sub-space Projections

Word embeddings are useful unsupervised techniques to attain initial model values or features prior to supervised learning. These models can be then retrained using the available labeled data.

Embedding provide a compact real valued representations of each word in a vocabulary.

Even then the total number of parameters in the model can be rather high. Very often a small amount of supervised data is available which can lead to severe overfitting.

Even if regularization is used to reduce the overfitting only a reduced subset of words willactually be present in the labeled dataset. Words not seen during training will never get their embeddings updated.

In the following slides, simple solution to this problem is explained.

Page 8: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Embedding Sub-space

The adapted embedding matrix is defined as the factorization 𝑆 ∙ 𝐸 , 𝑤ℎ𝑒𝑟𝑒 𝑆 ∈ 𝑅𝑠x𝑒 , 𝑤𝑖𝑡ℎ 𝑠 ≪ 𝑒

The parameters of the matrix S are estimated using labeled dataset, while E is kept fixed.

We determine the optimal projection of the embedding matrix E into a sub-space of determine s.

The idea of embedding sub-space rests on two fundamental principles:1). With dimensionality reduction of the embeddings, the model can better fit the

complexity of the task at hand or the amount of available data.2). Using a projection, all the embeddings are indirectly updated, not only those of

words present in the labeled dataset.

We assume that the words on the labeled dataset are, to some extent, representative ofthe words found in the unlabeled corpus.

Page 9: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Non-Linear Sub-space Embedding Model

The concept of embedding sub-space can be applied to log-linear classifiers ordeep learning architecture that uses embeddings.

The NLSE can be interpreted as a simple feed-forward neural network model with one single hidden layer utilizing the embedding sub-space approach.

Page 10: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Results

Page 11: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Twitter Sentiment Analysis

Average F-measure on the SemEval test sets varying with embedding sub-spacesize s. Sub-space size 0 used to denote the baseline (log-linear model)

Page 12: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Comparison of two baselines with two variations Performance of state-of-the-art systems for Twitter sentiment prediction

Page 13: Authors : Ramon F. Astudillo, Silvio Amir, Wang Lin, Mario ...home.iitk.ac.in/~aadilh/cs671/hw3/Learning word embeddings from scarce data.pdfWang Lin, Mario Silva, Isabel Trancoso

Thank You