21
INTRODUCTION TO Machine Learning ETHEM ALPAYDIN © The MIT Press, 2004 [email protected] http://www.cmpe.boun.edu.tr/~ethem/i2ml Lecture Slides for

INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

  • Upload
    others

  • View
    11

  • Download
    0

Embed Size (px)

Citation preview

Page 1: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

INTRODUCTION TO

Machine Learning

ETHEM ALPAYDIN© The MIT Press, 2004

[email protected]://www.cmpe.boun.edu.tr/~ethem/i2ml

Lecture Slides for

Page 2: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

CHAPTER 6:

Dimensionality Reduction

Page 3: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

3

Why Reduce Dimensionality?

1. Reduces time complexity: Less computation

2. Reduces space complexity: Less parameters

3. Saves the cost of observing the feature

4. Simpler models are more robust on small datasets

5. More interpretable; simpler explanation

6. Data visualization (structure, groups, outliers, etc) if plotted in 2 or 3 dimensions

Page 4: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

4

Feature Selection vs Extraction

Feature selection: Choosing k<d important features, ignoring the remaining d – k

Subset selection algorithms

Feature extraction: Project the

original xi , i =1,...,d dimensions to

new k<d dimensions, zj , j =1,...,k

Principal components analysis (PCA), linear discriminant analysis (LDA), factor analysis (FA)

Page 5: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

5

Subset Selection

There are 2d subsets of d featuresForward search: Add the best feature at each step

Set of features F initially Ø.At each iteration, find the best new feature

j = argmini E ( F ∪ xi )Add xj to F if E ( F ∪ xj ) < E ( F )

Hill-climbing O(d2) algorithmBackward search: Start with all features and remove

one at a time, if possible.Floating search (Add k, remove l)

Page 6: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

6

Principal Components Analysis (PCA)

Find a low-dimensional space such that when x is projected there, information loss is minimized.The projection of x on the direction of w is: z = wTxFind w such that Var(z) is maximized

Var(z) = Var(wTx) = E[(wTx – wTµ)2] = E[(wTx – wTµ)(wTx – wTµ)]= E[wT(x – µ)(x – µ)Tw]

= wT E[(x – µ)(x –µ)T]w = wT∑w

where Var(x)= E[(x – µ)(x –µ)T] = ∑

Page 7: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

7

Maximize Var(z) subject to ||w||=1

∑w1 = αw1 that is, w1 is an eigenvector of ∑

Choose the one with the largest eigenvalue for Var(z) to be max

Second principal component: Max Var(z2), s.t., ||w2||=1 and orthogonal to w1

∑ w2 = α w2 that is, w2 is another eigenvector of ∑

and so on.

( )1max 11111

−α− wwwww

TTΣ

( ) ( )01max 1222222

−β−−α− wwwwwww

TTTΣ

Page 8: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

8

What PCA doesz = WT(x – m)

where the columns of W are the eigenvectors of ∑, and m is sample mean

Centers the data at the origin and rotates the axes

Page 9: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

9

How to choose k ?

Proportion of Variance (PoV) explained

when λi are sorted in descending order

Typically, stop at PoV>0.9

Scree graph plots of PoV vs k, stop at “elbow”

dk

k

λ++λ++λ+λλ++λ+λLL

L

21

21

Page 10: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

10

Page 11: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

11

Page 12: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

12

Factor Analysis

Find a small number of factors z, which when combined generate x :

xi – µi = vi1z1 + vi2z2 + ... + vikzk + εi

where zj, j =1,...,k are the latent factors with E[ zj ]=0, Var(zj)=1, Cov(zi ,, zj)=0, i ≠ j ,

εi are the noise sourcesE[ εi ]= ψi, Cov(εi , εj) =0, i ≠ j, Cov(εi , zj) =0 ,

and vij are the factor loadings

Page 13: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

13

PCA vs FA

PCA From x to z z = WT(x – µ)

FA From z to x x – µ = Vz + ε

x z

z x

Page 14: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

14

Factor Analysis

In FA, factors zj are stretched, rotated and translated to generate x

Page 15: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

15

Multidimensional Scaling

Given pairwise distances between N points,

dij, i,j =1,...,N

place on a low-dim map s.t. distances are preserved.

z = g (x | θ ) Find θ that min Sammon stress

( )( )

( ) ( )( )∑

−−θ−θ=

−−−=θ

s,r sr

srsr

s,r sr

srsr

E

2

2

2

2

||

|

xx

xxxgxg

xx

xxzzX

Page 16: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

16

Map of Europe by MDS

Map from CIA – The World Factbook: http://www.cia.gov/

Page 17: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

17

Linear Discriminant Analysis

Find a low-dimensional space such that when xis projected, classes are well-separated.

Find w that maximizes

( ) ( )

( )∑∑∑ −==

+−

=

t

ttT

t

tt

ttT

rmsr

rm

ssmm

J

2

1211

22

21

221

xwxw

w

Page 18: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

18

Between-class scatter:

Within-class scatter:

( ) ( )( )( )

( )( )TBBT

TT

TTmm

2121

2121

2

212

21

where mmmmww

wmmmmw

mwmw

−−==

−−=

−=−

SS

( )( )( )

( )( )21

21

21

111

111

2

121

where

where

SSSS

S

S

+==+

−−=

=−−=

−=

∑∑∑

WWT

t

t

Ttt

Tt

t

TttT

t

t

tT

ss

r

r

rms

ww

mxmx

wwwmxmxw

xw

Page 19: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

19

Fisher’s Linear Discriminant

Find w that max

LDA soln:

Parametric soln:

( )211 mmw −⋅= −

Wc S

( )( )

ww

mmw

wwww

wW

T

T

WT

BT

JSS

S2

21 −==

( )( ) ( )Σ

−Σ= −

,~Cp ii µµµ

N| n whe21

1

x

w

Page 20: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

20

K>2 Classes

Within-class scatter:

Between-class scatter:

Find W that max

( )( )Tit

it

t

tii

K

iiW r mxmx −−== ∑∑

=

SSS 1

( )( ) ∑∑==

=−−=K

ii

K

i

TiiiB K

N11

1 mmmmmmS

( )WSW

WSWW

WT

BT

J =The largest eigenvectors of SW

-1SBMaximum rank of K-1

Page 21: INTRODUCTION TO Machine Learningethem/i2ml/slides/v1-1/i2ml-chap6-v… · Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) 4 Feature Selection

Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)

21