33
S LICED I NVERSE R EGRESSION 5 th March 2015, Grenoble A LESSANDRO C HIANCONE S TEPHANE G IRARD J OCELYN C HANUSSOT

Presentation 5 march persyval

Embed Size (px)

Citation preview

Page 1: Presentation 5 march persyval

SLICED INVERSE REGRESSION5th March 2015, Grenoble

ALESSANDRO CHIANCONE

STEPHANE GIRARD JOCELYN CHANUSSOT

Page 2: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

1

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

Page 3: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

1

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

Page 4: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

1

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

Page 5: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

2

Suppose k = 1 for simplicity: Y = f(�TX, ✏)

We want to find the direction � that best explains Y.

In other words if Y is fixed then �TX should not vary. Consequently our goal

is to find a direction � which minimizes the variations of �TX given Y .

To estimate �TX|Y we arrange Y in h slices each with the same number of

samples, our goal is to find a direction � which minimizes the within-slice

variance of �TX under the constraint var(�TX) = 1.

Since

ˆ

⌃ =

ˆB +

ˆW , where

ˆ

⌃, ˆB, ˆW are respectively the sample covariance

matrix, the between-slice covariance matrix and the within-slice covariance

matrix an equivalent approach is to maximize the between-slice variance under

the same constraint.

Page 6: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

Y

�TX

2

Page 7: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

S1

S2

Whithin-slice variance for slice S10

Y

�TX

2

Page 8: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

2

Suppose k = 1 for simplicity: Y = f(�TX, ✏)

We want to find the direction � that best explains Y.

In other words if Y is fixed then �TX should not vary. Consequently our goal

is to find a direction � which minimizes the variations of �TX given Y .

To estimate �TX|Y we arrange Y in h slices each with the same number of

samples, our goal is to find a direction � which minimizes the within-slice

variance of �TX under the constraint var(�TX) = 1.

Since

ˆ

⌃ =

ˆB +

ˆW , where

ˆ

⌃, ˆB, ˆW are respectively the sample covariance

matrix, the between-slice covariance matrix and the within-slice covariance

matrix an equivalent approach is to maximize the between-slice variance under

the same constraint.

Page 9: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

2

Suppose k = 1 for simplicity: Y = f(�TX, ✏)

We want to find the direction � that best explains Y.

In other words if Y is fixed then �TX should not vary. Consequently our goal

is to find a direction � which minimizes the variations of �TX given Y .

To estimate �TX|Y we arrange Y in h slices each with the same number of

samples, our goal is to find a direction � which minimizes the within-slice

variance of �TX under the constraint var(�TX) = 1.

Since

ˆ

⌃ =

ˆB +

ˆW , where

ˆ

⌃, ˆB, ˆW are respectively the sample covariance

matrix, the between-slice covariance matrix and the within-slice covariance

matrix an equivalent approach is to maximize the between-slice variance under

the same constraint.

Page 10: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

3

SIR IN PRACTICE

SIR is quite simple to implement:

• Split Y in h slices

• Estimate � using the slices,

ˆ

� is the between-slice covariance matrix

• Compute the eigendecomposition of the matrix

ˆ

�1ˆ

�, where

ˆ

⌃ is the

empirical covariance matrix of X

• Select the eigenvectors corresponding to the highest eigenvalues.

Page 11: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

4

SIR IN THEORY

Let us go back to the general case Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏).

Since f is unknown it is not possible to directly retrieve �1,�2, ...,�k but a

basis of the e.d.r. space S. It can be shown that the previous SIR methodology

solves this problem if the so-called Linearity Design Condition is verified:

(LDC) 8b 2 Rp E(bTX|�T1 X,�T

2 X, ...,�Tk X) = c0 + c1�T

1 X + ...+ ck�Tk X for

some constants c0, ..., ck.

Page 12: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

5

Recently many papers pointed out limitations of SIR since the

eigendecomposition can be challenging when the covariance matrix ⌃ is ill

conditioned. Many solutions have been proposed starting from preprocessing

the data using PCA

[1]to a more comprehensive approach to regularize SIR

[2].

The weakest point of SIR is the (LDC) which cannot be verified in practice

because it depends on the true directions �1, ...,�k. The condition holds in

case of elliptic symmetry and more generally it has been shown

[3]that if the

dimension p tends to infinity the condition is always approximately verified.

[2]BERNARD-MICHEL, Caroline; GARDES, Laurent; GIRARD, Stéphane. Gaussian regularized sliced inverse regression. Statistics and Computing, 2009, 19.1: 85-98.

[1]LI, Lexin; LI, Hongzhe. Dimension reduction methods for microarrays with application to censored survival data. Bioinformatics, 2004, 20.18: 3406-3412.

Page 13: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

5

Recently many papers pointed out limitations of SIR since the

eigendecomposition can be challenging when the covariance matrix ⌃ is ill

conditioned. Many solutions have been proposed starting from preprocessing

the data using PCA

[1]to a more comprehensive approach to regularize SIR

[2].

The weakest point of SIR is the (LDC) which cannot be verified in practice

because it depends on the true directions �1, ...,�k. The condition holds in

case of elliptic symmetry and more generally it has been shown

[3]that if the

dimension p tends to infinity the condition is always approximately verified.

[2]BERNARD-MICHEL, Caroline; GARDES, Laurent; GIRARD, Stéphane. Gaussian regularized sliced inverse regression. Statistics and Computing, 2009, 19.1: 85-98.

[3]HALL, Peter; LI, Ker-Chau. On almost linearity of low dimensional projections from high dimensional data. The annals of Statistics, 1993, 867-889.

[1]LI, Lexin; LI, Hongzhe. Dimension reduction methods for microarrays with application to censored survival data. Bioinformatics, 2004, 20.18: 3406-3412.

Page 14: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

6

CLUSTER SIR

[1]KUENTZ, Vanessa; SARACCO, Jérôme. Cluster-based sliced inverse regression. Journal of the Korean Statistical Society, 2010, 39.2: 251-267.

The LDC is verified when X follows an elliptically symmetric distribution(e.g. multi normality of X).

When X follows a Gaussian mixture model the condition does not globallyhold but it is verified locally (i.e. in each mixture).

Kuentz & Saracco [1] clusterized X to force the condition to hold locally ineach cluster and then combine the result in each cluster to obtain the finale.d.r. directions

Our work is based on this intuition. We first clusterize the predictor space Xthen a greedy merging algorithm is proposed to assign each cluster to its e.d.rspace taking into account the size of the cluster on which SIR is performed.

Page 15: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

6

CLUSTER SIR

[1]KUENTZ, Vanessa; SARACCO, Jérôme. Cluster-based sliced inverse regression. Journal of the Korean Statistical Society, 2010, 39.2: 251-267.

The LDC is verified when X follows an elliptically symmetric distribution(e.g. multi normality of X).

When X follows a Gaussian mixture model the condition does not globallyhold but it is verified locally (i.e. in each mixture).

Kuentz & Saracco [1] clusterized X to force the condition to hold locally ineach cluster and then combine the result in each cluster to obtain the finale.d.r. directions

Our work is based on this intuition. We first clusterize the predictor space Xthen a greedy merging algorithm is proposed to assign each cluster to its e.d.rspace taking into account the size of the cluster on which SIR is performed.

Page 16: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

SIR - CLUSTER SIR k = 1

SIR - Y = f(�TX)

Cluster SIR- X = X1 [X2 [ .... [Xc,

where c is the number of clusters

Yi = f(�TXi), i = 1, ..., c

The e.d.r. direction and the linking

function do not change depending

on the cluster.

7

Page 17: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

SIR - CLUSTER SIR COLLABORATIVE SIRk = 1

X = X1 [X2 [ ... [Xc

Yi = fi(�Ti Xi), �i 2 {�1, ...,�D}

The number D (D c) of e.d.r.directions is unknown.

The e.d.r. directions and the link

function may change depending

on the cluster.

A merging algorithm is introduced to

infer the number D based on the

collinearity of the vectors �i.

SIR - Y = f(�TX)

Cluster SIR- X = X1 [X2 [ .... [Xc,

where c is the number of clusters

Yi = f(�TXi), i = 1, ..., c

The e.d.r. direction and the linking

function do not change depending

on the cluster.

7

Page 18: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

8

The directions �1, ..., �c are obtained applying SIR independently in each

cluster.

A hierarchical structure is built to infer the number D of e.d.r. directions

using a proximity criteria.

The most collinear vector to a set of vectors A = {�1, ..., �c} given the

proximity criterion m(a, b) = cos

2(a, b) = (a

Tb)

2is the solution of the

following problem:

�(A) = max

a2Rp

X

�t2A

wtm(�t, a) s.t. kak = 1

= largest eigenvalue of

X

�t2A

wt�t�Tt

where wt are weights and sum to one.

To build the hierarchy we consider the following iterative algorithm initialized

with the set A = {{�1}, ..., {�c}}:

while card(A) 6= 1

Let a, b 2 A such that �(a [ b) > �(c [ d)8c, d 2 A

A = (A \ {a, b}) [ a [ b

end

at each step the cardinality of the set A decreases merging the most collinear

sets of directions. Therefore it is possible to infer the number D of underlying

e.d.r. spaces analyzing the values of � in the hierarchy.

Page 19: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

9

�1 �12�2 �11...

�11 �12...

...

...

�1 , ..., �12

�1 , �2

�1 , �2 �11 , �12

number of merges

Page 20: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

10

The directions �1, ..., �c are obtained applying SIR independently in each

cluster.

A hierarchical structure is built to infer the number D of e.d.r. directions

using a proximity criteria.

The most collinear vector to a set of vectors A = {�1, ..., �c} given the

proximity criterion m(a, b) = cos

2(a, b) = (a

Tb)

2is the solution of the

following problem:

�(A) = max

a2Rp

X

�t2A

wtm(�t, a) s.t. kak = 1

= largest eigenvalue of

X

�t2A

wt�t�Tt

where wt are weights and sum to one.

To build the hierarchy we consider the following iterative algorithm initialized

with the set A = {{�1}, ..., {�c}}:

while card(A) 6= 1

Let a, b 2 A such that �(a [ b) > �(c [ d)8c, d 2 A

A = (A \ {a, b}) [ a [ b

end

at each step the cardinality of the set A decreases merging the most collinear

sets of directions. Therefore it is possible to infer the number D of underlying

e.d.r. spaces analyzing the values of � in the hierarchy.

Page 21: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

11

REGULARIZATION

After merging, each cluster is assigned one of the D e.d.r. directions

ˆ

�1, ...,ˆ

�D

For each Xi, i = 1, ..., c we consider the D e.d.r directions and we analyze the

two-dimensional distributions:

(Yi,ˆ

Tj Xi) 8j = 1, ..., D

Then we select the direction

ˆ

�j⇤ , j⇤= min

j=1,...,D�2,j , where �2,j is the second

eigenvalue of the covariance matrix Cov(Yi,ˆ

Tj Xi).

This step reconsiders the data to select the best direction from the pool of the

D estimated directions

Page 22: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

12

EXPERIMENTAL RESULTS

[1]CHAVENT, Marie, et al. A sliced inverse regression approach for data stream. Computational Statistics, 2013, 1-24.

We simulated 100 di↵erent datasets following a gaussian mixture model:

• 10 mixture components

• uniform mixing proportions

• 2500 samples

• dimension p = 240

The response variable Y is generated using the hyperbolic sin link function:

Yi = sinh(�Ti Xi) + ✏, ✏ independent of X.

We will show the case where �i 2 {�1,�2} i.e. the number of e.d.r. directions

D = 2

The proximity criteria m(

ˆ�,�) = cos

2(

ˆ�,�) = (

ˆ�T�)2 is evaluated to assess

the quality of the estimation

[1]

Page 23: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

13

−10

−5

0

5

10

−6−4−202468−6

−4

−2

0

2

4

6

8

Projection on the first two principal components

Page 24: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

14

The proximity criteria is computed using the estimation obtained in each

cluster independently, PC, and after collaborative SIR, PCM , for each run of

K-means:

The average PC is 0.4499 with a standard deviation of 0.0690

The average PCM is 0.7939 with a standard deviation of 0.0960

1 2 3 4 5 6 7 8 90

10

20

30

40

50

60

0.2 0.25 0.3 0.35 0.4 0.45 0.50

5

10

15

20

25

PCM � PC

Page 25: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

15

Y

projection

Page 26: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

16

REAL DATASET: HORSE MUSSELS

We consider the data example of horse mussels present in cluster SIR. The

observations correspond to n = 192 horse mussels captured in the Malborough

Sounds at the Northeast of New Zealand’s South Island. The response variable

Y is the muscle mass, the edible portion of the mussel, in grams. The

predictor X is of dimension p = 4 and measures numerical characteristics of

the shell: length, width, height, each in mm, and mass in grams.

We repeated the following algorithm 100 times:

(1) Randomly select 80% of training and 20% of test.

(2) Apply SIR, cluster SIR and collaborative SIR on the training.

(3) Regress the functions using the training samples

(4) Compute the Mean Absolute Relative Error (MARE) on the test

Page 27: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

17

MARE

Page 28: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

18

Page 29: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

19

REAL DATASET: GALAXIES

We consider the data example of Galaxies. The observations correspond to

n = 292766 di↵erent galaxies. The response variable Y is the stellar formation

rate. The predictor X is of dimension p = 46 and is composed of spectral

characteristics of the galaxies

We applied Collaborative SIR on the whole dataset to investigate the presence

of subgroups and di↵erent directions.

After di↵erent runs and number k of clusters we can observe that there are

two di↵erent subgroups and hence directions �1,�2.

Page 30: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

20

Stellarformationrate

XT�i, i = 1, 2

Page 31: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

21

�1

�2

Di↵erences between �1 and �2

Page 32: Presentation 5 march persyval

SLICED INVERSE REGRESSIONALESSANDRO CHIANCONE

1 2 3

ALESSANDRO CHIANCONE

SEQUENTIAL DIMENSION REDUCTION

STEPHANE GIRARD JOCELYN CHANUSSOT

Given a response continuous variable Y 2 R and X 2 Rp explanatory variableswe look for a functional relationship between the two:

Y = f(X, ✏), ✏ independent of X

It is reasonable to think that X contains more information that what we needfor our task of predicting Y , in addition when the dimension p is largeregression is a very hard task when we do not have information about f .

Given that a possible solution is to make a strong assumption:

Y = f(�T1 X,�T

2 X, ...,�Tk X, ✏)

The space S = Span{�1,�2, ...,�k} is called e↵ective dimension reduction(e.d.r.) space. Only few linear combinations of the predictors, k ⌧ p, aresu�cient to regress Y .

22

CONCLUSIONS

Slices Inverse Regression is definitely and underestimated method, give it a try!

Since data nowadays are complex and rich, our Collaborative SIR may catch the subgroups and allows to reduce dimensionality in an easy and fast way

ROBUSTSIR

FUTURE WORK

PublicationsChini, Marco, Alessandro Chiancone, and Salvatore Stramondo. “Scale Object Selection (SOS) through a hierarchical segmentation by a

multi-spectral per-pixel classification.” Pattern Recognition Letters 49 (2014): 214-223.\

Papers in preparationAlessandro Chiancone, Stephane Girard and Jocelyn Chanussot. “Collaborative Sliced Inverse Regression.”

Page 33: Presentation 5 march persyval

[4]KUENTZ, Vanessa; SARACCO, Jérôme. Cluster-based sliced inverse regression. Journal of the Korean Statistical Society, 2010, 39.2: 251-267.

[2]Bernard-Michel, C., Gardes, L., & Girard, S. (2009). Gaussian regularized sliced inverse regression. Statistics and Computing, 19(1), 85-98.

[3]HALL, Peter; LI, Ker-Chau. On almost linearity of low dimensional projections from high dimensional data. The annals of Statistics, 1993, 867-889.

[1]LI, Lexin; LI, Hongzhe. Dimension reduction methods for microarrays with application to censored survival data. Bioinformatics, 2004, 20.18: 3406-3412.

[1]CHAVENT, Marie, et al. A sliced inverse regression approach for data stream. Computational Statistics, 2013, 1-24.

CLASSIFICATION

SEGMENTATIONREGRESSION

thank you