22
Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University http:// www.ntu.edu.sg/home/dongxu [email protected]

Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Embed Size (px)

Citation preview

Page 1: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Enhancing Tensor Subspace

Learning by Element Rearrangement

1

Dong XU

School of Computer Engineering

Nanyang Technological Universityhttp://www.ntu.edu.sg/home/dongxu

[email protected]

Page 2: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Outline

• Summary of our recent works on Tensor (or Bilinear) Subspace Learning

• Element Rearrangement for Tensor (or Bilinear) Subspace Learning

Page 3: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

What is Tensor?

Tensors are arrays of numbers which transform in certain ways under coordinate transformations.

1m

2m3m

1m

Vector Matrix 3rd-order Tensor

2m

1m

1 2 3m m m X R1 2m mX R1mxR

Bilinear (or 2D) Subspace Learning: each image is represented as a 2nd-order tensor (i.e., a matrix)

Tensor Subspace Learning (more general case): each image is represented as a higher order tensor

Page 4: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

2

1

m

ij ik kjk

Y X U

100

100

100

100

100

100

.

.

.

100

10

100

10

100

10

100

10

100

10

100

10

=

=

=

.

.

.

Definition of Mode-k Product

(100)

2 )'(10m2 (100)m

2m(100)

1m(100)

1m

2 )'(10m

1m

2 (100)m

1(100)m

3 (40)m

2 )'(10m

2 (100)m

k U Y XNotation:

Product for two Matrices

Original Matrix

New Matrix

= 1(100)m

3 (40)m

2 )'(10m

Y XUProjection

Matrix

Original Tensor

New Tensor

Projection Matrix

Projection: high-dimensional space -> low-dimensional space

Reconstruction: low-dimensional space -> high-dimensional space

Page 5: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Definition of Mode-k Flattening

...1(100)m

2 3 (100*40)m m

Tensor Matrix

1m

2 (100)m

1(100)m

3 (40)m

Potential Assumption in Previous Tensor-based Subspace Learning:

Intra-tensor correlations: Correlations along column vectors of mode-k flattened matrices.

Page 6: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Data Representation in Dimensionality Reduction

Vector Matrix 3rd-order Tensor

High Dimension

Low Dimension

Examples

PCA, LDA Rank-1 Decomposition, 2001

A. Shashua and A. Levin,Tensorface, 2002

M. Vasilescu andD. Terzopoulos,

Our WorkXu et al., 2005

Yan et al., 2005…

...

...

...

Filtered Image Video SequenceGray-level Image

Page 7: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

What is Gabor Features?Gabor features can improve recognition performance in comparison to grayscale features. Chengjun Liu T-IP, 2002

Gabor Wavelet Kernels

Eight Orientations

Five

Scales

Input: Grayscale

Image Output: 40 Gabor-filtered

Images

Page 8: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Why Represent Image Objects as Tensors instead of Vectors?

Natural Representation Gray-level Images (2D structure) Videos (3D structure) Gabor-filtered Images (3D structure)

Enhance Learnability in Real Application Curse of Dimensionality (Gabor-filtered image: 100*100*40 -> Vector: 400,000)

Small sample size problem (less than 5,000 images in common face databases)

Reduce Computation Cost

... ...

Page 9: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Concurrent Subspace Analysis (CSA) as an Example(Criterion: Optimal Reconstruction)

1m

100

40

100

1U

2U

The reconstructed sample

Input sample

Projection Matrices?

Sample in Low- dimensional space

3U

1m

10

10

10

Dimensionality Reduction

1m

100

40

100

Reconstruction

D. Xu, S. Yan, Lei Zhang, H. Zhang et al., CVPR 2005 and T-CSVT 2008 3

1

* 31

21 1 1 3 3 3

|

( | )

arg min || ... ||k k

k k

i iiU

U

U U U U

X X

Objective Function:

Page 10: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Tensorization - New Research Direction: Other Related Works

• Discriminant Analysis with Tensor Representation (DATER): CVPR 2005 and T-IP 2007

• Coupled Kernel-based Subspace Learning (CKDA): CVPR 2005 • Rank-one Projections with Adaptive Margins (RPAM): CVPR 2006 and T-SMC-B

2007• Enhancing Tensor Subspace Learning by Element Rearrangement: CVPR 2007

and T-PAMI 2009• Discriminant Locally Linear Embedding with High Order Tensor Data (DLLE/T):

T-SMC-B 2008• Convergent 2D Subspace Learning with Null Space Analysis (NS2DLDA): T-

CSVT 2008• Semi-supervised Bilinear Subspace Learning: T-IP 2009• Applications in Human Gait Recognition

– CSA+DATER: T-CSVT 2006– Tensor Marginal Fisher Analysis (TMFA): T-IP 2007

Note: Other researchers also published several papers along this direction!!!

Page 11: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Tensorization - New Research Direction Tensorization in Graph Embedding Framework

Direct Graph Embedding

1minT

T

y B yy Ly

Original PCA & LDA,ISOMAP, LLE,

Laplacian Eigenmap

Linearization

PCA, LDA, LPP, MFA

wXy T

Kernelization

KPCA, KDA, KMFA

)( iii xw

Tensorization

CSA, DATER, TMFA

nnii wwwy 2

21

1X

Type

Formulation

Example

S. Yan, D. Xu, H. Zhang et al., CVPR, 2005 and T-PAMI,2007

Google Citation: 174 (until 15-Sep-2009)

Page 12: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Element Rearrangement: Motivations

• The success of tensor-based subspace learning relies on the redundancy among the unfolded vector

• However, such correlation/redundancy is usually not strong for real data

• Our Solution: Element rearrangement is employed as a preprocessing step to increase the intra-tensor correlations for existing tensor subspace learning methods

12

Page 13: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Motivations-Continued

Sets of highlycorrelated pixels

Columns of highlycorrelated pixels

Element Rearrangement

Low correlation High correlation

Intra-tensor correlations: Correlations among the features within certain tensor dimensions, such as rows, columns and Gabor features…

Page 14: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Problem Definition

• The task of enhancing correlation/redundancy among 2nd–order tensor is to search for a pixel rearrangement operator R, such that

14

* 2

,1

arg min{ min || || }N

R T R Ti iR U V

i

R X UU X VV

1. is the rearranged matrix from sample2. The column numbers of U and V are predefined

iXRiX

After the pixel rearrangement, we can use the rearranged tensors as input for tensor subspace learning algorithms!

Page 15: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Solution to Pixel Rearrangement Problem

15

Compute reconstructed matrices

1, 1 1 1 1

nRRec T Ti n n n i n nX U U X V V

Optimize operator R

2,

1

arg min || ||N

R Recn i i nR

i

R X X

Optimize U and V

n=n+1

Initialize U0, V0

2

,1

( , ) arg min || ||n n

NR RT T

n n i iU Vi

U V X UU X VV

2 2, 1 1 1 1

1 1

: || || || ||n n n

N NR R RRec T Ti i n i n n i n n

i i

Note X X X U U X V V

Page 16: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

• It is Integer Programming problem

Step for Optimizing R

16

* 2,

1

arg min || ||N

R Reci i nR

i

R X X

,

min .

1: 0 1; 2 : 1; 3 : 1

pq pqR

p q

pq pq pqp q

c R st

R R R

2,

1

| ( ) ( ) |N

Recpq i i n

i

where c X p X q

1. Linear programming problem in Earth Mover’s Distance (EMD) has integer solution.

2. We constrain the rearrangement within spatially local neighborhood or feature-based neighborhood for speedup.

p

Original matrix

Reconstructed matrix

q

pqc

Sender

Receiver

Page 17: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Convergence Speed

17

Page 18: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Rearrangement Results

18

Page 19: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Reconstruction Visualization

19

Page 20: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Reconstruction Visualization

20

Page 21: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Classification Accuracy

Page 22: Enhancing Tensor Subspace Learning by Element Rearrangement 1 Dong XU School of Computer Engineering Nanyang Technological University

Summary

.

• Our papers published in CVPR 2005 are the first works to address dimensionality reduction with the image objects represented as high-order tensors of arbitrary order.

• Our papers published in CVPR 2005 opens a new research direction. We also published a series of works along this direction.

• Element arrangement can further improve data compression performance and classification accuracy.