24
Minshan Cui, Saurabh Prasad, Majid Mahroogy, Lori Mann Bruce, James Aanstoos GENETIC ALGORITHMS AND LINEAR DISCRIMINANT ANALYSIS BASED DIMENSIONALITY REDUCTION FOR REMOTELY SENSED IMAGE ANALYSIS

Minshan Cui, Saurabh Prasad, Majid Mahroogy , Lori Mann Bruce, James Aanstoos

  • Upload
    shika

  • View
    38

  • Download
    0

Embed Size (px)

DESCRIPTION

Minshan Cui, Saurabh Prasad, Majid Mahroogy , Lori Mann Bruce, James Aanstoos. Genetic Algorithms and Linear Discriminant Analysis based Dimensionality Reduction for Remotely Sensed Image Analysis. Stepwise LDA (S-LDA), (DAFE) - PowerPoint PPT Presentation

Citation preview

Page 1: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Minshan Cui, Saurabh Prasad, Majid Mahroogy, Lori Mann Bruce, James Aanstoos

GENETIC ALGORITHMS AND LINEAR DISCRIMINANT ANALYSIS BASED DIMENSIONALITY REDUCTION FOR

REMOTELY SENSED IMAGE ANALYSIS

Page 2: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Stepwise LDA (S-LDA), (DAFE)• A preliminary forward selection and backward rejection is

employed to discard less relevant features.• A Linear Discriminant Analysis (LDA) projection is

applied on this reduced subset of features to further reduce the dimensionality of the feature space.

Drawbacks• In forward selection, one is unable to reevaluate the

features that become irrelevant after adding some other features.

• In backward rejection, one is unable to reevaluate the features after they have been discarded.

Traditional Approaches(Stepwise Selection, Greedy Search, …)

Page 3: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Genetic algorithms are a class of optimization techniques that search for the global minimum of a fitness function.

This typically involves four steps – evaluation, reproduction, recombination, and mutation.

Genetic Algorithm

Page 4: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Population

Genetic Algorithm(Select 4 bands out of 10 bands)

Rank

Fitness Functio

n

Fitness Value

Reproduction

Crossover

Mutation

Next Generation

Repeating this process until one of stopping criteria is met

2

3

1

4

5

2 10 7 5

9 7 8 3

5 1 2 8

10 4 3 6

2 5 9 4

5 1 2 8

2 10 7 5

9 7 8 3

10 4 3 62 10 7 5 9 7 8 3

2 5 9 4 2 10 7 5

2 10 8 3 2 10 9 4 10 4 7 59 5 8 3

2 10 8 3

2 10 9 4

10 4 7 5

9 5 8 3

Page 5: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Using genetic algorithm with BD or Fisher’s ratio as a fitness function to select most relevant features in a dataset.

• Bhattacharyya distance (BD)

• Fisher’s ratio

Applying linear discriminant analysis on the selected features to further extract features.

Genetic Algorithm based Linear Discriminant analysis

||||

|2

|ln

21)(

1

2)(

81

21

21

1221

12

TBD

wSwwSwwJ

WT

BT

)(

Page 6: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Genetic Algorithm based Linear Discriminant analysis

Genetic Algorithm

Fitness Function

Bhattacharyya Distance

Fisher’s Ratio

or

OriginalFeatures

Linear Discriminant Analysis

ExtractedFeatures

SelectedFeatures

Page 7: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Hyperspectral Imagery (HSI)

Experimental Hyperspectral Dataset

• Using NASA’s AVIRIS sensor

• 145x145 pixels and 220 bands in the 400 to 2450 nm region of the visible and infrared spectrum.

400 600 800 1000 1200 1400 1600 1800 2000 2200 24001000

2000

3000

4000

5000

6000

7000

8000

Wavelength in micrometer

Ref

lect

ance

Signatures of AVIRIS Indian Pines

corn notillcorn mingrass pasturehay windrowedsoybeans notillsoybeans minsoybeans cleanwoods

Figure 1: A plot of reflectance versus wavelength for eight classes of spectral signatures from AVIRIS Indian Pines data.

• Ground truth of HSI data

20 40 60 80 100 120 140

20

40

60

80

100

120

14020 40 60 80 100 120 140

20

40

60

80

100

120

14020 40 60 80 100 120 140

20

40

60

80

100

120

140 20 40 60 80 100 120 140

20

40

60

80

100

120

140

• Feature layers

20 40 60 80 100 120 140

20

40

60

80

100

120

140

Page 8: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Untreated Check3 days after spray

0.43 kg ae/ha

21 days after spray

3 days after spray

21 days after spray

Check

0.11 kg ae/ha

0.43 kg ae/ha0.22 kg ae/ha

0.05 kg ae/ha0.03 kg ae/ha0.02 kg ae/ha0.01 kg ae/ha

Experimental Hyperspectral Dataset

Page 9: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Experimental Hyperspectral Dataset

Page 10: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Synthetic Aperture Radar (SAR)

Experimental Synthetic Aperture Radar Dataset

Parameter Value

Frequency L-band

Bandwidth 80 MHz

Range Resolution 1.8 m

Polarization Full quad polarization

Quantization 12 bits

Antenna size 0.5 m range/1.5 azimuth

Power > 2.0 kW

Table 1: Illustrating some salient characteristics of UAVSAR

10 20 30 40 50 60 70 80 90

10

20

30

40

50

60

70

8010 20 30 40 50 60 70 80 90

10

20

30

40

50

60

70

8010 20 30 40 50 60 70 80 90

10

20

30

40

50

60

70

8010 20 30 40 50 60 70 80 90

10

20

30

40

50

60

70

80

• Feature layers via GLCM

• From NASA Jet Propulsion Laboratory’s Unmanned Aerial Vehicle Synthetic Aperture Radar (UAVSAR)

• Two classes – healthy levees and levees with landslides on them

• Ground truth of SAR data

• Breached Levee

Page 11: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

HSI and SAR analysis using:• LDA

• Stepwise LDA (S-LDA)

• GA-LDA-Fisher (Using Fisher’s ratio as a fitness function in GA.)

• GA-LDA-BD(Using Bhattacharyya distance as a fitness function in GA.)

Performance measures:• Overall recognition accuracies

Experiments

Page 12: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

HSI Data Experimental Results

0 20 40 60 80 100 120 140 160 180 20030

35

40

45

50

55

60

65

70

75

80(HSI Data) Number of features = 50

Number of training samples/class

Ove

rall

Acc

urac

y(%

)

GA-LDA-FisherGA-LDA-BDSLDALDA

0 20 40 60 80 100 120 140 160 180 20030

35

40

45

50

55

60

65

70

75

80(HSI Data) Number of features = 25

Number of training samples/class

Ove

rall

Acc

urac

y(%

)

GA-LDA-FisherGA-LDA-BDSLDALDA

Page 13: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

SAR Data Experimental Results

0 20 40 60 80 100 120 140 160 180 20050

55

60

65

70

75

80

85

90

95

100(SAR Data) Number of features = 10

Number of training samples/class

Ove

rall

Acc

urac

y(%

)

GA-LDA-FisherGA-LDA-BDSLDALDA

0 20 40 60 80 100 120 140 160 180 20050

55

60

65

70

75

80

85

90

95

100(SAR Data) Number of features = 15

Number of training samples/class

Ove

rall

Acc

urac

y(%

)

GA-LDA-FisherGA-LDA-BDSLDALDA

Page 14: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

GA search is very effective at selecting the most pertinent features.

Given a moderate feature space dimensionality and sufficient training samples, LDA is a good projection based dimensionality reduction strategy.

As the number of features increases and the training-sample-size decreases, methods such as GA-LDA can assist by providing a robust intermediate step of pruning away redundant and less useful features.

Conclusions

Page 15: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

[1] Ho-Duck Kim, Chang-Hyun Park, Hyun-Chang Yang, Kwee-Bo Sim, “Genetic Algorithm Based Feature Selection Method Development for Pattern Recognition,” in SICE-ICASE, 2006.[2] Chulhee Lee and Daesik Hong, “Feature Extraction Using the Bhattacharyya Distance,” in IEEE International on Systems, Man, and Cybernetics, 1997[3] Tran Huy Dat, Cuntai Guan, “Feature selection based on fisher ratio and mutual information analysis for robust brain computer interface,” in IEEE International Conference on Acoustics, Speech and Signal Processing, 2007.[4] R.O. Duda, P.E. Stark, D.G. Stork, Pattern Classification, Wiley Inter-science, October 2000.[5] S. Prasad and L. M. Bruce, “Limitations of Principal Components Analysis for Hyperspectral Target Recognition,” in IEEE Geoscience and Remote Sensing Letters, vol. 5, pp. 625-629, 2008.[6] S. Kumar, J. Ghosh, M.M. Crawford, “Best-bases feature extraction algorithms for classification of hyperspectral data,” in IEEE Transactions on Geoscience and Remote Sensing, Vol. 39, No. 7, pp 1368-1379, July 2001.[7] Nakariyakul, S. ,Casasent, D.P., “Improved forward floating selection algorithm for feature subset selection,” in Proceedings of the 2008 International Conference on Wavelet Analysis and Pattern Recognition, HongKong, 30-31 Aug. 2008.[8] K.S. Tang, K.F. Man, S. Kwong, Q. He, “Genetic algorithms and their applications,” in IEEE Signal Processing Magazine, Vol. 13, Nov 1996.[9] NASA Jet Propulsion Laboratory Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) web page http://aviris.jpl.nasa.gov/[10] Kevin Wheeler, Scott Hensley, Yunling Lou, Tim Miller, Jim Hoffman, "An L-band SAR for repeat pass deformation measurements on a UAV platform", 2004 IEEE Radar Conference, Philadelphia, PA, April 2004. (classifying health levee from landslide in a UAVSAR image).

References

Page 16: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Thank You

Questions - Comments - Suggestions

Minshan [email protected]

Page 17: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Elite countSpecifies the number of individuals that are guaranteed to survive to the next

generation.

Crossover fraction Specifies the fraction of the next generation, other than elite children, that are

produced by crossover.

Ex) Assume population size =10, elite count = 2 and crossover fraction = 0.8

• nEliteKids = 2• nCrossoverKids = round(CrossoverFraction × (10 - nEliteKids))

= round( 0.8 × (10 – 2)) = 6• nMutateKids = 10 - nEliteKids - nCrossoverKids

= 10 - 2 - 6 = 2

How many elite, crossover and mutate kids will be produced in next generation?

Page 18: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Since 2 parents produce 1 crossover kid and 1 parent produce 1 mutate kid, GA will need:

nParents = 2 × nCrossoverKids + nMutateKids = 2 × 6 + 2 = 14.

GA will need 12 parents to produce 6 crossover kids and 2 parents to produce 2 mutate kids.

How many parents GA need to produce crossover kids and mutate kids?

Page 19: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Feature number: 1 2 3 4 5 6 7 8 9 10

Fitness Value: 3.68 17.24 21.46 - 9.26 7.59 7.92 104.22 6.47 13.25 12.22

1. The space of each individual having is proportional to its fitness (or rank). 2. Place 14 equally spaced arrows in this line.3. Individuals with arrows placed will be selected as parents.

Selected parents = 1 1 2 3 4 4 4 5 6 7 8 8 9 10

How to select individuals to be parents?

1 2 3 4 5 6 7 8 9 10

Page 20: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

First we need to randomize the selected parents.parents = 3 4 4 2 1 5 10 4 1 8 6 9 6 8

12 parents selected to produce 6 crossover kids.parents = [3 4] [4 2] [1 5] [10 4] [1 8] [6 9]

2 parents selected to produce 2 mutate kids.parents = [6] [8]

How to produce crossover and mutate kids?

Page 21: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Single point Two Parents ( Individual 3 & 4 ) 193.2 19.736 215.74 129.85 55.08 142.92 183.95 11.855 98.849 155.64 103.73 15.757 142.21 95.922 2.9524 95.908 184.02 63.021 179.7 183.07Crossover kid

193.2 19.736 215.74 129.85 55.08 95.908 184.02 63.021 179.7 183.07

ScatteredRandomly produce binary strings. 1 means change, 0 means reserve. 1 0 0 1 1 0 1 0 0 1Two Parents ( Individual 3 & 4 ) 193.2 19.736 215.74 129.85 55.08 142.92 183.95 11.855 98.849 155.64 103.73 15.757 142.21 95.922 2.9524 95.908 184.02 63.021 179.7 183.07Crossover kid

103.73 19.736 215.74 95.922 2.9524 142.92 184.02 11.855 183.07 98.849

Crossover

Page 22: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

Mutation-gaussianAdds a random number taken from a Gaussian distribution with mean 0 to each

entry of the parent vector. Parent

103.17 192.15 51.61 210.8 78.211 188.76 177.62 125.36 198.09 140.78Mutate kid169.97 211.92 4.7027 82.935 172.73 23.559 123.35 32.11 158.63 214.26

Mutation-uniform 1. First, the algorithm selects a fraction of the vector entries of an individual for

mutation, where each entry has a probability Rate of being mutated. 2. In the second step, the algorithm replaces each selected entry by a random number

selected uniformly from the range for that entry.Parent

103.17 192.15 51.61 210.8 78.211 188.76 177.62 125.36 198.09 140.78Mutate kid

103.17 213.45 51.61 210.8 45.231 188.76 177.62 97.56 198.09 140.78

Mutatation

Page 23: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

nextGeneration = [ eliteKids, crossoverKids, mutateKids ]

Next generation

Page 24: Minshan Cui, Saurabh Prasad, Majid  Mahroogy ,  Lori Mann Bruce, James Aanstoos

1. Generations — Specifies the maximum number of iterations for the genetic algorithm to perform. The default is 100.

2. Time limit — Specifies the maximum time in seconds the genetic algorithm runs before stopping.

3. Fitness limit — The algorithm stops if the best fitness value is less than or equal to the value of Fitness limit.

4. Stall generations — The algorithm stops if the weighted average change in the fitness function value over Stall generations is less than Function tolerance.

5. Stall time limit — The algorithm stops if there is no improvement in the best fitness value for an interval of time in seconds specified by Stall time.

6. Function tolerance — The algorithm runs until the cumulative change in the fitness function value over Stall generations is less than or equal to Function Tolerance.

Stopping criteria