50
Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Embed Size (px)

Citation preview

Page 1: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Evolutionary Search

Artificial Intelligence

CMSC 25000

January 25, 2007

Page 2: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Agenda

• Motivation:– Evolving a solution

• Genetic Algorithms– Modelling search as evolution

• Mutation • Crossover• Survival of the fittest• Survival of the most diverse

• Conclusions

Page 3: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Motivation: Evolution

• Evolution through natural selection– Individuals pass on traits to offspring– Individuals have different traits– Fittest individuals survive to produce more

offspring– Over time, variation can accumulate

• Leading to new species

Page 4: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Simulated Evolution

• Evolving a solution• Begin with population of individuals

– Individuals = candidate solutions ~chromosomes

• Produce offspring with variation– Mutation: change features– Crossover: exchange features between individuals

• Apply natural selection– Select “best” individuals to go on to next generation

• Continue until satisfied with solution

Page 5: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Genetic Algorithms Applications

• Search parameter space for optimal assignment– Not guaranteed to find optimal, but can approach

• Classic optimization problems:– E.g. Travelling Salesman Problem

• Program design (“Genetic Programming”)

• Aircraft carrier landings

Page 6: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Genetic Algorithm Example• Cookie recipes (Winston, AI, 1993)

• As evolving populations

• Individual = batch of cookies• Quality: 0-9

– Chromosomes = 2 genes: 1 chromosome each • Flour Quantity, Sugar Quantity: 1-9

• Mutation: – Randomly select Flour/Sugar: +/- 1 [1-9]

• Crossover:– Split 2 chromosomes & rejoin; keeping both

Page 7: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Fitness

• Natural selection: Most fit survive

• Fitness= Probability of survival to next gen

• Question: How do we measure fitness?– “Standard method”: Relate fitness to quality

• :0-1; :1-9:

Chromosome Quality Fitness

1 43 11 21 1

4321

0.40.30.20.1

if iq j

jii qqf /

Page 8: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

GA Design Issues

• Genetic design: – Identify sets of features = genes; Constraints?

• Population: How many chromosomes?– Too few => inbreeding; Too many=>too slow

• Mutation: How frequent?– Too few=>slow change; Too many=> wild

• Crossover: Allowed? How selected?• Duplicates?

Page 9: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

GA Design: Basic Cookie GA• Genetic design:

– Identify sets of features: 2 genes: flour+sugar;1-9

• Population: How many chromosomes?– 1 initial, 4 max

• Mutation: How frequent?– 1 gene randomly selected, randomly mutated

• Crossover: Allowed? No

• Duplicates? No

• Survival: Standard method

Page 10: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Basic Cookie GA Results

• Results are for 1000 random trials– Initial state: 1 1-1, quality 1 chromosome

• On average, reaches max quality (9) in 16 generations

• Best: max quality in 8 generations

• Conclusion:– Low dimensionality search

• Successful even without crossover

Page 11: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Basic Cookie GA+Crossover Results

• Results are for 1000 random trials– Initial state: 1 1-1, quality 1 chromosome

• On average, reaches max quality (9) in 14 generations

• Conclusion:– Faster with crossover: combine good in each gene– Key: Global max achievable by maximizing each

dimension independently - reduce dimensionality

Page 12: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Solving the Moat Problem

• Problem:– No single step mutation

can reach optimal values using standard fitness (quality=0 => probability=0)

• Solution A:– Crossover can combine fit

parents in EACH gene

• However, still slow: 155 generations on average

1 2 3 4 5 4 3 2 123454321

0 0 0 0 0 0 0 20 0 0 0 0 0 0 30 0 7 8 7 0 0 40 0 8 9 8 0 0 50 0 7 8 7 0 0 40 0 0 0 0 0 0 30 0 0 0 0 0 0 22 3 4 5 4 3 2 1

Page 13: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Questions

• How can we avoid the 0 quality problem?

• How can we avoid local maxima?

Page 14: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Rethinking Fitness

• Goal: Explicit bias to best– Remove implicit biases based on quality

scale

• Solution: Rank method– Ignore actual quality values except for ranking

• Step 1: Rank candidates by quality• Step 2: Probability of selecting ith candidate, given

that i-1 candidate not selected, is constant p. – Step 2b: Last candidate is selected if no other has been

• Step 3: Select candidates using the probabilities

Page 15: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Rank Method

Chromosome Quality Rank Std. Fitness Rank Fitness

1 41 31 25 27 5

4 3 2 1 0

1 2345

0.4 0.30.20.10.0

0.667 0.2220.0740.0250.012

Results: Average over 1000 random runs on Moat problem- 75 Generations (vs 155 for standard method)

No 0 probability entries: Based on rank not absolute quality

Page 16: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Diversity

• Diversity: – Degree to which chromosomes exhibit

different genes– Rank & Standard methods look only at quality– Need diversity: escape local min, variety for

crossover– “As good to be different as to be fit”

Page 17: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Rank-Space Method

• Combines diversity and quality in fitness• Diversity measure:

– Sum of inverse squared distances in genes

• Diversity rank: Avoids inadvertent bias• Rank-space:

– Sort on sum of diversity AND quality ranks– Best: lower left: high diversity & quality

i id

1

Page 18: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Rank-Space Method

Chromosome Q D D Rank Q Rank Comb Rank R-S Fitness

1 43 11 21 17 5

4 3 2 1 0

1 5342

1 2345

0.667 0.0250.2220.0120.074

Diversity rank breaks tiesAfter select others, sum distances to bothResults: Average (Moat) 15 generations

0.04 0.25 0.059 0.062 0.05

1 4253

W.r.t. highest ranked 5-1

Page 19: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

GA’s and Local Maxima

• Quality metrics only:– Susceptible to local max problems

• Quality + Diversity:– Can populate all local maxima

• Including global max

– Key: Population must be large enough

Page 20: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

GA Discussion

• Similar to stochastic local beam search– Beam: Population size– Stochastic: selection & mutation– Local: Each generation from single previous– Key difference: Crossover – 2 sources!

• Why crossover?– Schema: Partial local subsolutions

• E.g. 2 halves of TSP tour

Page 21: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Question

• Traveling Salesman Problem– CSP-style Iterative refinement

– Genetic Algorithm

• N-Queens– CSP-style Iterative refinement

– Genetic Algorithm

Page 22: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Iterative Improvement Example

• TSP– Start with some valid tour

• E.g. find greedy solution

– Make incremental change to tour• E.g. hill-climbing - take change that produces

greatest improvement– Problem: Local minima– Solution: Randomize to search other parts of space

• Other methods: Simulated annealing, Genetic alg’s

Page 23: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Machine Learning:Nearest Neighbor &

Information Retrieval SearchArtificial Intelligence

CMSC 25000

January 25, 2007

Page 24: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Agenda

• Machine learning: Introduction• Nearest neighbor techniques

– Applications:• Credit rating• Text Classification

– K-nn– Issues:

• Distance, dimensions, & irrelevant attributes• Efficiency:

– k-d trees, parallelism

Page 25: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Machine Learning

• Learning: Acquiring a function, based on past inputs and values, from new inputs to values.

• Learn concepts, classifications, values– Identify regularities in data

Page 26: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Machine Learning Examples

• Pronunciation: – Spelling of word => sounds

• Speech recognition:– Acoustic signals => sentences

• Robot arm manipulation:– Target => torques

• Credit rating:– Financial data => loan qualification

Page 27: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Complexity & Generalization

• Goal: Predict values accurately on new inputs• Problem:

– Train on sample data– Can make arbitrarily complex model to fit– BUT, will probably perform badly on NEW data

• Strategy:– Limit complexity of model (e.g. degree of equ’n)– Split training and validation sets

• Hold out data to check for overfitting

Page 28: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor

• Memory- or case- based learning

• Supervised method: Training– Record labeled instances and feature-value vectors

• For each new, unlabeled instance– Identify “nearest” labeled instance– Assign same label

• Consistency heuristic: Assume that a property is the same as that of the nearest reference case.

Page 29: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor Example

• Credit Rating:– Classifier: Good /

Poor– Features:

• L = # late payments/yr; • R = Income/Expenses

Name L R G/P

A 0 1.2 G

B 25 0.4 P

C 5 0.7 G

D 20 0.8 PE 30 0.85 P

F 11 1.2 G

G 7 1.15 GH 15 0.8 P

Page 30: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor Example

Name L R G/P

A 0 1.2 G

B 25 0.4 P

C 5 0.7 G

D 20 0.8 PE 30 0.85 P

F 11 1.2 G

G 7 1.15 GH 15 0.8 P L

R

302010

1 A

B

C D E

FG

H

Page 31: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor Example

L 302010

1 A

B

C D E

FG

HR

Name L R G/P

I 6 1.15

J 22 0.45

K 15 1.2

G

IP

J

??

K

Distance Measure:

Sqrt ((L1-L2)^2 + [sqrt(10)*(R1-R2)]^2))

- Scaled distance

Page 32: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor Analysis

• Problem: – Ambiguous labeling, Training Noise

• Solution:– K-nearest neighbors

• Not just single nearest instance

• Compare to K nearest neighbors– Label according to majority of K

• What should K be?– Often 3, can train as well

Page 33: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Text Classification

Page 34: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Matching Topics and Documents

• Two main perspectives:– Pre-defined, fixed, finite topics:

• “Text Classification”

– Arbitrary topics, typically defined by statement of information need (aka query)

• “Information Retrieval”

Page 35: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Vector Space Information Retrieval

• Task:– Document collection– Query specifies information need: free text– Relevance judgments: 0/1 for all docs

• Word evidence: Bag of words– No ordering information

Page 36: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Vector Space Model

Computer

Tv

Program

Two documents: computer program, tv programQuery: computer program : matches 1 st doc: exact: distance=2 vs 0 educational program: matches both equally: distance=1

Page 37: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Vector Space Model

• Represent documents and queries as– Vectors of term-based features

• Features: tied to occurrence of terms in collection

– E.g.

• Solution 1: Binary features: t=1 if present, 0 otherwise– Similiarity: number of terms in common

• Dot product

),...,,();,...,,( ,,2,1,,2,1 kNkkkjNjjj tttqtttd

ji

N

ikijk ttdqsim ,

1,),(

Page 38: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Vector Space Model II

• Problem: Not all terms equally interesting– E.g. the vs dog vs Levow

• Solution: Replace binary term features with weights– Document collection: term-by-document matrix

– View as vector in multidimensional space• Nearby vectors are related

– Normalize for vector length

),...,,();,...,,( ,,2,1,,2,1 kNkkkjNjjj wwwqwwwd

Page 39: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Vector Similarity Computation

• Similarity = Dot product

• Normalization:– Normalize weights in advance– Normalize post-hoc

ji

N

ikijkjk wwdqdqsim ,

1,),(

N

i ji

N

i ki

N

i jikijk

ww

wwdqsim

1

2,1

2,

1 ,,),(

Page 40: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Term Weighting

• “Aboutness”– To what degree is this term what document is about?– Within document measure– Term frequency (tf): # occurrences of t in doc j

• “Specificity”– How surprised are you to see this term?– Collection frequency– Inverse document frequency (idf):

)log(i

i n

Nidf

ijiji idftfw ,,

Page 41: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Term Selection & Formation

• Selection:– Some terms are truly useless

• Too frequent, no content– E.g. the, a, and,…

– Stop words: ignore such terms altogether

• Creation:– Too many surface forms for same concepts

• E.g. inflections of words: verb conjugations, plural

– Stem terms: treat all forms as same underlying

Page 42: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Efficient Implementations

• Classification cost:– Find nearest neighbor: O(n)

• Compute distance between unknown and all instances

• Compare distances

– Problematic for large data sets

• Alternative:– Use binary search to reduce to O(log n)

Page 43: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Efficient Implementation: K-D Trees

• Divide instances into sets based on features– Binary branching: E.g. > value– 2^d leaves with d split path = n

• d= O(log n)

– To split cases into sets,• If there is one element in the set, stop• Otherwise pick a feature to split on

– Find average position of two middle objects on that dimension

» Split remaining objects based on average position

» Recursively split subsets

Page 44: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

K-D Trees: Classification

R > 0.825?

L > 17.5? L > 9 ?

No Yes

R > 0.6? R > 0.75? R > 1.025 ?R > 1.175 ?

NoYes No Yes

No

Poor Good

Yes No Yes

Good Poor

No Yes

Good Good

No

Poor

Yes

Good

Page 45: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Efficient Implementation:Parallel Hardware

• Classification cost:– # distance computations

• Const time if O(n) processors

– Cost of finding closest• Compute pairwise minimum, successively• O(log n) time

Page 46: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor: Issues

• Prediction can be expensive if many features

• Affected by classification, feature noise– One entry can change prediction

• Definition of distance metric– How to combine different features

• Different types, ranges of values

• Sensitive to feature selection

Page 47: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor: Analysis

• Issue: – What is a good distance metric?– How should features be combined?

• Strategy:– (Typically weighted) Euclidean distance– Feature scaling: Normalization

• Good starting point: – (Feature - Feature_mean)/Feature_standard_deviation– Rescales all values - Centered on 0 with std_dev 1

Page 48: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor: Analysis

• Issue: – What features should we use?

• E.g. Credit rating: Many possible features– Tax bracket, debt burden, retirement savings, etc..

– Nearest neighbor uses ALL – Irrelevant feature(s) could mislead

• Fundamental problem with nearest neighbor

Page 49: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Nearest Neighbor: Advantages

• Fast training:– Just record feature vector - output value set

• Can model wide variety of functions– Complex decision boundaries– Weak inductive bias

• Very generally applicable

Page 50: Evolutionary Search Artificial Intelligence CMSC 25000 January 25, 2007

Summary

• Machine learning:– Acquire function from input features to value

• Based on prior training instances

– Supervised vs Unsupervised learning• Classification and Regression

– Inductive bias: • Representation of function to learn• Complexity, Generalization, & Validation