46
Chapter 2: Concept Learning and the General-to- Specific Ordering

Chapter 2: Concept Learning and the General-to-Specific Ordering

  • Upload
    kendis

  • View
    34

  • Download
    1

Embed Size (px)

DESCRIPTION

Chapter 2: Concept Learning and the General-to-Specific Ordering. Concept of Concepts. Examples of Concepts “birds”, “car”, “situations in which I should study more in order to pass the exam” Concept Some subset of objects or events defined over a larger set, or - PowerPoint PPT Presentation

Citation preview

Page 1: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Chapter 2: Concept Learning and the General-to-Specific Ordering

Page 2: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Concept of Concepts

Examples of Concepts “birds”, “car”, “situations in which I should study

more in order to pass the exam” Concept

Some subset of objects or events defined over a larger set, or

A boolean-valued function defined over this larger set.

Concept “birds” is the subset of animals that constitute birds.

Page 3: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Concept Learning

Learning Inducing general functions from specific

training examples Concept learning

Acquiring the definition of a general category given a sample of positive and negative training examples of the category

Inferring a boolean-valued function from training examples of its input and output.

Page 4: Chapter 2: Concept Learning and  the General-to-Specific Ordering

A Concept Learning Task

Target concept EnjoySport “days on which Aldo enjoys water sport”

Hypothesis A vector of 6 constraints, specifying the values of

the six attributes Sky, AirTemp, Humidity, Wind, Water, and Forecast.

For each attribute the hypo will either “?”, single value (e.g. Warm) , or “0”

<?, Cold, High, ?, ?, ?> expresses the hypothesis that Aldo enjoys his favorite sport only on cold days with high humidity.

Page 5: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Training Examples for EnjoySport

Instance Sky AirTemp Humidity Wind Water Forecast EnjoySport

A Sunny Warm Normal Strong Warm Same No B Sunny Warm High Strong Warm Same Yes C Rainy Cold High Strong Warm Change No D Sunny Warm High Strong Cool Change Yes

Training examples for the target concept EnjoySport

Page 6: Chapter 2: Concept Learning and  the General-to-Specific Ordering

The Learning Task

Given: Instances X: set of items over which the concept is defined. Hypotheses H: conjunction of constraints on attributes. Target concept c: c : X → {0, 1} Training examples (positive/negative): <x,c(x)> Training set D: available training examples

Determine: A hypothesis h in H such that h(x) = c(x), for all x in X

Page 7: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Learning Hypothesis

Learning task is to determine h identical to c over the entire set of instances X.

But the only information about c is its value over D. Inductive learning algorithms can at best guarantee that

the induced h fits c over D. Assumption is that the best h regarding unseen instances is

the h that best fits the observed data in D.

Inductive learning hypothesis Any good hypothesis over a sufficiently large set of

training examples will also approximate the target function. well over unseen examples.

Page 8: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Concept Learning as Search

Search Find a hypothesis that best fits training examples Efficient search in hypothesis space (finite/infinite)

Search space in EnjoySport 3*2*2*2*2*2 = 96 distinct instances (eg. Sky={Sunny,

Cloudy, Rainy} 5*4*4*4*4*4 = 5120 syntactically distinct hypotheses within

H (considering 0 and ? in addition) 1+4*3*3*3*3*3 = 973 semantically distinct hypotheses (count

just one 0 for each attribute since every hypo having one or more 0 symbols is empty)

Page 9: Chapter 2: Concept Learning and  the General-to-Specific Ordering

General-to-Specific Ordering

General-to-specific ordering of hypotheses: x satisfies h ⇔ h(x)=1 More_general_than_or_equal_to relation

(Strictly) more_general_than relation

<Sunny,?,?,?,?,?> >g <Sunny,?,?,Strong,?,?>

)]1)(()1)()[(( xhxhXxhh jkkgj

)()( jgkkgjkgj hhhhhh

Page 10: Chapter 2: Concept Learning and  the General-to-Specific Ordering

More_General_Than Relation

Page 11: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Find-SFind-S: Finding a Maximally Specific Hypothes

is

1. Initialize h to the most specific hypothesis in H

2. For each positive training example x

For each attribute constraint ai in h

If the constraint ai is satisfied by x

Then do nothing

Else replace ai in h by the next more general

constraint satisfied by x

3. Output hypothesis h

Page 12: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Hypothesis Space Search by Find-SFind-S

Page 13: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Properties of Find-SFind-S

Ignores every negative example (no revision to h required in response to negative examples). Why? What’re the assumptions for this?

Guaranteed to output the most specific hypothesis consistent with the positive training examples (for conjunctive hypothesis space).

Final h also consistent with negative examples provided the target c is in H and no error in D.

Page 14: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Weaknesses of Find-S

Has the learner converged to the correct target concept? No way to know whether the solution is unique.

Why prefer the most specific hypothesis? How about the most general hypothesis?

Are the training examples consistent? Training sets containing errors or noise can severely mislead the algorithm Find-S.

What if there are several maximally specific consistent hypotheses? No backtrack to explore a different branch of partial ordering.

Page 15: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Version Spaces (VSs)

Output all hypotheses consistent with the training examples.

Version space Consistent(h,D) ⇔ ( <∀ x,c(x)> D) h(x) = c(x)

VSH,D ⇔ {h H | Consistent(h,D)} List-Then-Eliminate Algorithm

Lists all hypotheses, then removes inconsistent ones.

Applicable to finite H

Page 16: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Compact Representation of VSs

More compact representation for version spaces General boundary G

Specific boundary S

Version Space redefined with S and G

)]},'()')['(),(|{ DgConsistentggHgDgConsistentHgG g

)]},'()')[('(),(|{ DsConsistentssHsDsConsistentHsS g

)})()((|{, shgGgSsHhVS ggDH

Page 17: Chapter 2: Concept Learning and  the General-to-Specific Ordering

A Version Space with S and G Boundaries

Page 18: Chapter 2: Concept Learning and  the General-to-Specific Ordering

CE: Candidate-Elimination Algorithm

Initialize G to the set of maximally general hypotheses in H

Initialize S to the set of maximally specific hypotheses in H

For each training example d, do

If d is a positive example

Remove from G any hypothesis inconsistent with d

For each hypothesis s in S that is not consistent with d

Remove s from S

Add to S all minimal generalizations h of s such that

h is consistent with d, and some member of G is more general

than h

Remove from S any hypothesis that is more general than another

hypothesis in S

Page 19: Chapter 2: Concept Learning and  the General-to-Specific Ordering

If d is a negative example

Remove from S any hypothesis inconsistent with d

For each hypothesis g in G that is not consistent with d

Remove g from G

Add to G all minimal specializations h of g such that

h is consistent with d, and some member of S is more specific than h

Remove from G any hypothesis that is less general than another

hypothesis in G

Candidate-Elimination Algorithm

Page 20: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Given First Two Examples

Page 21: Chapter 2: Concept Learning and  the General-to-Specific Ordering

After the Third Example

Page 22: Chapter 2: Concept Learning and  the General-to-Specific Ordering

After the Fourth Example

Page 23: Chapter 2: Concept Learning and  the General-to-Specific Ordering

The Concept Learned

Page 24: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Remarks on Candidate Elimination

Will the CE algorithm converge to the correct hypothesis?

What training example should the learner request next?

How can partially learned concepts be used?

Page 25: Chapter 2: Concept Learning and  the General-to-Specific Ordering

When Does CE Converge?

Will the Candidate-Elimination algorithm converge to the correct hypothesis?

Prerequisites 1. No error in training examples

2. The target hypothesis exists which correctly

describes c(x). If S and G boundary sets converge to an empty

set, this means there is no hypothesis in H consistent with observed examples.

Page 26: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Who Provides Examples?

What training example should the learner request next? Two methods

Fully supervised learning: External teacher provides all training examples (input + correct output)

Learning by query: The learner generates instances (queries) by conducting experiments, then obtains the correct classification for this instance from an external oracle (nature or a teacher).

Negative training examples specializes G, positive ones generalize S.

Page 27: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Optimal Query Strategies

What would be a good query? The learner should attempt to discriminate among alternative competing hypotheses in its current version space.

A good query is the one that is classified positive by some of these hypos, but negative by others.

In general, the optimal query strategy for a concept learner is to generate instances that satisfy exactly half the hypos in the current version space.

Experiments needed to find the correct target concept:

VS2log

Page 28: Chapter 2: Concept Learning and  the General-to-Specific Ordering

How to Use Partially Learned Concepts?

Suppose the learner is asked to classify the four new instances shown in the following table.

Instance Sky AirTemp Humidity Wind Water Forecast EnjoySport

A Sunny Warm Normal Strong Cool Change ? B Rainy Cold Normal Light Warm Same ? C Sunny Warm Normal Light Warm Same ? D Sunny Cold Normal Strong Warm Same ?

A: classified as positive by all hypos in the current version space (Fig. 2.3)

B: classified as negative by all hypos

C: 3 positive, 3 negative

D: 2 positive, 4 negative (can be decided by majority vote, for example)

Page 29: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Partially Learned VS Revisited

Page 30: Chapter 2: Concept Learning and  the General-to-Specific Ordering

CE will converge toward the target concept provided that it is contained in its initial hypo space and training examples contain no errors.

What if the target concept is not contained in the hypo space?

One solution: Use a hypothesis space that includes every possible hypothesis (more expressive hypo space).

New problem: Generalize poorly or do not generalize at all.

Fundamental Questions for Inductive

Inference

Page 31: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias

EnjoySport: H contains only conjunctions of attribute values.

This H is unable to represent even simple disjunctive target concepts such as <Sunny,?,?,?,?,?> ∨ <Cloudy,?,?,?,?,?>

Given the following three training examples of this disjunctive hypothesis, CE would find that there are zero hypo in VS.

Page 32: Chapter 2: Concept Learning and  the General-to-Specific Ordering

The problem is that we have biased the learner to consider only conjunctive hypotheses. We require more expressive hypothesis space.

-

Example Sky AirTemp Humidity Wind Water Forecast EnjoySport

1 Sunny Warm Normal Strong Cool Change Yes 2 Cloudy Warm Normal Strong Cool Change Yes 3 Rainy Warm Normal Strong Cool Change No

A Biased Hypothesis Space

Page 33: Chapter 2: Concept Learning and  the General-to-Specific Ordering

An Unbiased Learner One solution: Provide H contains every teachable concept

(every possible subset of instances X). Power set of X: set of all subsets of a set X

EnjoySport: |X| = 96

Size of the power set: 2|X| = 296 = 1028 (the number of distinct target concepts)

In contrast, our conjunctive H contains only 973 (semantically distinct) of these.

New problem: unable to generalize beyond the observed examples. Observed examples are only unambiguously classified. Voting results in no majority or minority.

Page 34: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Futility of Bias-Free Learning

Fundamental property of inductive inference:

“A learner that makes no a priori assumptions regarding the identity of the target concept has no rational basis for classifying any unseen instances.”

Page 35: Chapter 2: Concept Learning and  the General-to-Specific Ordering

L: an arbitrary learning algorithm c: some arbitrary target concept Dc = { <x, c(x)> }: an arbitrary set of training

data L(xi, Dc): classification that L assigns to xi after

learning Dc. Inductive inference step performed by L:

(Dc ^ xi) I> L(xi, Dc)

Inductive Inference

Page 36: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias Formally Defined

Because L is an inductive learning algorithm, the result L(xi, Dc): will not in general provably correct; L need not follow deductively from Dc and xi.

However, additional assumptions can be added to Dc ^ xi so that L(xi, Dc) would follow deductively.

Definition: The inductive bias of L is any minimal set of assertions B (assumptions, background knowledge etc.) such that for any target concept c and corresponding training examples Dc

)],())[(( ciici DxLxDBXx f

Page 37: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias of CE Algorithm

Given the assumption c ∈ H, the inductive inference performed by the CE algorithm can be justified deductively. Why? If we assume c ∈ H, it follows deductively that c

∈ VSH,Dc.

Since we defined L(xi, Dc) to be unanimous vote of all hypos in VS, if L outputs the classification L(xi, Dc), it must be the case the every hypo in L(xi, Dc) also produces this classification, including the hypo c ∈ VSH,Dc.

Inductive bias of CE: The target concept c is contained in the given hypothesis space H.

Page 38: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive & Deductive Systems

Page 39: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Strength of Inductive Biases

(1) Rote-Learner: weakest (no bias)

(2) Candidate-Elimination Algorithm

(3) Find-S: strongest bias of the three

Page 40: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias of Rote-Learner

Simply stores each observed training example in memory.

New instances are classified by looking them up in memory: If it is found in memory, the stored classification is returned. Otherwise, the system refuses to classify the new instance.

No inductive bias: The classifications for new instances follow deductively from D with no additional assumptions required.

Page 41: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias of Cand.-Ellim.

New instances are classified only if all hypos in VS agree. Otherwise, it refuses to classify.

Inductive bias: The target concept can be represented in its hypothesis space.

This inductive bias is stronger than that of rote-learner since CE will classify some instances that the rote-learner will not.

Page 42: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Inductive Bias of Find-S

Find the most specific hypo consistent with D and uses this hypo to classify new instances.

Even stronger inductive biase The target concept can be described in its hypo spa

ce. All instances are negative unless opposite is entaile

d by its other knowledge (default reasoning)

Page 43: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Summary (1/3)

Concept learning can be cast as a problem of searching through a large predefined space of potential hypotheses.

General-to-specific partial ordering of hypotheses provides a useful structure for search.

Find-S algorithm performs specific-to-general search to find the most specific hypothesis.

Page 44: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Candidate-Elimination algorithm computes version space by incrementally computing the sets of maximally specific (S) and maximally general (G) hypotheses.

S and G delimit the entire set of hypotheses consistent with the data.

Version spaces and Candidate-Elimination algorithm provide a useful conceptual framework for studying concept learning.

Summary (2/3)

Page 45: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Summary (3/3)

Candidate-Elimination algorithm is not robust to noisy data or to situations where the unknown target concept is not expressible in the provided hypothesis space.

Inductive bias in Candidate-Elimination algorithm is that target concept exists in H

If the hypothesis space is enriched to the point where there is every possible hypothesis (the power set of instances), then this will remove the inductive bias of CE and thus remove the ability to classify any instance beyond the observed examples.

Page 46: Chapter 2: Concept Learning and  the General-to-Specific Ordering

Home Work

Exercise 2.1 Exercise 2.5