30
Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science Department Carnegie Mellon University

Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

  • View
    214

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Optimal Nonmyopic Value of Information in Graphical Models

Efficient Algorithms and Theoretical Limits

Andreas Krause, Carlos Guestrin

Computer Science DepartmentCarnegie Mellon University

Page 2: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Related applications

Medical expert systems

select among potential examinations Sensor scheduling

observations drain power, require storage Active learning, experimental design ...

Page 3: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Y1 Y2 Y3 Y4 Y5

Part-of-Speech Tagging

Andreas is giving a talk

X1 X2 X3 X4 X5

S P O S P O S P O S P O S P OS P O S P O S P O S P O S P O

Y3=P

observe

S P O S P O S P O S P O S P O

Y2=PY2=O

observe

S P O S P O S P O S P O S P O

reward: 0.8reward: 0.9reward: 0.3

Classify each word to belong to subject, predicate, object

Classification must respect sentence structureValues:

(S)ubject,

(P)redicate,

(O)bjectAsk expert k most informative questionsNeed to compute expected reward for any selection!

What does “most informative” mean?Which reward function should we use?

Our probabilistic model providescertain a priori classification accuracy.

What if we could ask an expert?

Page 4: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Reward functions

Depend on probability distributions: E[ R(X | O) ] := o P(o) R( P(X | O = o) )

In classification / prediction setting, rewards measure reduction of uncertainty Margin to runner-up:

confidence in most likely assignment Information gain:

uncertainty about hidden variables

In decision theoretic setting, reward measures the value of information

Page 5: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Reward functions:Value of Information (VOI)

Medical decision making: Utility depends on actual condition and chosen action

Actual condition unknown! Only know P(ill | O=o) EU(a | O=o) = P(ill | O=o) U(ill, a)

+ P(healthy | O=o) U(healthy, a)

VOI = expected maximum expected utility

healthy ill

Treatment -$$ $

No treatment 0 -$$$The more we know, the more effectively we can act

ConditionsActions

Page 6: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Local reward functions

Often, we want to evaluate rewards on multiple variables

Natural way of generalizing rewards to this setting: E[ R(X | O) ] := i E[ R(Xi | O) ]

Useful representation for many practical problems

Not fundamentally necessary in our approach

For any particular observation,local reward functions can be efficiently evaluated using probabilistic inference!

Page 7: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Costs and budgets

Each variable X can have a different cost c(X) Instead of only allowing k questions, we specify

integer budget B which we can spend Examples:

Medical domain: Cost of examinations Sensor networks: Power consumption Part-of-speech tagging: Fee for asking expert

Page 8: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

The subset selection problem

Consider myopically selecting

This can be seen as an attempt to nonmyopically maximize

Selected subset O is specified in advance (open loop)

E[R({O1})] , E[R({O2, O1})], ... , E[R({Ok,Ok-1 ... O1})]

mostinformativesingleton

most informative

(greedy)improvement

greedy selection

= X2O c(X)total cost of observing

budget

Often, we can acquire informationbased on earlier observations.

What about this closed loop setting?

Page 9: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

This outcome is inconsistent with our beliefs,so we better explore further by querying Y1

Y1 Y2 Y3 Y4 Y5

The conditional plan problem

Andreas is giving a talk

X1 X2 X3 X4 X5

Y2=PValues:

(S)ubject,

(P)redicate,

(O)bject

Assume, most informative query would be Y2

This outcome is consistent with our beliefs,so we can e.g. stop querying.

Y2=S

Now assume we observe a different outcome

Page 10: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

The conditional plan problem

Conditional plan selects different subset (s) for all outcomes S = s

Find conditional plan nonmyopically maximizing

Y2 = ?

Y5 = ? Y3 = ? Y5 = ?

SP

O

Y3 = ? Y3 = ? Y4 = ?

SP

O

Y1 = ? Y4 = ?

SP

O

stop

Nonmyopic planning implies that we construct the entire (exponentially large)

plan in advance!

Not clear if even compactly representable!

Page 11: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

A nonmyopic analysis

Problems intuitively seem hard Most previous approaches are myopic

Greedily select next best observation

In this paper, we present the first optimal nonmyopic algorithms for a non-trivial class of graphical models complexity theoretic hardness results

Page 12: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Inference in graphical models

Inference P(Xi = x | O = o) needed to compute local reward functions

Efficient inference possible for many graphical models:

X1 X2 X3 X3

X1

X2

X4

X5

X1 X3 X5

X2 X4 X6

What about optimizing value of information?

Page 13: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Chain graphical models

Filtering: Only use past observations Sensor scheduling, ...

Smoothing: Use all observations Structured classification, ...

Contains conditional chains HMMs, chain CRFs

X1 X2 X3 X4 X5

flow of informationflow of information

Page 14: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Key insight

X1 X2 X3 X4 X5 X6

Reward(1:3) Reward(3:6)

Reward(1:6) = Reward(1:3) + Reward(3:6) + const(3)

X3

Making observation

Expected Reward for subchain 1:3 whenobserving X1 and X3

Expected Reward for subchain 3:6 whenobserving X3 and X6

Reward functions decompose along chain!

Expected Reward for subchain 1:6 when

observing X1 , X3 and X6

Page 15: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Dynamic programming

Base case: 0 observations leftCompute expected reward for all sub-chains without making observations

Inductive case: k observations leftFind optimal observation (= split), optimally allocate budget (depending on observation)

Page 16: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Base case

X1 X2 X3 X4 X5 X6

Reward(1:2)

X1 X2 X3

Reward(1:3)

X1 X2 X3 X4

Reward(1:4)

X1 X2 X3 X4 X5

Reward(1:5)

X1 X2 X3 X4 X5 X6

Reward(1:6)

X1 X2 X3 X4 X5 X6

Reward(2:3)

X2 X3 X4 X5 X6

Reward(2:4)

X2 X3 X4 X5

Reward(2:5)

X2 X3 X4 X5 X6

Reward(2:6)

X1

1 2 3 4 5 6

2

3

4

5

6

0.8

1.7

2.4

3.0

2.9

2.4

0.7

1.8

3.0

Beginning of sub-chain

End

of

sub-

chai

n

Page 17: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

0 1 1

Inductive case

Reward =

E.g., compute value for spending first of three observations at X3; have 2 observations left

Compute expected reward for subchain a:b, making k observations, using expected rewards for all subchains with at most k-1 observations

2 1 0

1.0 + 3.0 = 4.0

X1 X2 X3 X4 X5 X6

spend obs. here spend obs. here

2.0 + 2.5 = 4.5 2.0 + 2.6 = 4.6 computed using base case and inductive case for 1,2

obs.

Can compute value of any split by optimally allocating budgets, referring to

base and earlier inductive cases.

For subset selection / filtering, speedups are possible.

Page 18: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Value of information for split at 3: 3.9, best: 3.9 Value of information for split at 4: 3.8, best: 3.9 Value of information for split at 5: 3.3, best: 3.9 Value of information for split at 2: 3.7, best: 3.7

Inductive case (continued)

X1 X2 X3 X4 X5 X6

Reward(1:2) Reward(2:6)

Reward(1:6) = Reward(1:2) + Reward(2:6) + const(2)

current best

X1 X2 X3 X4 X5 X6

Reward(1:3) Reward(3:6)

Reward(1:6) = Reward (1:3) + Reward (3:6) + const(3)

current best

X1 X2 X3 X4 X5 X6

Reward (1:4) Reward(4:6)

Reward(1:6) = Reward(1:4) + Reward(4:6) + const(4)

current best

X1 X2 X3 X4 X5 X6

Reward(1:5) Reward(5:6)

Reward(1:6) = Reward(1:5) + Reward(5:6) + const(5)

current best

Optimal VOI for subchain 1:6 and k observations to make = 3.9

1 2 3 4 5 6

2 0.8

3 2.1

4 2.8

5 3.4

6need to compute optimal VOI with k observation left

3.9

Compute expected reward for subchain a:b, making k observations, using expected rewards for all subchains with at most k-1 observations

End

of

sub-

chai

n

Beginning of sub-chainTracing back the maximal values

allows to recover the optimal subset or conditional plan!

Tables represent solution inpolynomial space!

Here we don’t needto allocate budget

Now we need to optimally allocate

our budget!

Page 19: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Results about optimal algorithms

Theorem: For chain graphical models, our algorithms compute• the nonmyopic optimal subset

in time O( d B n2 ) for filtering and

in time O( d2 B n3 ) for smoothing• the nonmyopic optimal conditional plan

in time O( d2 B n2 ) for filtering andin time O( d3 B2 n3 ) for smoothing

d: maximum domain size; B: budget we can spend for observations n: number of random variables

Page 20: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation of our algorithms

Three real-world data sets Sensor scheduling CpG-island detection Part-of-speech tagging

Goals: Compare optimal algorithms with (myopic) heuristics Relating objective values to prediction accuracy

Page 21: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation: Temperature

Temperature data from sensor deployment at Intel Research Berkeley

Task: Scheduling of single sensor Select k optimal times to observe

sensor during each day Optimize sum of residual entropiesS

ER

VE

R

LAB

KITC

HE

N

CO

PYE

LEC

PH

ON

EQ

UIE

T

ST

OR

AG

E

CO

NFE

RE

NC

E

OFFIC

EO

FFICE

50

51

52

53

54

46

48

49

47

43

45

44

42

41

37

39

38

36

33 3 6 10

11

12

13

14

15

16

17

19

20

21

22

24

25

26

28

30

32

31

27

29

23 1

8

9

5 8

74

34 1

235

40

Page 22: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Optimal algorithms significantly improve on commonly used myopic heuristics

Conditional plans give higher rewards than subsets

Evaluation: Temperature

Baseline:

Uniform spacing

of observations

24h0h

Page 23: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation: CpG-island detection

Annotated gene DNA sequences Task: Predict start and end of CpG island

ask expert to annotate k places in sequence optimize classification margin

Page 24: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation: CpG-island detection

Optimal algorithms provide better prediction accuracy Even small differences in objective value can lead to

improved prediction results

Page 25: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation: Reuters data

POS-Tagging CRF trained on Reuters news archive data

Task: Ask expert for k most informative tags Maximize classification margin

Page 26: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Evaluation: POS-Tagging

Optimizing classification margin leads to improved precision and recall

Page 27: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Can we generalize?

Many Graphical Models Tasks (e.g. Inference, MPE) which are efficiently solvable for chains can be generalized to polytrees

Even computing expected rewards is hard Optimization is a lot harder!

X1

X2

X3

X4

X5

Page 28: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Complexity Classes (Review)

P

NP – SAT

#P – #SAT

NPPP – E-MAJSAT

Probabilistic inferencein general graphical models

MAP assignmenton general GMs; Some planning

problems

Wildly more complex!!

Probabilistic inferencein polytrees

Page 29: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Hardness results

Proof by reduction from #3CNF-SAT and E-MAJSAT

Theorem: Even on discrete polytrees, • computing expected rewards is #P-complete• subset selection is NPPP-complete• computing conditional plans is NPPP-hard

As we presented last week at UAI, approximation algorithms with

strong guarantees available!

subset selectioncomputing rewards

Page 30: Optimal Nonmyopic Value of Information in Graphical Models Efficient Algorithms and Theoretical Limits Andreas Krause, Carlos Guestrin Computer Science

Summary We developed efficient optimal nonmyopic

algorithms for chain graphical models subset selection and conditional plans filtering + smoothing

Even on discrete polytrees, problems become wildly intractable! Chain is probably only graphical model we can hope to

solve optimally

Our algorithms improve prediction accuracy

Provide viable optimal approach for a wide range of value of information tasks