329
Today. Quick Review:

Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

  • Upload
    others

  • View
    0

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Today.

Quick Review:

experts framework/multiplicative weights algorithmFinish:

randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 2: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 3: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:

randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 4: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 5: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 6: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Notes.

Got to definition of Approximate Equilibrium for zero sumgames.

Page 7: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

The multiplicative weights framework.

Page 8: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 9: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 10: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 11: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 12: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 13: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 14: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 15: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 16: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 17: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 18: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 19: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..

never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 20: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never!

Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 21: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 22: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 23: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make?

Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 24: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 25: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 26: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 27: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 28: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:

makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 29: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.

”You could have done so well”...but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 30: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 31: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t!

ha..ha!

Analysis of Algorithms: do as well as possible!

Page 32: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..

ha!

Analysis of Algorithms: do as well as possible!

Page 33: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 34: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 35: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 36: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 37: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1

Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 38: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.

Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 39: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound:

every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 40: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 41: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 42: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 43: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 44: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 45: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?

(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 46: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1

At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 47: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 48: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 49: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts

mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 50: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect experts

mistake→ ≤ n/4 perfect experts...

mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 51: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 52: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...

mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 53: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 54: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 55: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert

→ at most logn mistakes!

Page 56: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 57: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 58: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 59: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm.

Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 60: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 61: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 62: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 63: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 64: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.

2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 65: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.

3. wi → wi/2 if wrong.

Page 66: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 67: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 68: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority

1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 69: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 70: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 71: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function:

∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 72: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi .

Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 73: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 74: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 75: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:

total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 76: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by

−1? −2? factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 77: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1?

−2? factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 78: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2?

factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 79: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 80: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 81: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 82: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 83: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 84: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 85: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 86: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 87: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes

M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 88: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 89: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 90: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 91: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 92: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)

≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 93: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 94: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 95: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 96: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 97: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 98: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 99: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 100: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 101: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 102: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 103: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 104: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 105: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization

!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 106: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization

!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 107: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 108: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 109: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 110: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 111: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 112: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 113: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 114: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly

optimal!

Page 115: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 116: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 117: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 118: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 119: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 120: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 121: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 122: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 123: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 124: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 125: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 126: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 127: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t .

W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 128: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) =

n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 129: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 130: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total.

→W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 131: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 132: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 133: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt)

Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 134: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 135: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:

W (t +1)≤∑i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 136: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi

= ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 137: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 138: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)

= W (t)(1− εLt)

Page 139: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 140: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 141: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 142: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 143: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 144: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 145: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 146: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 147: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε)

ish of the best expert!

No factor of 2 loss!

Page 148: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish

of the best expert!

No factor of 2 loss!

Page 149: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 150: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 151: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 152: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 153: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 154: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 155: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 156: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 157: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 158: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 159: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 160: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 161: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 162: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 163: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 164: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 165: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:

Choose proportional to weightsmultiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 166: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 167: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 168: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 169: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 170: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 171: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 172: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).

Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 173: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 174: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 175: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 176: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 177: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 178: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 179: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 180: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 181: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 182: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:

Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 183: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 184: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 185: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo.

Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 186: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 187: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 188: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 189: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 190: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo.

Value of C?

Page 191: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 192: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 193: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 194: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 195: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :

row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 196: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v

=⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 197: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .

column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 198: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v

=⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 199: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.

=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 200: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 201: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 202: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point!

and R = C!

Doesn’t matter who plays first!

Page 203: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 204: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 205: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later.

Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 206: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 207: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 208: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 209: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 210: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 211: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 212: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 213: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)

→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 214: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 215: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 216: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)

→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 217: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”

→ “Response x to y is within ε of best response”

Page 218: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 219: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 220: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.

(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 221: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.

(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 222: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.

(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 223: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 224: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C)

..and (D).Not hard. Even easy. Still, head scratching happens.

Page 225: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).

Not hard. Even easy. Still, head scratching happens.

Page 226: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard.

Even easy. Still, head scratching happens.

Page 227: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy.

Still, head scratching happens.

Page 228: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 229: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 230: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 231: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 232: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 233: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 234: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 235: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.

Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 236: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 237: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 238: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.

Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 239: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.

Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 240: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 241: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .

Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 242: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.

Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 243: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 244: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt

and x∗ = argminxtxtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 245: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 246: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 247: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:

xt minimizes the best column response is chosen. Clearly good for row.column best response is at least what it is against xt . Total loss, L is at least

column payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 248: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen.

Clearly good for row.column best response is at least what it is against xt . Total loss, L is at least

column payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 249: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 250: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt .

Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 251: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff.

Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 252: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.

Combine bounds. Done!

Page 253: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds.

Done!

Page 254: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 255: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 256: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt

and x∗ = argminxtxtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 257: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 258: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 259: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .

Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 260: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .

Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 261: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 262: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 263: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt

and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 264: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt

→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 265: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.

→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 266: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 267: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 268: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights:

L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 269: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 270: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε

→ C(x∗)≤ (1+ ε)R(y∗)+ lnnεT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 271: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 272: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 273: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1

→ C(x∗)−R(y∗)≤ 2ε.

Page 274: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 275: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 276: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt

and y∗ = 1T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 277: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 278: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 279: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .

Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 280: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).

Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 281: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .

Algorithm loss: ∑t xtAyt ≥ ∑t xtAyrL≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 282: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 283: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 284: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 285: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt

and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 286: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt

→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 287: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.

→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 288: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 289: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 290: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights:

L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 291: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 292: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε

→ C(x∗)≤ (1+ ε)R(y∗)+ lnnεT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 293: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 294: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 295: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1

→ C(x∗)−R(y∗)≤ 2ε.

Page 296: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 297: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 298: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist?

Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 299: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 300: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here?

Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 301: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 302: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 303: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?

T = lnnε2 → O(nm logn

ε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 304: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2

→ O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 305: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ).

Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 306: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 307: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m)

Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 308: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.

(Faster linear programming: O(√

n +m) linear solution solves.)Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 309: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 310: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower

... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 311: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 312: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 313: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 314: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 315: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 316: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: r

column for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 317: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 318: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 319: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)

Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 320: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.

Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 321: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 322: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.

Route: minimize max loaded on any edge.

Page 323: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 324: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 325: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 326: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 327: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 328: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 329: Today. - Peoplesatishr/cs270/sp13/slides/lec-5.pdf · Today. Quick Review: experts framework/multiplicative weights algorithm Finish: randomized multiplicative weights algorithm for

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.