329
Today. Quick Review:

people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

  • Upload
    others

  • View
    4

  • Download
    0

Embed Size (px)

Citation preview

Page 1: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Today.

Quick Review:

experts framework/multiplicative weights algorithmFinish:

randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 2: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 3: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:

randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 4: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 5: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Today.

Quick Review:experts framework/multiplicative weights algorithm

Finish:randomized multiplicative weights algorithm for experts framework.

Equilibrium for two person games:using experts framework/MW algorithm.

Page 6: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Notes.

Got to definition of Approximate Equilibrium for zero sumgames.

Page 7: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

The multiplicative weights framework.

Page 8: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 9: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 10: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 11: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 12: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 13: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 14: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Expert’s framework.

n experts.

Every day, each offers a prediction.

“Rain” or “Shine.”

Whose advise do you follow?

“The one who is correct most often.”

Sort of.

How well do you do?

Page 15: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 16: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 17: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 18: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 19: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..

never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 20: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never!

Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 21: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 22: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 23: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make?

Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 24: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 25: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 26: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Infallible expert.One of the expert’s is infallible!

Your strategy?

Choose any expert that has not made a mistake!

How long to find perfect expert?

Maybe..never! Never see a mistake.

Better model?

How many mistakes could you make? Mistake Bound.

(A) 1(B) 2(C) logn(D) n−1

Adversary designs setup to watch who you choose, and makethat expert make a mistake.

n−1!

Page 27: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 28: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:

makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 29: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.

”You could have done so well”...but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 30: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 31: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t!

ha..ha!

Analysis of Algorithms: do as well as possible!

Page 32: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..

ha!

Analysis of Algorithms: do as well as possible!

Page 33: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 34: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Concept Alert.

Note.

Adversary:makes you want to look bad.”You could have done so well”...

but you didn’t! ha..ha!

Analysis of Algorithms: do as well as possible!

Page 35: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 36: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 37: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1

Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 38: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.

Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 39: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound:

every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 40: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 41: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 42: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 43: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 44: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Back to mistake bound.

Infallible Experts.

Alg: Choose one of the perfect experts.

Mistake Bound: n−1Lower bound: adversary argument.Upper bound: every mistake finds fallible expert.

Better Algorithm?

Making decision, not trying to find expert!

Algorithm: Go with the majority of previously correct experts.

What you would do anyway!

Page 45: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?

(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 46: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1

At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 47: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 48: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 49: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts

mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 50: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect experts

mistake→ ≤ n/4 perfect experts...

mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 51: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 52: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...

mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 53: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 54: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 55: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert

→ at most logn mistakes!

Page 56: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Alg 2: find majority of the perfect

How many mistakes could you make?(A) 1(B) 2(C) logn(D) n−1At most logn!

When alg makes a mistake,|“perfect” experts| drops by a factor of two.

Initially n perfect experts mistake→ ≤ n/2 perfect expertsmistake→ ≤ n/4 perfect experts

...mistake→ ≤ 1 perfect expert

≥ 1 perfect expert→ at most logn mistakes!

Page 57: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 58: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 59: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm.

Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 60: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 61: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 62: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 63: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 64: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.

2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 65: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.

3. wi → wi/2 if wrong.

Page 66: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 67: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Imperfect Experts

Goal?

Do as well as the best expert!

Algorithm. Suggestions?

Go with majority?

Penalize inaccurate experts?

Best expert is penalized the least.

1. Initially: wi = 1.2. Predict with weighted majority of experts.3. wi → wi/2 if wrong.

Page 68: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority

1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 69: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 70: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 71: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function:

∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 72: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi .

Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 73: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 74: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 75: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:

total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 76: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by

−1? −2? factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 77: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1?

−2? factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 78: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2?

factor of 12?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 79: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?

each incorrect expert weight multiplied by 12 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 80: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !

total weight decreases byfactor of 1

2? factor of 34?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 81: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 82: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?

mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 83: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 84: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 85: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: weighted majority1. Initially: wi = 1.

2. Predict withweightedmajority ofexperts.

3. wi → wi/2 ifwrong.

Goal: Best expert makes m mistakes.

Potential function: ∑i wi . Initially n.

For best expert, b, wb ≥ 12m .

Each mistake:total weight of incorrect experts reduced by−1? −2? factor of 1

2?each incorrect expert weight multiplied by 1

2 !total weight decreases by

factor of 12? factor of 3

4?mistake→ ≥ half weight with incorrect experts.

Mistake→ potential function decreased by 34 .

We have

12m ≤∑

iwi ≤

(34

)M

n.

where M is number of algorithm mistakes.

Page 86: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 87: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes

M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 88: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 89: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 90: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 91: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 92: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)

≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 93: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 94: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 95: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 96: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 97: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 98: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis: continued.

12m ≤ ∑i wi ≤

(34

)Mn.

m - best expert mistakes M algorithm mistakes.

12m ≤

(34

)Mn.

Take log of both sides.

−m ≤−M log(4/3)+ logn.

Solve for M.M ≤ (m + logn)/ log(4/3)≤ 2.4(m + logn)

Multiple by 1− ε for incorrect experts...

(1− ε)m ≤(1− ε

2

)M n.

Massage...

M ≤ 2(1+ ε)m + 2lnnε

Approaches a factor of two of best expert performance!

Page 99: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 100: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 101: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 102: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 103: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 104: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Analysis?

Two experts: A,B

Bad example?

Which is worse?(A) A right on even, B right on odd.(B) A right first half of days, B right second

Best expert peformance: T/2 mistakes.

Pattern (A): T −1 mistakes.

Factor of (almost) two worse!

Page 105: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization

!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 106: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization

!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 107: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 108: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 109: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 110: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 111: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 112: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 113: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 114: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly

optimal!

Page 115: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomization!!!!

Better approach?

Use?

Randomization!

That is, choose expert i with prob ∝ wi

Bad example: A,B,A,B,A...

After a bit, A and B make nearly the same number of mistakes.

Choose each with approximately the same probabilty.

Make a mistake around 1/2 of the time.

Best expert makes T/2 mistakes.

Rougly optimal!

Page 116: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 117: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 118: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 119: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 120: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 121: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized analysis.

Some formulas:

For ε ≤ 1,x ∈ [0,1],

(1+ ε)x ≤ (1+ εx)(1− ε)x ≤ (1− εx)

For ε ∈ [0, 12 ],

−ε− ε2 ≤ ln(1− ε)≤−ε

ε− ε2 ≤ ln(1+ ε)≤ ε

Proof Idea: ln(1+x) = x− x2

2 + x3

3 −·· ·

Page 122: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 123: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 124: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 125: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 126: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 127: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t .

W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 128: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) =

n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 129: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 130: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total.

→W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 131: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 132: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 133: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt)

Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 134: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 135: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:

W (t +1)≤∑i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 136: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi

= ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 137: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 138: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)

= W (t)(1− εLt)

Page 139: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Randomized algorithmLosses in [0,1].

Expert i loses `ti ∈ [0,1] in round t.

1. Initially wi = 1 for expert i .

2. Choose expert i with prob wiW , W = ∑i wi .

3. wi ← wi(1− ε)`ti

W (t) sum of wi at time t . W (0) = n

Best expert, b, loses L∗ total. →W (T )≥ wb ≥ (1− ε)L∗ .

Lt = ∑iwi `

ti

W expected loss of alg. in time t .

Claim: W (t +1)≤W (t)(1− εLt) Loss→ weight loss.

Proof:W (t +1)≤∑

i(1− ε`t

i )wi = ∑i

wi − ε ∑i

wi`ti

= ∑i

wi

(1− ε

∑i wi`ti

∑i wi

)= W (t)(1− εLt)

Page 140: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 141: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 142: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 143: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 144: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 145: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 146: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 147: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε)

ish of the best expert!

No factor of 2 loss!

Page 148: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish

of the best expert!

No factor of 2 loss!

Page 149: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 150: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Analysis

(1− ε)L∗ ≤W (T )≤ n ∏t(1− εLt)

Take logs(L∗) ln(1− ε)≤ lnn +∑ ln(1− εLt)

Use −ε− ε2 ≤ ln(1− ε)≤−ε

−(L∗)(ε + ε2)≤ lnn− ε ∑Lt

And

∑t Lt ≤ (1+ ε)L∗+ lnnε

.

∑t Lt is total expected loss of algorithm.

Within (1+ ε) ish of the best expert!

No factor of 2 loss!

Page 151: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 152: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 153: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 154: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 155: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 156: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 157: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Gains.

Why so negative?

Each day, each expert gives gain in [0,1].

Multiplicative weights with (1+ ε)gti .

G ≥ (1− ε)G∗− lognε

where G∗ is payoff of best expert.

Scaling:

Not [0,1], say [0,ρ].

L≤ (1+ ε)L∗+ρ logn

ε

Page 158: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 159: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 160: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 161: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 162: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 163: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 164: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 165: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:

Choose proportional to weightsmultiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 166: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 167: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 168: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 169: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 170: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Summary: multiplicative weights.

Framework: n experts, each loses different amount every day.

Perfect Expert: logn mistakes.

Imperfect Expert: best makes m mistakes.

Deterministic Strategy: 2(1+ ε)m + lognε

Real numbered losses: Best loses L∗ total.

Randomized Strategy: (1+ ε)L∗+ lognε

Strategy:Choose proportional to weights

multiply weight by (1− ε)loss.

Multiplicative weights framework!

Applications next!

Page 171: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 172: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).

Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 173: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 174: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 175: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 176: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 177: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 178: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person zero sum games.m×n payoff matrix A.

Row mixed strategy: x = (x1, . . . ,xm).Column mixed strategy: y = (y1, . . . ,yn).

Payoff for strategy pair (x ,y):

p(x ,y) = x tAy

That is,

∑i

xi

(∑j

ai ,jyj

)= ∑

j

(∑i

xiai ,j

)yj .

Recall row minimizes, column maximizes.

Equilibrium pair: (x∗,y∗)?

(x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

Page 179: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 180: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 181: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Equilibrium.

Equilibrium pair: (x∗,y∗)?

p(x ,y) = (x∗)tAy∗ = maxy

(x∗)tAy = minx

x tAy∗.

(No better column strategy, no better row strategy.)

No row is better:mini A(i) ·y = (x∗)tAy∗. 1

No column is better:maxj(At)(j) ·x = (x∗)tAy∗.

1A(i) is i th row.

Page 182: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:

Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 183: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 184: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 185: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo.

Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 186: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 187: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 188: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 189: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 190: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo.

Value of C?

Page 191: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Best Response

Column goes first:Find y , where best row is not too low..

R = maxy

minx

(x tAy).

Note: x can be (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of R?

Row goes first:Find x , where best column is not high.

C = minx

maxy

(x tAy).

Agin: y of form (0,0, . . . ,1, . . .0).

Example: Roshambo. Value of C?

Page 192: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 193: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 194: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 195: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :

row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 196: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v

=⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 197: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .

column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 198: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v

=⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 199: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.

=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 200: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 201: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 202: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point!

and R = C!

Doesn’t matter who plays first!

Page 203: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 204: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Duality.

R = maxy

minx

(x tAy).

C = minx

maxy

(x tAy).

Weak Duality: R ≤ C.Proof: Better to go second.

At Equilibrium (x∗,y∗), payoff v :row payoffs (Ay∗) all ≥ v =⇒ R ≥ v .column payoffs ((x∗)tA) all ≤ v =⇒ v ≥ C.=⇒ R ≥ C

Equilibrium =⇒ R = C!

Strong Duality: There is an equilibrium point! and R = C!

Doesn’t matter who plays first!

Page 205: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later.

Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 206: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 207: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 208: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 209: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 210: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 211: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 212: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 213: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)

→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 214: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 215: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 216: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)

→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 217: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”

→ “Response x to y is within ε of best response”

Page 218: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 219: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of Equilibrium.

Later. Still later...

Aproximate equilibrium ...

C(x) = maxy x tAy

R(y) = minx x tAy

Always: R(y)≤ C(x)

Strategy pair: (x ,y)

Equilibrium: (x ,y)

R(y) = C(x)→ C(x)−R(y) = 0.

Approximate Equilibrium: C(x)−R(y)≤ ε.

With R(y)≤ C(x)→ “Response y to x is within ε of best response”→ “Response x to y is within ε of best response”

Page 220: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.

(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 221: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.

(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 222: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.

(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 223: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 224: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C)

..and (D).Not hard. Even easy. Still, head scratching happens.

Page 225: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).

Not hard. Even easy. Still, head scratching happens.

Page 226: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard.

Even easy. Still, head scratching happens.

Page 227: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy.

Still, head scratching happens.

Page 228: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Proof of approximate equilibrium.

How?

(A) Using geometry.(B) Using a fixed point theorem.(C) Using multiplicative weights.(D) By the skin of my teeth.

(C) ..and (D).Not hard. Even easy. Still, head scratching happens.

Page 229: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 230: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 231: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 232: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 233: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 234: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and experts

Again: find (x∗,y∗), such that

(maxy x∗Ay)− (minx x∗Ay∗)≤ ε

C(x∗) − R(y∗)≤ ε

Experts Framework: n Experts, T days, L∗ -total loss.

Multiplicative Weights Method yields loss L where

L≤ (1+ ε)L∗+ lognε

Page 235: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.

Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 236: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 237: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 238: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.

Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 239: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.

Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 240: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 241: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .

Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 242: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.

Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 243: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 244: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt

and x∗ = argminxtxtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 245: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 246: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 247: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:

xt minimizes the best column response is chosen. Clearly good for row.column best response is at least what it is against xt . Total loss, L is at least

column payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 248: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen.

Clearly good for row.column best response is at least what it is against xt . Total loss, L is at least

column payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 249: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 250: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt .

Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 251: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff.

Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 252: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.

Combine bounds. Done!

Page 253: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds.

Done!

Page 254: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Games and Experts.Assume: A has payoffs in [0,1].

For T = lognε2 days:

1) m pure row strategies are experts.Use multiplicative weights, produce row distribution.Let xt be distribution (row strategy) xt on day t .

2) Each day, adversary plays best column response to xt .Choose column of A that maximizes row’s expected loss.Let yt be indicator vector for this column.

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Proof Idea:xt minimizes the best column response is chosen. Clearly good for row.

column best response is at least what it is against xt . Total loss, L is at leastcolumn payoff. Best row payoff, L∗ is roughly less than L due to MW anlysis.Combine bounds. Done!

Page 255: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 256: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt

and x∗ = argminxtxtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 257: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 258: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 259: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .

Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 260: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .

Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 261: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 262: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 263: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt

and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 264: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt

→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 265: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.

→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 266: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 267: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 268: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights:

L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 269: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 270: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε

→ C(x∗)≤ (1+ ε)R(y∗)+ lnnεT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 271: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 272: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 273: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1

→ C(x∗)−R(y∗)≤ 2ε.

Page 274: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium!Experts: xt is strategy on day t , yt is best column against xt .

Let y∗ = 1T ∑t yt and x∗ = argminxt

xtAyt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Loss on day t , xtAyt ≥ C(x∗) by the choice of x .Thus, algorithm loss, L, is ≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 275: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 276: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt

and y∗ = 1T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 277: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 278: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 279: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .

Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 280: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).

Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 281: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .

Algorithm loss: ∑t xtAyt ≥ ∑t xtAyrL≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 282: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 283: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 284: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 285: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt

and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 286: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt

→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 287: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.

→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 288: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 289: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 290: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights:

L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 291: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 292: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε

→ C(x∗)≤ (1+ ε)R(y∗)+ lnnεT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 293: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT

→ C(x∗)−R(y∗)≤ εR(y∗)+ lnnεT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 294: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 295: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1

→ C(x∗)−R(y∗)≤ 2ε.

Page 296: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Approximate Equilibrium: notes!Experts: xt is strategy on day t , yt is best column against xt .

Let x∗ = 1T ∑t xt and y∗ = 1

T ∑t yt .

Claim: (x∗,y)∗ are 2ε-optimal for matrix A.

Column payoff: C(x∗) = maxy x∗Ay .Let yr be best response to C(x∗).Day t ,yt best response to xt → xtAyt ≥ xtAyr .Algorithm loss: ∑t xtAyt ≥ ∑t xtAyr

L≥ TC(x∗).

Best expert: L∗- best row against all the columns played.

best row against ∑t Ayt and Ty∗ = ∑t yt→ best row against TAy∗.→ L∗ ≤ TR(y∗).

Multiplicative Weights: L≤ (1+ ε)L∗+ lnnε

TC(x∗)≤ (1+ ε)TR(y∗)+ lnnε→ C(x∗)≤ (1+ ε)R(y∗)+ lnn

εT→ C(x∗)−R(y∗)≤ εR(y∗)+ lnn

εT .

T = lnnε2 , R(y∗)≤ 1→ C(x∗)−R(y∗)≤ 2ε.

Page 297: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 298: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist?

Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 299: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 300: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here?

Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 301: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 302: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 303: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?

T = lnnε2 → O(nm logn

ε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 304: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2

→ O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 305: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ).

Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 306: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 307: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m)

Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 308: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.

(Faster linear programming: O(√

n +m) linear solution solves.)Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 309: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 310: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower

... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 311: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 312: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 313: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 314: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Comments

For any ε, there exists an ε-Approximate Equilibrium.

Does an equilibrium exist? Yes.

Something about math here? Fixed point theorem.

Later: will use geometry, linear programming.

Complexity?T = lnn

ε2 → O(nm lognε2 ). Basically linear!

Versus Linear Programming: O(n3m) Basically quadratic.(Faster linear programming: O(

√n +m) linear solution solves.)

Still much slower ... and more complicated.

Dynamics: best response, update weight, best response.

Also works with both using multiplicative weights.

“In practice.”

Page 315: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 316: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: r

column for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 317: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 318: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 319: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)

Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 320: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.

Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 321: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 322: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.

Route: minimize max loaded on any edge.

Page 323: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 324: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Toll/Congestion

Given: G = (V ,E).Given (s1, t1) . . .(sk , tk ).Row: choose routing of all paths.Column: choose edge.Row pays if column chooses edge on any path.

Matrix:row for each routing: rcolumn for each edge: e

A[r ,e] is congestion on edge e by routing r

Offense: (Best Response.)Router: route along shortest paths.Toll: charge most loaded edge.

Defense: Toll: maximize shortest path under tolls.Route: minimize max loaded on any edge.

Page 325: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 326: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 327: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 328: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.

Page 329: people.eecs.berkeley.edusatishr/cs270/sp13/slides/lecture5… · Infallible expert. One of the expert’s is infallible! Your strategy? Choose any expert that has not made a mistake!

Two person game.

Row is router.

An exponential number of rows.

Two person game with experts won’t be so easy to implement.

Version with row and column flipped may work.

Next Time.