27
Computer vision: models, learning and inference Chapter 2 Introduction to probability Please send errata to [email protected]

Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

  • Upload
    others

  • View
    6

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Computer vision: models, learning and inference

Chapter 2

Introduction to probability

Please send errata to [email protected]

Page 2: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Random variables

• A random variable x denotes a quantity that is uncertain

• May be result of experiment (flipping a coin) or a real world measurements (measuring temperature)

• If observe several instances of x we get different values

• Some values occur more than others and this information is captured by a probability distribution

2 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 3: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Discrete Random Variables

3 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 4: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Continuous Random Variable

4 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 5: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Joint Probability

• Consider two random variables x and y

• If we observe multiple paired instances, then some combinations of outcomes are more likely than others

• This is captured in the joint probability distribution

• Written as Pr(x,y)

• Can read Pr(x,y) as “probability of x and y”

5 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 6: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Joint Probability

6 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 7: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Marginalization We can recover probability distribution of any variable in a joint distribution

by integrating (or summing) over the other variables

7 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 8: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Marginalization We can recover probability distribution of any variable in a joint distribution

by integrating (or summing) over the other variables

8 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 9: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Marginalization We can recover probability distribution of any variable in a joint distribution

by integrating (or summing) over the other variables

9 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 10: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Marginalization We can recover probability distribution of any variable in a joint distribution

by integrating (or summing) over the other variables

Works in higher dimensions as well – leaves joint distribution between whatever variables are left

10 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 11: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Conditional Probability

• Conditional probability of x given that y=y1 is relative propensity of variable x to take different outcomes given that y is fixed to be equal to y1.

• Written as Pr(x|y=y1)

11 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 12: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Conditional Probability • Conditional probability can be extracted from joint probability

• Extract appropriate slice and normalize

12 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 13: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Conditional Probability

• More usually written in compact form

• Can be re-arranged to give

13 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 14: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Conditional Probability

• This idea can be extended to more than two variables

14 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 15: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Bayes’ Rule From before:

Combining:

Re-arranging:

15 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 16: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Bayes’ Rule Terminology

Posterior – what we know about y after seeing x

Prior – what we know about y before seeing x

Likelihood – propensity for observing a certain value of x given a certain value of y

Evidence –a constant to ensure that the left hand side is a valid distribution

16 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 17: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Independence • If two variables x and y are independent then variable x tells

us nothing about variable y (and vice-versa)

17 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 18: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Independence • If two variables x and y are independent then variable x tells

us nothing about variable y (and vice-versa)

18 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 19: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Independence • When variables are independent, the joint factorizes into a

product of the marginals:

19 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 20: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation

Expectation tell us the expected or average value of some function f [x] taking into account the distribution of x

Definition:

20 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 21: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation

Expectation tell us the expected or average value of some function f [x] taking into account the distribution of x

Definition in two dimensions:

21 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 22: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation: Common Cases

22 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 23: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation: Rules

Rule 1: Expected value of a constant is the constant

23 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 24: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation: Rules

Rule 2: Expected value of constant times function is constant times expected value of function

24 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 25: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation: Rules

Rule 3: Expectation of sum of functions is sum of expectation of functions

25 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 26: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Expectation: Rules

Rule 4: Expectation of product of functions in variables x and y is product of expectations of functions if x and y are independent

26 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

Page 27: Computer vision: models, learning and inferencejmagee/cs201/slides/02_Intro_To_Probability.pdf · Computer vision: models, learning and inference. ©2011 Simon J.D. Prince 27 •

Conclusions

27 Computer vision: models, learning and inference. ©2011 Simon J.D. Prince

• Rules of probability are compact and simple

• Concepts of marginalization, joint and conditional probability, Bayes rule and expectation underpin all of the models in this book

• One remaining concept – conditional expectation – discussed later