40
18.175: Lecture 14 Weak convergence and characteristic functions Scott Sheffield MIT 18.175 Lecture 14

18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

  • Upload
    others

  • View
    3

  • Download
    0

Embed Size (px)

Citation preview

Page 1: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

18.175: Lecture 14

Weak convergence and characteristicfunctions

Scott Sheffield

MIT

18.175 Lecture 14

Page 2: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Outline

Weak convergence

Characteristic functions

18.175 Lecture 14

Page 3: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Outline

Weak convergence

Characteristic functions

18.175 Lecture 14

Page 4: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 5: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.

I Theorem: Xn =⇒ X∞ if and only if for every boundedcontinuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 6: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 7: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 8: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 9: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 14

Page 10: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 14

Page 11: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 14

Page 12: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 14

Page 13: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 14

Page 14: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Corresponds to L1 distance between density functions whenthese exist.

I Convergence in total variation norm is much stronger thanweak convergence. Discrete uniform random variable Un on(1/n, 2/n, 3/n, . . . , n/n) converges weakly to uniform randomvariable U on [0, 1]. But total variation distance between Un

and U is 1 for all n.

18.175 Lecture 14

Page 15: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Corresponds to L1 distance between density functions whenthese exist.

I Convergence in total variation norm is much stronger thanweak convergence. Discrete uniform random variable Un on(1/n, 2/n, 3/n, . . . , n/n) converges weakly to uniform randomvariable U on [0, 1]. But total variation distance between Un

and U is 1 for all n.

18.175 Lecture 14

Page 16: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Corresponds to L1 distance between density functions whenthese exist.

I Convergence in total variation norm is much stronger thanweak convergence. Discrete uniform random variable Un on(1/n, 2/n, 3/n, . . . , n/n) converges weakly to uniform randomvariable U on [0, 1]. But total variation distance between Un

and U is 1 for all n.

18.175 Lecture 14

Page 17: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Corresponds to L1 distance between density functions whenthese exist.

I Convergence in total variation norm is much stronger thanweak convergence. Discrete uniform random variable Un on(1/n, 2/n, 3/n, . . . , n/n) converges weakly to uniform randomvariable U on [0, 1]. But total variation distance between Un

and U is 1 for all n.

18.175 Lecture 14

Page 18: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Outline

Weak convergence

Characteristic functions

18.175 Lecture 14

Page 19: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Outline

Weak convergence

Characteristic functions

18.175 Lecture 14

Page 20: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 21: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 22: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 23: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 24: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 25: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 26: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 27: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic functions

I Let X be a random variable.

I The characteristic function of X is defined byφ(t) = φX (t) := E [e itX ].

I Recall that by definition e it = cos(t) + i sin(t).

I Characteristic function φX similar to moment generatingfunction MX .

I φX+Y = φXφY , just as MX+Y = MXMY , if X and Y areindependent.

I And φaX (t) = φX (at) just as MaX (t) = MX (at).

I And if X has an mth moment then E [Xm] = imφ(m)X (0).

I Characteristic functions are well defined at all t for all randomvariables X .

18.175 Lecture 14

Page 28: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function properties

I φ(0) = 1

I φ(−t) = φ(t)

I |φ(t)| = |Ee itX | ≤ E |e itX | = 1.

I |φ(t + h)− φ(t)| ≤ E |e ihX − 1|, so φ(t) uniformly continuouson (−∞,∞)

I Ee it(aX+b) = e itbφ(at)

18.175 Lecture 14

Page 29: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function properties

I φ(0) = 1

I φ(−t) = φ(t)

I |φ(t)| = |Ee itX | ≤ E |e itX | = 1.

I |φ(t + h)− φ(t)| ≤ E |e ihX − 1|, so φ(t) uniformly continuouson (−∞,∞)

I Ee it(aX+b) = e itbφ(at)

18.175 Lecture 14

Page 30: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function properties

I φ(0) = 1

I φ(−t) = φ(t)

I |φ(t)| = |Ee itX | ≤ E |e itX | = 1.

I |φ(t + h)− φ(t)| ≤ E |e ihX − 1|, so φ(t) uniformly continuouson (−∞,∞)

I Ee it(aX+b) = e itbφ(at)

18.175 Lecture 14

Page 31: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function properties

I φ(0) = 1

I φ(−t) = φ(t)

I |φ(t)| = |Ee itX | ≤ E |e itX | = 1.

I |φ(t + h)− φ(t)| ≤ E |e ihX − 1|, so φ(t) uniformly continuouson (−∞,∞)

I Ee it(aX+b) = e itbφ(at)

18.175 Lecture 14

Page 32: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function properties

I φ(0) = 1

I φ(−t) = φ(t)

I |φ(t)| = |Ee itX | ≤ E |e itX | = 1.

I |φ(t + h)− φ(t)| ≤ E |e ihX − 1|, so φ(t) uniformly continuouson (−∞,∞)

I Ee it(aX+b) = e itbφ(at)

18.175 Lecture 14

Page 33: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 34: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 35: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 36: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 37: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 38: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 39: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14

Page 40: 18.175: Lecture 14 .1in Weak convergence and characteristic …math.mit.edu/~sheffield/175/Lecture14.pdf · 2014-02-11 · 18.175 Lecture 14. Total variation norm I If we have two

Characteristic function examples

I Coin: If P(X = 1) = P(X = −1) = 1/2 thenφX (t) = (e it + e−it)/2 = cos t.

I That’s periodic. Do we always have periodicity if X is arandom integer?

I Poisson: If X is Poisson with parameter λ thenφX (t) =

∑∞k=0 e

−λ λke itkk! = exp(λ(e it − 1)).

I Why does doubling λ amount to squaring φX ?

I Normal: If X is standard normal, then φX (t) = e−t2/2.

I Is φX always real when the law of X is symmetric about zero?

I Exponential: If X is standard exponential (density e−x on(0,∞)) then φX (t) = 1/(1− it).

I Bilateral exponential: if fX (t) = e−|x |/2 on R thenφX (t) = 1/(1 + t2). Use linearity of fX → φX .

18.175 Lecture 14