85
18.175: Lecture 9 More large deviations Scott Sheffield MIT 18.175 Lecture 9

18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

  • Upload
    others

  • View
    2

  • Download
    0

Embed Size (px)

Citation preview

Page 1: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

18.175: Lecture 9

More large deviations

Scott Sheffield

MIT

18.175 Lecture 9

Page 2: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 3: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 4: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 5: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 6: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 7: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 8: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 9: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 10: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 11: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

DeMoivre-Laplace limit theorem

I Let Xi be i.i.d. random variables. Write Sn =∑n

i=1 Xn.

I Suppose each Xi is 1 with probability p and 0 with probabilityq = 1− p.

I DeMoivre-Laplace limit theorem:

limn→∞

Pa ≤ Sn − np√npq

≤ b → Φ(b)− Φ(a).

I Here Φ(b)− Φ(a) = Pa ≤ Z ≤ b when Z is a standardnormal random variable.

I Sn−np√npq describes “number of standard deviations that Sn is

above or below its mean”.

I Proof idea: use binomial coefficients and Stirling’s formula.

I Question: Does similar statement hold if Xi are i.i.d. fromsome other law?

I Central limit theorem: Yes, if they have finite variance.

18.175 Lecture 9

Page 12: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Local p = 1/2 DeMoivre-Laplace limit theorem

I Stirling: n! ∼ nne−n√

2πn where ∼ means ratio tends to one.

I Theorem: If 2k/√

2n→ x thenP(S2n = 2k) ∼ (πn)−1/2e−x

2/2.

I Recall P(S2n = 2k) =( 2nn+k

)2−2n = 2−2n (2n)!

(n+k)!(n−k)! .

18.175 Lecture 9

Page 13: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Local p = 1/2 DeMoivre-Laplace limit theorem

I Stirling: n! ∼ nne−n√

2πn where ∼ means ratio tends to one.

I Theorem: If 2k/√

2n→ x thenP(S2n = 2k) ∼ (πn)−1/2e−x

2/2.

I Recall P(S2n = 2k) =( 2nn+k

)2−2n = 2−2n (2n)!

(n+k)!(n−k)! .

18.175 Lecture 9

Page 14: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Local p = 1/2 DeMoivre-Laplace limit theorem

I Stirling: n! ∼ nne−n√

2πn where ∼ means ratio tends to one.

I Theorem: If 2k/√

2n→ x thenP(S2n = 2k) ∼ (πn)−1/2e−x

2/2.

I Recall P(S2n = 2k) =( 2nn+k

)2−2n = 2−2n (2n)!

(n+k)!(n−k)! .

18.175 Lecture 9

Page 15: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 16: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 17: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 18: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 19: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 20: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 21: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 22: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 23: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Weak convergence

I Let X be random variable, Xn a sequence of random variables.

I Say Xn converge in distribution or converge in law to X iflimn→∞ FXn(x) = FX (x) at all x ∈ R at which FX iscontinuous.

I Also say that the Fn = FXn converge weakly to F = FX .

I Example: Xi chosen from −1, 1 with i.i.d. fair coin tosses:then n−1/2

∑ni=1 Xi converges in law to a normal random

variable (mean zero, variance one) by DeMoivre-Laplace.

I Example: If Xn is equal to 1/n a.s. then Xn converge weaklyto an X equal to 0 a.s. Note that limn→∞ Fn(0) 6= F (0) inthis case.

I Example: If Xi are i.i.d. then the empirical distributionsconverge a.s. to law of X1 (Glivenko-Cantelli).

I Example: Let Xn be the nth largest of 2n + 1 points choseni.i.d. from fixed law.

18.175 Lecture 9

Page 24: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 25: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.

I Theorem: Xn =⇒ X∞ if and only if for every boundedcontinuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 26: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 27: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 28: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 29: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Convergence results

I Theorem: If Fn → F∞, then we can find correspondingrandom variables Yn on a common measure space so thatYn → Y∞ almost surely.

I Proof idea: Take Ω = (0, 1) and Yn = supy : Fn(y) < x.I Theorem: Xn =⇒ X∞ if and only if for every bounded

continuous g we have Eg(Xn)→ Eg(X∞).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

I Theorem: Suppose g is measurable and its set ofdiscontinuity points has µX measure zero. Then Xn =⇒ X∞implies g(Xn) =⇒ g(X ).

I Proof idea: Define Xn on common sample space so convergea.s., use bounded convergence theorem.

18.175 Lecture 9

Page 30: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 9

Page 31: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 9

Page 32: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 9

Page 33: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Compactness

I Theorem: Every sequence Fn of distribution has subsequenceconverging to right continuous nondecreasing F so thatlimFn(k)(y) = F (y) at all continuity points of F .

I Limit may not be a distribution function.

I Need a “tightness” assumption to make that the case. Say µnare tight if for every ε we can find an M so thatµn[−M,M] < ε for all n. Define tightness analogously forcorresponding real random variables or distributions functions.

I Theorem: Every subsequential limit of the Fn above is thedistribution function of a probability measure if and only if theFn are tight.

18.175 Lecture 9

Page 34: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Convergence in total variation norm is much stronger thanweak convergence.

18.175 Lecture 9

Page 35: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Convergence in total variation norm is much stronger thanweak convergence.

18.175 Lecture 9

Page 36: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Total variation norm

I If we have two probability measures µ and ν we define thetotal variation distance between them is||µ− ν|| := supB |µ(B)− ν(B)|.

I Intuitively, it two measures are close in the total variationsense, then (most of the time) a sample from one measurelooks like a sample from the other.

I Convergence in total variation norm is much stronger thanweak convergence.

18.175 Lecture 9

Page 37: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 38: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 39: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?I What’s the higher dimensional analog of rolling the tangent

line?

18.175 Lecture 9

Page 40: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?I What’s the higher dimensional analog of rolling the tangent

line?

18.175 Lecture 9

Page 41: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?

I What is the Legendre dual of x2? Of the function equal to 0at 0 and ∞ everywhere else?

I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?I What’s the higher dimensional analog of rolling the tangent

line?

18.175 Lecture 9

Page 42: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?

I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?I What’s the higher dimensional analog of rolling the tangent

line?

18.175 Lecture 9

Page 43: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?I How are derivatives of Λ and Λ∗ related?

I What is the Legendre dual of the Legendre dual of a convexfunction?

I What’s the higher dimensional analog of rolling the tangentline?

18.175 Lecture 9

Page 44: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?

I What’s the higher dimensional analog of rolling the tangentline?

18.175 Lecture 9

Page 45: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Legendre transform

I Define Legendre transform (or Legendre dual) of a functionΛ : Rd → R by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I Let’s describe the Legendre dual geometrically if d = 1: Λ∗(x)is where tangent line to Λ of slope x intersects the real axis.We can “roll” this tangent line around the convex hull of thegraph of Λ, to get all Λ∗ values.

I Is the Legendre dual always convex?I What is the Legendre dual of x2? Of the function equal to 0

at 0 and ∞ everywhere else?I How are derivatives of Λ and Λ∗ related?I What is the Legendre dual of the Legendre dual of a convex

function?I What’s the higher dimensional analog of rolling the tangent

line?18.175 Lecture 9

Page 46: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 47: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Outline

DeMoivre-Laplace limit theorem

Weak convergence

Legendre transform

Large deviations

18.175 Lecture 9

Page 48: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 49: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 50: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 51: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 52: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 53: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 54: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 55: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions

I Let X be a random variable.

I The moment generating function of X is defined byM(t) = MX (t) := E [etX ].

I When X is discrete, can write M(t) =∑

x etxpX (x). So M(t)

is a weighted average of countably many exponentialfunctions.

I When X is continuous, can write M(t) =∫∞−∞ etx f (x)dx . So

M(t) is a weighted average of a continuum of exponentialfunctions.

I We always have M(0) = 1.

I If b > 0 and t > 0 thenE [etX ] ≥ E [et minX ,b] ≥ PX ≥ betb.

I If X takes both positive and negative values with positiveprobability then M(t) grows at least exponentially fast in |t|as |t| → ∞.

18.175 Lecture 9

Page 56: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions for i.i.d. sums

I We showed that if Z = X + Y and X and Y are independent,then MZ (t) = MX (t)MY (t)

I If X1 . . .Xn are i.i.d. copies of X and Z = X1 + . . .+ Xn thenwhat is MZ?

I Answer: MnX .

18.175 Lecture 9

Page 57: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions for i.i.d. sums

I We showed that if Z = X + Y and X and Y are independent,then MZ (t) = MX (t)MY (t)

I If X1 . . .Xn are i.i.d. copies of X and Z = X1 + . . .+ Xn thenwhat is MZ?

I Answer: MnX .

18.175 Lecture 9

Page 58: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Recall: moment generating functions for i.i.d. sums

I We showed that if Z = X + Y and X and Y are independent,then MZ (t) = MX (t)MY (t)

I If X1 . . .Xn are i.i.d. copies of X and Z = X1 + . . .+ Xn thenwhat is MZ?

I Answer: MnX .

18.175 Lecture 9

Page 59: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Large deviations

I Consider i.i.d. random variables Xi . Can we show thatP(Sn ≥ na)→ 0 exponentially fast when a > E [Xi ]?

I Kind of a quantitative form of the weak law of large numbers.The empirical average An is very unlikely to ε away from itsexpected value (where “very” means with probability less thansome exponentially decaying function of n).

18.175 Lecture 9

Page 60: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Large deviations

I Consider i.i.d. random variables Xi . Can we show thatP(Sn ≥ na)→ 0 exponentially fast when a > E [Xi ]?

I Kind of a quantitative form of the weak law of large numbers.The empirical average An is very unlikely to ε away from itsexpected value (where “very” means with probability less thansome exponentially decaying function of n).

18.175 Lecture 9

Page 61: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 62: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 63: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 64: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 65: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.

I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 66: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?

18.175 Lecture 9

Page 67: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

General large deviation principle

I More general framework: a large deviation principle describeslimiting behavior as n→∞ of family µn of measures onmeasure space (X ,B) in terms of a rate function I .

I The rate function is a lower-semicontinuous mapI : X → [0,∞]. (The sets x : I (x) ≤ a are closed — ratefunction called “good” if these sets are compact.)

I DEFINITION: µn satisfy LDP with rate function I andspeed n if for all Γ ∈ B,

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ) ≤ lim sup

n→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x).

I INTUITION: when “near x” the probability density functionfor µn is tending to zero like e−I (x)n, as n→∞.

I Simple case: I is continuous, Γ is closure of its interior.I Question: How would I change if we replaced the measuresµn by weighted measures e(λn,·)µn?

I Replace I (x) by I (x)− (λ, x)? What is infx I (x)− (λ, x)?18.175 Lecture 9

Page 68: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj for i.i.d.

vectors X1,X2, . . . ,Xn in Rd with same law as X .

I Define log moment generating function of X by

Λ(λ) = ΛX (λ) = logMX (λ) = logEe(λ,X ),

where (·, ·) is inner product on Rd .

I Define Legendre transform of Λ by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction Λ∗.

18.175 Lecture 9

Page 69: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj for i.i.d.

vectors X1,X2, . . . ,Xn in Rd with same law as X .

I Define log moment generating function of X by

Λ(λ) = ΛX (λ) = logMX (λ) = logEe(λ,X ),

where (·, ·) is inner product on Rd .

I Define Legendre transform of Λ by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction Λ∗.

18.175 Lecture 9

Page 70: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj for i.i.d.

vectors X1,X2, . . . ,Xn in Rd with same law as X .

I Define log moment generating function of X by

Λ(λ) = ΛX (λ) = logMX (λ) = logEe(λ,X ),

where (·, ·) is inner product on Rd .

I Define Legendre transform of Λ by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction Λ∗.

18.175 Lecture 9

Page 71: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj for i.i.d.

vectors X1,X2, . . . ,Xn in Rd with same law as X .

I Define log moment generating function of X by

Λ(λ) = ΛX (λ) = logMX (λ) = logEe(λ,X ),

where (·, ·) is inner product on Rd .

I Define Legendre transform of Λ by

Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ).

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction Λ∗.

18.175 Lecture 9

Page 72: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Thinking about Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj .

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction

I (x) = Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ),

where Λ(λ) = logM(λ) = Ee(λ,X1).I This means that for all Γ ∈ B we have this asymptotic lower

bound on probabilities µn(Γ)

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ),

so (up to sub-exponential error) µn(Γ) ≥ e−n infx∈Γ0 I (x).I and this asymptotic upper bound on the probabilities µn(Γ)

lim supn→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x),

which says (up to subexponential error) µn(Γ) ≤ e−n infx∈Γ I (x).

18.175 Lecture 9

Page 73: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Thinking about Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj .

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction

I (x) = Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ),

where Λ(λ) = logM(λ) = Ee(λ,X1).

I This means that for all Γ ∈ B we have this asymptotic lowerbound on probabilities µn(Γ)

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ),

so (up to sub-exponential error) µn(Γ) ≥ e−n infx∈Γ0 I (x).I and this asymptotic upper bound on the probabilities µn(Γ)

lim supn→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x),

which says (up to subexponential error) µn(Γ) ≤ e−n infx∈Γ I (x).

18.175 Lecture 9

Page 74: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Thinking about Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj .

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction

I (x) = Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ),

where Λ(λ) = logM(λ) = Ee(λ,X1).I This means that for all Γ ∈ B we have this asymptotic lower

bound on probabilities µn(Γ)

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ),

so (up to sub-exponential error) µn(Γ) ≥ e−n infx∈Γ0 I (x).

I and this asymptotic upper bound on the probabilities µn(Γ)

lim supn→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x),

which says (up to subexponential error) µn(Γ) ≤ e−n infx∈Γ I (x).

18.175 Lecture 9

Page 75: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Thinking about Cramer’s theorem

I Let µn be law of empirical mean An = 1n

∑nj=1 Xj .

I CRAMER’S THEOREM: µn satisfy LDP with convex ratefunction

I (x) = Λ∗(x) = supλ∈Rd

(λ, x)− Λ(λ),

where Λ(λ) = logM(λ) = Ee(λ,X1).I This means that for all Γ ∈ B we have this asymptotic lower

bound on probabilities µn(Γ)

− infx∈Γ0

I (x) ≤ lim infn→∞

1

nlogµn(Γ),

so (up to sub-exponential error) µn(Γ) ≥ e−n infx∈Γ0 I (x).I and this asymptotic upper bound on the probabilities µn(Γ)

lim supn→∞

1

nlogµn(Γ) ≤ − inf

x∈ΓI (x),

which says (up to subexponential error) µn(Γ) ≤ e−n infx∈Γ I (x).18.175 Lecture 9

Page 76: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer upper bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).

I For simplicity, assume that Λ is defined for all x (whichimplies that X has moments of all orders and Λ and Λ∗ arestrictly convex, and the derivatives of Λ and Λ′ are inverses ofeach other). It is also enough to consider the case X hasmean zero, which implies that Λ(0) = 0 is a minimum of Λ,and Λ∗(0) = 0 is a minimum of Λ∗.

I We aim to show (up to subexponential error) thatµn(Γ) ≤ e−n infx∈Γ I (x).

I If Γ were singleton set x we could find the λ correspondingto x , so Λ∗(x) = (x , λ)− Λ(λ). Note then that

Ee(nλ,An) = Ee(λ,Sn) = MnX (λ) = enΛ(λ),

and also Ee(nλ,An) ≥ en(λ,x)µnx. Taking logs and dividingby n gives Λ(λ) ≥ 1

n logµn + (λ, x), so that1n logµn(Γ) ≤ −Λ∗(x), as desired.

I General Γ: cut into finitely many pieces, bound each piece?

18.175 Lecture 9

Page 77: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer upper bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I For simplicity, assume that Λ is defined for all x (which

implies that X has moments of all orders and Λ and Λ∗ arestrictly convex, and the derivatives of Λ and Λ′ are inverses ofeach other). It is also enough to consider the case X hasmean zero, which implies that Λ(0) = 0 is a minimum of Λ,and Λ∗(0) = 0 is a minimum of Λ∗.

I We aim to show (up to subexponential error) thatµn(Γ) ≤ e−n infx∈Γ I (x).

I If Γ were singleton set x we could find the λ correspondingto x , so Λ∗(x) = (x , λ)− Λ(λ). Note then that

Ee(nλ,An) = Ee(λ,Sn) = MnX (λ) = enΛ(λ),

and also Ee(nλ,An) ≥ en(λ,x)µnx. Taking logs and dividingby n gives Λ(λ) ≥ 1

n logµn + (λ, x), so that1n logµn(Γ) ≤ −Λ∗(x), as desired.

I General Γ: cut into finitely many pieces, bound each piece?

18.175 Lecture 9

Page 78: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer upper bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I For simplicity, assume that Λ is defined for all x (which

implies that X has moments of all orders and Λ and Λ∗ arestrictly convex, and the derivatives of Λ and Λ′ are inverses ofeach other). It is also enough to consider the case X hasmean zero, which implies that Λ(0) = 0 is a minimum of Λ,and Λ∗(0) = 0 is a minimum of Λ∗.

I We aim to show (up to subexponential error) thatµn(Γ) ≤ e−n infx∈Γ I (x).

I If Γ were singleton set x we could find the λ correspondingto x , so Λ∗(x) = (x , λ)− Λ(λ). Note then that

Ee(nλ,An) = Ee(λ,Sn) = MnX (λ) = enΛ(λ),

and also Ee(nλ,An) ≥ en(λ,x)µnx. Taking logs and dividingby n gives Λ(λ) ≥ 1

n logµn + (λ, x), so that1n logµn(Γ) ≤ −Λ∗(x), as desired.

I General Γ: cut into finitely many pieces, bound each piece?

18.175 Lecture 9

Page 79: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer upper bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I For simplicity, assume that Λ is defined for all x (which

implies that X has moments of all orders and Λ and Λ∗ arestrictly convex, and the derivatives of Λ and Λ′ are inverses ofeach other). It is also enough to consider the case X hasmean zero, which implies that Λ(0) = 0 is a minimum of Λ,and Λ∗(0) = 0 is a minimum of Λ∗.

I We aim to show (up to subexponential error) thatµn(Γ) ≤ e−n infx∈Γ I (x).

I If Γ were singleton set x we could find the λ correspondingto x , so Λ∗(x) = (x , λ)− Λ(λ). Note then that

Ee(nλ,An) = Ee(λ,Sn) = MnX (λ) = enΛ(λ),

and also Ee(nλ,An) ≥ en(λ,x)µnx. Taking logs and dividingby n gives Λ(λ) ≥ 1

n logµn + (λ, x), so that1n logµn(Γ) ≤ −Λ∗(x), as desired.

I General Γ: cut into finitely many pieces, bound each piece?

18.175 Lecture 9

Page 80: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer upper bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I For simplicity, assume that Λ is defined for all x (which

implies that X has moments of all orders and Λ and Λ∗ arestrictly convex, and the derivatives of Λ and Λ′ are inverses ofeach other). It is also enough to consider the case X hasmean zero, which implies that Λ(0) = 0 is a minimum of Λ,and Λ∗(0) = 0 is a minimum of Λ∗.

I We aim to show (up to subexponential error) thatµn(Γ) ≤ e−n infx∈Γ I (x).

I If Γ were singleton set x we could find the λ correspondingto x , so Λ∗(x) = (x , λ)− Λ(λ). Note then that

Ee(nλ,An) = Ee(λ,Sn) = MnX (λ) = enΛ(λ),

and also Ee(nλ,An) ≥ en(λ,x)µnx. Taking logs and dividingby n gives Λ(λ) ≥ 1

n logµn + (λ, x), so that1n logµn(Γ) ≤ −Λ∗(x), as desired.

I General Γ: cut into finitely many pieces, bound each piece?18.175 Lecture 9

Page 81: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer lower bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).

I We aim to show that asymptotically µn(Γ) ≥ e−n infx∈Γ0 I (x).

I It’s enough to show that for each given x ∈ Γ0, we have thatasymptotically µn(Γ) ≥ e−nI (x).

I Idea is to weight law of each Xi by e(λ,x) to get a newmeasure whose expectation is in the interior of x . In this newmeasure, An is “typically” in Γ for large Γ, so the probability isof order 1.

I But by how much did we have to modify the measure to makethis typical? Aren’t we weighting the law of An by aboute−nI (x) near x?

18.175 Lecture 9

Page 82: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer lower bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I We aim to show that asymptotically µn(Γ) ≥ e−n infx∈Γ0 I (x).

I It’s enough to show that for each given x ∈ Γ0, we have thatasymptotically µn(Γ) ≥ e−nI (x).

I Idea is to weight law of each Xi by e(λ,x) to get a newmeasure whose expectation is in the interior of x . In this newmeasure, An is “typically” in Γ for large Γ, so the probability isof order 1.

I But by how much did we have to modify the measure to makethis typical? Aren’t we weighting the law of An by aboute−nI (x) near x?

18.175 Lecture 9

Page 83: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer lower bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I We aim to show that asymptotically µn(Γ) ≥ e−n infx∈Γ0 I (x).

I It’s enough to show that for each given x ∈ Γ0, we have thatasymptotically µn(Γ) ≥ e−nI (x).

I Idea is to weight law of each Xi by e(λ,x) to get a newmeasure whose expectation is in the interior of x . In this newmeasure, An is “typically” in Γ for large Γ, so the probability isof order 1.

I But by how much did we have to modify the measure to makethis typical? Aren’t we weighting the law of An by aboute−nI (x) near x?

18.175 Lecture 9

Page 84: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer lower bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I We aim to show that asymptotically µn(Γ) ≥ e−n infx∈Γ0 I (x).

I It’s enough to show that for each given x ∈ Γ0, we have thatasymptotically µn(Γ) ≥ e−nI (x).

I Idea is to weight law of each Xi by e(λ,x) to get a newmeasure whose expectation is in the interior of x . In this newmeasure, An is “typically” in Γ for large Γ, so the probability isof order 1.

I But by how much did we have to modify the measure to makethis typical? Aren’t we weighting the law of An by aboute−nI (x) near x?

18.175 Lecture 9

Page 85: 18.175: Lecture 9 .1in More large deviationsmath.mit.edu/~sheffield/2016175/Lecture9.pdf18.175: Lecture 9 More large deviations Scott She eld MIT 18.175 Lecture 9

Proving Cramer lower bound

I Recall that I (x) = Λ∗(x) = supλ∈Rd(λ, x)− Λ(λ).I We aim to show that asymptotically µn(Γ) ≥ e−n infx∈Γ0 I (x).

I It’s enough to show that for each given x ∈ Γ0, we have thatasymptotically µn(Γ) ≥ e−nI (x).

I Idea is to weight law of each Xi by e(λ,x) to get a newmeasure whose expectation is in the interior of x . In this newmeasure, An is “typically” in Γ for large Γ, so the probability isof order 1.

I But by how much did we have to modify the measure to makethis typical? Aren’t we weighting the law of An by aboute−nI (x) near x?

18.175 Lecture 9