95
Lecture 1: Concepts of probability Prof. Dr. Svetlozar Rachev Institute for Statistics and Mathematical Economics University of Karlsruhe Portfolio and Asset Liability Management Summer Semester 2008 Prof. Dr. Svetlozar Rachev (University of Karlsruhe) Lecture 1: Concepts of probability 2008 1 / 95

Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Lecture 1: Concepts of probability

Prof. Dr. Svetlozar Rachev

Institute for Statistics and Mathematical EconomicsUniversity of Karlsruhe

Portfolio and Asset Liability Management

Summer Semester 2008

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 1 / 95

Page 2: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copyright

These lecture-notes cannot be copied and/or distributed withoutpermission.The material is based on the text-book:Svetlozar T. Rachev, Stoyan Stoyanov, and Frank J. FabozziAdvanced Stochastic Models, Risk Assessment, and PortfolioOptimization: The Ideal Risk, Uncertainty, and PerformanceMeasuresJohn Wiley, Finance, 2007

Prof. Svetlozar (Zari) T. RachevChair of Econometrics, Statisticsand Mathematical FinanceSchool of Economics and Business EngineeringUniversity of KarlsruheKollegium am Schloss, Bau II, 20.12, R210Postfach 6980, D-76128, Karlsruhe, GermanyTel. +49-721-608-7535, +49-721-608-2042(s)Fax: +49-721-608-3811http://www.statistik.uni-karslruhe.de

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 2 / 95

Page 3: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Basic concepts

An outcome for a random variable is the mutually exclusivepotential result that can occur. The accepted notation for anoutcome is the Greek letter ω.

A sample space is a set of all possible outcomes. The samplespace is denoted by Ω. The fact that a given outcome ωi belongsto the sample space is expressed by ωi ∈ Ω.

An event is a subset of the sample space and can be representedas a collection of some of the outcomes.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 3 / 95

Page 4: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Basic concepts

Example 1Consider Microsoft’s stock return over the next year.The sample space contains outcomes ranging from - 100% (all thefunds invested in Microsoft’s stock will be lost) to an extremely highpositive return.The sample space can be partitioned into two subsets: outcomeswhere the return is less than or equal to 10% and a subset where thereturn exceeds 10%. Consequently, a return greater than 10% is anevent since it is a subset of the sample space.

Example 2A 1-month LIBOR three months from now that exceeds 4% is an event.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 4 / 95

Page 5: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Basic concepts

The collection of all events is usually denoted by A.

In the theory of probability, we consider the sample space Ωtogether with the set of events A, usually written as (Ω, A),because the notion of “probability” is associated with an event.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 5 / 95

Page 6: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Discrete probability distributions

A discrete random variable limits the outcomes where the variablecan only take on discrete values.

ExampleConsider the default of a corporation on its debt obligations over thenext five years. This random variable has only two possible outcomes:default or nondefault. Hence, it is a discrete random variable.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 6 / 95

Page 7: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Discrete probability distributions

Consider an option contract where for an upfront payment (i.e.,the option price) of $50,000, the buyer of the contract receives thepayment given in Table below from the seller of the optiondepending on the return on the S&P 500 index.

In this case, the random variable is a discrete random variable buton the limited number of outcomes.

If S&P 500 return is Payment receivedLess than or equal to zero $0Greater than zero but less than 5% $10,000Greater than 5% but less than 10% $20,000Greater than or equal to 10% $100,000

Table: Option payments depending on the value of the S&P500 index.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 7 / 95

Page 8: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Bernoulli distribution

A random variable X is called Bernoulli-distributed with parameterp if it has only two possible outcomes, usually encoded as “1”(which might represent “success” or “default”) or “0” (which mightrepresent “failure” or “survival”).

One classical example for a Bernoulli distributed random variableoccurring in the field of finance is the default event of a company.We observe a company C in a specified time interval I, e.g.,January 1, 2007, until December 31, 2007. We define

X =

1 if C defaults in I0 else

The parameter p in this case would be the annualized probabilityof default of company C.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 8 / 95

Page 9: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Binomial distribution

We usually do not consider only one company but a whole basketC1, . . . , Cn of companies. Assuming that all these n companieshave the same annualized probability of default p, this leads to ageneralization of the Bernoulli distribution, called binomialdistribution.

A binomial distributed random variable Y with parameters n and pis obtained as the sum of n independent and identicallyBernoulli-distributed random variables X1, . . . , Xn.

In our example, Y represents the total number of defaultsoccurring in the year 2007 observed for companies C1, . . . , Cn.Given the two parameters, the probability of observing k ,0 ≤ k ≤ n defaults can be explicitly calculated as follows:

P(Y = k) =

(

nk

)

pk (1 − p)n−k where(

nk

)

=n!

(n − k)!k !

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 9 / 95

Page 10: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Poisson distribution

The Poisson distribution depends upon only one parameter λ andcan be interpreted as an approximation to the binomial distributionwhen the parameter p is a small number.

A Poisson-distributed random variable is usually used to describethe random number of events (number of defaults in our example)occurring over a certain time interval.

The parameter λ indicates the rate of occurrence of the randomevents, i.e., it tells us how many events occur on average per unitof time.

The probability distribution of a Poisson-distributed randomvariable N is described by the following equation:

P(N = k) =λk

k !e−λ, k = 0, 1, 2, . . .

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 10 / 95

Page 11: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Continuous probability distributions

If the random variable can take on any possible value within therange of outcomes, then the probability distribution is said to be acontinuous random variable.

When a random variable is either the price of or the return on afinancial asset or an interest rate, the random variable is assumedto be continuous.

For a continuous random variable, the calculation of probabilitiesis substantially different from the discrete case. The reason is thatif we want to derive the probability that the realization of therandom variable lays within some range (i.e., over a subset orsubinterval of the sample space), then we cannot proceed in asimilar way as in the discrete case: The number of values in aninterval is so large, that we cannot just add the probabilities of thesingle outcomes.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 11 / 95

Page 12: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

A probability distribution function P assigns a probability P(A) forevery event A, i.e., of realizing a value for the random value in anyspecified subset A of the sample space. For example, aprobability distribution function can assign a probability of realizinga monthly return that is negative or the probability of realizing amonthly return that is greater than 0.5% or the probability ofrealizing a monthly return that is between 0.4% and 1.0%.

To compute the probability, a mathematical function is needed torepresent the probability distribution function. There are severalpossibilities of representing a probability distribution by means ofa mathematical function. In the case of a continuous probabilitydistribution, the most popular way is to provide the so-calledprobability density function or simply density function.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 12 / 95

Page 13: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

In general, we will denote the density function for the randomvariable X as fX (x). Note that the letter x is used for the functionargument and the index denotes that the density functioncorresponds to the random variable X . The letter x is theconvention adopted to denote a particular value for the randomvariable.

The density function of a probability distribution is alwaysnon-negative and as its name indicates: Large values for fX (x) ofthe density function at some point x imply a relatively highprobability of realizing a value in the neighborhood of x , whereasfX (x) = 0 for all x in some interval (a, b) implies that theprobability for observing a realization in (a, b) is zero.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 13 / 95

Page 14: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

0

0.05

0.1

0.15

0.2

0.25

0.3

0.35

0.4

x

f X(x

)

a b

Figure: The probability of the event that a given random variable X isbetween two real numbers a and b is equal to the shaded area under thedensity function fX (x).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 14 / 95

Page 15: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

The shaded area is the probability of realizing a return less than band greater than a. As probabilities are represented by areasunder the density function, it follows that the probability for everysingle outcome of a continuous random variable always equals 0.

Computing the probability involves the integral calculus,calculating the area under a curve. Thus, the probability that arealization from a random variable is between two real numbers aand b is calculated according to the formula,

P(a ≤ X ≤ b) =

∫ b

afX (x)dx .

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 15 / 95

Page 16: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

The mathematical function that provides the cumulative probabilityof a probability distribution, i.e., the function which assigns toevery real value x the probability of getting an outcome less thanor equal to x , is called the cumulative distribution function orcumulative probability function or simply distribution function andis denoted mathematically by FX (x).

A cumulative distribution function is always non-negative,non-decreasing, and as it represents probabilities it takes onlyvalues between zero and one. An example of a distributionfunction is given on the next slide.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 16 / 95

Page 17: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

0

1

x

FX(x

)

a b

FX(a)

FX(b)

Figure: The probability of the event that a given random variable X isbetween two real numbers a and b is equal to the difference FX (b) − FX (a).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 17 / 95

Page 18: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

The mathematical connection between a probability densityfunction f , a probability distribution P, and a cumulativedistribution function F of some random variable X is given by:

P(X ≤ t) = FX (t) =

∫ t

−∞

fX (x)dx .

The density equals the first derivative of the distribution function,

fX (x) =dFX (x)

dx.

The CDF is another way to uniquely characterize an probabilitydistribution on the set of real numbers. The probability that therandom variable is between two real numbers a and b is given by

P(a < X ≤ b) = FX (b) − FX (a).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 18 / 95

Page 19: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probability distribution function, probability densityfunction, and cumulative distribution function

Not all distribution functions are continuous and differentiable.

Sometimes, a distribution function may have a jump for somevalue of the argument, or it can be composed of only jumps andflat sections. Such are the distribution functions of a discreterandom variable for example.

Figure on the next slide illustrates a more general case in whichFX (x) is differentiable except for the point x = a where there is ajump. It is often said that the distribution function has a point massat x = a because the value a happens with non-zero probability incontrast to the other outcomes, x 6= a. In fact, the probability that aoccurs is equal to the size of the jump of the distribution function.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 19 / 95

Page 20: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

0 0

1

x

FX(x

)

Jump size

a

Figure: A distribution function FX (x) with a jump at x = a.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 20 / 95

Page 21: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

The normal distribution

The class of normal distributions, or Gaussian distributions, iscertainly one of the most important probability distributions instatistics and due to some of its appealing properties also theclass which is used in most applications in finance. Here weintroduce some of its basic properties.

The random variable X is said to be normally distributed withparameters µ and σ, abbreviated by X ∈ N(µ, σ2), if the density ofthe random variable is given by the formula,

fX (x) =1√

2πσ2e−

(x−µ)2

2σ2 , x ∈ R.

The parameter µ is called a location parameter because themiddle of the distribution equals µ and σ is called a shapeparameter or a scale parameter. If µ = 0 and σ = 1, then X issaid to have a standard normal distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 21 / 95

Page 22: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

The normal distribution

An important property of the normal distribution is the so-calledlocation-scale invariance of the normal distribution.

Imagine you have random variable X which is normally distributedwith the parameters µ and σ. Now we consider the randomvariable Y which is obtained as Y = aX + b.

In general, the distribution of Y might substantially differ from thedistribution of X but in the case where X is normally distributed,the random variable Y is again normally distributed withparameters and µ = aµ + b and σ = aσ.

ExampleImagine that X measures the temperature at the top of the EmpireState Building on January 1, 2008, at 6 A.M. in degrees Celsius.Then Y = 9

5X + 32 will give the temperature in degrees Fahrenheit,and if X is normally distributed then Y will be too.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 22 / 95

Page 23: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

The normal distribution

Another interesting and important property of normal distributionsis their summation stability.

If you take the sum of several independent random variables whichare all normally distributed with location parameters µi and scaleparameters σi , then the sum again will be normally distributed.

The two parameters of the resulting distribution are obtained as

µ = µ1 + µ2 + . . . + µn

σ =√

σ21 + σ2

2 + . . . + σ2n

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 23 / 95

Page 24: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

The normal distribution

The normal distribution possesses a domain of attraction.

A mathematical result called the central limit theorem states that -under certain technical conditions - the distribution of a large sumof random variables behaves necessarily like a normal distribution.

The normal distribution is the class of stable distributions which isunique in the sense that a large sum of random variables can onlyconverge to a stable distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 24 / 95

Page 25: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Exponential distribution

The exponential distribution is often used to model waiting times inthe queuing theory.

It is concentrated on the positive real numbers and the densityfunction f and the cumulative distribution function F of anexponentially distributed random variable τ possess the followingform:

fτ (x) =1β

e−xβ , x > 0

Fτ (x) = 1 − e−xβ , x > 0

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 25 / 95

Page 26: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Exponential distribution

In credit risk modeling, the parameter λ = 1/β has a naturalinterpretation as hazard rate or default intensity.

Let τ denote an exponential distributed random variable, e.g., therandom time (counted in days and started on January 1, 2008) wehave to wait until Ford Motor Company defaults. Consider thefollowing expression:

λ(∆t) =P(τ ∈ (t , t + ∆t ]|τ > t)

∆t=

P(τ ∈ (t , t + ∆t ])∆tP(τ > t)

where ∆t denotes a small period of time.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 26 / 95

Page 27: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Exponential distribution

λ(∆t) represents a ratio of a probability and the quantity ∆t . Theprobability in the numerator represents the probability that defaultoccurs in the time interval (t , t + ∆t ] conditional upon the fact thatFord Motor Company survives until time t .

The ratio of the obtained probability and the length of theconsidered time interval can be denoted as a default rate ordefault intensity.

Now, letting ∆t tend to zero we finally obtain after some calculusthe desired relation λ = 1/β.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 27 / 95

Page 28: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Exponential distribution

There is the link between the Exponential and Poisson distributions:

Consider a sequence of independent and identical exponentiallydistributed random variables τ1, τ2, . . .

We can think of τ1, for example, as the time we have to wait until afirm in a high-yield bond portfolio defaults. τ2 will then representthe time between the first and the second default and so on.These waiting times are sometimes called interarrival times.

Now, let Nt denote the number of defaults which have occurreduntil time t ≥ 0. One important probabilistic result states that therandom variable Nt is Poisson distributed with parameter λ = t/β.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 28 / 95

Page 29: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Student’s t distribution

Student’s t distributions (or the t-distributions), are used in financeas probabilistic models of assets returns. The density function ofthe t-distribution is given by the following equation:

fX (x) =1√πn

Γ((n + 1)/2)

Γ(n/2)

(

1 +x2

n

)−n+1

2

, x ∈ R

where n is an integer valued parameter called degree of freedom.

For large values of n, the t-distribution doesn’t significantly differfrom a standard normal distribution. Usually, for values n > 30, thet-distribution is considered as equal to the standard normaldistribution

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 29 / 95

Page 30: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Extreme value distribution

The extreme value distribution (also denoted as Gumbel-typeextreme value distribution) occurs as the limit distribution of thelargest observation in a sample of increasing size. That’s why it isoften used in operational risk applications

Its density function f and distribution function F , respectively, isgiven by the following equations:

fX (x) =1b

e−x−a

b −e−x−a

b , x ∈ R

FX (x) = e−e−x−a

b , x ∈ R

where a denotes a real location parameter and b > 0 a positivereal shape parameter.

The class of extreme value distributions forms a location-scalefamily.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 30 / 95

Page 31: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Generalized extreme value distribution

Besides the previously mentioned (Gumbel type) extreme valuedistribution, there are two other types of distributions:

Weibull-type extreme value distributionFréchet-type extreme value distribution

All three can be represented as a three parameter distributionfamily referred to as a generalized extreme value distribution withthe following cumulative distribution function:

FX (x) = e−(1+ξ x−µσ

)−1/ξ, 1 + ξ

x − µ

σ> 0

where ξ and µ are real and σ is a positive real parameter.

If ξ tends to zero, we obtain the extreme value distributiondiscussed above. For positive values of ξ the distribution is calledFrechet-type and for negative values of ξ Weibull-type extremevalue distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 31 / 95

Page 32: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments and quantiles

In describing a probability distribution function, it is common tosummarize it by using various measures. The five most commonlyused measures are:

location

dispersion

asymmetry

concentration in tails

quantiles

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 32 / 95

Page 33: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Location

To find the central value or location several measures are used:the mean or average value, the median, or the mode.

The relationship among these three measures of location dependson the skewness of a probability distribution function.

The most commonly used measure of location is the mean and isdenoted by µ or EX or E(X ).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 33 / 95

Page 34: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Dispersion

Various measures of dispersion (or how spread out the values ofthe random variable) are the range, variance, and mean absolutedeviation.

The most commonly used measure is the variance (denoted byσ2). It measures the dispersion of the values that the randomvariable can realize relative to the mean. It is the average of thesquared deviations from the mean. Taking the square root of thevariance one obtains the standard deviation (denoted by σ).

In contrast to the variance, the mean absolute deviation takes theaverage of the absolute deviations from the mean.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 34 / 95

Page 35: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Asymmetry

A probability distribution may be symmetric or asymmetric aroundits mean. A popular measure for the asymmetry of a distribution iscalled its skewness.

A negative skewness measure indicates that the distribution isskewed to the left; that is, compared to the right tail, the left tail iselongated. A positive skewness measure indicates that thedistribution is skewed to the right; that is, compared to the left tail,the right tail is elongated.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 35 / 95

Page 36: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Positive skewnessNegative skewness

Figure: The density graphs of a positively and a negatively skeweddistribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 36 / 95

Page 37: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Concentration in tails

Additional information about a probability distribution function isprovided by measuring the concentration (mass) of potentialoutcomes in its tails.

The tails of a probability distribution function contain the extremevalues (e.g. financial ruin).

The fatness of the tails of the distribution is related to thepeakedness of the distribution around its mean or center. The jointmeasure of peakedness and tail fatness is called kurtosis.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 37 / 95

Page 38: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments

The four measures described above are called statistical moments orsimply moments:

The first moment - the mean, also referred to as the expectedvalue.

The variance is the second central moment.

Skewness is a rescaled is third central moment.

Kurtosis is a rescaled fourth central moment.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 38 / 95

Page 39: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments

Parameter Discrete distribution Continuous distribution

Mean EX =∑

i

xiP(X = xi) EX =

−∞

xfX (x)dx

Variance σ2 = E(X − EX )2 σ2 = E(X − EX )2

Skewness ζ =E(X − EX )3

(σ2)3/2ζ =

E(X − EX )3

(σ2)3/2

Kurtosis κ =E(X − EX )4

(σ2)4 κ =E(X − EX )4

(σ2)4

Table: General formula for parameters

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 39 / 95

Page 40: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments

The skewness measure reported in table above is the so-calledFisher’s skewness. Another possible way to define the measure isthe Pearson’s skewness which equals the square of the Fisher’sskewness.

The same holds true for the kurtosis, where we have reported thePearson’s kurtosis. Fishers’ kurtosis (sometimes denoted asexcess kurtosis) can be obtained by subtracting three fromPearson’s kurtosis.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 40 / 95

Page 41: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments

Generally, the moment of order n of a random variable is denotedby µn defined as

µn = EX n,

where n = 1, 2, . . . For a discrete probability distribution, themoment of order k is calculated according to the formula

µn =∑

i

xni P(X = xi),

and in the case of a continuous probability distribution, the formulais

µn =

−∞

xnfX (x)dx .

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 41 / 95

Page 42: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Statistical moments

The centered moment of order n is denoted by mn and is definedas

mn = E(X − EX )n

where n = 1, 2, . . .. For a discrete probability distribution, thecentered moment of order n is calculated according to the formula

mn =∑

i

(xi − EX )nP(X = xi),

and in the case of a continuous probability distribution, the formulais

mn =

−∞

(x − EX )nfX (x)dx .

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 42 / 95

Page 43: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Quantiles

To describe a probability distribution sometimes a concept calledα-quantile is used. It gives us information where the first α% ofthe distribution are located.

Given an arbitrary observation of the considered probabilitydistribution, this observation will be smaller than the α-quantile qα

in α% of the cases and larger in (100 − α)% of the cases.

The 25%-, 50%- and 75%-quantile are referred to as the firstquartile, second quartile, and third quartile, respectively. The 1%-,2%-, , 98%-, 99%-quantiles are called percentiles.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 43 / 95

Page 44: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Sample moments

In practical applications we are faced with the situation that weobserve realizations of a probability distribution (e.g., the dailyreturn of the S&P 500 index over the last two years), but we don’tknow the distribution which generates these returns.

Consequently we are not able to calculate the statistical moments.But, having the observations r1, . . . , rk , we can try to estimate the“true moments” (sample moments) out of the sample.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 44 / 95

Page 45: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Sample moments

The empirical analogue for the mean of a random variable is theaverage of the observations:

EX ≈ 1k

k∑

i=1

ri

For large n it is reasonable to expect that the average of theobservations will not be far from the mean of the probabilitydistribution.

All theoretical formulae for the calculation of the four statisticalmoments are expressed as “means of something”.

This simple and intuitive idea is based on a fundamental result inthe theory of probability known as the law of large numbers. Thisresult, together with the central limit theorem, forms the basics ofthe theory of statistics.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 45 / 95

Page 46: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Moment Sample moment

Mean r =1k

k∑

i=1

ri

Variance s2 =1k

k∑

i=1

(ri − r)2

Skewness ζ =1k

∑ki=1(ri − r)3

(s2)3/2

Kurtosis κ =1k

∑ki=1(ri − r)4

(s2)2

Table: Calculation of sample moments

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 46 / 95

Page 47: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Joint probability distributions

Here we move from the probability distribution of a single randomvariable (univariate distribution) to that of multiple randomvariables (multivariate distribution).

Understanding multivariate distributions is important becausefinancial theories such as portfolio selection theory andasset-pricing theory involve distributional properties of sets ofinvestment opportunities.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 47 / 95

Page 48: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Conditional probability

A useful concept in understanding the relationship betweenmultiple random variables is that of conditional probability.

Consider the returns on the stocks of two companies in one andthe same industry. The future return X on the stocks of company1 is not unrelated to the future return Y on the stocks of company2 because the future development of the two companies is drivento some extent by common factors since they are in one and thesame industry.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 48 / 95

Page 49: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Conditional probability

It is a reasonable question to ask what is the probability that thefuture return X is smaller than a given percentage, e.g. X ≤ −2%,on condition that Y realizes a huge loss, e.g. Y ≤ −10%?

Essentially, the conditional probability is calculating the probabilityof an event provided that another event happens. If we denote thefirst event by A and the second event by B, then the conditionalprobability of A provided that B happens, denoted by P(A|B), isgiven by the formula,

P(A|B) =P(A ∩ B)

P(B),

which is also known as the Bayes formula. According to theformula, we divide the probability that both events A and B occursimultaneously, denoted by A∩B, by the probability of the event B.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 49 / 95

Page 50: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Conditional probability

In the two-stock example, the formula is applied in the followingway,

P(X ≤ −2%|Y ≤ −10%) =P(X ≤ −2%, Y ≤ −10%)

P(Y ≤ −10%). (1)

Thus, in order to compute the conditional probability, we have tobe able to calculate the quantity

P(X ≤ −2%, Y ≤ −10%)

which represents the joint probability of the two events.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 50 / 95

Page 51: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Definition of joint probability distributions

In real-world applications the interest is in the joint probabilitydistribution or joint distribution of more than one random variable.

For example, suppose that a portfolio consists of a position in twoassets, asset 1 and asset 2. Then there will be a probabilitydistribution for (1) asset 1, (2) asset 2, and (3) asset 1 and asset 2.

The first two distributions are referred to as the marginalprobability distributions or marginal distributions. The distributionfor asset 1 and asset 2 is called the joint probability distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 51 / 95

Page 52: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Definition of joint probability distributions

Like in the univariate case, there is a mathematical connectionbetween the probability distribution P, the cumulative distributionfunction F , and the density function f of a multivariate randomvariable (also called a random vector ) X = (X1, . . . , Xn):

P(X1 ≤ t1, . . . , Xn ≤ tn) = FX (t1, . . . , tn)

=

∫ t1

−∞

. . .

∫ tn

−∞

fX (x1, . . . , xn)dx1 . . . dxn

The joint probability that the first random variable realizes a valueless than or equal to t1 and the second less than or equal to t2 andso on is given by the cumulative distribution function F .

The value can be obtained by calculating “the volume” under thedensity function f . Because there are n random variables, wehave now n arguments for both functions: the density function andthe cumulative distribution function.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 52 / 95

Page 53: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Marginal distributions

It is also possible to express the density function in terms of thedistribution function by computing sequentially the first-orderpartial derivatives of the distribution function with respect to allvariables,

fX (x1, . . . , xn) =∂nFX (x1, . . . , xn)

∂x1 . . . ∂xn. (2)

Beside this joint distribution, we can consider the abovementioned marginal distributions, i.e., the distribution of one singlerandom variable Xi . The marginal density fi of Xi is obtained byintegrating the joint density over all variables which are not takeninto consideration:

fXi (x) =

−∞

. . .

−∞

fX (x1, . . . , xi−1, x , xi+1, . . . , xn)dx1 . . . dxi−1dxi+1 . . . dxn

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 53 / 95

Page 54: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Dependence of random variables

Considering multivariate distributions we are faced with inferencebetween the distributions; that is, large values of one randomvariable imply large values of another random variable or smallvalues of a third random variable. This property is denoted as thedependence of random variables.

The inverse case of no dependence is denoted as stochasticindependence. More precisely, two random variables areindependently distributed if and only if their joint distribution givenin terms of the joint cumulative distribution function F or the jointdensity function f equals the product of their marginaldistributions:

FX (x1, . . . , xn) = FX1(x1) . . . FXn(xn)

fX (x1, . . . , xn) = fX1(x1) . . . fXn(xn)

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 54 / 95

Page 55: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Dependence of random variables

In the special case of n = 2, we can say that two random variablesare said to be independently distributed, if knowing the value ofone random variable does not provide any information about theother random variable.

For instance, if we assume that the two events X ≤ −2% andY ≤ −10% are independent, then the conditional probability inequation (1) equals

P(X ≤ −2%|Y ≤ −10%) =P(X ≤ −2%)P(Y ≤ −10%)

P(Y ≤ −10%)

= P(X ≤ −2%).

Indeed, under the assumption of independence, the eventY ≤ −10% has no influence on the probability of the other event.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 55 / 95

Page 56: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Covariance and correlation

There are two strongly related measures among many that arecommonly used to measure how two random variables tend tomove together, the covariance and the correlation. Letting

σX denote the standard deviation of XσY denote the standard deviation of YσXY denote the covariance between X and YρXY denote the correlation between X and Y

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 56 / 95

Page 57: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Covariance and correlation

The relationship between the correlation, which is also denoted byρXY = corr(X , Y ), and covariance is as follows:

ρXY =σXY

σX σY.

Here the covariance, also denoted by σXY = cov(X , Y ), is definedas,

σXY = E(X − EX )(Y − EY )

= E(XY ) − EXEY

It can be shown that the correlation can only have values from −1to +1. When the correlation is zero, the two random variables aresaid to be uncorrelated.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 57 / 95

Page 58: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Covariance and correlation

If we add two random variables, X + Y , the expected value (firstcentral moment) is simply the sum of the expected value of thetwo random variables. That is,

E(X + Y ) = EX + EY .

The variance of the sum of two random variables, denoted byσ2

X+Y , is

σ2X+Y = σ2

X + σ2Y + 2σXY

Here the last term accounts for the fact that there might be adependence between X and Y measured through the covariance.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 58 / 95

Page 59: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

In finance, it is common to assume that the random variables arenormally distributed. The joint distribution is then referred to as amultivariate normal distribution.

Consider first n independent standard normal random variablesX1, . . . , Xn. Their common density function can be written as theproduct of their individual density functions and so we obtain thefollowing expression as the density function of the random vectorX = X1, . . . , Xn:

fX (x1, . . . , xn) =1

(√

2π)ne−

x′x2

where the vector notation x ′x denotes the sum of the componentsof the vector x raised to the second power, x ′x =

∑ni=1 x2

i .

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 59 / 95

Page 60: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

Now consider n vectors with n real components arranged in amatrix A. In this case, it is often said that the matrix A has a n × ndimension.

The random variable

Y = AX + µ, (3)

in which AX denotes the n × n matrix A multiplied by the randomvector X and µ is a vector of n constants, has a generalmultivariate normal distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 60 / 95

Page 61: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

The density function of Y can now be expressed as

fY (y1, . . . , yn) =1

(π|Σ|)n/2e−

(y−µ)′Σ−1(y−µ)2

where |Σ| denotes the determinant of the matrix Σ and Σ−1

denotes the inverse of Σ.

The matrix Σ can be calculated from the matrix A, Σ = AA′. Theelements of Σ = σijn

i,j=1 are the covariances between thecomponents of the vector Y ,

σij = cov(Yi , Yj).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 61 / 95

Page 62: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

−20

2

−2

0

2

0

0.1

0.2

0.3

y1

y2

f Y(y

)

Figure: The probability density function of a two-dimensional normaldistribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 62 / 95

Page 63: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

The figure above contains the probability density function of atwo-dimensional normal distribution with a covariance matrix,

Σ =

(

1 0.80.8 1

)

and mean µ = (0, 0).The matrix A from the representation given in formula (3) equals

A =

(

1 00.8 0.6

)

.

The correlation between the two components of the random vectorY is equal to 0.8, corr(Y1, Y2) = 0.8, because in this example thevariances of the two components are equal to 1. This is a strongpositive correlation which means that the realizations of therandom vector Y will cluster along the diagonal splitting the firstand the third quadrant.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 63 / 95

Page 64: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

0.05

0.05

0.05

0.05

0.05

0.1

0.1

0.1

0.1

0.150.15

0.15

0.2

0.2

0.25

y1

y 2

−3 −2 −1 0 1 2 3−3

−2

−1

0

1

2

3

Figure: The level lines of the two-dimensional probability density functionplotted on slide 62.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 64 / 95

Page 65: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Multivariate normal distribution

The contour lines on the previous slide are ellipses centered at themean µ = (0, 0) of the random vector Y with their major axes lyingalong the diagonal of the first quadrant.

The contour lines indicate that realizations of the random vector Yroughly take the form of an elongated ellipse as the ones shown inFigure 6 which means that large values of Y1 will correspond tolarge values of Y2 in a given pair of observations.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 65 / 95

Page 66: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Elliptical distributions

A generalization of the multivariate normal distribution is given bythe class of elliptical distributions.

Simply speaking, a n-dimensional random vector X with densityfunction f is called spherically distributed if all the level curves, i.e.,the set of all points where the density function f admits a certainvalue c, possesses the form of a sphere.

In the special case when n = 2, the density function can beplotted and the level curves look like circles. Analogously, an-dimensional random vector X with density function f is calledelliptically distributed if the form of all level curves equals the oneof an ellipse.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 66 / 95

Page 67: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Elliptical distributions

One can think of elliptical distributions as a special class ofsymmetric distributions which possess a number of desirableproperties.

Examples of elliptically distributed random variables include allmultivariate normal distributions, multivariate t-distributions,logistic distributions, Lapace distributions, and a part of themultivariate stable distributions.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 67 / 95

Page 68: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Elliptical distributions

Elliptical distributions with existing density function can bedescribed by a triple (µ,Σ, g), where µ and Σ play similar roles asthe mean vector and the variance-covariance matrix in themultivariate normal setting. The function g is the so-called densitygenerator.

All three together define the density function of the distribution as:

fX (x) =c

|Σ|g((x − µ)′Σ−1(x − µ))

where c is a normalizing constant.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 68 / 95

Page 69: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

A major drawback of correlation is that it is not invariant undernonlinear strictly increasing transformations. In general

corr(T (X ), T (Y )) 6= corr(X , Y ),

where T (x) is such transformation.

One example which explains this technical requirement is thefollowing:

Assume that X and Y represent the continuous return (log-return)of two assets over the period [0, t ], where t denotes some point oftime in the future.

If you know the correlation of these two random variables, thisdoes not imply that you know the dependence structure betweenthe asset prices itself.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 69 / 95

Page 70: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

The asset prices (P and Q for asset X and Y , respectively) areobtained by Pt = P0 exp(X ) and Qt = Q0 exp(Y ), where P0 andQ0 denote the corresponding asset prices at time 0.

The asset prices are strictly increasing functions of the return butthe correlation structure is not maintained by this transformation.

This observation implies that the return could be uncorrelatedwhereas the prices are strongly correlated and vice versa.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 70 / 95

Page 71: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

A more prevalent approach which overcomes this disadvantage isto model dependency using copulas.

As noted by Patton (2004): “The word copula comes from Latin fora “link” or “bond”, and was coined by Sklar (1959), who first provedthe theorem that a collection of marginal distributions can be‘coupled’ together via a copula to form a multivariate distribution.”

The description of the joint distribution of a random vector isdivided into two parts:

1 the specification of the marginal distributions and2 the specification of the dependence structure by means of a special

function, called copula.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 71 / 95

Page 72: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

The use of copulas offers the following advantages:

The nature of dependency that can be modeled is more general.In comparison, only linear dependence can be explained by thecorrelation.

Dependence of extreme events might be modeled

Copulas are indifferent to continuously increasing transformations(not only linear as it is true for correlations).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 72 / 95

Page 73: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

From a mathematical viewpoint, a copula function C is nothingmore than a probability distribution function on the n-dimensionalhypercube In = [0, 1] × [0, 1] × . . . × [0, 1]:

C : Id → [0, 1]

(u1, . . . , un) → C(u1, . . . , un)

Any multivariate probability distribution function FY of somerandom vector Y = (Y1, ..., Yn) can be represented with the helpof a copula function C in the following form:

FY (y1, . . . , yn) = P(Y1 ≤ y1, . . . , Yn ≤ yn) = C(P(Y1 ≤ y1), . . . , P(Yn ≤ yn))

= C(FY1(y1), . . . , FYn(yn))

where FYi (yi), i = 1, . . . , n denote the marginal distribution functionsof the random variables Yi , i = 1, . . . , n.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 73 / 95

Page 74: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

The copula function makes the bridge between the univariatedistribution of the individual random variables and their jointprobability distribution.

The copula function creates uniquely the dependence whereasthe probability distribution of the involved random variables isprovided by their marginal distribution. By fixing the marginaldistributions and varying the copula function, we obtain allpossible joint distributions with the given marginals.

The links between marginal distributions and joint distributions areuseful in understanding the notion of a minimal probability metric.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 74 / 95

Page 75: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

We noted that the copula is just a probability distribution functionand, therefore, it can be characterized by means of a cumulativedistribution function or a probability density function.

Given a copula function C, the density is computed according toformula (2),

c(u1, . . . , un) =∂nC(u1, . . . , un)

∂u1 . . . ∂un.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 75 / 95

Page 76: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

In this way, using the relationship between the copula and thedistribution function, the density of the copula can be expressedby means of the density of the random variable. This is done byapplying the chain rule of differentiation,

c(FY1(y1), . . . , FYn(yn)) =fY (y1, . . . , yn)

fY1(y1) . . . fYn(yn). (4)

The numerator contains the density of the random variable Y andon the denominator we find the density of the Y but under theassumption that components of Y are independent randomvariables.

The left hand-side corresponds to the copula density buttransformed to the sample space by means of the marginaldistribution functions FYi (yi), i = 1, 2, . . . , n.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 76 / 95

Page 77: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

The copula density of a two-dimensional normal distribution with themean µ = (0, 0) and covariance matrix

Σ =

(

1 0.80.8 1

)

is plotted below.

0.20.4

0.60.8

0.20.4

0.60.8

0

1

2

3

4

u1

u2

c(u 1,u

2)

Figure: The copula density of a two-dimensional normal distribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 77 / 95

Page 78: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

The contour lines of the copula density transformed in the samplespace through the marginal distribution functions is given on thenext slide.

Plots of the probability density function and the contour lines ofthe probability density function are given on slides 62 and 64.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 78 / 95

Page 79: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

0.5

0.5

0.5

0.5

1

1

1

1

1

1

1.5

1.5

1.5

1.5

1.5

1.5

2

2

2

2

2.5

2.5

3

3

y1

y 2

−1 −0.5 0 0.5 1

−1

−0.5

0

0.5

1

Figure: The contour lines of a copula density of a two-dimensional normaldistribution transformed in the sample space.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 79 / 95

Page 80: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

Equation (4) reveals that, if the random variable Y hasindependent components, then the density of the correspondingcopula, denoted by c0, is a constant in the unit hypercube

c0(u1, . . . , un) = 1

and the copula C0 has the following simple form,

C0(u1, . . . , un) = u1 . . . un.

This copula characterizes stochastic independence.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 80 / 95

Page 81: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

Now let us consider a density c of some copula C:

The formula in equation (4) is a ratio of two positive quantitiesbecause the density function can only take non-negative values.

For each value of the vector of arguments y = (y1, . . . , yn),equation (4) provides information about the degree of dependencebetween the events that simultaneously Yi is in a smallneighborhood of yi for i = 1, 2, . . . , n.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 81 / 95

Page 82: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

That is, the copula density provides information about the localstructure of the dependence.

With respect to the copula density c0 characterizing the notion ofindependence, the arbitrary copula density function can be eitherabove 1, or below 1.

Suppose that for some vector y , the right hand-side of equation(4) is close to zero. This means that the numerator is muchsmaller than the denominator,

fY (y1, . . . , yn) < fY1(y1) . . . fYn(yn).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 82 / 95

Page 83: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

As a consequence, the joint probability of the events that Yi is in asmall neighborhood of yi for i = 1, 2, . . . , n is much smaller thanwhat it would if the corresponding events were independent.

This case corresponds to these events being almost disjoint; thatis, with a very small probability of occurring simultaneously.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 83 / 95

Page 84: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

Suppose that the converse holds, the numerator in equation (4) ismuch larger than the denominator and, as a result, the copuladensity is larger than 1. In this case,

fY (y1, . . . , yn) > fY1(y1) . . . fYn(yn),

It means that the joint probability of the events that Yi is in a smallneighborhood of yi for i = 1, 2, . . . , n is larger than what it would ifthe corresponding events were independent.

Therefore, copula density values larger than 1 mean that thecorresponding events are more likely to happen simultaneously.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 84 / 95

Page 85: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

This analysis indicates that the copula density function providesinformation about the local dependence structure of amultidimensional random variable Y relative to the case ofstochastic independence.

Figure on the slide 79 provides an illustration is thetwo-dimensional case. It shows the contour lines of the surfacecalculated according to formula (4) for the two-dimensional normaldistribution.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 85 / 95

Page 86: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Copula functions

All points which have an elevation above 1 have a localdependence implying that the events Y1 ∈ (y1, y1 + ǫ) andY2 ∈ (y2, y2 + ǫ) for a small ǫ > 0 are likely to occur jointly. Thismeans that in a large sample of observations, we will observe thetwo events happening together more often than implied by theindependence assumption.

In contrast, all points with an elevation below 1 have a localdependence implying that the events Y1 ∈ (y1, y1 + ǫ) andY2 ∈ (y2, y2 + ǫ) for a small ǫ > 0 are likely to occur disjointly. Thismeans that in a large sample of observations we will observe thetwo events happening less frequently than implied by theindependence assumption.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 86 / 95

Page 87: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Chebyshev’s inequality

Chebyshev’s inequality provides a way to estimate theapproximate probability of deviation of a random variable from itsmean. Its most simple form concerns positive random variables.

Suppose that X is a positive random variable, X > 0. Thefollowing inequality is known as Chebyshev’s inequality,

P(X ≥ ǫ) ≤ EXǫ

, (5)

where ǫ > 0.

In this form, equation (5) can be used to estimate the probability ofobserving a large observation by means of the mathematicalexpectation and the level ǫ.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 87 / 95

Page 88: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Chebyshev’s inequality

Chebyshev’s inequality is rough as demonstrated geometrically inthe following way. The mathematical expectation of a positivecontinuous random variable admits the representation,

EX =

0P(X ≥ x)dx ,

which means that it equals the area closed between thedistribution function and the upper limit of the distribution function.This area is illustrated in Figure on the next slide as the shadedarea above the distribution function.

On the other hand, the quantity ǫP(X ≥ ǫ) = ǫ(1 − FX (x)) is equalto the area of the rectangle in the upper left corner of the graph.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 88 / 95

Page 89: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Chebyshev’s inequality

0 0

1

x

FX(x)

ε

P(X≥ ε)

ε

Figure: Chebyshev’s inequality, a geometric illustration. The area of therectangle in the upper left corner is smaller than the shaded area.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 89 / 95

Page 90: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Chebyshev’s inequality

In effect, the inequality

ǫP(X ≥ ǫ) ≤ EX

admits the following geometric interpretation — the area of therectangle is smaller than the shaded area in the figure above.

For an arbitrary random variable, Chebychev’s inequality takes theform

P(|X − EX | ≥ ǫσX ) ≤ 1ǫ2 ,

where σX is the standard deviation of X and ǫ > 0.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 90 / 95

Page 91: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Fréchet-Hoeffding inequality

Consider a n-dimensional random vector Y with a distributionfunction FY (y1, . . . , yn). Denote by

W (y1, . . . , yn) = max(FY1(y1) + . . . + FYn(yn) + 1 − n, 0)

and byM(y1, . . . , yn) = min(FY1(y1), . . . , FYn(yn)),

in which FYi (yi) stands for the distribution function of the i-thmarginal.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 91 / 95

Page 92: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Fréchet-Hoeffding inequality

The following inequality is known as Fréchet-Hoeffding inequality,

W (y1, . . . , yn) ≤ FY (y1, . . . , yn) ≤ M(y1, . . . , yn). (6)

The quantities W (y1, . . . , yn) and M(y1, . . . , yn) are also called theFréchet lower bound and the Fréchet upper bound.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 92 / 95

Page 93: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Fréchet-Hoeffding inequality

Since copulas are essentially probability distributions defined onthe unit hypercube, Fréchet-Hoeffding inequality holds for them aswell. In this case, it has a simpler form because the marginaldistributions are uniform.

The lower and the upper Fréchet bounds equal

W (u1, . . . , un) = max(u1 + . . . + un + 1 − n, 0)

andM(u1, . . . , un) = min(u1, . . . , un)

respectively.

Fréchet-Hoeffding inequality is given by

W (u1, . . . , un) ≤ C(u1, . . . , un) ≤ M(u1, . . . , un).

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 93 / 95

Page 94: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Probabilistic inequalities - Fréchet-Hoeffding inequality

In the two-dimensional case, the inequality reduces to

max(u1 + u2 − 1, 0) ≤ C(u1, u2) ≤ min(u1, u2).

In the two-dimensional case only, the lower Fréchet bound,sometimes referred to as the minimal copula, represents perfectnegative dependence between the two random variables.

In a similar way, the upper Fréchet bound, sometimes referred toas the maximal copula, represents perfect positive dependencebetween the two random variables.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 94 / 95

Page 95: Lecture 1: Concepts of probabilitystatistik.econ.kit.edu/download/doc_secure1/1_StochModels.pdf · Basic concepts An outcome for a random variable is the mutually exclusive potential

Svetlozar T. Rachev, Stoyan Stoyanov, and Frank J. FabozziAdvanced Stochastic Models, Risk Assessment, and PortfolioOptimization: The Ideal Risk, Uncertainty, and PerformanceMeasuresJohn Wiley, Finance, 2007.

Chapter 1.

Prof. Dr. Svetlozar Rachev (University of Karlsruhe)Lecture 1: Concepts of probability 2008 95 / 95