koc1(dba)

206
Lectures on Lévy Processes and Stochastic Calculus (Koc University) Lecture 1: Infinite Divisibility David Applebaum School of Mathematics and Statistics, University of Sheffield, UK 5th December 2011 Dave Applebaum (Sheffield UK) Lecture 1 December 2011 1 / 40

Upload: serhat-yucel

Post on 01-Nov-2014

1.401 views

Category:

Technology


0 download

DESCRIPTION

 

TRANSCRIPT

Page 1: Koc1(dba)

Lectures on Lévy Processes and StochasticCalculus (Koc University)

Lecture 1: Infinite Divisibility

David Applebaum

School of Mathematics and Statistics, University of Sheffield, UK

5th December 2011

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 1 / 40

Page 2: Koc1(dba)

Introduction

A Lévy process is essentially a stochastic process with stationary andindependent increments.

The basic theory was developed, principally by Paul Lévy in the 1930s.In the past 20 years there has been a renaissance of interest and aplethora of books, articles and conferences. Why ?

There are both theoretical and practical reasons.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 2 / 40

Page 3: Koc1(dba)

Introduction

A Lévy process is essentially a stochastic process with stationary andindependent increments.

The basic theory was developed, principally by Paul Lévy in the 1930s.In the past 20 years there has been a renaissance of interest and aplethora of books, articles and conferences. Why ?

There are both theoretical and practical reasons.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 2 / 40

Page 4: Koc1(dba)

Introduction

A Lévy process is essentially a stochastic process with stationary andindependent increments.

The basic theory was developed, principally by Paul Lévy in the 1930s.In the past 20 years there has been a renaissance of interest and aplethora of books, articles and conferences. Why ?

There are both theoretical and practical reasons.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 2 / 40

Page 5: Koc1(dba)

Introduction

A Lévy process is essentially a stochastic process with stationary andindependent increments.

The basic theory was developed, principally by Paul Lévy in the 1930s.In the past 20 years there has been a renaissance of interest and aplethora of books, articles and conferences. Why ?

There are both theoretical and practical reasons.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 2 / 40

Page 6: Koc1(dba)

THEORETICAL

Lévy processes are simplest generic class of process which have(a.s.) continuous paths interspersed with random jumps ofarbitrary size occurring at random times.

Lévy processes comprise a natural subclass of semimartingalesand of Markov processes of Feller type.

There are many interesting examples - Brownian motion, simpleand compound Poisson processes, α-stable processes,subordinated processes, financial processes, relativistic process,Riemann zeta process . . .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 3 / 40

Page 7: Koc1(dba)

THEORETICAL

Lévy processes are simplest generic class of process which have(a.s.) continuous paths interspersed with random jumps ofarbitrary size occurring at random times.

Lévy processes comprise a natural subclass of semimartingalesand of Markov processes of Feller type.

There are many interesting examples - Brownian motion, simpleand compound Poisson processes, α-stable processes,subordinated processes, financial processes, relativistic process,Riemann zeta process . . .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 3 / 40

Page 8: Koc1(dba)

THEORETICAL

Lévy processes are simplest generic class of process which have(a.s.) continuous paths interspersed with random jumps ofarbitrary size occurring at random times.

Lévy processes comprise a natural subclass of semimartingalesand of Markov processes of Feller type.

There are many interesting examples - Brownian motion, simpleand compound Poisson processes, α-stable processes,subordinated processes, financial processes, relativistic process,Riemann zeta process . . .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 3 / 40

Page 9: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 10: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 11: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 12: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 13: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 14: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 15: Koc1(dba)

Noise. Lévy processes are a good model of “noise” in randomdynamical systems.

Input + Noise = Output

Attempts to describe this differentially leads to stochastic calculus.A large class of Markov processes can be built as solutions ofstochastic differential equations (SDEs) driven by Lévy noise.

Lévy driven stochastic partial differential equations (SPDEs) arecurrently being studied with some intensity.

Robust structure. Most applications utilise Lévy processes takingvalues in Euclidean space but this can be replaced by a Hilbertspace, a Banach space (both of these are important for SPDEs), alocally compact group, a manifold. Quantised versions arenon-commutative Lévy processes on quantum groups.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 4 / 40

Page 16: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 17: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 18: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 19: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 20: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 21: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 22: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 23: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 24: Koc1(dba)

APPLICATIONS

These include:

Turbulence via Burger’s equation (Bertoin)New examples of quantum field theories (Albeverio, Gottshalk,Wu)Viscoelasticity (Bouleau)Time series - Lévy driven CARMA models (Brockwell, Marquardt)Stochastic resonance in non-linear signal processing (Patel,Kosco, Applebaum)Finance and insurance (a cast of thousands)

The biggest explosion of activity has been in mathematical finance.Two major areas of activity are:

option pricing in incomplete markets.interest rate modelling.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 5 / 40

Page 25: Koc1(dba)

Some Basic Ideas of Probability

Notation. Our state space is Euclidean space Rd . The inner productbetween two vectors x = (x1, . . . , xd ) and y = (y1, . . . , yd ) is

(x , y) =d∑

i=1

xiyi .

The associated norm (length of a vector) is

|x | = (x , x)12 =

(d∑

i=1

x2i

) 12

.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 6 / 40

Page 26: Koc1(dba)

Some Basic Ideas of Probability

Notation. Our state space is Euclidean space Rd . The inner productbetween two vectors x = (x1, . . . , xd ) and y = (y1, . . . , yd ) is

(x , y) =d∑

i=1

xiyi .

The associated norm (length of a vector) is

|x | = (x , x)12 =

(d∑

i=1

x2i

) 12

.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 6 / 40

Page 27: Koc1(dba)

Some Basic Ideas of Probability

Notation. Our state space is Euclidean space Rd . The inner productbetween two vectors x = (x1, . . . , xd ) and y = (y1, . . . , yd ) is

(x , y) =d∑

i=1

xiyi .

The associated norm (length of a vector) is

|x | = (x , x)12 =

(d∑

i=1

x2i

) 12

.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 6 / 40

Page 28: Koc1(dba)

Let (Ω,F ,P) be a probability space, so that Ω is a set, F is a σ-algebraof subsets of Ω and P is a probability measure defined on (Ω,F).

Random variables are measurable functions X : Ω→ Rd . The law of Xis pX , where for each A ∈ F ,pX (A) = P(X ∈ A).

(Xn,n ∈ N) are independent if for alli1, i2, . . . ir ∈ N,Ai1 ,Ai2 , . . . ,Air ∈ B(Rd ),

P(Xi1 ∈ A1,Xi2 ∈ A2, . . . ,Xir ∈ Ar )

= P(Xi1 ∈ A1)P(Xi2 ∈ A2) · · ·P(Xir ∈ Ar ).

If X and Y are independent, the law of X + Y is given by convolution ofmeasures

pX+Y = pX ∗ pY , where (pX ∗ pY )(A) =

∫Rd

pX (A− y)pY (dy).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 7 / 40

Page 29: Koc1(dba)

Let (Ω,F ,P) be a probability space, so that Ω is a set, F is a σ-algebraof subsets of Ω and P is a probability measure defined on (Ω,F).

Random variables are measurable functions X : Ω→ Rd . The law of Xis pX , where for each A ∈ F ,pX (A) = P(X ∈ A).

(Xn,n ∈ N) are independent if for alli1, i2, . . . ir ∈ N,Ai1 ,Ai2 , . . . ,Air ∈ B(Rd ),

P(Xi1 ∈ A1,Xi2 ∈ A2, . . . ,Xir ∈ Ar )

= P(Xi1 ∈ A1)P(Xi2 ∈ A2) · · ·P(Xir ∈ Ar ).

If X and Y are independent, the law of X + Y is given by convolution ofmeasures

pX+Y = pX ∗ pY , where (pX ∗ pY )(A) =

∫Rd

pX (A− y)pY (dy).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 7 / 40

Page 30: Koc1(dba)

Let (Ω,F ,P) be a probability space, so that Ω is a set, F is a σ-algebraof subsets of Ω and P is a probability measure defined on (Ω,F).

Random variables are measurable functions X : Ω→ Rd . The law of Xis pX , where for each A ∈ F ,pX (A) = P(X ∈ A).

(Xn,n ∈ N) are independent if for alli1, i2, . . . ir ∈ N,Ai1 ,Ai2 , . . . ,Air ∈ B(Rd ),

P(Xi1 ∈ A1,Xi2 ∈ A2, . . . ,Xir ∈ Ar )

= P(Xi1 ∈ A1)P(Xi2 ∈ A2) · · ·P(Xir ∈ Ar ).

If X and Y are independent, the law of X + Y is given by convolution ofmeasures

pX+Y = pX ∗ pY , where (pX ∗ pY )(A) =

∫Rd

pX (A− y)pY (dy).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 7 / 40

Page 31: Koc1(dba)

Let (Ω,F ,P) be a probability space, so that Ω is a set, F is a σ-algebraof subsets of Ω and P is a probability measure defined on (Ω,F).

Random variables are measurable functions X : Ω→ Rd . The law of Xis pX , where for each A ∈ F ,pX (A) = P(X ∈ A).

(Xn,n ∈ N) are independent if for alli1, i2, . . . ir ∈ N,Ai1 ,Ai2 , . . . ,Air ∈ B(Rd ),

P(Xi1 ∈ A1,Xi2 ∈ A2, . . . ,Xir ∈ Ar )

= P(Xi1 ∈ A1)P(Xi2 ∈ A2) · · ·P(Xir ∈ Ar ).

If X and Y are independent, the law of X + Y is given by convolution ofmeasures

pX+Y = pX ∗ pY , where (pX ∗ pY )(A) =

∫Rd

pX (A− y)pY (dy).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 7 / 40

Page 32: Koc1(dba)

Equivalently∫Rd

g(y)(pX ∗ pY )(dy) =

∫Rd

∫Rd

g(x + y)pX (dx)pY (dy),

for all g ∈ Bb(Rd ) (the bounded Borel measurable functions on Rd ).If X and Y are independent with densities fX and fY , respectively, thenX + Y has density fX+Y given by convolution of functions:

fX+Y = fX ∗ fY , where (fX ∗ fY )(x) =

∫Rd

fX (x − y)fY (y)dy .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 8 / 40

Page 33: Koc1(dba)

Equivalently∫Rd

g(y)(pX ∗ pY )(dy) =

∫Rd

∫Rd

g(x + y)pX (dx)pY (dy),

for all g ∈ Bb(Rd ) (the bounded Borel measurable functions on Rd ).If X and Y are independent with densities fX and fY , respectively, thenX + Y has density fX+Y given by convolution of functions:

fX+Y = fX ∗ fY , where (fX ∗ fY )(x) =

∫Rd

fX (x − y)fY (y)dy .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 8 / 40

Page 34: Koc1(dba)

Equivalently∫Rd

g(y)(pX ∗ pY )(dy) =

∫Rd

∫Rd

g(x + y)pX (dx)pY (dy),

for all g ∈ Bb(Rd ) (the bounded Borel measurable functions on Rd ).If X and Y are independent with densities fX and fY , respectively, thenX + Y has density fX+Y given by convolution of functions:

fX+Y = fX ∗ fY , where (fX ∗ fY )(x) =

∫Rd

fX (x − y)fY (y)dy .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 8 / 40

Page 35: Koc1(dba)

The characteristic function of X is φX : Rd → C, where

φX (u) =

∫Rd

ei(u,x)pX (dx).

Theorem (Kac’s theorem)X1, . . . ,Xn are independent if and only if

E

exp

in∑

j=1

(uj ,Xj)

= φX1(u1) · · ·φXn (un)

for all u1, . . . ,un ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 9 / 40

Page 36: Koc1(dba)

The characteristic function of X is φX : Rd → C, where

φX (u) =

∫Rd

ei(u,x)pX (dx).

Theorem (Kac’s theorem)X1, . . . ,Xn are independent if and only if

E

exp

in∑

j=1

(uj ,Xj)

= φX1(u1) · · ·φXn (un)

for all u1, . . . ,un ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 9 / 40

Page 37: Koc1(dba)

The characteristic function of X is φX : Rd → C, where

φX (u) =

∫Rd

ei(u,x)pX (dx).

Theorem (Kac’s theorem)X1, . . . ,Xn are independent if and only if

E

exp

in∑

j=1

(uj ,Xj)

= φX1(u1) · · ·φXn (un)

for all u1, . . . ,un ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 9 / 40

Page 38: Koc1(dba)

More generally, the characteristic function of a probability measure µon Rd is

φµ(u) =

∫Rd

ei(u,x)µ(dx).

Important properties are:-1 φµ(0) = 1.

2 φµ is positive definite i.e.∑

i,j ci cjφµ(ui − uj) ≥ 0, for allci ∈ C,ui ∈ Rd ,1 ≤ i , j ≤ n,n ∈ N.

3 φµ is uniformly continuous - Hint: Look at |φµ(u + h)− φµ(u)| and usedominated convergence)).

Also µ→ φµ is injective.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 10 / 40

Page 39: Koc1(dba)

More generally, the characteristic function of a probability measure µon Rd is

φµ(u) =

∫Rd

ei(u,x)µ(dx).

Important properties are:-1 φµ(0) = 1.

2 φµ is positive definite i.e.∑

i,j ci cjφµ(ui − uj) ≥ 0, for allci ∈ C,ui ∈ Rd ,1 ≤ i , j ≤ n,n ∈ N.

3 φµ is uniformly continuous - Hint: Look at |φµ(u + h)− φµ(u)| and usedominated convergence)).

Also µ→ φµ is injective.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 10 / 40

Page 40: Koc1(dba)

More generally, the characteristic function of a probability measure µon Rd is

φµ(u) =

∫Rd

ei(u,x)µ(dx).

Important properties are:-1 φµ(0) = 1.

2 φµ is positive definite i.e.∑

i,j ci cjφµ(ui − uj) ≥ 0, for allci ∈ C,ui ∈ Rd ,1 ≤ i , j ≤ n,n ∈ N.

3 φµ is uniformly continuous - Hint: Look at |φµ(u + h)− φµ(u)| and usedominated convergence)).

Also µ→ φµ is injective.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 10 / 40

Page 41: Koc1(dba)

More generally, the characteristic function of a probability measure µon Rd is

φµ(u) =

∫Rd

ei(u,x)µ(dx).

Important properties are:-1 φµ(0) = 1.

2 φµ is positive definite i.e.∑

i,j ci cjφµ(ui − uj) ≥ 0, for allci ∈ C,ui ∈ Rd ,1 ≤ i , j ≤ n,n ∈ N.

3 φµ is uniformly continuous - Hint: Look at |φµ(u + h)− φµ(u)| and usedominated convergence)).

Also µ→ φµ is injective.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 10 / 40

Page 42: Koc1(dba)

More generally, the characteristic function of a probability measure µon Rd is

φµ(u) =

∫Rd

ei(u,x)µ(dx).

Important properties are:-1 φµ(0) = 1.

2 φµ is positive definite i.e.∑

i,j ci cjφµ(ui − uj) ≥ 0, for allci ∈ C,ui ∈ Rd ,1 ≤ i , j ≤ n,n ∈ N.

3 φµ is uniformly continuous - Hint: Look at |φµ(u + h)− φµ(u)| and usedominated convergence)).

Also µ→ φµ is injective.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 10 / 40

Page 43: Koc1(dba)

Conversely Bochner’s theorem states that if φ : Rd → C satisfies (1),(2) and is continuous at u = 0, then it is the characteristic function ofsome probability measure µ on Rd .

ψ : Rd → C is conditionally positive definite if for all n ∈ N andc1, . . . , cn ∈ C for which

∑nj=1 cj = 0 we have

n∑j,k=1

cj ckψ(uj − uk ) ≥ 0,

for all u1, . . . ,un ∈ Rd .

Note: conditionally positive definite is sometimes called negativedefinite (with a minus sign).

ψ : Rd → C will be said to be hermitian if ψ(u) = ψ(−u), for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 11 / 40

Page 44: Koc1(dba)

Conversely Bochner’s theorem states that if φ : Rd → C satisfies (1),(2) and is continuous at u = 0, then it is the characteristic function ofsome probability measure µ on Rd .

ψ : Rd → C is conditionally positive definite if for all n ∈ N andc1, . . . , cn ∈ C for which

∑nj=1 cj = 0 we have

n∑j,k=1

cj ckψ(uj − uk ) ≥ 0,

for all u1, . . . ,un ∈ Rd .

Note: conditionally positive definite is sometimes called negativedefinite (with a minus sign).

ψ : Rd → C will be said to be hermitian if ψ(u) = ψ(−u), for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 11 / 40

Page 45: Koc1(dba)

Conversely Bochner’s theorem states that if φ : Rd → C satisfies (1),(2) and is continuous at u = 0, then it is the characteristic function ofsome probability measure µ on Rd .

ψ : Rd → C is conditionally positive definite if for all n ∈ N andc1, . . . , cn ∈ C for which

∑nj=1 cj = 0 we have

n∑j,k=1

cj ckψ(uj − uk ) ≥ 0,

for all u1, . . . ,un ∈ Rd .

Note: conditionally positive definite is sometimes called negativedefinite (with a minus sign).

ψ : Rd → C will be said to be hermitian if ψ(u) = ψ(−u), for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 11 / 40

Page 46: Koc1(dba)

Conversely Bochner’s theorem states that if φ : Rd → C satisfies (1),(2) and is continuous at u = 0, then it is the characteristic function ofsome probability measure µ on Rd .

ψ : Rd → C is conditionally positive definite if for all n ∈ N andc1, . . . , cn ∈ C for which

∑nj=1 cj = 0 we have

n∑j,k=1

cj ckψ(uj − uk ) ≥ 0,

for all u1, . . . ,un ∈ Rd .

Note: conditionally positive definite is sometimes called negativedefinite (with a minus sign).

ψ : Rd → C will be said to be hermitian if ψ(u) = ψ(−u), for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 11 / 40

Page 47: Koc1(dba)

Conversely Bochner’s theorem states that if φ : Rd → C satisfies (1),(2) and is continuous at u = 0, then it is the characteristic function ofsome probability measure µ on Rd .

ψ : Rd → C is conditionally positive definite if for all n ∈ N andc1, . . . , cn ∈ C for which

∑nj=1 cj = 0 we have

n∑j,k=1

cj ckψ(uj − uk ) ≥ 0,

for all u1, . . . ,un ∈ Rd .

Note: conditionally positive definite is sometimes called negativedefinite (with a minus sign).

ψ : Rd → C will be said to be hermitian if ψ(u) = ψ(−u), for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 11 / 40

Page 48: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 49: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 50: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 51: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 52: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 53: Koc1(dba)

Theorem (Schoenberg correspondence)

ψ : Rd → C is hermitian and conditionally positive definite if and only ifetψ is positive definite for each t > 0.

Proof. We only give the easy part here.Suppose that etψ is positive definite for all t > 0. Fix n ∈ N and choosec1, . . . , cn and u1, . . . ,un as above.We then find that for each t > 0,

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0,

and so

n∑j,k=1

cj ckψ(uj − uk ) = limt→0

1t

n∑j,k=1

cj ck (etψ(uj−uk ) − 1) ≥ 0.

2Dave Applebaum (Sheffield UK) Lecture 1 December 2011 12 / 40

Page 54: Koc1(dba)

Infinite Divisibility

We study this first because a Lévy process is infinite divisibility inmotion, i.e. infinite divisibility is the underlying probabilistic idea whicha Lévy process embodies dynamically.

Let µ be a probability measure on Rd . Define µ∗n

= µ ∗ · · · ∗ µ (ntimes). We say that µ has a convolution nth root, if there exists aprobability measure µ

1n for which (µ

1n )∗

n= µ.

µ is infinitely divisible if it has a convolution nth root for all n ∈ N. In thiscase µ

1n is unique.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 13 / 40

Page 55: Koc1(dba)

Infinite Divisibility

We study this first because a Lévy process is infinite divisibility inmotion, i.e. infinite divisibility is the underlying probabilistic idea whicha Lévy process embodies dynamically.

Let µ be a probability measure on Rd . Define µ∗n

= µ ∗ · · · ∗ µ (ntimes). We say that µ has a convolution nth root, if there exists aprobability measure µ

1n for which (µ

1n )∗

n= µ.

µ is infinitely divisible if it has a convolution nth root for all n ∈ N. In thiscase µ

1n is unique.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 13 / 40

Page 56: Koc1(dba)

Infinite Divisibility

We study this first because a Lévy process is infinite divisibility inmotion, i.e. infinite divisibility is the underlying probabilistic idea whicha Lévy process embodies dynamically.

Let µ be a probability measure on Rd . Define µ∗n

= µ ∗ · · · ∗ µ (ntimes). We say that µ has a convolution nth root, if there exists aprobability measure µ

1n for which (µ

1n )∗

n= µ.

µ is infinitely divisible if it has a convolution nth root for all n ∈ N. In thiscase µ

1n is unique.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 13 / 40

Page 57: Koc1(dba)

Infinite Divisibility

We study this first because a Lévy process is infinite divisibility inmotion, i.e. infinite divisibility is the underlying probabilistic idea whicha Lévy process embodies dynamically.

Let µ be a probability measure on Rd . Define µ∗n

= µ ∗ · · · ∗ µ (ntimes). We say that µ has a convolution nth root, if there exists aprobability measure µ

1n for which (µ

1n )∗

n= µ.

µ is infinitely divisible if it has a convolution nth root for all n ∈ N. In thiscase µ

1n is unique.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 13 / 40

Page 58: Koc1(dba)

Infinite Divisibility

We study this first because a Lévy process is infinite divisibility inmotion, i.e. infinite divisibility is the underlying probabilistic idea whicha Lévy process embodies dynamically.

Let µ be a probability measure on Rd . Define µ∗n

= µ ∗ · · · ∗ µ (ntimes). We say that µ has a convolution nth root, if there exists aprobability measure µ

1n for which (µ

1n )∗

n= µ.

µ is infinitely divisible if it has a convolution nth root for all n ∈ N. In thiscase µ

1n is unique.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 13 / 40

Page 59: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 60: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 61: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 62: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 63: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 64: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 65: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 66: Koc1(dba)

Theoremµ is infinitely divisible iff for all n ∈ N, there exists a probability measureµn with characteristic function φn such that

φµ(u) = (φn(u))n,

for all u ∈ Rd . Moreover µn = µ1n .

Proof. If µ is infinitely divisible, take φn = φµ

1n. Conversely, for each

n ∈ N, by Fubini’s theorem,

φµ(u) =

∫Rd· · ·∫Rd

ei(u,y1+···+yn)µn(dy1) · · ·µn(dyn)

=

∫Rd

ei(u,y)µ∗n

n (dy).

But φµ(u) =∫Rd ei(u,y)µ(dy) and φ determines µ uniquely. Hence

µ = µ∗n

n . 2

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 14 / 40

Page 67: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 68: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 69: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 70: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 71: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 72: Koc1(dba)

- If µ and ν are each infinitely divisible, then so is µ ∗ ν.

- If (µn,n ∈ N) are infinitely divisible and µnw⇒ µ, then µ is infinitely

divisible.

[Note: Weak convergence. µnw⇒ µ means

limn→∞

∫Rd

f (x)µn(dx) =

∫Rd

f (x)µ(dx),

for each f ∈ Cb(Rd ).]

A random variable X is infinitely divisible if its law pX is infinitelydivisible, X d

= Y (n)1 + · · ·+ Y (n)

n , where Y (n)1 , . . . ,Y (n)

n are i.i.d., for eachn ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 15 / 40

Page 73: Koc1(dba)

Examples of Infinite Divisibility

In the following, we will demonstrate infinite divisibility of a randomvariable X by finding i.i.d. Y (n)

1 , . . . ,Y (n)n such that

X d= Y (n)

1 + · · ·+ Y (n)n , for each n ∈ N.

Example 1 - Gaussian Random Variables

Let X = (X1, . . . ,Xd ) be a random vector.We say that it is (non − degenerate)Gaussian if it there exists a vectorm ∈ Rd and a strictly positive-definite symmetric d × d matrix A suchthat X has a pdf (probability density function) of the form:-

f (x) =1

(2π)d2√

det(A)exp

(−1

2(x −m,A−1(x −m))

), (1.1)

for all x ∈ Rd .In this case we will write X ∼ N(m,A). The vector m is the mean of X ,so m = E(X ) and A is the covariance matrix so thatA = E((X −m)(X −m)T ).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 16 / 40

Page 74: Koc1(dba)

Examples of Infinite Divisibility

In the following, we will demonstrate infinite divisibility of a randomvariable X by finding i.i.d. Y (n)

1 , . . . ,Y (n)n such that

X d= Y (n)

1 + · · ·+ Y (n)n , for each n ∈ N.

Example 1 - Gaussian Random Variables

Let X = (X1, . . . ,Xd ) be a random vector.We say that it is (non − degenerate)Gaussian if it there exists a vectorm ∈ Rd and a strictly positive-definite symmetric d × d matrix A suchthat X has a pdf (probability density function) of the form:-

f (x) =1

(2π)d2√

det(A)exp

(−1

2(x −m,A−1(x −m))

), (1.1)

for all x ∈ Rd .In this case we will write X ∼ N(m,A). The vector m is the mean of X ,so m = E(X ) and A is the covariance matrix so thatA = E((X −m)(X −m)T ).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 16 / 40

Page 75: Koc1(dba)

Examples of Infinite Divisibility

In the following, we will demonstrate infinite divisibility of a randomvariable X by finding i.i.d. Y (n)

1 , . . . ,Y (n)n such that

X d= Y (n)

1 + · · ·+ Y (n)n , for each n ∈ N.

Example 1 - Gaussian Random Variables

Let X = (X1, . . . ,Xd ) be a random vector.We say that it is (non − degenerate)Gaussian if it there exists a vectorm ∈ Rd and a strictly positive-definite symmetric d × d matrix A suchthat X has a pdf (probability density function) of the form:-

f (x) =1

(2π)d2√

det(A)exp

(−1

2(x −m,A−1(x −m))

), (1.1)

for all x ∈ Rd .In this case we will write X ∼ N(m,A). The vector m is the mean of X ,so m = E(X ) and A is the covariance matrix so thatA = E((X −m)(X −m)T ).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 16 / 40

Page 76: Koc1(dba)

Examples of Infinite Divisibility

In the following, we will demonstrate infinite divisibility of a randomvariable X by finding i.i.d. Y (n)

1 , . . . ,Y (n)n such that

X d= Y (n)

1 + · · ·+ Y (n)n , for each n ∈ N.

Example 1 - Gaussian Random Variables

Let X = (X1, . . . ,Xd ) be a random vector.We say that it is (non − degenerate)Gaussian if it there exists a vectorm ∈ Rd and a strictly positive-definite symmetric d × d matrix A suchthat X has a pdf (probability density function) of the form:-

f (x) =1

(2π)d2√

det(A)exp

(−1

2(x −m,A−1(x −m))

), (1.1)

for all x ∈ Rd .In this case we will write X ∼ N(m,A). The vector m is the mean of X ,so m = E(X ) and A is the covariance matrix so thatA = E((X −m)(X −m)T ).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 16 / 40

Page 77: Koc1(dba)

Examples of Infinite Divisibility

In the following, we will demonstrate infinite divisibility of a randomvariable X by finding i.i.d. Y (n)

1 , . . . ,Y (n)n such that

X d= Y (n)

1 + · · ·+ Y (n)n , for each n ∈ N.

Example 1 - Gaussian Random Variables

Let X = (X1, . . . ,Xd ) be a random vector.We say that it is (non − degenerate)Gaussian if it there exists a vectorm ∈ Rd and a strictly positive-definite symmetric d × d matrix A suchthat X has a pdf (probability density function) of the form:-

f (x) =1

(2π)d2√

det(A)exp

(−1

2(x −m,A−1(x −m))

), (1.1)

for all x ∈ Rd .In this case we will write X ∼ N(m,A). The vector m is the mean of X ,so m = E(X ) and A is the covariance matrix so thatA = E((X −m)(X −m)T ).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 16 / 40

Page 78: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 79: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 80: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 81: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 82: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 83: Koc1(dba)

A standard calculation yields

φX (u) = exp i(m,u)− 12

(u,Au), (1.2)

and hence

(φX (u))1n = exp

i(m

n,u)− 1

2

(u,

1n

Au)

,

so we see that X is infinitely divisible with each Y (n)j ∼ N(m

n ,1n A) for

each 1 ≤ j ≤ n.We say that X is a standard normal whenever X ∼ N(0, σ2I) for someσ > 0.

We say that X is degenerate Gaussian if (1.2) holds with det(A) = 0,and these random variables are also infinitely divisible.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 17 / 40

Page 84: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 85: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 86: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 87: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 88: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 89: Koc1(dba)

Example 2 - Poisson Random Variables

In this case, we take d = 1 and consider a random variable X takingvalues in the set n ∈ N ∪ 0. We say that is Poisson if there existsc > 0 for which

P(X = n) =cn

n!e−c .

In this case we will write X ∼ π(c). We have E(X ) = Var(X ) = c. It iseasy to verify that

φX (u) = exp[c(eiu − 1)],

from which we deduce that X is infinitely divisible with eachY (n)

j ∼ π( cn ), for 1 ≤ j ≤ n,n ∈ N.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 18 / 40

Page 90: Koc1(dba)

Example 3 - Compound Poisson Random Variables

Let (Z (n),n ∈ N) be a sequence of i.i.d. random variables takingvalues in Rd with common law µZ and let N ∼ π(c) be a Poissonrandom variable which is independent of all the Z (n)’s.The compoundPoisson random variable X is defined as follows:-

X :=

0 if N = 0

Z (1) + · · ·+ Z (N) if N > 0.

Theorem

For each u ∈ Rd ,

φX (u) = exp[∫

Rd(ei(u,y) − 1)cµZ (dy)

].

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 19 / 40

Page 91: Koc1(dba)

Example 3 - Compound Poisson Random Variables

Let (Z (n),n ∈ N) be a sequence of i.i.d. random variables takingvalues in Rd with common law µZ and let N ∼ π(c) be a Poissonrandom variable which is independent of all the Z (n)’s.The compoundPoisson random variable X is defined as follows:-

X :=

0 if N = 0

Z (1) + · · ·+ Z (N) if N > 0.

Theorem

For each u ∈ Rd ,

φX (u) = exp[∫

Rd(ei(u,y) − 1)cµZ (dy)

].

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 19 / 40

Page 92: Koc1(dba)

Example 3 - Compound Poisson Random Variables

Let (Z (n),n ∈ N) be a sequence of i.i.d. random variables takingvalues in Rd with common law µZ and let N ∼ π(c) be a Poissonrandom variable which is independent of all the Z (n)’s.The compoundPoisson random variable X is defined as follows:-

X :=

0 if N = 0

Z (1) + · · ·+ Z (N) if N > 0.

Theorem

For each u ∈ Rd ,

φX (u) = exp[∫

Rd(ei(u,y) − 1)cµZ (dy)

].

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 19 / 40

Page 93: Koc1(dba)

Example 3 - Compound Poisson Random Variables

Let (Z (n),n ∈ N) be a sequence of i.i.d. random variables takingvalues in Rd with common law µZ and let N ∼ π(c) be a Poissonrandom variable which is independent of all the Z (n)’s.The compoundPoisson random variable X is defined as follows:-

X :=

0 if N = 0

Z (1) + · · ·+ Z (N) if N > 0.

Theorem

For each u ∈ Rd ,

φX (u) = exp[∫

Rd(ei(u,y) − 1)cµZ (dy)

].

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 19 / 40

Page 94: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 95: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 96: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 97: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 98: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 99: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 100: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 101: Koc1(dba)

Proof. Let φZ be the common characteristic function of the Zn’s. Byconditioning and using independence we find,

φX (u) =∞∑

n=0

E(ei(u,Z (1)+···+Z (N))|N = n)P(N = n)

=∞∑

n=0

E(ei(u,Z (1))+···+Z (n)))e−c cn

n!

= e−c∞∑

n=0

[cφZ (u)]n

n!

= exp[c(φZ (u)− 1)],

and the result follows on writing φZ (u) =∫

ei(u,y)µZ (dy). 2

If X is compound Poisson as above, we write X ∼ π(c, µZ ). It is clearlyinfinitely divisible with each Y (n)

j ∼ π( cn , µZ ), for 1 ≤ j ≤ n.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 20 / 40

Page 102: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 103: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 104: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 105: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 106: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 107: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 108: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 109: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 110: Koc1(dba)

The Lévy-Khintchine Formula

de Finetti (1920’s) suggested that the most general infinitely divisiblerandom variable could be written X = Y + W , where Y and W areindependent, Y ∼ N(m,A),W ∼ π(c, µZ ). Then φX (u) = eη(u), where

η(u) = i(m,u)− 12

(u,Au) +

∫Rd

(ei(u,y) − 1)cµZ (dy). (1.3)

This is WRONG! ν(·) = cµZ (·) is a finite measure here.Lévy and Khintchine showed that ν can be σ-finite, provided it is whatis now called a Lévy measure on Rd − 0 = x ∈ Rd , x 6= 0, i.e.∫

(|y |2 ∧ 1)ν(dy) <∞, (1.4)

(where a ∧ b := mina,b, for a,b ∈ R). Since |y |2 ∧ ε ≤ |y |2 ∧ 1whenever 0 < ε ≤ 1, it follows from (1.4) that

ν((−ε, ε)c) <∞ for all ε > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 21 / 40

Page 111: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 112: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 113: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 114: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 115: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 116: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 117: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 118: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 119: Koc1(dba)

Here is the fundamental result of this lecture:-

Theorem (Lévy-Khintchine)

A Borel probability measure µ on Rd is infinitely divisible if there existsa vector b ∈ Rd , a non-negative symmetric d × d matrix A and a Lévymeasure ν on Rd − 0 such that for all u ∈ Rd ,

φµ(u) = exp[i(b,u)− 1

2(u,Au) (1.5)

+

∫Rd−0

(ei(u,y) − 1− i(u, y)1B(y))ν(dy)

]. (1.6)

where B = B1(0) = y ∈ Rd ; |y | < 1.Conversely, any mapping of the form (1.5) is the characteristic functionof an infinitely divisible probability measure on Rd .

The triple (b,A, ν) is called the characteristics of the infinitely divisiblerandom variable X . Define η = logφµ, where we take the principal partof the logarithm. η is called the Lévy symbol.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 22 / 40

Page 120: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 121: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 122: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 123: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 124: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 125: Koc1(dba)

η may also be called the characteristic exponent.We’re not going to prove this result here. To understand it, it isinstructive to let (Un,n ∈ N) be a sequence of Borel sets in B1(0) withUn ↓ 0. Observe that

η(u) = limn→∞

ηn(u) where each

ηn(u) = i[(

b −∫

Ucn∩B

yν(dy),u)]− 1

2(u,Au) +

∫Uc

n

(ei(u,y) − 1)ν(dy),

so η is in some sense (to be made more precise later) the limit of asequence of sums of Gaussians and independent compoundPoissons. Interesting phenomena appear in the limit as we’ll seebelow.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 23 / 40

Page 126: Koc1(dba)

First, we classify Lévy symbols analytically:-

Theoremη is a Lévy symbol if and only if it is a continuous, hermitianconditionally positive definite function for which η(0) = 0.

Hilbert Space. The Lévy-Khintchine formula extends in an obviousmanner to separable Hilbert spaces H. We interpret (·, ·) as the Hilbertspace inner product. The characteristics (b,A, ν) are defined similarly -A is a positive, symmetric linear operator on H.

Banach space. We can also extend to a separable Banach space Ehaving dual E∗. Define Φµ : E∗ → C by

Φµ(u) =

∫E

ei〈u,x〉µ(dx),

where 〈·, ·〉 is duality. Replacing (·, ·) with 〈·, ·〉 the Lévy-Khintchineformula extends naturally but care must be taken in defining a Lévymeasure!

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 24 / 40

Page 127: Koc1(dba)

First, we classify Lévy symbols analytically:-

Theoremη is a Lévy symbol if and only if it is a continuous, hermitianconditionally positive definite function for which η(0) = 0.

Hilbert Space. The Lévy-Khintchine formula extends in an obviousmanner to separable Hilbert spaces H. We interpret (·, ·) as the Hilbertspace inner product. The characteristics (b,A, ν) are defined similarly -A is a positive, symmetric linear operator on H.

Banach space. We can also extend to a separable Banach space Ehaving dual E∗. Define Φµ : E∗ → C by

Φµ(u) =

∫E

ei〈u,x〉µ(dx),

where 〈·, ·〉 is duality. Replacing (·, ·) with 〈·, ·〉 the Lévy-Khintchineformula extends naturally but care must be taken in defining a Lévymeasure!

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 24 / 40

Page 128: Koc1(dba)

First, we classify Lévy symbols analytically:-

Theoremη is a Lévy symbol if and only if it is a continuous, hermitianconditionally positive definite function for which η(0) = 0.

Hilbert Space. The Lévy-Khintchine formula extends in an obviousmanner to separable Hilbert spaces H. We interpret (·, ·) as the Hilbertspace inner product. The characteristics (b,A, ν) are defined similarly -A is a positive, symmetric linear operator on H.

Banach space. We can also extend to a separable Banach space Ehaving dual E∗. Define Φµ : E∗ → C by

Φµ(u) =

∫E

ei〈u,x〉µ(dx),

where 〈·, ·〉 is duality. Replacing (·, ·) with 〈·, ·〉 the Lévy-Khintchineformula extends naturally but care must be taken in defining a Lévymeasure!

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 24 / 40

Page 129: Koc1(dba)

First, we classify Lévy symbols analytically:-

Theoremη is a Lévy symbol if and only if it is a continuous, hermitianconditionally positive definite function for which η(0) = 0.

Hilbert Space. The Lévy-Khintchine formula extends in an obviousmanner to separable Hilbert spaces H. We interpret (·, ·) as the Hilbertspace inner product. The characteristics (b,A, ν) are defined similarly -A is a positive, symmetric linear operator on H.

Banach space. We can also extend to a separable Banach space Ehaving dual E∗. Define Φµ : E∗ → C by

Φµ(u) =

∫E

ei〈u,x〉µ(dx),

where 〈·, ·〉 is duality. Replacing (·, ·) with 〈·, ·〉 the Lévy-Khintchineformula extends naturally but care must be taken in defining a Lévymeasure!

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 24 / 40

Page 130: Koc1(dba)

First, we classify Lévy symbols analytically:-

Theoremη is a Lévy symbol if and only if it is a continuous, hermitianconditionally positive definite function for which η(0) = 0.

Hilbert Space. The Lévy-Khintchine formula extends in an obviousmanner to separable Hilbert spaces H. We interpret (·, ·) as the Hilbertspace inner product. The characteristics (b,A, ν) are defined similarly -A is a positive, symmetric linear operator on H.

Banach space. We can also extend to a separable Banach space Ehaving dual E∗. Define Φµ : E∗ → C by

Φµ(u) =

∫E

ei〈u,x〉µ(dx),

where 〈·, ·〉 is duality. Replacing (·, ·) with 〈·, ·〉 the Lévy-Khintchineformula extends naturally but care must be taken in defining a Lévymeasure!

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 24 / 40

Page 131: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 132: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 133: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 134: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 135: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 136: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 137: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 138: Koc1(dba)

Stable Laws

This is one of the most important subclasses of infinitely divisible laws.We consider the general central limit problem in dimension d = 1, solet (Yn,n ∈ N) be a sequence of real valued i.i.d. random variables andconsider the rescaled partial sums

Sn =Y1 + Y2 + · · ·+ Yn − bn

σn,

where (bn,n ∈ N) is an arbitrary sequence of real numbers and(σn,n ∈ N) an arbitrary sequence of positive numbers. We areinterested in the case where there exists a random variable X for which

limn→∞

P(Sn ≤ x) = P(X ≤ x), (1.7)

for all x ∈ R, i.e. (Sn,n ∈ N) converges in distribution to X . If eachbn = nm and σn =

√nσ for fixed m ∈ R, σ > 0 then X ∼ N(m, σ2) by

the usual Laplace - de-Moivre central limit theorem.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 25 / 40

Page 139: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 140: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 141: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 142: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 143: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 144: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 145: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 146: Koc1(dba)

More generally a random variable is said to be stable if it arises as alimit as in (1.7). It is not difficult to show that (1.7) is equivalent to thefollowing:-

There exist real valued sequences (cn,n ∈ N) and (dn,n ∈ N) witheach cn > 0 such that

X1 + X2 + · · ·+ Xnd= cnX + dn (1.8)

where X1, . . . ,Xn are independent copies of X . X is said to be strictlystable if each dn = 0.To see that (1.8)⇒ (1.7) take each Yj = Xj ,bn = dn and σn = cn. Infact it can be shown that the only possible choice of cn in (1.8) iscn = σn

1α , where 0 < α ≤ 2 and σ > 0. The parameter α plays a key

role in the investigation of stable random variables and is called theindex of stability.Note that (1.8) can also be expressed in the equivalent form

φX (u)n = eiudnφX (cnu),

for each u ∈ R.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 26 / 40

Page 147: Koc1(dba)

It follows immediately from (1.8) that all stable random variables areinfinitely divisible and the characteristics in the Lévy-Khintchineformula are given as follows:

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 27 / 40

Page 148: Koc1(dba)

It follows immediately from (1.8) that all stable random variables areinfinitely divisible and the characteristics in the Lévy-Khintchineformula are given as follows:

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 27 / 40

Page 149: Koc1(dba)

Theorem

If X is a stable real-valued random variable, then its characteristicsmust take one of the two following forms.

1 When α = 2, ν = 0 (so X ∼ N(b,A)).

2 When α 6= 2, A = 0 andν(dx) =

c1

x1+α1(0,∞)(x)dx +c2

|x |1+α1(−∞,0)(x)dx,

where c1 ≥ 0, c2 ≥ 0 and c1 + c2 > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 28 / 40

Page 150: Koc1(dba)

Theorem

If X is a stable real-valued random variable, then its characteristicsmust take one of the two following forms.

1 When α = 2, ν = 0 (so X ∼ N(b,A)).

2 When α 6= 2, A = 0 andν(dx) =

c1

x1+α1(0,∞)(x)dx +c2

|x |1+α1(−∞,0)(x)dx,

where c1 ≥ 0, c2 ≥ 0 and c1 + c2 > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 28 / 40

Page 151: Koc1(dba)

Theorem

If X is a stable real-valued random variable, then its characteristicsmust take one of the two following forms.

1 When α = 2, ν = 0 (so X ∼ N(b,A)).

2 When α 6= 2, A = 0 andν(dx) =

c1

x1+α1(0,∞)(x)dx +c2

|x |1+α1(−∞,0)(x)dx,

where c1 ≥ 0, c2 ≥ 0 and c1 + c2 > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 28 / 40

Page 152: Koc1(dba)

Theorem

If X is a stable real-valued random variable, then its characteristicsmust take one of the two following forms.

1 When α = 2, ν = 0 (so X ∼ N(b,A)).

2 When α 6= 2, A = 0 andν(dx) =

c1

x1+α1(0,∞)(x)dx +c2

|x |1+α1(−∞,0)(x)dx,

where c1 ≥ 0, c2 ≥ 0 and c1 + c2 > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 28 / 40

Page 153: Koc1(dba)

A careful transformation of the integrals in the Lévy-Khintchine formulagives a different form for the characteristic function which is often moreconvenient.

Theorem

A real-valued random variable X is stable if and only if there existsσ > 0,−1 ≤ β ≤ 1 and µ ∈ R such that for all u ∈ R,

1

φX (u) = exp[iµu − 1

2σ2u2

]when α = 2

2

φX (u) = exp[iµu − σα|u|α

(1− iβsgn(u) tan(

πα

2))]

when α 6= 1,2.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 29 / 40

Page 154: Koc1(dba)

A careful transformation of the integrals in the Lévy-Khintchine formulagives a different form for the characteristic function which is often moreconvenient.

Theorem

A real-valued random variable X is stable if and only if there existsσ > 0,−1 ≤ β ≤ 1 and µ ∈ R such that for all u ∈ R,

1

φX (u) = exp[iµu − 1

2σ2u2

]when α = 2

2

φX (u) = exp[iµu − σα|u|α

(1− iβsgn(u) tan(

πα

2))]

when α 6= 1,2.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 29 / 40

Page 155: Koc1(dba)

A careful transformation of the integrals in the Lévy-Khintchine formulagives a different form for the characteristic function which is often moreconvenient.

Theorem

A real-valued random variable X is stable if and only if there existsσ > 0,−1 ≤ β ≤ 1 and µ ∈ R such that for all u ∈ R,

1

φX (u) = exp[iµu − 1

2σ2u2

]when α = 2

2

φX (u) = exp[iµu − σα|u|α

(1− iβsgn(u) tan(

πα

2))]

when α 6= 1,2.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 29 / 40

Page 156: Koc1(dba)

A careful transformation of the integrals in the Lévy-Khintchine formulagives a different form for the characteristic function which is often moreconvenient.

Theorem

A real-valued random variable X is stable if and only if there existsσ > 0,−1 ≤ β ≤ 1 and µ ∈ R such that for all u ∈ R,

1

φX (u) = exp[iµu − 1

2σ2u2

]when α = 2

2

φX (u) = exp[iµu − σα|u|α

(1− iβsgn(u) tan(

πα

2))]

when α 6= 1,2.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 29 / 40

Page 157: Koc1(dba)

A careful transformation of the integrals in the Lévy-Khintchine formulagives a different form for the characteristic function which is often moreconvenient.

Theorem

A real-valued random variable X is stable if and only if there existsσ > 0,−1 ≤ β ≤ 1 and µ ∈ R such that for all u ∈ R,

1

φX (u) = exp[iµu − 1

2σ2u2

]when α = 2

2

φX (u) = exp[iµu − σα|u|α

(1− iβsgn(u) tan(

πα

2))]

when α 6= 1,2.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 29 / 40

Page 158: Koc1(dba)

Theorem3

φX (u) = exp[iµu − σ|u|

(1 + iβ

sgn(u) log(|u|))]

when α = 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 30 / 40

Page 159: Koc1(dba)

Theorem3

φX (u) = exp[iµu − σ|u|

(1 + iβ

sgn(u) log(|u|))]

when α = 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 30 / 40

Page 160: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 161: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 162: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 163: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 164: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 165: Koc1(dba)

It can be shown that E(X 2) <∞ if and only if α = 2 (i.e. X isGaussian) and E(|X |) <∞ if and only if 1 < α ≤ 2.All stable random variables have densities fX , which can in general beexpressed in series form. In three important cases, there are closedforms.

1 The Normal Distribution

α = 2, X ∼ N(µ, σ2).

2 The Cauchy Distribution

α = 1, β = 0 fX (x) =σ

π[(x − µ)2 + σ2].

3 The Lévy Distribution

α =12, β = 1 fX (x) =

( σ2π

) 12 1

(x − µ)32

exp(− σ

2(x − µ)

),

for x > µ.Dave Applebaum (Sheffield UK) Lecture 1 December 2011 31 / 40

Page 166: Koc1(dba)

In general the series representations are given in terms of a realvalued parameter λ.

For x > 0 and 0 < α < 1:

fX (x , λ) =1πx

∞∑k=1

Γ(kα + 1)

k !(−x−α)k sin

(kπ2

(λ− α)

)

For x > 0 and 1 < α < 2,

fX (x , λ) =1πx

∞∑k=1

Γ(kα−1 + 1)

k !(−x)k sin

(kπ2α

(λ− α)

)

In each case the formula for negative x is obtained by using

fX (−x , λ) = fX (x ,−λ), for x > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 32 / 40

Page 167: Koc1(dba)

In general the series representations are given in terms of a realvalued parameter λ.

For x > 0 and 0 < α < 1:

fX (x , λ) =1πx

∞∑k=1

Γ(kα + 1)

k !(−x−α)k sin

(kπ2

(λ− α)

)

For x > 0 and 1 < α < 2,

fX (x , λ) =1πx

∞∑k=1

Γ(kα−1 + 1)

k !(−x)k sin

(kπ2α

(λ− α)

)

In each case the formula for negative x is obtained by using

fX (−x , λ) = fX (x ,−λ), for x > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 32 / 40

Page 168: Koc1(dba)

In general the series representations are given in terms of a realvalued parameter λ.

For x > 0 and 0 < α < 1:

fX (x , λ) =1πx

∞∑k=1

Γ(kα + 1)

k !(−x−α)k sin

(kπ2

(λ− α)

)

For x > 0 and 1 < α < 2,

fX (x , λ) =1πx

∞∑k=1

Γ(kα−1 + 1)

k !(−x)k sin

(kπ2α

(λ− α)

)

In each case the formula for negative x is obtained by using

fX (−x , λ) = fX (x ,−λ), for x > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 32 / 40

Page 169: Koc1(dba)

In general the series representations are given in terms of a realvalued parameter λ.

For x > 0 and 0 < α < 1:

fX (x , λ) =1πx

∞∑k=1

Γ(kα + 1)

k !(−x−α)k sin

(kπ2

(λ− α)

)

For x > 0 and 1 < α < 2,

fX (x , λ) =1πx

∞∑k=1

Γ(kα−1 + 1)

k !(−x)k sin

(kπ2α

(λ− α)

)

In each case the formula for negative x is obtained by using

fX (−x , λ) = fX (x ,−λ), for x > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 32 / 40

Page 170: Koc1(dba)

In general the series representations are given in terms of a realvalued parameter λ.

For x > 0 and 0 < α < 1:

fX (x , λ) =1πx

∞∑k=1

Γ(kα + 1)

k !(−x−α)k sin

(kπ2

(λ− α)

)

For x > 0 and 1 < α < 2,

fX (x , λ) =1πx

∞∑k=1

Γ(kα−1 + 1)

k !(−x)k sin

(kπ2α

(λ− α)

)

In each case the formula for negative x is obtained by using

fX (−x , λ) = fX (x ,−λ), for x > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 32 / 40

Page 171: Koc1(dba)

Note that if a stable random variable is symmetric then Theorem 8yields

φX (u) = exp(−ρα|u|α) for all 0 < α ≤ 2, (1.9)

where ρ = σ(0 < α < 2) and ρ = σ√2, when α = 2, and we will write

X ∼ SαS in this case.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 33 / 40

Page 172: Koc1(dba)

Note that if a stable random variable is symmetric then Theorem 8yields

φX (u) = exp(−ρα|u|α) for all 0 < α ≤ 2, (1.9)

where ρ = σ(0 < α < 2) and ρ = σ√2, when α = 2, and we will write

X ∼ SαS in this case.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 33 / 40

Page 173: Koc1(dba)

Note that if a stable random variable is symmetric then Theorem 8yields

φX (u) = exp(−ρα|u|α) for all 0 < α ≤ 2, (1.9)

where ρ = σ(0 < α < 2) and ρ = σ√2, when α = 2, and we will write

X ∼ SαS in this case.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 33 / 40

Page 174: Koc1(dba)

One of the reasons why stable laws are so important in applications isthe nice decay properties of the tails. The case α = 2 is special in thatwe have exponential decay, indeed for a standard normal X there isthe elementary estimate

P(X > y) ∼ e−12 y2

√2πy

as y →∞,

When α 6= 2 we have the slower polynomial decay as expressed in thefollowing,

limy→∞

yαP(X > y) = Cα1 + β

2σα,

limy→∞

yαP(X < −y) = Cα1− β

2σα,

where Cα > 1. The relatively slow decay of the tails for non-Gaussianstable laws (“heavy tails”) makes them ideally suited for modelling awide range of interesting phenomena, some of which exhibit“long-range dependence”.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 34 / 40

Page 175: Koc1(dba)

One of the reasons why stable laws are so important in applications isthe nice decay properties of the tails. The case α = 2 is special in thatwe have exponential decay, indeed for a standard normal X there isthe elementary estimate

P(X > y) ∼ e−12 y2

√2πy

as y →∞,

When α 6= 2 we have the slower polynomial decay as expressed in thefollowing,

limy→∞

yαP(X > y) = Cα1 + β

2σα,

limy→∞

yαP(X < −y) = Cα1− β

2σα,

where Cα > 1. The relatively slow decay of the tails for non-Gaussianstable laws (“heavy tails”) makes them ideally suited for modelling awide range of interesting phenomena, some of which exhibit“long-range dependence”.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 34 / 40

Page 176: Koc1(dba)

One of the reasons why stable laws are so important in applications isthe nice decay properties of the tails. The case α = 2 is special in thatwe have exponential decay, indeed for a standard normal X there isthe elementary estimate

P(X > y) ∼ e−12 y2

√2πy

as y →∞,

When α 6= 2 we have the slower polynomial decay as expressed in thefollowing,

limy→∞

yαP(X > y) = Cα1 + β

2σα,

limy→∞

yαP(X < −y) = Cα1− β

2σα,

where Cα > 1. The relatively slow decay of the tails for non-Gaussianstable laws (“heavy tails”) makes them ideally suited for modelling awide range of interesting phenomena, some of which exhibit“long-range dependence”.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 34 / 40

Page 177: Koc1(dba)

One of the reasons why stable laws are so important in applications isthe nice decay properties of the tails. The case α = 2 is special in thatwe have exponential decay, indeed for a standard normal X there isthe elementary estimate

P(X > y) ∼ e−12 y2

√2πy

as y →∞,

When α 6= 2 we have the slower polynomial decay as expressed in thefollowing,

limy→∞

yαP(X > y) = Cα1 + β

2σα,

limy→∞

yαP(X < −y) = Cα1− β

2σα,

where Cα > 1. The relatively slow decay of the tails for non-Gaussianstable laws (“heavy tails”) makes them ideally suited for modelling awide range of interesting phenomena, some of which exhibit“long-range dependence”.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 34 / 40

Page 178: Koc1(dba)

One of the reasons why stable laws are so important in applications isthe nice decay properties of the tails. The case α = 2 is special in thatwe have exponential decay, indeed for a standard normal X there isthe elementary estimate

P(X > y) ∼ e−12 y2

√2πy

as y →∞,

When α 6= 2 we have the slower polynomial decay as expressed in thefollowing,

limy→∞

yαP(X > y) = Cα1 + β

2σα,

limy→∞

yαP(X < −y) = Cα1− β

2σα,

where Cα > 1. The relatively slow decay of the tails for non-Gaussianstable laws (“heavy tails”) makes them ideally suited for modelling awide range of interesting phenomena, some of which exhibit“long-range dependence”.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 34 / 40

Page 179: Koc1(dba)

Deeper mathematical investigations of heavy tails require themathematical technique of regular variation.

The generalisation of stability to random vectors is straightforward - justreplace X1, . . . ,Xn, X and each dn in (1.8) by vectors and the formulain Theorem 7 extends directly. Note however that when α 6= 2 in therandom vector version of Theorem 7, the Lévy measure takes the form

ν(dx) =c

|x |d+αdx

where c > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 35 / 40

Page 180: Koc1(dba)

Deeper mathematical investigations of heavy tails require themathematical technique of regular variation.

The generalisation of stability to random vectors is straightforward - justreplace X1, . . . ,Xn, X and each dn in (1.8) by vectors and the formulain Theorem 7 extends directly. Note however that when α 6= 2 in therandom vector version of Theorem 7, the Lévy measure takes the form

ν(dx) =c

|x |d+αdx

where c > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 35 / 40

Page 181: Koc1(dba)

Deeper mathematical investigations of heavy tails require themathematical technique of regular variation.

The generalisation of stability to random vectors is straightforward - justreplace X1, . . . ,Xn, X and each dn in (1.8) by vectors and the formulain Theorem 7 extends directly. Note however that when α 6= 2 in therandom vector version of Theorem 7, the Lévy measure takes the form

ν(dx) =c

|x |d+αdx

where c > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 35 / 40

Page 182: Koc1(dba)

Deeper mathematical investigations of heavy tails require themathematical technique of regular variation.

The generalisation of stability to random vectors is straightforward - justreplace X1, . . . ,Xn, X and each dn in (1.8) by vectors and the formulain Theorem 7 extends directly. Note however that when α 6= 2 in therandom vector version of Theorem 7, the Lévy measure takes the form

ν(dx) =c

|x |d+αdx

where c > 0.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 35 / 40

Page 183: Koc1(dba)

We can generalise the definition of stable random variables if weweaken the conditions on the random variables (Y (n),n ∈ N) in thegeneral central limit problem by requiring these to be independent, butno longer necessarily identically distributed.In this case (subject to a technical growth restriction) the limitingrandom variables are called self-decomposable (or of class L) and theyare also infinitely divisible.Alternatively a random variable X is self-decomposable if and only iffor each 0 < a < 1, there exists a random variable Ya which isindependent of X such that

X d= aX + Ya ⇔ φX (u) = φX (au)φYa(u),

for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 36 / 40

Page 184: Koc1(dba)

We can generalise the definition of stable random variables if weweaken the conditions on the random variables (Y (n),n ∈ N) in thegeneral central limit problem by requiring these to be independent, butno longer necessarily identically distributed.In this case (subject to a technical growth restriction) the limitingrandom variables are called self-decomposable (or of class L) and theyare also infinitely divisible.Alternatively a random variable X is self-decomposable if and only iffor each 0 < a < 1, there exists a random variable Ya which isindependent of X such that

X d= aX + Ya ⇔ φX (u) = φX (au)φYa(u),

for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 36 / 40

Page 185: Koc1(dba)

We can generalise the definition of stable random variables if weweaken the conditions on the random variables (Y (n),n ∈ N) in thegeneral central limit problem by requiring these to be independent, butno longer necessarily identically distributed.In this case (subject to a technical growth restriction) the limitingrandom variables are called self-decomposable (or of class L) and theyare also infinitely divisible.Alternatively a random variable X is self-decomposable if and only iffor each 0 < a < 1, there exists a random variable Ya which isindependent of X such that

X d= aX + Ya ⇔ φX (u) = φX (au)φYa(u),

for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 36 / 40

Page 186: Koc1(dba)

We can generalise the definition of stable random variables if weweaken the conditions on the random variables (Y (n),n ∈ N) in thegeneral central limit problem by requiring these to be independent, butno longer necessarily identically distributed.In this case (subject to a technical growth restriction) the limitingrandom variables are called self-decomposable (or of class L) and theyare also infinitely divisible.Alternatively a random variable X is self-decomposable if and only iffor each 0 < a < 1, there exists a random variable Ya which isindependent of X such that

X d= aX + Ya ⇔ φX (u) = φX (au)φYa(u),

for all u ∈ Rd .

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 36 / 40

Page 187: Koc1(dba)

An infinitely divisible law is self-decomposable if and only if the Lévymeasure is of the form:

ν(dx) =k(x)

|x |dx ,

where k is decreasing on (0,∞) and increasing on (−∞,0). There hasrecently been increasing interest in these distributions both from atheoretical and applied perspective. Examples include gamma, Pareto,Student-t , F and log-normal distributions.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 37 / 40

Page 188: Koc1(dba)

An infinitely divisible law is self-decomposable if and only if the Lévymeasure is of the form:

ν(dx) =k(x)

|x |dx ,

where k is decreasing on (0,∞) and increasing on (−∞,0). There hasrecently been increasing interest in these distributions both from atheoretical and applied perspective. Examples include gamma, Pareto,Student-t , F and log-normal distributions.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 37 / 40

Page 189: Koc1(dba)

An infinitely divisible law is self-decomposable if and only if the Lévymeasure is of the form:

ν(dx) =k(x)

|x |dx ,

where k is decreasing on (0,∞) and increasing on (−∞,0). There hasrecently been increasing interest in these distributions both from atheoretical and applied perspective. Examples include gamma, Pareto,Student-t , F and log-normal distributions.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 37 / 40

Page 190: Koc1(dba)

Heavy Tails - Subexponentiality and Regular Variation

Consider the aggregate behaviour of the sum of a large number of i.i.d.random variables,

This can be co-operative so that no one r.v. dominates→ familiarGaussian C.L.T. .

This can be dominated by the behaviour of a single randomvariable→ C.L.T. for stable laws.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 38 / 40

Page 191: Koc1(dba)

Heavy Tails - Subexponentiality and Regular Variation

Consider the aggregate behaviour of the sum of a large number of i.i.d.random variables,

This can be co-operative so that no one r.v. dominates→ familiarGaussian C.L.T. .

This can be dominated by the behaviour of a single randomvariable→ C.L.T. for stable laws.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 38 / 40

Page 192: Koc1(dba)

Heavy Tails - Subexponentiality and Regular Variation

Consider the aggregate behaviour of the sum of a large number of i.i.d.random variables,

This can be co-operative so that no one r.v. dominates→ familiarGaussian C.L.T. .

This can be dominated by the behaviour of a single randomvariable→ C.L.T. for stable laws.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 38 / 40

Page 193: Koc1(dba)

Heavy Tails - Subexponentiality and Regular Variation

Consider the aggregate behaviour of the sum of a large number of i.i.d.random variables,

This can be co-operative so that no one r.v. dominates→ familiarGaussian C.L.T. .

This can be dominated by the behaviour of a single randomvariable→ C.L.T. for stable laws.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 38 / 40

Page 194: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 195: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 196: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 197: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 198: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 199: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 200: Koc1(dba)

To capture the second type of behaviour, we say that a non-negativerandom variable X is subexponential if as x →∞

P(X1 + · · ·+ Xn > x) ∼ P(maxX1, . . . ,Xn > x),

where X1, . . . ,Xn are independent copies of X .A natural and more easily handled subclass of subexponential randomvariables are those of regular variation of index α.X ∈ R−α where α > 0 if

limx→∞

FX (cx)

FX (x)= cα,

where FX (x) := P(X > x).

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 39 / 40

Page 201: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40

Page 202: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40

Page 203: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40

Page 204: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40

Page 205: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40

Page 206: Koc1(dba)

e.g. X Pareto - parameters K , α > 0, FX (x) =

(K

K + x

)α. X is

self-decomposable.

Intuition: “Heavy tails” (subexponentiality) are due to “large jumps”.

Large jumps are governed by the tail of the Lévy measure (see Lecture3)

Fact (Tail equivalence): If X is infinitely divisible then FX ∈ R−α if andonly if ν ∈ R−α.

In this case limx→∞FX (x)

ν(x)= 1.

Dave Applebaum (Sheffield UK) Lecture 1 December 2011 40 / 40