exploratory statistics with r

338
New operational instruments for statistical exploration (=NOISE) New operational instruments for statistical exploration (=NOISE) Christian P. Robert Universit´ e Paris Dauphine http://www.ceremade.dauphine.fr/ ~ xian Licence MI2E, 2008–2009

Upload: christian-robert

Post on 07-May-2015

2.003 views

Category:

Education


0 download

DESCRIPTION

These are the slides in English of my exploratory statistics course in Paris Dauphine

TRANSCRIPT

Page 1: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

New operational instrumentsfor statistical exploration

(=NOISE)

Christian P. Robert

Universite Paris Dauphinehttp://www.ceremade.dauphine.fr/~xian

Licence MI2E, 2008–2009

Page 2: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Outline

1 Simulation of random variables

2 Monte Carlo Method and EM algorithm

3 Bootstrap Method

4 Rudiments of Nonparametric Statistics

Page 3: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Chapter 1 :Simulation of random variables

IntroductionRandom generatorNon-uniform distributions (1)Non-uniform distributions (2)Markovian methods

Page 4: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Introduction

Necessity to ”reproduce chance” on a computer

Evaluation of the behaviour of a complex system (network,computer program, queue, particle system, atmosphere,epidemics, economic actions, &tc)

Page 5: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Introduction

Necessity to ”reproduce chance” on a computer

Evaluation of the behaviour of a complex system (network,computer program, queue, particle system, atmosphere,epidemics, economic actions, &tc)

Determine probabilistic properties of a new statisticalprocedure or under an unknown distribution [bootstrap]

Page 6: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Introduction

Necessity to ”reproduce chance” on a computer

Evaluation of the behaviour of a complex system (network,computer program, queue, particle system, atmosphere,epidemics, economic actions, &tc)

Determine probabilistic properties of a new statisticalprocedure or under an unknown distribution [bootstrap]

Validation of a probabilistic model

Page 7: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Introduction

Necessity to ”reproduce chance” on a computer

Evaluation of the behaviour of a complex system (network,computer program, queue, particle system, atmosphere,epidemics, economic actions, &tc)

Determine probabilistic properties of a new statisticalprocedure or under an unknown distribution [bootstrap]

Validation of a probabilistic model

Approximation of an expectation/integral for a non-standarddistribution [Law of Large Numbers]

Page 8: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Introduction

Necessity to ”reproduce chance” on a computer

Evaluation of the behaviour of a complex system (network,computer program, queue, particle system, atmosphere,epidemics, economic actions, &tc)

Determine probabilistic properties of a new statisticalprocedure or under an unknown distribution [bootstrap]

Validation of a probabilistic model

Approximation of an expectation/integral for a non-standarddistribution [Law of Large Numbers]

Maximisation of a weakly regular function/likelihood

Page 9: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Example (TCL for the binomial distribution)

IfXn ∼ B(n, p) ,

Xn converges in distribution to the normal distribution:

√n (Xn/n − p)

n→∞ N

(0,

p(1 − p)

n

)

Page 10: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

n= 4

0.0 0.2 0.4 0.6 0.8 1.0

010

2030

n= 8

0.0 0.2 0.4 0.6 0.8 1.0

05

1015

2025

n= 16

0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9

05

1015

20

n= 32

0.2 0.3 0.4 0.5 0.6 0.7 0.8

02

46

810

14

n= 64

0.3 0.4 0.5 0.6

05

1015

2025

n= 128

0.35 0.40 0.45 0.50 0.55 0.60 0.65

05

1015

n= 256

0.40 0.45 0.50 0.55 0.60

05

1020

30

n= 512

0.44 0.46 0.48 0.50 0.52 0.54 0.56 0.58

05

1015

2025

n= 1024

0.46 0.48 0.50 0.52 0.54

010

2030

4050

Page 11: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Example (Stochastic minimisation)

Consider the function

h(x, y) = (x sin(20y) + y sin(20x))2 cosh(sin(10x)x)

+ (x cos(10y) − y sin(10x))2 cosh(cos(20y)y) ,

to be minimised.

Page 12: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Example (Stochastic minimisation)

Consider the function

h(x, y) = (x sin(20y) + y sin(20x))2 cosh(sin(10x)x)

+ (x cos(10y) − y sin(10x))2 cosh(cos(20y)y) ,

to be minimised. (I know that the global minimum is 0 for(x, y) = (0, 0).)

Page 13: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

-1

-0.5

0

0.5

1

X

-1

-0.5

0

0.5

1

Y

01

23

45

6Z

Page 14: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Example (Stochastic minimisation (2))

Instead of solving the first order equations

∂h(x, y)

∂x= 0 ,

∂h(x, y)

∂y= 0

and of checking that the second order conditions are met, we cangenerate a random sequence in R2

θj+1 = θj +αj

2βj∆h(θj , βjζj) ζj

Page 15: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Example (Stochastic minimisation (2))

Instead of solving the first order equations

∂h(x, y)

∂x= 0 ,

∂h(x, y)

∂y= 0

and of checking that the second order conditions are met, we cangenerate a random sequence in R2

θj+1 = θj +αj

2βj∆h(θj , βjζj) ζj

where

⋄ the ζj ’s are uniform on the unit circle x2 + y2 = 1;

⋄ ∆h(θ, ζ) = h(θ + ζ) − h(θ − ζ);

⋄ (αj) and (βj) converge to 0

Page 16: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

-0.2 0.0 0.2 0.4 0.6

0.2

0.4

0.6

0.8

Case when αj = 1/10 log(1 + j) et βj = 1/j

Page 17: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

The traveling salesman problem

A classical allocation problem:

Salesman who needs to visitn cities

Page 18: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

The traveling salesman problem

A classical allocation problem:

Salesman who needs to visitn cities

Traveling costs betweenpairs of cities known [anddifferent]

Page 19: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

The traveling salesman problem

A classical allocation problem:

Salesman who needs to visitn cities

Traveling costs betweenpairs of cities known [anddifferent]

Search of the optimumcircuit

Page 20: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

The traveling salesman problem

A classical allocation problem:

Salesman who needs to visitn cities

Traveling costs betweenpairs of cities known [anddifferent]

Search of the optimumcircuit

Page 21: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

An NP-complete problem

The traveling salesmanproblem is an example ofmathematical problems thatrequire explosive resolutiontimes

Page 22: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

An NP-complete problem

The traveling salesmanproblem is an example ofmathematical problems thatrequire explosive resolutiontimesNumber of possible circuits n!and exact solutions availablein O(2n) time

Page 23: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

An NP-complete problem

The traveling salesmanproblem is an example ofmathematical problems thatrequire explosive resolutiontimesNumber of possible circuits n!and exact solutions availablein O(2n) timeNumerous practicalconsequences (networks,integrated circuit design,genomic sequencing, &tc.)

Procter & Gamblecompetition, 1962

Page 24: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

An open problem

Exact solution for 15, 112German cities found in 2001 in22.6 CPU years.

Page 25: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

An open problem

Exact solution for 15, 112German cities found in 2001 in22.6 CPU years.

Exact solution for the 24, 978Sweedish cities found in 2004 in84.8 CPU years.

Page 26: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Resolution via simulation

The simulated annealing algorithm:Repeat

Random modifications of parts of the original circuit with costC0

Page 27: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Resolution via simulation

The simulated annealing algorithm:Repeat

Random modifications of parts of the original circuit with costC0

Evaluation of the cost C of the new circuit

Page 28: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Resolution via simulation

The simulated annealing algorithm:Repeat

Random modifications of parts of the original circuit with costC0

Evaluation of the cost C of the new circuit

Acceptation of the new circuit with probability

exp

{C0 − C

T

}∧ 1

Page 29: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Resolution via simulation

The simulated annealing algorithm:Repeat

Random modifications of parts of the original circuit with costC0

Evaluation of the cost C of the new circuit

Acceptation of the new circuit with probability

exp

{C0 − C

T

}∧ 1

T , temperature, is progressively reduced[Metropolis, 1953]

Page 30: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Illustration

Example (400 cities)

T = 1.2

Page 31: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Illustration

Example (400 cities)

T = 0.8

Page 32: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Illustration

Example (400 cities)

T = 0.4

Page 33: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Illustration

Example (400 cities)

T = 0.0

Page 34: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Option pricing

Complicated computation of expectations/average values ofoptions, E[CT ], necessary to evaluate the entry price(1 + r)−T E[CT ]

Page 35: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Option pricing

Complicated computation of expectations/average values ofoptions, E[CT ], necessary to evaluate the entry price(1 + r)−T E[CT ]

Example (European options)

Case whenCT = (ST − K)+

with

ST = S0 × Y1 × · · · × YT , Pr(Yi = u) = 1 − Pr(Yi = d) = p .

Page 36: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Option pricing

Complicated computation of expectations/average values ofoptions, E[CT ], necessary to evaluate the entry price(1 + r)−T E[CT ]

Example (European options)

Case whenCT = (ST − K)+

with

ST = S0 × Y1 × · · · × YT , Pr(Yi = u) = 1 − Pr(Yi = d) = p .

Resolution via the simulation of the binomial rv’s Yi

Page 37: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Option pricing (cont’d)

Example (Asian options)

Continuous time model where

CT =

(1

T

∫ T

0S(t)dt − K

)+

≈(

1

T

T∑

n=1

S(n) − K

)+

,

with

S(n + 1) = S(n) × exp {∆X(n + 1)} , ∆X(n)iid∼ N (0, σ2) .

Page 38: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Introduction

Option pricing (cont’d)

Example (Asian options)

Continuous time model where

CT =

(1

T

∫ T

0S(t)dt − K

)+

≈(

1

T

T∑

n=1

S(n) − K

)+

,

with

S(n + 1) = S(n) × exp {∆X(n + 1)} , ∆X(n)iid∼ N (0, σ2) .

Resolution via the simulation of the normal rv’s ∆Xi

Page 39: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

Pseudo-random generatorPivotal element of simulation techniques: they all require theavailability of uniform U (0, 1) random variables viatransformations

Page 40: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

Pseudo-random generatorPivotal element of simulation techniques: they all require theavailability of uniform U (0, 1) random variables viatransformations

Definition (Pseudo-random generator)

Un Pseudo-random generator is a deterministic Ψ from ]0, 1[ to]0, 1[ such that, for any starting value u0 and any n, the sequence

{u0, Ψ(u0), Ψ(Ψ(u0)), . . . ,Ψn(u0)}

behaves (statistically) like an iid sequence U (0, 1)

Page 41: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

Pseudo-random generatorPivotal element of simulation techniques: they all require theavailability of uniform U (0, 1) random variables viatransformations

Definition (Pseudo-random generator)

Un Pseudo-random generator is a deterministic Ψ from ]0, 1[ to]0, 1[ such that, for any starting value u0 and any n, the sequence

{u0, Ψ(u0), Ψ(Ψ(u0)), . . . ,Ψn(u0)}

behaves (statistically) like an iid sequence U (0, 1)

¡Paradox!

While avoiding randomness, the deterministic sequence(u0, u1 = Ψ(u0), . . . , un = Ψ(un−1))must resemble a random sequence!

Page 42: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

In R, use of the procedure

runif( )

Description:‘runif’ generates random deviates.Example:u = runif(20)‘Random.seed’ is an integer vector, containing the random numbergenerator (RNG) state for random number generation in R. It canbe saved and restored, but should not be altered by the user.

Page 43: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

500 520 540 560 580 600

0.0

0.2

0.4

0.6

0.8

1.0

uniform sample

0.0 0.2 0.4 0.6 0.8 1.0

0.0

0.5

1.0

1.5

Page 44: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

In C, use of the procedure

rand() / random()

SYNOPSIS# include <stdlib.h>long int random(void);DESCRIPTIONThe random() function uses a non-linear additive feedback randomnumber generator employing a default table of size 31 longintegers to return successive pseudo-random numbers in the rangefrom 0 to RAND MAX. The period of this random generator isvery large, approximately 16*((2**31)-1).RETURN VALUErandom() returns a value between 0 and RAND MAX.

Page 45: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

En Scilab, use of the procedure

rand()

rand() : with no arguments gives a scalar whose value changeseach time it is referenced. By default, random numbers areuniformly distributed in the interval (0,1). rand(’normal’) switchesto a normal distribution with mean 0 and variance 1.rand(’uniform’) switches back to the uniform distributionEXAMPLEx=rand(10,10,’uniform’)

Page 46: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

Example (A standard uniform generator)

The congruencial generator

D(x) = (ax + b) mod (M + 1).

has a period of M for proper choices of (a, b) and becomes agenerator on ]0, 1[ when dividing by M + 2

v = u*69069069 (1)

0.0 0.2 0.4 0.6 0.8 1.0

0.00.2

0.40.6

0.81.0

1.2

0.0 0.2 0.4 0.6 0.8 1.0

0.00.2

0.40.6

0.81.0

t

t+1

0.0 0.2 0.4 0.6 0.8 1.0

0.00.2

0.40.6

0.81.0

t

t+5

0.0 0.2 0.4 0.6 0.8 1.0

0.00.2

0.40.6

0.81.0

t

t+10

Page 47: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Random generator

Conclusion :

Use the appropriate random generator on the computer or thesoftware at hand instead of constructing a random generator ofpoor quality

Page 48: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Distributions different from the uniform distribution (1)

A problem formaly solved since

Page 49: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Distributions different from the uniform distribution (1)

A problem formaly solved since

Theorem (Generic inversion)

If U is a uniform random variable on [0, 1) and if FX is the cdf ofthe random variable X, then F−1

X (U) is distributed like X

Page 50: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Distributions different from the uniform distribution (1)

A problem formaly solved since

Theorem (Generic inversion)

If U is a uniform random variable on [0, 1) and if FX is the cdf ofthe random variable X, then F−1

X (U) is distributed like X

Proof. Indeed,

P (F−1X (U) ≤ x) = P (U ≤ FX(x)) = FX(x)

Page 51: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Distributions different from the uniform distribution (1)

A problem formaly solved since

Theorem (Generic inversion)

If U is a uniform random variable on [0, 1) and if FX is the cdf ofthe random variable X, then F−1

X (U) is distributed like X

Proof. Indeed,

P (F−1X (U) ≤ x) = P (U ≤ FX(x)) = FX(x)

Note. When FX is not strictly increasing, we can take

F−1X (u) = inf {x; FX(x) ≥ u}

Page 52: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Applications...

Binomial distribution, B(n, p),

FX(x) =∑

i≤x

(n

i

)pi(1 − p)n−i

and F−1X (u) can be obtained numerically

Page 53: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Applications...

Binomial distribution, B(n, p),

FX(x) =∑

i≤x

(n

i

)pi(1 − p)n−i

and F−1X (u) can be obtained numerically

Exponential distribution, E xp(λ),

FX(x) = 1 − exp(λx) et F−1X (u) = − log(u)/λ

Page 54: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Applications...

Binomial distribution, B(n, p),

FX(x) =∑

i≤x

(n

i

)pi(1 − p)n−i

and F−1X (u) can be obtained numerically

Exponential distribution, E xp(λ),

FX(x) = 1 − exp(λx) et F−1X (u) = − log(u)/λ

Cauchy distribution, C (0, 1),

FX(x) =1

πarctan(x)+

1

2et F−1

X (u) = tan(π(u−1/2))

Page 55: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Other transformations...

[Hint]

Find transforms linking the distribution of interest withsimpler/know distributions

Page 56: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Other transformations...

[Hint]

Find transforms linking the distribution of interest withsimpler/know distributions

Example (Box-Muller transform)

For the normal distribution N (0, 1), if X1, X2i.i.d.∼ N (0, 1),

X21 + X2

2 ∼ χ22, arctan(X1/X2) ∼ U ([0, 2π])

[Jacobian]Since the χ2

2 distribution is the same as the E xp(1/2) distribution,using a cdf inversion produces

X1 =√−2 log(U1) sin(2πU2) X2 =

√−2 log(U1) cos(2πU2)

Page 57: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Example

Student’s t and Fisher’s F distributions are natural byproducts ofthe generation of the normal and of the chi-square distributions.

Page 58: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Example

Student’s t and Fisher’s F distributions are natural byproducts ofthe generation of the normal and of the chi-square distributions.

Example

The Cauchy distribution can be derived from the normal

distribution as: if X1, X2i.i.d.∼ N (0, 1), then X1/X2 ∼ C (0, 1)

Page 59: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Example

The Beta distribution B(α, β), with density

fX(x) =Γ(α + β)

Γ(α)Γ(β)xα−1(1 − x)β−1 ,

can be derived from the Gamma distribution by: if X1 ∼ G a(α, 1),X2 ∼ G a(β, 1), then

X1

X1 + X2∼ B(α, β)

Page 60: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Multidimensional distributions

Consider the generation of

(X1, . . . , Xp) ∼ f(x1, . . . , xp)

in Rp with components that are not necessarily independent

Cascade rule

f(x1, . . . , xp) = f1(x1) × f2|1(x2|x1) . . . × fp|−p(xp|x1, . . . , xp−1)

Page 61: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (1)

Implementation

Simulate for t = 1, . . . , T

1 X1 ∼ f1(x1)

2 X2 ∼ f2|1(x2|x1)

...

p. Xp ∼ fp|−p(xp|x1, . . . , xp−1)

Page 62: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Distributions different from the uniform distribution (2)

F−1X rarely available

Page 63: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Distributions different from the uniform distribution (2)

F−1X rarely available

implemented algorithm in a resident software only forstandard distributions

Page 64: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Distributions different from the uniform distribution (2)

F−1X rarely available

implemented algorithm in a resident software only forstandard distributions

inversion lemma does not apply in larger dimensions

Page 65: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Distributions different from the uniform distribution (2)

F−1X rarely available

implemented algorithm in a resident software only forstandard distributions

inversion lemma does not apply in larger dimensions

new distributions may require fast resolution

Page 66: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

The Accept-Reject Algorithm

Given a distribution with density f to be simulated

Page 67: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

The Accept-Reject Algorithm

Given a distribution with density f to be simulated

Theorem (Fundamental theorem of simulation)

The uniform distribution on thesub-graph

Sf = {(x, u); 0 ≤ u ≤ f(x)}

produces a marginal in x withdensity f .

0 2 4 6 8 100.0

00.0

50.1

00.1

50.2

00.2

5

x

f(x)

Page 68: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Proof :

Marginal density given by

∫ ∞

0I0≤u≤f(x)du = f(x)

and independence from the normalisation constant

Page 69: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Proof :

Marginal density given by

∫ ∞

0I0≤u≤f(x)du = f(x)

and independence from the normalisation constant

Example

For a normal distribution, we just need to simulate (u, x) atrandom in

{(u, x); 0 ≤ u ≤ exp(−x2/2)}

Page 70: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Accept-reject algorithm

1 Find a density g that can be simulated and such that

supx

f(x)

g(x)= M < ∞

Page 71: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Accept-reject algorithm

1 Find a density g that can be simulated and such that

supx

f(x)

g(x)= M < ∞

2 Generate

Y1, Y2, . . .i.i.d.∼ g , U1, U2, . . .

i.i.d.∼ U ([0, 1])

Page 72: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Accept-reject algorithm

1 Find a density g that can be simulated and such that

supx

f(x)

g(x)= M < ∞

2 Generate

Y1, Y2, . . .i.i.d.∼ g , U1, U2, . . .

i.i.d.∼ U ([0, 1])

3 Take X = Yk where

k = inf{n ; Un ≤ f(Yn)/Mg(Yn)}

Page 73: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Theorem (Accept–reject)

The random variable produced by the above stopping rule isdistributed form fX

Page 74: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Theorem (Accept–reject)

The random variable produced by the above stopping rule isdistributed form fX

Proof (1) : We have

P (X ≤ x) =∞∑

k=1

P (X = Yk , Yk ≤ x)

=∞∑

k=1

(1 − 1

M

)k−1

P (Uk ≤ f(Yk)/Mg(Yk) , Yk ≤ x)

=∞∑

k=1

(1 − 1

M

)k−1 ∫ x

−∞

∫ f(y)/Mg(y)

0du g(y)dy

=∞∑

k=1

(1 − 1

M

)k−1 1

M

∫ x

−∞f(y)dy

Page 75: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Proof (2)

If (X, U) is uniform on A ⊃ B,the distribution of (X, U)restricted to B is uniform on B.

−4 −2 0 2 4

01

23

45

Page 76: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Properties

Does not require a normalisation constant

Page 77: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Properties

Does not require a normalisation constant

Does not require an exact upper bound M

Page 78: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Properties

Does not require a normalisation constant

Does not require an exact upper bound M

Allows for the recycling of the Yk’s for another density f (notethat rejected Yk’s are no longer distributed from g)

Page 79: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Properties

Does not require a normalisation constant

Does not require an exact upper bound M

Allows for the recycling of the Yk’s for another density f (notethat rejected Yk’s are no longer distributed from g)

Requires on average M Yk’s for one simulated X (efficiencymeasure)

Page 80: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Example

Take f(x) = exp(−x2/2) et g(x) = 1/(1 + x2)

f(x)

g(x)= (1 + x2) e−x2/2 ≤ 2/

√e

Probability of acceptance√

e/2π = 0.66

Page 81: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Theorem (Envelope)

If there exists a density gm, a function gl and a constant M suchthat

gl(x) ≤ f(x) ≤ Mgm(x) ,

then

1 Generate X ∼ gm(x), U ∼ U[0,1];

2 Accept X if U ≤ gl(X)/Mgm(X);

3 else, accept X if U ≤ f(X)/Mgm(X)

produces random variable distributed from f .

Page 82: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Uniform ratio algorithmsSlice sampler

Result :

Uniform simulation on

{(u, v); 0 ≤ u ≤√

2f(v/u)}

producesX = V/U ∼ f

Page 83: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Uniform ratio algorithmsSlice sampler

Result :

Uniform simulation on

{(u, v); 0 ≤ u ≤√

2f(v/u)}

producesX = V/U ∼ f

Proof :

Change of variable (u, v) → (x, u) with Jacobian u and marginaldistribution of x provided by

x ∼∫ √

2f(x)

0u du =

√2f(x)

2

2= f(x)

Page 84: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Non-uniform distributions (2)

Example

For a normal distribution,simulate (u, v) at random in

0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4

0.0

0.2

0.4

0.6

u

v

{(u, v); 0 ≤ u ≤√

2 e−v2/4u2} = {(u, v); v2 ≤ −4 u2 log(u/√

2)}

Page 85: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Slice sampler

If a uniform simulation on

G = {(u, x); 0 ≤ u ≤ f(x)}

is too complex [because of the inversion of x into u ≤ f(x)], wecan use instead a random walk on G:

Page 86: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Slice sampler

Simulate for t = 1, . . . , T

1 ω(t+1) ∼ U[0,f(x(t))];

2 x(t+1) ∼ UG(t+1) , where

G(t+1) = {y; f(y) ≥ ω(t+1)}.

0.0 0.2 0.4 0.6 0.8 1.0

0.0

00

0.0

02

0.0

04

0.0

06

0.0

08

0.0

10

xf(

x)

Page 87: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Justification

The random walk is exploring uniformly G:

If(U (t), X(t)) ∼ UG ,

then(U (t+1), X(t+1)) ∼ UG .

Page 88: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Proof:

Pr((U (t+1), X(t+1)) ∈ A × B)

=

∫ ∫ ∫

B

AI0≤u≤f(x)

I0≤u′≤f(x)

f(x)

If(x′)≥u′(x′)∫If(y)≥u′dy

d(x, u, x′, u′)

=

∫ ∫

B

Af(x)

I0≤u′≤f(x)

f(x)

If(x′)≥u′(x′)∫If(y)≥u′dy

d(x, x′, u′)

=

∫If(x)≥u′dx

B

A

If(x′)≥u′(x′)∫If(y)≥u′dy

d(x′, u′)

=

B

AIf(x′)≥u′≥0 d(x′, u′)

Page 89: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Normal distribution)

For the standard normal distribution,

f(x) ∝ exp(−x2/2),

a slice sampler is

ω|x ∼ U[0,exp(−x2/2)] ,

X|ω ∼ U[−√

−2 log(ω),√

−2 log(ω)]

Page 90: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Note

The technique also operates when f is replaced with

ϕ(x) ∝ f(x)

Page 91: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Note

The technique also operates when f is replaced with

ϕ(x) ∝ f(x)

It can be generalised to the case when f is decomposed in

f(x) =

p∏

i=1

fi(x)

Page 92: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Truncated normal distribution)

If we consider instead the truncated N (−3, 1) distributionrestricted to [0, 1], with density

f(x) =exp(−(x + 3)2/2)√

2π[Φ(4) − Φ(3)]∝ exp(−(x + 3)2/2) = ϕ(x) ,

Page 93: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Truncated normal distribution)

If we consider instead the truncated N (−3, 1) distributionrestricted to [0, 1], with density

f(x) =exp(−(x + 3)2/2)√

2π[Φ(4) − Φ(3)]∝ exp(−(x + 3)2/2) = ϕ(x) ,

a slice sampler is

ω|x ∼ U[0,exp(−(x+3)2/2)] ,

X|ω ∼ U[0,1∧{−3+

√−2 log(ω)}]

Page 94: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

The Metropolis–Hastings algorithm

Generalisation of the slice sampler to situations when the slicesampler cannot be easily implemented

Idea

Create a sequence (Xn)n such that, for n ‘large enough’, thedensity of Xn is close to f

Page 95: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

The Metropolis–Hastings algorithm (2)

If f is the density of interest, we pick a proposal conditional density

q(y|x)

such that

Page 96: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

The Metropolis–Hastings algorithm (2)

If f is the density of interest, we pick a proposal conditional density

q(y|x)

such that

it is easy to simulate

it is positive everywhere f is positive

Page 97: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Metropolis–Hastings

For a current value X(t) = x(t),

1 Generate Yt ∼ q(y|x(t)).

Page 98: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Metropolis–Hastings

For a current value X(t) = x(t),

1 Generate Yt ∼ q(y|x(t)).

2 Take

X(t+1) =

{Yt with proba. ρ(x(t), Yt),

x(t) with proba. 1 − ρ(x(t), Yt),

where

ρ(x, y) = min

{f(y)

f(x)

q(x|y)

q(y|x), 1

}.

Page 99: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accept moves to yt’s such that

f(yt)

q(yt|xt)≥ f(xt)

q(xt|yt)

Page 100: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accept moves to yt’s such that

f(yt)

q(yt|xt)≥ f(xt)

q(xt|yt)

Does not depend on normalising constants for both f andq(·|x) (if the later is independent from x)

Page 101: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accept moves to yt’s such that

f(yt)

q(yt|xt)≥ f(xt)

q(xt|yt)

Does not depend on normalising constants for both f andq(·|x) (if the later is independent from x)

Never accept values of yt such that f(yt) = 0

Page 102: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accept moves to yt’s such that

f(yt)

q(yt|xt)≥ f(xt)

q(xt|yt)

Does not depend on normalising constants for both f andq(·|x) (if the later is independent from x)

Never accept values of yt such that f(yt) = 0

The sequence (x(t))t can take repeatedly the same value

Page 103: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accept moves to yt’s such that

f(yt)

q(yt|xt)≥ f(xt)

q(xt|yt)

Does not depend on normalising constants for both f andq(·|x) (if the later is independent from x)

Never accept values of yt such that f(yt) = 0

The sequence (x(t))t can take repeatedly the same value

The X(t)’s are dependent (Marakovian) random variables

links

Page 104: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Justification

Joint distribution of (X(t), X(t+1))

If X(t) ∼ f(x(t)),

(X(t), X(t+1)) ∼ f(x(t)){

ρ(x(t), x(t+1)) × q(x(t+1)|x(t))

[Yt accepted]

+

∫ [1 − ρ(x(t), y)

]q(y|x(t)) dy Ix(t)(x(t+1))

}

[Yt rejected]

Page 105: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Balance condition

f(x) × ρ(x, y) × q(y|x) = f(x) min

{f(y)

f(x)

q(x|y)

q(y|x), 1

}q(y|x)

= min {f(y)q(x|y), f(x)q(y|x)}= f(y) × ρ(y, x) × q(x|y)

Page 106: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Balance condition

f(x) × ρ(x, y) × q(y|x) = f(x) min

{f(y)

f(x)

q(x|y)

q(y|x), 1

}q(y|x)

= min {f(y)q(x|y), f(x)q(y|x)}= f(y) × ρ(y, x) × q(x|y)

Thus the distribution of (X(t), X(t+1)) as the distribution of(X(t+1), X(t)) : if X(t) has the density f , then so does X(t+1)

Page 107: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Link with slice samplingThe slice sampler is a very special case of Metropolis-Hastingsalgorithm where the acceptance probability is always 1

Page 108: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Link with slice samplingThe slice sampler is a very special case of Metropolis-Hastingsalgorithm where the acceptance probability is always 1

1 for the generation of U ,

I0≤u′≤f(x)

I0≤u≤f(x)×

f(x)−1 I0≤u′≤f(x)

f(x)−1 I0≤u≤f(x)= 1

[joint density] [conditional density]

Page 109: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Link with slice samplingThe slice sampler is a very special case of Metropolis-Hastingsalgorithm where the acceptance probability is always 1

1 for the generation of U ,

I0≤u′≤f(x)

I0≤u≤f(x)×

f(x)−1 I0≤u′≤f(x)

f(x)−1 I0≤u≤f(x)= 1

[joint density] [conditional density]

2 pour la generation de X,

I0≤u≤f(y)

I0≤u≤f(x)×

I{z;u≤f(z)}(x)

I{z;u≤f(z)}(y)

∫{z;u≤f(z)} f(z) dz∫{z;u≤f(z)} f(z) dz

= 1

[joint density] [conditional density]

Page 110: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Independent proposals

Proposal q independent from X(t), denoted g as in Accept-Rejectalgorithms.

Page 111: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Independent proposals

Proposal q independent from X(t), denoted g as in Accept-Rejectalgorithms.

Independent Metropolis-Hastings

For the current value X(t) = x(t),

1 Generate Yt ∼ g(y)

2 Take

X(t+1) =

Yt with proba. min

{f(Yt) g(x(t))

f(x(t)) g(Yt), 1

},

x(t) otherwise.

Page 112: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Alternative to Accept-Reject

Page 113: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Alternative to Accept-Reject

Avoids the computation of max f(x)/g(x)

Page 114: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Alternative to Accept-Reject

Avoids the computation of max f(x)/g(x)

Accepts more often than Accept-Reject

Page 115: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Alternative to Accept-Reject

Avoids the computation of max f(x)/g(x)

Accepts more often than Accept-Reject

If xt achieves max f(x)/g(x), this is almost identical toAccept-Reject

Page 116: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Alternative to Accept-Reject

Avoids the computation of max f(x)/g(x)

Accepts more often than Accept-Reject

If xt achieves max f(x)/g(x), this is almost identical toAccept-Reject

Except that the sequence (xt) is not independent

Page 117: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Gamma distribution)

Generate a distribution Ga(α, β) from a proposalGa(⌊α⌋, b = ⌊α⌋/α), where ⌊α⌋ is the integer part of α (this is asum of exponentials)

1 Generate Yt ∼ Ga(⌊α⌋, ⌊α⌋/α)

2 Take

X(t+1) =

Yt with prob.

(Yt

x(t)exp

{x(t) − Yt

α

})α−⌊α⌋

x(t) else.

Page 118: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Random walk Metropolis–Hastings

ProposalYt = X(t) + εt,

where εt ∼ g, independent from X(t), and g symmetrical

Page 119: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Random walk Metropolis–Hastings

ProposalYt = X(t) + εt,

where εt ∼ g, independent from X(t), and g symmetrical

Instrumental distribution with density

g(y − x)

Motivation

local perturbation of X(t) / exploration of its neighbourhood

Page 120: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Random walk Metropolis–Hastings

Starting from X(t) = x(t)

1 Generate Yt ∼ g(y − x(t))

2 Take

X(t+1) =

Yt with proba. min

{1,

f(Yt)

f(x(t))

},

[symmetry of g]

x(t) otherwise

Page 121: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accepts higher point and sometimes lower points (seegradient algorithm)

Page 122: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accepts higher point and sometimes lower points (seegradient algorithm)

Depends on the dispersion de g

Page 123: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Properties

Always accepts higher point and sometimes lower points (seegradient algorithm)

Depends on the dispersion de g

Average robability of acceptance

=

∫ ∫min{f(x), f(y)}g(y − x) dxdy

close to 1 if g has a small variance [Danger!]far from 1 if g has a large variance [Re-Danger!]

Page 124: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Normal distribution)

Generate N (0, 1) based on a uniform perturbation on [−δ, δ]

Yt = X(t) + δωt

Page 125: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Normal distribution)

Generate N (0, 1) based on a uniform perturbation on [−δ, δ]

Yt = X(t) + δωt

Acceptance probability

ρ(x(t), yt) = exp{(x(t)2 − y2t )/2} ∧ 1.

Page 126: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Normal distribution (2))

Statistics based on 15000 simulations

δ 0.1 0.5 1.0

mean 0.399 −0.111 0.10variance 0.698 1.11 1.06

When δ ↑, faster exploration of the support of f .

Page 127: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Normal distribution (2))

Statistics based on 15000 simulations

δ 0.1 0.5 1.0

mean 0.399 −0.111 0.10variance 0.698 1.11 1.06

When δ ↑, faster exploration of the support of f .

-1 0 1 2

050

100150

200250

(a)

-1.5-1.0

-0.50.0

0.5

-2 0 2

0100

200300

400

(b)

-1.5-1.0

-0.50.0

0.5-3 -2 -1 0 1 2 3

0100

200300

400

(c)

-1.5-1.0

-0.50.0

0.5

3 samples with δ = 0.1, 0.5 and 1.0, with convergence ofempirical averages (over 15000 simulations).

Page 128: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Missing variable models

Special case when the density to simulate can be written as

f(x) =

Zf(x, z)dz

The random variable Z is then called missing data

Page 129: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Completion principe

Idea

Simulate f produces simulations from f

Page 130: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Completion principe

Idea

Simulate f produces simulations from f

If(X, Z) ∼ f(x, z) ,

marginalyX ∼ f(x)

Page 131: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Data Augmentation

Starting from x(t),

1. Simulate Z(t+1) ∼ fZ|X(z|x(t)) ;

2. Simuleater X(t+1) ∼ fX|Z(x|z(t+1)) .

Page 132: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Example (Mixture of distributions)

Consider the simulation (in R2) of the density f(µ1, µ2)proportional to

e−µ21−µ2

2 ×100∏

i=1

{0.3 e−(xi−µ1)2/2 + 0.7 e−(xi−µ2)2/2

}

when the xi’s are given/observed.

Page 133: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Echantillon de 0.3 N(2.5,1)+ 0.7 N(0,1)

x

−2 0 2 4

0.00

0.05

0.10

0.15

0.20

0.25

0.30

0.35

−0.4 −0.2 0.0 0.2 0.4

2.0

2.2

2.4

2.6

2.8

3.0

µ1

µ 2

Histogram of the xi’s and level set of f(µ1, µ2)

Page 134: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Completion (1)

Replace every sum in the density with an integral:

0.3 e−(xi−µ1)2/2 + 0.7 e−(xi−µ2)2/2 =

∫ (I[0,0.3 e−(xi−µ1)2/2]

(ui)

+I[0.3 e−(xi−µ1)2/2,0.3 e−(xi−µ1)2/2+0.7 e−(xi−µ2)2/2]

(ui))

dui

and simulate ((µ1, µ2), (U1, . . . , Un)) = (X, Z) via DataAugmentation

Page 135: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Completion (2)

Replace the Ui’s by the ξi’s , where

ξi =

{1 si Ui ≤ 0.3 e−(xi−µ1)2/2,

2 sinon

Then

Pr (ξi = 1|µ1, µ2) =0.3 e−(xi−µ1)2/2

0.3 e−(xi−µ1)2/2 + 0.7 e−(xi−µ2)2/2

= 1 − Pr (ξi = 2|µ1, µ2)

Page 136: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Conditioning (1)

The conditional distribution of Z = (ξ1, . . . , ξn) givenX = (µ1, µ2) is given by

Pr (ξi = 1|µ1, µ2) =0.3 e−(xi−µ1)2/2

0.3 e−(xi−µ1)2/2 + 0.7 e−(xi−µ2)2/2

= 1 − Pr (ξi = 2|µ1, µ2)

Page 137: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Simulation of random variables

Markovian methods

Conditioning (2)

The conditional distribution of X = (µ1, µ2) givenZ = (ξ1, . . . , ξn) is given by

(µ1, µ2)|Z ∼ e−µ21−µ2

2 ×∏

{i;ξi=1}

e−(xi−µ1)2/2 ×∏

{i;ξi=2}

e−(xi−µ2)2/2

∝ exp

{−(n1 + 2)

(µ1 −

n1µ1

n1 + 2

)2

/2

}

× exp

{−(n2 + 2)

(µ2 −

n2µ2

n2 + 2

)2

/2

}

where nj is the number of ξi’s equal to j and njµj is the sum ofthe xi’s associated with those ξi equal to j

[Easy!]

Page 138: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Chapter 2 :Monte Carlo Methods &

EM algorithm

IntroductionIntegration by Monte Carlo methodImportance functionsAcceleration methods

Page 139: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Uses of simulation

1 integration

I = Ef [h(X)] =

∫h(x)f(x)dx

Page 140: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Uses of simulation

1 integration

I = Ef [h(X)] =

∫h(x)f(x)dx

2 limiting behaviour/stationarity of complex systems

Page 141: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Uses of simulation

1 integration

I = Ef [h(X)] =

∫h(x)f(x)dx

2 limiting behaviour/stationarity of complex systems

3 optimisation

arg minx

h(x) = arg maxx

exp{−βh(x)} β > 0

Page 142: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Example (Propagation of an epidemic)

On a grid representing a region, a point is given by its coordinatesx, y

Page 143: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Example (Propagation of an epidemic)

On a grid representing a region, a point is given by its coordinatesx, yThe probability to catch a disease is

Px,y =exp(α + β · nx,y)

1 + exp(α + β · nx,y)Inx,y>0

if nx,y denotes the number of neighbours of (x, y) who alread havethis disease

Page 144: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Example (Propagation of an epidemic)

On a grid representing a region, a point is given by its coordinatesx, yThe probability to catch a disease is

Px,y =exp(α + β · nx,y)

1 + exp(α + β · nx,y)Inx,y>0

if nx,y denotes the number of neighbours of (x, y) who alread havethis diseaseThe probability to get healed is

Qx,y =exp(δ + γ · nx,y)

1 + exp(δ + γ · nx,y)

Page 145: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Introduction

Example (Propagation of an epidemic (2))

Question

Given (α, β, γ, δ), what is the speed of propagation of thisepidemic? the average duration? the number of infected persons?

Page 146: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Monte Carlo integration

Law of large numbers

If X1, . . . , Xn simulated from f ,

In =1

n

n∑

i=1

h(Xi) −→ I

Page 147: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Central Limit Theorem

Evaluation of the error b

σ2n =

1

n2

n∑

i=1

(h(Xi) − I)2

andIn ≈ N (I, σ2

n)

Page 148: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal)

For a Gaussian distribution, E[X4] = 3. Via Monte Carlointegration,

n 5 50 500 5000 50,000 500,000

In 1.65 5.69 3.24 3.13 3.038 3.029

5 10 50 100 500 1000 5000 10000 50000

0.00.5

1.01.5

2.02.5

3.0

n

In

Page 149: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Cauchy / Normal)

Consider the joint model

X|θ ∼ N (θ, 1), θ ∼ C(0, 1)

Once X is observed, θ is estimated by

δπ(x) =

∫ ∞

−∞

θ

1 + θ2e−(x−θ)2/2dθ

∫ ∞

−∞

1

1 + θ2e−(x−θ)2/2dθ

Page 150: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Cauchy / Normal (2))

This representation of δπ suggests using iid variables

θ1, · · · , θm ∼ N (x, 1)

and to compute

δπm(x) =

∑mi=1

θi

1 + θ2i

∑mi=1

1

1 + θ2i

.

Page 151: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Cauchy / Normal (2))

This representation of δπ suggests using iid variables

θ1, · · · , θm ∼ N (x, 1)

and to compute

δπm(x) =

∑mi=1

θi

1 + θ2i

∑mi=1

1

1 + θ2i

.

By vurtue of the Law of Large Numbers,

δπm(x) −→ δπ(x) quand m −→ ∞.

Page 152: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal cdf)

Approximation of the normal cdf

Φ(t) =

∫ t

−∞

1√2π

e−y2/2dy

by

Φ(t) =1

n

n∑

i=1

IXi≤t,

based on a sample of size n (X1, . . . , Xn), generated by thealgorithm of Box-Muller.

Page 153: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal cdf(2))

• VarianceΦ(t)(1 − Φ(t))/n,

since the variables IXi≤t are iid Bernoulli(Φ(t)).

Page 154: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal cdf(2))

• VarianceΦ(t)(1 − Φ(t))/n,

since the variables IXi≤t are iid Bernoulli(Φ(t)).

• For t close to t = 0 thea variance is about 1/4n:a precision of four decimals requires on average

√n =

√2 104

simulations, thus, 200 millions of iterations.

Page 155: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal cdf(2))

• VarianceΦ(t)(1 − Φ(t))/n,

since the variables IXi≤t are iid Bernoulli(Φ(t)).

• For t close to t = 0 thea variance is about 1/4n:a precision of four decimals requires on average

√n =

√2 104

simulations, thus, 200 millions of iterations.

• Larger [absolute] precision in the tails

Page 156: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Integration by Monte Carlo method

Example (Normal cdf(3))

n 0.0 0.67 0.84 1.28 1.65 2.32 2.58 3.09 3.72

102 0.485 0.74 0.77 0.9 0.945 0.985 0.995 1 1

103 0.4925 0.7455 0.801 0.902 0.9425 0.9885 0.9955 0.9985 1

104 0.4962 0.7425 0.7941 0.9 0.9498 0.9896 0.995 0.999 0.9999

105 0.4995 0.7489 0.7993 0.9003 0.9498 0.9898 0.995 0.9989 0.9999

106 0.5001 0.7497 0.8 0.9002 0.9502 0.99 0.995 0.999 0.9999

107 0.5002 0.7499 0.8 0.9001 0.9501 0.99 0.995 0.999 0.9999

108 0.5 0.75 0.8 0.9 0.95 0.99 0.995 0.999 0.9999

Evaluation of normal quantiles by Monte Carlo based on nnormal generations

Page 157: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Importance functions

Alternative representation :

I =

∫h(x)f(x)dx =

∫h(x)

f(x)

g(x)g(x)dx

Page 158: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Importance functions

Alternative representation :

I =

∫h(x)f(x)dx =

∫h(x)

f(x)

g(x)g(x)dx

Thus, if Y1, . . . , Yn simuated from g,

In =1

n

n∑

i=1

h(Yi)f(Yi)

g(Yi)−→ I

Page 159: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Appeal

Works for all g’s such that

supp(g) ⊃ supp(f)

Page 160: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Appeal

Works for all g’s such that

supp(g) ⊃ supp(f)

Possible improvement of the variance

Page 161: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Appeal

Works for all g’s such that

supp(g) ⊃ supp(f)

Possible improvement of the variance

Recycling of simulations Yi ∼ g for other densities f

Page 162: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Appeal

Works for all g’s such that

supp(g) ⊃ supp(f)

Possible improvement of the variance

Recycling of simulations Yi ∼ g for other densities f

Usage of simple distributions g

Page 163: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Normal)

For the normal distribution and the approximation of E[X4],

∫ ∞

−∞x4e−x2/2dx

[y=x2]= 2

∫ ∞

0y3/2 1

2e−y/2dy

suggests using g(y) = exp(−y/2)/2n 5 50 500 5000 50000

In 3.29 2.89 3.032 2.97 3.041

5 10 50 100 500 1000 5000 10000 50000

−0.1

0.00.1

0.20.3

0.40.5

n

In

Page 164: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Choice of the importance function

The “best” g function depends on the density f and on the hfunction

Page 165: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Choice of the importance function

The “best” g function depends on the density f and on the hfunction

Theorem (Optimal importance)

The choice of g that minimises the variance of In is

g⋆(x) =|h(x)|f(x)

I

Page 166: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Remarks

Finite variance only if

Ef

[h2(X)

f(X)

g(X)

]=

Xh2(x)

f(X)

g(X)dx < ∞ .

Page 167: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Remarks

Finite variance only if

Ef

[h2(X)

f(X)

g(X)

]=

Xh2(x)

f(X)

g(X)dx < ∞ .

Null variance for g⋆ if h s positive (!!)

Page 168: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Remarks

Finite variance only if

Ef

[h2(X)

f(X)

g(X)

]=

Xh2(x)

f(X)

g(X)dx < ∞ .

Null variance for g⋆ if h s positive (!!)

g⋆ depends on the very I we are trying to estimate (??)

Page 169: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Remarks

Finite variance only if

Ef

[h2(X)

f(X)

g(X)

]=

Xh2(x)

f(X)

g(X)dx < ∞ .

Null variance for g⋆ if h s positive (!!)

g⋆ depends on the very I we are trying to estimate (??)

Replacement of In by the harmonic mean

In =

∑ni=1 h(yi)/|h(yi)|∑n

i=1 1/|h(yi)|

(numerator and denominator are convergent)often poor (infinite variance)

Page 170: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Normal)

For the normal distribution and the approximation of E[X4],g⋆(x) ∝ x4 exp(−x2/2), distribution of the squared root of aG a(5/2, 1/2) rv

[Exercise]

n 5 50 500 5,000 50,000 500,000

In 4.877 2.566 2.776 2.317 2.897 3.160

1e+01 1e+02 1e+03 1e+04 1e+05

−10

12

n

In

Page 171: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Student’s t)

X ∼ T (ν, θ, σ2), with density

f(x) =Γ((ν + 1)/2)

σ√

νπ Γ(ν/2)

(1 +

(x − θ)2

νσ2

)−(ν+1)/2

.

Take θ = 0, σ = 1 and

I =

∫ ∞

2.1x5f(x)dx

is the integral of interest

Page 172: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Student’s t (2))

• Choice of importancefunctions

◦ f , since f = N (0,1)√χ2

ν/ν

◦ Cauchy C(0, 1)◦ Normal N (0, 1)◦ U ([0, 1/2.1])

Page 173: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Student’s t (2))

• Choice of importancefunctions

◦ f , since f = N (0,1)√χ2

ν/ν

◦ Cauchy C(0, 1)◦ Normal N (0, 1)◦ U ([0, 1/2.1])

Results:

◦ Uniform optimal

◦ Cauchy OK

◦ f and Normal poor

Page 174: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Importance functions

Example (Student’s t (2))

• Choice of importancefunctions

◦ f , since f = N (0,1)√χ2

ν/ν

◦ Cauchy C(0, 1)◦ Normal N (0, 1)◦ U ([0, 1/2.1])

Results:

◦ Uniform optimal

◦ Cauchy OK

◦ f and Normal poor

0 10000 20000 30000 40000 50000

5.05.5

6.06.5

7.0

0 10000 20000 30000 40000 50000

5.05.5

6.06.5

7.0

0 10000 20000 30000 40000 50000

5.05.5

6.06.5

7.0

0 10000 20000 30000 40000 50000

5.05.5

6.06.5

7.0

Page 175: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Correlated simulations

Negative correlation...

Two samples (X1, . . . , Xm) and (Y1, . . . , Ym) distributed from f inorder to estimate

I =

R

h(x)f(x)dx .

Both

I1 =1

m

m∑

i=1

h(Xi) et I2 =1

m

m∑

i=1

h(Yi)

have mean I and variance σ2

Page 176: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Correlated simulations (2)

...reduices the variance

The variance of the average is

Page 177: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Correlated simulations (2)

...reduices the variance

The variance of the average is

var

(I1 + I2

2

)=

σ2

2+

1

2cov(I1, I2).

Page 178: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Correlated simulations (2)

...reduices the variance

The variance of the average is

var

(I1 + I2

2

)=

σ2

2+

1

2cov(I1, I2).

Therefore, if both samples are negatively correlated,

cov(I1, I2) ≤ 0 ,

they do better than two independent samples with the same size

Page 179: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Antithetic variables

Construction of negatively correlated variables

1 If f symmetric about µ, take Yi = 2µ − Xi

2 If Xi = F−1(Ui), take Yi = F−1(1 − Ui)

3 If (Ai)i is a partition of X , partitionned sampling takes Xj ’sin each Ai (requires the knowledge of Pr(Ai))

Page 180: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Control variates

Take

I =

∫h(x)f(x)dx

to be computer and

I0 =

∫h0(x)f(x)dx

already knownWe nonetheless estimate I0 by I0 (and I by I)

Page 181: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Control variates (2)

Combined estimator

I∗ = I + β(I0 − I0)

I∗ is unbiased for I et

var(I∗) = var(I) + β2var(I) + 2βcov(I, I0)

Page 182: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Control variates (3)

Optimal choice of β

β⋆ = −cov(I, I0)

var(I0),

withvar(I⋆) = (1 − ρ2) var(I) ,

where ρ correlation between I and I0

Page 183: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Approximation of quantiles)

Consider the evaluation of

= Pr(X > a) =

∫ ∞

af(x)dx

by

ˆ =1

n

n∑

i=1

I(Xi > a), Xiiid∼ f

with Pr(X > µ) = 12

Page 184: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Approximation of quantiles (2))

The control variate

1

n

n∑

i=1

I(Xi > a) + β

(1

n

n∑

i=1

I(Xi > µ) − Pr(X > µ)

)

improves upon ˆ if

β < 0 et |β| < 2cov(δ1, δ3)

var(δ3)= 2

Pr(X > a)

Pr(X > µ).

Page 185: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Integration by conditioningTake advantage of the inequality

var(E[δ(X)|Y]) ≤ var(δ(X))

also called Rao-Blackwell Theorem

Consequence :

If I is an unbiased estimator of I = Ef [h(X)], with X simulatedfrom the joint density f(x, y), where

∫f(x, y)dy = f(x),

Page 186: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Integration by conditioningTake advantage of the inequality

var(E[δ(X)|Y]) ≤ var(δ(X))

also called Rao-Blackwell Theorem

Consequence :

If I is an unbiased estimator of I = Ef [h(X)], with X simulatedfrom the joint density f(x, y), where

∫f(x, y)dy = f(x),

the estimatorI∗ = Ef [I|Y1, . . . , Yn]

dominates I(X1, . . . , Xn) in terms of variance (and is alsounbiased)

Page 187: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Mean of a Student’s t)

Consider

E[h(x)] = E[exp(−x2)] avec X ∼ T (ν, 0, σ2)

Page 188: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Mean of a Student’s t)

Consider

E[h(x)] = E[exp(−x2)] avec X ∼ T (ν, 0, σ2)

Student’s t distribution can be simulated as

X|y ∼ N (µ, σ2y) and Y −1 ∼ χ2ν .

Page 189: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Mean of a Student’s t (2))

The empirical average

1

m

m∑

j=1

exp(−X2j ) ,

can be improved based on the joint sample

((X1, Y1), . . . , (Xm, Ym))

since

1

m

m∑

j=1

E[exp(−X2)|Yj ] =1

m

m∑

j=1

1√2σ2Yj + 1

is the conditional expectation

Page 190: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Monte Carlo Method and EM algorithm

Acceleration methods

Example (Mean of a Student’s t (3))

In this special case, the precision is ten times higher

0 2000 4000 6000 8000 10000

0.50

0.52

0.54

0.56

0.58

0.60

0 2000 4000 6000 8000 10000

0.50

0.52

0.54

0.56

0.58

0.60

Estimators of E[exp(−X2)]: empirical average (full lines)versus conditional expectation (dotted line) for(ν, µ, σ) = (4.6, 0, 1).

Page 191: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Chapter 3 :The Bootstrap Method

IntroductionGlivenko-Cantelli’s TheoremBootstrapParametric Bootstrap

Page 192: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Intrinsic randomness

Estimation from a random sample means uncertainty

Page 193: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Intrinsic randomness

Estimation from a random sample means uncertainty

Since based on a random sample, an estimator

δ(X1, . . . , Xn)

also is a random variable

Page 194: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Average variation

Question 1 :

How much does δ(X1, . . . , Xn) vary when the sample varies?

Page 195: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Average variation

Question 1 :

How much does δ(X1, . . . , Xn) vary when the sample varies?

Question 2 :

What is the variance of δ(X1, . . . , Xn) ?

Page 196: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Average variation

Question 1 :

How much does δ(X1, . . . , Xn) vary when the sample varies?

Question 2 :

What is the variance of δ(X1, . . . , Xn) ?

Question 3 :

What is the distribution of δ(X1, . . . , Xn) ?

Page 197: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Example (Normal sample)

Take X1, . . . , X100 a random sample from N (θ, 1). Its mean θ isestimated by

θ =1

100

100∑

i=1

Xi

Page 198: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Example (Normal sample)

Take X1, . . . , X100 a random sample from N (θ, 1). Its mean θ isestimated by

θ =1

100

100∑

i=1

Xi

Moyennes de 100 points pour 200 echantillons

x

−0.2 −0.1 0.0 0.1 0.2 0.3

01

23

45

6

Page 199: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Example (Normal sample)

Take X1, . . . , X100 a random sample from N (θ, 1). Its mean θ isestimated by

θ =1

100

100∑

i=1

Xi

Moyennes de 100 points pour 200 echantillons

x

−0.2 −0.1 0.0 0.1 0.2 0.3

01

23

45

6

Variation compatible with the (known) distributionθ ∼ N (θ, 1/100)

Page 200: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Associated problems

Observation of a single sample in most cases

Page 201: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Associated problems

Observation of a single sample in most cases

The sampling distribution is often unknown

Page 202: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Associated problems

Observation of a single sample in most cases

The sampling distribution is often unknown

The evaluation of the average variation of δ(X1, . . . , Xn) isparamount for the construction of confidence intervals and fortesting/answering questions like

H0 : θ ≤ 0

Page 203: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Associated problems

Observation of a single sample in most cases

The sampling distribution is often unknown

The evaluation of the average variation of δ(X1, . . . , Xn) isparamount for the construction of confidence intervals and fortesting/answering questions like

H0 : θ ≤ 0

In the normal case, the true θ stands with high probability inthe interval

[θ − 2σ, θ + 2σ] .

Page 204: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Introduction

Associated problems

Observation of a single sample in most cases

The sampling distribution is often unknown

The evaluation of the average variation of δ(X1, . . . , Xn) isparamount for the construction of confidence intervals and fortesting/answering questions like

H0 : θ ≤ 0

In the normal case, the true θ stands with high probability inthe interval

[θ − 2σ, θ + 2σ] .

Quid of σ ?!

Page 205: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Estimation of the repartition function

Extension/application of the LLN to the approximation of the cdf:

Page 206: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Estimation of the repartition function

Extension/application of the LLN to the approximation of the cdf:For a sample X1, . . . , Xn, if

Fn(x) =1

n

n∑

i=1

I]−∞,Xi](x)

=card {Xi; Xi ≤ x}

n,

Page 207: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Estimation of the repartition function

Extension/application of the LLN to the approximation of the cdf:For a sample X1, . . . , Xn, if

Fn(x) =1

n

n∑

i=1

I]−∞,Xi](x)

=card {Xi; Xi ≤ x}

n,

Fn(x) is a convergent estimator of the cdf F (x)[Glivenko–Cantelli]

Fn(x) −→ Pr(X ≤ x)

Page 208: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Example (Normal sample)

−2 −1 0 1 2

0.00.2

0.40.6

0.81.0

−2 −1 0 1 2

0.00.2

0.40.6

0.81.0

Estimation of the cdf F from a normal sample of 100 pointsand variation of this estimation over 200 normal samples

Page 209: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Properties

Estimator of a non-parametric nature : it is not necessary toknow the distribution or the shape of the distribution of thesample to derive this estimatorc© it is always available

Page 210: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Properties

Estimator of a non-parametric nature : it is not necessary toknow the distribution or the shape of the distribution of thesample to derive this estimatorc© it is always available

Robustess versus efficiency: If the [parameterised] shape ofthe distribution is known, there exists a better approximationbased on this shape, but if the shape is wrong, the result canbe completely off!

Page 211: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Example (Normal sample)

cdf of N (θ, 1), Φ(x − θ)

−2 −1 0 1 2

0.00.2

0.40.6

0.81.0

−2 −1 0 1 2

0.00.2

0.40.6

0.81.0

Estimation of Φ(· − θ) by Fn and by Φ(· − θ) based on 100points and maximal variation of thoses estimations over 200replications

Page 212: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Example (Non-normal sample)

Sample issued from

0.3N (0, 1) + 0.7N (2.5, 1)

wrongly allocated to a normal distribution Φ(· − θ)

Page 213: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

−2 −1 0 1 2 3 4

0.0

0.2

0.4

0.6

0.8

−2 −1 0 1 2 3 4

0.0

0.2

0.4

0.6

0.8

Estimation of Φ(· − θ) by Fn and by Φ(· − θ) based on 100points and maximal variation of thoses estimations over 200replications

Page 214: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Extension to functionals of F

For any quantity of the form

θ(F ) =

∫h(x) dF (x) ,

[Functional of the cdf]

Page 215: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Extension to functionals of F

For any quantity of the form

θ(F ) =

∫h(x) dF (x) ,

[Functional of the cdf]use of the approximation

θ(F ) = θ(Fn)

=

∫h(x) dFn(x)

=1

n

n∑

i=1

h(Xi)

[Moment estimator]

Page 216: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Example (Normal sample)

Since θ also is the median of N (θ, 1), θ can be chosen as themedian of Fn, equal to the median of X1, . . . , Xn, namely X(n/2)

Page 217: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Glivenko-Cantelli’s Theorem

Example (Normal sample)

Since θ also is the median of N (θ, 1), θ can be chosen as themedian of Fn, equal to the median of X1, . . . , Xn, namely X(n/2)

−0.4 −0.2 0.0 0.2 0.4

01

23

−0.4 −0.2 0.0 0.2 0.4

01

23

Histogramme des medianes

−0.4 −0.2 0.0 0.2 0.4

01

23

Histogramme des moyennes

Comparison of the variations of sample means and samplemedians over 200 normal samples

Page 218: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

How can one approximate the distribution of θ(Fn) ?

Principle

Since

θ(Fn) = θ(X1, . . . , Xn) with X1, . . . , Xni.i.d.∼ F

Page 219: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

How can one approximate the distribution of θ(Fn) ?

Principle

Since

θ(Fn) = θ(X1, . . . , Xn) with X1, . . . , Xni.i.d.∼ F

replace F with Fn :

θ(Fn) ≈ θ(X∗1 , . . . , X∗

n) with X∗1 , . . . , X∗

ni.i.d.∼ Fn

Page 220: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Implementation

Since Fn is known, it is possible to simulate from Fn, thereforeone can approximate the distribution of θ(X∗

1 , . . . , X∗n) [instead of

θ(X1, . . . , Xn)]The distribution corresponding to

Fn(x) = card {Xi; Xi ≤ x}/n

allocates a probability of 1/n to each point in {x1, . . . , xn} :

PrFn(X∗ = xi) = 1/n

Simulating from Fn is equivalent to sampling with replacementin (X1, . . . , Xn)

[in R, sample(x,n,replace=T)]

Page 221: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Monte Carlo implementation

1 For b = 1, . . . , B,1 generate a sample Xb

1, . . . ,Xbn from Fn

Page 222: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Monte Carlo implementation

1 For b = 1, . . . , B,1 generate a sample Xb

1, . . . ,Xbn from Fn

2 construct the corresponding value

θb = θ(Xb1, . . . ,X

bn)

Page 223: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Monte Carlo implementation

1 For b = 1, . . . , B,1 generate a sample Xb

1, . . . ,Xbn from Fn

2 construct the corresponding value

θb = θ(Xb1, . . . ,X

bn)

2 Use the sampleθ1, . . . , θB

to approximate the distribution of

θ(X1, . . . , Xn)

Page 224: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Notesbootstrap

the sample itself is used to build an evaluation of its distribution[Adventures of the Munchausen Baron ]

Page 225: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Notesbootstrap

the sample itself is used to build an evaluation of its distribution[Adventures of the Munchausen Baron ]

a bootstrap sample is obtained via n samplins withreplacement in (X1, . . . , Xn)

Page 226: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Notesbootstrap

the sample itself is used to build an evaluation of its distribution[Adventures of the Munchausen Baron ]

a bootstrap sample is obtained via n samplins withreplacement in (X1, . . . , Xn)

this sample can then take nn values (or(2n−1

n

)values if the

order does not matter)

Page 227: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Sample 0.3N (0, 1) + 0.7N (2.5, 1))

1.4 1.6 1.8 2.0 2.2

0.00.5

1.01.5

2.02.5

3.0

Variation of the empirical means over 200 bootstrap samplesversus observed average

Page 228: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Derivation of the average variation)

For an estimator θ(X1, . . . , Xn), the standard deviation is given by

η(F ) =

√EF [(θ(X1, . . . , Xn) − EF [θ(X1, . . . , Xn)])2]

Page 229: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Derivation of the average variation)

For an estimator θ(X1, . . . , Xn), the standard deviation is given by

η(F ) =

√EF [(θ(X1, . . . , Xn) − EF [θ(X1, . . . , Xn)])2]

and its bootstrap approximation is

η(Fn) =

√EFn [(θ(X1, . . . , Xn) − EFn [θ(X1, . . . , Xn)])2]

Page 230: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Derivation of the average variation (2))

Approximation itself approximated by

η(Fn) =

(1

B

B∑

b=1

(θ(Xb1, . . . , X

bn) − θ)2

)1/2

where

θ =1

B

B∑

b=1

θ(Xb1, . . . , X

bn)

Page 231: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Sample 0.3N (0, 1) + 0.7N (2.5, 1))

1.4 1.6 1.8 2.0 2.2

0.00.5

1.01.5

2.02.5

3.0

Interval of bootstrap variation at ±2η(Fn) and average of theobserved sample

Page 232: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Normal sample)

Sample

(X1, . . . , X100)i.i.d.∼ N (θ, 1)

Page 233: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

Example (Normal sample)

Sample

(X1, . . . , X100)i.i.d.∼ N (θ, 1)

Comparison of the confidence intervals

[x − 2 ∗ σx/10, x + 2 ∗ σx/10] = [−0.113, 0.327]

[normal approximation]

[x∗ − 2 ∗ σ∗, x∗ + 2 ∗ σ∗] = [−0.116, 0.336]

[normal bootstrap approximation]

[q∗(0.025), q∗(0.975)] = [−0.112, 0.336]

[generic bootstrap approximation]

Page 234: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Bootstrap

−0.2 −0.1 0.0 0.1 0.2 0.3 0.4

01

23

4

Approximation normale

Intervalle normal

Intervalle bootstrap

normal

Intervalle bootstrap generique

Variation ranges at 95% for a sample of 100 points and 200bootstrap replications

Page 235: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Parametric Bootstrap

If the parametric shape of F is known,

F (·) = Φλ(·) λ ∈ Λ ,

Page 236: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Parametric Bootstrap

If the parametric shape of F is known,

F (·) = Φλ(·) λ ∈ Λ ,

an evaluation of F more efficient than Fn is provided by

Φλn

where λn is a convergent estimator of λ[Cf Example 46]

Page 237: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Parametric Bootstrap

Approximation of the distribution of

θ(X1, . . . , Xn)

by the distribution of

θ(X∗1 , . . . , X∗

n) X∗1 , . . . , X∗

ni.i.d.∼ Φλn

Page 238: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Parametric Bootstrap

Approximation of the distribution of

θ(X1, . . . , Xn)

by the distribution of

θ(X∗1 , . . . , X∗

n) X∗1 , . . . , X∗

ni.i.d.∼ Φλn

May avoid simulation approximations in some cases

Page 239: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample )

TakeX1, . . . , Xn

i.i.d.∼ Exp(λ)

and λ = 1/Eλ[X] to be estimated

Page 240: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample )

TakeX1, . . . , Xn

i.i.d.∼ Exp(λ)

and λ = 1/Eλ[X] to be estimatedA possible estimator is

λ(x1, . . . , xn) =n∑n

i=1 xi

but this estimator is biased

Eλ[λ(X1, . . . , Xn)] 6= λ

Page 241: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (2))

Questions :

What is the bias

λ − Eλ[λ(X1, . . . , Xn)]

of this estimator ?

Page 242: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (2))

Questions :

What is the bias

λ − Eλ[λ(X1, . . . , Xn)]

of this estimator ?

What is the distribution of this estimator ?

Page 243: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Bootstrap evaluation of the bias

Example (Exponential Sample (3))

λ(x1, . . . , xn) − Eλ(x1,...,xn)[λ(X1, . . . , Xn)]

[parametric version]

λ(x1, . . . , xn) − EFn[λ(X1, . . . , Xn)]

[non-parametric version]

Page 244: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (4))

In the first (parametric) version,

1/λ(X1, . . . , Xn) ∼ Ga(n, nλ)

andEλ[λ(X1, . . . , Xn)] =

n

n − 1λ

Page 245: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (4))

In the first (parametric) version,

1/λ(X1, . . . , Xn) ∼ Ga(n, nλ)

andEλ[λ(X1, . . . , Xn)] =

n

n − 1λ

therefore the bias is analytically evaluated as

−λ/n − 1

and estimated by

− λ(X1, . . . , Xn)

n − 1= −0.00787

Page 246: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (5))

In the second (nonparametric) version, evaluation by Monte Carlo,

λ(x1, . . . , xn) − EFn[λ(X1, . . . , Xn)] = 0.00142

which achieves the “wrong” sign

Page 247: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (6))

Construction of a confidence interval on λBy parametric bootstrap,

Prλ

(λ1 ≤ λ ≤ λ2

)= Pr

(ω1 ≤ λ/λ ≤ ω2

)= 0.95

can be deduced from

λ/λ ∼ Ga(n, n)

[In R, qgamma(0.975,n,1/n)]

[λ1, λ2] = [0.452, 0.580]

Page 248: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (7))

In nonarametric bootstrap, one replaces

PrF (q(.025) ≤ λ(F ) ≤ q(.975)) = 0.95

withPrFn

(q∗(.025) ≤ λ(Fn) ≤ q∗(.975)

)= 0.95

Page 249: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Exponential Sample (7))

In nonarametric bootstrap, one replaces

PrF (q(.025) ≤ λ(F ) ≤ q(.975)) = 0.95

withPrFn

(q∗(.025) ≤ λ(Fn) ≤ q∗(.975)

)= 0.95

Approximation of quantiles q∗(.025) and q∗(.975) of λ(Fn) bybootstrap (Monte Carlo) sampling

[q∗(.025), q∗(.975)] = [0.454, 0.576]

Page 250: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

0.45 0.50 0.55 0.60

02

46

810

1214

Intervalle bootstrap

parametrique

Intervalle bootstrap

non−parametrique

Page 251: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Student Sample)

Take

X1, . . . , Xni.i.d.∼ T(5, µ, τ2)

def= µ + τ

N (0, 1)√χ2

5/5

µ and τ could be estimated by

µn =1

n

n∑

i=1

Xi τn =

√5 − 2

5

√√√√ 1

n

n∑

i=1

(Xi − µ)2

=

√5 − 2

5σn

Page 252: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Student Sample (2))

Problem

µn is not distributed from a Student T(5, µ, τ2/n) distributionThe distribution of µn ccan be reproduced by bootstrap sampling

Page 253: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

Example (Student Sample (3))

Comparison of confidence intervals

[µn − 2 ∗ σn/10, µn + 2 ∗ σn/10] = [−0.068, 0.319]

[normal approximation]

[q∗(0.05), q∗(0.95)] = [−0.056, 0.305]

[parametric boostrap approximation]

[q∗(0.05), q∗(0.95)] = [−0.094, 0.344]

[non parametric boostrap approximation]

Page 254: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Bootstrap Method

Parametric Bootstrap

−0.2 −0.1 0.0 0.1 0.2 0.3 0.4

01

23

45

6

Intervalle normal a 2 SD

Intervalle bootstrap

nonparametrique

01

23

45

Intervalle normal a 2 SD

Intervalle bootstrap

parametrique

95% variation interval for a 150 points sample with 400bootstrap replicas (top) nonparametric and (bottom)parametric

Page 255: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Chapter 4 :Rudiments of Nonparametric Statistics

IntroductionDensity EstimationNonparametric tests

Page 256: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Introduction

Probleme :

How could one conduct a statistical inference when the distributionof the data X1, . . . , Xn is unknown?

X1, . . . , Xni.i.d.∼ F

with F unknown

Page 257: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Introduction

Probleme :

How could one conduct a statistical inference when the distributionof the data X1, . . . , Xn is unknown?

X1, . . . , Xni.i.d.∼ F

with F unknown

Nonparametric setting in opposition to the parametric casewhen F (·) = Gθ(·) with only θ unknown

Page 258: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Introduction

Nonparametric Statistical Inference

Estimation of a quantity that depends on F

θ(F ) =

∫h(x) dF (x)

Page 259: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Introduction

Nonparametric Statistical Inference

Estimation of a quantity that depends on F

θ(F ) =

∫h(x) dF (x)

Decision on an hypothesis about F

F ∈ F0 ? F == F0 ? θ(F ) ∈ Θ0 ?

Page 260: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Introduction

Nonparametric Statistical Inference

Estimation of a quantity that depends on F

θ(F ) =

∫h(x) dF (x)

Decision on an hypothesis about F

F ∈ F0 ? F == F0 ? θ(F ) ∈ Θ0 ?

Estimation of functionals of F

F f(x) =dF

dx(x) EF [h(X1)|X2 = x]

Page 261: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Density Estimation

To estimate

f(x) =dF

dx(x)

[density of X]

Page 262: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Density Estimation

To estimate

f(x) =dF

dx(x)

[density of X]a natural solution is

fn(x) =dFn

dx(x)

Page 263: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Density Estimation

To estimate

f(x) =dF

dx(x)

[density of X]a natural solution is

fn(x) =dFn

dx(x)

butFn cannot be differentiated!

Page 264: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Histogram Estimation

A first solution is to reproduce the stepwise constant structure ofFn pour f

fn(x) =k∑

i=1

ωiI[ai,ai+1[(x) a1 < . . . < ak+1

Page 265: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Histogram Estimation

A first solution is to reproduce the stepwise constant structure ofFn pour f

fn(x) =k∑

i=1

ωiI[ai,ai+1[(x) a1 < . . . < ak+1

by picking the ωi’s such that

k∑

i=1

ωi(ai+1 − ai) = 1 et ωi(ai+1 − ai) = PF (X ∈ [ai, ai+1[)

Page 266: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Histogram Estimation (cont’d)

For instance,

ωi(ai+1 − ai) =1

n

n∑

i=1

I[ai,ai+1[(Xi)

= Fn(ai+1) − Fn(ai)

[bootstrap]

Page 267: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Histogram Estimation (cont’d)

For instance,

ωi(ai+1 − ai) =1

n

n∑

i=1

I[ai,ai+1[(Xi)

= Fn(ai+1) − Fn(ai)

[bootstrap]is a converging estimator of PF (X ∈ [ai, ai+1[)

Page 268: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Histogram Estimation (cont’d)

For instance,

ωi(ai+1 − ai) =1

n

n∑

i=1

I[ai,ai+1[(Xi)

= Fn(ai+1) − Fn(ai)

[bootstrap]is a converging estimator of PF (X ∈ [ai, ai+1[)

[Warning: side effects!]

Page 269: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

hist(x)$density

With R, hist(x)$density provides the values of ωi andhist(x)$breaks the values of the ai’s

It is better to use the valuesproduced by hist(x)$density to buildup a stepwise linear function byplot(hist(x)$density) rather than touse a stepwise constant function.

−2 −1 0 1 2 30

.00

.10

.20

.30

.40

.5

Histogram estimator fork = 45 and 450 normalobservations

Page 270: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Probabilist Interpretation

Starting with stepwise constant functions, the resultingapproximation of the distribution is a weighted sum of uniforms

k∑

i=1

πiU([ai, ai+1])

Page 271: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Probabilist Interpretation

Starting with stepwise constant functions, the resultingapproximation of the distribution is a weighted sum of uniforms

k∑

i=1

πiU([ai, ai+1])

Equivalent to a stepwise linear approximation of the cdf

Fn(x) =

n∑

i=1

πix − ai

ai+1 − aiI[ai,ai+1[(x)

Page 272: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Drawbacks

Depends on the choice of the partition (ai)i, often based onthe data itself (see R)

Page 273: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Drawbacks

Depends on the choice of the partition (ai)i, often based onthe data itself (see R)

Problem of the endpoints a1 and ak+1 : while not infinite(why?), they still must approximate the support of f

Page 274: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Drawbacks

Depends on the choice of the partition (ai)i, often based onthe data itself (see R)

Problem of the endpoints a1 and ak+1 : while not infinite(why?), they still must approximate the support of f

k and (ai)i must depend on n to allow for the convergence offn toward f

Page 275: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Drawbacks

Depends on the choice of the partition (ai)i, often based onthe data itself (see R)

Problem of the endpoints a1 and ak+1 : while not infinite(why?), they still must approximate the support of f

k and (ai)i must depend on n to allow for the convergence offn toward f

but... ai+1 − ai must not decrease too fast to 0 to allow forthe convergence of πi: there must be enough observations ineach interval [ai, ai+1]

Page 276: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Scott bandwidth

“Optimal” selection of the width of the classes :

hn = 3.5 σ n−1/3 et hn = 2.15 σ n−1/5

provide the right width ai+1 − ai (nclass = range(x) / h) for astepwise constant fn and a stepwise linear fn, respectively. (In thesense that they ensure the convergence of fn toward f when ngoes to ∞.)

Page 277: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Scott bandwidth

“Optimal” selection of the width of the classes :

hn = 3.5 σ n−1/3 et hn = 2.15 σ n−1/5

provide the right width ai+1 − ai (nclass = range(x) / h) for astepwise constant fn and a stepwise linear fn, respectively. (In thesense that they ensure the convergence of fn toward f when ngoes to ∞.)

[nclass=9 and nclass=12 in the next example]

Page 278: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

−2 −1 0 1 2 3

0.00

0.10

0.20

0.30

k = 5

−2 −1 0 1 2 3

0.0

0.1

0.2

0.3

0.4

k = 15

−2 −1 0 1 2 3

0.0

0.1

0.2

0.3

0.4

k = 25

−2 −1 0 1 2 3

0.0

0.1

0.2

0.3

0.4

k = 35

−2 −1 0 1 2 3

0.0

0.2

0.4

k = 45

−2 −1 0 1 2 3

0.0

0.2

0.4

k = 55

−2 −1 0 1 2 3

0.0

0.2

0.4

k = 65

−2 −1 0 1 2 3

0.0

0.2

0.4

k = 75

−2 −1 0 1 2 3

0.0

0.2

0.4

0.6

k = 85

Variation of the histogram estimators as a function of k for anormal sample with 450 observations

Page 279: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Kernel EstimatorStarting with the definition

f(x) =dF

dx(x) ,

we can also use the approximation

f(x) =Fn(x + δ) − Fn(x − δ)

=1

2δn

n∑

i=1

{IXi<x+δ − IXi<x−δ}

=1

2δn

n∑

i=1

I[−δ,δ](x − Xi)

when δ is small enough.[Positive point : f is a density]

Page 280: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Analytical and probabilistic interpretation

With this approximation

fn(x) =# observations close to x

2δn

Particular case of an histogram estimator where the ai’s are likeXj ± δ

Page 281: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Analytical and probabilistic interpretation

With this approximation

fn(x) =# observations close to x

2δn

Particular case of an histogram estimator where the ai’s are likeXj ± δ

Representation of fn as a weighted sum of uniforms

1

n

n∑

i=1

U([Xi − δ, Xi + δ])

[Note connection with bootstrap]

Page 282: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

−2 −1 0 1 2 3 4

0.0

0.2

0.4

0.6

0.8

bandwith 0.1

−2 0 2 4

0.0

0.1

0.2

0.3

0.4

bandwith 0.5

−2 0 2 4

0.00

0.10

0.20

0.30

bandwith 1

−10 −5 0 5 10

0.00

0.04

0.08

0.12

bandwith 10

Variation of uniform kernel estimators as a function of δ for anon-normal sample of 200 observations

Page 283: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Extension

Instead of a uniform approximation around each Xi, we can use asmoother distribution:

f(x) =1

δn

n∑

i=1

K

(x − Xi

δ

)

where K is a probability density (kernel) and δ a scale factor thatis small enough.

Page 284: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Extension

Instead of a uniform approximation around each Xi, we can use asmoother distribution:

f(x) =1

δn

n∑

i=1

K

(x − Xi

δ

)

where K is a probability density (kernel) and δ a scale factor thatis small enough.

With R, density(x)

Page 285: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Kernel selection

All densities are a priori acceptable. In practice (and with R, usageof

the normal/Gaussian kernel [kernel=”gaussian” or ”g”]

the Epanechnikov’s kernel [kernel=”epanechnikov” or ”e”]

K(y) = C {1 − y2}2 I[−1,1](y)

the triangular kernel [kernel=”triangular” or ”t”]

K(y) = (1 + y)I[−1,0](y) + (1 − y)I[0,1](y)

Page 286: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Kernel selection

All densities are a priori acceptable. In practice (and with R, usageof

the normal/Gaussian kernel [kernel=”gaussian” or ”g”]

the Epanechnikov’s kernel [kernel=”epanechnikov” or ”e”]

K(y) = C {1 − y2}2 I[−1,1](y)

the triangular kernel [kernel=”triangular” or ”t”]

K(y) = (1 + y)I[−1,0](y) + (1 − y)I[0,1](y)

Conclusion : Very little influence on the estimation of f (exceptfor the uniform kernel [kernel=”rectangular” or ”r”]).

Page 287: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

−4 −2 0 2 4 6

0.00

0.05

0.10

0.15

0.20

0.25

Noyau uniforme

−4 −2 0 2 4 6

0.00

0.05

0.10

0.15

0.20

0.25

Noyau triangulaire

−4 −2 0 2 4 6

0.00

0.05

0.10

0.15

0.20

0.25

Noyau normal

−4 −2 0 2 4 6

0.00

0.05

0.10

0.15

0.20

0.25

Noyau d’Epanechnikov

Variation of the kernel estimates with the kernel for anon-normal sample of 200 observations

Page 288: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Convergence to f

The choice of the bandwidth δ is crucial!

If δ large, many Xi contribute to the estimation of f(x)[Over-smoothing]

Page 289: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Convergence to f

The choice of the bandwidth δ is crucial!

If δ large, many Xi contribute to the estimation of f(x)[Over-smoothing]

If δ small, few Xi contribuent to the estimation of f(x)[Under-smoothing]

Page 290: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

−2 0 2 4

0.00

0.10

0.20

0.30

bandwith 0.5

−2 0 2 4

0.00

0.10

0.20

0.30

bandwith 1

−4 −2 0 2 4 6

0.00

0.05

0.10

0.15

0.20

0.25

bandwith 2.5

−6 −4 −2 0 2 4 6 8

0.00

0.05

0.10

0.15

0.20

bandwith 5

Variation of fn as a function of δ for a non-normal sample of200 observations

Page 291: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Optimal bandwidth

When considering the averaged integrated error

d(f, fn) = E

[∫{f(x) − fn(x)}2 dx

],

there exists an optimal choice for the bandwidth δ, denoted hn toindicate its dependance on n.

Page 292: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Optimal bandwidth (cont’d)

Using the decomposition

∫ {f(x) − E

[f(x)

]}2dx +

∫var{f(x)}dx ,

[Bias2+variance]and the approximations

f(x) − E

[f(x)

]≃ f ′′(x)

2h2

n

E

[exp{−(Xi − x)2/2h2

n}√2πhn

]≃ f(x) ,

[Exercise]

Page 293: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Optimal bandwidth (cont’d)

we deduce that the bias is of order

∫ {f ′′(x)

2

}2

dx h4n

and that the variance is approximately

1

nhn

√2π

∫f(x) dx =

1

nhn

√2π

[Exercise]

Page 294: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Optimal bandwidth (end’d)

Therefore, the error goes to 0 when n goes to ∞ if

1 hn goes to 0 and

2 nhn goes to infinity.

Page 295: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Optimal bandwidth (end’d)

Therefore, the error goes to 0 when n goes to ∞ if

1 hn goes to 0 and

2 nhn goes to infinity.

The optimal bandwidth is given by

h⋆n =

(√2π

∫ {f ′′(x)

}2dx n

)−1/5

Page 296: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Empirical bandwidth

Since the optimal bandwidth depends on f , unknown, we can usean approximation like

hn =0.9 min(σ, q75 − q25)

(1.34n)1/5,

where σ is the empirical standard deviation and q25 and q75 are theestimated 25% and 75% quantiles of X.

Page 297: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Empirical bandwidth

Since the optimal bandwidth depends on f , unknown, we can usean approximation like

hn =0.9 min(σ, q75 − q25)

(1.34n)1/5,

where σ is the empirical standard deviation and q25 and q75 are theestimated 25% and 75% quantiles of X.

Note : The values 0.9 and 1.34 are chose for the normal case.

Page 298: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Density Estimation

Empirical bandwidth

Since the optimal bandwidth depends on f , unknown, we can usean approximation like

hn =0.9 min(σ, q75 − q25)

(1.34n)1/5,

where σ is the empirical standard deviation and q25 and q75 are theestimated 25% and 75% quantiles of X.

Note : The values 0.9 and 1.34 are chose for the normal case.

Warning! This is not the defect bandwidth in R

Page 299: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests

Given a question about F , such asIs F equal to F0, a known distribution ?

Page 300: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests

Given a question about F , such asIs F equal to F0, a known distribution ?

the statistical answer is based on the data

X1, . . . , Xn ∼ F

to decide whether yes or no the question [the hypothesis] iscompatible with this data

Page 301: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests (cont’d)

A test procedure (or statistical test) ϕ(x1, . . . , xn) is takingvalues in {0, 1} (for yes/no)

Page 302: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests (cont’d)

A test procedure (or statistical test) ϕ(x1, . . . , xn) is takingvalues in {0, 1} (for yes/no)

When deciding about the question on F , there are two types oferrors:

1 refuse the hypothesis erroneously (Type I)

Page 303: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests (cont’d)

A test procedure (or statistical test) ϕ(x1, . . . , xn) is takingvalues in {0, 1} (for yes/no)

When deciding about the question on F , there are two types oferrors:

1 refuse the hypothesis erroneously (Type I)

2 accept the hypothesis erroneously (Type II)

Page 304: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests (cont’d)

A test procedure (or statistical test) ϕ(x1, . . . , xn) is takingvalues in {0, 1} (for yes/no)

When deciding about the question on F , there are two types oferrors:

1 refuse the hypothesis erroneously (Type I)

2 accept the hypothesis erroneously (Type II)

Both types of errors must then be balanced

Page 305: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The perspective of statistical tests (cont’d)

In pratice, a choice is made to concentrate upon type I errors andto reject the hypothesis only when the data is significantlyincompatibles with this hypothesis.

0 1 2 3 4

0.0

0.1

0.2

0.3

0.4

Acceptation

Rejet

To accept an hypothesis after a test only means that thedata has not rejected this hypothesis !!!

Page 306: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Comparison of distributions

Example (Two equal distributions?)

Given two samples X1, . . . , Xn and Y1, . . . , Ym, with respectivedistributions F and G, both unknownWhat is the answer to the question

F == G ?

Page 307: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Comparison of distributions (contd)

Example (Two equal distributions?)

Idea :

Compare the estimates of F and of G,

Fn(x) =1

n

n∑

i=1

IXi≤x et Gm(x) =1

m

m∑

i=1

IYi≤x

Page 308: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The Kolmogorov–Smirnov Statistics

−4 −2 0 2 4 6

0.0

0.2

0.4

0.6

0.8

1.0

Meme distribution difference maximale 0.05

−4 −2 0 2 4 60.

00.

20.

40.

60.

81.

0

Deux distributions difference maximale 0.14

Evaluation via the difference

K(m, n) = maxx

∣∣∣Fn(x) − Gm(x)∣∣∣ = max

Xi,Yj

∣∣∣Fn(x) − Gm(x)∣∣∣

Page 309: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

−4 −2 0 2 4 6

−0.0

20.

000.

020.

04

Meme distribution difference maximale 0.05

−4 −2 0 2 4 6

0.00

0.02

0.04

0.06

0.08

0.10

0.12

0.14

Deux distributions difference maximale 0.14

Evolution of the difference Fn(x) − Gm(x) in two cases

Page 310: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The Kolmogorov–Smirnov Statistics (2)

Usage :

If K(m, n) “large”, the distributions F and G are significativelydifferent.

Page 311: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The Kolmogorov–Smirnov Statistics (2)

Usage :

If K(m, n) “large”, the distributions F and G are significativelydifferent.If K(m, n) “small”, they cannot be distinguished on the dataX1, . . . , Xn and Y1, . . . , Ym, therefore F = G is acceptable

[Kolmogorov–Smirnov test]

Page 312: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

The Kolmogorov–Smirnov Statistics (2)

Usage :

If K(m, n) “large”, the distributions F and G are significativelydifferent.If K(m, n) “small”, they cannot be distinguished on the dataX1, . . . , Xn and Y1, . . . , Ym, therefore F = G is acceptable

[Kolmogorov–Smirnov test]

With R, ks.test(x,y)

Page 313: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Calibration of the test

For m and n fixed, if F = G, K(m, n) has a fixed distribution forall F ’s.

Page 314: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Calibration of the test

For m and n fixed, if F = G, K(m, n) has a fixed distribution forall F ’s.It is thus always possible to reduce the problem to the comparisonof two uniform samples and to use simulation to approximate thedistribution of K(m, n) and of its quantiles.

m=200,n=200

0.05 0.10 0.15

05

1015 Valeur

observee

Quantile a 95%

Page 315: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Calibration of the test (cont’d)

If the observed K(m, n) is above the 90 or 95 % quantile ofK(m, n) under H0 the value is very unlikely

if F = G

and the hypothesis of equality of both distributions is rejected.

Page 316: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Calibration of the test (cont’d)

Example of R output:Two-sample Kolmogorov-Smirnov testdata: z[, 1] and z[, 2]D = 0.05, p-value = 0.964alternative hypothesis: two.sided

Page 317: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Calibration of the test (cont’d)

Example of R output:Two-sample Kolmogorov-Smirnov testdata: z[, 1] and z[, 2]D = 0.05, p-value = 0.964alternative hypothesis: two.sided

p-value = 0.964 means that the probability that K(m, n) is largerthan the observed value D = 0.05 is 0.964, thus that the observedvalue is small under the distribution of K(m, n) : we thus acceptthe equality hypothesis.

Page 318: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Test of independance

Example (Independence)

Testing for independance between two rs’s X and Y based on theobservation of the pairs (X1, Y1), . . . , (Xn, Yn)

Page 319: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Test of independance

Example (Independence)

Testing for independance between two rs’s X and Y based on theobservation of the pairs (X1, Y1), . . . , (Xn, Yn)Question

X ⊥ Y ?

Page 320: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Rank test

Idea:

If the Xi’s are ordered

X(1) ≤ . . . X(n)

the ranks Ri (orders after the ranking of the Xi’s) of thecorresponding Yi’s

Y[1], . . . , Y[n],

must be completely random.

In R, rank(y[order(x)])

Page 321: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Rank test (cont’d)

Rank: The vectorR = (R1, . . . , Rn)

is called the rank statistic of the sample (Y[1], . . . Y[n])

Page 322: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Rank test (cont’d)

Rank: The vectorR = (R1, . . . , Rn)

is called the rank statistic of the sample (Y[1], . . . Y[n])Spearman’s statistic is

Sn =n∑

i=1

i Ri

[Correlation between i and Ri]

Page 323: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Rank test (cont’d)

Rank: The vectorR = (R1, . . . , Rn)

is called the rank statistic of the sample (Y[1], . . . Y[n])Spearman’s statistic is

Sn =n∑

i=1

i Ri

[Correlation between i and Ri]It is possible to prove that, if X ⊥ Y ,

E[Sn] =n(n + 1)2

4var(Sn) =

n2(n + 1)2(n − 1)

144

Page 324: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Spearman’s statistic

The distribution of Sn is available via [uniform] simulation or vianormal approximation

Distribution de S sur 500 echantillons de 200 points

−2 −1 0 1 2 3

0.0

0.1

0.2

0.3

0.4

Recentred version of Spearman’s statistics and normalapproximation

Page 325: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Spearman’s statistic (cont’d)

It is therefore possible to find the 5% and 95% quantiles of Sn

through simulation and to decide if the observed value of Sn isin-between those quantiles ( = Accept independance) or outside (= Reject independance)

Page 326: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Multinomial tests

Example (Chi-square test)

An histogram representation brings a robustified answer to testingproblems, like

Is the sample X1, . . . , Xn normal N (0, 1) ?

Idea:

Replace the original problem by its discretised version on intervals[ai, ai+1]

Is it true that

P (Xi ∈ [ai, ai+1]) =

∫ ai+1

ai

exp(−x2/2)√2π

dxdef= pi ?

Page 327: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Principle

Multinomial modelling

The problem is always expressed through a multinomial distribution

Mk

(p01, . . . , p

0k

)

or a family of multinomial distributions

Mk (p1(θ), . . . , pk(θ)) θ ∈ Θ

Page 328: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Examples

For testing the adequation to a standard normal distribution,N (0, 1), k is determined by the number of intervals [ai, ai+1]and the p0

i ’s by

p0i =

∫ ai+1

ai

exp(−x2/2)√2π

dx

Page 329: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Examples

For testing the adequation to a standard normal distribution,N (0, 1), k is determined by the number of intervals [ai, ai+1]and the p0

i ’s by

p0i =

∫ ai+1

ai

exp(−x2/2)√2π

dx

For testing the adequation to a normal distribution, N (θ, 1),the pi(θ) are given by

pi(θ) =

∫ ai+1

ai

exp(−(x − θ)2/2)√2π

dx

Page 330: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Examples (cont’d)

For testing the independance between two random variables,X et Y ,

X ⊥ Y ?

k is the number of cubes [ai, ai+1] × [bi, bi+1], θ is defined by

θ1i = P (X ∈ [ai, ai+1]) θ2i = P (Y ∈ [bi, bi+1])

and

pi,j(θ)def= P (X ∈ [ai, ai+1], Y ∈ [bi, bi+1])

= θ1i × θ2j

Page 331: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Chi-square testA natural estimator for the pi’s is

pi = P (X ∈ [ai, ai+1]) = Fn(ai+1) − Fn(ai)

[See bootstrap]

Page 332: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Chi-square testA natural estimator for the pi’s is

pi = P (X ∈ [ai, ai+1]) = Fn(ai+1) − Fn(ai)

[See bootstrap]The chi-square statistic is

Sn = nk∑

i=1

(pi − p0i )

2

p0i

=

k∑

i=1

(ni − np0i )

2

np0i

when testing the adequation to a multinomial distribution

Mk

(p01, . . . , p

0k

)

Page 333: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Chi-square test (cont’d)

and

Sn = nk∑

i=1

(pi − pi(θ))2

pi(θ)

=

k∑

i=1

(ni − npi(θ))2

npi(θ)

when testing the adequation to a family of multinomialdistributions

Mk (p1(θ), . . . , pk(θ)) θ ∈ Θ

Page 334: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Approximated distribution

For the adequation to a multinomial distribution, the distributionof Sn is approximately (for large n’s)

Sn ∼ χ2k−1

and for the adequation to a family of multinomial distributions,with dim(θ) = p,

Sn ∼ χ2k−p−1

Page 335: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

0 5 10 15 20

0.00

0.05

0.10

0.15

0.20

0.25

Distributions of Sn

Distribution of Sn for 200 normal samples of 100 points and atest of adequation to N (0, 1) with k = 4

Page 336: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Use and limits

The hypothesis under scrutiny is rejected if Sn is too large for aχ2

k−1 or χ2k−p−1 distribution

[In R, pchisq(S)]

Page 337: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

Use and limits

The hypothesis under scrutiny is rejected if Sn is too large for aχ2

k−1 or χ2k−p−1 distribution

[In R, pchisq(S)]Convergence (in n) to a χ2

k−1 (or χ2k−p−1) distribution is only

established for fixed k and (ai). In pratice, k and (ai) aredetermined by the observations, which reduces the validity of theapproximation.

Page 338: Exploratory Statistics with R

New operational instruments for statistical exploration (=NOISE)

Rudiments of Nonparametric Statistics

Nonparametric tests

−4 −2 0 2 4

−2−1

01

23

Normal Q−Q Plot

Quantile normal

Quan

tile o

bser

ve

0 5000 10000 15000 200000

5010

015

020

0

n

S n

QQ-plot of a non-normal sample and evolution of Sn as afunction of n for this sample