approximating bayes factors

125
On approximating Bayes factors On approximating Bayes factors Christian P. Robert Universit´ e Paris Dauphine & CREST-INSEE http://www.ceremade.dauphine.fr/ ~ xian CRiSM workshop on model uncertainty U. Warwick, May 30, 2010 Joint works with N. Chopin, J.-M. Marin, K. Mengersen & D. Wraith

Upload: christian-robert

Post on 02-Dec-2014

2.950 views

Category:

Technology


1 download

DESCRIPTION

Talk at CRiSM, Warwick, May 30

TRANSCRIPT

Page 1: Approximating Bayes Factors

On approximating Bayes factors

On approximating Bayes factors

Christian P. Robert

Universite Paris Dauphine & CREST-INSEEhttp://www.ceremade.dauphine.fr/~xian

CRiSM workshop on model uncertaintyU. Warwick, May 30, 2010

Joint works with N. Chopin, J.-M. Marin, K. Mengersen& D. Wraith

Page 2: Approximating Bayes Factors

On approximating Bayes factors

Outline

1 Introduction

2 Importance sampling solutions compared

3 Nested sampling

Page 3: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Model choice as model comparison

Choice between models

Several models available for the same observation

Mi : x ∼ fi(x|θi), i ∈ I

where I can be finite or infinite

Replace hypotheses with models

Page 4: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Model choice as model comparison

Choice between models

Several models available for the same observation

Mi : x ∼ fi(x|θi), i ∈ I

where I can be finite or infinite

Replace hypotheses with models

Page 5: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Bayesian model choice

Probabilise the entire model/parameter space

allocate probabilities pi to all models Mi

define priors πi(θi) for each parameter space Θi

compute

π(Mi|x) =pi

∫Θi

fi(x|θi)πi(θi)dθi∑j

pj

∫Θj

fj(x|θj)πj(θj)dθj

take largest π(Mi|x) to determine “best” model,

Page 6: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Bayesian model choice

Probabilise the entire model/parameter space

allocate probabilities pi to all models Mi

define priors πi(θi) for each parameter space Θi

compute

π(Mi|x) =pi

∫Θi

fi(x|θi)πi(θi)dθi∑j

pj

∫Θj

fj(x|θj)πj(θj)dθj

take largest π(Mi|x) to determine “best” model,

Page 7: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Bayesian model choice

Probabilise the entire model/parameter space

allocate probabilities pi to all models Mi

define priors πi(θi) for each parameter space Θi

compute

π(Mi|x) =pi

∫Θi

fi(x|θi)πi(θi)dθi∑j

pj

∫Θj

fj(x|θj)πj(θj)dθj

take largest π(Mi|x) to determine “best” model,

Page 8: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Model choice

Bayesian model choice

Probabilise the entire model/parameter space

allocate probabilities pi to all models Mi

define priors πi(θi) for each parameter space Θi

compute

π(Mi|x) =pi

∫Θi

fi(x|θi)πi(θi)dθi∑j

pj

∫Θj

fj(x|θj)πj(θj)dθj

take largest π(Mi|x) to determine “best” model,

Page 9: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Bayes factor

Bayes factor

Definition (Bayes factors)

For comparing model M0 with θ ∈ Θ0 vs. M1 with θ ∈ Θ1, underpriors π0(θ) and π1(θ), central quantity

B01 =π(Θ0|x)π(Θ1|x)

/π(Θ0)π(Θ1)

=

∫Θ0

f0(x|θ0)π0(θ0)dθ0∫Θ1

f1(x|θ)π1(θ1)dθ1

[Jeffreys, 1939]

Page 10: Approximating Bayes Factors

On approximating Bayes factors

Introduction

Evidence

Evidence

Problems using a similar quantity, the evidence

Zk =∫

Θk

πk(θk)Lk(θk) dθk,

aka the marginal likelihood.[Jeffreys, 1939]

Page 11: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

A comparison of importance sampling solutions

1 Introduction

2 Importance sampling solutions comparedRegular importanceBridge samplingHarmonic meansMixtures to bridgeChib’s solutionThe Savage–Dickey ratio

3 Nested sampling

[Marin & Robert, 2010]

Page 12: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Bayes factor approximation

When approximating the Bayes factor

B01 =

∫Θ0

f0(x|θ0)π0(θ0)dθ0∫Θ1

f1(x|θ1)π1(θ1)dθ1

use of importance functions $0 and $1 and

B01 =n−1

0

∑n0i=1 f0(x|θi0)π0(θi0)/$0(θi0)

n−11

∑n1i=1 f1(x|θi1)π1(θi1)/$1(θi1)

Page 13: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Diabetes in Pima Indian women

Example (R benchmark)

“A population of women who were at least 21 years old, of PimaIndian heritage and living near Phoenix (AZ), was tested fordiabetes according to WHO criteria. The data were collected bythe US National Institute of Diabetes and Digestive and KidneyDiseases.”200 Pima Indian women with observed variables

plasma glucose concentration in oral glucose tolerance test

diastolic blood pressure

diabetes pedigree function

presence/absence of diabetes

Page 14: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Probit modelling on Pima Indian women

Probability of diabetes function of above variables

P(y = 1|x) = Φ(x1β1 + x2β2 + x3β3) ,

Test of H0 : β3 = 0 for 200 observations of Pima.tr based on ag-prior modelling:

β ∼ N3(0, n(XTX)−1

)[Marin & Robert, 2007]

Page 15: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Probit modelling on Pima Indian women

Probability of diabetes function of above variables

P(y = 1|x) = Φ(x1β1 + x2β2 + x3β3) ,

Test of H0 : β3 = 0 for 200 observations of Pima.tr based on ag-prior modelling:

β ∼ N3(0, n(XTX)−1

)[Marin & Robert, 2007]

Page 16: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

MCMC 101 for probit models

Use of either a random walk proposal

β′ = β + ε

in a Metropolis-Hastings algorithm (since the likelihood isavailable) or of a Gibbs sampler that takes advantage of themissing/latent variable

z|y, x, β ∼ N (xTβ, 1){

Iyz≥0 × I1−yz≤0

}(since β|y,X, z is distributed as a standard normal)

[Gibbs three times faster]

Page 17: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

MCMC 101 for probit models

Use of either a random walk proposal

β′ = β + ε

in a Metropolis-Hastings algorithm (since the likelihood isavailable) or of a Gibbs sampler that takes advantage of themissing/latent variable

z|y, x, β ∼ N (xTβ, 1){

Iyz≥0 × I1−yz≤0

}(since β|y,X, z is distributed as a standard normal)

[Gibbs three times faster]

Page 18: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

MCMC 101 for probit models

Use of either a random walk proposal

β′ = β + ε

in a Metropolis-Hastings algorithm (since the likelihood isavailable) or of a Gibbs sampler that takes advantage of themissing/latent variable

z|y, x, β ∼ N (xTβ, 1){

Iyz≥0 × I1−yz≤0

}(since β|y,X, z is distributed as a standard normal)

[Gibbs three times faster]

Page 19: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Importance sampling for the Pima Indian dataset

Use of the importance function inspired from the MLE estimatedistribution

β ∼ N (β, Σ)

R Importance sampling codemodel1=summary(glm(y~-1+X1,family=binomial(link="probit")))

is1=rmvnorm(Niter,mean=model1$coeff[,1],sigma=2*model1$cov.unscaled)

is2=rmvnorm(Niter,mean=model2$coeff[,1],sigma=2*model2$cov.unscaled)

bfis=mean(exp(probitlpost(is1,y,X1)-dmvlnorm(is1,mean=model1$coeff[,1],

sigma=2*model1$cov.unscaled))) / mean(exp(probitlpost(is2,y,X2)-

dmvlnorm(is2,mean=model2$coeff[,1],sigma=2*model2$cov.unscaled)))

Page 20: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Importance sampling for the Pima Indian dataset

Use of the importance function inspired from the MLE estimatedistribution

β ∼ N (β, Σ)

R Importance sampling codemodel1=summary(glm(y~-1+X1,family=binomial(link="probit")))

is1=rmvnorm(Niter,mean=model1$coeff[,1],sigma=2*model1$cov.unscaled)

is2=rmvnorm(Niter,mean=model2$coeff[,1],sigma=2*model2$cov.unscaled)

bfis=mean(exp(probitlpost(is1,y,X1)-dmvlnorm(is1,mean=model1$coeff[,1],

sigma=2*model1$cov.unscaled))) / mean(exp(probitlpost(is2,y,X2)-

dmvlnorm(is2,mean=model2$coeff[,1],sigma=2*model2$cov.unscaled)))

Page 21: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Regular importance

Diabetes in Pima Indian womenComparison of the variation of the Bayes factor approximationsbased on 100 replicas for 20, 000 simulations from the prior andthe above MLE importance sampler

●●

Monte Carlo Importance sampling

23

45

Page 22: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Bridge sampling

Special case:If

π1(θ1|x) ∝ π1(θ1|x)π2(θ2|x) ∝ π2(θ2|x)

live on the same space (Θ1 = Θ2), then

B12 ≈1n

n∑i=1

π1(θi|x)π2(θi|x)

θi ∼ π2(θ|x)

[Gelman & Meng, 1998; Chen, Shao & Ibrahim, 2000]

Page 23: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Bridge sampling variance

The bridge sampling estimator does poorly if

var(B12)B2

12

≈ 1n

E

[(π1(θ)− π2(θ)

π2(θ)

)2]

is large, i.e. if π1 and π2 have little overlap...

Page 24: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Bridge sampling variance

The bridge sampling estimator does poorly if

var(B12)B2

12

≈ 1n

E

[(π1(θ)− π2(θ)

π2(θ)

)2]

is large, i.e. if π1 and π2 have little overlap...

Page 25: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

(Further) bridge sampling

General identity:

B12 =

∫π2(θ|x)α(θ)π1(θ|x)dθ∫π1(θ|x)α(θ)π2(θ|x)dθ

∀ α(·)

1n1

n1∑i=1

π2(θ1i|x)α(θ1i)

1n2

n2∑i=1

π1(θ2i|x)α(θ2i)θji ∼ πj(θ|x)

Page 26: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Optimal bridge sampling

The optimal choice of auxiliary function is

α? =n1 + n2

n1π1(θ|x) + n2π2(θ|x)

leading to

B12 ≈

1n1

n1∑i=1

π2(θ1i|x)n1π1(θ1i|x) + n2π2(θ1i|x)

1n2

n2∑i=1

π1(θ2i|x)n1π1(θ2i|x) + n2π2(θ2i|x)

Back later!

Page 27: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Optimal bridge sampling (2)

Reason:

Var(B12)B2

12

≈ 1n1n2

{∫π1(θ)π2(θ)[n1π1(θ) + n2π2(θ)]α(θ)2 dθ(∫

π1(θ)π2(θ)α(θ) dθ)2 − 1

}

(by the δ method)

Drawback: Dependence on the unknown normalising constantssolved iteratively

Page 28: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Optimal bridge sampling (2)

Reason:

Var(B12)B2

12

≈ 1n1n2

{∫π1(θ)π2(θ)[n1π1(θ) + n2π2(θ)]α(θ)2 dθ(∫

π1(θ)π2(θ)α(θ) dθ)2 − 1

}

(by the δ method)

Drawback: Dependence on the unknown normalising constantssolved iteratively

Page 29: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Extension to varying dimensions

When dim(Θ1) 6= dim(Θ2), e.g. θ2 = (θ1, ψ), introduction of apseudo-posterior density, ω(ψ|θ1, x), augmenting π1(θ1|x) intojoint distribution

π1(θ1|x)× ω(ψ|θ1, x)

on Θ2 so that

B12 =

∫π1(θ1|x)α(θ1, ψ)π2(θ1, ψ|x)dθ1ω(ψ|θ1, x) dψ∫π2(θ1, ψ|x)α(θ1, ψ)π1(θ1|x)dθ1 ω(ψ|θ1, x) dψ

= Eπ2

[π1(θ1)ω(ψ|θ1)π2(θ1, ψ)

]=

Eϕ [π1(θ1)ω(ψ|θ1)/ϕ(θ1, ψ)]Eϕ [π2(θ1, ψ)/ϕ(θ1, ψ)]

for any conditional density ω(ψ|θ1) and any joint density ϕ.

Page 30: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Extension to varying dimensions

When dim(Θ1) 6= dim(Θ2), e.g. θ2 = (θ1, ψ), introduction of apseudo-posterior density, ω(ψ|θ1, x), augmenting π1(θ1|x) intojoint distribution

π1(θ1|x)× ω(ψ|θ1, x)

on Θ2 so that

B12 =

∫π1(θ1|x)α(θ1, ψ)π2(θ1, ψ|x)dθ1ω(ψ|θ1, x) dψ∫π2(θ1, ψ|x)α(θ1, ψ)π1(θ1|x)dθ1 ω(ψ|θ1, x) dψ

= Eπ2

[π1(θ1)ω(ψ|θ1)π2(θ1, ψ)

]=

Eϕ [π1(θ1)ω(ψ|θ1)/ϕ(θ1, ψ)]Eϕ [π2(θ1, ψ)/ϕ(θ1, ψ)]

for any conditional density ω(ψ|θ1) and any joint density ϕ.

Page 31: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Illustration for the Pima Indian dataset

Use of the MLE induced conditional of β3 given (β1, β2) as apseudo-posterior and mixture of both MLE approximations on β3

in bridge sampling estimate

R bridge sampling codecova=model2$cov.unscaled

expecta=model2$coeff[,1]

covw=cova[3,3]-t(cova[1:2,3])%*%ginv(cova[1:2,1:2])%*%cova[1:2,3]

probit1=hmprobit(Niter,y,X1)

probit2=hmprobit(Niter,y,X2)

pseudo=rnorm(Niter,meanw(probit1),sqrt(covw))

probit1p=cbind(probit1,pseudo)

bfbs=mean(exp(probitlpost(probit2[,1:2],y,X1)+dnorm(probit2[,3],meanw(probit2[,1:2]),

sqrt(covw),log=T))/ (dmvnorm(probit2,expecta,cova)+dnorm(probit2[,3],expecta[3],

cova[3,3])))/ mean(exp(probitlpost(probit1p,y,X2))/(dmvnorm(probit1p,expecta,cova)+

dnorm(pseudo,expecta[3],cova[3,3])))

Page 32: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Illustration for the Pima Indian dataset

Use of the MLE induced conditional of β3 given (β1, β2) as apseudo-posterior and mixture of both MLE approximations on β3

in bridge sampling estimate

R bridge sampling codecova=model2$cov.unscaled

expecta=model2$coeff[,1]

covw=cova[3,3]-t(cova[1:2,3])%*%ginv(cova[1:2,1:2])%*%cova[1:2,3]

probit1=hmprobit(Niter,y,X1)

probit2=hmprobit(Niter,y,X2)

pseudo=rnorm(Niter,meanw(probit1),sqrt(covw))

probit1p=cbind(probit1,pseudo)

bfbs=mean(exp(probitlpost(probit2[,1:2],y,X1)+dnorm(probit2[,3],meanw(probit2[,1:2]),

sqrt(covw),log=T))/ (dmvnorm(probit2,expecta,cova)+dnorm(probit2[,3],expecta[3],

cova[3,3])))/ mean(exp(probitlpost(probit1p,y,X2))/(dmvnorm(probit1p,expecta,cova)+

dnorm(pseudo,expecta[3],cova[3,3])))

Page 33: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Bridge sampling

Diabetes in Pima Indian women (cont’d)Comparison of the variation of the Bayes factor approximationsbased on 100× 20, 000 simulations from the prior (MC), the abovebridge sampler and the above importance sampler

●●

MC Bridge IS

23

45

Page 34: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

The original harmonic mean estimator

When θki ∼ πk(θ|x),

1T

T∑t=1

1L(θkt|x)

is an unbiased estimator of 1/mk(x)[Newton & Raftery, 1994]

Highly dangerous: Most often leads to an infinite variance!!!

Page 35: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

The original harmonic mean estimator

When θki ∼ πk(θ|x),

1T

T∑t=1

1L(θkt|x)

is an unbiased estimator of 1/mk(x)[Newton & Raftery, 1994]

Highly dangerous: Most often leads to an infinite variance!!!

Page 36: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

“The Worst Monte Carlo Method Ever”

“The good news is that the Law of Large Numbers guarantees thatthis estimator is consistent ie, it will very likely be very close to thecorrect answer if you use a sufficiently large number of points fromthe posterior distribution.The bad news is that the number of points required for thisestimator to get close to the right answer will often be greaterthan the number of atoms in the observable universe. The evenworse news is that it’s easy for people to not realize this, and tonaıvely accept estimates that are nowhere close to the correctvalue of the marginal likelihood.”

[Radford Neal’s blog, Aug. 23, 2008]

Page 37: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

“The Worst Monte Carlo Method Ever”

“The good news is that the Law of Large Numbers guarantees thatthis estimator is consistent ie, it will very likely be very close to thecorrect answer if you use a sufficiently large number of points fromthe posterior distribution.The bad news is that the number of points required for thisestimator to get close to the right answer will often be greaterthan the number of atoms in the observable universe. The evenworse news is that it’s easy for people to not realize this, and tonaıvely accept estimates that are nowhere close to the correctvalue of the marginal likelihood.”

[Radford Neal’s blog, Aug. 23, 2008]

Page 38: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Approximating Zk from a posterior sample

Use of the [harmonic mean] identity

Eπk[

ϕ(θk)πk(θk)Lk(θk)

∣∣∣∣x] =∫

ϕ(θk)πk(θk)Lk(θk)

πk(θk)Lk(θk)Zk

dθk =1Zk

no matter what the proposal ϕ(·) is.[Gelfand & Dey, 1994; Bartolucci et al., 2006]

Direct exploitation of the MCMC output

Page 39: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Approximating Zk from a posterior sample

Use of the [harmonic mean] identity

Eπk[

ϕ(θk)πk(θk)Lk(θk)

∣∣∣∣x] =∫

ϕ(θk)πk(θk)Lk(θk)

πk(θk)Lk(θk)Zk

dθk =1Zk

no matter what the proposal ϕ(·) is.[Gelfand & Dey, 1994; Bartolucci et al., 2006]

Direct exploitation of the MCMC output

Page 40: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Comparison with regular importance sampling

Harmonic mean: Constraint opposed to usual importance samplingconstraints: ϕ(θ) must have lighter (rather than fatter) tails thanπk(θk)Lk(θk) for the approximation

Z1k = 1

/1T

T∑t=1

ϕ(θ(t)k )

πk(θ(t)k )Lk(θ

(t)k )

to have a finite variance.E.g., use finite support kernels (like Epanechnikov’s kernel) for ϕ

Page 41: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Comparison with regular importance sampling

Harmonic mean: Constraint opposed to usual importance samplingconstraints: ϕ(θ) must have lighter (rather than fatter) tails thanπk(θk)Lk(θk) for the approximation

Z1k = 1

/1T

T∑t=1

ϕ(θ(t)k )

πk(θ(t)k )Lk(θ

(t)k )

to have a finite variance.E.g., use finite support kernels (like Epanechnikov’s kernel) for ϕ

Page 42: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Comparison with regular importance sampling (cont’d)

Compare Z1k with a standard importance sampling approximation

Z2k =1T

T∑t=1

πk(θ(t)k )Lk(θ

(t)k )

ϕ(θ(t)k )

where the θ(t)k ’s are generated from the density ϕ(·) (with fatter

tails like t’s)

Page 43: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

HPD indicator as ϕUse the convex hull of MCMC simulations corresponding to the10% HPD region (easily derived!) and ϕ as indicator:

ϕ(θ) =10T

∑t∈HPD

Id(θ,θ(t))≤ε

[Robert & Wraith, 2009]

Page 44: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Harmonic means

Diabetes in Pima Indian women (cont’d)Comparison of the variation of the Bayes factor approximationsbased on 100 replicas for 20, 000 simulations for a simulation fromthe above harmonic mean sampler and importance samplers

●●

Harmonic mean Importance sampling

3.102

3.104

3.106

3.108

3.110

3.112

3.114

3.116

Page 45: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Approximating Zk using a mixture representation

Bridge sampling redux

Design a specific mixture for simulation [importance sampling]purposes, with density

ϕk(θk) ∝ ω1πk(θk)Lk(θk) + ϕ(θk) ,

where ϕ(·) is arbitrary (but normalised)Note: ω1 is not a probability weight

[Chopin & Robert, 2010]

Page 46: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Approximating Zk using a mixture representation

Bridge sampling redux

Design a specific mixture for simulation [importance sampling]purposes, with density

ϕk(θk) ∝ ω1πk(θk)Lk(θk) + ϕ(θk) ,

where ϕ(·) is arbitrary (but normalised)Note: ω1 is not a probability weight

[Chopin & Robert, 2010]

Page 47: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Approximating Z using a mixture representation (cont’d)

Corresponding MCMC (=Gibbs) sampler

At iteration t

1 Take δ(t) = 1 with probability

ω1πk(θ(t−1)k )Lk(θ

(t−1)k )

/(ω1πk(θ

(t−1)k )Lk(θ

(t−1)k ) + ϕ(θ(t−1)

k ))

and δ(t) = 2 otherwise;

2 If δ(t) = 1, generate θ(t)k ∼ MCMC(θ(t−1)

k , θk) whereMCMC(θk, θ′k) denotes an arbitrary MCMC kernel associatedwith the posterior πk(θk|x) ∝ πk(θk)Lk(θk);

3 If δ(t) = 2, generate θ(t)k ∼ ϕ(θk) independently

Page 48: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Approximating Z using a mixture representation (cont’d)

Corresponding MCMC (=Gibbs) sampler

At iteration t

1 Take δ(t) = 1 with probability

ω1πk(θ(t−1)k )Lk(θ

(t−1)k )

/(ω1πk(θ

(t−1)k )Lk(θ

(t−1)k ) + ϕ(θ(t−1)

k ))

and δ(t) = 2 otherwise;

2 If δ(t) = 1, generate θ(t)k ∼ MCMC(θ(t−1)

k , θk) whereMCMC(θk, θ′k) denotes an arbitrary MCMC kernel associatedwith the posterior πk(θk|x) ∝ πk(θk)Lk(θk);

3 If δ(t) = 2, generate θ(t)k ∼ ϕ(θk) independently

Page 49: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Approximating Z using a mixture representation (cont’d)

Corresponding MCMC (=Gibbs) sampler

At iteration t

1 Take δ(t) = 1 with probability

ω1πk(θ(t−1)k )Lk(θ

(t−1)k )

/(ω1πk(θ

(t−1)k )Lk(θ

(t−1)k ) + ϕ(θ(t−1)

k ))

and δ(t) = 2 otherwise;

2 If δ(t) = 1, generate θ(t)k ∼ MCMC(θ(t−1)

k , θk) whereMCMC(θk, θ′k) denotes an arbitrary MCMC kernel associatedwith the posterior πk(θk|x) ∝ πk(θk)Lk(θk);

3 If δ(t) = 2, generate θ(t)k ∼ ϕ(θk) independently

Page 50: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Evidence approximation by mixtures

Rao-Blackwellised estimate

ξ =1T

T∑t=1

ω1πk(θ(t)k )Lk(θ

(t)k )/ω1πk(θ

(t)k )Lk(θ

(t)k ) + ϕ(θ(t)

k ) ,

converges to ω1Zk/{ω1Zk + 1}Deduce Z3k from ω1Z3k/{ω1Z3k + 1} = ξ ie

Z3k =

∑Tt=1 ω1πk(θ

(t)k )Lk(θ

(t)k )/ω1π(θ(t)

k )Lk(θ(t)k ) + ϕ(θ(t)

k )

∑Tt=1 ϕ(θ(t)

k )/ω1πk(θ

(t)k )Lk(θ

(t)k ) + ϕ(θ(t)

k )

[Bridge sampler]

Page 51: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Mixtures to bridge

Evidence approximation by mixtures

Rao-Blackwellised estimate

ξ =1T

T∑t=1

ω1πk(θ(t)k )Lk(θ

(t)k )/ω1πk(θ

(t)k )Lk(θ

(t)k ) + ϕ(θ(t)

k ) ,

converges to ω1Zk/{ω1Zk + 1}Deduce Z3k from ω1Z3k/{ω1Z3k + 1} = ξ ie

Z3k =

∑Tt=1 ω1πk(θ

(t)k )Lk(θ

(t)k )/ω1π(θ(t)

k )Lk(θ(t)k ) + ϕ(θ(t)

k )

∑Tt=1 ϕ(θ(t)

k )/ω1πk(θ

(t)k )Lk(θ

(t)k ) + ϕ(θ(t)

k )

[Bridge sampler]

Page 52: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Chib’s representation

Direct application of Bayes’ theorem: given x ∼ fk(x|θk) andθk ∼ πk(θk),

Zk = mk(x) =fk(x|θk)πk(θk)

πk(θk|x)

Use of an approximation to the posterior

Zk = mk(x) =fk(x|θ∗k)πk(θ∗k)

πk(θ∗k|x).

Page 53: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Chib’s representation

Direct application of Bayes’ theorem: given x ∼ fk(x|θk) andθk ∼ πk(θk),

Zk = mk(x) =fk(x|θk)πk(θk)

πk(θk|x)

Use of an approximation to the posterior

Zk = mk(x) =fk(x|θ∗k)πk(θ∗k)

πk(θ∗k|x).

Page 54: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Case of latent variables

For missing variable z as in mixture models, natural Rao-Blackwellestimate

πk(θ∗k|x) =1T

T∑t=1

πk(θ∗k|x, z(t)k ) ,

where the z(t)k ’s are Gibbs sampled latent variables

Page 55: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Label switching

A mixture model [special case of missing variable model] isinvariant under permutations of the indices of the components.E.g., mixtures

0.3N (0, 1) + 0.7N (2.3, 1)

and0.7N (2.3, 1) + 0.3N (0, 1)

are exactly the same!c© The component parameters θi are not identifiablemarginally since they are exchangeable

Page 56: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Label switching

A mixture model [special case of missing variable model] isinvariant under permutations of the indices of the components.E.g., mixtures

0.3N (0, 1) + 0.7N (2.3, 1)

and0.7N (2.3, 1) + 0.3N (0, 1)

are exactly the same!c© The component parameters θi are not identifiablemarginally since they are exchangeable

Page 57: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Connected difficulties

1 Number of modes of the likelihood of order O(k!):c© Maximization and even [MCMC] exploration of the

posterior surface harder

2 Under exchangeable priors on (θ,p) [prior invariant underpermutation of the indices], all posterior marginals areidentical:c© Posterior expectation of θ1 equal to posterior expectation

of θ2

Page 58: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Connected difficulties

1 Number of modes of the likelihood of order O(k!):c© Maximization and even [MCMC] exploration of the

posterior surface harder

2 Under exchangeable priors on (θ,p) [prior invariant underpermutation of the indices], all posterior marginals areidentical:c© Posterior expectation of θ1 equal to posterior expectation

of θ2

Page 59: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

License

Since Gibbs output does not produce exchangeability, the Gibbssampler has not explored the whole parameter space: it lacksenergy to switch simultaneously enough component allocations atonce

0 100 200 300 400 500

−10

12

3

n

µ i

−1 0 1 2 3

0.20.3

0.40.5

µi

p i

0 100 200 300 400 500

0.20.3

0.40.5

n

p i

0.2 0.3 0.4 0.5

0.40.6

0.81.0

pi

σ i

0 100 200 300 400 500

0.40.6

0.81.0

n

σ i

0.4 0.6 0.8 1.0

−10

12

3

σi

µ i

Page 60: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Label switching paradox

We should observe the exchangeability of the components [labelswitching] to conclude about convergence of the Gibbs sampler.If we observe it, then we do not know how to estimate theparameters.If we do not, then we are uncertain about the convergence!!!

Page 61: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Label switching paradox

We should observe the exchangeability of the components [labelswitching] to conclude about convergence of the Gibbs sampler.If we observe it, then we do not know how to estimate theparameters.If we do not, then we are uncertain about the convergence!!!

Page 62: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Label switching paradox

We should observe the exchangeability of the components [labelswitching] to conclude about convergence of the Gibbs sampler.If we observe it, then we do not know how to estimate theparameters.If we do not, then we are uncertain about the convergence!!!

Page 63: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Compensation for label switching

For mixture models, z(t)k usually fails to visit all configurations in a

balanced way, despite the symmetry predicted by the theory

πk(θk|x) = πk(σ(θk)|x) =1k!

∑σ∈S

πk(σ(θk)|x)

for all σ’s in Sk, set of all permutations of {1, . . . , k}.Consequences on numerical approximation, biased by an order k!Recover the theoretical symmetry by using

πk(θ∗k|x) =1T k!

∑σ∈Sk

T∑t=1

πk(σ(θ∗k)|x, z(t)k ) .

[Berkhof, Mechelen, & Gelman, 2003]

Page 64: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Compensation for label switching

For mixture models, z(t)k usually fails to visit all configurations in a

balanced way, despite the symmetry predicted by the theory

πk(θk|x) = πk(σ(θk)|x) =1k!

∑σ∈S

πk(σ(θk)|x)

for all σ’s in Sk, set of all permutations of {1, . . . , k}.Consequences on numerical approximation, biased by an order k!Recover the theoretical symmetry by using

πk(θ∗k|x) =1T k!

∑σ∈Sk

T∑t=1

πk(σ(θ∗k)|x, z(t)k ) .

[Berkhof, Mechelen, & Gelman, 2003]

Page 65: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Galaxy dataset

n = 82 galaxies as a mixture of k normal distributions with bothmean and variance unknown.

[Roeder, 1992]Average density

data

Rel

ativ

e F

requ

ency

−2 −1 0 1 2 3

0.0

0.2

0.4

0.6

0.8

Page 66: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Galaxy dataset (k)Using only the original estimate, with θ∗k as the MAP estimator,

log(mk(x)) = −105.1396

for k = 3 (based on 103 simulations), while introducing thepermutations leads to

log(mk(x)) = −103.3479

Note that−105.1396 + log(3!) = −103.3479

k 2 3 4 5 6 7 8

mk(x) -115.68 -103.35 -102.66 -101.93 -102.88 -105.48 -108.44

Estimations of the marginal likelihoods by the symmetrised Chib’sapproximation (based on 105 Gibbs iterations and, for k > 5, 100permutations selected at random in Sk).

[Lee, Marin, Mengersen & Robert, 2008]

Page 67: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Galaxy dataset (k)Using only the original estimate, with θ∗k as the MAP estimator,

log(mk(x)) = −105.1396

for k = 3 (based on 103 simulations), while introducing thepermutations leads to

log(mk(x)) = −103.3479

Note that−105.1396 + log(3!) = −103.3479

k 2 3 4 5 6 7 8

mk(x) -115.68 -103.35 -102.66 -101.93 -102.88 -105.48 -108.44

Estimations of the marginal likelihoods by the symmetrised Chib’sapproximation (based on 105 Gibbs iterations and, for k > 5, 100permutations selected at random in Sk).

[Lee, Marin, Mengersen & Robert, 2008]

Page 68: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Galaxy dataset (k)Using only the original estimate, with θ∗k as the MAP estimator,

log(mk(x)) = −105.1396

for k = 3 (based on 103 simulations), while introducing thepermutations leads to

log(mk(x)) = −103.3479

Note that−105.1396 + log(3!) = −103.3479

k 2 3 4 5 6 7 8

mk(x) -115.68 -103.35 -102.66 -101.93 -102.88 -105.48 -108.44

Estimations of the marginal likelihoods by the symmetrised Chib’sapproximation (based on 105 Gibbs iterations and, for k > 5, 100permutations selected at random in Sk).

[Lee, Marin, Mengersen & Robert, 2008]

Page 69: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Case of the probit model

For the completion by z,

π(θ|x) =1T

∑t

π(θ|x, z(t))

is a simple average of normal densities

R Bridge sampling codegibbs1=gibbsprobit(Niter,y,X1)

gibbs2=gibbsprobit(Niter,y,X2)

bfchi=mean(exp(dmvlnorm(t(t(gibbs2$mu)-model2$coeff[,1]),mean=rep(0,3),

sigma=gibbs2$Sigma2)-probitlpost(model2$coeff[,1],y,X2)))/

mean(exp(dmvlnorm(t(t(gibbs1$mu)-model1$coeff[,1]),mean=rep(0,2),

sigma=gibbs1$Sigma2)-probitlpost(model1$coeff[,1],y,X1)))

Page 70: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Case of the probit model

For the completion by z,

π(θ|x) =1T

∑t

π(θ|x, z(t))

is a simple average of normal densities

R Bridge sampling codegibbs1=gibbsprobit(Niter,y,X1)

gibbs2=gibbsprobit(Niter,y,X2)

bfchi=mean(exp(dmvlnorm(t(t(gibbs2$mu)-model2$coeff[,1]),mean=rep(0,3),

sigma=gibbs2$Sigma2)-probitlpost(model2$coeff[,1],y,X2)))/

mean(exp(dmvlnorm(t(t(gibbs1$mu)-model1$coeff[,1]),mean=rep(0,2),

sigma=gibbs1$Sigma2)-probitlpost(model1$coeff[,1],y,X1)))

Page 71: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

Chib’s solution

Diabetes in Pima Indian women (cont’d)Comparison of the variation of the Bayes factor approximationsbased on 100 replicas for 20, 000 simulations for a simulation fromthe above Chib’s and importance samplers

●●

Chib's method importance sampling

0.024

00.0

245

0.025

00.0

255

Page 72: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

The Savage–Dickey ratio

Special representation of the Bayes factor used for simulation

Original version (Dickey, AoMS, 1971)

Page 73: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Savage’s density ratio theorem

Given a test H0 : θ = θ0 in a model f(x|θ, ψ) with a nuisanceparameter ψ, under priors π0(ψ) and π1(θ, ψ) such that

π1(ψ|θ0) = π0(ψ)

then

B01 =π1(θ0|x)π1(θ0)

,

with the obvious notations

π1(θ) =∫π1(θ, ψ)dψ , π1(θ|x) =

∫π1(θ, ψ|x)dψ ,

[Dickey, 1971; Verdinelli & Wasserman, 1995]

Page 74: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Rephrased

“Suppose that f0(θ) = f1(θ|φ = φ0). As f0(x|θ) = f1(x|θ, φ = φ0),

f0(x) =

Zf1(x|θ, φ = φ0)f1(θ|φ = φ0) dθ = f1(x|φ = φ0) ,

i.e., the denumerator of the Bayes factor is the value of f1(x|φ) at φ = φ0, while the denominator is an averageof the values of f1(x|φ) for φ 6= φ0, weighted by the prior distribution f1(φ) under the augmented model.Applying Bayes’ theorem to the right-hand side of [the above] we get

f0(x) = f1(φ0|x)f1(x)‹f1(φ0)

and hence the Bayes factor is given by

B = f0(x)‹f1(x) = f1(φ0|x)

‹f1(φ0) .

the ratio of the posterior to prior densities at φ = φ0 under the augmented model.”

[O’Hagan & Forster, 1996]

Page 75: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Rephrased

“Suppose that f0(θ) = f1(θ|φ = φ0). As f0(x|θ) = f1(x|θ, φ = φ0),

f0(x) =

Zf1(x|θ, φ = φ0)f1(θ|φ = φ0) dθ = f1(x|φ = φ0) ,

i.e., the denumerator of the Bayes factor is the value of f1(x|φ) at φ = φ0, while the denominator is an averageof the values of f1(x|φ) for φ 6= φ0, weighted by the prior distribution f1(φ) under the augmented model.Applying Bayes’ theorem to the right-hand side of [the above] we get

f0(x) = f1(φ0|x)f1(x)‹f1(φ0)

and hence the Bayes factor is given by

B = f0(x)‹f1(x) = f1(φ0|x)

‹f1(φ0) .

the ratio of the posterior to prior densities at φ = φ0 under the augmented model.”

[O’Hagan & Forster, 1996]

Page 76: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Measure-theoretic difficulty

Representation depends on the choice of versions of conditionaldensities:

B01 =∫π0(ψ)f(x|θ0, ψ) dψ∫

π1(θ, ψ)f(x|θ, ψ) dψdθ[by definition]

=∫π1(ψ|θ0)f(x|θ0, ψ) dψ π1(θ0)∫π1(θ, ψ)f(x|θ, ψ) dψdθ π1(θ0)

[specific version of π1(ψ|θ0)

and arbitrary version of π1(θ0)]

=∫π1(θ0, ψ)f(x|θ0, ψ) dψ

m1(x)π1(θ0)[specific version of π1(θ0, ψ)]

=π1(θ0|x)π1(θ0)

[version dependent]

Page 77: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Measure-theoretic difficulty

Representation depends on the choice of versions of conditionaldensities:

B01 =∫π0(ψ)f(x|θ0, ψ) dψ∫

π1(θ, ψ)f(x|θ, ψ) dψdθ[by definition]

=∫π1(ψ|θ0)f(x|θ0, ψ) dψ π1(θ0)∫π1(θ, ψ)f(x|θ, ψ) dψdθ π1(θ0)

[specific version of π1(ψ|θ0)

and arbitrary version of π1(θ0)]

=∫π1(θ0, ψ)f(x|θ0, ψ) dψ

m1(x)π1(θ0)[specific version of π1(θ0, ψ)]

=π1(θ0|x)π1(θ0)

[version dependent]

Page 78: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Choice of density version

c© Dickey’s (1971) condition is not a condition:

Ifπ1(θ0|x)π1(θ0)

=∫π0(ψ)f(x|θ0, ψ) dψ

m1(x)

is chosen as a version, then Savage–Dickey’s representation holds

Page 79: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Choice of density version

c© Dickey’s (1971) condition is not a condition:

Ifπ1(θ0|x)π1(θ0)

=∫π0(ψ)f(x|θ0, ψ) dψ

m1(x)

is chosen as a version, then Savage–Dickey’s representation holds

Page 80: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Savage–Dickey paradox

Verdinelli-Wasserman extension:

B01 =π1(θ0|x)π1(θ0)

Eπ1(ψ|x,θ0,x)

[π0(ψ)π1(ψ|θ0)

]similarly depends on choices of versions...

...but Monte Carlo implementation relies on specific versions of alldensities without making mention of it

[Chen, Shao & Ibrahim, 2000]

Page 81: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Savage–Dickey paradox

Verdinelli-Wasserman extension:

B01 =π1(θ0|x)π1(θ0)

Eπ1(ψ|x,θ0,x)

[π0(ψ)π1(ψ|θ0)

]similarly depends on choices of versions...

...but Monte Carlo implementation relies on specific versions of alldensities without making mention of it

[Chen, Shao & Ibrahim, 2000]

Page 82: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Computational implementation

Starting from the (new) prior

π1(θ, ψ) = π1(θ)π0(ψ)

define the associated posterior

π1(θ, ψ|x) = π0(ψ)π1(θ)f(x|θ, ψ)/m1(x)

and imposeπ1(θ0|x)π0(θ0)

=∫π0(ψ)f(x|θ0, ψ) dψ

m1(x)

to hold.Then

B01 =π1(θ0|x)π1(θ0)

m1(x)m1(x)

Page 83: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Computational implementation

Starting from the (new) prior

π1(θ, ψ) = π1(θ)π0(ψ)

define the associated posterior

π1(θ, ψ|x) = π0(ψ)π1(θ)f(x|θ, ψ)/m1(x)

and imposeπ1(θ0|x)π0(θ0)

=∫π0(ψ)f(x|θ0, ψ) dψ

m1(x)

to hold.Then

B01 =π1(θ0|x)π1(θ0)

m1(x)m1(x)

Page 84: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

First ratio

If (θ(1), ψ(1)), . . . , (θ(T ), ψ(T )) ∼ π(θ, ψ|x), then

1T

∑t

π1(θ0|x, ψ(t))

converges to π1(θ0|x) (if the right version is used in θ0).

π1(θ0|x, ψ) =π1(θ0)f(x|θ0, ψ)∫π1(θ)f(x|θ, ψ) dθ

Page 85: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Rao–Blackwellisation with latent variablesWhen π1(θ0|x, ψ) unavailable, replace with

1T

T∑t=1

π1(θ0|x, z(t), ψ(t))

via data completion by latent variable z such that

f(x|θ, ψ) =∫f(x, z|θ, ψ) dz

and that π1(θ, ψ, z|x) ∝ π0(ψ)π1(θ)f(x, z|θ, ψ) available in closedform, including the normalising constant, based on version

π1(θ0|x, z, ψ)π1(θ0)

=f(x, z|θ0, ψ)∫

f(x, z|θ, ψ)π1(θ) dθ.

Page 86: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Rao–Blackwellisation with latent variablesWhen π1(θ0|x, ψ) unavailable, replace with

1T

T∑t=1

π1(θ0|x, z(t), ψ(t))

via data completion by latent variable z such that

f(x|θ, ψ) =∫f(x, z|θ, ψ) dz

and that π1(θ, ψ, z|x) ∝ π0(ψ)π1(θ)f(x, z|θ, ψ) available in closedform, including the normalising constant, based on version

π1(θ0|x, z, ψ)π1(θ0)

=f(x, z|θ0, ψ)∫

f(x, z|θ, ψ)π1(θ) dθ.

Page 87: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Bridge revival (1)

Since m1(x)/m1(x) is unknown, apparent failure!bridge identity

Eπ1(θ,ψ|x)

[π1(θ, ψ)f(x|θ, ψ)

π0(ψ)π1(θ)f(x|θ, ψ)

]= Eπ1(θ,ψ|x)

[π1(ψ|θ)π0(ψ)

]=m1(x)m1(x)

to (biasedly) estimate m1(x)/m1(x) by

T/ T∑

t=1

π1(ψ(t)|θ(t))π0(ψ(t))

based on the same sample from π1.

Page 88: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Bridge revival (1)

Since m1(x)/m1(x) is unknown, apparent failure!bridge identity

Eπ1(θ,ψ|x)

[π1(θ, ψ)f(x|θ, ψ)

π0(ψ)π1(θ)f(x|θ, ψ)

]= Eπ1(θ,ψ|x)

[π1(ψ|θ)π0(ψ)

]=m1(x)m1(x)

to (biasedly) estimate m1(x)/m1(x) by

T/ T∑

t=1

π1(ψ(t)|θ(t))π0(ψ(t))

based on the same sample from π1.

Page 89: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Bridge revival (2)Alternative identity

Eπ1(θ,ψ|x)

[π0(ψ)π1(θ)f(x|θ, ψ)π1(θ, ψ)f(x|θ, ψ)

]= Eπ1(θ,ψ|x)

[π0(ψ)π1(ψ|θ)

]=m1(x)m1(x)

suggests using a second sample (θ(1), ψ(1), z(1)), . . . ,(θ(T ), ψ(T ), z(T )) ∼ π1(θ, ψ|x) and the ratio estimate

1T

T∑t=1

π0(ψ(t))/π1(ψ(t)|θ(t))

Resulting unbiased estimate:

B01 =1T

∑t π1(θ0|x, z(t), ψ(t))

π1(θ0)1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ(t))

Page 90: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Bridge revival (2)Alternative identity

Eπ1(θ,ψ|x)

[π0(ψ)π1(θ)f(x|θ, ψ)π1(θ, ψ)f(x|θ, ψ)

]= Eπ1(θ,ψ|x)

[π0(ψ)π1(ψ|θ)

]=m1(x)m1(x)

suggests using a second sample (θ(1), ψ(1), z(1)), . . . ,(θ(T ), ψ(T ), z(T )) ∼ π1(θ, ψ|x) and the ratio estimate

1T

T∑t=1

π0(ψ(t))/π1(ψ(t)|θ(t))

Resulting unbiased estimate:

B01 =1T

∑t π1(θ0|x, z(t), ψ(t))

π1(θ0)1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ(t))

Page 91: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Difference with Verdinelli–Wasserman representation

The above leads to the representation

B01 =π1(θ0|x)π1(θ0)

Eπ1(θ,ψ|x)

[π0(ψ)π1(ψ|θ)

]which shows how our approach differs from Verdinelli andWasserman’s

B01 =π1(θ0|x)π1(θ0)

Eπ1(ψ|x,θ0,x)

[π0(ψ)π1(ψ|θ0)

]

Page 92: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Difference with Verdinelli–Wasserman approximation

In terms of implementation,

B01MR

(x) =1T

∑t π1(θ0|x, z(t), ψ(t))

π1(θ0)1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ(t))

formaly resembles

B01VW

(x) =1T

T∑t=1

π1(θ0|x, z(t), ψ(t))π1(θ0)

1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ0)

.

But the simulated sequences differ: first average involvessimulations from π1(θ, ψ, z|x) and from π1(θ, ψ, z|x), while secondaverage relies on simulations from π1(θ, ψ, z|x) and fromπ1(ψ, z|x, θ0),

Page 93: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Difference with Verdinelli–Wasserman approximation

In terms of implementation,

B01MR

(x) =1T

∑t π1(θ0|x, z(t), ψ(t))

π1(θ0)1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ(t))

formaly resembles

B01VW

(x) =1T

T∑t=1

π1(θ0|x, z(t), ψ(t))π1(θ0)

1T

T∑t=1

π0(ψ(t))π1(ψ(t)|θ0)

.

But the simulated sequences differ: first average involvessimulations from π1(θ, ψ, z|x) and from π1(θ, ψ, z|x), while secondaverage relies on simulations from π1(θ, ψ, z|x) and fromπ1(ψ, z|x, θ0),

Page 94: Approximating Bayes Factors

On approximating Bayes factors

Importance sampling solutions compared

The Savage–Dickey ratio

Diabetes in Pima Indian women (cont’d)Comparison of the variation of the Bayes factor approximationsbased on 100 replicas for 20, 000 simulations for a simulation fromthe above importance, Chib’s, Savage–Dickey’s and bridgesamplers

IS Chib Savage−Dickey Bridge

2.83.0

3.23.4

Page 95: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Properties of nested sampling

1 Introduction

2 Importance sampling solutions compared

3 Nested samplingPurposeImplementationError ratesImpact of dimensionConstraintsImportance variantA mixture comparison

[Chopin & Robert, 2010]

Page 96: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Purpose

Nested sampling: Goal

Skilling’s (2007) technique using the one-dimensionalrepresentation:

Z = Eπ[L(θ)] =∫ 1

0ϕ(x) dx

withϕ−1(l) = P π(L(θ) > l).

Note; ϕ(·) is intractable in most cases.

Page 97: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Nested sampling: First approximation

Approximate Z by a Riemann sum:

Z =j∑i=1

(xi−1 − xi)ϕ(xi)

where the xi’s are either:

deterministic: xi = e−i/N

or random:

x0 = 1, xi+1 = tixi, ti ∼ Be(N, 1)

so that E[log xi] = −i/N .

Page 98: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Extraneous white noise

Take

Z =∫e−θ dθ =

∫1δe−(1−δ)θ e−δθ = Eδ

[1δe−(1−δ)θ

]Z =

1N

N∑i=1

δ−1 e−(1−δ)θi(xi−1 − xi) , θi ∼ E(δ) I(θi ≤ θi−1)

N deterministic random50 4.64 10.5

4.65 10.5100 2.47 4.9

2.48 5.02500 .549 1.01

.550 1.14

Comparison of variances and MSEs

Page 99: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Extraneous white noise

Take

Z =∫e−θ dθ =

∫1δe−(1−δ)θ e−δθ = Eδ

[1δe−(1−δ)θ

]Z =

1N

N∑i=1

δ−1 e−(1−δ)θi(xi−1 − xi) , θi ∼ E(δ) I(θi ≤ θi−1)

N deterministic random50 4.64 10.5

4.65 10.5100 2.47 4.9

2.48 5.02500 .549 1.01

.550 1.14

Comparison of variances and MSEs

Page 100: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Extraneous white noise

Take

Z =∫e−θ dθ =

∫1δe−(1−δ)θ e−δθ = Eδ

[1δe−(1−δ)θ

]Z =

1N

N∑i=1

δ−1 e−(1−δ)θi(xi−1 − xi) , θi ∼ E(δ) I(θi ≤ θi−1)

N deterministic random50 4.64 10.5

4.65 10.5100 2.47 4.9

2.48 5.02500 .549 1.01

.550 1.14

Comparison of variances and MSEs

Page 101: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Nested sampling: Second approximation

Replace (intractable) ϕ(xi) by ϕi, obtained by

Nested sampling

Start with N values θ1, . . . , θN sampled from πAt iteration i,

1 Take ϕi = L(θk), where θk is the point with smallestlikelihood in the pool of θi’s

2 Replace θk with a sample from the prior constrained toL(θ) > ϕi: the current N points are sampled from priorconstrained to L(θ) > ϕi.

Page 102: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Nested sampling: Second approximation

Replace (intractable) ϕ(xi) by ϕi, obtained by

Nested sampling

Start with N values θ1, . . . , θN sampled from πAt iteration i,

1 Take ϕi = L(θk), where θk is the point with smallestlikelihood in the pool of θi’s

2 Replace θk with a sample from the prior constrained toL(θ) > ϕi: the current N points are sampled from priorconstrained to L(θ) > ϕi.

Page 103: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Nested sampling: Second approximation

Replace (intractable) ϕ(xi) by ϕi, obtained by

Nested sampling

Start with N values θ1, . . . , θN sampled from πAt iteration i,

1 Take ϕi = L(θk), where θk is the point with smallestlikelihood in the pool of θi’s

2 Replace θk with a sample from the prior constrained toL(θ) > ϕi: the current N points are sampled from priorconstrained to L(θ) > ϕi.

Page 104: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Implementation

Nested sampling: Third approximation

Iterate the above steps until a given stopping iteration j isreached: e.g.,

observe very small changes in the approximation Z;

reach the maximal value of L(θ) when the likelihood isbounded and its maximum is known;

truncate the integral Z at level ε, i.e. replace∫ 1

0ϕ(x) dx with

∫ 1

εϕ(x) dx

Page 105: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Error rates

Approximation error

Error = Z− Z

=j∑i=1

(xi−1 − xi)ϕi −∫ 1

0ϕ(x) dx = −

∫ ε

0ϕ(x) dx (Truncation Error)

+

[j∑i=1

(xi−1 − xi)ϕ(xi)−∫ 1

εϕ(x) dx

](Quadrature Error)

+

[j∑i=1

(xi−1 − xi) {ϕi − ϕ(xi)}

](Stochastic Error)

[Dominated by Monte Carlo!]

Page 106: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Error rates

A CLT for the Stochastic Error

The (dominating) stochastic error is OP (N−1/2):

N1/2 {Stochastic Error} D→ N (0, V )

with

V = −∫s,t∈[ε,1]

sϕ′(s)tϕ′(t) log(s ∨ t) ds dt.

[Proof based on Donsker’s theorem]

The number of simulated points equals the number of iterations j,and is a multiple of N : if one stops at first iteration j such thate−j/N < ε, then: j = Nd− log εe.

Page 107: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Error rates

A CLT for the Stochastic Error

The (dominating) stochastic error is OP (N−1/2):

N1/2 {Stochastic Error} D→ N (0, V )

with

V = −∫s,t∈[ε,1]

sϕ′(s)tϕ′(t) log(s ∨ t) ds dt.

[Proof based on Donsker’s theorem]

The number of simulated points equals the number of iterations j,and is a multiple of N : if one stops at first iteration j such thate−j/N < ε, then: j = Nd− log εe.

Page 108: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Impact of dimension

Curse of dimension

For a simple Gaussian-Gaussian model of dimension dim(θ) = d,the following 3 quantities are O(d):

1 asymptotic variance of the NS estimator;

2 number of iterations (necessary to reach a given truncationerror);

3 cost of one simulated sample.

Therefore, CPU time necessary for achieving error level e is

O(d3/e2)

Page 109: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Impact of dimension

Curse of dimension

For a simple Gaussian-Gaussian model of dimension dim(θ) = d,the following 3 quantities are O(d):

1 asymptotic variance of the NS estimator;

2 number of iterations (necessary to reach a given truncationerror);

3 cost of one simulated sample.

Therefore, CPU time necessary for achieving error level e is

O(d3/e2)

Page 110: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Impact of dimension

Curse of dimension

For a simple Gaussian-Gaussian model of dimension dim(θ) = d,the following 3 quantities are O(d):

1 asymptotic variance of the NS estimator;

2 number of iterations (necessary to reach a given truncationerror);

3 cost of one simulated sample.

Therefore, CPU time necessary for achieving error level e is

O(d3/e2)

Page 111: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Impact of dimension

Curse of dimension

For a simple Gaussian-Gaussian model of dimension dim(θ) = d,the following 3 quantities are O(d):

1 asymptotic variance of the NS estimator;

2 number of iterations (necessary to reach a given truncationerror);

3 cost of one simulated sample.

Therefore, CPU time necessary for achieving error level e is

O(d3/e2)

Page 112: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Constraints

Sampling from constr’d priors

Exact simulation from the constrained prior is intractable in mostcases!

Skilling (2007) proposes to use MCMC, but:

this introduces a bias (stopping rule).

if MCMC stationary distribution is unconst’d prior, more andmore difficult to sample points such that L(θ) > l as lincreases.

If implementable, then slice sampler can be devised at the samecost!

[Thanks, Gareth!]

Page 113: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Constraints

Sampling from constr’d priors

Exact simulation from the constrained prior is intractable in mostcases!

Skilling (2007) proposes to use MCMC, but:

this introduces a bias (stopping rule).

if MCMC stationary distribution is unconst’d prior, more andmore difficult to sample points such that L(θ) > l as lincreases.

If implementable, then slice sampler can be devised at the samecost!

[Thanks, Gareth!]

Page 114: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Constraints

Sampling from constr’d priors

Exact simulation from the constrained prior is intractable in mostcases!

Skilling (2007) proposes to use MCMC, but:

this introduces a bias (stopping rule).

if MCMC stationary distribution is unconst’d prior, more andmore difficult to sample points such that L(θ) > l as lincreases.

If implementable, then slice sampler can be devised at the samecost!

[Thanks, Gareth!]

Page 115: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Importance variant

A IS variant of nested sampling

Consider instrumental prior π and likelihood L, weight function

w(θ) =π(θ)L(θ)

π(θ)L(θ)

and weighted NS estimator

Z =j∑i=1

(xi−1 − xi)ϕiw(θi).

Then choose (π, L) so that sampling from π constrained toL(θ) > l is easy; e.g. N (c, Id) constrained to ‖c− θ‖ < r.

Page 116: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

Importance variant

A IS variant of nested sampling

Consider instrumental prior π and likelihood L, weight function

w(θ) =π(θ)L(θ)

π(θ)L(θ)

and weighted NS estimator

Z =j∑i=1

(xi−1 − xi)ϕiw(θi).

Then choose (π, L) so that sampling from π constrained toL(θ) > l is easy; e.g. N (c, Id) constrained to ‖c− θ‖ < r.

Page 117: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Benchmark: Target distribution

Posterior distribution on (µ, σ) associated with the mixture

pN (0, 1) + (1− p)N (µ, σ) ,

when p is known

Page 118: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Experiment

n observations withµ = 2 and σ = 3/2,

Use of a uniform priorboth on (−2, 6) for µand on (.001, 16) forlog σ2.

occurrences of posteriorbursts for µ = xi

computation of thevarious estimates of Z

Page 119: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Experiment (cont’d)

MCMC sample for n = 16observations from the mixture.

Nested sampling sequencewith M = 1000 starting points.

Page 120: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Experiment (cont’d)

MCMC sample for n = 50observations from the mixture.

Nested sampling sequencewith M = 1000 starting points.

Page 121: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Comparison

Monte Carlo and MCMC (=Gibbs) outputs based on T = 104

simulations and numerical integration based on a 850× 950 grid inthe (µ, σ) parameter space.Nested sampling approximation based on a starting sample ofM = 1000 points followed by at least 103 further simulations fromthe constr’d prior and a stopping rule at 95% of the observedmaximum likelihood.Constr’d prior simulation based on 50 values simulated by randomwalk accepting only steps leading to a lik’hood higher than thebound

Page 122: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Comparison (cont’d)

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

●●

V1 V2 V3 V4

0.85

0.90

0.95

1.00

1.05

1.10

1.15

Graph based on a sample of 10 observations for µ = 2 andσ = 3/2 (150 replicas).

Page 123: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Comparison (cont’d)

●●

V1 V2 V3 V4

0.90

0.95

1.00

1.05

1.10

Graph based on a sample of 50 observations for µ = 2 andσ = 3/2 (150 replicas).

Page 124: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Comparison (cont’d)

●●

●●●

●●

V1 V2 V3 V4

0.85

0.90

0.95

1.00

1.05

1.10

1.15

Graph based on a sample of 100 observations for µ = 2 andσ = 3/2 (150 replicas).

Page 125: Approximating Bayes Factors

On approximating Bayes factors

Nested sampling

A mixture comparison

Comparison (cont’d)

Nested sampling gets less reliable as sample size increasesMost reliable approach is mixture Z3 although harmonic solutionZ1 close to Chib’s solution [taken as golden standard]Monte Carlo method Z2 also producing poor approximations to Z

(Kernel φ used in Z2 is a t non-parametric kernel estimate withstandard bandwidth estimation.)