SlideShare a Scribd company logo
Unbiased MCMC with couplings
Pierre E. Jacob
Department of Statistics, Harvard University
joint work with John O’Leary, Yves Atchad´e, and other fantastic
people acknowledged throughout
Department of Statistics, University of Oxford
July 19, 2019
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Setting
Continuous or discrete space of dimension d.
Target probability distribution π,
with probability density function x → π(x).
Goal: approximate π, e.g. approximate
Eπ[h(X)] = h(x)π(x)dx = π(h),
for a class of “test” functions h.
Pierre E. Jacob Unbiased MCMC
Markov chain Monte Carlo
Initially, X0 ∼ π0, then Xt|Xt−1 ∼ P(Xt−1, ·) for t = 1, . . . , T.
Estimator:
1
T − b
T
t=b+1
h(Xt),
where b iterations are discarded as burn-in.
Might converge to Eπ[h(X)] as T → ∞ by the ergodic theorem.
Biased for any fixed b, T, since π0 = π.
Averaging independent copies of such estimators for fixed T, b
would not provide a consistent estimator of Eπ[h(X)]
as the number of independent copies goes to infinity.
Pierre E. Jacob Unbiased MCMC
Example: Metropolis–Hastings kernel P
Initialize: X0 ∼ π0.
At each iteration t ≥ 0, with Markov chain at state Xt,
1 propose X ∼ k(Xt, ·),
2 sample U ∼ U([0, 1]),
3 if
U ≤
π(X )k(X , Xt)
π(Xt)k(Xt, X )
,
set Xt+1 = X , otherwise set Xt+1 = Xt.
Hastings, Monte Carlo sampling methods using Markov chains and
their applications, Biometrika, 1970.
Pierre E. Jacob Unbiased MCMC
MCMC path
π = N(0, 1), MH with Normal proposal std = 0.5, π0 = N(10, 32
)
Pierre E. Jacob Unbiased MCMC
MCMC marginal distributions
π = N(0, 1), RWMH with Normal proposal std = 0.5, π0 = N(10, 32
)
Pierre E. Jacob Unbiased MCMC
Parallel computing
R processors generate unbiased estimators, independently.
The number of estimators completed by time t is random.
There are different ways of aggregating estimators,
either discarding on-going calculations at time t or not.
Different regimes can be considered, e.g.
as R → ∞ for fixed t,
as t → ∞ for fixed R,
as t and R both go to infinity.
Glynn & Heidelberger, Analysis of parallel replicated
simulations under a completion time constraint, 1991.
Pierre E. Jacob Unbiased MCMC
Parallel computing
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Coupled chains
Generate two Markov chains (Xt) and (Yt) as follows,
sample X0 and Y0 from π0 (independently or not),
sample X1|X0 ∼ P(X0, ·),
for t ≥ 1, sample (Xt+1, Yt)|(Xt, Yt−1) ∼ ¯P ((Xt, Yt−1), ·),
¯P is such that Xt+1|Xt ∼ P(Xt, ·) and Yt|Yt−1 ∼ P(Yt−1, ·),
so each chain marginally evolves according to P.
⇒ Xt and Yt have the same distribution for all t ≥ 0.
¯P is also such that ∃τ such that Xτ = Yτ−1,
and the chains are faithful.
Pierre E. Jacob Unbiased MCMC
Debiasing idea (one slide version)
Limit as a telescopic sum, for all k ≥ 0,
Eπ[h(X)] = lim
t→∞
E[h(Xt)] = E[h(Xk)] +
∞
t=k+1
E[h(Xt) − h(Xt−1)].
Since for all t ≥ 0, Xt and Yt have the same distribution,
= E[h(Xk)] +
∞
t=k+1
E[h(Xt) − h(Yt−1)].
If we can swap expectation and limit,
= E[h(Xk) +
∞
t=k+1
(h(Xt) − h(Yt−1))],
so h(Xk) + ∞
t=k+1(h(Xt) − h(Yt−1)) is unbiased.
Pierre E. Jacob Unbiased MCMC
Unbiased estimators
Unbiased estimator is given by
Hk(X, Y ) = h(Xk) +
τ−1
t=k+1
(h(Xt) − h(Yt−1)),
with the convention τ−1
t=k+1{·} = 0 if τ − 1 < k + 1.
h(Xk) alone is biased; the other terms correct for the bias.
Cost: τ − 1 calls to ¯P and 1 + max(0, k − τ) calls to P.
Glynn & Rhee, Exact estimation for Markov chain equilibrium
expectations, 2014.
Note: same reasoning would work with arbitrary lags L ≥ 1.
Pierre E. Jacob Unbiased MCMC
Conditions
Jacob, O’Leary, Atchad´e, Unbiased MCMC with couplings.
1 Marginal chain converges:
E[h(Xt)] → Eπ[h(X)],
and h(Xt) has (2 + η)-finite moments for all t.
2 Meeting time τ has geometric tails:
∃C < +∞ ∃δ ∈ (0, 1) ∀t ≥ 0 P(τ > t) ≤ Cδt
.
3 Chains stay together: Xt = Yt−1 for all t ≥ τ.
Condition 2 itself implied by e.g. geometric drift condition.
Under these conditions, Hk(X, Y ) is unbiased, has finite
expected cost and finite variance, for all k.
Pierre E. Jacob Unbiased MCMC
Conditions: update
Middleton, Deligiannidis, Doucet, Jacob, Unbiased MCMC for
intractable target distributions.
1 Marginal chain converges:
E[h(Xt)] → Eπ[h(X)],
and h(Xt) has (2 + η)-finite moments for all t.
2 Meeting time τ has polynomial tails:
∃C < +∞ ∃δ > 2(2η−1
+ 1) ∀t ≥ 0 P(τ > t) ≤ Ct−δ
.
3 Chains stay together: Xt = Yt−1 for all t ≥ τ.
Condition 2 itself implied by e.g. polynomial drift condition.
Pierre E. Jacob Unbiased MCMC
Improved unbiased estimators
Efficiency matters, thus in practice we recommend a variation
of the previous estimator, defined for integers k ≤ m as
Hk:m(X, Y ) =
1
m − k + 1
m
t=k
Ht(X, Y )
which can also be written
1
m − k + 1
m
t=k
h(Xt)+
τ−1
t=k+1
min 1,
t − k
m − k + 1
(h(Xt)−h(Yt−1)),
i.e. standard MCMC average + bias correction term.
As k → ∞, bias correction is zero with increasing probability.
Note: changing the lag is another way of modifying efficiency
while preserving the lack of bias.
Pierre E. Jacob Unbiased MCMC
Efficiency
Writing Hk:m(X, Y ) = MCMCk:m + BCk:m,
then, denoting the MSE of MCMCk:m by MSEk:m,
V[Hk:m(X, Y )] ≤ MSEk:m+2 MSEk:m E BC2
k:m +E BC2
k:m .
Under geometric drift condition and regularity assumptions on
h, for some δ < 1, C < ∞,
E[BC2
k:m] ≤
Cδk
(m − k + 1)2
,
and δ is directly related to tails of the meeting time.
Similarly under polynomial drift, we obtain a term of the form
k−δ in the numerator instead of δk.
Pierre E. Jacob Unbiased MCMC
Signed measure estimator
Replacing function evaluations by delta masses leads to
ˆπ(·) =
1
m − k + 1
m
t=k
δXt (·)
+
τ−1
t=k+1
min 1,
t − k
m − k + 1
(δXt (·) − δYt−1 (·)),
which is of the form
ˆπ(·) =
N
n=1
ωnδZn (·),
where N
n=1 ωn = 1 but some ωn might be negative.
Unbiasedness reads: E[Eˆπ[h(X)]] = Eπ[h(X)].
Pierre E. Jacob Unbiased MCMC
Assessing convergence of MCMC
Total variation distance between Xk ∼ πk and π = limk→∞ πk:
πk − π TV =
1
2
sup
h:|h|≤1
|E[h(Xk)] − Eπ[h(X)]|
=
1
2
sup
h:|h|≤1
|E[
τ−1
t=k+1
h(Xt) − h(Yt−1)]|
≤ E[max(0, τ − k − 1)].
0.00
0.01
0.02
0.03
0.04
0 50 100 150 200
meeting time
density
1e−03
1e−02
1e−01
1e+00
1e+01
0 50 100
k
upperbound
Pierre E. Jacob Unbiased MCMC
Assessing convergence of MCMC
With L-lag couplings, τ(L) = inf{t ≥ L : Xt = Yt−L},
dTV (πk, π) ≤ E 0 ∨ (τ(L)
− L − k)/L ,
dW (πk, π) ≤ E
(τ(L)−L−k)/L
j=1
Xk+jL − Yk+(j−1)L 1 .
0.00
0.25
0.50
0.75
1.00
1e+01 1e+02 1e+03 1e+04 1e+05 1e+06
iterations
dTV
SSG PT
Biswas & Jacob, Estimating Convergence of Markov chains with
L-Lag Couplings, 2019.
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Designing coupled chains
To implement the proposed unbiased estimators,
we need to sample from a Markov kernel ¯P,
such that, when (Xt+1, Yt) is sampled from ¯P ((Xt, Yt−1), ·),
marginally Xt+1|Xt ∼ P(Xt, ·), and Yt|Yt−1 ∼ P(Yt−1, ·),
it is possible that Xt+1 = Yt exactly for some t ≥ 0,
if Xt = Yt−1, then Xt+1 = Yt almost surely.
Pierre E. Jacob Unbiased MCMC
Couplings of MCMC algorithms
Many practical couplings in the literature. . .
Propp & Wilson, Exact sampling with coupled Markov chains
and applications to statistical mechanics, Random Structures &
Algorithms, 1996.
Johnson, Studying convergence of Markov chain Monte Carlo
algorithms using coupled sample paths, JASA, 1996.
Neal, Circularly-coupled Markov chain sampling, UoT tech
report, 1999.
Pinto & Neal, Improving Markov chain Monte Carlo estimators
by coupling to an approximating chain, UoT tech report, 2001.
Glynn & Rhee, Exact estimation for Markov chain equilibrium
expectations, Journal of Applied Probability, 2014.
Pierre E. Jacob Unbiased MCMC
Couplings
(X, Y ) follows a coupling of p and q if X ∼ p and Y ∼ q.
The coupling inequality states that
P(X = Y ) ≤ 1 − p − q TV,
for any coupling, with p − q TV = 1
2 |p(x) − q(x)|dx.
Maximal couplings achieve the bound.
Pierre E. Jacob Unbiased MCMC
Maximal coupling of Gamma and Normal
Pierre E. Jacob Unbiased MCMC
Maximal coupling: sampling algorithm
Requires: evaluations of p and q, sampling from p and q.
1 Sample X ∼ p and W ∼ U([0, 1]).
If W ≤ q(X)/p(X), output (X, X).
2 Otherwise, sample Y ∼ q and W ∼ U([0, 1])
until W > p(Y )/q(Y ), and output (X, Y ).
Output: a pair (X, Y ) such that X ∼ p, Y ∼ q
and P(X = Y ) is maximal.
Pierre E. Jacob Unbiased MCMC
Back to Metropolis–Hastings (kernel P)
At each iteration t, Markov chain at state Xt,
1 propose X ∼ k(Xt, ·),
2 sample U ∼ U([0, 1]),
3 if
U ≤
π(X )k(X , Xt)
π(Xt)k(Xt, X )
,
set Xt+1 = X , otherwise set Xt+1 = Xt.
How to propagate two MH chains from states Xt and Yt−1
such that {Xt+1 = Yt} can happen?
Pierre E. Jacob Unbiased MCMC
Coupling of Metropolis–Hastings (kernel ¯P)
At each iteration t, two Markov chains at states Xt, Yt−1,
1 propose (X , Y ) from max coupling of k(Xt, ·), k(Yt−1, ·),
2 sample U ∼ U([0, 1]),
3 if
U ≤
π(X )k(X , Xt)
π(Xt)k(Xt, X )
,
set Xt+1 = X , otherwise set Xt+1 = Xt,
if
U ≤
π(Y )k(Y , Yt−1)
π(Yt−1)k(Yt−1, Y )
,
set Yt = Y , otherwise set Yt = Yt−1.
Pierre E. Jacob Unbiased MCMC
Coupling of Gibbs
Gibbs sampler: update component i of the chain,
leaving π(dxi|x1, . . . , xi−1, xi+1, . . . , xd) invariant.
For instance, we can propose X ∼ k(Xi
t, ·) to replace Xi
t,
and accept or not with a Metropolis–Hastings step.
These proposals can be maximally coupled across two chains,
at each component update.
The chains meet when all components have met.
Likewise we can couple parallel tempering chains,
and meeting occurs when entire ensembles of chains meet.
Pierre E. Jacob Unbiased MCMC
Hamiltonian Monte Carlo
Introduce potential energy U(q) = − log π(q),
and total energy E(q, p) = U(q) + 1
2|p|2.
Hamiltonian dynamics for (q(s), p(s)), where s ≥ 0:
d
ds
q(s) = pE(q(s), p(s)),
d
ds
p(s) = − qE(q(s), p(s)).
Solving Hamiltonian dynamics exactly is not feasible,
but discretization + Metropolis–Hastings correction ensure that
π remains invariant.
Common random numbers can make two HMC chains contract,
under assumptions on the target such as strong log-concavity.
Heng & Jacob, Unbiased HMC with couplings, Biometrika, 2019.
Pierre E. Jacob Unbiased MCMC
Coupling of HMC
Figure 2 of Mangoubi & Smith, Rapid mixing of HMC strongly
log-concave distributions, 2017.
Coupling two copies X1, X2, . . . (blue) and Y1, Y2, . . .
(green) of HMC by choosing same momentum pi at
every step.
See also Bou-Rabee, Eberle & Zimmer, Coupling and Convergence
for Hamiltonian Monte Carlo, 2018.
Pierre E. Jacob Unbiased MCMC
Particle Independent Metropolis–Hastings
Initialization: run SMC, obtain ˆπ(0)(·), ˆZ(0),
where E[ ˆZ(0)] = Z, the normalizing constant of π.
At each iteration t, given (ˆπ(t)(·), ˆZ(t)),
1 run SMC, obtain (π (·), Z ),
2 sample U ∼ U([0, 1]),
3 if
U ≤
Z
ˆZ(t)
,
set (t + 1)-th state to (π (·), Z ),
otherwise set (t + 1)-th state to (ˆπ(t)(·), ˆZ(t)).
For any number of particles, T−1 T
t=1 ˆπ(t)(·) →T→∞ π.
Andrieu, Doucet, Holenstein, Particle MCMC, JRSS B, 2010.
Pierre E. Jacob Unbiased MCMC
Coupled Particle Independent Metropolis–Hastings
At each iteration t, given (ˆπ(t)(·), ˆZ(t)), and (˜π(t−1)(·), ˜Z(t−1)),
1 run SMC, obtain (π (·), Z ),
2 sample U ∼ U([0, 1]),
3 if
U ≤
Z
ˆZ(t)
,
set (t + 1)-th “1st” state to (π (·), Z ),
otherwise set (t + 1)-th “1st” state to (ˆπ(t)(·), ˆZ(t)).
4 if
U ≤
Z
˜Z(t−1)
,
set t-th “2nd” state to (π (·), Z ),
otherwise set t-th “2nd” state to (˜π(t−1)(·), ˜Z(t−1)).
Pierre E. Jacob Unbiased MCMC
Unbiased SMC
PIMH turns any SMC sampler into an MCMC scheme
that can be debiased using the coupled chains machinery.
Thus any MCMC algorithm that can be plugged in an
SMC sampler, can then be subsequently de-biased.
No need for couplings of the MCMC chains themselves.
Middleton, Deligiannidis, Doucet, Jacob, Unbiased Smoothing using
Particle Independent Metropolis-Hastings, AISTATS, 2019.
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Bimodal target
Target is mixture of univariate Normal distributions:
π = 0.5 · N(−4, 1) + 0.5 · N(+4, 1).
MCMC: random walk Metropolis–Hastings,
with proposal standard deviation σ that will vary.
Initial distribution π0 will vary.
Pierre E. Jacob Unbiased MCMC
Bimodal target
With σ = 3, π0 = N(10, 102). . .
0.00
0.05
0.10
0.15
0.20
−10 −5 0 5 10
x
density
Pierre E. Jacob Unbiased MCMC
Bimodal target
With σ = 1, π0 = N(10, 102). . .
0.00
0.05
0.10
0.15
0.20
−10 −5 0 5 10
x
density
Pierre E. Jacob Unbiased MCMC
Bimodal target
With σ = 1, π0 = N(10, 12). . .
0.0
0.1
0.2
0.3
−10 −5 0 5 10
x
density
Pierre E. Jacob Unbiased MCMC
Outline
1 It is about Monte Carlo and bias
2 Unbiased estimators from Markov kernels
3 Design of coupled Markov chains
4 It can fail
5 Beyond parallel computation and burn-in
Pierre E. Jacob Unbiased MCMC
Unbiased property
Unbiased MCMC estimators can be used to tackle problems for
which standard MCMC methods are ill-suited.
For instance, problems where one needs to approximate
h(x1, x2)π2(x2|x1)dx2 π1(x1)dx1,
or equivalently
E1 [E2[h(X1, X2)|X1]] .
Pierre E. Jacob Unbiased MCMC
Example 1: cut distribution
First model, parameter θ1, data Y1, posterior π1(θ1|Y1).
Second model, parameter θ2, data Y2, conditional posterior:
π2(θ2|Y2, θ1) =
p2(θ2|θ1)p2(Y2|θ1, θ2)
p2(Y2|θ1)
.
Plummer, Cuts in Bayesian graphical models, 2015.
Jacob, Murray, Holmes, Robert, Better together? Statistical learning
in models made of modules.
Pierre E. Jacob Unbiased MCMC
Example: epidemiological study
Model of virus prevalence
∀i = 1, . . . , I Zi ∼ Binomial(Ni, ϕi),
Zi is number of women infected with high-risk HPV in a
sample of size Ni in country i.
Impact of prevalence onto cervical cancer occurrence
∀i = 1, . . . , I Yi ∼ Poisson(λiTi), log(λi) = θ2,1 + θ2,2ϕi,
Yi is number of cancer cases arising from Ti woman-years of
follow-up in country i.
Plummer, Cuts in Bayesian graphical models, 2015.
Pierre E. Jacob Unbiased MCMC
Example: epidemiological study
Approximations of the cut distribution marginals:
0
1
2
3
−2.5 −2.0 −1.5
θ2,1
density
0.00
0.05
0.10
0.15
10 15 20 25
θ2,2
density
Red curves were obtained by long MCMC run targeting
π2(θ2|Y2, θ1), for draws θ1 from first posterior π1(θ1|Y1).
Black bars were obtained with unbiased MCMC.
Pierre E. Jacob Unbiased MCMC
Example 2: normalizing constants
We might be interested in normalizing constant Z of
π(x) ∝ exp(−U(x)), defined as Z = exp(−U(x))dx.
Introduce
∀λ ∈ [0, 1] πλ(x) ∝ exp(−Uλ(x)).
Thermodynamics integration or path sampling identity:
log
Z1
Z0
= −
1
0
λUλ(x)πλ(dx)
q(λ)
q(dλ),
where q(dλ) is any distribution supported on [0, 1].
Rischard, Jacob, Pillai, Unbiased estimation of log normalizing
constants with applications to Bayesian cross-validation.
Pierre E. Jacob Unbiased MCMC
Example 3: Bayesian cross-validation
Data y = {y1, . . . , yn}, made of n units.
Partition y into training T and validation V sets.
For each split y = {T, V }, assess predictive performance with
log p(V |T) = log p(V |θ, T)p(dθ|T).
Note, log p(V |T) is a log-ratio of normalizing constants.
Finally we might want to approximate
CV =
n
nT
−1
T,V
log p(V |T).
Rischard, Jacob, Pillai, Unbiased estimation of log normalizing
constants with applications to Bayesian cross-validation.
Pierre E. Jacob Unbiased MCMC
Discussion
Perfect samplers, designed to sample i.i.d. from π, would
yield the same benefits and more.
What about regeneration?
Implications of lack of bias are numerous.
Proposed estimators can be obtained
using coupled MCMC kernels,
or using the “MCMC → SMC → coupled PIMH” route.
Cannot work if underlying MCMC doesn’t work. . .
but parallelism allows for more expensive MCMC kernels.
Thank you for listening!
Pierre E. Jacob Unbiased MCMC
References
with John O’Leary, Yves F. Atchad´e
Unbiased Markov chain Monte Carlo with couplings, 2019.
with Fredrik Lindsten, Thomas Sch¨on
Smoothing with Couplings of Conditional Particle Filters, 2018.
with Jeremy Heng
Unbiased Hamiltonian Monte Carlo with couplings, 2019.
with Lawrence Middleton, George Deligiannidis, Arnaud
Doucet
Unbiased Markov chain Monte Carlo for intractable target
distributions, 2019.
Unbiased Smoothing using Particle Independent
Metropolis-Hastings, 2019.
Pierre E. Jacob Unbiased MCMC
References
with Maxime Rischard, Natesh Pillai
Unbiased estimation of log normalizing constants with
applications to Bayesian cross-validation, 2019.
with Niloy Biswas
Estimating Convergence of Markov chains with L-Lag Couplings,
2019.
Funding provided by the National Science Foundation, grants
DMS-1712872 and DMS-1844695.
Pierre E. Jacob Unbiased MCMC

More Related Content

What's hot

Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Christian Robert
 
Richard Everitt's slides
Richard Everitt's slidesRichard Everitt's slides
Richard Everitt's slides
Christian Robert
 
ABC in Venezia
ABC in VeneziaABC in Venezia
ABC in Venezia
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Complexity of exact solutions of many body systems: nonequilibrium steady sta...
Complexity of exact solutions of many body systems: nonequilibrium steady sta...Complexity of exact solutions of many body systems: nonequilibrium steady sta...
Complexity of exact solutions of many body systems: nonequilibrium steady sta...
Lake Como School of Advanced Studies
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Bayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal modelsBayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal modelskhbrodersen
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Introduction to MCMC methods
Introduction to MCMC methodsIntroduction to MCMC methods
Introduction to MCMC methods
Christian Robert
 
Introduction to Diffusion Monte Carlo
Introduction to Diffusion Monte CarloIntroduction to Diffusion Monte Carlo
Introduction to Diffusion Monte Carlo
Claudio Attaccalite
 
MCMC and likelihood-free methods
MCMC and likelihood-free methodsMCMC and likelihood-free methods
MCMC and likelihood-free methods
Christian Robert
 
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithmno U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Approximate Bayesian Computation with Quasi-Likelihoods
Approximate Bayesian Computation with Quasi-LikelihoodsApproximate Bayesian Computation with Quasi-Likelihoods
Approximate Bayesian Computation with Quasi-Likelihoods
Stefano Cabras
 
Omiros' talk on the Bernoulli factory problem
Omiros' talk on the  Bernoulli factory problemOmiros' talk on the  Bernoulli factory problem
Omiros' talk on the Bernoulli factory problem
BigMC
 
Bayesian regression models and treed Gaussian process models
Bayesian regression models and treed Gaussian process modelsBayesian regression models and treed Gaussian process models
Bayesian regression models and treed Gaussian process models
Tommaso Rigon
 

What's hot (20)

Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithmsRao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
Rao-Blackwellisation schemes for accelerating Metropolis-Hastings algorithms
 
Richard Everitt's slides
Richard Everitt's slidesRichard Everitt's slides
Richard Everitt's slides
 
ABC in Venezia
ABC in VeneziaABC in Venezia
ABC in Venezia
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Complexity of exact solutions of many body systems: nonequilibrium steady sta...
Complexity of exact solutions of many body systems: nonequilibrium steady sta...Complexity of exact solutions of many body systems: nonequilibrium steady sta...
Complexity of exact solutions of many body systems: nonequilibrium steady sta...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Bayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal modelsBayesian inversion of deterministic dynamic causal models
Bayesian inversion of deterministic dynamic causal models
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Introduction to MCMC methods
Introduction to MCMC methodsIntroduction to MCMC methods
Introduction to MCMC methods
 
Introduction to Diffusion Monte Carlo
Introduction to Diffusion Monte CarloIntroduction to Diffusion Monte Carlo
Introduction to Diffusion Monte Carlo
 
MCMC and likelihood-free methods
MCMC and likelihood-free methodsMCMC and likelihood-free methods
MCMC and likelihood-free methods
 
Lesage
LesageLesage
Lesage
 
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithmno U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
no U-turn sampler, a discussion of Hoffman & Gelman NUTS algorithm
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Approximate Bayesian Computation with Quasi-Likelihoods
Approximate Bayesian Computation with Quasi-LikelihoodsApproximate Bayesian Computation with Quasi-Likelihoods
Approximate Bayesian Computation with Quasi-Likelihoods
 
Omiros' talk on the Bernoulli factory problem
Omiros' talk on the  Bernoulli factory problemOmiros' talk on the  Bernoulli factory problem
Omiros' talk on the Bernoulli factory problem
 
Bayesian regression models and treed Gaussian process models
Bayesian regression models and treed Gaussian process modelsBayesian regression models and treed Gaussian process models
Bayesian regression models and treed Gaussian process models
 

Similar to Unbiased MCMC with couplings

Unbiased Markov chain Monte Carlo
Unbiased Markov chain Monte CarloUnbiased Markov chain Monte Carlo
Unbiased Markov chain Monte Carlo
JeremyHeng10
 
Unbiased Markov chain Monte Carlo
Unbiased Markov chain Monte CarloUnbiased Markov chain Monte Carlo
Unbiased Markov chain Monte Carlo
JeremyHeng10
 
Adaptive Restore algorithm & importance Monte Carlo
Adaptive Restore algorithm & importance Monte CarloAdaptive Restore algorithm & importance Monte Carlo
Adaptive Restore algorithm & importance Monte Carlo
Christian Robert
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
Pierre Jacob
 
Talk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniquesTalk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniques
Pierre Jacob
 
The Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal FunctionThe Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal Function
Radboud University Medical Center
 
Monte Carlo methods for some not-quite-but-almost Bayesian problems
Monte Carlo methods for some not-quite-but-almost Bayesian problemsMonte Carlo methods for some not-quite-but-almost Bayesian problems
Monte Carlo methods for some not-quite-but-almost Bayesian problems
Pierre Jacob
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
Christian Robert
 
Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3
Fabian Pedregosa
 
Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte CarloUnbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo
JeremyHeng10
 
Nonlinear Stochastic Optimization by the Monte-Carlo Method
Nonlinear Stochastic Optimization by the Monte-Carlo MethodNonlinear Stochastic Optimization by the Monte-Carlo Method
Nonlinear Stochastic Optimization by the Monte-Carlo Method
SSA KPI
 
Probabilistic Control of Switched Linear Systems with Chance Constraints
Probabilistic Control of Switched Linear Systems with Chance ConstraintsProbabilistic Control of Switched Linear Systems with Chance Constraints
Probabilistic Control of Switched Linear Systems with Chance Constraints
Leo Asselborn
 
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic ReachabilityProbabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Leo Asselborn
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
Alexander Litvinenko
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo
JeremyHeng10
 
Metodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang LandauMetodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang Landau
angely alcendra
 
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Leo Asselborn
 
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
The Statistical and Applied Mathematical Sciences Institute
 
Litvinenko low-rank kriging +FFT poster
Litvinenko low-rank kriging +FFT  posterLitvinenko low-rank kriging +FFT  poster
Litvinenko low-rank kriging +FFT poster
Alexander Litvinenko
 

Similar to Unbiased MCMC with couplings (20)

Unbiased Markov chain Monte Carlo
Unbiased Markov chain Monte CarloUnbiased Markov chain Monte Carlo
Unbiased Markov chain Monte Carlo
 
Unbiased Markov chain Monte Carlo
Unbiased Markov chain Monte CarloUnbiased Markov chain Monte Carlo
Unbiased Markov chain Monte Carlo
 
Adaptive Restore algorithm & importance Monte Carlo
Adaptive Restore algorithm & importance Monte CarloAdaptive Restore algorithm & importance Monte Carlo
Adaptive Restore algorithm & importance Monte Carlo
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Talk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniquesTalk at CIRM on Poisson equation and debiasing techniques
Talk at CIRM on Poisson equation and debiasing techniques
 
The Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal FunctionThe Gaussian Hardy-Littlewood Maximal Function
The Gaussian Hardy-Littlewood Maximal Function
 
Monte Carlo methods for some not-quite-but-almost Bayesian problems
Monte Carlo methods for some not-quite-but-almost Bayesian problemsMonte Carlo methods for some not-quite-but-almost Bayesian problems
Monte Carlo methods for some not-quite-but-almost Bayesian problems
 
Delayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithmsDelayed acceptance for Metropolis-Hastings algorithms
Delayed acceptance for Metropolis-Hastings algorithms
 
Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3Random Matrix Theory and Machine Learning - Part 3
Random Matrix Theory and Machine Learning - Part 3
 
Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte CarloUnbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo
 
Nonlinear Stochastic Optimization by the Monte-Carlo Method
Nonlinear Stochastic Optimization by the Monte-Carlo MethodNonlinear Stochastic Optimization by the Monte-Carlo Method
Nonlinear Stochastic Optimization by the Monte-Carlo Method
 
Probabilistic Control of Switched Linear Systems with Chance Constraints
Probabilistic Control of Switched Linear Systems with Chance ConstraintsProbabilistic Control of Switched Linear Systems with Chance Constraints
Probabilistic Control of Switched Linear Systems with Chance Constraints
 
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic ReachabilityProbabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
Probabilistic Control of Uncertain Linear Systems Using Stochastic Reachability
 
Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...Hierarchical matrices for approximating large covariance matries and computin...
Hierarchical matrices for approximating large covariance matries and computin...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo Unbiased Hamiltonian Monte Carlo
Unbiased Hamiltonian Monte Carlo
 
Metodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang LandauMetodo Monte Carlo -Wang Landau
Metodo Monte Carlo -Wang Landau
 
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
Robust Control of Uncertain Switched Linear Systems based on Stochastic Reach...
 
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
2018 MUMS Fall Course - Statistical Representation of Model Input (EDITED) - ...
 
Litvinenko low-rank kriging +FFT poster
Litvinenko low-rank kriging +FFT  posterLitvinenko low-rank kriging +FFT  poster
Litvinenko low-rank kriging +FFT poster
 

More from Pierre Jacob

ISBA 2022 Susie Bayarri lecture
ISBA 2022 Susie Bayarri lectureISBA 2022 Susie Bayarri lecture
ISBA 2022 Susie Bayarri lecture
Pierre Jacob
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
Pierre Jacob
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
Pierre Jacob
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
Pierre Jacob
 
Current limitations of sequential inference in general hidden Markov models
Current limitations of sequential inference in general hidden Markov modelsCurrent limitations of sequential inference in general hidden Markov models
Current limitations of sequential inference in general hidden Markov models
Pierre Jacob
 
On non-negative unbiased estimators
On non-negative unbiased estimatorsOn non-negative unbiased estimators
On non-negative unbiased estimators
Pierre Jacob
 
Path storage in the particle filter
Path storage in the particle filterPath storage in the particle filter
Path storage in the particle filter
Pierre Jacob
 
SMC^2: an algorithm for sequential analysis of state-space models
SMC^2: an algorithm for sequential analysis of state-space modelsSMC^2: an algorithm for sequential analysis of state-space models
SMC^2: an algorithm for sequential analysis of state-space models
Pierre Jacob
 
PAWL - GPU meeting @ Warwick
PAWL - GPU meeting @ WarwickPAWL - GPU meeting @ Warwick
PAWL - GPU meeting @ Warwick
Pierre Jacob
 
Presentation of SMC^2 at BISP7
Presentation of SMC^2 at BISP7Presentation of SMC^2 at BISP7
Presentation of SMC^2 at BISP7
Pierre Jacob
 
Presentation MCB seminar 09032011
Presentation MCB seminar 09032011Presentation MCB seminar 09032011
Presentation MCB seminar 09032011
Pierre Jacob
 

More from Pierre Jacob (11)

ISBA 2022 Susie Bayarri lecture
ISBA 2022 Susie Bayarri lectureISBA 2022 Susie Bayarri lecture
ISBA 2022 Susie Bayarri lecture
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Current limitations of sequential inference in general hidden Markov models
Current limitations of sequential inference in general hidden Markov modelsCurrent limitations of sequential inference in general hidden Markov models
Current limitations of sequential inference in general hidden Markov models
 
On non-negative unbiased estimators
On non-negative unbiased estimatorsOn non-negative unbiased estimators
On non-negative unbiased estimators
 
Path storage in the particle filter
Path storage in the particle filterPath storage in the particle filter
Path storage in the particle filter
 
SMC^2: an algorithm for sequential analysis of state-space models
SMC^2: an algorithm for sequential analysis of state-space modelsSMC^2: an algorithm for sequential analysis of state-space models
SMC^2: an algorithm for sequential analysis of state-space models
 
PAWL - GPU meeting @ Warwick
PAWL - GPU meeting @ WarwickPAWL - GPU meeting @ Warwick
PAWL - GPU meeting @ Warwick
 
Presentation of SMC^2 at BISP7
Presentation of SMC^2 at BISP7Presentation of SMC^2 at BISP7
Presentation of SMC^2 at BISP7
 
Presentation MCB seminar 09032011
Presentation MCB seminar 09032011Presentation MCB seminar 09032011
Presentation MCB seminar 09032011
 

Recently uploaded

EY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptxEY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptx
AlguinaldoKong
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
RenuJangid3
 
Comparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebratesComparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebrates
sachin783648
 
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
NathanBaughman3
 
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdfSCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SELF-EXPLANATORY
 
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Ana Luísa Pinho
 
Richard's entangled aventures in wonderland
Richard's entangled aventures in wonderlandRichard's entangled aventures in wonderland
Richard's entangled aventures in wonderland
Richard Gill
 
ESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptxESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptx
muralinath2
 
insect taxonomy importance systematics and classification
insect taxonomy importance systematics and classificationinsect taxonomy importance systematics and classification
insect taxonomy importance systematics and classification
anitaento25
 
platelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptxplatelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptx
muralinath2
 
Orion Air Quality Monitoring Systems - CWS
Orion Air Quality Monitoring Systems - CWSOrion Air Quality Monitoring Systems - CWS
Orion Air Quality Monitoring Systems - CWS
Columbia Weather Systems
 
role of pramana in research.pptx in science
role of pramana in research.pptx in sciencerole of pramana in research.pptx in science
role of pramana in research.pptx in science
sonaliswain16
 
extra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdfextra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdf
DiyaBiswas10
 
general properties of oerganologametal.ppt
general properties of oerganologametal.pptgeneral properties of oerganologametal.ppt
general properties of oerganologametal.ppt
IqrimaNabilatulhusni
 
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
Sérgio Sacani
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
silvermistyshot
 
Mammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also FunctionsMammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also Functions
YOGESH DOGRA
 
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Sérgio Sacani
 
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
Sérgio Sacani
 
Structural Classification Of Protein (SCOP)
Structural Classification Of Protein  (SCOP)Structural Classification Of Protein  (SCOP)
Structural Classification Of Protein (SCOP)
aishnasrivastava
 

Recently uploaded (20)

EY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptxEY - Supply Chain Services 2018_template.pptx
EY - Supply Chain Services 2018_template.pptx
 
Leaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdfLeaf Initiation, Growth and Differentiation.pdf
Leaf Initiation, Growth and Differentiation.pdf
 
Comparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebratesComparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebrates
 
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
Astronomy Update- Curiosity’s exploration of Mars _ Local Briefs _ leadertele...
 
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdfSCHIZOPHRENIA Disorder/ Brain Disorder.pdf
SCHIZOPHRENIA Disorder/ Brain Disorder.pdf
 
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
Deep Behavioral Phenotyping in Systems Neuroscience for Functional Atlasing a...
 
Richard's entangled aventures in wonderland
Richard's entangled aventures in wonderlandRichard's entangled aventures in wonderland
Richard's entangled aventures in wonderland
 
ESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptxESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptx
 
insect taxonomy importance systematics and classification
insect taxonomy importance systematics and classificationinsect taxonomy importance systematics and classification
insect taxonomy importance systematics and classification
 
platelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptxplatelets- lifespan -Clot retraction-disorders.pptx
platelets- lifespan -Clot retraction-disorders.pptx
 
Orion Air Quality Monitoring Systems - CWS
Orion Air Quality Monitoring Systems - CWSOrion Air Quality Monitoring Systems - CWS
Orion Air Quality Monitoring Systems - CWS
 
role of pramana in research.pptx in science
role of pramana in research.pptx in sciencerole of pramana in research.pptx in science
role of pramana in research.pptx in science
 
extra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdfextra-chromosomal-inheritance[1].pptx.pdfpdf
extra-chromosomal-inheritance[1].pptx.pdfpdf
 
general properties of oerganologametal.ppt
general properties of oerganologametal.pptgeneral properties of oerganologametal.ppt
general properties of oerganologametal.ppt
 
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...
 
Lateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensiveLateral Ventricles.pdf very easy good diagrams comprehensive
Lateral Ventricles.pdf very easy good diagrams comprehensive
 
Mammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also FunctionsMammalian Pineal Body Structure and Also Functions
Mammalian Pineal Body Structure and Also Functions
 
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
Observation of Io’s Resurfacing via Plume Deposition Using Ground-based Adapt...
 
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
 
Structural Classification Of Protein (SCOP)
Structural Classification Of Protein  (SCOP)Structural Classification Of Protein  (SCOP)
Structural Classification Of Protein (SCOP)
 

Unbiased MCMC with couplings

  • 1. Unbiased MCMC with couplings Pierre E. Jacob Department of Statistics, Harvard University joint work with John O’Leary, Yves Atchad´e, and other fantastic people acknowledged throughout Department of Statistics, University of Oxford July 19, 2019 Pierre E. Jacob Unbiased MCMC
  • 2. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 3. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 4. Setting Continuous or discrete space of dimension d. Target probability distribution π, with probability density function x → π(x). Goal: approximate π, e.g. approximate Eπ[h(X)] = h(x)π(x)dx = π(h), for a class of “test” functions h. Pierre E. Jacob Unbiased MCMC
  • 5. Markov chain Monte Carlo Initially, X0 ∼ π0, then Xt|Xt−1 ∼ P(Xt−1, ·) for t = 1, . . . , T. Estimator: 1 T − b T t=b+1 h(Xt), where b iterations are discarded as burn-in. Might converge to Eπ[h(X)] as T → ∞ by the ergodic theorem. Biased for any fixed b, T, since π0 = π. Averaging independent copies of such estimators for fixed T, b would not provide a consistent estimator of Eπ[h(X)] as the number of independent copies goes to infinity. Pierre E. Jacob Unbiased MCMC
  • 6. Example: Metropolis–Hastings kernel P Initialize: X0 ∼ π0. At each iteration t ≥ 0, with Markov chain at state Xt, 1 propose X ∼ k(Xt, ·), 2 sample U ∼ U([0, 1]), 3 if U ≤ π(X )k(X , Xt) π(Xt)k(Xt, X ) , set Xt+1 = X , otherwise set Xt+1 = Xt. Hastings, Monte Carlo sampling methods using Markov chains and their applications, Biometrika, 1970. Pierre E. Jacob Unbiased MCMC
  • 7. MCMC path π = N(0, 1), MH with Normal proposal std = 0.5, π0 = N(10, 32 ) Pierre E. Jacob Unbiased MCMC
  • 8. MCMC marginal distributions π = N(0, 1), RWMH with Normal proposal std = 0.5, π0 = N(10, 32 ) Pierre E. Jacob Unbiased MCMC
  • 9. Parallel computing R processors generate unbiased estimators, independently. The number of estimators completed by time t is random. There are different ways of aggregating estimators, either discarding on-going calculations at time t or not. Different regimes can be considered, e.g. as R → ∞ for fixed t, as t → ∞ for fixed R, as t and R both go to infinity. Glynn & Heidelberger, Analysis of parallel replicated simulations under a completion time constraint, 1991. Pierre E. Jacob Unbiased MCMC
  • 10. Parallel computing Pierre E. Jacob Unbiased MCMC
  • 11. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 12. Coupled chains Generate two Markov chains (Xt) and (Yt) as follows, sample X0 and Y0 from π0 (independently or not), sample X1|X0 ∼ P(X0, ·), for t ≥ 1, sample (Xt+1, Yt)|(Xt, Yt−1) ∼ ¯P ((Xt, Yt−1), ·), ¯P is such that Xt+1|Xt ∼ P(Xt, ·) and Yt|Yt−1 ∼ P(Yt−1, ·), so each chain marginally evolves according to P. ⇒ Xt and Yt have the same distribution for all t ≥ 0. ¯P is also such that ∃τ such that Xτ = Yτ−1, and the chains are faithful. Pierre E. Jacob Unbiased MCMC
  • 13. Debiasing idea (one slide version) Limit as a telescopic sum, for all k ≥ 0, Eπ[h(X)] = lim t→∞ E[h(Xt)] = E[h(Xk)] + ∞ t=k+1 E[h(Xt) − h(Xt−1)]. Since for all t ≥ 0, Xt and Yt have the same distribution, = E[h(Xk)] + ∞ t=k+1 E[h(Xt) − h(Yt−1)]. If we can swap expectation and limit, = E[h(Xk) + ∞ t=k+1 (h(Xt) − h(Yt−1))], so h(Xk) + ∞ t=k+1(h(Xt) − h(Yt−1)) is unbiased. Pierre E. Jacob Unbiased MCMC
  • 14. Unbiased estimators Unbiased estimator is given by Hk(X, Y ) = h(Xk) + τ−1 t=k+1 (h(Xt) − h(Yt−1)), with the convention τ−1 t=k+1{·} = 0 if τ − 1 < k + 1. h(Xk) alone is biased; the other terms correct for the bias. Cost: τ − 1 calls to ¯P and 1 + max(0, k − τ) calls to P. Glynn & Rhee, Exact estimation for Markov chain equilibrium expectations, 2014. Note: same reasoning would work with arbitrary lags L ≥ 1. Pierre E. Jacob Unbiased MCMC
  • 15. Conditions Jacob, O’Leary, Atchad´e, Unbiased MCMC with couplings. 1 Marginal chain converges: E[h(Xt)] → Eπ[h(X)], and h(Xt) has (2 + η)-finite moments for all t. 2 Meeting time τ has geometric tails: ∃C < +∞ ∃δ ∈ (0, 1) ∀t ≥ 0 P(τ > t) ≤ Cδt . 3 Chains stay together: Xt = Yt−1 for all t ≥ τ. Condition 2 itself implied by e.g. geometric drift condition. Under these conditions, Hk(X, Y ) is unbiased, has finite expected cost and finite variance, for all k. Pierre E. Jacob Unbiased MCMC
  • 16. Conditions: update Middleton, Deligiannidis, Doucet, Jacob, Unbiased MCMC for intractable target distributions. 1 Marginal chain converges: E[h(Xt)] → Eπ[h(X)], and h(Xt) has (2 + η)-finite moments for all t. 2 Meeting time τ has polynomial tails: ∃C < +∞ ∃δ > 2(2η−1 + 1) ∀t ≥ 0 P(τ > t) ≤ Ct−δ . 3 Chains stay together: Xt = Yt−1 for all t ≥ τ. Condition 2 itself implied by e.g. polynomial drift condition. Pierre E. Jacob Unbiased MCMC
  • 17. Improved unbiased estimators Efficiency matters, thus in practice we recommend a variation of the previous estimator, defined for integers k ≤ m as Hk:m(X, Y ) = 1 m − k + 1 m t=k Ht(X, Y ) which can also be written 1 m − k + 1 m t=k h(Xt)+ τ−1 t=k+1 min 1, t − k m − k + 1 (h(Xt)−h(Yt−1)), i.e. standard MCMC average + bias correction term. As k → ∞, bias correction is zero with increasing probability. Note: changing the lag is another way of modifying efficiency while preserving the lack of bias. Pierre E. Jacob Unbiased MCMC
  • 18. Efficiency Writing Hk:m(X, Y ) = MCMCk:m + BCk:m, then, denoting the MSE of MCMCk:m by MSEk:m, V[Hk:m(X, Y )] ≤ MSEk:m+2 MSEk:m E BC2 k:m +E BC2 k:m . Under geometric drift condition and regularity assumptions on h, for some δ < 1, C < ∞, E[BC2 k:m] ≤ Cδk (m − k + 1)2 , and δ is directly related to tails of the meeting time. Similarly under polynomial drift, we obtain a term of the form k−δ in the numerator instead of δk. Pierre E. Jacob Unbiased MCMC
  • 19. Signed measure estimator Replacing function evaluations by delta masses leads to ˆπ(·) = 1 m − k + 1 m t=k δXt (·) + τ−1 t=k+1 min 1, t − k m − k + 1 (δXt (·) − δYt−1 (·)), which is of the form ˆπ(·) = N n=1 ωnδZn (·), where N n=1 ωn = 1 but some ωn might be negative. Unbiasedness reads: E[Eˆπ[h(X)]] = Eπ[h(X)]. Pierre E. Jacob Unbiased MCMC
  • 20. Assessing convergence of MCMC Total variation distance between Xk ∼ πk and π = limk→∞ πk: πk − π TV = 1 2 sup h:|h|≤1 |E[h(Xk)] − Eπ[h(X)]| = 1 2 sup h:|h|≤1 |E[ τ−1 t=k+1 h(Xt) − h(Yt−1)]| ≤ E[max(0, τ − k − 1)]. 0.00 0.01 0.02 0.03 0.04 0 50 100 150 200 meeting time density 1e−03 1e−02 1e−01 1e+00 1e+01 0 50 100 k upperbound Pierre E. Jacob Unbiased MCMC
  • 21. Assessing convergence of MCMC With L-lag couplings, τ(L) = inf{t ≥ L : Xt = Yt−L}, dTV (πk, π) ≤ E 0 ∨ (τ(L) − L − k)/L , dW (πk, π) ≤ E (τ(L)−L−k)/L j=1 Xk+jL − Yk+(j−1)L 1 . 0.00 0.25 0.50 0.75 1.00 1e+01 1e+02 1e+03 1e+04 1e+05 1e+06 iterations dTV SSG PT Biswas & Jacob, Estimating Convergence of Markov chains with L-Lag Couplings, 2019. Pierre E. Jacob Unbiased MCMC
  • 22. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 23. Designing coupled chains To implement the proposed unbiased estimators, we need to sample from a Markov kernel ¯P, such that, when (Xt+1, Yt) is sampled from ¯P ((Xt, Yt−1), ·), marginally Xt+1|Xt ∼ P(Xt, ·), and Yt|Yt−1 ∼ P(Yt−1, ·), it is possible that Xt+1 = Yt exactly for some t ≥ 0, if Xt = Yt−1, then Xt+1 = Yt almost surely. Pierre E. Jacob Unbiased MCMC
  • 24. Couplings of MCMC algorithms Many practical couplings in the literature. . . Propp & Wilson, Exact sampling with coupled Markov chains and applications to statistical mechanics, Random Structures & Algorithms, 1996. Johnson, Studying convergence of Markov chain Monte Carlo algorithms using coupled sample paths, JASA, 1996. Neal, Circularly-coupled Markov chain sampling, UoT tech report, 1999. Pinto & Neal, Improving Markov chain Monte Carlo estimators by coupling to an approximating chain, UoT tech report, 2001. Glynn & Rhee, Exact estimation for Markov chain equilibrium expectations, Journal of Applied Probability, 2014. Pierre E. Jacob Unbiased MCMC
  • 25. Couplings (X, Y ) follows a coupling of p and q if X ∼ p and Y ∼ q. The coupling inequality states that P(X = Y ) ≤ 1 − p − q TV, for any coupling, with p − q TV = 1 2 |p(x) − q(x)|dx. Maximal couplings achieve the bound. Pierre E. Jacob Unbiased MCMC
  • 26. Maximal coupling of Gamma and Normal Pierre E. Jacob Unbiased MCMC
  • 27. Maximal coupling: sampling algorithm Requires: evaluations of p and q, sampling from p and q. 1 Sample X ∼ p and W ∼ U([0, 1]). If W ≤ q(X)/p(X), output (X, X). 2 Otherwise, sample Y ∼ q and W ∼ U([0, 1]) until W > p(Y )/q(Y ), and output (X, Y ). Output: a pair (X, Y ) such that X ∼ p, Y ∼ q and P(X = Y ) is maximal. Pierre E. Jacob Unbiased MCMC
  • 28. Back to Metropolis–Hastings (kernel P) At each iteration t, Markov chain at state Xt, 1 propose X ∼ k(Xt, ·), 2 sample U ∼ U([0, 1]), 3 if U ≤ π(X )k(X , Xt) π(Xt)k(Xt, X ) , set Xt+1 = X , otherwise set Xt+1 = Xt. How to propagate two MH chains from states Xt and Yt−1 such that {Xt+1 = Yt} can happen? Pierre E. Jacob Unbiased MCMC
  • 29. Coupling of Metropolis–Hastings (kernel ¯P) At each iteration t, two Markov chains at states Xt, Yt−1, 1 propose (X , Y ) from max coupling of k(Xt, ·), k(Yt−1, ·), 2 sample U ∼ U([0, 1]), 3 if U ≤ π(X )k(X , Xt) π(Xt)k(Xt, X ) , set Xt+1 = X , otherwise set Xt+1 = Xt, if U ≤ π(Y )k(Y , Yt−1) π(Yt−1)k(Yt−1, Y ) , set Yt = Y , otherwise set Yt = Yt−1. Pierre E. Jacob Unbiased MCMC
  • 30. Coupling of Gibbs Gibbs sampler: update component i of the chain, leaving π(dxi|x1, . . . , xi−1, xi+1, . . . , xd) invariant. For instance, we can propose X ∼ k(Xi t, ·) to replace Xi t, and accept or not with a Metropolis–Hastings step. These proposals can be maximally coupled across two chains, at each component update. The chains meet when all components have met. Likewise we can couple parallel tempering chains, and meeting occurs when entire ensembles of chains meet. Pierre E. Jacob Unbiased MCMC
  • 31. Hamiltonian Monte Carlo Introduce potential energy U(q) = − log π(q), and total energy E(q, p) = U(q) + 1 2|p|2. Hamiltonian dynamics for (q(s), p(s)), where s ≥ 0: d ds q(s) = pE(q(s), p(s)), d ds p(s) = − qE(q(s), p(s)). Solving Hamiltonian dynamics exactly is not feasible, but discretization + Metropolis–Hastings correction ensure that π remains invariant. Common random numbers can make two HMC chains contract, under assumptions on the target such as strong log-concavity. Heng & Jacob, Unbiased HMC with couplings, Biometrika, 2019. Pierre E. Jacob Unbiased MCMC
  • 32. Coupling of HMC Figure 2 of Mangoubi & Smith, Rapid mixing of HMC strongly log-concave distributions, 2017. Coupling two copies X1, X2, . . . (blue) and Y1, Y2, . . . (green) of HMC by choosing same momentum pi at every step. See also Bou-Rabee, Eberle & Zimmer, Coupling and Convergence for Hamiltonian Monte Carlo, 2018. Pierre E. Jacob Unbiased MCMC
  • 33. Particle Independent Metropolis–Hastings Initialization: run SMC, obtain ˆπ(0)(·), ˆZ(0), where E[ ˆZ(0)] = Z, the normalizing constant of π. At each iteration t, given (ˆπ(t)(·), ˆZ(t)), 1 run SMC, obtain (π (·), Z ), 2 sample U ∼ U([0, 1]), 3 if U ≤ Z ˆZ(t) , set (t + 1)-th state to (π (·), Z ), otherwise set (t + 1)-th state to (ˆπ(t)(·), ˆZ(t)). For any number of particles, T−1 T t=1 ˆπ(t)(·) →T→∞ π. Andrieu, Doucet, Holenstein, Particle MCMC, JRSS B, 2010. Pierre E. Jacob Unbiased MCMC
  • 34. Coupled Particle Independent Metropolis–Hastings At each iteration t, given (ˆπ(t)(·), ˆZ(t)), and (˜π(t−1)(·), ˜Z(t−1)), 1 run SMC, obtain (π (·), Z ), 2 sample U ∼ U([0, 1]), 3 if U ≤ Z ˆZ(t) , set (t + 1)-th “1st” state to (π (·), Z ), otherwise set (t + 1)-th “1st” state to (ˆπ(t)(·), ˆZ(t)). 4 if U ≤ Z ˜Z(t−1) , set t-th “2nd” state to (π (·), Z ), otherwise set t-th “2nd” state to (˜π(t−1)(·), ˜Z(t−1)). Pierre E. Jacob Unbiased MCMC
  • 35. Unbiased SMC PIMH turns any SMC sampler into an MCMC scheme that can be debiased using the coupled chains machinery. Thus any MCMC algorithm that can be plugged in an SMC sampler, can then be subsequently de-biased. No need for couplings of the MCMC chains themselves. Middleton, Deligiannidis, Doucet, Jacob, Unbiased Smoothing using Particle Independent Metropolis-Hastings, AISTATS, 2019. Pierre E. Jacob Unbiased MCMC
  • 36. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 37. Bimodal target Target is mixture of univariate Normal distributions: π = 0.5 · N(−4, 1) + 0.5 · N(+4, 1). MCMC: random walk Metropolis–Hastings, with proposal standard deviation σ that will vary. Initial distribution π0 will vary. Pierre E. Jacob Unbiased MCMC
  • 38. Bimodal target With σ = 3, π0 = N(10, 102). . . 0.00 0.05 0.10 0.15 0.20 −10 −5 0 5 10 x density Pierre E. Jacob Unbiased MCMC
  • 39. Bimodal target With σ = 1, π0 = N(10, 102). . . 0.00 0.05 0.10 0.15 0.20 −10 −5 0 5 10 x density Pierre E. Jacob Unbiased MCMC
  • 40. Bimodal target With σ = 1, π0 = N(10, 12). . . 0.0 0.1 0.2 0.3 −10 −5 0 5 10 x density Pierre E. Jacob Unbiased MCMC
  • 41. Outline 1 It is about Monte Carlo and bias 2 Unbiased estimators from Markov kernels 3 Design of coupled Markov chains 4 It can fail 5 Beyond parallel computation and burn-in Pierre E. Jacob Unbiased MCMC
  • 42. Unbiased property Unbiased MCMC estimators can be used to tackle problems for which standard MCMC methods are ill-suited. For instance, problems where one needs to approximate h(x1, x2)π2(x2|x1)dx2 π1(x1)dx1, or equivalently E1 [E2[h(X1, X2)|X1]] . Pierre E. Jacob Unbiased MCMC
  • 43. Example 1: cut distribution First model, parameter θ1, data Y1, posterior π1(θ1|Y1). Second model, parameter θ2, data Y2, conditional posterior: π2(θ2|Y2, θ1) = p2(θ2|θ1)p2(Y2|θ1, θ2) p2(Y2|θ1) . Plummer, Cuts in Bayesian graphical models, 2015. Jacob, Murray, Holmes, Robert, Better together? Statistical learning in models made of modules. Pierre E. Jacob Unbiased MCMC
  • 44. Example: epidemiological study Model of virus prevalence ∀i = 1, . . . , I Zi ∼ Binomial(Ni, ϕi), Zi is number of women infected with high-risk HPV in a sample of size Ni in country i. Impact of prevalence onto cervical cancer occurrence ∀i = 1, . . . , I Yi ∼ Poisson(λiTi), log(λi) = θ2,1 + θ2,2ϕi, Yi is number of cancer cases arising from Ti woman-years of follow-up in country i. Plummer, Cuts in Bayesian graphical models, 2015. Pierre E. Jacob Unbiased MCMC
  • 45. Example: epidemiological study Approximations of the cut distribution marginals: 0 1 2 3 −2.5 −2.0 −1.5 θ2,1 density 0.00 0.05 0.10 0.15 10 15 20 25 θ2,2 density Red curves were obtained by long MCMC run targeting π2(θ2|Y2, θ1), for draws θ1 from first posterior π1(θ1|Y1). Black bars were obtained with unbiased MCMC. Pierre E. Jacob Unbiased MCMC
  • 46. Example 2: normalizing constants We might be interested in normalizing constant Z of π(x) ∝ exp(−U(x)), defined as Z = exp(−U(x))dx. Introduce ∀λ ∈ [0, 1] πλ(x) ∝ exp(−Uλ(x)). Thermodynamics integration or path sampling identity: log Z1 Z0 = − 1 0 λUλ(x)πλ(dx) q(λ) q(dλ), where q(dλ) is any distribution supported on [0, 1]. Rischard, Jacob, Pillai, Unbiased estimation of log normalizing constants with applications to Bayesian cross-validation. Pierre E. Jacob Unbiased MCMC
  • 47. Example 3: Bayesian cross-validation Data y = {y1, . . . , yn}, made of n units. Partition y into training T and validation V sets. For each split y = {T, V }, assess predictive performance with log p(V |T) = log p(V |θ, T)p(dθ|T). Note, log p(V |T) is a log-ratio of normalizing constants. Finally we might want to approximate CV = n nT −1 T,V log p(V |T). Rischard, Jacob, Pillai, Unbiased estimation of log normalizing constants with applications to Bayesian cross-validation. Pierre E. Jacob Unbiased MCMC
  • 48. Discussion Perfect samplers, designed to sample i.i.d. from π, would yield the same benefits and more. What about regeneration? Implications of lack of bias are numerous. Proposed estimators can be obtained using coupled MCMC kernels, or using the “MCMC → SMC → coupled PIMH” route. Cannot work if underlying MCMC doesn’t work. . . but parallelism allows for more expensive MCMC kernels. Thank you for listening! Pierre E. Jacob Unbiased MCMC
  • 49. References with John O’Leary, Yves F. Atchad´e Unbiased Markov chain Monte Carlo with couplings, 2019. with Fredrik Lindsten, Thomas Sch¨on Smoothing with Couplings of Conditional Particle Filters, 2018. with Jeremy Heng Unbiased Hamiltonian Monte Carlo with couplings, 2019. with Lawrence Middleton, George Deligiannidis, Arnaud Doucet Unbiased Markov chain Monte Carlo for intractable target distributions, 2019. Unbiased Smoothing using Particle Independent Metropolis-Hastings, 2019. Pierre E. Jacob Unbiased MCMC
  • 50. References with Maxime Rischard, Natesh Pillai Unbiased estimation of log normalizing constants with applications to Bayesian cross-validation, 2019. with Niloy Biswas Estimating Convergence of Markov chains with L-Lag Couplings, 2019. Funding provided by the National Science Foundation, grants DMS-1712872 and DMS-1844695. Pierre E. Jacob Unbiased MCMC