SlideShare a Scribd company logo
Inference via Bayesian Synthetic Likelihoods for
a Mixed-Effects SDE Model of Tumor Growth
Umberto Picchini
Centre for Mathematical Sciences,
Lund University
European Meeting of Statisticians
Helsinki 24-28 July 2017
Umberto Picchini (umberto@maths.lth.se)
This is joint ongoing work with Julie Lyng Forman (Biostatistics unit,
University of Copenhagen).
This presentation is based on the working paper:
P. and Forman (2017). Stochastic differential equation mixed effects
models for tumor growth and response to treatment,
arXiv:1607.02633.
Umberto Picchini (umberto@maths.lth.se)
In this talk we have three main goals:
Introduce a state-space model for tumor growth in mice, with
dynamics driven by a stochastic differential equation (SDE).
Formulate a mixed-effects SDE model for population estimation.
Show how to produce approximate Bayesian inference for our
mixed-effects SDE model using synthetic likelihoods.
Should we decide to make our model more complex, we can
seriously consider the synthetic likelihood approach for
non-state-space models having intractable likelihoods.
Umberto Picchini (umberto@maths.lth.se)
In this talk we have three main goals:
Introduce a state-space model for tumor growth in mice, with
dynamics driven by a stochastic differential equation (SDE).
Formulate a mixed-effects SDE model for population estimation.
Show how to produce approximate Bayesian inference for our
mixed-effects SDE model using synthetic likelihoods.
Should we decide to make our model more complex, we can
seriously consider the synthetic likelihood approach for
non-state-space models having intractable likelihoods.
Umberto Picchini (umberto@maths.lth.se)
Nowadays there are several ways to deal with “intractable likelihoods”.
“Plug-and-play methods”: the only requirements is the ability to simulate
from the data-generating-model.
1 particle marginal methods (PMMH, PMCMC) based on SMC filters
[Andrieu and Roberts 2009, Andrieu et al 2010].
2 (improved) Iterated filtering [Ionides et al. 2015]
3 approximate Bayesian computation (ABC) [Marin et al. 2012].
4 Synthetic likelihoods [Wood 2010].
(1)-(2) easily accommodate models of state-space type (Markovian
dynamics, conditionally independent measurements).
(3)-(4) do not impose any structure on the model. You only need to be able
to generate realizations from the model.
In the following I focus on Synthetic Likelihoods.
Umberto Picchini (umberto@maths.lth.se)
Our experiment: a tumor xenography study
a skin tumor is grown in each mouse in the study.
3 groups of mice: 2 groups getting an experimental treatment; 1
control group (no treatment).
experimental groups get treated with chemio or radiation therapy.
we wish to assess the effect of the treatments on tumor growth,
that is estimate model parameters.
Only 5–8 mice per group. Data are sparse.
group 1: chemio therapy;
group 3: combined chemio-radio therapy;
group 5: no treatment
Umberto Picchini (umberto@maths.lth.se)
Three experimental groups
0 5 10 15 20 25 30 35 40
days
3.5
4
4.5
5
5.5
6
6.5
7
7.5
logvolume(mm3
)
group 1
0 5 10 15 20 25 30 35 40
days
2.5
3
3.5
4
4.5
5
5.5
6
6.5
7
7.5
logvolume(mm
3
)
group 3
0 5 10 15 20 25 30 35
days
3
3.5
4
4.5
5
5.5
6
6.5
7
7.5
8
logvolume(mm3
)
group 5
Figure: Data of log-volumes (mm3
) for the three groups.
Umberto Picchini (umberto@maths.lth.se)
Figure: Source http://www.nature.com/articles/srep04384
Umberto Picchini (umberto@maths.lth.se)
Mixed-effects modelling
Repeated measurements taken on a series of individuals/animals play
an important role in biomedical research.
Say that we have measurements on M subjects (mice).
Assume responses following the same model form for all
subjects
Each subject has its own individual parameters φi
φi ∼ p(φ|θ), i = 1, ..., M
θ are fixed (yet unknown) population parameters.
it may be desirable to consider random variations into individual
process dynamics (⇒ stochastic differential equations)
Umberto Picchini (umberto@maths.lth.se)
We formulate a state-space model accounting for:
intra-individual variation: explained via an SDE;
between-individuals variation: modelled by assuming “mixed
effects” φi ∼ p(φ|θ). Interest is on estimating θ.
residual variation.
Our data represent the size of the total volume Vi(t) at time t for
subject i = 1, ..., M.
For subject i, a fraction αi of the tumor volume has cells killed by the
treatment, 0 αi 1.



Vi(t) = Vsurv
i (t) + Vkill
i (t) dynamics are in the following slides
Vkill
i (0) = αivi,0 fraction of killed tumor volume
Vsurv
i (0) = (1 − αi)vi,0 fraction of survived tumor volume
vi,0 = 100 [mm3] known starting tumor volume
Umberto Picchini (umberto@maths.lth.se)
We formulate a state-space model accounting for:
intra-individual variation: explained via an SDE;
between-individuals variation: modelled by assuming “mixed
effects” φi ∼ p(φ|θ). Interest is on estimating θ.
residual variation.
Our data represent the size of the total volume Vi(t) at time t for
subject i = 1, ..., M.
For subject i, a fraction αi of the tumor volume has cells killed by the
treatment, 0 αi 1.



Vi(t) = Vsurv
i (t) + Vkill
i (t) dynamics are in the following slides
Vkill
i (0) = αivi,0 fraction of killed tumor volume
Vsurv
i (0) = (1 − αi)vi,0 fraction of survived tumor volume
vi,0 = 100 [mm3] known starting tumor volume
Umberto Picchini (umberto@maths.lth.se)
Umberto Picchini (umberto@maths.lth.se)
SDE mixed effects model
For subject i we take ni measurements.
Yij = log(Vij) + εij, i = 1, ..., M; j = 1, ..., ni
Vi(t) = Vsurv
i (t) + Vkill
i (t),
dVsurv
i (t) = (βi + γ2
/2)Vsurv
i (t)dt + γVsurv
i (t)dBi(t), Vsurv
i (0) = (1 − αi)vi,0
dVkill
i (t) = (−δi + τ2
/2)Vkill
i (t)dt + τVkill
i (t)dWi(t), Vkill
i (0) = αivi,0.
We assume Gaussian random effects, one realization per individual:
βi ∼ N( ¯β, σ2
β); δi ∼ N(¯δ, σ2
δ); αi ∼ N(0,1)( ¯α, σ2
α)
hence
φi = (βi, δi, αi)
And Gaussian residual variation (independent of everything else)
εij ∼iid N(0, σ2
ε)
Umberto Picchini (umberto@maths.lth.se)
Data Yij|Vi(tj) are conditionally independent.
Latent state {Vi(t)} is Markovian, conditionally on random effects.
The model is of state space type.
We wish to fit the model to the entire pool of data for M subjects.
Notice that data are very sparse, which makes inference challenging.
We estimate all population parameters and residual variation:
θ = ( ¯β, ¯δ, ¯α
means random effects
, σ2
β, σ2
δ, σ2
α
variances random effects
, γ, τ
intra-subj variation
, σ2
ε
residual variance
)
Umberto Picchini (umberto@maths.lth.se)
For random effect φi = (βi, δi, αi) and data yi = {yij} for subject i the
intractable likelihood for subject i is:
p(yi|θ) = p(yi|φi; θ)p(φi|θ)dφi
= p(yi|xi; θ)p(xi|φi; θ)dxi p(φi|θ)dφi
=
ni
j=1
p(yij|xij, φi, θ)p(xi,j|xi,j−1, φi; θ) p(xi0|φi, θ)dxi
× p(φi|θ)dφi
and the full likelihood for all subjects y = (y1, ..., yM) is
p(y|θ) =
M
i=1
p(yi|θ)
Umberto Picchini (umberto@maths.lth.se)
The previous intractable likelihood is manageable via particle filters
(sequential Monte Carlo).
What if the model is not of state-space type? Then the likelihood
would be even more intractable!
Umberto Picchini (umberto@maths.lth.se)
Synthetic Likelihoods (Wood, 2010)
Regardless the specific application, assume the following:
y: observed data, from static or dynamic models
s(y): (vector of) summary statistics of data, e.g. mean,
autocorrelations, marginal quantiles etc.
assume
s(y) ∼ N(µθ, Σθ)
an assumption justifiable via second order Taylor expansion
(same as in Laplace approximations).
µθ and Σθ unknown: estimate them via simulations.
Approach justifiable for very noisy models. Summary statistics
retain essential features of the data. Also useful for near-chaotic
models (very irregular likelihood).
Umberto Picchini (umberto@maths.lth.se)
nature09319-f2.2.jpg (JPEG Image, 946 × 867 pixels) - Scaled (84%) http://www.nature.com/nature/journal/v466/n7310/images/nature09319...
Figure: Figure from Wood 2010.
Umberto Picchini (umberto@maths.lth.se)
For fixed θ we simulate N artificial datasets y∗
1 , ..., y∗
N and compute
corresponding (possibly vector valued) summaries s∗
1 , ..., s∗
N.
compute
ˆµθ =
1
N
N
i=1
s∗
i , ˆΣθ =
1
N − 1
N
i=1
(s∗
i − ˆµθ)(s∗
i − ˆµθ)
compute the statistics sobs for the observed data y.
evaluate a multivariate Gaussian likelihood at sobs
LN(sobs|θ) := exp(lN(sobs|θ)) ∝
1
| ˆΣθ|
e−(sobs− ˆµθ) ˆΣ−1
θ (sobs− ˆµθ)/2
This synthetic likelihood can be maximized w.r.t. θ or be plugged in a
(marginal) MCMC algorithm for Bayesian inference
πN(θ|sobs) ∝ LN(sobs|θ)π(θ)
Umberto Picchini (umberto@maths.lth.se)
Bayesian synthetic likelihoods
Actually we follow Price et al 2017.1
Construct an unbiased estimator ˜LN for a Gaussian likelihood
(Ghurye and Olkin, 1969), this implies that for any statistic s
E(˜LN(s|θ)) = L(s|θ) = N(s; µθ, Σθ)
plug ˜LN(sobs|θ) into a MCMC algorithm for inference on θ.
resulting draws have stationary distribution π(θ|sobs) not
πN(θ|sobs), whenever sobs is Gaussian.
The above is true regardless of the value of N.
The latter follows from Beaumont 2003, Andrieu and Roberts 2009.
1
Price, Drovandi, Lee and Nott. Bayesian synthetic likelihood. 2017, JCGS.
Umberto Picchini (umberto@maths.lth.se)
Recall we have not one but M subjects to fit simultaneously.
Data are y = (y1, ..., yM).
We construct the following vector-statistics:
s = (sindiv
1 , ..., sindiv
M , sbetween
)
For subject i individual summaries sindiv
i contain:
mean absolute deviation for subject i;
slope of the line segment connecting the first and the last
observation, (yi(tni ) − yi(t1))/(tni − t1);
first two measurement values yi(t1), yi(t2);
the estimated slope ˆβi1 from the autoregression
E(yij) = βi0 + βi1yi,j−1
Umberto Picchini (umberto@maths.lth.se)
Inter-individuals summaries sbetween include:
MAD{yi1}i=1:M, the mean absolute deviation between subjects at
the first time point;
the same as above but for the second time point.
These are useful to understand the “width” of the variability between
trajectories. Remember
0 5 10 15 20 25 30 35 40
days
2.5
3
3.5
4
4.5
5
5.5
6
6.5
7
7.5
logvolume(mm3
)
group 3
Umberto Picchini (umberto@maths.lth.se)
Therefore to run a single iteration of an MCMC algorithm using
synthetic likelihoods we must:
simulate M independent realizations of the random effects, and
corresponding M subjects trajectories;
do the above N times (can be done in parallel);
compute summary statistics for the M × N trajectories.
Umberto Picchini (umberto@maths.lth.se)
A particle marginal algorithm for exact Bayesian inference
for i = 1, ..., M do
draw φl
i ∼ p(φi|θ)
if j = 1 then
Sample xl
i1 ∼ p(xi1|x0, φl
i; θ).
Compute wl
i1 = p(yi1|xl
i1) and ˆp(yi1) = L
l=1 wl
i1/L.
Normalization: ˜wl
i1 := wl
i1/ L
l=1 wl
i1.
Resampling: sample L times with replacement from {xl
i1, ˜wl
i1}. Denote the
sampled particles with ˜xl
i1.
end if
for j = 2, ..., ni do
Forward propagation: sample xl
ij ∼ p(xij|˜xl
i,j−1, φl
i; θ).
Compute wl
ij = p(yij|xl
ij) and normalise ˜wl
ij := wl
ij/ L
l=1 wl
ij
Compute ˆp(yij|yi,1:j−1) = L
l=1 wl
ij/L
Resample L times with replacement from {xl
ij, ˜wl
ij}. Sampled particles are ˜xl
ij.
end for
end for
Umberto Picchini (umberto@maths.lth.se)
Each iteration of the previous for loop gives an unbiased ˆp(yi|θ).
Since E[ˆp(yi|θ)] = p(yi|θ)
and since all the ˆp(yi|θ)) are independent one of the other
then E[ M
i=1 ˆp(yi|θ)] = M
i=1 p(yi|θ)
The above means that the overall likelihood for our mixed effects
model can be estimated unbiasedly.
Therefore exact Bayesian inference can be obtained using
pseudo-marginal arguments (e.g. Andrieu and Roberts 20092):
We can sample exactly from π(θ|y) via Metropolis-Hastings.
2
Andrieu and Roberts 2009. The pseudo-marginal approach for efficient Monte
Carlo computations. The Annals of Statistics: 697-725.
Umberto Picchini (umberto@maths.lth.se)
Group 3 results: marginal posteriors
Solid lines: exact Bayesian (particle MCMC) targeting π(θ|y).
Dashed lines: synthetic likelihoods posteriors π(θ|sobs).
Dotted lines: priors.
-1 -0.5 0 0.5 1 1.5 2 2.5
0
0.5
1
1.5
2
2.5
3
3.5
(a) log ¯β
-2 -1 0 1 2 3
0
0.2
0.4
0.6
0.8
1
(b) log ¯δ
0 0.2 0.4 0.6 0.8 1
0
0.5
1
1.5
2
2.5
3
(c) ¯α
0 1 2 3 4
0
0.5
1
1.5
2
2.5
(d) γ
Umberto Picchini (umberto@maths.lth.se)
0 1 2 3 4 5 6
0
0.2
0.4
0.6
0.8
1
1.2
(e) τ
0 0.5 1 1.5 2 2.5 3 3.5
0
0.5
1
1.5
2
2.5
3
3.5
(f) σβ
0 0.5 1 1.5 2 2.5 3
0
0.5
1
1.5
2
2.5
3
3.5
(g) σδ
0 0.5 1 1.5
0
0.5
1
1.5
2
2.5
3
3.5
4
4.5
(h) σα
60
Umberto Picchini (umberto@maths.lth.se)
Identify treatment efficacy with more subjects
The previous poor-data scenario shows difficulties in identify
between-subjects variability with (too) few subjects.
We now perform simulation studies with M = 17 subjects.
D1: a simulated dataset with 17 subjects assigned to a low
efficacy treatment, α = 0.37.
D2: a simulated dataset with 17 subjects assigned to a treatment
with high efficacy, α = 0.75.
We use Bayesian synthetic likelihoods: N = 6, 000 simulated
summaries, R = 15, 000 MCMC iterations.
Umberto Picchini (umberto@maths.lth.se)
Dashed curves: from low efficacy treatment (D1), α = 0.37.
Solid curves: from high efficacy treatment (D2), α = 0.75.
-1 -0.5 0 0.5 1 1.5 2 2.5
0
2
4
6
8
10
(j) log ¯β
0 0.2 0.4 0.6 0.8 1
0
2
4
6
8
10
(k) ¯α
A larger number of subjects enables identification of treatment
efficacy.
Umberto Picchini (umberto@maths.lth.se)
Summary
Bayesian synthetic likelihoods work well for SDE mixed effects
models provided a not too small number of subjects is given.
This gives us confidence for the possibility to perform inference in
non-state-space mixed effects models.
Reference:
P. and Forman (2017). Stochastic differential equation mixed effects
models for tumor growth and response to treatment,
arXiv:1607.02633.
Umberto Picchini (umberto@maths.lth.se)
Appendix
Umberto Picchini (umberto@maths.lth.se)
Unbiased Gaussian likelihood estimate
Price et al. 2017 note than plugging-in the estimates ˆµ(θ) and ˆΣ(θ) into the
Gaussian likelihood p(s|θ) results in a biased estimate, while one could
instead use an unbiased estimator of a Gaussian likelihood (Ghurye and
Olkin, 1969) given by
ˆp(s|θ) = (2π)−d/2 c(d, N − 2)
c(d, N − 1)(1 − 1/N)d/2
|(N − 1) ˆΣN(θ)|−(n−d−2)/2
× ψ (N − 1) ˆΣN(θ) − (s − ˆµN(θ))(s − ˆµN(θ)) /(1 − 1/N)
(N−d−3)/2
where d = dim(s), π denotes the mathematical constant, N > d + 3, and for
a square matrix A the function ψ(A) is defined as ψ(A) = |A| if A is positive
definite and ψ(A) = 0 otherwise. Finally
c(k, v) = 2−kv/2
π−k(k−1)/4
/ k
i=1 Γ(1
2 (v − i + 1)).
Umberto Picchini (umberto@maths.lth.se)
Marginal distribution of simulated summaries
-4 -2 0 2 4
sintra
1
-1
0
1
2
-4 -2 0 2 4
sintra
2
-20
0
20
40
-4 -2 0 2 4
sintra
3
0
5
10
-4 -2 0 2 4
sintra
4
0
5
10
-4 -2 0 2 4
sintra
5
-2
0
2
4
-4 -2 0 2 4
sinter
1
0
1
2
3
-4 -2 0 2 4
sinter
2
0
1
2
3
Umberto Picchini (umberto@maths.lth.se)

More Related Content

What's hot

Computing near-optimal policies from trajectories by solving a sequence of st...
Computing near-optimal policies from trajectories by solving a sequence of st...Computing near-optimal policies from trajectories by solving a sequence of st...
Computing near-optimal policies from trajectories by solving a sequence of st...Université de Liège (ULg)
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsChristian Robert
 
Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Umberto Picchini
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forestsChristian Robert
 
Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Umberto Picchini
 
(Approximate) Bayesian computation as a new empirical Bayes (something)?
(Approximate) Bayesian computation as a new empirical Bayes (something)?(Approximate) Bayesian computation as a new empirical Bayes (something)?
(Approximate) Bayesian computation as a new empirical Bayes (something)?Christian Robert
 
ABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsUmberto Picchini
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityChristian Robert
 
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Christian Robert
 
Testing as estimation: the demise of the Bayes factor
Testing as estimation: the demise of the Bayes factorTesting as estimation: the demise of the Bayes factor
Testing as estimation: the demise of the Bayes factorChristian Robert
 
Machine Learning for Actuaries
Machine Learning for ActuariesMachine Learning for Actuaries
Machine Learning for ActuariesArthur Charpentier
 
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...SYRTO Project
 
Nber slides11 lecture2
Nber slides11 lecture2Nber slides11 lecture2
Nber slides11 lecture2NBER
 
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...IOSRJM
 
Finding statistically significant interactions between continuous features (I...
Finding statistically significant interactions between continuous features (I...Finding statistically significant interactions between continuous features (I...
Finding statistically significant interactions between continuous features (I...Thien Q. Tran
 
Pattern Recognition
Pattern RecognitionPattern Recognition
Pattern RecognitionEunho Lee
 
testing as a mixture estimation problem
testing as a mixture estimation problemtesting as a mixture estimation problem
testing as a mixture estimation problemChristian Robert
 

What's hot (20)

Computing near-optimal policies from trajectories by solving a sequence of st...
Computing near-optimal policies from trajectories by solving a sequence of st...Computing near-optimal policies from trajectories by solving a sequence of st...
Computing near-optimal policies from trajectories by solving a sequence of st...
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in Statistics
 
Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)Intro to Approximate Bayesian Computation (ABC)
Intro to Approximate Bayesian Computation (ABC)
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forests
 
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
 
Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...
 
(Approximate) Bayesian computation as a new empirical Bayes (something)?
(Approximate) Bayesian computation as a new empirical Bayes (something)?(Approximate) Bayesian computation as a new empirical Bayes (something)?
(Approximate) Bayesian computation as a new empirical Bayes (something)?
 
ABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space models
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard University
 
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
 
Testing as estimation: the demise of the Bayes factor
Testing as estimation: the demise of the Bayes factorTesting as estimation: the demise of the Bayes factor
Testing as estimation: the demise of the Bayes factor
 
Machine Learning for Actuaries
Machine Learning for ActuariesMachine Learning for Actuaries
Machine Learning for Actuaries
 
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
 
Nber slides11 lecture2
Nber slides11 lecture2Nber slides11 lecture2
Nber slides11 lecture2
 
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...
Application of Semiparametric Non-Linear Model on Panel Data with Very Small ...
 
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
2018 MUMS Fall Course - Issue Arising in Several Working Groups: Probabilisti...
 
Finding statistically significant interactions between continuous features (I...
Finding statistically significant interactions between continuous features (I...Finding statistically significant interactions between continuous features (I...
Finding statistically significant interactions between continuous features (I...
 
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
 
Pattern Recognition
Pattern RecognitionPattern Recognition
Pattern Recognition
 
testing as a mixture estimation problem
testing as a mixture estimation problemtesting as a mixture estimation problem
testing as a mixture estimation problem
 

Viewers also liked

Stanの事後処理 LTver
Stanの事後処理 LTverStanの事後処理 LTver
Stanの事後処理 LTverdaiki hojo
 
楽にggplotを描く・整える
楽にggplotを描く・整える楽にggplotを描く・整える
楽にggplotを描く・整えるdaiki hojo
 
Stanの便利な事後処理関数
Stanの便利な事後処理関数Stanの便利な事後処理関数
Stanの便利な事後処理関数daiki hojo
 
いいからベイズ推定してみる
いいからベイズ推定してみるいいからベイズ推定してみる
いいからベイズ推定してみるMakoto Hirakawa
 
glmmstanパッケージを作ってみた
glmmstanパッケージを作ってみたglmmstanパッケージを作ってみた
glmmstanパッケージを作ってみたHiroshi Shimizu
 
状態空間モデルの実行方法と実行環境の比較
状態空間モデルの実行方法と実行環境の比較状態空間モデルの実行方法と実行環境の比較
状態空間モデルの実行方法と実行環境の比較Hiroki Itô
 

Viewers also liked (6)

Stanの事後処理 LTver
Stanの事後処理 LTverStanの事後処理 LTver
Stanの事後処理 LTver
 
楽にggplotを描く・整える
楽にggplotを描く・整える楽にggplotを描く・整える
楽にggplotを描く・整える
 
Stanの便利な事後処理関数
Stanの便利な事後処理関数Stanの便利な事後処理関数
Stanの便利な事後処理関数
 
いいからベイズ推定してみる
いいからベイズ推定してみるいいからベイズ推定してみる
いいからベイズ推定してみる
 
glmmstanパッケージを作ってみた
glmmstanパッケージを作ってみたglmmstanパッケージを作ってみた
glmmstanパッケージを作ってみた
 
状態空間モデルの実行方法と実行環境の比較
状態空間モデルの実行方法と実行環境の比較状態空間モデルの実行方法と実行環境の比較
状態空間モデルの実行方法と実行環境の比較
 

Similar to Inference via Bayesian Synthetic Likelihoods for a Mixed-Effects SDE Model of Tumor Growth

Bayesian inference for mixed-effects models driven by SDEs and other stochast...
Bayesian inference for mixed-effects models driven by SDEs and other stochast...Bayesian inference for mixed-effects models driven by SDEs and other stochast...
Bayesian inference for mixed-effects models driven by SDEs and other stochast...Umberto Picchini
 
Mixed Model Analysis for Overdispersion
Mixed Model Analysis for OverdispersionMixed Model Analysis for Overdispersion
Mixed Model Analysis for Overdispersiontheijes
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modulesChristian Robert
 
revision-notes-introductory-econometrics-lecture-1-11.pdf
revision-notes-introductory-econometrics-lecture-1-11.pdfrevision-notes-introductory-econometrics-lecture-1-11.pdf
revision-notes-introductory-econometrics-lecture-1-11.pdfMrDampha
 
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...Kazuki Yoshida
 
Using model-based statistical inference to learn about evolution
Using model-based statistical inference to learn about evolutionUsing model-based statistical inference to learn about evolution
Using model-based statistical inference to learn about evolutionErick Matsen
 
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos SYRTO Project
 
Sampling Distribution and Simulation in R
Sampling Distribution and Simulation in RSampling Distribution and Simulation in R
Sampling Distribution and Simulation in RPremier Publishers
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data sciencepujashri1975
 
A Causal Framework for Meta-Analysis, drafty
A Causal Framework for Meta-Analysis, drafty A Causal Framework for Meta-Analysis, drafty
A Causal Framework for Meta-Analysis, drafty Wei Wang
 
Linear Modeling Survival Analysis Statistics Assignment Help
Linear Modeling Survival Analysis Statistics Assignment HelpLinear Modeling Survival Analysis Statistics Assignment Help
Linear Modeling Survival Analysis Statistics Assignment HelpStatistics Assignment Experts
 
Descriptive Statistics Formula Sheet Sample Populatio.docx
Descriptive Statistics Formula Sheet    Sample Populatio.docxDescriptive Statistics Formula Sheet    Sample Populatio.docx
Descriptive Statistics Formula Sheet Sample Populatio.docxsimonithomas47935
 

Similar to Inference via Bayesian Synthetic Likelihoods for a Mixed-Effects SDE Model of Tumor Growth (20)

Bayesian inference for mixed-effects models driven by SDEs and other stochast...
Bayesian inference for mixed-effects models driven by SDEs and other stochast...Bayesian inference for mixed-effects models driven by SDEs and other stochast...
Bayesian inference for mixed-effects models driven by SDEs and other stochast...
 
Lausanne 2019 #1
Lausanne 2019 #1Lausanne 2019 #1
Lausanne 2019 #1
 
Statistical analysis by iswar
Statistical analysis by iswarStatistical analysis by iswar
Statistical analysis by iswar
 
Mixed Model Analysis for Overdispersion
Mixed Model Analysis for OverdispersionMixed Model Analysis for Overdispersion
Mixed Model Analysis for Overdispersion
 
Ch01_03.ppt
Ch01_03.pptCh01_03.ppt
Ch01_03.ppt
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modules
 
revision-notes-introductory-econometrics-lecture-1-11.pdf
revision-notes-introductory-econometrics-lecture-1-11.pdfrevision-notes-introductory-econometrics-lecture-1-11.pdf
revision-notes-introductory-econometrics-lecture-1-11.pdf
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
 
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...
Matching Weights to Simultaneously Compare Three Treatment Groups: a Simulati...
 
isi
isiisi
isi
 
Using model-based statistical inference to learn about evolution
Using model-based statistical inference to learn about evolutionUsing model-based statistical inference to learn about evolution
Using model-based statistical inference to learn about evolution
 
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos
A Dynamic Factor Model: Inference and Empirical Application. Ioannis Vrontos
 
Talk 5
Talk 5Talk 5
Talk 5
 
Sampling Distribution and Simulation in R
Sampling Distribution and Simulation in RSampling Distribution and Simulation in R
Sampling Distribution and Simulation in R
 
Module-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data scienceModule-2_Notes-with-Example for data science
Module-2_Notes-with-Example for data science
 
Data science
Data scienceData science
Data science
 
A Causal Framework for Meta-Analysis, drafty
A Causal Framework for Meta-Analysis, drafty A Causal Framework for Meta-Analysis, drafty
A Causal Framework for Meta-Analysis, drafty
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Linear Modeling Survival Analysis Statistics Assignment Help
Linear Modeling Survival Analysis Statistics Assignment HelpLinear Modeling Survival Analysis Statistics Assignment Help
Linear Modeling Survival Analysis Statistics Assignment Help
 
Descriptive Statistics Formula Sheet Sample Populatio.docx
Descriptive Statistics Formula Sheet    Sample Populatio.docxDescriptive Statistics Formula Sheet    Sample Populatio.docx
Descriptive Statistics Formula Sheet Sample Populatio.docx
 

Recently uploaded

Large scale production of streptomycin.pptx
Large scale production of streptomycin.pptxLarge scale production of streptomycin.pptx
Large scale production of streptomycin.pptxCherry
 
Lab report on liquid viscosity of glycerin
Lab report on liquid viscosity of glycerinLab report on liquid viscosity of glycerin
Lab report on liquid viscosity of glycerinossaicprecious19
 
Citrus Greening Disease and its Management
Citrus Greening Disease and its ManagementCitrus Greening Disease and its Management
Citrus Greening Disease and its Managementsubedisuryaofficial
 
Multi-source connectivity as the driver of solar wind variability in the heli...
Multi-source connectivity as the driver of solar wind variability in the heli...Multi-source connectivity as the driver of solar wind variability in the heli...
Multi-source connectivity as the driver of solar wind variability in the heli...Sérgio Sacani
 
GBSN - Biochemistry (Unit 5) Chemistry of Lipids
GBSN - Biochemistry (Unit 5) Chemistry of LipidsGBSN - Biochemistry (Unit 5) Chemistry of Lipids
GBSN - Biochemistry (Unit 5) Chemistry of LipidsAreesha Ahmad
 
Anemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditionsAnemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditionsmuralinath2
 
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptxGLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptxSultanMuhammadGhauri
 
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingAreesha Ahmad
 
ESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptxESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptxmuralinath2
 
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...muralinath2
 
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...Sérgio Sacani
 
FAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS imagesFAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS imagesAlex Henderson
 
Comparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebratesComparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebratessachin783648
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...Scintica Instrumentation
 
Transport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSETransport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSEjordanparish425
 
INSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere UniversityINSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere UniversitySteffi Friedrichs
 
Hemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptxHemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptxmuralinath2
 
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATIONPRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATIONChetanK57
 
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiologyBLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiologyNoelManyise1
 

Recently uploaded (20)

Large scale production of streptomycin.pptx
Large scale production of streptomycin.pptxLarge scale production of streptomycin.pptx
Large scale production of streptomycin.pptx
 
Lab report on liquid viscosity of glycerin
Lab report on liquid viscosity of glycerinLab report on liquid viscosity of glycerin
Lab report on liquid viscosity of glycerin
 
Citrus Greening Disease and its Management
Citrus Greening Disease and its ManagementCitrus Greening Disease and its Management
Citrus Greening Disease and its Management
 
Multi-source connectivity as the driver of solar wind variability in the heli...
Multi-source connectivity as the driver of solar wind variability in the heli...Multi-source connectivity as the driver of solar wind variability in the heli...
Multi-source connectivity as the driver of solar wind variability in the heli...
 
GBSN - Biochemistry (Unit 5) Chemistry of Lipids
GBSN - Biochemistry (Unit 5) Chemistry of LipidsGBSN - Biochemistry (Unit 5) Chemistry of Lipids
GBSN - Biochemistry (Unit 5) Chemistry of Lipids
 
Anemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditionsAnemia_ different types_causes_ conditions
Anemia_ different types_causes_ conditions
 
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptxGLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
GLOBAL AND LOCAL SCENARIO OF FOOD AND NUTRITION.pptx
 
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.
 
GBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram StainingGBSN- Microbiology (Lab 3) Gram Staining
GBSN- Microbiology (Lab 3) Gram Staining
 
ESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptxESR_factors_affect-clinic significance-Pathysiology.pptx
ESR_factors_affect-clinic significance-Pathysiology.pptx
 
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...
Circulatory system_ Laplace law. Ohms law.reynaults law,baro-chemo-receptors-...
 
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
Gliese 12 b: A Temperate Earth-sized Planet at 12 pc Ideal for Atmospheric Tr...
 
FAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS imagesFAIRSpectra - Towards a common data file format for SIMS images
FAIRSpectra - Towards a common data file format for SIMS images
 
Comparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebratesComparative structure of adrenal gland in vertebrates
Comparative structure of adrenal gland in vertebrates
 
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
(May 29th, 2024) Advancements in Intravital Microscopy- Insights for Preclini...
 
Transport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSETransport in plants G1.pptx Cambridge IGCSE
Transport in plants G1.pptx Cambridge IGCSE
 
INSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere UniversityINSIGHT Partner Profile: Tampere University
INSIGHT Partner Profile: Tampere University
 
Hemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptxHemoglobin metabolism_pathophysiology.pptx
Hemoglobin metabolism_pathophysiology.pptx
 
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATIONPRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
PRESENTATION ABOUT PRINCIPLE OF COSMATIC EVALUATION
 
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiologyBLOOD AND BLOOD COMPONENT- introduction to blood physiology
BLOOD AND BLOOD COMPONENT- introduction to blood physiology
 

Inference via Bayesian Synthetic Likelihoods for a Mixed-Effects SDE Model of Tumor Growth

  • 1. Inference via Bayesian Synthetic Likelihoods for a Mixed-Effects SDE Model of Tumor Growth Umberto Picchini Centre for Mathematical Sciences, Lund University European Meeting of Statisticians Helsinki 24-28 July 2017 Umberto Picchini (umberto@maths.lth.se)
  • 2. This is joint ongoing work with Julie Lyng Forman (Biostatistics unit, University of Copenhagen). This presentation is based on the working paper: P. and Forman (2017). Stochastic differential equation mixed effects models for tumor growth and response to treatment, arXiv:1607.02633. Umberto Picchini (umberto@maths.lth.se)
  • 3. In this talk we have three main goals: Introduce a state-space model for tumor growth in mice, with dynamics driven by a stochastic differential equation (SDE). Formulate a mixed-effects SDE model for population estimation. Show how to produce approximate Bayesian inference for our mixed-effects SDE model using synthetic likelihoods. Should we decide to make our model more complex, we can seriously consider the synthetic likelihood approach for non-state-space models having intractable likelihoods. Umberto Picchini (umberto@maths.lth.se)
  • 4. In this talk we have three main goals: Introduce a state-space model for tumor growth in mice, with dynamics driven by a stochastic differential equation (SDE). Formulate a mixed-effects SDE model for population estimation. Show how to produce approximate Bayesian inference for our mixed-effects SDE model using synthetic likelihoods. Should we decide to make our model more complex, we can seriously consider the synthetic likelihood approach for non-state-space models having intractable likelihoods. Umberto Picchini (umberto@maths.lth.se)
  • 5. Nowadays there are several ways to deal with “intractable likelihoods”. “Plug-and-play methods”: the only requirements is the ability to simulate from the data-generating-model. 1 particle marginal methods (PMMH, PMCMC) based on SMC filters [Andrieu and Roberts 2009, Andrieu et al 2010]. 2 (improved) Iterated filtering [Ionides et al. 2015] 3 approximate Bayesian computation (ABC) [Marin et al. 2012]. 4 Synthetic likelihoods [Wood 2010]. (1)-(2) easily accommodate models of state-space type (Markovian dynamics, conditionally independent measurements). (3)-(4) do not impose any structure on the model. You only need to be able to generate realizations from the model. In the following I focus on Synthetic Likelihoods. Umberto Picchini (umberto@maths.lth.se)
  • 6. Our experiment: a tumor xenography study a skin tumor is grown in each mouse in the study. 3 groups of mice: 2 groups getting an experimental treatment; 1 control group (no treatment). experimental groups get treated with chemio or radiation therapy. we wish to assess the effect of the treatments on tumor growth, that is estimate model parameters. Only 5–8 mice per group. Data are sparse. group 1: chemio therapy; group 3: combined chemio-radio therapy; group 5: no treatment Umberto Picchini (umberto@maths.lth.se)
  • 7. Three experimental groups 0 5 10 15 20 25 30 35 40 days 3.5 4 4.5 5 5.5 6 6.5 7 7.5 logvolume(mm3 ) group 1 0 5 10 15 20 25 30 35 40 days 2.5 3 3.5 4 4.5 5 5.5 6 6.5 7 7.5 logvolume(mm 3 ) group 3 0 5 10 15 20 25 30 35 days 3 3.5 4 4.5 5 5.5 6 6.5 7 7.5 8 logvolume(mm3 ) group 5 Figure: Data of log-volumes (mm3 ) for the three groups. Umberto Picchini (umberto@maths.lth.se)
  • 9. Mixed-effects modelling Repeated measurements taken on a series of individuals/animals play an important role in biomedical research. Say that we have measurements on M subjects (mice). Assume responses following the same model form for all subjects Each subject has its own individual parameters φi φi ∼ p(φ|θ), i = 1, ..., M θ are fixed (yet unknown) population parameters. it may be desirable to consider random variations into individual process dynamics (⇒ stochastic differential equations) Umberto Picchini (umberto@maths.lth.se)
  • 10. We formulate a state-space model accounting for: intra-individual variation: explained via an SDE; between-individuals variation: modelled by assuming “mixed effects” φi ∼ p(φ|θ). Interest is on estimating θ. residual variation. Our data represent the size of the total volume Vi(t) at time t for subject i = 1, ..., M. For subject i, a fraction αi of the tumor volume has cells killed by the treatment, 0 αi 1.    Vi(t) = Vsurv i (t) + Vkill i (t) dynamics are in the following slides Vkill i (0) = αivi,0 fraction of killed tumor volume Vsurv i (0) = (1 − αi)vi,0 fraction of survived tumor volume vi,0 = 100 [mm3] known starting tumor volume Umberto Picchini (umberto@maths.lth.se)
  • 11. We formulate a state-space model accounting for: intra-individual variation: explained via an SDE; between-individuals variation: modelled by assuming “mixed effects” φi ∼ p(φ|θ). Interest is on estimating θ. residual variation. Our data represent the size of the total volume Vi(t) at time t for subject i = 1, ..., M. For subject i, a fraction αi of the tumor volume has cells killed by the treatment, 0 αi 1.    Vi(t) = Vsurv i (t) + Vkill i (t) dynamics are in the following slides Vkill i (0) = αivi,0 fraction of killed tumor volume Vsurv i (0) = (1 − αi)vi,0 fraction of survived tumor volume vi,0 = 100 [mm3] known starting tumor volume Umberto Picchini (umberto@maths.lth.se)
  • 13. SDE mixed effects model For subject i we take ni measurements. Yij = log(Vij) + εij, i = 1, ..., M; j = 1, ..., ni Vi(t) = Vsurv i (t) + Vkill i (t), dVsurv i (t) = (βi + γ2 /2)Vsurv i (t)dt + γVsurv i (t)dBi(t), Vsurv i (0) = (1 − αi)vi,0 dVkill i (t) = (−δi + τ2 /2)Vkill i (t)dt + τVkill i (t)dWi(t), Vkill i (0) = αivi,0. We assume Gaussian random effects, one realization per individual: βi ∼ N( ¯β, σ2 β); δi ∼ N(¯δ, σ2 δ); αi ∼ N(0,1)( ¯α, σ2 α) hence φi = (βi, δi, αi) And Gaussian residual variation (independent of everything else) εij ∼iid N(0, σ2 ε) Umberto Picchini (umberto@maths.lth.se)
  • 14. Data Yij|Vi(tj) are conditionally independent. Latent state {Vi(t)} is Markovian, conditionally on random effects. The model is of state space type. We wish to fit the model to the entire pool of data for M subjects. Notice that data are very sparse, which makes inference challenging. We estimate all population parameters and residual variation: θ = ( ¯β, ¯δ, ¯α means random effects , σ2 β, σ2 δ, σ2 α variances random effects , γ, τ intra-subj variation , σ2 ε residual variance ) Umberto Picchini (umberto@maths.lth.se)
  • 15. For random effect φi = (βi, δi, αi) and data yi = {yij} for subject i the intractable likelihood for subject i is: p(yi|θ) = p(yi|φi; θ)p(φi|θ)dφi = p(yi|xi; θ)p(xi|φi; θ)dxi p(φi|θ)dφi = ni j=1 p(yij|xij, φi, θ)p(xi,j|xi,j−1, φi; θ) p(xi0|φi, θ)dxi × p(φi|θ)dφi and the full likelihood for all subjects y = (y1, ..., yM) is p(y|θ) = M i=1 p(yi|θ) Umberto Picchini (umberto@maths.lth.se)
  • 16. The previous intractable likelihood is manageable via particle filters (sequential Monte Carlo). What if the model is not of state-space type? Then the likelihood would be even more intractable! Umberto Picchini (umberto@maths.lth.se)
  • 17. Synthetic Likelihoods (Wood, 2010) Regardless the specific application, assume the following: y: observed data, from static or dynamic models s(y): (vector of) summary statistics of data, e.g. mean, autocorrelations, marginal quantiles etc. assume s(y) ∼ N(µθ, Σθ) an assumption justifiable via second order Taylor expansion (same as in Laplace approximations). µθ and Σθ unknown: estimate them via simulations. Approach justifiable for very noisy models. Summary statistics retain essential features of the data. Also useful for near-chaotic models (very irregular likelihood). Umberto Picchini (umberto@maths.lth.se)
  • 18. nature09319-f2.2.jpg (JPEG Image, 946 × 867 pixels) - Scaled (84%) http://www.nature.com/nature/journal/v466/n7310/images/nature09319... Figure: Figure from Wood 2010. Umberto Picchini (umberto@maths.lth.se)
  • 19. For fixed θ we simulate N artificial datasets y∗ 1 , ..., y∗ N and compute corresponding (possibly vector valued) summaries s∗ 1 , ..., s∗ N. compute ˆµθ = 1 N N i=1 s∗ i , ˆΣθ = 1 N − 1 N i=1 (s∗ i − ˆµθ)(s∗ i − ˆµθ) compute the statistics sobs for the observed data y. evaluate a multivariate Gaussian likelihood at sobs LN(sobs|θ) := exp(lN(sobs|θ)) ∝ 1 | ˆΣθ| e−(sobs− ˆµθ) ˆΣ−1 θ (sobs− ˆµθ)/2 This synthetic likelihood can be maximized w.r.t. θ or be plugged in a (marginal) MCMC algorithm for Bayesian inference πN(θ|sobs) ∝ LN(sobs|θ)π(θ) Umberto Picchini (umberto@maths.lth.se)
  • 20. Bayesian synthetic likelihoods Actually we follow Price et al 2017.1 Construct an unbiased estimator ˜LN for a Gaussian likelihood (Ghurye and Olkin, 1969), this implies that for any statistic s E(˜LN(s|θ)) = L(s|θ) = N(s; µθ, Σθ) plug ˜LN(sobs|θ) into a MCMC algorithm for inference on θ. resulting draws have stationary distribution π(θ|sobs) not πN(θ|sobs), whenever sobs is Gaussian. The above is true regardless of the value of N. The latter follows from Beaumont 2003, Andrieu and Roberts 2009. 1 Price, Drovandi, Lee and Nott. Bayesian synthetic likelihood. 2017, JCGS. Umberto Picchini (umberto@maths.lth.se)
  • 21. Recall we have not one but M subjects to fit simultaneously. Data are y = (y1, ..., yM). We construct the following vector-statistics: s = (sindiv 1 , ..., sindiv M , sbetween ) For subject i individual summaries sindiv i contain: mean absolute deviation for subject i; slope of the line segment connecting the first and the last observation, (yi(tni ) − yi(t1))/(tni − t1); first two measurement values yi(t1), yi(t2); the estimated slope ˆβi1 from the autoregression E(yij) = βi0 + βi1yi,j−1 Umberto Picchini (umberto@maths.lth.se)
  • 22. Inter-individuals summaries sbetween include: MAD{yi1}i=1:M, the mean absolute deviation between subjects at the first time point; the same as above but for the second time point. These are useful to understand the “width” of the variability between trajectories. Remember 0 5 10 15 20 25 30 35 40 days 2.5 3 3.5 4 4.5 5 5.5 6 6.5 7 7.5 logvolume(mm3 ) group 3 Umberto Picchini (umberto@maths.lth.se)
  • 23. Therefore to run a single iteration of an MCMC algorithm using synthetic likelihoods we must: simulate M independent realizations of the random effects, and corresponding M subjects trajectories; do the above N times (can be done in parallel); compute summary statistics for the M × N trajectories. Umberto Picchini (umberto@maths.lth.se)
  • 24. A particle marginal algorithm for exact Bayesian inference for i = 1, ..., M do draw φl i ∼ p(φi|θ) if j = 1 then Sample xl i1 ∼ p(xi1|x0, φl i; θ). Compute wl i1 = p(yi1|xl i1) and ˆp(yi1) = L l=1 wl i1/L. Normalization: ˜wl i1 := wl i1/ L l=1 wl i1. Resampling: sample L times with replacement from {xl i1, ˜wl i1}. Denote the sampled particles with ˜xl i1. end if for j = 2, ..., ni do Forward propagation: sample xl ij ∼ p(xij|˜xl i,j−1, φl i; θ). Compute wl ij = p(yij|xl ij) and normalise ˜wl ij := wl ij/ L l=1 wl ij Compute ˆp(yij|yi,1:j−1) = L l=1 wl ij/L Resample L times with replacement from {xl ij, ˜wl ij}. Sampled particles are ˜xl ij. end for end for Umberto Picchini (umberto@maths.lth.se)
  • 25. Each iteration of the previous for loop gives an unbiased ˆp(yi|θ). Since E[ˆp(yi|θ)] = p(yi|θ) and since all the ˆp(yi|θ)) are independent one of the other then E[ M i=1 ˆp(yi|θ)] = M i=1 p(yi|θ) The above means that the overall likelihood for our mixed effects model can be estimated unbiasedly. Therefore exact Bayesian inference can be obtained using pseudo-marginal arguments (e.g. Andrieu and Roberts 20092): We can sample exactly from π(θ|y) via Metropolis-Hastings. 2 Andrieu and Roberts 2009. The pseudo-marginal approach for efficient Monte Carlo computations. The Annals of Statistics: 697-725. Umberto Picchini (umberto@maths.lth.se)
  • 26. Group 3 results: marginal posteriors Solid lines: exact Bayesian (particle MCMC) targeting π(θ|y). Dashed lines: synthetic likelihoods posteriors π(θ|sobs). Dotted lines: priors. -1 -0.5 0 0.5 1 1.5 2 2.5 0 0.5 1 1.5 2 2.5 3 3.5 (a) log ¯β -2 -1 0 1 2 3 0 0.2 0.4 0.6 0.8 1 (b) log ¯δ 0 0.2 0.4 0.6 0.8 1 0 0.5 1 1.5 2 2.5 3 (c) ¯α 0 1 2 3 4 0 0.5 1 1.5 2 2.5 (d) γ Umberto Picchini (umberto@maths.lth.se)
  • 27. 0 1 2 3 4 5 6 0 0.2 0.4 0.6 0.8 1 1.2 (e) τ 0 0.5 1 1.5 2 2.5 3 3.5 0 0.5 1 1.5 2 2.5 3 3.5 (f) σβ 0 0.5 1 1.5 2 2.5 3 0 0.5 1 1.5 2 2.5 3 3.5 (g) σδ 0 0.5 1 1.5 0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 (h) σα 60 Umberto Picchini (umberto@maths.lth.se)
  • 28. Identify treatment efficacy with more subjects The previous poor-data scenario shows difficulties in identify between-subjects variability with (too) few subjects. We now perform simulation studies with M = 17 subjects. D1: a simulated dataset with 17 subjects assigned to a low efficacy treatment, α = 0.37. D2: a simulated dataset with 17 subjects assigned to a treatment with high efficacy, α = 0.75. We use Bayesian synthetic likelihoods: N = 6, 000 simulated summaries, R = 15, 000 MCMC iterations. Umberto Picchini (umberto@maths.lth.se)
  • 29. Dashed curves: from low efficacy treatment (D1), α = 0.37. Solid curves: from high efficacy treatment (D2), α = 0.75. -1 -0.5 0 0.5 1 1.5 2 2.5 0 2 4 6 8 10 (j) log ¯β 0 0.2 0.4 0.6 0.8 1 0 2 4 6 8 10 (k) ¯α A larger number of subjects enables identification of treatment efficacy. Umberto Picchini (umberto@maths.lth.se)
  • 30. Summary Bayesian synthetic likelihoods work well for SDE mixed effects models provided a not too small number of subjects is given. This gives us confidence for the possibility to perform inference in non-state-space mixed effects models. Reference: P. and Forman (2017). Stochastic differential equation mixed effects models for tumor growth and response to treatment, arXiv:1607.02633. Umberto Picchini (umberto@maths.lth.se)
  • 32. Unbiased Gaussian likelihood estimate Price et al. 2017 note than plugging-in the estimates ˆµ(θ) and ˆΣ(θ) into the Gaussian likelihood p(s|θ) results in a biased estimate, while one could instead use an unbiased estimator of a Gaussian likelihood (Ghurye and Olkin, 1969) given by ˆp(s|θ) = (2π)−d/2 c(d, N − 2) c(d, N − 1)(1 − 1/N)d/2 |(N − 1) ˆΣN(θ)|−(n−d−2)/2 × ψ (N − 1) ˆΣN(θ) − (s − ˆµN(θ))(s − ˆµN(θ)) /(1 − 1/N) (N−d−3)/2 where d = dim(s), π denotes the mathematical constant, N > d + 3, and for a square matrix A the function ψ(A) is defined as ψ(A) = |A| if A is positive definite and ψ(A) = 0 otherwise. Finally c(k, v) = 2−kv/2 π−k(k−1)/4 / k i=1 Γ(1 2 (v − i + 1)). Umberto Picchini (umberto@maths.lth.se)
  • 33. Marginal distribution of simulated summaries -4 -2 0 2 4 sintra 1 -1 0 1 2 -4 -2 0 2 4 sintra 2 -20 0 20 40 -4 -2 0 2 4 sintra 3 0 5 10 -4 -2 0 2 4 sintra 4 0 5 10 -4 -2 0 2 4 sintra 5 -2 0 2 4 -4 -2 0 2 4 sinter 1 0 1 2 3 -4 -2 0 2 4 sinter 2 0 1 2 3 Umberto Picchini (umberto@maths.lth.se)