An application of the "data cloning" method for parameter estimation via MLE aided by Approximate Bayesian Computation. The relevant paper is http://arxiv.org/abs/1505.06318
Un cours bien détaillé qui traite les filtres analogiques 1er et 2ème ordres.
Pour plus d'informations je suis entièrement disponible.
Sabirhamzaa@gmail.com
Un cours bien détaillé qui traite les filtres analogiques 1er et 2ème ordres.
Pour plus d'informations je suis entièrement disponible.
Sabirhamzaa@gmail.com
ELE2611 Classe 3 - Filtres analogiques linéaires IJerome LE NY
Approximations rationnelles classiques, dénormalisation de fonction de transfert.
Slides for the class 3 of the course ELE2611 (Circuits II) at Polytechnique Montreal, in French. Videos here: https://www.youtube.com/playlist?list=PLDKmox2v5e7tKNXeRBaLjCLIdv6d3X-82
Sensor Capasitive Construction
Sensor kapasitif merupakan sensor elektronika yang bekerja berdasarkan konsep kapasitif. Sensor ini bekerja berdasarkan perubahan muatan energi listrik yang dapat disimpan oleh sensor akibat perubahan jarak lempeng, perubahan luas penampang dan perubahan volume dielektrikan sensor kapasitif tersebut.
Sensor kapasitif terdiri dari dua komponen utama dua plat sebagai elektrode yaitu sensing elektrode dan referense elektrode.
My data are incomplete and noisy: Information-reduction statistical methods f...Umberto Picchini
We review parameter inference for stochastic modelling in complex scenario, such as bad parameters initialization and near-chaotic dynamics. We show how state-of-art methods for state-space models can fail while, in some situations, reducing data to summary statistics (information reduction) enables robust estimation. Wood's synthetic likelihoods method is reviewed and the lecture closes with an example of approximate Bayesian computation methodology.
Accompanying code is available at https://github.com/umbertopicchini/pomp-ricker and https://github.com/umbertopicchini/abc_g-and-k
Readership lecture given at Lund University on 7 June 2016. The lecture is of popular science nature hence mathematical detail is kept to a minimum. However numerous links and references are offered for further reading.
ELE2611 Classe 3 - Filtres analogiques linéaires IJerome LE NY
Approximations rationnelles classiques, dénormalisation de fonction de transfert.
Slides for the class 3 of the course ELE2611 (Circuits II) at Polytechnique Montreal, in French. Videos here: https://www.youtube.com/playlist?list=PLDKmox2v5e7tKNXeRBaLjCLIdv6d3X-82
Sensor Capasitive Construction
Sensor kapasitif merupakan sensor elektronika yang bekerja berdasarkan konsep kapasitif. Sensor ini bekerja berdasarkan perubahan muatan energi listrik yang dapat disimpan oleh sensor akibat perubahan jarak lempeng, perubahan luas penampang dan perubahan volume dielektrikan sensor kapasitif tersebut.
Sensor kapasitif terdiri dari dua komponen utama dua plat sebagai elektrode yaitu sensing elektrode dan referense elektrode.
My data are incomplete and noisy: Information-reduction statistical methods f...Umberto Picchini
We review parameter inference for stochastic modelling in complex scenario, such as bad parameters initialization and near-chaotic dynamics. We show how state-of-art methods for state-space models can fail while, in some situations, reducing data to summary statistics (information reduction) enables robust estimation. Wood's synthetic likelihoods method is reviewed and the lecture closes with an example of approximate Bayesian computation methodology.
Accompanying code is available at https://github.com/umbertopicchini/pomp-ricker and https://github.com/umbertopicchini/abc_g-and-k
Readership lecture given at Lund University on 7 June 2016. The lecture is of popular science nature hence mathematical detail is kept to a minimum. However numerous links and references are offered for further reading.
A likelihood-free version of the stochastic approximation EM algorithm (SAEM)...Umberto Picchini
I show how to obtain approximate maximum likelihood inference for "complex" models having some latent (unobservable) component. With "complex" I mean models having a so-called intractable likelihood, where the latter is unavailable in closed for or is too difficult to approximate. I construct a version of SAEM (and EM-type algorithm) that makes it possible to conduct inference for complex models. Traditionally SAEM is implementable only for models that are fairly tractable analytically. By introducing the concept of synthetic likelihood, where information is captured by a series of user-defined summary statistics (as in approximate Bayesian computation), it is possible to automatize SAEM to run on any model having some latent-component.
Approximate Bayesian computation and machine learning (BigMC 2014)Pierre Pudlo
The talk I was not able to give because strike actions by rail unions had been disrupting the rail traffic between Paris and Montpellier. But Chrisitian P. Robert replaced me.
Inference for stochastic differential equations via approximate Bayesian comp...Umberto Picchini
Despite the title the methods are appropriate for more general dynamical models (including state-space models). Presentation given at Nordstat 2012, Umeå. Relevant research paper at http://arxiv.org/abs/1204.5459 and software code at https://sourceforge.net/projects/abc-sde/
Approximate Bayesian computation for the Ising/Potts modelMatt Moores
Bayes’ formula involves the likelihood function, p(y|theta), which is a problem when the likelihood is unavailable in closed form. ABC is a method for approximating the posterior p(theta|y) without evaluating the likelihood. Instead, pseudo-data is simulated from a generative model and compared with the observations. This talk will give an introduction to ABC algorithms: rejection sampling, ABC-MCMC and ABC-SMC. Application of these algorithms to image analysis will be presented as an illustrative example. These methods have been implemented in the R package bayesImageS.
This is joint work with Christian Robert (Warwick/Dauphine), Kerrie Mengersen and Christopher Drovandi (QUT).
We apply tensor train (TT) data format to solve an elliptic PDE with uncertain coefficients. We reduce complexity and storage from exponential to linear. Post-processing in TT format is also provided.
Efficient Analysis of high-dimensional data in tensor formatsAlexander Litvinenko
We solve a PDE with uncertain coefficients. The solution is approximated in the Karhunen Loeve/PCE basis. How to compute maximum ? frequency? probability density function? with almost linear complexity? We offer various methods.
Stratified Monte Carlo and bootstrapping for approximate Bayesian computationUmberto Picchini
Presented on 7 May 2020 at "One World Approximate Bayesian Computation (ABC) Seminar". A video is available at https://youtu.be/IOPnRfAJ_W8
Approximate Bayesian computation (ABC) is computationally intensive for complex model simulators. To exploit expensive simulations, data-resampling via bootstrapping was used with success in [1] to obtain many artificial datasets at little cost and construct a synthetic likelihood. When using the same approach within ABC to produce a pseudo-marginal ABC-MCMC algorithm, the posterior variance is inflated, thus producing biased posterior inference. Here we use stratified Monte Carlo to considerably reduce the bias induced by data resampling. We also show that it is possible to obtain reliable inference using a larger than usual ABC threshold, by employing stratified Monte Carlo. Finally, we show that with stratified sampling we obtain a less variable ABC likelihood. In our paper [2] we consider simulation studies for static (Gaussian, g-and-k distribution, Ising model) and dynamic models (Lotka-Volterra). For the Lotka-Volterra case study, we compare our results against a standard pseudo-Marginal ABC and find that our approach is four times more efficient and, given limited computational budget, it explores the posterior surface more thoroughly. A comparison against state-of-art sequential Monte Carlo ABC is also reported.
References
[1] R. G. Everitt (2017). Bootstrapped synthetic likelihood. arXiv:1711.05825.
[2] U. Picchini, R.G. Everitt (2019). Stratified sampling and resampling for approximateBayesian computation. arXiv:1905.07976
Richard's entangled aventures in wonderlandRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Richard's aventures in two entangled wonderlandsRichard Gill
Since the loophole-free Bell experiments of 2020 and the Nobel prizes in physics of 2022, critics of Bell's work have retreated to the fortress of super-determinism. Now, super-determinism is a derogatory word - it just means "determinism". Palmer, Hance and Hossenfelder argue that quantum mechanics and determinism are not incompatible, using a sophisticated mathematical construction based on a subtle thinning of allowed states and measurements in quantum mechanics, such that what is left appears to make Bell's argument fail, without altering the empirical predictions of quantum mechanics. I think however that it is a smoke screen, and the slogan "lost in math" comes to my mind. I will discuss some other recent disproofs of Bell's theorem using the language of causality based on causal graphs. Causal thinking is also central to law and justice. I will mention surprising connections to my work on serial killer nurse cases, in particular the Dutch case of Lucia de Berk and the current UK case of Lucy Letby.
Multi-source connectivity as the driver of solar wind variability in the heli...Sérgio Sacani
The ambient solar wind that flls the heliosphere originates from multiple
sources in the solar corona and is highly structured. It is often described
as high-speed, relatively homogeneous, plasma streams from coronal
holes and slow-speed, highly variable, streams whose source regions are
under debate. A key goal of ESA/NASA’s Solar Orbiter mission is to identify
solar wind sources and understand what drives the complexity seen in the
heliosphere. By combining magnetic feld modelling and spectroscopic
techniques with high-resolution observations and measurements, we show
that the solar wind variability detected in situ by Solar Orbiter in March
2022 is driven by spatio-temporal changes in the magnetic connectivity to
multiple sources in the solar atmosphere. The magnetic feld footpoints
connected to the spacecraft moved from the boundaries of a coronal hole
to one active region (12961) and then across to another region (12957). This
is refected in the in situ measurements, which show the transition from fast
to highly Alfvénic then to slow solar wind that is disrupted by the arrival of
a coronal mass ejection. Our results describe solar wind variability at 0.5 au
but are applicable to near-Earth observatories.
A brief information about the SCOP protein database used in bioinformatics.
The Structural Classification of Proteins (SCOP) database is a comprehensive and authoritative resource for the structural and evolutionary relationships of proteins. It provides a detailed and curated classification of protein structures, grouping them into families, superfamilies, and folds based on their structural and sequence similarities.
Earliest Galaxies in the JADES Origins Field: Luminosity Function and Cosmic ...Sérgio Sacani
We characterize the earliest galaxy population in the JADES Origins Field (JOF), the deepest
imaging field observed with JWST. We make use of the ancillary Hubble optical images (5 filters
spanning 0.4−0.9µm) and novel JWST images with 14 filters spanning 0.8−5µm, including 7 mediumband filters, and reaching total exposure times of up to 46 hours per filter. We combine all our data
at > 2.3µm to construct an ultradeep image, reaching as deep as ≈ 31.4 AB mag in the stack and
30.3-31.0 AB mag (5σ, r = 0.1” circular aperture) in individual filters. We measure photometric
redshifts and use robust selection criteria to identify a sample of eight galaxy candidates at redshifts
z = 11.5 − 15. These objects show compact half-light radii of R1/2 ∼ 50 − 200pc, stellar masses of
M⋆ ∼ 107−108M⊙, and star-formation rates of SFR ∼ 0.1−1 M⊙ yr−1
. Our search finds no candidates
at 15 < z < 20, placing upper limits at these redshifts. We develop a forward modeling approach to
infer the properties of the evolving luminosity function without binning in redshift or luminosity that
marginalizes over the photometric redshift uncertainty of our candidate galaxies and incorporates the
impact of non-detections. We find a z = 12 luminosity function in good agreement with prior results,
and that the luminosity function normalization and UV luminosity density decline by a factor of ∼ 2.5
from z = 12 to z = 14. We discuss the possible implications of our results in the context of theoretical
models for evolution of the dark matter halo mass function.
Cancer cell metabolism: special Reference to Lactate PathwayAADYARAJPANDEY1
Normal Cell Metabolism:
Cellular respiration describes the series of steps that cells use to break down sugar and other chemicals to get the energy we need to function.
Energy is stored in the bonds of glucose and when glucose is broken down, much of that energy is released.
Cell utilize energy in the form of ATP.
The first step of respiration is called glycolysis. In a series of steps, glycolysis breaks glucose into two smaller molecules - a chemical called pyruvate. A small amount of ATP is formed during this process.
Most healthy cells continue the breakdown in a second process, called the Kreb's cycle. The Kreb's cycle allows cells to “burn” the pyruvates made in glycolysis to get more ATP.
The last step in the breakdown of glucose is called oxidative phosphorylation (Ox-Phos).
It takes place in specialized cell structures called mitochondria. This process produces a large amount of ATP. Importantly, cells need oxygen to complete oxidative phosphorylation.
If a cell completes only glycolysis, only 2 molecules of ATP are made per glucose. However, if the cell completes the entire respiration process (glycolysis - Kreb's - oxidative phosphorylation), about 36 molecules of ATP are created, giving it much more energy to use.
IN CANCER CELL:
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
Unlike healthy cells that "burn" the entire molecule of sugar to capture a large amount of energy as ATP, cancer cells are wasteful.
Cancer cells only partially break down sugar molecules. They overuse the first step of respiration, glycolysis. They frequently do not complete the second step, oxidative phosphorylation.
This results in only 2 molecules of ATP per each glucose molecule instead of the 36 or so ATPs healthy cells gain. As a result, cancer cells need to use a lot more sugar molecules to get enough energy to survive.
introduction to WARBERG PHENOMENA:
WARBURG EFFECT Usually, cancer cells are highly glycolytic (glucose addiction) and take up more glucose than do normal cells from outside.
Otto Heinrich Warburg (; 8 October 1883 – 1 August 1970) In 1931 was awarded the Nobel Prize in Physiology for his "discovery of the nature and mode of action of the respiratory enzyme.
WARNBURG EFFECT : cancer cells under aerobic (well-oxygenated) conditions to metabolize glucose to lactate (aerobic glycolysis) is known as the Warburg effect. Warburg made the observation that tumor slices consume glucose and secrete lactate at a higher rate than normal tissues.
The increased availability of biomedical data, particularly in the public domain, offers the opportunity to better understand human health and to develop effective therapeutics for a wide range of unmet medical needs. However, data scientists remain stymied by the fact that data remain hard to find and to productively reuse because data and their metadata i) are wholly inaccessible, ii) are in non-standard or incompatible representations, iii) do not conform to community standards, and iv) have unclear or highly restricted terms and conditions that preclude legitimate reuse. These limitations require a rethink on data can be made machine and AI-ready - the key motivation behind the FAIR Guiding Principles. Concurrently, while recent efforts have explored the use of deep learning to fuse disparate data into predictive models for a wide range of biomedical applications, these models often fail even when the correct answer is already known, and fail to explain individual predictions in terms that data scientists can appreciate. These limitations suggest that new methods to produce practical artificial intelligence are still needed.
In this talk, I will discuss our work in (1) building an integrative knowledge infrastructure to prepare FAIR and "AI-ready" data and services along with (2) neurosymbolic AI methods to improve the quality of predictions and to generate plausible explanations. Attention is given to standards, platforms, and methods to wrangle knowledge into simple, but effective semantic and latent representations, and to make these available into standards-compliant and discoverable interfaces that can be used in model building, validation, and explanation. Our work, and those of others in the field, creates a baseline for building trustworthy and easy to deploy AI models in biomedicine.
Bio
Dr. Michel Dumontier is the Distinguished Professor of Data Science at Maastricht University, founder and executive director of the Institute of Data Science, and co-founder of the FAIR (Findable, Accessible, Interoperable and Reusable) data principles. His research explores socio-technological approaches for responsible discovery science, which includes collaborative multi-modal knowledge graphs, privacy-preserving distributed data mining, and AI methods for drug discovery and personalized medicine. His work is supported through the Dutch National Research Agenda, the Netherlands Organisation for Scientific Research, Horizon Europe, the European Open Science Cloud, the US National Institutes of Health, and a Marie-Curie Innovative Training Network. He is the editor-in-chief for the journal Data Science and is internationally recognized for his contributions in bioinformatics, biomedical informatics, and semantic technologies including ontologies and linked data.
THE IMPORTANCE OF MARTIAN ATMOSPHERE SAMPLE RETURN.Sérgio Sacani
The return of a sample of near-surface atmosphere from Mars would facilitate answers to several first-order science questions surrounding the formation and evolution of the planet. One of the important aspects of terrestrial planet formation in general is the role that primary atmospheres played in influencing the chemistry and structure of the planets and their antecedents. Studies of the martian atmosphere can be used to investigate the role of a primary atmosphere in its history. Atmosphere samples would also inform our understanding of the near-surface chemistry of the planet, and ultimately the prospects for life. High-precision isotopic analyses of constituent gases are needed to address these questions, requiring that the analyses are made on returned samples rather than in situ.
ABC with data cloning for MLE in state space models
1. Maximum likelihood estimation of state-space
SDE models using data-cloning approximate
Bayesian computation
Umberto Picchini
Centre for Mathematical Sciences,
Lund University
AMS-EMS-SPM 2015, Porto
Umberto Picchini (umberto@maths.lth.se)
2. Nowadays there are several ways to deal with “intractable
likelihoods”, that is models for which an explicit likelihood function
is unavailable.
“Plug-and-play methods”: the only requirements is the ability to
simulate from the data-generating-model.
particle marginal methods (PMMH, PMCMC) based on SMC
filters [Andrieu et al. 2010].
Iterated filtering [Ionides et al. 2011]
approximate Bayesian computation (ABC) [Marin et al. 2012].
In the following I will focus on ABC methods.
Andrieu, Doucet and Holenstein 2010. Particle Markov chain Monte Carlo methods.
JRSS-B.
Ionides, Bhadra, Atchade and King 2011. Iterated filtering. Ann. Stat.
Marin, Pudlo, Robert and Ryder 2012. Approximate Bayesian computational methods.
Stat. Comput.
Umberto Picchini (umberto@maths.lth.se)
3. A state-space model (SSM)
Yt ∼ f(yt|Xt, φ), t t0
Xt ∼ g(xt|xt−1, η).
(1)
We have data y = (y0, y1, ..., yn) from (1) at discrete time-points
0 t0 < ... < tn.
Transition densities g(xt|xt−1, η) are typically unknown.
We are interested in inference for the vector parameter θ = (φ, η),
however the likelihood function is intractable
p(y|θ) =
T
t=1
p(yt|xt; θ)p(x1)
T
t=2
p(xt|xt−1; θ)
unavailable
dx1:T
Umberto Picchini (umberto@maths.lth.se)
4. Approximate Bayesian computation (ABC)
Consider the posterior distribution of θ:
π(θ|y) ∝ p(y|θ)π(θ)
Purpose of ABC is to obtain an approximation πδ(θ|y) to the true
posterior π(θ|y).
Here δ > 0 is a tolerance value. The smaller δ the better the
approximation to π(θ|y).
In practice inference is carried via some Monte Carlo sampling from
πδ(θ|y).
However for a “small” δ sampling from πδ(θ|y) can be difficult (high
rejection rates).
Umberto Picchini (umberto@maths.lth.se)
5. ABC gives a way to approximate a posterior distribution
π(θ|y) ∝ p(y|θ)π(θ)
key to the success of ABC is the ability to bypass the explicit
calculation of the likelihood p(y|θ)
...only forward-simulation from the model is required!
Simulate artificial-data y∗ from the SSM model (1):
y∗
∼ p(y|θ)
for SDEs, use numerical discretization (arbitrarily accurate as the
stepsize h → 0) or exact simulation (see
Beskos,Roberts,Fearnhead,Papaspiliopulos).
ABC had an incredible success in genetic studies since mid 90’s
(Tavare et al ’97, Pritchard et al. ’99). Now is everywhere.
Umberto Picchini (umberto@maths.lth.se)
6. ABC basics
Generate θ∗ ∼ π(θ), x∗
t ∼ p(X|θ∗), y∗ ∼ f(yt|x∗
t , θ∗).
proposal θ∗ is accepted if y∗ is “close” to data y, according to a
threshold δ > 0.
The above generate draws from the augmented approximated
posterior
πδ(θ, y∗
|y) ∝ Jδ(y, y∗
; θ) p(y∗
|θ)π(θ)
∝π(θ|y∗)
Jδ(·) weights the intractable posterior π(θ|y∗) ∝ p(y∗|θ)π(θ) with
high values when y∗ ≈ y.
Rationale: if Jδ(·) constant when δ = 0 (y = y∗) recover the exact
posterior π(θ|y).
Example: Jδ(y, y∗; θ) ∝ n
i=1
1
δe−
y∗
i −yi
2
2δ2
Umberto Picchini (umberto@maths.lth.se)
7. ABC within MCMC (Marjoram et al. 2003)
Data: y ∈ Y. Realizations y∗ from the SSM, y∗ ∈ Y.
Algorithm 1 a generic iteration of ABC-MCMC (fixed threshold δ)
At r-th iteration
1. generate θ∗ ∼ q(θ|θr), e.g. using Gaussian random walk
2. simulate x∗|θ∗ ∼ p(x|θ∗) and y∗ ∼ p(y|x∗, θ∗)
3. accept (θ∗, y∗) with probability
min 1, Jδ(y,y∗;θ∗)p(y∗|θ∗)π(θ∗)
Jδ(y,yr;θr)p(yr|θr)π(θr)
q(θr|θ∗)
q(θ∗|θr)
p(yr|θr)
p(y∗|θ∗)
then set r = r + 1 and go to 1.
Umberto Picchini (umberto@maths.lth.se)
8. ABC within MCMC (Marjoram et al. 2003)
Data: y ∈ Y. Realizations y∗ from the SSM, y∗ ∈ Y.
Algorithm 2 a generic iteration of ABC-MCMC (fixed threshold δ)
At r-th iteration
1. generate θ∗ ∼ q(θ|θr), e.g. using Gaussian random walk
2. simulate x∗|θ∗ ∼ p(x|θ∗) and y∗ ∼ p(y|x∗, θ∗)
3. accept (θ∗, y∗) with probability
min 1, Jδ(y,y∗;θ∗)p(y∗|θ∗)π(θ∗)
Jδ(y,yr;θr)p(yr|θr)π(θr)
q(θr|θ∗)
q(θ∗|θr)
p(yr|θr)
p(y∗|θ∗)
then set r = r + 1 and go to 1.
Samples are from πδ(θ|y)
or from the exact posterior when δ = 0.
Umberto Picchini (umberto@maths.lth.se)
9. ABC within MCMC (Marjoram et al. 2003)
Data: y ∈ Y. Realizations y∗ from the SSM, y∗ ∈ Y.
Algorithm 3 a generic iteration of ABC-MCMC (fixed threshold δ)
At r-th iteration
1. generate θ∗ ∼ q(θ|θr), e.g. using Gaussian random walk
2. simulate x∗|θ∗ ∼ p(x|θ∗) and y∗ ∼ p(y|x∗, θ∗)
3. accept (θ∗, y∗) with probability
min 1, Jδ(y,y∗;θ∗)p(y∗|θ∗)π(θ∗)
Jδ(y,yr;θr)p(yr|θr)π(θr)
q(θr|θ∗)
q(θ∗|θr)
p(yr|θr)
p(y∗|θ∗)
then set r = r + 1 and go to 1.
Samples are from πδ(θ|y)
or from the exact posterior when δ = 0.
Umberto Picchini (umberto@maths.lth.se)
10. a completely made-up illustration
green: the target posterior; prior distribution is uniform.
Let’s decrease δ progressively...
Umberto Picchini (umberto@maths.lth.se)
0 2 4 6 8 10
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
11. Typically we cannot reduce δ as much as we like.
When incurring into high rejection rates we might have to stop at the
pink approximation.
0 2 4 6 8 10
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
For the “best feasible δ” (pink) we get the MAP pretty much ok.
Tails are awful though...
Umberto Picchini (umberto@maths.lth.se)
12. Suppose we are in a scenario where it’s not feasible to decrease δ
further...What to do?
Here I am borrowing the data cloning idea.
data-cloning was independently introduced in:
1 Doucet, Godsill, Robert. Statistics and Computing (2002)
2 Jacquier, Johannes, Polson. J. Econometrics (2007)
3 popularized in ecology by Lele, Dennis, Lutscher. Ecology
Letters (2007).
Umberto Picchini (umberto@maths.lth.se)
13. Suppose we are in a scenario where it’s not feasible to decrease δ
further...What to do?
Here I am borrowing the data cloning idea.
data-cloning was independently introduced in:
1 Doucet, Godsill, Robert. Statistics and Computing (2002)
2 Jacquier, Johannes, Polson. J. Econometrics (2007)
3 popularized in ecology by Lele, Dennis, Lutscher. Ecology
Letters (2007).
Umberto Picchini (umberto@maths.lth.se)
14. “data cloning” for state-space models
(forget about ABC for the moment)
data: y
likelihood: L(θ; y)
choose an integer K 1 and stack K copies of your data
y(K)
= (y, y, ..., y)
K times
The corresponding posterior is
π(θ|y(K)
) ∝ (L(θ; y(K)
))π(θ)
Consider K independent realizations X(1)
, ..., X(K)
of {Xt}, with
X(k)
= (X
(k)
0 , ..., X
(k)
n ) , k = 1, ..., K
L(θ; y(K)
) =
K
k=1
f(y|X(k)
, θ)p(X(k)
|θ)dX(k)
= (L(θ; y))K
.
use MCMC to sample from π(θ|y(K)
) for “large” K.
Umberto Picchini (umberto@maths.lth.se)
15. Asymptotics, K → ∞ (Jacquier et al. 2007; Lele et al. 2007)
K is the # of data “clones”
when K → ∞ we have...
¯θ = sample mean of MCMC draws from π(θ|y(K)) ⇒ ˆθmle
(whatever the prior!)
K× [sample covariance of draws] from π(θ|y(K)) ⇒ I−1
ˆθmle
the
inverse of the Fisher information of the MLE.
¯θ ⇒ N ˆθmle, K−1 · I−1
ˆθmle
1 Jacquier, Johannes, Polson. J. Econometrics (2007)
2 Lele, Dennis, Lutscher. Ecology Letters (2007).
Umberto Picchini (umberto@maths.lth.se)
16. Our idea
Compensate for the inability to decrease δ by increasing K.
1 Run ABC-MCMC for decreasing δ (fix K = 1, no data-cloning);
2 Stop decreasing δ and start increasing K 1 (data-cloning).
3 distribution shrinks around the MLE (tick vertical line)
Umberto Picchini (umberto@maths.lth.se)
0 2 4 6 8 10
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
initial δ
18. lim
K→∞
lim
δ→0
πδ(θ|y(K)
) = N(ˆθmle, K−1
· I−1
ˆθmle
)
Now:
of course we can’t really let both δ → 0 and K → ∞
these two criteria compete! Computationally not feasible to
satisfy both.
I have no proof for the quality of the estimates for δ 0 and K
finite.
Umberto Picchini (umberto@maths.lth.se)
19. in Summary:
non-ABC (augmented) target posterior for a SSM:
π(θ, ˜X(K)
|y(K)
) ∝
K
k=1
f(y|X(k)
, θ)p(X(k)
|θ) π(θ)
here ˜X(K) = (X(1), ..., X(K)), each X(k) ∼ p(X|θ) i.i.d.
my ABC data-cloned posterior for a SSM:
πδ(θ, y∗(K)
|y(K)
) ∝
K
k=1
Jδ(y, y∗(k)
, θ)p(X(k)
|θ) π(θ)
as an example: Jδ(y, y∗(k)
; θ) := n
i=1
1
δe−
y∗(k)
i −yi
2
2δ2
Umberto Picchini (umberto@maths.lth.se)
20. Main problem with ABC: for complex models it is difficult to
obtain a decent acceptance rate during ABC-MCMC when δ
“small”.
Idea: set δ to a large (manageable) value, and compensate by
“powering up” the posterior → data-cloning. That is...
1 Preliminary step: use a typical ABC-MCMC with K = 1.
Determine the main mode ˜θ of πδ(θ|y) with δ “not-too-small”
(5% acceptance rate).
2 Start a further ABC-MCMC with K 1 by drawing proposal
using independence Metropolis centred at ˜θ.
3 Increase K progressively...
Umberto Picchini (umberto@maths.lth.se)
21. Algorithm 4 data-cloning ABC (P. 2015)
ABC-MCMC stage K = 1 using adaptive Metropolis random walk AMRW
1. Generate X∗
from p(X|θ∗
) and a corresponding y∗
from SSM. Compute
Jδ(y, y∗
; θ∗
).
2. Generate θ#
:= AMRW(θ∗
, Σ). Generate X#
’s from p(X|θ#
) and corresponding
y#
. Compute Jδ(y, y#
; θ#
).
3. Accept θ∗
with probability
α = min 1,
Jδ(y, y#
; θ#
)
Jδ(y, y∗; θ∗)
×
u1(θ∗
|θ#
, Σ)
u1(θ#|θ∗, Σ)
×
π(θ#
)
π(θ∗)
Data-cloning stage using a Metropolis independent sampler MIS
4. Fetch the maximum ˜θ from ABC-MCMC then do as above but proposing using
θ#
:= MIS(˜θ, ˆΣ).
5. Increase K := K + 1. Generate independently y#(1)
, ..., y#(K)
from p(y|θ#
)
6. Accept proposal with probability
α = min 1,
K
k=1 Jδ(y, y#(k)
; θ#
)
K
k=1 Jδ(y, y∗(k); θ∗)
×
u2(θ∗
|˜θ, ˆΣ)
u2(θ#|˜θ, ˆΣ)
×
π(θ#
)
π(θ∗)
.
Umberto Picchini (umberto@maths.lth.se)
22. Stochastic Gompertz model
dXt = BCe−Ct
Xtdt + σXtdWt, X0 = Ae−B
Used in ecology for population growth, e.g. chicken growth data [Donnet,
Foulley, Samson 2010]
0 5 10 15 20 25 30 35 40
0
1
2
3
4
5
6
7
8
9
12 observations from {log Xt}. X0 assumed known.
We wish to estimate θ = (A, B, C, σ)
Exact MLE available as transition densities are known.
Umberto Picchini (umberto@maths.lth.se)
25. Gompertz state-space model
Yti = log(Xti ) + εti εti ∼ N(0, σ2
ε)
dXt = BCe−CtXtdt + σXtdWt, X0 = Ae−B
12 observations from {Yti }. State {Xt} is unobserved. X0 assumed
known.
Wish to estimate θ = (A, B, C, σ, σε)
Umberto Picchini (umberto@maths.lth.se)
26. 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
0
1
2
3
4
5
6
7
8
9
t
Figure: data and three sample trajectories from the estimated state-space model.
True values ABC-DC ((K, δ) = (4, 0.8))
log A 8.01 8.01 (0.567)
log B(*) 1.609 1.611
log C 2.639 3.152 (0.982)
log σ 0 -0.080 (0.258)
log σ −0.799 -0.577 (0.176)
Umberto Picchini (umberto@maths.lth.se)
27. Take-home message
1 Sometimes we want to do MLE but we are unable to...
2 Sometimes we want to go full Bayesian but we can’t...
3 Sometimes even ABC is challenging...
4 There are endless possibilities out there (EP, VB and more...)
5 Working paper:
P. (2015) “Approximate maximum likelihood estimation using
data-cloning ABC‘”, arXiv:1505.06318.
6 blog discussion by Christian P. Robert (2 June)
https://xianblog.wordpress.com
Thank You
Umberto Picchini (umberto@maths.lth.se)
29. Appendix
“Likelihood free” Metropolis-Hastings
Suppose at a given iteration of Metropolis-Hastings we are in the
(augmented)-state position (θ#, x#) and wonder whether to move (or
not) to a new state (θ , x ). The move is generated via a proposal
distribution “q((θ#, x#) → (x , θ ))”.
e.g. “q((θ#, x#) → (x , θ ))” = u(θ |θ#)v(x | θ );
move “(θ#, x#) → (θ , x )” accepted with probability
α(θ#,x#)→(x ,θ ) = min 1,
π(θ )π(x |θ )π(y|x , θ )q((θ , x ) → (θ#, x#))
π(θ#)π(x#|θ#)π(y|x#, θ#)q((θ#, x#) → (θ , x ))
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )v(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)v(x | θ )
now choose v(x | θ) ≡ π(x | θ)
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )
π(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)
π(x | θ )
This is likelihood–free! And we only need to know how to generate xUmberto Picchini (umberto@maths.lth.se)
30. Appendix
“Likelihood free” Metropolis-Hastings
Suppose at a given iteration of Metropolis-Hastings we are in the
(augmented)-state position (θ#, x#) and wonder whether to move (or
not) to a new state (θ , x ). The move is generated via a proposal
distribution “q((θ#, x#) → (x , θ ))”.
e.g. “q((θ#, x#) → (x , θ ))” = u(θ |θ#)v(x | θ );
move “(θ#, x#) → (θ , x )” accepted with probability
α(θ#,x#)→(x ,θ ) = min 1,
π(θ )π(x |θ )π(y|x , θ )q((θ , x ) → (θ#, x#))
π(θ#)π(x#|θ#)π(y|x#, θ#)q((θ#, x#) → (θ , x ))
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )v(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)v(x | θ )
now choose v(x | θ) ≡ π(x | θ)
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )
π(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)
π(x | θ )
This is likelihood–free! And we only need to know how to generate xUmberto Picchini (umberto@maths.lth.se)
31. Appendix
“Likelihood free” Metropolis-Hastings
Suppose at a given iteration of Metropolis-Hastings we are in the
(augmented)-state position (θ#, x#) and wonder whether to move (or
not) to a new state (θ , x ). The move is generated via a proposal
distribution “q((θ#, x#) → (x , θ ))”.
e.g. “q((θ#, x#) → (x , θ ))” = u(θ |θ#)v(x | θ );
move “(θ#, x#) → (θ , x )” accepted with probability
α(θ#,x#)→(x ,θ ) = min 1,
π(θ )π(x |θ )π(y|x , θ )q((θ , x ) → (θ#, x#))
π(θ#)π(x#|θ#)π(y|x#, θ#)q((θ#, x#) → (θ , x ))
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )v(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)v(x | θ )
now choose v(x | θ) ≡ π(x | θ)
= min 1,
π(θ )π(x |θ )π(y|x , θ )u(θ#|θ )
π(x# | θ#)
π(θ#)π(x#|θ#)π(y|x#, θ#)u(θ |θ#)
π(x | θ )
This is likelihood–free! And we only need to know how to generate xUmberto Picchini (umberto@maths.lth.se)