This document presents a method for improving the scalability of approximate Bayesian computation (ABC) for latent graphical models like the hidden Potts model used in image analysis. It does this by pre-computing an auxiliary model that approximates the relationship between model parameters and summary statistics, avoiding the need to simulate pseudo-data during ABC model fitting. Experimental results on both simulated and satellite image data show the method reduces ABC runtime from weeks to hours while maintaining accuracy of parameter estimates.
Bayesian modelling and computation for Raman spectroscopyMatt Moores
Raman spectroscopy can be used to identify molecules by the characteristic scattering of light from a laser. Each Raman-active dye label has a unique spectral signature, comprised by the locations and amplitudes of the peaks. The Raman spectrum is discretised into a multivariate observation that is highly collinear, hence it lends itself to a reduced-rank representation. We introduce a sequential Monte Carlo (SMC) algorithm to separate this signal into a series of peaks plus a smoothly-varying baseline, corrupted by additive white noise. By incorporating this representation into a Bayesian functional regression, we can quantify the relationship between dye concentration and peak intensity. We also estimate the model evidence using SMC to investigate long-range dependence between peaks. These methods have been implemented as an R package, using RcppEigen and OpenMP.
Approximate Bayesian computation for the Ising/Potts modelMatt Moores
Bayes’ formula involves the likelihood function, p(y|theta), which is a problem when the likelihood is unavailable in closed form. ABC is a method for approximating the posterior p(theta|y) without evaluating the likelihood. Instead, pseudo-data is simulated from a generative model and compared with the observations. This talk will give an introduction to ABC algorithms: rejection sampling, ABC-MCMC and ABC-SMC. Application of these algorithms to image analysis will be presented as an illustrative example. These methods have been implemented in the R package bayesImageS.
This is joint work with Christian Robert (Warwick/Dauphine), Kerrie Mengersen and Christopher Drovandi (QUT).
Bayesian modelling and computation for Raman spectroscopyMatt Moores
Raman spectroscopy can be used to identify molecules by the characteristic scattering of light from a laser. Each Raman-active dye label has a unique spectral signature, comprised by the locations and amplitudes of the peaks. The Raman spectrum is discretised into a multivariate observation that is highly collinear, hence it lends itself to a reduced-rank representation. We introduce a sequential Monte Carlo (SMC) algorithm to separate this signal into a series of peaks plus a smoothly-varying baseline, corrupted by additive white noise. By incorporating this representation into a Bayesian functional regression, we can quantify the relationship between dye concentration and peak intensity. We also estimate the model evidence using SMC to investigate long-range dependence between peaks. These methods have been implemented as an R package, using RcppEigen and OpenMP.
Approximate Bayesian computation for the Ising/Potts modelMatt Moores
Bayes’ formula involves the likelihood function, p(y|theta), which is a problem when the likelihood is unavailable in closed form. ABC is a method for approximating the posterior p(theta|y) without evaluating the likelihood. Instead, pseudo-data is simulated from a generative model and compared with the observations. This talk will give an introduction to ABC algorithms: rejection sampling, ABC-MCMC and ABC-SMC. Application of these algorithms to image analysis will be presented as an illustrative example. These methods have been implemented in the R package bayesImageS.
This is joint work with Christian Robert (Warwick/Dauphine), Kerrie Mengersen and Christopher Drovandi (QUT).
Pre-computation for ABC in image analysisMatt Moores
MCMSki IV (the 5th IMS-ISBA joint meeting)
January 2014
Chamonix Mont-Blanc, France
The associated journal article has now been uploaded to arXiv: http://arxiv.org/abs/1403.4359
Slides: On the Chi Square and Higher-Order Chi Distances for Approximating f-...Frank Nielsen
Slides for the paper:
On the Chi Square and Higher-Order Chi Distances for Approximating f-Divergences
published in IEEE SPL:
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6654274
We are interested in finding a permutation of the entries of a given square matrix so that the maximum number of its nonzero entries are moved to one of the corners in a L-shaped fashion.
If we interpret the nonzero entries of the matrix as the edges of a graph, this problem boils down to the so-called core–periphery structure, consisting of two sets: the core, a set of nodes that is highly connected across the whole graph, and the periphery, a set of nodes that is well connected only to the nodes that are in the core.
Matrix reordering problems have applications in sparse factorizations and preconditioning, while revealing core–periphery structures in networks has applications in economic, social and communication networks.
Core–periphery detection in networks with nonlinear Perron eigenvectorsFrancesco Tudisco
Core–periphery detection is a highly relevant task in exploratory network analysis. Given a network of nodes and edges, one is interested in revealing the presence and measuring the consistency of a core–periphery structure using only the network topology. This mesoscale network structure consists of two sets: the core, a set of nodes that is highly connected across the whole network, and the periphery, a set of nodes that is well connected only to the nodes that are in the core. Networks with such a core–periphery structure have been observed in several applications, including economic, social, communication and citation networks.
In this talk we discuss a new core–periphery detection model based on the optimization of a class of core–periphery quality functions. While the quality measures are highly nonconvex in general and thus hardly treatable, we show that the global solution coincides with the nonlinear Perron eigenvector of a suitably defined parameter dependent matrix M(x), i.e. the positive solution to the nonlinear eigenvector problem M(x)x=λx. Using recent advances in nonlinear Perron–Frobeniustheory, we discuss uniqueness of the global solution and we propose a nonlinear power method-type scheme that (a) allows us to solve the optimization problem with global convergence guarantees and (b) effectively scales to very large and sparse networks. Finally, we present several numerical experiments showing that the new method largely out-performs state-of-the-art techniques for core-periphery detection.
Small updates of matrix functions used for network centralityFrancesco Tudisco
Many relevant measures of importance for nodes and edges of a network are defined in terms of suitable entries of matrix functions $f(A)$, for different choices of $f$ and $A$. Addressing the entries of $f(A)$ can be computationally challenging and this is particularly prohibitive when $A$ undergoes a perturbation $A+\delta A$ and the entries of $f(A)$ have to be updated. Given the adjacency matrix $A$ of a graph $G=(V,E)$, in this talk we consider the case where $\delta A$ is a sparse matrix that yields a small perturbation of the edge structure of $G$.
In particular, we present a bound showing that the variation of the entry $f(A)_{u,v}$ decays exponentially with the distance in $G$ that separates either $u$ or $v$ from the set of nodes touched by the edges that are perturbed. Our bound depends only on the distances in the original graph $G$ and on the field of values of the perturbed matrix $A+\delta A$. We show several numerical examples in support of the proposed result.
Talk presented at the IMA Numerical Analysis and Optimization conference, Birmingham 2018
The talk is based on the paper:
S. Pozza and F. Tudisco, On the stability of network indices defined by means of matrix functions, SIMAX, 2018
Optimal interval clustering: Application to Bregman clustering and statistica...Frank Nielsen
We present a generic dynamic programming method to compute the optimal clustering of n scalar elements into k pairwise disjoint intervals. This case includes 1D Euclidean k-means, k-medoids, k-medians, k-centers, etc. We extend the method to incorporate cluster size constraints and show how to choose the appropriate k by model selection. Finally, we illustrate and refine the method on two case studies: Bregman clustering and statistical mixture learning maximizing the complete likelihood.
http://arxiv.org/abs/1403.2485
After we applied the stochastic Galerkin method to solve stochastic PDE, and solve large linear system, we obtain stochastic solution (random field), which is represented in Karhunen Loeve and PCE basis. No sampling error is involved, only algebraic truncation error. Now we would like to escape classical MCMC path to compute the posterior. We develop an Bayesian* update formula for KLE-PCE coefficients.
Efficient Simulations for Contamination of Groundwater Aquifers under Uncerta...Alexander Litvinenko
1. Solved time-dependent density driven flow problem with uncertain porosity and permeability in 2D and 3D
2. Computed propagation of uncertainties in porosity into the mass fraction.
3. Computed the mean, variance, exceedance probabilities, quantiles, risks.
4. Such QoIs as the number of fingers, their size, shape, propagation time can be unstable
5. For moderate perturbations, our gPCE surrogate results are similar to qMC results.
6. Used highly scalable solver on up to 800 computing nodes,
Pre-computation for ABC in image analysisMatt Moores
MCMSki IV (the 5th IMS-ISBA joint meeting)
January 2014
Chamonix Mont-Blanc, France
The associated journal article has now been uploaded to arXiv: http://arxiv.org/abs/1403.4359
Slides: On the Chi Square and Higher-Order Chi Distances for Approximating f-...Frank Nielsen
Slides for the paper:
On the Chi Square and Higher-Order Chi Distances for Approximating f-Divergences
published in IEEE SPL:
http://ieeexplore.ieee.org/xpl/articleDetails.jsp?arnumber=6654274
We are interested in finding a permutation of the entries of a given square matrix so that the maximum number of its nonzero entries are moved to one of the corners in a L-shaped fashion.
If we interpret the nonzero entries of the matrix as the edges of a graph, this problem boils down to the so-called core–periphery structure, consisting of two sets: the core, a set of nodes that is highly connected across the whole graph, and the periphery, a set of nodes that is well connected only to the nodes that are in the core.
Matrix reordering problems have applications in sparse factorizations and preconditioning, while revealing core–periphery structures in networks has applications in economic, social and communication networks.
Core–periphery detection in networks with nonlinear Perron eigenvectorsFrancesco Tudisco
Core–periphery detection is a highly relevant task in exploratory network analysis. Given a network of nodes and edges, one is interested in revealing the presence and measuring the consistency of a core–periphery structure using only the network topology. This mesoscale network structure consists of two sets: the core, a set of nodes that is highly connected across the whole network, and the periphery, a set of nodes that is well connected only to the nodes that are in the core. Networks with such a core–periphery structure have been observed in several applications, including economic, social, communication and citation networks.
In this talk we discuss a new core–periphery detection model based on the optimization of a class of core–periphery quality functions. While the quality measures are highly nonconvex in general and thus hardly treatable, we show that the global solution coincides with the nonlinear Perron eigenvector of a suitably defined parameter dependent matrix M(x), i.e. the positive solution to the nonlinear eigenvector problem M(x)x=λx. Using recent advances in nonlinear Perron–Frobeniustheory, we discuss uniqueness of the global solution and we propose a nonlinear power method-type scheme that (a) allows us to solve the optimization problem with global convergence guarantees and (b) effectively scales to very large and sparse networks. Finally, we present several numerical experiments showing that the new method largely out-performs state-of-the-art techniques for core-periphery detection.
Small updates of matrix functions used for network centralityFrancesco Tudisco
Many relevant measures of importance for nodes and edges of a network are defined in terms of suitable entries of matrix functions $f(A)$, for different choices of $f$ and $A$. Addressing the entries of $f(A)$ can be computationally challenging and this is particularly prohibitive when $A$ undergoes a perturbation $A+\delta A$ and the entries of $f(A)$ have to be updated. Given the adjacency matrix $A$ of a graph $G=(V,E)$, in this talk we consider the case where $\delta A$ is a sparse matrix that yields a small perturbation of the edge structure of $G$.
In particular, we present a bound showing that the variation of the entry $f(A)_{u,v}$ decays exponentially with the distance in $G$ that separates either $u$ or $v$ from the set of nodes touched by the edges that are perturbed. Our bound depends only on the distances in the original graph $G$ and on the field of values of the perturbed matrix $A+\delta A$. We show several numerical examples in support of the proposed result.
Talk presented at the IMA Numerical Analysis and Optimization conference, Birmingham 2018
The talk is based on the paper:
S. Pozza and F. Tudisco, On the stability of network indices defined by means of matrix functions, SIMAX, 2018
Optimal interval clustering: Application to Bregman clustering and statistica...Frank Nielsen
We present a generic dynamic programming method to compute the optimal clustering of n scalar elements into k pairwise disjoint intervals. This case includes 1D Euclidean k-means, k-medoids, k-medians, k-centers, etc. We extend the method to incorporate cluster size constraints and show how to choose the appropriate k by model selection. Finally, we illustrate and refine the method on two case studies: Bregman clustering and statistical mixture learning maximizing the complete likelihood.
http://arxiv.org/abs/1403.2485
After we applied the stochastic Galerkin method to solve stochastic PDE, and solve large linear system, we obtain stochastic solution (random field), which is represented in Karhunen Loeve and PCE basis. No sampling error is involved, only algebraic truncation error. Now we would like to escape classical MCMC path to compute the posterior. We develop an Bayesian* update formula for KLE-PCE coefficients.
Efficient Simulations for Contamination of Groundwater Aquifers under Uncerta...Alexander Litvinenko
1. Solved time-dependent density driven flow problem with uncertain porosity and permeability in 2D and 3D
2. Computed propagation of uncertainties in porosity into the mass fraction.
3. Computed the mean, variance, exceedance probabilities, quantiles, risks.
4. Such QoIs as the number of fingers, their size, shape, propagation time can be unstable
5. For moderate perturbations, our gPCE surrogate results are similar to qMC results.
6. Used highly scalable solver on up to 800 computing nodes,
R package bayesImageS: Scalable Inference for Intractable LikelihoodsMatt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm and approximate Bayesian computation (ABC). A serious drawback of these algorithms is that they do not scale well for models with a large state space. Markov random fields, such as the Ising/Potts model and exponential random graph model (ERGM), are particularly challenging because the number of discrete variables increases linearly with the size of the image or graph. The likelihood of these models cannot be computed directly, due to the presence of an intractable normalising constant. In this context, it is necessary to employ algorithms that provide a suitable compromise between accuracy and computational cost.
Bayesian indirect likelihood (BIL) is a class of methods that approximate the likelihood function using a surrogate model. This model can be trained using a pre-computation step, utilising massively parallel hardware to simulate auxiliary variables. We review various types of surrogate model that can be used in BIL. In the case of the Potts model, we introduce a parametric approximation to the score function that incorporates its known properties, such as heteroskedasticity and critical temperature. We demonstrate this method on 2D satellite remote sensing and 3D computed tomography (CT) images. We achieve a hundredfold improvement in the elapsed runtime, compared to the exchange algorithm or ABC. Our algorithm has been implemented in the R package “bayesImageS,” which is available from CRAN.
Propagation of Uncertainties in Density Driven Groundwater FlowAlexander Litvinenko
Major Goal: estimate risks of the pollution in a subsurface flow.
How?: we solve density-driven groundwater flow with uncertain porosity and permeability.
We set up density-driven groundwater flow problem,
review stochastic modeling and stochastic methods, use UG4 framework (https://gcsc.uni-frankfurt.de/simulation-and-modelling/ug4),
model uncertainty in porosity and permeability,
2D and 3D numerical experiments.
bayesImageS: Bayesian computation for medical Image Segmentation using a hidd...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism.
In many countries, groundwater is the strategic reserve, which is used as drinking water and as an irrigation resource. Therefore, accurate modeling of the pollution of the soil and groundwater aquifer is highly important. As a model, we consider a density-driven groundwater flow problem with uncertain porosity and permeability. This problem may arise in geothermal reservoir simulation, natural saline-disposal basins, modeling of contaminant plumes and subsurface flow.
This strongly non-linear problem describes how salt or polluted water streams down building ''fingers". The solving process requires a very fine unstructured mesh and, therefore, high computational resources. Consequently, we run the parallel multigrid solver UG4 (https://github.com/UG4/ughub.wiki.git) on Shaheen II supercomputer.
The parallelization is done in both - the physical space and the stochastic space. The novelty of this work is the estimation of risks that the pollution will achieve a specific critical concentration. Additionally, we demonstrate how the multigrid UG4 solver can be run in a black-box fashion for testing different scenarios in the density-driven flow.
We solve Elder's problem in 2D and 3D domains, where unknown porosity and permeability are modeled by random fields. For approximations in the stochastic space, we use the generalized polynomial chaos expansion. We compute different quantities of interest such as the mean, variance and exceedance probabilities of the concentration. As a reference solution, we use the solution, obtained from the quasi-Monte Carlo method.
Hyperon and charm baryons masses from twisted mass Lattice QCDChristos Kallidonis
Talk given at the University of Bonn, Germany. We present results on the masses of all forty light, strange and charm baryons from Lattice QCD simulations. We elaborate on the various methods and techniques followed and examine systematic uncertainties related to isospin breaking effects and finite lattice spacing.
Major Goal: estimate risks of the pollution in a subsurface flow.
How? We solve density-driven groundwater flow with uncertain porosity and permeability.
1. We set up density-driven groundwater flow problem
2. Review stochastic modeling and stochastic methods
3. Modeling of uncertainty in porosity and permeability
4. Numerical methods to solve deterministic problem
5. 2D and 3D examples with 0.5-8 Millions mesh points.
Bayesian Inference and Uncertainty Quantification for Inverse ProblemsMatt Moores
So-called “inverse” problems arise when the parameters of a physical system cannot be directly observed. The mapping between these latent parameters and the space of noisy observations is represented as a mathematical model, often involving a system of differential equations. We seek to infer the parameter values that best fit our observed data. However, it is also vital to obtain accurate quantification of the uncertainty involved with these parameters, particularly when the output of the model will be used for forecasting. Bayesian inference provides well-calibrated uncertainty estimates, represented by the posterior distribution over the parameters. In this talk, I will give a brief introduction to Markov chain Monte Carlo (MCMC) algorithms for sampling from the posterior distribution and describe how they can be combined with numerical solvers for the forward model. We apply these methods to two examples of ODE models: growth curves in ecology, and thermogravimetric analysis (TGA) in chemistry. This is joint work with Matthew Berry, Mark Nelson, Brian Monaghan and Raymond Longbottom.
bayesImageS: an R package for Bayesian image analysisMatt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism. I will also discuss the process of releasing this software as an open source R package on the CRAN repository.
Accelerating Pseudo-Marginal MCMC using Gaussian ProcessesMatt Moores
The grouped independence Metropolis-Hastings (GIMH) and Markov chain within Metropolis (MCWM) algorithms are pseudo-marginal methods used to perform Bayesian inference in latent variable models. These methods replace intractable likelihood calculations with unbiased estimates within Markov chain Monte Carlo algorithms. The GIMH method has the posterior of interest as its limiting distribution, but suffers from poor mixing if it is too computationally intensive to obtain high-precision likelihood estimates. The MCWM algorithm has better mixing properties, but less theoretical support. In this paper we accelerate the GIMH method by using a Gaussian process (GP) approximation to the log-likelihood and train this GP using a short pilot run of the MCWM algorithm. Our new method, GP-GIMH, is illustrated on simulated data from a stochastic volatility and a gene network model. Our approach produces reasonable estimates of the univariate and bivariate posterior distributions, and the posterior correlation matrix in these examples with at least an order of magnitude improvement in computing time.
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...Matt Moores
There are many approaches to Bayesian computation with intractable likelihoods, including the exchange algorithm, approximate Bayesian computation (ABC), thermodynamic integration, and composite likelihood. These approaches vary in accuracy as well as scalability for datasets of significant size. The Potts model is an example where such methods are required, due to its intractable normalising constant. This model is a type of Markov random field, which is commonly used for image segmentation. The dimension of its parameter space increases linearly with the number of pixels in the image, making this a challenging application for scalable Bayesian computation. My talk will introduce various algorithms in the context of the Potts model and describe their implementation in C++, using OpenMP for parallelism. I will also discuss the process of releasing this software as an open source R package on the CRAN repository.
Informative Priors for Segmentation of Medical ImagesMatt Moores
There is an abundance of prior information available for image-guided radiotherapy, making it ideally suited for Bayesian techniques. I will demonstrate some results from applying the method of Teo, Sapiro & Wandell (1997) to cone-beam computed tomography (CT). A previous CT scan of the same object forms the prior expectation. The posterior probabilities of class membership are smoothed by diffusion, before labeling each pixel according to the maximum a posteriori (MAP) estimate. The effect of the prior and of the smoothing is discussed and some potential extensions to this method are proposed.
Chatty Kathy - UNC Bootcamp Final Project Presentation - Final Version - 5.23...John Andrews
SlideShare Description for "Chatty Kathy - UNC Bootcamp Final Project Presentation"
Title: Chatty Kathy: Enhancing Physical Activity Among Older Adults
Description:
Discover how Chatty Kathy, an innovative project developed at the UNC Bootcamp, aims to tackle the challenge of low physical activity among older adults. Our AI-driven solution uses peer interaction to boost and sustain exercise levels, significantly improving health outcomes. This presentation covers our problem statement, the rationale behind Chatty Kathy, synthetic data and persona creation, model performance metrics, a visual demonstration of the project, and potential future developments. Join us for an insightful Q&A session to explore the potential of this groundbreaking project.
Project Team: Jay Requarth, Jana Avery, John Andrews, Dr. Dick Davis II, Nee Buntoum, Nam Yeongjin & Mat Nicholas
Explore our comprehensive data analysis project presentation on predicting product ad campaign performance. Learn how data-driven insights can optimize your marketing strategies and enhance campaign effectiveness. Perfect for professionals and students looking to understand the power of data analysis in advertising. for more details visit: https://bostoninstituteofanalytics.org/data-science-and-artificial-intelligence/
Techniques to optimize the pagerank algorithm usually fall in two categories. One is to try reducing the work per iteration, and the other is to try reducing the number of iterations. These goals are often at odds with one another. Skipping computation on vertices which have already converged has the potential to save iteration time. Skipping in-identical vertices, with the same in-links, helps reduce duplicate computations and thus could help reduce iteration time. Road networks often have chains which can be short-circuited before pagerank computation to improve performance. Final ranks of chain nodes can be easily calculated. This could reduce both the iteration time, and the number of iterations. If a graph has no dangling nodes, pagerank of each strongly connected component can be computed in topological order. This could help reduce the iteration time, no. of iterations, and also enable multi-iteration concurrency in pagerank computation. The combination of all of the above methods is the STICD algorithm. [sticd] For dynamic graphs, unchanged components whose ranks are unaffected can be skipped altogether.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
StarCompliance is a leading firm specializing in the recovery of stolen cryptocurrency. Our comprehensive services are designed to assist individuals and organizations in navigating the complex process of fraud reporting, investigation, and fund recovery. We combine cutting-edge technology with expert legal support to provide a robust solution for victims of crypto theft.
Our Services Include:
Reporting to Tracking Authorities:
We immediately notify all relevant centralized exchanges (CEX), decentralized exchanges (DEX), and wallet providers about the stolen cryptocurrency. This ensures that the stolen assets are flagged as scam transactions, making it impossible for the thief to use them.
Assistance with Filing Police Reports:
We guide you through the process of filing a valid police report. Our support team provides detailed instructions on which police department to contact and helps you complete the necessary paperwork within the critical 72-hour window.
Launching the Refund Process:
Our team of experienced lawyers can initiate lawsuits on your behalf and represent you in various jurisdictions around the world. They work diligently to recover your stolen funds and ensure that justice is served.
At StarCompliance, we understand the urgency and stress involved in dealing with cryptocurrency theft. Our dedicated team works quickly and efficiently to provide you with the support and expertise needed to recover your assets. Trust us to be your partner in navigating the complexities of the crypto world and safeguarding your investments.
How can I successfully sell my pi coins in Philippines?DOT TECH
Even tho pi not launched globally, crypto whales, holders, investors are looking forward to hold up to 20,000 pi coins before mainnet launch in 2026.
All a miner or pioneer has to do to sell is to get in contact with a legitimate pi vendor ( a person that buys pi coins from miners and resell them to investors)
I will leave the telegram contact of my personal pi vendor:
@Pi_vendor_247
#pi network
#pi 2024
#sell pi
Precomputation for SMC-ABC with undirected graphical models
1. Background Pre-computation Experimental Results Conclusion
Pre-processing for SMC-ABC
with undirected graphical models
Matt Moores1,2 Chris Drovandi1 Kerrie Mengersen1,2
Antonietta Mira3 Christian Robert4,5
1Mathematical Sciences School, Queensland University of Technology,
Brisbane, Australia
2Institute for Health and Biomedical Innovation, QUT Kelvin Grove
3Institute of Finance, Università della Svizzera italiana, Switzerland
4CEREMADE, Université Paris Dauphine, France
5Department of Statistics, University of Warwick, UK
ABC in Sydney
July 4, 2014
3. Background Pre-computation Experimental Results Conclusion
Background
Image analysis often involves:
Large datasets, with millions of pixels
Multiple images with similar characteristics
For example: satellite remote sensing (Landsat), computed
tomography (CT)
Table: Scale of common types of images
Number Landsat CT slices
of pixels (90m2/px) (512×512)
26 0.06km2
. . .
56 14.06km2
0.1
106 900.00km2
3.8
156 10251.56km2
43.5
4. Background Pre-computation Experimental Results Conclusion
Motivation
Computational cost is dominated by simulation of pseudo-data
e.g. Hidden Potts model in image analysis
(Grelaud et al. 2009, Everitt 2012)
Model fitting with ABC can be separated into:
Learning about the summary statistic, given the parameter
f (s(w) | θ) π(θ)
Choosing parameter values, given a summary statistic
π (θ | δ(s(y), s(w)) )
For latent models, an additional step of learning about the
summary statistic, given the data: s(z) | y, θ
Grelaud, Robert, Marin, Rodolphe Taly (2009) Bayesian Analysis 4(2)
Everitt (2012) JCGS 21(4)
5. Background Pre-computation Experimental Results Conclusion
hidden Markov random field
Joint distribution of observed pixel intensities yi ∈ y
and latent labels zi ∈ z:
Pr(y, z|µ, σ2
, β) ∝ L(y|µ, σ2
, z)π(z|β) (1)
Additive Gaussian noise:
yi|zi =j
iid
∼ N µj, σ2
j
(2)
Potts model:
π(zi|zi∼`, β) =
exp {β
P
i∼` δ(zi, z`)}
Pk
j=1 exp {β
P
i∼` δ(j, z`)}
(3)
Potts (1952) Proceedings of the Cambridge Philosophical Society 48(1)
7. Background Pre-computation Experimental Results Conclusion
Doubly-intractable likelihood
p(β|z) = C(β)−1
π(β) exp {β S(z)} (4)
The normalising constant of the Potts model has computational
complexity of O(n2kn), since it involves a sum over all possible
combinations of the labels z ∈ Z:
C(β) =
X
z∈Z
exp {β S(z)} (5)
S(z) is the sufficient statistic of the Potts model:
S(z) =
X
i∼`∈L
δ(zi, z`) (6)
where L is the set of all unique neighbour pairs.
8. Background Pre-computation Experimental Results Conclusion
Expectation of S(z)
exact expectation of S(z) for n=12 and k=
β
E
(
S
(
z
))
5
10
15
1 2 3 4
2
3
4
(a) n = 12 k ∈ 2, 3, 4
exact expectation of S(z) for k=3 and n=
β
E
(
S
(
z
))
5
10
15
1 2 3 4
4
6
9
12
(b) k = 3 n ∈ 4, 6, 9, 12
Figure: Distribution of Ez|β[S(z)]
9. Background Pre-computation Experimental Results Conclusion
Standard deviation of S(z)
exact standard deviation of S(z) for n=12 and k=
β
σ
(
S
(
z
))
0.0
0.5
1.0
1.5
2.0
2.5
3.0
1 2 3 4
2
3
4
(a) n = 12 k ∈ 2, 3, 4
exact standard deviation of S(z) for k=3 and n=
β
σ
(
S
(
z
))
0.0
0.5
1.0
1.5
2.0
2.5
1 2 3 4
4
6
9
12
(b) k = 3 n ∈ 4, 6, 9, 12
Figure: Distribution of σz|β[S(z)]
10. Background Pre-computation Experimental Results Conclusion
Pre-computation
The distribution of the summary statistics f(s(w)|θ) is
independent of the observed data y
By simulating pseudo-data for values of θ, we can create a
binding function φ(θ) for an auxiliary model fA(s(w)|φ(θ))
This binding function can be reused across multiple datasets,
amortising its computational cost
By replacing s(w) with pseudo summary statistics drawn from our
auxiliary model, we avoid the need to simulate pseudo-data during
model fitting.
Wood (2010) Nature 466
Cabras, Castellanos Ruli (2014) Metron (to appear)
11. Background Pre-computation Experimental Results Conclusion
Simulation from f(·|β)
0.0 0.5 1.0 1.5 2.0 2.5 3.0
10
15
20
25
30
β
E
(
S
(
z
))
(a) Ez|β (S(w))
0.0 0.5 1.0 1.5 2.0 2.5 3.0
0
1
2
3
4
β
σ
(
S
(
z
))
(b) σz|β (S(w))
Figure: Approximation of S(w)|β using 1000 iterations of
Swendsen-Wang (discarding 500 as burn-in)
Swendsen Wang (1987) Physical Review Letters 58
12. Background Pre-computation Experimental Results Conclusion
Piecewise linear model
0.0 0.5 1.0 1.5 2.0 2.5 3.0
10000
15000
20000
25000
30000
β
E
S
(
z
)
(a) φ̂µ(β)
0.0 0.5 1.0 1.5 2.0 2.5 3.0
0
50
100
150
200
250
300
350
β
σ
S
(
z
)
(b) φ̂σ(β)
Figure: Binding functions for S(w) | β with n = 56
, k = 3
13. Background Pre-computation Experimental Results Conclusion
Approximations (A2
BC?)
This method introduces additional layers of approximation error:
finite sampling points for β
(even more of a problem in higher dimensions)
piecewise linear model for the binding functions φ̂µ(β) and
φ̂σ(β)
Gaussian kernel for Ŝ(w)|β ∼ N
φ̂µ(β), φ̂σ(β)2
But it enables us to scale up ABC for real world problems, for
which it was previously infeasible
14. Background Pre-computation Experimental Results Conclusion
Scalable SMC-ABC for the hidden Potts model
Algorithm 1 SMC-ABC using precomputed fA(s(w)|φ(θ))
1: Draw N particles β0
i ∼ π0(β)
2: Draw N × M statistics Ŝ(wi,m) ∼ N
φ̂µ(β0
i), φ̂σ(β0
i)2
3: repeat
4: Update S(zt)|y, πt(β)
5: Adaptively select ABC tolerance t
6: Update importance weights ωi for each particle
7: if effective sample size (ESS) Nmin then
8: Resample particles according to their weights
9: end if
10: Update particles using random walk proposal
(with adaptive RWMH bandwidth σ2
t )
11: until
naccept
N 0.015 or t 10−9 or t ≥ 100
15. Background Pre-computation Experimental Results Conclusion
Additive Gaussian noise
SMC iteration
S
(
z
)
0 20 40 60 80 100
15000
15500
16000
16500
17000
17500
18000
Figure: Change in the value of S(z) according to the current distribution
of πt(β|z) and L(y|µ, σ2
, z)
23. Background Pre-computation Experimental Results Conclusion
Results
Precomputation of fA(s(w)|φ(θ)) took 13h 23m
for 987 values of β
Model fitting took 1h 42m for 40 SMC iterations
In contrast:
Generating M = 50 pseudo-datasets per particle
takes 89 hours for a single iteration
(∴ 20 weeks for 40 iterations)
25. Background Pre-computation Experimental Results Conclusion
Summary
Scalability of SMC-ABC can be improved by pre-computing an
auxiliary model fA(s(w)|φ(θ))
Pre-computation took 1.4 hours on a 16 core Xeon server
for 987 values of β with 15,625 pixels
(13.4 hours for 978,380 pixels)
Average runtime for SMC-ABC improved from 71.4 hours to
39 minutes with 15,625 pixels
(1.7 hours for 978,380 pixels)
The binding functions represent the nonlinear, heteroskedastic
relationship between the parameter and the summary statistic.
This method could be extended to multivariate applications, such
as estimating both β and k for the hidden Potts model, or
estimating θ for an ERGM.
26. Appendix
For Further Reading I
Matt Moores, Chris Drovandi, Kerrie Mengersen Christian Robert
Pre-processing for approximate Bayesian computation in image analysis.
arXiv:1403.4359 [stat.CO], 2014.
Simon Wood
Statistical inference for noisy nonlinear ecological dynamic systems.
Nature, 466: 1102–04, 2010.
Stefano Cabras, Marı́a Eugenia Castellanos Erlis Ruli
A Quasi likelihood approximation of posterior distributions for
likelihood-intractable complex models.
To appear in Metron, 2014.
Christopher Drovandi, Anthony Pettitt Malcolm Faddy
Approximate Bayesian computation using indirect inference.
J. R. Stat. Soc. Ser. C 60(3): 317–37, 2011.
Richard Everitt
Bayesian Parameter Estimation for Latent Markov Random Fields and
Social Networks.
J. Comput. Graph. Stat., 21(4): 940–60, 2012.
27. Appendix
For Further Reading II
Christopher Drovandi, Anthony Pettitt Anthony Lee
Bayesian indirect inference using a parametric auxiliary model.
http://eprints.qut.edu.au/63767/3/63767.pdf
Pierre Del Moral, Arnaud Doucet Ajay Jasra
An adaptive sequential Monte Carlo method for approximate Bayesian
computation.
Statistics Computing, 22(5): 1009–20, 2012.
A. Grelaud, C. P. Robert, J.-M. Marin, F. Rodolphe J.-F. Taly
ABC likelihood-free methods for model choice in Gibbs random fields.
Bayesian Analysis, 4(2): 317–36, 2009.
Renfrey B. Potts
Some generalized order-disorder transformations.
Proc. Cambridge Philosophical Society, 48(1): 106–9, 1952.
R. H. Swendsen J.-S. Wang
Nonuniversal critical dynamics in Monte Carlo simulations.
Physical Review Letters, 58: 86–8, 1987.