The document proposes a new approach called successively quadratic interpolation for polynomial interpolation that is more efficient than Neville's and Aitken's algorithms. It involves iteratively computing quadratic interpolation polynomials using three points rather than linear interpolation using two points. The new algorithm reduces computational costs by about 20% compared to Neville's algorithm. Numerical experiments on test functions show the new algorithm has lower CPU time than Neville's algorithm while achieving the same solutions, demonstrating its improved efficiency.
We approach the screening problem - i.e. detecting which inputs of a computer model significantly impact the output - from a formal Bayesian model selection point of view. That is, we place a Gaussian process prior on the computer model and consider the $2^p$ models that result from assuming that each of the subsets of the $p$ inputs affect the response. The goal is to obtain the posterior probabilities of each of these models. In this talk, we focus on the specification of objective priors on the model-specific parameters and on convenient ways to compute the associated marginal likelihoods. These two problems that normally are seen as unrelated, have challenging connections since the priors proposed in the literature are specifically designed to have posterior modes in the boundary of the parameter space, hence precluding the application of approximate integration techniques based on e.g. Laplace approximations. We explore several ways of circumventing this difficulty, comparing different methodologies with synthetic examples taken from the literature.
Authors: Gonzalo Garcia-Donato (Universidad de Castilla-La Mancha) and Rui Paulo (Universidade de Lisboa)
The document discusses various techniques for classifying pictures using neural networks, including convolutional neural networks. It describes how convolutional neural networks can be used to classify images by breaking them into overlapping tiles, applying small neural networks to each tile, and pooling the results. The document also discusses using recurrent neural networks to classify videos by treating them as higher-dimensional tensors.
This document discusses various regularization techniques for deep learning models. It defines regularization as any modification to a learning algorithm intended to reduce generalization error without affecting training error. It then describes several specific regularization methods, including weight decay, norm penalties, dataset augmentation, early stopping, dropout, adversarial training, and tangent propagation. The goal of regularization is to reduce overfitting and improve generalizability of deep learning models.
This document discusses nonstationary covariance modeling. It begins by introducing concepts of covariance, correlation, and how correlation affects estimation and prediction. It then discusses properties of random fields like second-order stationarity and intrinsic stationarity. The difference between these two is explained. Common parametric models for isotropic covariance and variogram functions are presented, including spherical, exponential, Gaussian, rational quadratic, and Matérn models. Parameters and properties of these models like range, smoothness, and valid dimensionality are described. Examples of each model type are shown graphically.
The document discusses uncertainty quantification (UQ) using quasi-Monte Carlo (QMC) integration methods. It introduces parametric operator equations for modeling input uncertainty in partial differential equations. Both forward and inverse UQ problems are considered. QMC methods like interlaced polynomial lattice rules are discussed for approximating high-dimensional integrals arising in UQ, with convergence rates superior to standard Monte Carlo. Algorithms for single-level and multilevel QMC are presented for solving forward and inverse UQ problems.
The document proposes a new approach called successively quadratic interpolation for polynomial interpolation that is more efficient than Neville's and Aitken's algorithms. It involves iteratively computing quadratic interpolation polynomials using three points rather than linear interpolation using two points. The new algorithm reduces computational costs by about 20% compared to Neville's algorithm. Numerical experiments on test functions show the new algorithm has lower CPU time than Neville's algorithm while achieving the same solutions, demonstrating its improved efficiency.
We approach the screening problem - i.e. detecting which inputs of a computer model significantly impact the output - from a formal Bayesian model selection point of view. That is, we place a Gaussian process prior on the computer model and consider the $2^p$ models that result from assuming that each of the subsets of the $p$ inputs affect the response. The goal is to obtain the posterior probabilities of each of these models. In this talk, we focus on the specification of objective priors on the model-specific parameters and on convenient ways to compute the associated marginal likelihoods. These two problems that normally are seen as unrelated, have challenging connections since the priors proposed in the literature are specifically designed to have posterior modes in the boundary of the parameter space, hence precluding the application of approximate integration techniques based on e.g. Laplace approximations. We explore several ways of circumventing this difficulty, comparing different methodologies with synthetic examples taken from the literature.
Authors: Gonzalo Garcia-Donato (Universidad de Castilla-La Mancha) and Rui Paulo (Universidade de Lisboa)
The document discusses various techniques for classifying pictures using neural networks, including convolutional neural networks. It describes how convolutional neural networks can be used to classify images by breaking them into overlapping tiles, applying small neural networks to each tile, and pooling the results. The document also discusses using recurrent neural networks to classify videos by treating them as higher-dimensional tensors.
This document discusses various regularization techniques for deep learning models. It defines regularization as any modification to a learning algorithm intended to reduce generalization error without affecting training error. It then describes several specific regularization methods, including weight decay, norm penalties, dataset augmentation, early stopping, dropout, adversarial training, and tangent propagation. The goal of regularization is to reduce overfitting and improve generalizability of deep learning models.
This document discusses nonstationary covariance modeling. It begins by introducing concepts of covariance, correlation, and how correlation affects estimation and prediction. It then discusses properties of random fields like second-order stationarity and intrinsic stationarity. The difference between these two is explained. Common parametric models for isotropic covariance and variogram functions are presented, including spherical, exponential, Gaussian, rational quadratic, and Matérn models. Parameters and properties of these models like range, smoothness, and valid dimensionality are described. Examples of each model type are shown graphically.
The document discusses uncertainty quantification (UQ) using quasi-Monte Carlo (QMC) integration methods. It introduces parametric operator equations for modeling input uncertainty in partial differential equations. Both forward and inverse UQ problems are considered. QMC methods like interlaced polynomial lattice rules are discussed for approximating high-dimensional integrals arising in UQ, with convergence rates superior to standard Monte Carlo. Algorithms for single-level and multilevel QMC are presented for solving forward and inverse UQ problems.
This document discusses various methods for estimating normalizing constants that arise when evaluating integrals numerically. It begins by noting there are many computational methods for approximating normalizing constants across different communities. It then lists the topics that will be covered in the upcoming workshop, including discussions on estimating constants using Monte Carlo methods and Bayesian versus frequentist approaches. The document provides examples of estimating normalizing constants using Monte Carlo integration, reverse logistic regression, and Xiao-Li Meng's maximum likelihood estimation approach. It concludes by discussing some of the challenges in bringing a statistical framework to constant estimation problems.
This document provides an introduction to Approximate Bayesian Computation (ABC), a likelihood-free method for approximating posterior distributions when the likelihood function is unavailable or computationally intractable. It describes the ABC rejection sampling algorithm and key concepts like tolerance levels, distance functions, summary statistics, and improvements like ABC-MCMC and ABC-SMC. ABC is presented as an alternative to traditional Bayesian inference methods for models where direct likelihood evaluation is impossible or too expensive.
This document provides a summary of spatial data modeling and analysis techniques. It begins with an outline of the topics to be covered, including additive statistical models for spatial data, spatial covariance functions, the multivariate normal distribution, kriging for prediction and uncertainty, and the likelihood function for parameter estimation. It then introduces the key concepts and equations for modeling spatial processes as Gaussian random fields with specified covariance functions. Examples are given of commonly used covariance functions and the types of random surfaces they generate. Kriging is described as a best linear unbiased prediction technique that uses a spatial covariance function and observations to make predictions at unknown locations. The document concludes with examples of parameter estimation via maximum likelihood and using the fitted model to make predictions and conditional simulations
The document discusses methods for performing spatial statistics on large datasets. Standard maximum likelihood estimation is computationally infeasible for datasets with tens of thousands of observations due to the need to compute and store large covariance matrices. The document outlines several approximation methods that can accommodate large datasets, including variogram fitting, pairwise likelihood approximations, independent block approximations, tapering of the covariance function, low-rank approximations using basis functions, and approximations based on stochastic partial differential equations. These methods allow inference for large spatial datasets by avoiding direct computation and storage of large covariance matrices.
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
The document proposes a delayed acceptance method for accelerating Metropolis-Hastings algorithms. It begins with a motivating example of non-informative inference for mixture models where computing the prior density is costly. It then introduces the delayed acceptance approach which splits the acceptance probability into pieces that are evaluated sequentially, avoiding computing the full acceptance ratio each time. It validates that the delayed acceptance chain is reversible and provides bounds on its spectral gap and asymptotic variance compared to the original chain. Finally, it discusses optimizing the delayed acceptance approach by considering the expected square jump distance and cost per iteration to maximize efficiency.
The document describes a new method called component-wise approximate Bayesian computation (ABC) that combines ABC with Gibbs sampling. It aims to improve ABC's ability to efficiently explore parameter spaces when the number of parameters is large. The method works by alternating sampling from each parameter's ABC posterior conditional distribution given current values of other parameters and the observed data. The method is proven to converge to a stationary distribution under certain assumptions, especially for hierarchical models where conditional distributions are often simplified. Numerical experiments on toy examples demonstrate the method can provide a better approximation of the true posterior than vanilla ABC.
In this talk, we discuss some recent advances in probabilistic schemes for high-dimensional PIDEs. It is known that traditional PDE solvers, e.g., finite element, finite difference methods, do not scale well with the increase of dimension. The idea of probabilistic schemes is to link a wide class of nonlinear parabolic PIDEs to stochastic Levy processes based on nonlinear version of the Feynman-Kac theory. As such, the solution of the PIDE can be represented by a conditional expectation (i.e., a high-dimensional integral) with respect to a stochastic dynamical system driven by Levy processes. In other words, we can solve the PIDEs by performing high-dimensional numerical integration. A variety of quadrature methods could be applied, including MC, QMC, sparse grids, etc. The probabilistic schemes have been used in many application problems, e.g., particle transport in plasmas (e.g., Vlasov-Fokker-Planck equations), nonlinear filtering (e.g., Zakai equations), and option pricing, etc.
This document provides an outline and overview of key concepts for estimating curves and surfaces from data using basis functions and penalized least squares regression. It discusses representing a curve or surface using basis functions, fitting the coefficients using ordinary least squares, and adding a penalty term to the least squares objective function to produce a smoothed estimate. The smoothing parameter λ controls the tradeoff between fit to the data and smoothness of the estimate. Cross-validation can be used to choose λ.
comments on exponential ergodicity of the bouncy particle samplerChristian Robert
The document summarizes recent work on establishing theoretical convergence rates for the bouncy particle sampler (BPS), a non-reversible Markov chain Monte Carlo algorithm. The main results show that under certain conditions on the target distribution, including having exponentially decaying tails, the BPS exhibits exponential ergodicity. A central limit theorem is also established. The analysis considers different cases for thin-tailed, thick-tailed, and transformed target distributions.
1) The document discusses detection and attribution in climate science, which refers to statistical techniques used to identify the contributions of different forcing factors (like greenhouse gases or solar activity) to changes in climate signals over time.
2) It provides context on the history and development of detection and attribution methods, beginning with early work in the 1970s-1990s and more recent Bayesian approaches.
3) A key paper discussed is one by Katzfuss, Hammerling and Smith (2017) that introduced a Bayesian hierarchical model for climate change detection and attribution to help address uncertainties.
The document discusses measuring sample quality using kernels. It introduces the kernel Stein discrepancy (KSD) as a new quality measure for comparing samples approximating a target distribution. The KSD is based on Stein's method and uses reproducing kernels. It can detect when a sample sequence is converging to the target distribution or not. Computing the KSD reduces to pairwise evaluations of kernel functions and is feasible. The KSD converges to zero if and only if the sample sequence converges to the target distribution for certain choices of kernels like the inverse multiquadric kernel with parameter between -1 and 0.
The document discusses achieving higher-order convergence for integration on RN using quasi-Monte Carlo (QMC) rules. It describes the problem that when using tensor product QMC rules on truncated domains, the convergence rate scales with the dimension s as (α log N)sN-α. The goal is to obtain a convergence rate independent of the dimension s. The document proposes using a multivariate decomposition method (MDM) to decompose an infinite-dimensional integral into a sum of finite-dimensional integrals, then applying QMC rules to each integral to achieve the desired higher-order convergence rate.
This document summarizes a talk given by Heiko Strathmann on using partial posterior paths to estimate expectations from large datasets without full posterior simulation. The key ideas are:
1. Construct a path of "partial posteriors" by sequentially adding mini-batches of data and computing expectations over these posteriors.
2. "Debias" the path of expectations to obtain an unbiased estimator of the true posterior expectation using a technique from stochastic optimization literature.
3. This approach allows estimating posterior expectations with sub-linear computational cost in the number of data points, without requiring full posterior simulation or imposing restrictions on the likelihood.
Experiments on synthetic and real-world examples demonstrate competitive performance versus standard M
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
This document discusses an empirical Bayesian approach for estimating regularization parameters in inverse problems using maximum likelihood estimation. It proposes the Stochastic Optimization with Unadjusted Langevin (SOUL) algorithm, which uses Markov chain sampling to approximate gradients in a stochastic projected gradient descent scheme for optimizing the regularization parameter. The algorithm is shown to converge to the maximum likelihood estimate under certain conditions on the log-likelihood and prior distributions.
Likelihood approximation with parallel hierarchical matrices for large spatia...Alexander Litvinenko
First, we use hierarchical matrices to approximate large Matern covariance matrices and the loglikelihood. Second, we find a maximum of loglikelihood and estimate 3 unknown parameters (covariance length, smoothness and variance).
Alexander Litvinenko's research interests include developing efficient numerical methods for solving stochastic PDEs using low-rank tensor approximations. He has made contributions in areas such as fast techniques for solving stochastic PDEs using tensor approximations, inexpensive functional approximations of Bayesian updating formulas, and modeling uncertainties in parameters, coefficients, and computational geometry using probabilistic methods. His current research focuses on uncertainty quantification, Bayesian updating techniques, and developing scalable and parallel methods using hierarchical matrices.
This document discusses various methods for estimating normalizing constants that arise when evaluating integrals numerically. It begins by noting there are many computational methods for approximating normalizing constants across different communities. It then lists the topics that will be covered in the upcoming workshop, including discussions on estimating constants using Monte Carlo methods and Bayesian versus frequentist approaches. The document provides examples of estimating normalizing constants using Monte Carlo integration, reverse logistic regression, and Xiao-Li Meng's maximum likelihood estimation approach. It concludes by discussing some of the challenges in bringing a statistical framework to constant estimation problems.
This document provides an introduction to Approximate Bayesian Computation (ABC), a likelihood-free method for approximating posterior distributions when the likelihood function is unavailable or computationally intractable. It describes the ABC rejection sampling algorithm and key concepts like tolerance levels, distance functions, summary statistics, and improvements like ABC-MCMC and ABC-SMC. ABC is presented as an alternative to traditional Bayesian inference methods for models where direct likelihood evaluation is impossible or too expensive.
This document provides a summary of spatial data modeling and analysis techniques. It begins with an outline of the topics to be covered, including additive statistical models for spatial data, spatial covariance functions, the multivariate normal distribution, kriging for prediction and uncertainty, and the likelihood function for parameter estimation. It then introduces the key concepts and equations for modeling spatial processes as Gaussian random fields with specified covariance functions. Examples are given of commonly used covariance functions and the types of random surfaces they generate. Kriging is described as a best linear unbiased prediction technique that uses a spatial covariance function and observations to make predictions at unknown locations. The document concludes with examples of parameter estimation via maximum likelihood and using the fitted model to make predictions and conditional simulations
The document discusses methods for performing spatial statistics on large datasets. Standard maximum likelihood estimation is computationally infeasible for datasets with tens of thousands of observations due to the need to compute and store large covariance matrices. The document outlines several approximation methods that can accommodate large datasets, including variogram fitting, pairwise likelihood approximations, independent block approximations, tapering of the covariance function, low-rank approximations using basis functions, and approximations based on stochastic partial differential equations. These methods allow inference for large spatial datasets by avoiding direct computation and storage of large covariance matrices.
Delayed acceptance for Metropolis-Hastings algorithmsChristian Robert
The document proposes a delayed acceptance method for accelerating Metropolis-Hastings algorithms. It begins with a motivating example of non-informative inference for mixture models where computing the prior density is costly. It then introduces the delayed acceptance approach which splits the acceptance probability into pieces that are evaluated sequentially, avoiding computing the full acceptance ratio each time. It validates that the delayed acceptance chain is reversible and provides bounds on its spectral gap and asymptotic variance compared to the original chain. Finally, it discusses optimizing the delayed acceptance approach by considering the expected square jump distance and cost per iteration to maximize efficiency.
The document describes a new method called component-wise approximate Bayesian computation (ABC) that combines ABC with Gibbs sampling. It aims to improve ABC's ability to efficiently explore parameter spaces when the number of parameters is large. The method works by alternating sampling from each parameter's ABC posterior conditional distribution given current values of other parameters and the observed data. The method is proven to converge to a stationary distribution under certain assumptions, especially for hierarchical models where conditional distributions are often simplified. Numerical experiments on toy examples demonstrate the method can provide a better approximation of the true posterior than vanilla ABC.
In this talk, we discuss some recent advances in probabilistic schemes for high-dimensional PIDEs. It is known that traditional PDE solvers, e.g., finite element, finite difference methods, do not scale well with the increase of dimension. The idea of probabilistic schemes is to link a wide class of nonlinear parabolic PIDEs to stochastic Levy processes based on nonlinear version of the Feynman-Kac theory. As such, the solution of the PIDE can be represented by a conditional expectation (i.e., a high-dimensional integral) with respect to a stochastic dynamical system driven by Levy processes. In other words, we can solve the PIDEs by performing high-dimensional numerical integration. A variety of quadrature methods could be applied, including MC, QMC, sparse grids, etc. The probabilistic schemes have been used in many application problems, e.g., particle transport in plasmas (e.g., Vlasov-Fokker-Planck equations), nonlinear filtering (e.g., Zakai equations), and option pricing, etc.
This document provides an outline and overview of key concepts for estimating curves and surfaces from data using basis functions and penalized least squares regression. It discusses representing a curve or surface using basis functions, fitting the coefficients using ordinary least squares, and adding a penalty term to the least squares objective function to produce a smoothed estimate. The smoothing parameter λ controls the tradeoff between fit to the data and smoothness of the estimate. Cross-validation can be used to choose λ.
comments on exponential ergodicity of the bouncy particle samplerChristian Robert
The document summarizes recent work on establishing theoretical convergence rates for the bouncy particle sampler (BPS), a non-reversible Markov chain Monte Carlo algorithm. The main results show that under certain conditions on the target distribution, including having exponentially decaying tails, the BPS exhibits exponential ergodicity. A central limit theorem is also established. The analysis considers different cases for thin-tailed, thick-tailed, and transformed target distributions.
1) The document discusses detection and attribution in climate science, which refers to statistical techniques used to identify the contributions of different forcing factors (like greenhouse gases or solar activity) to changes in climate signals over time.
2) It provides context on the history and development of detection and attribution methods, beginning with early work in the 1970s-1990s and more recent Bayesian approaches.
3) A key paper discussed is one by Katzfuss, Hammerling and Smith (2017) that introduced a Bayesian hierarchical model for climate change detection and attribution to help address uncertainties.
The document discusses measuring sample quality using kernels. It introduces the kernel Stein discrepancy (KSD) as a new quality measure for comparing samples approximating a target distribution. The KSD is based on Stein's method and uses reproducing kernels. It can detect when a sample sequence is converging to the target distribution or not. Computing the KSD reduces to pairwise evaluations of kernel functions and is feasible. The KSD converges to zero if and only if the sample sequence converges to the target distribution for certain choices of kernels like the inverse multiquadric kernel with parameter between -1 and 0.
The document discusses achieving higher-order convergence for integration on RN using quasi-Monte Carlo (QMC) rules. It describes the problem that when using tensor product QMC rules on truncated domains, the convergence rate scales with the dimension s as (α log N)sN-α. The goal is to obtain a convergence rate independent of the dimension s. The document proposes using a multivariate decomposition method (MDM) to decompose an infinite-dimensional integral into a sum of finite-dimensional integrals, then applying QMC rules to each integral to achieve the desired higher-order convergence rate.
This document summarizes a talk given by Heiko Strathmann on using partial posterior paths to estimate expectations from large datasets without full posterior simulation. The key ideas are:
1. Construct a path of "partial posteriors" by sequentially adding mini-batches of data and computing expectations over these posteriors.
2. "Debias" the path of expectations to obtain an unbiased estimator of the true posterior expectation using a technique from stochastic optimization literature.
3. This approach allows estimating posterior expectations with sub-linear computational cost in the number of data points, without requiring full posterior simulation or imposing restrictions on the likelihood.
Experiments on synthetic and real-world examples demonstrate competitive performance versus standard M
Maximum likelihood estimation of regularisation parameters in inverse problem...Valentin De Bortoli
This document discusses an empirical Bayesian approach for estimating regularization parameters in inverse problems using maximum likelihood estimation. It proposes the Stochastic Optimization with Unadjusted Langevin (SOUL) algorithm, which uses Markov chain sampling to approximate gradients in a stochastic projected gradient descent scheme for optimizing the regularization parameter. The algorithm is shown to converge to the maximum likelihood estimate under certain conditions on the log-likelihood and prior distributions.
Likelihood approximation with parallel hierarchical matrices for large spatia...Alexander Litvinenko
First, we use hierarchical matrices to approximate large Matern covariance matrices and the loglikelihood. Second, we find a maximum of loglikelihood and estimate 3 unknown parameters (covariance length, smoothness and variance).
Alexander Litvinenko's research interests include developing efficient numerical methods for solving stochastic PDEs using low-rank tensor approximations. He has made contributions in areas such as fast techniques for solving stochastic PDEs using tensor approximations, inexpensive functional approximations of Bayesian updating formulas, and modeling uncertainties in parameters, coefficients, and computational geometry using probabilistic methods. His current research focuses on uncertainty quantification, Bayesian updating techniques, and developing scalable and parallel methods using hierarchical matrices.
My PhD talk "Application of H-matrices for computing partial inverse"Alexander Litvinenko
This document describes a hierarchical domain decomposition (HDD) method for solving stochastic elliptic boundary value problems with oscillatory or jumping coefficients. HDD constructs mappings between boundary and interface values that allow the solution to be computed locally in each subdomain. These mappings are represented as H-matrices to reduce computational costs. The total storage cost of HDD is O(kn log2nh) and complexity is O(k2nh log3nh), where n is the number of degrees of freedom, k is the H-matrix rank, and h is the mesh size. HDD can also be used to compute solutions when the right-hand side is represented on a coarser grid.
The document summarizes a dissertation on applying hierarchical matrices to solve multiscale problems. The dissertation proposes a new hierarchical domain decomposition (HDD) method that combines hierarchical matrices and domain decomposition. HDD allows efficiently computing solution mappings and functionals, and solving problems on coarser grids or with multiple right-hand sides. Complexity analyses show HDD has lower complexity than other methods. Numerical tests on problems with oscillatory and jumping coefficients demonstrate HDD achieves the expected error bounds and is independent of frequency.
We combined: low-rank tensor techniques and FFT to compute kriging, estimate variance, compute conditional covariance. We are able to solve 3D problems with very high resolution
Low-rank tensor methods for stochastic forward and inverse problemsAlexander Litvinenko
The document discusses low-rank tensor methods for solving partial differential equations (PDEs) with uncertain coefficients. It covers two parts: (1) using the stochastic Galerkin method to discretize an elliptic PDE with uncertain diffusion coefficient represented by tensors, and (2) computing quantities of interest like the maximum value from the tensor solution in a efficient way. Specifically, it describes representing the diffusion coefficient, forcing term, and solution of the discretized PDE using tensors, and computing the maximum value and corresponding indices by solving an eigenvalue problem involving the tensor solution.
Possible applications of low-rank tensors in statistics and UQ (my talk in Bo...Alexander Litvinenko
Just some ideas how low-rank matrices/tensors can be useful in spatial and environmental statistics, where one usually has to deal with very large data
This document is a dissertation submitted by Alexander Litvinenko to the Faculty of Mathematics and Computer Science at the University of Leipzig in partial fulfillment of the requirements for the degree of Doctor of Natural Sciences. The dissertation proposes the application of hierarchical matrices (H-matrices) to solve multiscale problems using the hierarchical domain decomposition (HDD) method. It begins with an introduction and literature review of multiscale problems and existing solution methods. It then describes the classical finite element method, the HDD method, and H-matrices. The main body of the dissertation focuses on applying H-matrices within the HDD method to efficiently solve problems involving multiple spatial and temporal scales. Numerical results demonstrate the effectiveness of the proposed approach.
Response Surface in Tensor Train format for Uncertainty QuantificationAlexander Litvinenko
We apply low-rank Tensor Train format to solve PDEs with uncertain coefficients. First, we approximate uncertain permeability coefficient in TT format, then the operator and then apply iterations to solve stochastic Galerkin system.
My paper for Domain Decomposition Conference in Strobl, Austria, 2005Alexander Litvinenko
We did a first step in solving, so-called, skin problem. We developed an efficient H-matrix preconditioner to solve diffusion problem with jumping coefficients
Application H-matrices for solving PDEs with multi-scale coefficients, jumpin...Alexander Litvinenko
We develop hierarchical domain decomposition method to compute a part of the solution, a part of the inverse operator with O(n log n) storage and computing cost.
Hierarchical matrix approximation of large covariance matricesAlexander Litvinenko
We research class of Matern covariance matrices and their approximability in the H-matrix format. Further tasks are compute H-Cholesky factorization, trace, determinant, quadratic form, loglikelihood. Later H-matrices can be applied in kriging.
Low-rank methods for analysis of high-dimensional data (SIAM CSE talk 2017) Alexander Litvinenko
Overview of our latest works in applying low-rank tensor techniques to a) solving PDEs with uncertain coefficients (or multi-parametric PDEs) b) postprocessing high-dimensional data c) compute the largest element, level sets, TOP5% elelments
We research how we can use Scalable hierarchical algorithms for solving stochastic PDEs and for Uncertainty Quantification. Particularly we are interested in approximating large covariance matrices in H-matrix format, Hierarchical Cholesky factorization and computing Karhunen-Loeve expansion
Minimum mean square error estimation and approximation of the Bayesian updateAlexander Litvinenko
This document discusses methods for approximating the Bayesian update used in parameter identification problems with partial differential equations containing uncertain coefficients. It presents:
1) Deriving the Bayesian update from conditional expectation and proposing polynomial chaos expansions to approximate the full Bayesian update.
2) Describing minimum mean square error estimation to find estimators that minimize the error between the true parameter and its estimate given measurements.
3) Providing an example of applying these methods to identify an uncertain coefficient in a 1D elliptic PDE using measurements at two points.
Computation of Electromagnetic Fields Scattered from Dielectric Objects of Un...Alexander Litvinenko
1) The document describes a method called Multilevel Monte Carlo (MLMC) to efficiently compute electromagnetic fields scattered from dielectric objects of uncertain shapes. MLMC balances statistical errors from random sampling and numerical errors from geometry discretization to reduce computational time.
2) A surface integral equation solver is used to model scattering from dielectric objects. Random geometries are generated by perturbing surfaces with random fields defined by spherical harmonics.
3) MLMC is shown to estimate scattering cross sections accurately while requiring fewer overall computations compared to traditional Monte Carlo methods. This is achieved by optimally allocating samples across discretization levels.
After we applied the stochastic Galerkin method to solve stochastic PDE, and solve large linear system, we obtain stochastic solution (random field), which is represented in Karhunen Loeve and PCE basis. No sampling error is involved, only algebraic truncation error. Now we would like to escape classical MCMC path to compute the posterior. We develop an Bayesian* update formula for KLE-PCE coefficients.
This document discusses using bootstrap methods to create confidence intervals for time series forecasts. It provides examples of time series data and introduces the AR(1) model. The document describes an algorithm for calculating a bootstrap confidence interval for forecasting from an AR(1) model. It then discusses a simulation study comparing empirical coverage rates of bootstrap confidence intervals under different parameters. Finally, it applies the bootstrap method to forecasting Gross National Product growth, comparing the results to a parametric approach.
This document discusses using bootstrap methods to create confidence intervals for time series forecasts. It provides background on time series models and the autoregressive (AR) process. It then presents an algorithm for calculating a bootstrap confidence interval for forecasts from an AR(1) model. A simulation study compares coverage rates for bootstrap confidence intervals under different parameters. Finally, the method is applied to US Gross National Product data to forecast and construct confidence intervals.
More on randomization semi-definite programming and derandomizationAbner Chih Yi Huang
This document summarizes a presentation on derandomization techniques and semidefinite programming. It begins with an overview of derandomization using the method of conditional probabilities and a weighted MAXSAT algorithm example. It then discusses semidefinite programming, how it can solve certain problems more tightly than linear programming, and how it enables improved approximation algorithms, such as a 0.878 approximation for MAXCUT using a Goemans-Williamson random hyperplane rounding technique.
This document summarizes a research paper about using hierarchical deterministic quadrature methods for option pricing under the rough Bergomi model. It discusses the rough Bergomi model and challenges in pricing options under this model numerically. It then describes the methodology used, which involves analytic smoothing, adaptive sparse grids quadrature, quasi Monte Carlo, and coupling these with hierarchical representations and Richardson extrapolation. Several figures are included to illustrate the adaptive construction of sparse grids and simulation of the rough Bergomi dynamics.
This document discusses algorithms for predictive modeling, including logistic regression. It presents a medical dataset containing measurements of heart patients and whether they survived. Logistic regression is applied to predict survival using maximum likelihood estimation. Numerical optimization techniques like BFGS and Fisher's algorithm are discussed for maximum likelihood estimation of logistic regression. Iteratively reweighted least squares is also presented as an alternative approach.
1) The document describes methods for optimizing the widths of radial basis functions in regression analysis models.
2) It presents an efficient computational method for re-estimating the regularization parameter based on generalized cross-validation that utilizes eigendecomposition.
3) The method is also extended to optimize the basis function width by testing multiple trial values and selecting the width with the smallest cross-validation value. Testing on practical problems showed the method improved prediction performance over fixed-width approaches.
Pre-computation for ABC in image analysisMatt Moores
MCMSki IV (the 5th IMS-ISBA joint meeting)
January 2014
Chamonix Mont-Blanc, France
The associated journal article has now been uploaded to arXiv: http://arxiv.org/abs/1403.4359
Markov chain Monte Carlo methods and some attempts at parallelizing themPierre Jacob
Markov chain Monte Carlo (MCMC) methods are commonly used to approximate properties of target probability distributions. However, MCMC estimators are generally biased for any fixed number of samples. The document discusses various techniques for constructing unbiased estimators from MCMC output, including regeneration, sequential Monte Carlo samplers, and coupled Markov chains. Specifically, running two Markov chains in parallel and taking the difference in their values at meeting times can yield an unbiased estimator, though certain conditions must hold.
- Bayesian adjustment for confounding (BAC) in Bayesian propensity score estimation accounts for uncertainty in propensity score modeling and model selection.
- A prognostic score model is used to inform a prior on propensity score model selection, favoring inclusion of true confounders and exclusion of instruments.
- Simulation results found the informative prior was not able to adequately shape model selection; a penalty term was proposed to make the prior more influential.
- With the penalty term, the informative prior influenced inclusion of instruments in propensity score models without distorting inclusion of other variables.
Bayesian Experimental Design for Stochastic Kinetic ModelsColin Gillespie
In recent years, the use of the Bayesian paradigm for estimating the optimal experimental design has increased. However, standard techniques are
computationally intensive for even relatively small stochastic kinetic models. One solution to this problem is to couple cloud computing with a model emulator.
By running simulations simultaneously in the cloud, the large design space can be explored. A Gaussian process is then fitted to this output, enabling the
optimal design parameters to be estimated.
Linear regression [Theory and Application (In physics point of view) using py...ANIRBANMAJUMDAR18
Machine-learning models are behind many recent technological advances, including high-accuracy translations of the text and self-driving cars. They are also increasingly used by researchers to help in solving physics problems, like Finding new phases of matter, Detecting interesting outliers
in data from high-energy physics experiments, Founding astronomical objects are known as gravitational lenses in maps of the night sky etc. The rudimentary algorithm that every Machine Learning enthusiast starts with is a linear regression algorithm. In statistics, linear regression is a linear approach to modelling the relationship between a scalar response (or dependent variable) and one or more explanatory variables (or independent
variables). Linear regression analysis (least squares) is used in a physics lab to prepare the computer-aided report and to fit data. In this article, the application is made to experiment: 'DETERMINATION OF DIELECTRIC CONSTANT OF NON-CONDUCTING LIQUIDS'. The entire computation is made through Python 3.6 programming language in this article.
This document discusses automatic Bayesian cubature for numerical integration. It begins with an introduction to multivariate integration and the challenges it poses. It then describes an automatic cubature algorithm that generates sample points and computes error bounds iteratively until a tolerance threshold is met. Next, it covers Bayesian cubature, which treats integrands as random functions to obtain probabilistic error bounds. It defines a Bayesian trio identity relating the integration error to discrepancies, variations, and alignments. The document concludes with discussions of future work.
A Monte Carlo strategy for structure multiple-step-head time series predictionGianluca Bontempi
The document proposes a Monte Carlo approach called SMC (Structured Monte Carlo) for multiple-step-ahead time series forecasting that takes into account the structural dependencies between predictions. It generates samples using a direct forecasting approach and weights them based on how well they satisfy dependencies identified by an iterated approach. Experiments on three benchmark datasets show the SMC approach achieves more accurate forecasts as measured by SMAPE than iterated, direct, or other comparison methods for most prediction horizons tested.
The document discusses estimation of multi-Granger network causal models from time series data. It proposes a joint modeling approach to estimate vector autoregressive (VAR) models for multiple time series datasets simultaneously. The key steps are:
1. Estimate the inverse covariance matrices for each dataset using a factor model approach.
2. Use the estimated inverse covariance matrices in a generalized fused lasso optimization to jointly estimate the VAR coefficient matrices for each dataset.
Simulation results show the joint modeling approach improves estimation of the VAR coefficients and reduces forecasting error compared to estimating the models separately, especially when the number of time points is small. The factor modeling approach also provides a better estimate of the inverse covariance than using the empirical estimate.
Inference for stochastic differential equations via approximate Bayesian comp...Umberto Picchini
Despite the title the methods are appropriate for more general dynamical models (including state-space models). Presentation given at Nordstat 2012, Umeå. Relevant research paper at http://arxiv.org/abs/1204.5459 and software code at https://sourceforge.net/projects/abc-sde/
The asynchronous parallel algorithms are developed to solve massive optimization problems in a distributed data system, which can be run in parallel on multiple nodes with little or no synchronization. Recently they have been successfully implemented to solve a range of difficult problems in practice. However, the existing theories are mostly based on fairly restrictive assumptions on the delays, and cannot explain the convergence and speedup properties of such algorithms. In this talk we will give an overview on distributed optimization, and discuss some new theoretical results on the convergence of asynchronous parallel stochastic gradient algorithm with unbounded delays. Simulated and real data will be used to demonstrate the practical implication of these theoretical results.
The smile calibration problem is a mathematical conundrum in finance that has challenged quantitative analysts for decades. Through his research, Aitor Muguruza has discovered a novel resolution to this classic problem.
Poster to be presented at Stochastic Numerics and Statistical Learning: Theory and Applications Workshop 2024, Kaust, Saudi Arabia, https://cemse.kaust.edu.sa/stochnum/events/event/snsl-workshop-2024.
In this work we have considered a setting that mimics the Henry problem \cite{Simpson2003,Simpson04_Henry}, modeling seawater intrusion into a 2D coastal aquifer. The pure water recharge from the ``land side'' resists the salinisation of the aquifer due to the influx of saline water through the ``sea side'', thereby achieving some equilibrium in the salt concentration. In our setting, following \cite{GRILLO2010}, we consider a fracture on the sea side that significantly increases the permeability of the porous medium.
The flow and transport essentially depend on the geological parameters of the porous medium, including the fracture. We investigated the effects of various uncertainties on saltwater intrusion. We assumed uncertainties in the fracture width, the porosity of the bulk medium, its permeability and the pure water recharge from the land side. The porosity and permeability were modeled by random fields, the recharge by a random but periodic intensity and the thickness by a random variable. We calculated the mean and variance of the salt mass fraction, which is also uncertain.
The main question we investigated in this work was how well the MLMC method can be used to compute statistics of different QoIs. We found that the answer depends on the choice of the QoI. First, not every QoI requires a hierarchy of meshes and MLMC. Second, MLMC requires stable convergence rates for $\EXP{g_{\ell} - g_{\ell-1}}$ and $\Var{g_{\ell} - g_{\ell-1}}$. These rates should be independent of $\ell$. If these convergence rates vary for different $\ell$, then it will be hard to estimate $L$ and $m_{\ell}$, and MLMC will either not work or be suboptimal. We were not able to get stable convergence rates for all levels $\ell=1,\ldots,5$ when the QoI was an integral as in \eqref{eq:integral_box}. We found that for $\ell=1,\ldots 4$ and $\ell=5$ the rate $\alpha$ was different. Further investigation is needed to find the reason for this. Another difficulty is the dependence on time, i.e. the number of levels $L$ and the number of sums $m_{\ell}$ depend on $t$. At the beginning the variability is small, then it increases, and after the process of mixing salt and fresh water has stopped, the variance decreases again.
The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level. These estimates depend on the minimisation function in the MLMC algorithm.
To achieve the efficiency of the MLMC approach presented in this work, it is essential that the complexity of the numerical solution of each random realisation is proportional to the number of grid vertices on the grid levels.
We investigated the applicability and efficiency of the MLMC approach to the Henry-like problem with uncertain porosity, permeability and recharge. These uncertain parameters were modelled by random fields with three independent random variables. Permeability is a function of porosity. Both functions are time-dependent, have multi-scale behaviour and are defined for two layers. The numerical solution for each random realisation was obtained using the well-known ug4 parallel multigrid solver. The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level.
The MLMC method was used to compute the expected value and variance of several QoIs, such as the solution at a few preselected points $(t,\bx)$, the solution integrated over a small subdomain, and the time evolution of the freshwater integral. We have found that some QoIs require only 2-3 mesh levels and samples from finer meshes would not significantly improve the result. Other QoIs require more grid levels.
1. Investigated efficiency of MLMC for Henry problem with
uncertain porosity, permeability, and recharge.
2. Uncertainties are modeled by random fields.
3. MLMC could be much faster than MC, 3200 times faster !
4. The time dependence is challenging.
Remarks:
1. Check if MLMC is needed.
2. The optimal number of samples depends on the point (t;x)
3. An advanced MLMC may give better estimates of L and m`.
Density Driven Groundwater Flow with Uncertain Porosity and PermeabilityAlexander Litvinenko
In this work, we solved the density driven groundwater flow problem with uncertain porosity and permeability. An accurate solution of this time-dependent and non-linear problem is impossible because of the presence of natural uncertainties in the reservoir such as porosity and permeability.
Therefore, we estimated the mean value and the variance of the solution, as well as the propagation of uncertainties from the random input parameters to the solution.
We started by defining the Elder-like problem. Then we described the multi-variate polynomial approximation (\gPC) approach and used it to estimate the required statistics of the mass fraction.
Utilizing the \gPC method allowed us
to reduce the computational cost compared to the classical quasi Monte Carlo method.
\gPC assumes that the output function $\sol(t,\bx,\thetab)$ is square-integrable and smooth w.r.t uncertain input variables $\btheta$.
Many factors, such as non-linearity, multiple solutions, multiple stationary states, time dependence and complicated solvers, make the investigation of the convergence of the \gPC method a non-trivial task.
We used an easy-to-implement, but only sub-optimal \gPC technique to quantify the uncertainty. For example, it is known that by increasing the degree of global polynomials (Hermite, Langange and similar), Runge's phenomenon appears. Here, probably local polynomials, splines or their mixtures would be better. Additionally, we used an easy-to-parallelise quadrature rule, which was also only suboptimal. For instance, adaptive choice of sparse grid (or collocation) points \cite{ConradMarzouk13,nobile-sg-mc-2015,Sudret_sparsePCE,CONSTANTINE12,crestaux2009polynomial} would be better, but we were limited by the usage of parallel methods. Adaptive quadrature rules are not (so well) parallelisable. In conclusion, we can report that: a) we developed a highly parallel method to quantify uncertainty in the Elder-like problem; b) with the \gPC of degree 4 we can achieve similar results as with the \QMC method.
In the numerical section we considered two different aquifers - a solid parallelepiped and a solid elliptic cylinder. One of our goals was to see how the domain geometry influences the formation, the number and the shape of fingers.
Since the considered problem is nonlinear,
a high variance in the porosity may result in totally different solutions; for instance, the number of fingers, their intensity and shape, the propagation time, and the velocity may vary considerably.
The number of cells in the presented experiments varied from $241{,}152$ to $15{,}433{,}728$ for the cylindrical domain and from $524{,}288$ to $4{,}194{,}304$ for the parallelepiped. The maximal number of parallel processing units was $600\times 32$, where $600$ is the number of parallel nodes and $32$ is the number of computing cores on each node. The total computing time varied from 2 hours for the coarse mesh to 24 hours for the finest mesh.
Saltwater intrusion occurs when sea levels rise and saltwater moves onto the land. Usually, this occurs during storms, high tides, droughts, or when saltwater penetrates freshwater aquifers and raises the groundwater table. Since groundwater is an essential nutrition and irrigation resource, its salinization may lead to catastrophic consequences. Many acres of farmland may be lost because they can become too wet or salty to grow crops. Therefore, accurate modeling of different scenarios of saline flow is essential to help farmers and researchers develop strategies to improve the soil quality and decrease saltwater intrusion effects.
Saline flow is density-driven and described by a system of time-dependent nonlinear partial differential equations (PDEs). It features convection dominance and can demonstrate very complicated behavior.
As a specific model, we consider a Henry-like problem with uncertain permeability and porosity.
These parameters may strongly affect the flow and transport of salt.
We consider a class of density-driven flow problems. We are particularly interested in the problem of the salinization of coastal aquifers. We consider the Henry saltwater intrusion problem with uncertain porosity, permeability, and recharge parameters as a test case.
The reason for the presence of uncertainties is the lack of knowledge, inaccurate measurements,
and inability to measure parameters at each spatial or time location. This problem is nonlinear and time-dependent. The solution is the salt mass fraction, which is uncertain and changes in time. Uncertainties in porosity, permeability, recharge, and mass fraction are modeled using random fields. This work investigates the applicability of the well-known multilevel Monte Carlo (MLMC) method for such problems. The MLMC method can reduce the total computational and storage costs. Moreover, the MLMC method runs multiple scenarios on different spatial and time meshes and then estimates the mean value of the mass fraction.
The parallelization is performed in both the physical space and stochastic space. To solve every deterministic scenario, we run the parallel multigrid solver ug4 in a black-box fashion.
We use the solution obtained from the quasi-Monte Carlo method as a reference solution.
We investigated the applicability and efficiency of the MLMC approach for the Henry-like problem with uncertain porosity, permeability, and recharge. These uncertain parameters were modeled by random fields with three independent random variables. The numerical solution for each random realization was obtained using the well-known ug4 parallel multigrid solver. The number of required random samples on each level was estimated by computing the decay of the variances and computational costs for each level. We also computed the expected value and variance of the mass fraction in the whole domain, the evolution of the pdfs, the solutions at a few preselected points $(t,\bx)$, and the time evolution of the freshwater integral value. We have found that some QoIs require only 2-3 of the coarsest mesh levels, and samples from finer meshes would not significantly improve the result. Note that a different type of porosity may lead to a different conclusion.
The results show that the MLMC method is faster than the QMC method at the finest mesh. Thus, sampling at different mesh levels makes sense and helps to reduce the overall computational cost.
Here the interest is mainly to compute characterisations like the entropy,
the Kullback-Leibler divergence, more general $f$-divergences, or other such characteristics based on
the probability density. The density is often not available directly,
and it is a computational challenge to just represent it in a numerically
feasible fashion in case the dimension is even moderately large. It
is an even stronger numerical challenge to then actually compute said characteristics
in the high-dimensional case.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.\
$O(d n r^2 )$ for the TT format. Here $n$ is the number of discretisation
points in one direction, $r<<n$ is the maximal tensor rank, and $d$ the problem dimension.
This document proposes a method for weakly supervised regression on uncertain datasets. It combines graph Laplacian regularization and cluster ensemble methodology. The method solves an auxiliary minimization problem to determine the optimal solution for predicting uncertain parameters. It is tested on artificial data to predict target values using a mixture of normal distributions with labeled, inaccurately labeled, and unlabeled samples. The method is shown to outperform a simplified version by reducing mean Wasserstein distance between predicted and true values.
Computing f-Divergences and Distances of High-Dimensional Probability Density...Alexander Litvinenko
Poster presented on Stochastic Numerics and Statistical Learning: Theory and Applications Workshop in KAUST, Saudi Arabia.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
Even for moderate dimension $d$, the full storage and computation with such objects become very quickly infeasible.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.
O(d n r^2) for the TT format. Here $n$ is the number of discretisation
points in one direction, r<n is the maximal tensor rank, and d the problem dimension.
The particular data format is rather unimportant,
any of the well-known tensor formats (CP, Tucker, hierarchical Tucker, tensor-train (TT)) can be used,
and we used the TT data format. Much of the presentation and in fact the central train
of discussion and thought is actually independent of the actual representation.
In the beginning it was motivated through three possible ways how one may
arrive at such a representation of the pdf. One was if the pdf was given in some approximate
analytical form, e.g. like a function tensor product of lower-dimensional pdfs with a
product measure, or from an analogous representation of the pcf and subsequent use of the
Fourier transform, or from a low-rank functional representation of a high-dimensional
RV, again via its pcf.
The theoretical underpinnings of the relation between pdfs and pcfs as well as their
properties were recalled in Section: Theory, as they are important to be preserved in the
discrete approximation. This also introduced the concepts of the convolution and of
the point-wise multiplication Hadamard algebra, concepts which become especially important if
one wants to characterise sums of independent RVs or mixture models,
a topic we did not touch on for the sake of brevity but which follows very naturally from
the developments here. Especially the Hadamard algebra is also
important for the algorithms to compute various point-wise functions in the sparse formats.
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Alexander Litvinenko
Talk presented on SIAM IS 2022 conference.
Very often, in the course of uncertainty quantification tasks or
data analysis, one has to deal with high-dimensional random variables (RVs)
(with values in $\Rd$). Just like any other RV,
a high-dimensional RV can be described by its probability density (\pdf) and/or
by the corresponding probability characteristic functions (\pcf),
or a more general representation as
a function of other, known, random variables.
Here the interest is mainly to compute characterisations like the entropy, the Kullback-Leibler, or more general
$f$-divergences. These are all computed from the \pdf, which is often not available directly,
and it is a computational challenge to even represent it in a numerically
feasible fashion in case the dimension $d$ is even moderately large. It
is an even stronger numerical challenge to then actually compute said characterisations
in the high-dimensional case.
In this regard, in order to achieve a computationally feasible task, we propose
to approximate density by a low-rank tensor.
Low rank tensor approximation of probability density and characteristic funct...Alexander Litvinenko
This document summarizes a presentation on computing divergences and distances between high-dimensional probability density functions (pdfs) represented using tensor formats. It discusses:
1) Motivating the problem using examples from stochastic PDEs and functional representations of uncertainties.
2) Computing Kullback-Leibler divergence and other divergences when pdfs are not directly available.
3) Representing probability characteristic functions and approximating pdfs using tensor decompositions like CP and TT formats.
4) Numerical examples computing Kullback-Leibler divergence and Hellinger distance between Gaussian and alpha-stable distributions using these tensor approximations.
Identification of unknown parameters and prediction of missing values. Compar...Alexander Litvinenko
H-matrix approximation of large Mat\'{e}rn covariance matrices, Gaussian log-likelihoods.
Identifying unknown parameters and making predictions
Comparison with machine learning methods.
kNN is easy to implement and shows promising results.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
This document describes using the Continuation Multi-Level Monte Carlo (CMLMC) method to compute electromagnetic fields scattered from dielectric objects of uncertain shapes. CMLMC optimally balances statistical and discretization errors using fewer samples on fine meshes and more on coarse meshes. The method is tested by computing scattering cross sections for randomly perturbed spheres under plane wave excitation and comparing results to the unperturbed sphere. Computational costs and errors are analyzed to demonstrate the efficiency of CMLMC for this scattering problem with uncertain geometry.
Identification of unknown parameters and prediction with hierarchical matrice...Alexander Litvinenko
We compare four numerical methods for the prediction of missing values in four different datasets.
These methods are 1) the hierarchical maximum likelihood estimation (H-MLE), and three machine learning (ML) methods, which include 2) k-nearest neighbors (kNN), 3) random forest, and 4) Deep Neural Network (DNN).
From the ML methods, the best results (for considered datasets) were obtained by the kNN method with three (or seven) neighbors.
On one dataset, the MLE method showed a smaller error than the kNN method, whereas, on another, the kNN method was better.
The MLE method requires a lot of linear algebra computations and works fine on almost all datasets. Its result can be improved by taking a smaller threshold and more accurate hierarchical matrix arithmetics. To our surprise, the well-known kNN method produces similar results as H-MLE and worked much faster.
1. Motivation: why do we need low-rank tensors
2. Tensors of the second order (matrices)
3. CP, Tucker and tensor train tensor formats
4. Many classical kernels have (or can be approximated in ) low-rank tensor format
5. Post processing: Computation of mean, variance, level sets, frequency
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Propagation of Uncertainties in Density Driven Groundwater FlowAlexander Litvinenko
Major Goal: estimate risks of the pollution in a subsurface flow.
How?: we solve density-driven groundwater flow with uncertain porosity and permeability.
We set up density-driven groundwater flow problem,
review stochastic modeling and stochastic methods, use UG4 framework (https://gcsc.uni-frankfurt.de/simulation-and-modelling/ug4),
model uncertainty in porosity and permeability,
2D and 3D numerical experiments.
A workshop hosted by the South African Journal of Science aimed at postgraduate students and early career researchers with little or no experience in writing and publishing journal articles.
June 3, 2024 Anti-Semitism Letter Sent to MIT President Kornbluth and MIT Cor...Levi Shapiro
Letter from the Congress of the United States regarding Anti-Semitism sent June 3rd to MIT President Sally Kornbluth, MIT Corp Chair, Mark Gorenberg
Dear Dr. Kornbluth and Mr. Gorenberg,
The US House of Representatives is deeply concerned by ongoing and pervasive acts of antisemitic
harassment and intimidation at the Massachusetts Institute of Technology (MIT). Failing to act decisively to ensure a safe learning environment for all students would be a grave dereliction of your responsibilities as President of MIT and Chair of the MIT Corporation.
This Congress will not stand idly by and allow an environment hostile to Jewish students to persist. The House believes that your institution is in violation of Title VI of the Civil Rights Act, and the inability or
unwillingness to rectify this violation through action requires accountability.
Postsecondary education is a unique opportunity for students to learn and have their ideas and beliefs challenged. However, universities receiving hundreds of millions of federal funds annually have denied
students that opportunity and have been hijacked to become venues for the promotion of terrorism, antisemitic harassment and intimidation, unlawful encampments, and in some cases, assaults and riots.
The House of Representatives will not countenance the use of federal funds to indoctrinate students into hateful, antisemitic, anti-American supporters of terrorism. Investigations into campus antisemitism by the Committee on Education and the Workforce and the Committee on Ways and Means have been expanded into a Congress-wide probe across all relevant jurisdictions to address this national crisis. The undersigned Committees will conduct oversight into the use of federal funds at MIT and its learning environment under authorities granted to each Committee.
• The Committee on Education and the Workforce has been investigating your institution since December 7, 2023. The Committee has broad jurisdiction over postsecondary education, including its compliance with Title VI of the Civil Rights Act, campus safety concerns over disruptions to the learning environment, and the awarding of federal student aid under the Higher Education Act.
• The Committee on Oversight and Accountability is investigating the sources of funding and other support flowing to groups espousing pro-Hamas propaganda and engaged in antisemitic harassment and intimidation of students. The Committee on Oversight and Accountability is the principal oversight committee of the US House of Representatives and has broad authority to investigate “any matter” at “any time” under House Rule X.
• The Committee on Ways and Means has been investigating several universities since November 15, 2023, when the Committee held a hearing entitled From Ivory Towers to Dark Corners: Investigating the Nexus Between Antisemitism, Tax-Exempt Universities, and Terror Financing. The Committee followed the hearing with letters to those institutions on January 10, 202
How to Add Chatter in the odoo 17 ERP ModuleCeline George
In Odoo, the chatter is like a chat tool that helps you work together on records. You can leave notes and track things, making it easier to talk with your team and partners. Inside chatter, all communication history, activity, and changes will be displayed.
This slide is special for master students (MIBS & MIFB) in UUM. Also useful for readers who are interested in the topic of contemporary Islamic banking.
Biological screening of herbal drugs: Introduction and Need for
Phyto-Pharmacological Screening, New Strategies for evaluating
Natural Products, In vitro evaluation techniques for Antioxidants, Antimicrobial and Anticancer drugs. In vivo evaluation techniques
for Anti-inflammatory, Antiulcer, Anticancer, Wound healing, Antidiabetic, Hepatoprotective, Cardio protective, Diuretics and
Antifertility, Toxicity studies as per OECD guidelines
A Strategic Approach: GenAI in EducationPeter Windle
Artificial Intelligence (AI) technologies such as Generative AI, Image Generators and Large Language Models have had a dramatic impact on teaching, learning and assessment over the past 18 months. The most immediate threat AI posed was to Academic Integrity with Higher Education Institutes (HEIs) focusing their efforts on combating the use of GenAI in assessment. Guidelines were developed for staff and students, policies put in place too. Innovative educators have forged paths in the use of Generative AI for teaching, learning and assessments leading to pockets of transformation springing up across HEIs, often with little or no top-down guidance, support or direction.
This Gasta posits a strategic approach to integrating AI into HEIs to prepare staff, students and the curriculum for an evolving world and workplace. We will highlight the advantages of working with these technologies beyond the realm of teaching, learning and assessment by considering prompt engineering skills, industry impact, curriculum changes, and the need for staff upskilling. In contrast, not engaging strategically with Generative AI poses risks, including falling behind peers, missed opportunities and failing to ensure our graduates remain employable. The rapid evolution of AI technologies necessitates a proactive and strategic approach if we are to remain relevant.
This presentation includes basic of PCOS their pathology and treatment and also Ayurveda correlation of PCOS and Ayurvedic line of treatment mentioned in classics.
This presentation was provided by Steph Pollock of The American Psychological Association’s Journals Program, and Damita Snow, of The American Society of Civil Engineers (ASCE), for the initial session of NISO's 2024 Training Series "DEIA in the Scholarly Landscape." Session One: 'Setting Expectations: a DEIA Primer,' was held June 6, 2024.
How to Build a Module in Odoo 17 Using the Scaffold MethodCeline George
Odoo provides an option for creating a module by using a single line command. By using this command the user can make a whole structure of a module. It is very easy for a beginner to make a module. There is no need to make each file manually. This slide will show how to create a module using the scaffold method.
Thinking of getting a dog? Be aware that breeds like Pit Bulls, Rottweilers, and German Shepherds can be loyal and dangerous. Proper training and socialization are crucial to preventing aggressive behaviors. Ensure safety by understanding their needs and always supervising interactions. Stay safe, and enjoy your furry friends!
1. Center for Uncertainty
Quantification
Solving inverse problem via
non-linear update of PCE coefficients
A. Litvinenko1, H.G. Matthies2, B. Rosic2, E. Zander2
1
CEMSE Division, KAUST, 2
TU Braunschweig, Germany
alexander.litvinenko@kaust.edu.sa
Center for Uncertainty
Quantification
Center for Uncertainty
Quantification
Motivation
Notation: Given a physical system modeled
by a PDE or ODE with uncertain coefficient
q(x, ω):
A(u, q(x, ω)) = f.
A solution operator S: u = u(x, ω) = S(q, f).
A measurement operator: y = M(u). Noisy
observations z(ω) = ˆy + ε(ω) with random
measurement error ε and ‘truth’ ˆy.
Very often, after applying stochastic Galerkin
or building a surrogate, we have all ingredi-
ents in gPCE basis. It would be nice to do
data assimilation/compute ’Bayesian’ update
with gPCE coefficients (to avoid sampling).
Aim: given noisy observations z(ω), to iden-
tify q(ω).
How?: To identify q(ω) we derived non-linear
approximation of the Bayesian update from
the variational problem associated with con-
ditional expectation. To reduce cost of the
’Bayesian’ update we offer a functional ap-
proximation, e.g. gPCE.
New: We apply ’Bayesian’ update to gPCE
coefficients of q(ω) (not to the probability den-
sity function of q).
1. Minimum Mean Square Error
Estimation (MMSE)
Let q(ξ) : Ω → RNq
be the (a priori) stochastic model
of some unknown QoI (e.g., uncertain parameter q),
Y : Ω → RNy
be the stochastic model (e.g. of mea-
surement forecast Y = M(q(ξ)) + ε(ξM)). We search
for a function ϕ : RNy
→ RNq
. The best estimator ˆϕ for
Y given q is
ˆϕ = argminϕ E[ q(ξ) − ϕ(Y + ε(ξM)) 2
2], (1)
where the expectation needs to be taken over Ω =
Ω × ΩM. Writing ξ = (ξ, ξM) ∈ Ω the best estimator
(or predictor) of q given the measurement model is
qM(ξ ) = ˆϕ(Y + ε(ξM)). (2)
Suppose the actual measurements are: yM(ξ ) =
¯yM + εM(ξ ), where ¯yM are the measured values and
εM(ξ ) is some assumed error model, then
q(ξ ) = ˆϕ(¯yM + εM(ξ )). (3)
2. Generalized PCE of mapping ϕ
Let us represent ϕ as a gPCE
ϕ ≈ ˆϕ = y →
γ∈J
ϕγΨγ(y(ξ))
γ - multi-index and J a multi-index set.
Compute unknown coefficients ϕγ by minimizing MSE,
by taking derivative w.r.t. ϕγ:
∂
∂ϕγ
E[(q −
γ∈J
ϕγΨγ(Y ))2
] = 0, (4)
2
γ∈J
E [Ψγ(y)Ψδ(y)] ϕγ − E [qΨδ(y)]
= 0, ∀δ ∈ J ,
γ∈J
ϕγE[Ψγ(Y )Ψδ(Y )] = E[qΨδ(Y )], (5)
Aγδ := E[Ψγ(Y )Ψδ(Y )] ≈
NA
k=1
wA
k Ψγ(Y (ξk))Ψδ(Y (ξk)),
E[qΨδ(Y )] ≈
Nb
k=1
wb
kq(ξk)Ψδ(Y (ξk)), or in a matrix form,
(6)
V[diag(...wA
k ...)]VT
...
ϕβ
...
= W
wb
1q(ξ1)
...
wb
Nb
q(ξNb
)
, (7)
V := [..., Ψγ, ...]T
∈ R|Jγ|×NA
, [diag(...wA
k ...)] ∈ RNA×NA
,
W ∈ R|Jα|×Nb
, [wb
0q(ξ0)...wb
Nb
q(ξNb
)] ∈ RNb
.
Solving Eq. 7, obtain vector of coefficients (...ϕβ...) for
all β and then compute the update:
qnew(ξ ) = ˆϕ(¯yM + εM(ξ )). (8)
Example 1. The mapping ϕ does not exist in the
Hermite basis. y(ξ) = ξ2
, q(ξ) = ξ3
. PCE coefficients
are (1, 0, 1, 0): ξ2
= 1·H0(ξ)+0·H1(ξ)+1·H2(ξ)+0·H3(ξ)
and (0, 3, 0, 1): ξ3
= 0·H0(ξ)+3·H1(ξ)+0·H2(ξ)+1·H3(ξ).
Mapping ϕ does not exist. The matrix A is close to
singular. Support of Hermite polynomials (used for
Gaussian RVs) is (−∞, ∞).
Example 2. The mapping ϕ does exist in the La-
guerre basis.
y(ξ) = ξ2
, q(ξ) = ξ3
. gPCE coefficients are (2, −4, 2, 0)
and (6, −18, 18, −6). Mapping ϕ of order 8 and higher
produces a very accurate result. Support of Laguerre
polynomials (used for Gamma RVs) is [0, ∞).
In [1-5] we demonstrated that linear ϕ corresponds to
the well-known Kalman Filter.
Implementation: implementation in Matlab using the
Stochastic Galerkin library sglib by E. Zander [7].
3. Numerics
10 0 10
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
x
20 0 20
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
y
0 10 20
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
z
xf
x
a
yf
y
a
zf
z
a
Figure 1: Lorenz-84. Linear measurement
(x(t), y(t), z(t)) at t = 10: prior and posterior af-
ter one update.
10 5 0
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
x
x1
x2
15 10 5
0
0.05
0.1
0.15
0.2
0.25
0.3
0.35
0.4
0.45
0.5
y
y1
y2
5 10 15
0
0.1
0.2
0.3
0.4
0.5
0.6
z
z1
z2
Figure 2: Lorenz-84. Quadratic measurement
(x(t)2
, y(t)2
, z(t)2
) at t = 10: Comparison posterior for
LBU and NLBU after one update.
Example 3. Diffusion with uncertain coefficients:
− · (κ(x, ξ) u(x, ξ)) = f(x, ξ), D = [0, 1]. (9)
Measurements are in x1 = 0.2 and x2 = 0.8 with values
of y1 = 10 and y2 = 5 and noise with st. deviations of
σ1 = 0.5 and σ2 = 1.5.
0 0.2 0.4 0.6 0.8 1
−20
−10
0
10
20
30
0 0.2 0.4 0.6 0.8 1
−20
−10
0
10
20
30
Figure 3: Updating of the solution u. Left - original so-
lution u(ξ) and right the updated solution u (ξ ). Shown
are the mean +/− one to three standard deviations,
plus additionally 20 sample realisations. Uncertainty
decreases in the measurement points (0.2, 0.8).
0 0.2 0.4 0.6 0.8 1
−1
−0.5
0
0.5
1
1.5
2
0 0.2 0.4 0.6 0.8 1
−1
−0.5
0
0.5
1
1.5
2
Figure 4: Updating of the parameter κ. Left is the pa-
rameter model κ(ξ) and right the updated parameter
model κ (ξ ). Uncertainty decreases in the measure-
ment points (0.2, 0.8).
0.6 0.7 0.8 0.9 1
0.2
0.3
0.4
0.5
0.6 0.7 0.8 0.9 1
0.2
0.3
0.4
0.5
Figure 5: MMSE estimation with increasing polyno-
mial degrees pϕ = 1 and 3 from left to right. True
values X are marked by o, and estimated values
ˆX = ˆϕ(Y ) are marked by x.
Conclusion
1.We take into account the available mea-
surements and compute an update/a pos-
teriori gPCE coefficients of QoI (e.g. un-
certain coefficients).
2.We minimize MMSE and compute condi-
tional expectation
3.We developed a cheap gPCE based ap-
proximation of the Bayesian Update (see
[8]).
4.Introduced a way to derive MMSE ϕ (as a
linear, quadratic, cubic etc approximation,
i. e. compute conditional expectation of q,
given measurements .
5.Linear ϕ is equivalent to the Kalman filter.
Acknowledgements: A. Litvinenko is a member of the KAUST
ECRC and SRI UQ centers in Computational Science and Engi-
neering.
References
1. H.G. Matthies, E. Zander, B.V. Rosi´c, A. Litvinenko, O. Pajonk,
Inverse Problems in a Bayesian Setting, arXiv:1511.00524,
(2015).
2. A. Litvinenko, H. G. Matthies, Inverse problems and uncer-
tainty quantification, arXiv:1312.5048, (2013).
3. B. V. Rosi´c, A. Kucerova, J. Sykora, O. Pajonk, A. Litvinenko,
H. G. Matthies, Parameter Identification in a Probabilistic Set-
ting, Engineering Structures (2013).
4. O. Pajonk, B. V. Rosic, A. Litvinenko, H. G. Matthies, A Deter-
ministic Filter for Non-Gaussian Bayesian Estimation, - appli-
cations to dynamical system estimation with noisy measure-
ments. Physica D: Nonlinear Phenomena, Vol. 241(7), pp.
775-788, (2012).
5. B.V. Rosi´c, A. Litvinenko, O. Pajonk, H.G. Matthies, Sampling-
free linear Bayesian update of polynomial chaos representa-
tions, Journal of Computational Physics 231 (17), 5761-5787,
(2012).
6. H.G. Matthies, A. Litvinenko, O. Pajonk, B.V. Rosi´c, E. Zander,
Parametric and uncertainty computations with tensor product
representations, Uncertainty Quantification in Scientific Com-
puting, 139-150, (2012).
7. E. Zander, Stochastic Galerkin library
https://github.com/ezander/sglib
8. O. G. Ernst, B. Sprungk, H.-J. Starkloff, Analysis of the Ensem-
ble and Polynomial Chaos Kalman Filters in Bayesian Inverse
Problems, SIAM/ASA JUQ 3(1), (2015)