We apply tensor train (TT) data format to solve an elliptic PDE with uncertain coefficients. We reduce complexity and storage from exponential to linear. Post-processing in TT format is also provided.
We start with motivation, few examples of uncertainties. Then we discretize elliptic PDE with uncertain coefficients, apply TT format for permeability, the stochastic operator and for the solution. We compare sparse multi-index set approach with full multi-index+TT.
Tensor Train format allows us to keep the whole multi-index set, without any multi-index set truncation.
Reinforcement learning: hidden theory, and new super-fast algorithms
Lecture presented at the Center for Systems and Control (CSC@USC) and Ming Hsieh Institute for Electrical Engineering,
February 21, 2018
Stochastic Approximation algorithms are used to approximate solutions to fixed point equations that involve expectations of functions with respect to possibly unknown distributions. The most famous examples today are TD- and Q-learning algorithms. The first half of this lecture will provide an overview of stochastic approximation, with a focus on optimizing the rate of convergence. A new approach to optimize the rate of convergence leads to the new Zap Q-learning algorithm. Analysis suggests that its transient behavior is a close match to a deterministic Newton-Raphson implementation, and numerical experiments confirm super fast convergence.
Based on
@article{devmey17a,
Title = {Fastest Convergence for {Q-learning}},
Author = {Devraj, Adithya M. and Meyn, Sean P.},
Journal = {NIPS 2017 and ArXiv e-prints},
Year = 2017}
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsSean Meyn
A tutorial, and very new algorithms -- more details on arXiv and at NIPS 2017 https://arxiv.org/abs/1707.03770
Part of the Data Science Summer School at École Polytechnique: http://www.ds3-datascience-polytechnique.fr/program/
---------
2018 Updates:
See Zap slides from ISMP 2018 for new inverse-free optimal algorithms
Simons tutorial, March 2018 [one month before most discoveries announced at ISMP]
Part I (Basics, with focus on variance of algorithms)
https://www.youtube.com/watch?v=dhEF5pfYmvc
Part II (Zap Q-learning)
https://www.youtube.com/watch?v=Y3w8f1xIb6s
Big 2017 survey on variance in SA:
Fastest convergence for Q-learning
https://arxiv.org/abs/1707.03770
You will find the infinite-variance Q result there.
Our NIPS 2017 paper is distilled from this.
Typically quantifying uncertainty requires many evaluations of a computational model or simulator. If a simulator is computationally expensive and/or high-dimensional, working directly with a simulator often proves intractable. Surrogates of expensive simulators are popular and powerful tools for overcoming these challenges. I will give an overview of surrogate approaches from an applied math perspective and from a statistics perspective with the goal of setting the stage for the "other" community.
We start with motivation, few examples of uncertainties. Then we discretize elliptic PDE with uncertain coefficients, apply TT format for permeability, the stochastic operator and for the solution. We compare sparse multi-index set approach with full multi-index+TT.
Tensor Train format allows us to keep the whole multi-index set, without any multi-index set truncation.
Reinforcement learning: hidden theory, and new super-fast algorithms
Lecture presented at the Center for Systems and Control (CSC@USC) and Ming Hsieh Institute for Electrical Engineering,
February 21, 2018
Stochastic Approximation algorithms are used to approximate solutions to fixed point equations that involve expectations of functions with respect to possibly unknown distributions. The most famous examples today are TD- and Q-learning algorithms. The first half of this lecture will provide an overview of stochastic approximation, with a focus on optimizing the rate of convergence. A new approach to optimize the rate of convergence leads to the new Zap Q-learning algorithm. Analysis suggests that its transient behavior is a close match to a deterministic Newton-Raphson implementation, and numerical experiments confirm super fast convergence.
Based on
@article{devmey17a,
Title = {Fastest Convergence for {Q-learning}},
Author = {Devraj, Adithya M. and Meyn, Sean P.},
Journal = {NIPS 2017 and ArXiv e-prints},
Year = 2017}
Reinforcement Learning: Hidden Theory and New Super-Fast AlgorithmsSean Meyn
A tutorial, and very new algorithms -- more details on arXiv and at NIPS 2017 https://arxiv.org/abs/1707.03770
Part of the Data Science Summer School at École Polytechnique: http://www.ds3-datascience-polytechnique.fr/program/
---------
2018 Updates:
See Zap slides from ISMP 2018 for new inverse-free optimal algorithms
Simons tutorial, March 2018 [one month before most discoveries announced at ISMP]
Part I (Basics, with focus on variance of algorithms)
https://www.youtube.com/watch?v=dhEF5pfYmvc
Part II (Zap Q-learning)
https://www.youtube.com/watch?v=Y3w8f1xIb6s
Big 2017 survey on variance in SA:
Fastest convergence for Q-learning
https://arxiv.org/abs/1707.03770
You will find the infinite-variance Q result there.
Our NIPS 2017 paper is distilled from this.
Typically quantifying uncertainty requires many evaluations of a computational model or simulator. If a simulator is computationally expensive and/or high-dimensional, working directly with a simulator often proves intractable. Surrogates of expensive simulators are popular and powerful tools for overcoming these challenges. I will give an overview of surrogate approaches from an applied math perspective and from a statistics perspective with the goal of setting the stage for the "other" community.
Simplified Runtime Analysis of Estimation of Distribution AlgorithmsPK Lehre
We describe how to estimate the optimisation time of the UMDA, an estimation of distribution algorithm, using the level-based theorem. The paper was presented at GECCO 2015 in Madrid.
Seminar at IEEE Computational Intelligence Society, Singapore Chapter at School of Electrical and Electronic Engineering, NTU, Singapore, 20 February 2019
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
ICML 2021 tutorial on random matrix theory and machine learning.
Part 3 covers: 1. Motivation: Average-case versus worst-case in high dimensions 2. Algorithm halting times (runtimes) 3. Outlook
Multi-linear algebra and different tensor formats with applications Alexander Litvinenko
A short overview of well-known tensor formats, elliptic PDE with uncertain coefficients, some academic examples of separable functions, post-processing in tensor format
Simplified Runtime Analysis of Estimation of Distribution AlgorithmsPK Lehre
We describe how to estimate the optimisation time of the UMDA, an estimation of distribution algorithm, using the level-based theorem. The paper was presented at GECCO 2015 in Madrid.
Seminar at IEEE Computational Intelligence Society, Singapore Chapter at School of Electrical and Electronic Engineering, NTU, Singapore, 20 February 2019
Random Matrix Theory and Machine Learning - Part 3Fabian Pedregosa
ICML 2021 tutorial on random matrix theory and machine learning.
Part 3 covers: 1. Motivation: Average-case versus worst-case in high dimensions 2. Algorithm halting times (runtimes) 3. Outlook
Multi-linear algebra and different tensor formats with applications Alexander Litvinenko
A short overview of well-known tensor formats, elliptic PDE with uncertain coefficients, some academic examples of separable functions, post-processing in tensor format
We consider an elliptic BVP.
How to compute a part of the solution? For instance, solution on the interface, solution in s subdomain in a point without computing the whole solution and with O(n log n) complexity/storage.
My paper for Domain Decomposition Conference in Strobl, Austria, 2005Alexander Litvinenko
We did a first step in solving, so-called, skin problem. We developed an efficient H-matrix preconditioner to solve diffusion problem with jumping coefficients
Application H-matrices for solving PDEs with multi-scale coefficients, jumpin...Alexander Litvinenko
We develop hierarchical domain decomposition method to compute a part of the solution, a part of the inverse operator with O(n log n) storage and computing cost.
Response Surface in Tensor Train format for Uncertainty QuantificationAlexander Litvinenko
We apply low-rank Tensor Train format to solve PDEs with uncertain coefficients. First, we approximate uncertain permeability coefficient in TT format, then the operator and then apply iterations to solve stochastic Galerkin system.
Hierarchical matrix approximation of large covariance matricesAlexander Litvinenko
We research class of Matern covariance matrices and their approximability in the H-matrix format. Further tasks are compute H-Cholesky factorization, trace, determinant, quadratic form, loglikelihood. Later H-matrices can be applied in kriging.
We research how we can use Scalable hierarchical algorithms for solving stochastic PDEs and for Uncertainty Quantification. Particularly we are interested in approximating large covariance matrices in H-matrix format, Hierarchical Cholesky factorization and computing Karhunen-Loeve expansion
Minimum mean square error estimation and approximation of the Bayesian updateAlexander Litvinenko
We develop a Bayesian update surrogate. Our formula allows us to update polynomial chaos coefficients. In contrast to classical Bayesian approach, we suggest to update PCE coefficients. We show that classical Kalman filter is a particular case of our update.
Computation of Electromagnetic Fields Scattered from Dielectric Objects of Un...Alexander Litvinenko
We research how input uncertainties in the geometry shape propagate through the electromagnetic model to electro-magnetic fields. We use multi-level Monte Carlo methods.
After we applied the stochastic Galerkin method to solve stochastic PDE, and solve large linear system, we obtain stochastic solution (random field), which is represented in Karhunen Loeve and PCE basis. No sampling error is involved, only algebraic truncation error. Now we would like to escape classical MCMC path to compute the posterior. We develop an Bayesian* update formula for KLE-PCE coefficients.
Gracheva Inessa - Fast Global Image Denoising Algorithm on the Basis of Nonst...AIST
Gracheva Inessa, Kopylov Andrey, Krasotkina Olga,
(Tula State University, Tula, Russia) - Fast Global Image Denoising Algorithm on the Basis of Nonstationary Gamma-Normal Statistical Model
AIST 2015 Conference
Efficient Analysis of high-dimensional data in tensor formatsAlexander Litvinenko
We solve a PDE with uncertain coefficients. The solution is approximated in the Karhunen Loeve/PCE basis. How to compute maximum ? frequency? probability density function? with almost linear complexity? We offer various methods.
Possible applications of low-rank tensors in statistics and UQ (my talk in Bo...Alexander Litvinenko
Just some ideas how low-rank matrices/tensors can be useful in spatial and environmental statistics, where one usually has to deal with very large data
Research internship on optimal stochastic theory with financial application u...Asma Ben Slimene
This is a presntation of my second year intership on optimal stochastic theory and how we can apply it on some financial application then how we can solve such problems using finite differences methods!
Enjoy it !
Presentation on stochastic control problem with financial applications (Merto...Asma Ben Slimene
This is an introductory to optimal stochastic control theory with two applications in finance: Merton portfolio problem and Investement/consumption problem with numerical results using finite differences approach
Distributed solution of stochastic optimal control problem on GPUsPantelis Sopasakis
Stochastic optimal control problems arise in many
applications and are, in principle,
large-scale involving up to millions of decision variables. Their
applicability in control applications is often limited by the
availability of algorithms that can solve them efficiently and within
the sampling time of the controlled system.
In this paper we propose a dual accelerated proximal
gradient algorithm which is amenable to parallelization and
demonstrate that its GPU implementation affords high speed-up
values (with respect to a CPU implementation) and greatly outperforms
well-established commercial optimizers such as Gurobi.
ABC with data cloning for MLE in state space modelsUmberto Picchini
An application of the "data cloning" method for parameter estimation via MLE aided by Approximate Bayesian Computation. The relevant paper is http://arxiv.org/abs/1505.06318
In this talk, we discuss some recent advances in probabilistic schemes for high-dimensional PIDEs. It is known that traditional PDE solvers, e.g., finite element, finite difference methods, do not scale well with the increase of dimension. The idea of probabilistic schemes is to link a wide class of nonlinear parabolic PIDEs to stochastic Levy processes based on nonlinear version of the Feynman-Kac theory. As such, the solution of the PIDE can be represented by a conditional expectation (i.e., a high-dimensional integral) with respect to a stochastic dynamical system driven by Levy processes. In other words, we can solve the PIDEs by performing high-dimensional numerical integration. A variety of quadrature methods could be applied, including MC, QMC, sparse grids, etc. The probabilistic schemes have been used in many application problems, e.g., particle transport in plasmas (e.g., Vlasov-Fokker-Planck equations), nonlinear filtering (e.g., Zakai equations), and option pricing, etc.
Inference for stochastic differential equations via approximate Bayesian comp...Umberto Picchini
Despite the title the methods are appropriate for more general dynamical models (including state-space models). Presentation given at Nordstat 2012, Umeå. Relevant research paper at http://arxiv.org/abs/1204.5459 and software code at https://sourceforge.net/projects/abc-sde/
Similar to Tensor train to solve stochastic PDEs (20)
Poster to be presented at Stochastic Numerics and Statistical Learning: Theory and Applications Workshop 2024, Kaust, Saudi Arabia, https://cemse.kaust.edu.sa/stochnum/events/event/snsl-workshop-2024.
In this work we have considered a setting that mimics the Henry problem \cite{Simpson2003,Simpson04_Henry}, modeling seawater intrusion into a 2D coastal aquifer. The pure water recharge from the ``land side'' resists the salinisation of the aquifer due to the influx of saline water through the ``sea side'', thereby achieving some equilibrium in the salt concentration. In our setting, following \cite{GRILLO2010}, we consider a fracture on the sea side that significantly increases the permeability of the porous medium.
The flow and transport essentially depend on the geological parameters of the porous medium, including the fracture. We investigated the effects of various uncertainties on saltwater intrusion. We assumed uncertainties in the fracture width, the porosity of the bulk medium, its permeability and the pure water recharge from the land side. The porosity and permeability were modeled by random fields, the recharge by a random but periodic intensity and the thickness by a random variable. We calculated the mean and variance of the salt mass fraction, which is also uncertain.
The main question we investigated in this work was how well the MLMC method can be used to compute statistics of different QoIs. We found that the answer depends on the choice of the QoI. First, not every QoI requires a hierarchy of meshes and MLMC. Second, MLMC requires stable convergence rates for $\EXP{g_{\ell} - g_{\ell-1}}$ and $\Var{g_{\ell} - g_{\ell-1}}$. These rates should be independent of $\ell$. If these convergence rates vary for different $\ell$, then it will be hard to estimate $L$ and $m_{\ell}$, and MLMC will either not work or be suboptimal. We were not able to get stable convergence rates for all levels $\ell=1,\ldots,5$ when the QoI was an integral as in \eqref{eq:integral_box}. We found that for $\ell=1,\ldots 4$ and $\ell=5$ the rate $\alpha$ was different. Further investigation is needed to find the reason for this. Another difficulty is the dependence on time, i.e. the number of levels $L$ and the number of sums $m_{\ell}$ depend on $t$. At the beginning the variability is small, then it increases, and after the process of mixing salt and fresh water has stopped, the variance decreases again.
The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level. These estimates depend on the minimisation function in the MLMC algorithm.
To achieve the efficiency of the MLMC approach presented in this work, it is essential that the complexity of the numerical solution of each random realisation is proportional to the number of grid vertices on the grid levels.
We investigated the applicability and efficiency of the MLMC approach to the Henry-like problem with uncertain porosity, permeability and recharge. These uncertain parameters were modelled by random fields with three independent random variables. Permeability is a function of porosity. Both functions are time-dependent, have multi-scale behaviour and are defined for two layers. The numerical solution for each random realisation was obtained using the well-known ug4 parallel multigrid solver. The number of random samples required at each level was estimated by calculating the decay of the variances and the computational cost for each level.
The MLMC method was used to compute the expected value and variance of several QoIs, such as the solution at a few preselected points $(t,\bx)$, the solution integrated over a small subdomain, and the time evolution of the freshwater integral. We have found that some QoIs require only 2-3 mesh levels and samples from finer meshes would not significantly improve the result. Other QoIs require more grid levels.
1. Investigated efficiency of MLMC for Henry problem with
uncertain porosity, permeability, and recharge.
2. Uncertainties are modeled by random fields.
3. MLMC could be much faster than MC, 3200 times faster !
4. The time dependence is challenging.
Remarks:
1. Check if MLMC is needed.
2. The optimal number of samples depends on the point (t;x)
3. An advanced MLMC may give better estimates of L and m`.
Density Driven Groundwater Flow with Uncertain Porosity and PermeabilityAlexander Litvinenko
In this work, we solved the density driven groundwater flow problem with uncertain porosity and permeability. An accurate solution of this time-dependent and non-linear problem is impossible because of the presence of natural uncertainties in the reservoir such as porosity and permeability.
Therefore, we estimated the mean value and the variance of the solution, as well as the propagation of uncertainties from the random input parameters to the solution.
We started by defining the Elder-like problem. Then we described the multi-variate polynomial approximation (\gPC) approach and used it to estimate the required statistics of the mass fraction.
Utilizing the \gPC method allowed us
to reduce the computational cost compared to the classical quasi Monte Carlo method.
\gPC assumes that the output function $\sol(t,\bx,\thetab)$ is square-integrable and smooth w.r.t uncertain input variables $\btheta$.
Many factors, such as non-linearity, multiple solutions, multiple stationary states, time dependence and complicated solvers, make the investigation of the convergence of the \gPC method a non-trivial task.
We used an easy-to-implement, but only sub-optimal \gPC technique to quantify the uncertainty. For example, it is known that by increasing the degree of global polynomials (Hermite, Langange and similar), Runge's phenomenon appears. Here, probably local polynomials, splines or their mixtures would be better. Additionally, we used an easy-to-parallelise quadrature rule, which was also only suboptimal. For instance, adaptive choice of sparse grid (or collocation) points \cite{ConradMarzouk13,nobile-sg-mc-2015,Sudret_sparsePCE,CONSTANTINE12,crestaux2009polynomial} would be better, but we were limited by the usage of parallel methods. Adaptive quadrature rules are not (so well) parallelisable. In conclusion, we can report that: a) we developed a highly parallel method to quantify uncertainty in the Elder-like problem; b) with the \gPC of degree 4 we can achieve similar results as with the \QMC method.
In the numerical section we considered two different aquifers - a solid parallelepiped and a solid elliptic cylinder. One of our goals was to see how the domain geometry influences the formation, the number and the shape of fingers.
Since the considered problem is nonlinear,
a high variance in the porosity may result in totally different solutions; for instance, the number of fingers, their intensity and shape, the propagation time, and the velocity may vary considerably.
The number of cells in the presented experiments varied from $241{,}152$ to $15{,}433{,}728$ for the cylindrical domain and from $524{,}288$ to $4{,}194{,}304$ for the parallelepiped. The maximal number of parallel processing units was $600\times 32$, where $600$ is the number of parallel nodes and $32$ is the number of computing cores on each node. The total computing time varied from 2 hours for the coarse mesh to 24 hours for the finest mesh.
Saltwater intrusion occurs when sea levels rise and saltwater moves onto the land. Usually, this occurs during storms, high tides, droughts, or when saltwater penetrates freshwater aquifers and raises the groundwater table. Since groundwater is an essential nutrition and irrigation resource, its salinization may lead to catastrophic consequences. Many acres of farmland may be lost because they can become too wet or salty to grow crops. Therefore, accurate modeling of different scenarios of saline flow is essential to help farmers and researchers develop strategies to improve the soil quality and decrease saltwater intrusion effects.
Saline flow is density-driven and described by a system of time-dependent nonlinear partial differential equations (PDEs). It features convection dominance and can demonstrate very complicated behavior.
As a specific model, we consider a Henry-like problem with uncertain permeability and porosity.
These parameters may strongly affect the flow and transport of salt.
We consider a class of density-driven flow problems. We are particularly interested in the problem of the salinization of coastal aquifers. We consider the Henry saltwater intrusion problem with uncertain porosity, permeability, and recharge parameters as a test case.
The reason for the presence of uncertainties is the lack of knowledge, inaccurate measurements,
and inability to measure parameters at each spatial or time location. This problem is nonlinear and time-dependent. The solution is the salt mass fraction, which is uncertain and changes in time. Uncertainties in porosity, permeability, recharge, and mass fraction are modeled using random fields. This work investigates the applicability of the well-known multilevel Monte Carlo (MLMC) method for such problems. The MLMC method can reduce the total computational and storage costs. Moreover, the MLMC method runs multiple scenarios on different spatial and time meshes and then estimates the mean value of the mass fraction.
The parallelization is performed in both the physical space and stochastic space. To solve every deterministic scenario, we run the parallel multigrid solver ug4 in a black-box fashion.
We use the solution obtained from the quasi-Monte Carlo method as a reference solution.
We investigated the applicability and efficiency of the MLMC approach for the Henry-like problem with uncertain porosity, permeability, and recharge. These uncertain parameters were modeled by random fields with three independent random variables. The numerical solution for each random realization was obtained using the well-known ug4 parallel multigrid solver. The number of required random samples on each level was estimated by computing the decay of the variances and computational costs for each level. We also computed the expected value and variance of the mass fraction in the whole domain, the evolution of the pdfs, the solutions at a few preselected points $(t,\bx)$, and the time evolution of the freshwater integral value. We have found that some QoIs require only 2-3 of the coarsest mesh levels, and samples from finer meshes would not significantly improve the result. Note that a different type of porosity may lead to a different conclusion.
The results show that the MLMC method is faster than the QMC method at the finest mesh. Thus, sampling at different mesh levels makes sense and helps to reduce the overall computational cost.
Here the interest is mainly to compute characterisations like the entropy,
the Kullback-Leibler divergence, more general $f$-divergences, or other such characteristics based on
the probability density. The density is often not available directly,
and it is a computational challenge to just represent it in a numerically
feasible fashion in case the dimension is even moderately large. It
is an even stronger numerical challenge to then actually compute said characteristics
in the high-dimensional case.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.\
$O(d n r^2 )$ for the TT format. Here $n$ is the number of discretisation
points in one direction, $r<<n$ is the maximal tensor rank, and $d$ the problem dimension.
Talk presented on this workshop "Workshop: Imaging With Uncertainty Quantification (IUQ), September 2022",
https://people.compute.dtu.dk/pcha/CUQI/IUQworkshop.html
We consider a weakly supervised classification problem. It
is a classification problem where the target variable can be unknown
or uncertain for some subset of samples. This problem appears when
the labeling is impossible, time-consuming, or expensive. Noisy measurements
and lack of data may prevent accurate labeling. Our task
is to build an optimal classification function. For this, we construct and
minimize a specific objective function, which includes the fitting error on
labeled data and a smoothness term. Next, we use covariance and radial AQ1
basis functions to define the degree of similarity between points. The further
process involves the repeated solution of an extensive linear system
with the graph Laplacian operator. To speed up this solution process,
we introduce low-rank approximation techniques. We call the resulting
algorithm WSC-LR. Then we use the WSC-LR algorithm for analysis
CT brain scans to recognize ischemic stroke disease. We also compare
WSC-LR with other well-known machine learning algorithms.
Computing f-Divergences and Distances of High-Dimensional Probability Density...Alexander Litvinenko
Poster presented on Stochastic Numerics and Statistical Learning: Theory and Applications Workshop in KAUST, Saudi Arabia.
The task considered here was the numerical computation of characterising statistics of
high-dimensional pdfs, as well as their divergences and distances,
where the pdf in the numerical implementation was assumed discretised on some regular grid.
Even for moderate dimension $d$, the full storage and computation with such objects become very quickly infeasible.
We have demonstrated that high-dimensional pdfs,
pcfs, and some functions of them
can be approximated and represented in a low-rank tensor data format.
Utilisation of low-rank tensor techniques helps to reduce the computational complexity
and the storage cost from exponential $\C{O}(n^d)$ to linear in the dimension $d$, e.g.
O(d n r^2) for the TT format. Here $n$ is the number of discretisation
points in one direction, r<n is the maximal tensor rank, and d the problem dimension.
The particular data format is rather unimportant,
any of the well-known tensor formats (CP, Tucker, hierarchical Tucker, tensor-train (TT)) can be used,
and we used the TT data format. Much of the presentation and in fact the central train
of discussion and thought is actually independent of the actual representation.
In the beginning it was motivated through three possible ways how one may
arrive at such a representation of the pdf. One was if the pdf was given in some approximate
analytical form, e.g. like a function tensor product of lower-dimensional pdfs with a
product measure, or from an analogous representation of the pcf and subsequent use of the
Fourier transform, or from a low-rank functional representation of a high-dimensional
RV, again via its pcf.
The theoretical underpinnings of the relation between pdfs and pcfs as well as their
properties were recalled in Section: Theory, as they are important to be preserved in the
discrete approximation. This also introduced the concepts of the convolution and of
the point-wise multiplication Hadamard algebra, concepts which become especially important if
one wants to characterise sums of independent RVs or mixture models,
a topic we did not touch on for the sake of brevity but which follows very naturally from
the developments here. Especially the Hadamard algebra is also
important for the algorithms to compute various point-wise functions in the sparse formats.
Computing f-Divergences and Distances of\\ High-Dimensional Probability Densi...Alexander Litvinenko
Talk presented on SIAM IS 2022 conference.
Very often, in the course of uncertainty quantification tasks or
data analysis, one has to deal with high-dimensional random variables (RVs)
(with values in $\Rd$). Just like any other RV,
a high-dimensional RV can be described by its probability density (\pdf) and/or
by the corresponding probability characteristic functions (\pcf),
or a more general representation as
a function of other, known, random variables.
Here the interest is mainly to compute characterisations like the entropy, the Kullback-Leibler, or more general
$f$-divergences. These are all computed from the \pdf, which is often not available directly,
and it is a computational challenge to even represent it in a numerically
feasible fashion in case the dimension $d$ is even moderately large. It
is an even stronger numerical challenge to then actually compute said characterisations
in the high-dimensional case.
In this regard, in order to achieve a computationally feasible task, we propose
to approximate density by a low-rank tensor.
Low rank tensor approximation of probability density and characteristic funct...Alexander Litvinenko
Very often one has to deal with high-dimensional random variables (RVs). A high-dimensional RV can be described by its probability density (\pdf) and/or by the corresponding probability characteristic functions (\pcf), or by a function representation. Here the interest is mainly to compute characterisations like the entropy, or
relations between two distributions, like their Kullback-Leibler divergence, or more general measures such as $f$-divergences,
among others. These are all computed from the \pdf, which is often not available directly, and it is a computational challenge to even represent it in a numerically feasible fashion in case the dimension $d$ is even moderately large. It is an even stronger numerical challenge to then actually compute said characterisations in the high-dimensional case.
In this regard, in order to achieve a computationally feasible task, we propose to represent the density by a high order tensor product, and approximate this in a low-rank format.
Identification of unknown parameters and prediction of missing values. Compar...Alexander Litvinenko
H-matrix approximation of large Mat\'{e}rn covariance matrices, Gaussian log-likelihoods.
Identifying unknown parameters and making predictions
Comparison with machine learning methods.
kNN is easy to implement and shows promising results.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
We develop fast and efficient stochastic methods for characterizing scattering
from objects of uncertain shapes. This is highly needed in the
fields of electromagnetics, optics, and photonics.
The continuation multilevel Monte Carlo (CMLMC) method is
used together with a surface integral equation solver. The
CMLMC method optimally balances statistical errors due to
sampling of the parametric space, and numerical errors due
to the discretization of the geometry using a hierarchy of
discretizations, from coarse to fine. The number of realizations
of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational
work. Consequently, the total execution time is significantly
reduced, in comparison to the standard MC scheme.
Identification of unknown parameters and prediction with hierarchical matrice...Alexander Litvinenko
We compare four numerical methods for the prediction of missing values in four different datasets.
These methods are 1) the hierarchical maximum likelihood estimation (H-MLE), and three machine learning (ML) methods, which include 2) k-nearest neighbors (kNN), 3) random forest, and 4) Deep Neural Network (DNN).
From the ML methods, the best results (for considered datasets) were obtained by the kNN method with three (or seven) neighbors.
On one dataset, the MLE method showed a smaller error than the kNN method, whereas, on another, the kNN method was better.
The MLE method requires a lot of linear algebra computations and works fine on almost all datasets. Its result can be improved by taking a smaller threshold and more accurate hierarchical matrix arithmetics. To our surprise, the well-known kNN method produces similar results as H-MLE and worked much faster.
1. Motivation: why do we need low-rank tensors
2. Tensors of the second order (matrices)
3. CP, Tucker and tensor train tensor formats
4. Many classical kernels have (or can be approximated in ) low-rank tensor format
5. Post processing: Computation of mean, variance, level sets, frequency
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Computation of electromagnetic fields scattered from dielectric objects of un...Alexander Litvinenko
Computational tools for characterizing electromagnetic scattering from objects with uncertain shapes are needed in various applications ranging from remote sensing at microwave frequencies to Raman spectroscopy at optical frequencies. Often, such computational tools use the Monte Carlo (MC) method to sample a parametric space describing geometric uncertainties. For each sample, which corresponds to a realization of the geometry, a deterministic electromagnetic solver computes the scattered fields. However, for an accurate statistical characterization the number of MC samples has to be large. In this work, to address this challenge, the continuation multilevel Monte Carlo (\CMLMC) method is used together with a surface integral equation solver.
The \CMLMC method optimally balances statistical errors due to sampling of
the parametric space, and numerical errors due to the discretization of the geometry using a hierarchy of discretizations, from coarse to fine.
The number of realizations of finer discretizations can be kept low, with most samples
computed on coarser discretizations to minimize computational cost.
Consequently, the total execution time is significantly reduced, in comparison to the standard MC scheme.
Propagation of Uncertainties in Density Driven Groundwater FlowAlexander Litvinenko
Major Goal: estimate risks of the pollution in a subsurface flow.
How?: we solve density-driven groundwater flow with uncertain porosity and permeability.
We set up density-driven groundwater flow problem,
review stochastic modeling and stochastic methods, use UG4 framework (https://gcsc.uni-frankfurt.de/simulation-and-modelling/ug4),
model uncertainty in porosity and permeability,
2D and 3D numerical experiments.
The French Revolution, which began in 1789, was a period of radical social and political upheaval in France. It marked the decline of absolute monarchies, the rise of secular and democratic republics, and the eventual rise of Napoleon Bonaparte. This revolutionary period is crucial in understanding the transition from feudalism to modernity in Europe.
For more information, visit-www.vavaclasses.com
Introduction to AI for Nonprofits with Tapp NetworkTechSoup
Dive into the world of AI! Experts Jon Hill and Tareq Monaur will guide you through AI's role in enhancing nonprofit websites and basic marketing strategies, making it easy to understand and apply.
Acetabularia Information For Class 9 .docxvaibhavrinwa19
Acetabularia acetabulum is a single-celled green alga that in its vegetative state is morphologically differentiated into a basal rhizoid and an axially elongated stalk, which bears whorls of branching hairs. The single diploid nucleus resides in the rhizoid.
Model Attribute Check Company Auto PropertyCeline George
In Odoo, the multi-company feature allows you to manage multiple companies within a single Odoo database instance. Each company can have its own configurations while still sharing common resources such as products, customers, and suppliers.
2024.06.01 Introducing a competency framework for languag learning materials ...Sandy Millin
http://sandymillin.wordpress.com/iateflwebinar2024
Published classroom materials form the basis of syllabuses, drive teacher professional development, and have a potentially huge influence on learners, teachers and education systems. All teachers also create their own materials, whether a few sentences on a blackboard, a highly-structured fully-realised online course, or anything in between. Despite this, the knowledge and skills needed to create effective language learning materials are rarely part of teacher training, and are mostly learnt by trial and error.
Knowledge and skills frameworks, generally called competency frameworks, for ELT teachers, trainers and managers have existed for a few years now. However, until I created one for my MA dissertation, there wasn’t one drawing together what we need to know and do to be able to effectively produce language learning materials.
This webinar will introduce you to my framework, highlighting the key competencies I identified from my research. It will also show how anybody involved in language teaching (any language, not just English!), teacher training, managing schools or developing language learning materials can benefit from using the framework.
Palestine last event orientationfvgnh .pptxRaedMohamed3
An EFL lesson about the current events in Palestine. It is intended to be for intermediate students who wish to increase their listening skills through a short lesson in power point.
Read| The latest issue of The Challenger is here! We are thrilled to announce that our school paper has qualified for the NATIONAL SCHOOLS PRESS CONFERENCE (NSPC) 2024. Thank you for your unwavering support and trust. Dive into the stories that made us stand out!
Macroeconomics- Movie Location
This will be used as part of your Personal Professional Portfolio once graded.
Objective:
Prepare a presentation or a paper using research, basic comparative analysis, data organization and application of economic information. You will make an informed assessment of an economic climate outside of the United States to accomplish an entertainment industry objective.
The Roman Empire A Historical Colossus.pdfkaushalkr1407
The Roman Empire, a vast and enduring power, stands as one of history's most remarkable civilizations, leaving an indelible imprint on the world. It emerged from the Roman Republic, transitioning into an imperial powerhouse under the leadership of Augustus Caesar in 27 BCE. This transformation marked the beginning of an era defined by unprecedented territorial expansion, architectural marvels, and profound cultural influence.
The empire's roots lie in the city of Rome, founded, according to legend, by Romulus in 753 BCE. Over centuries, Rome evolved from a small settlement to a formidable republic, characterized by a complex political system with elected officials and checks on power. However, internal strife, class conflicts, and military ambitions paved the way for the end of the Republic. Julius Caesar’s dictatorship and subsequent assassination in 44 BCE created a power vacuum, leading to a civil war. Octavian, later Augustus, emerged victorious, heralding the Roman Empire’s birth.
Under Augustus, the empire experienced the Pax Romana, a 200-year period of relative peace and stability. Augustus reformed the military, established efficient administrative systems, and initiated grand construction projects. The empire's borders expanded, encompassing territories from Britain to Egypt and from Spain to the Euphrates. Roman legions, renowned for their discipline and engineering prowess, secured and maintained these vast territories, building roads, fortifications, and cities that facilitated control and integration.
The Roman Empire’s society was hierarchical, with a rigid class system. At the top were the patricians, wealthy elites who held significant political power. Below them were the plebeians, free citizens with limited political influence, and the vast numbers of slaves who formed the backbone of the economy. The family unit was central, governed by the paterfamilias, the male head who held absolute authority.
Culturally, the Romans were eclectic, absorbing and adapting elements from the civilizations they encountered, particularly the Greeks. Roman art, literature, and philosophy reflected this synthesis, creating a rich cultural tapestry. Latin, the Roman language, became the lingua franca of the Western world, influencing numerous modern languages.
Roman architecture and engineering achievements were monumental. They perfected the arch, vault, and dome, constructing enduring structures like the Colosseum, Pantheon, and aqueducts. These engineering marvels not only showcased Roman ingenuity but also served practical purposes, from public entertainment to water supply.
Thesis Statement for students diagnonsed withADHD.ppt
Tensor train to solve stochastic PDEs
1. Numerical methods for solving stochastic partial
differential equations in the Tensor Train format
Alexander Litvinenko1
(joint work with Sergey Dolgov2,3, Boris Khoromskij3 and
Hermann G. Matthies4)
1 SRI UQ and Extreme Computing Research Center KAUST, 2
Max-Planck-Institut f¨ur Mathematik in den Naturwissenschaften,
Leipzig, MPI for dynamics of complex systems in 3 Magdeburg,
4 TU Braunschweig, Germany
Center for Uncertainty
Quantification
ntification Logo Lock-up
http://sri-uq.kaust.edu.sa/
2. 4*
Motivation for UQ
Nowadays computational algorithms, run on
supercomputers, can simulate and resolve very
complex phenomena. But how reliable are these
predictions? Can we trust to these results?
Some parameters/coefficients are unknown, lack of
data, very few measurements → uncertainty.
Center for Uncertainty
Quantification
tion Logo Lock-up
2 / 58
3. 4*
Notation, problem setup
Consider
A(u; q) = f ⇒ u = S(f ; q),
where S is a solution operator.
Uncertain Input:
1. Parameter q := q(ω) (assume moments/cdf/pdf/quantiles of
q are given)
2. Boundary and initial conditions, right-hand side
3. Geometry of the domain
Uncertain solution:
1. mean value and variance of u
2. exceedance probabilities P(u > u∗)
3. probability density functions (pdf) of u.
Center for Uncertainty
Quantification
tion Logo Lock-up
3 / 58
4. 4*
KAUST
Figure : KAUST campus, 5 years
old, approx. 7000 people (include
1400 kids), 100 nations.
Center for Uncertainty
Quantification
tion Logo Lock-up
4 / 58
6. 4*
Stochastic Numerics Group at KAUST
Figure : SRI UQ Group
Center for Uncertainty
Quantification
tion Logo Lock-up
6 / 58
7. 4*
3rd UQ Workshop ”Advances in UQ Methods, Alg. & Appl.”
Center for Uncertainty
Quantification
tion Logo Lock-up
7 / 58
8. 4*
PDE with uncertain diffusion coefficients
PART 1. Stochastic Forward Problems
Center for Uncertainty
Quantification
tion Logo Lock-up
8 / 58
9. 4*
PDE with uncertain diffusion coefficients
Consider
− div(κ(x, ω) u(x, ω)) = f (x, ω) in G × Ω, G ⊂ R2,
u = 0 on ∂G,
(1)
where κ(x, ω) - uncertain diffusion coefficient. Since κ positive,
usually κ(x, ω) = eγ(x,ω).
For well-posedness see [Sarkis 09, Gittelson 10, H.J.Starkloff 11,
Ullmann 10].
Further we will assume that covκ(x, y) is given (or estimated from
the available data).
Center for Uncertainty
Quantification
tion Logo Lock-up
9 / 58
10. 4*
Our previous work
After applying the stochastic Galerkin method, obtain:
Ku = f, where all ingredients are represented in a tensor format
Solve for u. Compute max{u}, var(u), level sets of u, pdf, cdf,
1. Efficient Analysis of High Dimensional Data in Tensor Formats,
[Espig, Hackbusch, A.L., Matthies and Zander, 2012]
Research rank of K (from which ingredients it depends)
2. Efficient low-rank approximation of the stochastic Galerkin
matrix in tensor formats, [W¨ahnert, Espig, Hackbusch, A.L., Matthies, 2013]
Center for Uncertainty
Quantification
tion Logo Lock-up
10 / 58
11. 4*
Smooth transformation of Gaussian RF
Step 1: We assume κ = φ(γ) -a smooth transformation of the
Gaussian random field γ(x, ω), e.g. φ(γ) = exp(γ).
[see PhD of E. Zander 2013, or PhD of A. Keese, 2005]
Step 2: Given the covariance matrix of κ(x, ω), we derive the
covariance matrix of γ(x, ω). After that the KLE may be
computed,
γ(x, ω) =
∞
m=1
gm(x)θm(ω),
D
covγ(x, y)gm(y)dy = λmgm(x),
(2)
Center for Uncertainty
Quantification
tion Logo Lock-up
11 / 58
12. 4*
Full JM,p and sparse J sp
M,p multi-index sets
M-dimensional PCE approximation of κ writes (α = (α1, ..., αM))
κ(x, ω) ≈
α∈JM
κα(x)Hα(θ(ω)), Hα(θ) := hα1 (θ1) · · · hαM
(θM)
(3)
Definition
The full multi-index is defined by restricting each component
independently,
JM,p = {0, 1, . . . , p1}⊗· · ·⊗{0, 1, . . . , pM}, where p = (p1, . . . , pM)
is a shortcut for the tuple of order limits.
Definition
The sparse multi-index is defined by restricting the sum of
components,
J sp
M,p = {α = (α1, . . . , αM) : α ≥ 0, α1 + · · · + αM ≤ p} .
Center for Uncertainty
Quantification
tion Logo Lock-up
12 / 58
13. 4*
TT compression of PCE coeffs
The Galerkin coefficients κα are evaluated as follows [Thm 3.10,
PhD of E. Zander 13],
κα(x) =
(α1 + · · · + αM)!
α1! · · · αM!
φα1+···+αM
M
m=1
gαm
m (x), (4)
where φ|α| := φα1+···+αM
is the Galerkin coefficient of the
transform function, and gαm
m (x) means just the αm-th power of the
KLE function value gm(x).
Center for Uncertainty
Quantification
tion Logo Lock-up
13 / 58
14. 4*
Complexity reduction
Complexity reduction in Eq. (4) can be achieved with help of KLE
of κ(x, ω):
κ(x, ω) ≈ ¯κ(x) +
L
=1
√
µ v (x)η (ω) (5)
with the normalized spatial functions v (x).
Instead of using κα(x), (4), directly, we compute
˜κ (α) =
(α1 + · · · + αM)!
α1! · · · αM!
φα1+···+αM
D
M
m=1
gαm
m (x)v (x)dx.
Note that L N. Then we restore the approximate coefficients
κα(x) ≈ ¯κ(x) +
L
=1
v (x)˜κ (α).
Center for Uncertainty
Quantification
tion Logo Lock-up
14 / 58
15. 4*
Construction of the stochastic Galerkin operator
Given KLE of κ, assemble for i, j = 1, . . . , N, = 1, . . . , L:
K0(i, j) =
D
¯κ(x) ϕi (x)· ϕj (x)dx, K (i, j) =
D
v (x) ϕi (x)· ϕj (x)d
(6)
K
(ω)
(α, β) =
RM
Hα(θ)Hβ(θ)
ν∈JM
˜κ (ν)Hν(θ)ρ(θ)dθ
=
ν∈JM
∆α,β,ν ˜κ (ν),
∆α,β,ν = ∆α1,β1,ν1 · · · ∆αM ,βM ,νM
,
∆αm,βm,νm =
R
hαm (θ)hβm (θ)hνm (θ)ρ(θ)dθ,
Center for Uncertainty
Quantification
tion Logo Lock-up
15 / 58
16. 4*
Stochastic Galerkin operator
Putting together previous formulas, obtain the stochastic Galerkin
operator,
K = K
(x)
0 ⊗ ∆0 +
L
=1
K
(x)
⊗ K
(ω)
, (7)
with K ∈ RN(p+1)M ×N(p+1)M
in case of full JM,p.
IDEA: If PCE coefficients of κ are computed in the tensor product
format, the direct product in ∆ (15) allows to exploit the same
format for (7), and build the operator easily.
Center for Uncertainty
Quantification
tion Logo Lock-up
16 / 58
17. 4*
Tensor Train
Two tensor Train examples
Center for Uncertainty
Quantification
tion Logo Lock-up
17 / 58
19. 4*
Examples:
TT rank(f )=2
f = sin(x1 + x2 + ... + xd )
= (sin x1, cos x1)
cos x2 − sin x2
sin x2 cos x2
...
cos xd−1 − sin xd−1
sin xd−1 cos xd−1
cos xd
sin xd−1
Center for Uncertainty
Quantification
tion Logo Lock-up
19 / 58
20. 4*
Low-rank response surface: PCE in the TT format
Calculation of
˜κ (α) =
(α1 + · · · + αM)!
α1! · · · αM!
φα1+···+αM
D
M
m=1
gαm
m (x)v (x)dx.
in TT format needs:
a procedure to compute each element of a tensor, e.g.
˜κα1,...,αM
.
build a TT approximation ˜κα ≈ κ(1)(α1) · · · κ(M)(αM) using a
feasible amount of elements (i.e. much less than (p + 1)M).
Such procedure exists, and relies on the cross interpolation of
matrices, generalized to a higher-dimensional case [Oseledets, Tyrtyshnikov
2010; Savostyanov 13; Grasedyck; Bebendorf].
Center for Uncertainty
Quantification
tion Logo Lock-up
20 / 58
21. PCE coefficients ˜κ (α) are :
˜κ (α) =
s1,...,sM−1
κ
(1)
,s1
(α1)κ
(2)
s1,s2 (α2) · · · κ
(M)
sM−1 (αM). (8)
Collect the spatial components into the “zeroth” TT block,
κ(0)
(x) = κ
(0)
(x)
L
=0
= ¯κ(x) v1(x) · · · vL(x) , (9)
then the PCE writes as the following TT format,
κ(x, α) =
,s1,...,sM−1
κ
(0)
(x)κ
(1)
,s1
(α1) · · · κ
(M)
sM−1 (αM). (10)
Center for Uncertainty
Quantification
tion Logo Lock-up
21 / 58
22. 4*
Stochastic Galerkin matrix in TT format
Given κα(x), (10), we split the whole sum over ν in K,(7):
ν∈JM,p
∆α,β,ν ˜κ (ν) =
s1,...,sM−1
K
(1)
,s1
(α1, β1)K
(2)
s1,s2 (α2, β2) · · · K
(M)
sM−1 (αM,
K(m)
(αm, βm) =
pm
νm=0
∆αm,βm,νm κ(m)
(νm), m = 1, . . . , M.
(11)
then the TT representation for the operator writes
K =
,s1,...,sM−1
K
(0)
⊗K
(1)
,s1
⊗· · ·⊗K
(M)
sM−1 ∈ R(N·#JM,p)×(N·#JM,p)
,
(12)
Center for Uncertainty
Quantification
tion Logo Lock-up
22 / 58
23. 4*
Solving and Post-processing:
Solve the linear system Ku = f by alternating optimization
methods [Dolgov, Savostyanov 14] with a mean-field
preconditioned. Obtain the solution u in the TT format.
u(x, α) =
s0,...,sM−1
u
(0)
s0 (x)u
(1)
s0,s1 (α1) · · · u
(M)
sM−1 (αM). (13)
u(x, θ) =
s0,...,sM−1
u
(0)
s0 (x)
p
α1=0
hα1 (θ1)u
(1)
s0,s1 (α1) · · · (14)
p
αM =0
hαM
(θM)u
(M)
sM−1 (αM)
. (15)
Then compute: mean, co(variance), exceedance probabilities
Center for Uncertainty
Quantification
tion Logo Lock-up
23 / 58
24. 4*
Numerics: Main steps
1. Use sglib (E. Zander, TU BS) for discretization and solution
with J sp
M,p.
2. Compute PCE (sglib) of the coefficients κ(x, ω) in the TT
format by new block adaptive cross algorithm (TT toolbox)
3. Use TT-Toolbox for full JM,p,
4. amen cross.m for TT approximation of ˜κα,
5. Compute stochastic Galerkin matrix K in TT format,
6. Replace high-dimensional calculations by the TT-toolbox.
7. Compute solution of the linear system in TT (alternating
minimal energy, tAMEn)
8. Post-processing in TT format
Center for Uncertainty
Quantification
tion Logo Lock-up
24 / 58
25. 4*
Numerical experiments, errors, accuracy
D = [−1, 1]2[0, 1]2. f = f (x) = 1, log-normal and beta
distributions for κ. 557, 2145, 8417 dofs.
Eκ =
1
Nmc
Nmc
z=1
N
i=1 (κ(xi , θz) − κ∗(xi , θz))2
N
i=1 κ2
∗(xi , θz)
where {θz}Nmc
z=1 are normally distributed random samples and
κ∗(xi , θz) = φ (γ(xi , θz)) is the reference coefficient computed
without using the PCE for φ.
E¯u =
¯u − ¯u∗ L2(D)
¯u∗ L2(D)
, Evaru =
varu − varu∗ L2(D)
varu∗ L2(D)
.
Center for Uncertainty
Quantification
tion Logo Lock-up
25 / 58
26. 4*
More numerics
We compute the maximizer of the mean solution,
xmax : ¯u(xmax) ≥ ¯u(x) ∀x ∈ D.
umax(θ) = u(xmax, θ), and ˆu = ¯u(xmax).
Taking some τ > 1, we compute
P = P (umax(θ) > τˆu) =
RM
χumax(θ)>τˆu(θ)ρ(θ)dθ. (16)
By P∗ we will also denote the probability computed from the
Monte Carlo method, and estimate the error as EP = |P − P∗| /P∗.
Center for Uncertainty
Quantification
tion Logo Lock-up
26 / 58
27. 4*
Sparse J sp
M,p or full JM,p ?
What is better sparse J sp
M,p or full JM,p multi-index set ?
Center for Uncertainty
Quantification
tion Logo Lock-up
27 / 58
28. 4*
CPU times (sec.) versus p, log-normal distribution
TT (full index set JM,p) Sparse (index set J sp
M,p)
p Tκ Top Tu Tκ Top Tu
1 9.6 0.2 1.7 0.5 0.3 0.65
2 14.7 0.2 3 0.5 3.2 1.4
3 19.1 0.2 3.4 0.7 1028 18
4 24.4 0.2 4.2 2.2 — —
5 30.9 0.32 5.3 9.8 — —
Center for Uncertainty
Quantification
tion Logo Lock-up
28 / 58
29. 4*
How does polynomial order influence the ranks ?
How does the max. polynomial order p influence the ranks ?
Center for Uncertainty
Quantification
tion Logo Lock-up
29 / 58
30. 4*
Performance versus p, log-normal distribution
p CPU time, sec. rκ ru r ˆχ Eκ Eu P
TT Sparse ˆχ TT Sparse TT Sparse TT
1 11 1.4 0.2 32 42 1 4e-3 1.7e-1 1e-2 1e-1 0
2 18 5.1 0.3 32 49 1 1e-4 1.1e-1 5e-4 5e-2 0
3 23 1046 83 32 49 462 6e-5 2.e-3 3e-4 5e-4 2.8e-4
4 29 — 70 32 50 416 6e-5 — 1e-4 — 1.2e-4
5 37 — 103 32 49 410 6e-5 — 1e-4 — 6.2e-4
Take τ = 1.2:
P = P (umax(θ) > τˆu) =
RM
χumax(θ)>τˆu(θ)ρ(θ)dθ. (17)
Center for Uncertainty
Quantification
tion Logo Lock-up
30 / 58
31. 4*
How does stochastic dimension M influence the ranks ?
How does the stochastic dimension M influence the TT ranks ?
Center for Uncertainty
Quantification
tion Logo Lock-up
31 / 58
32. 4*
Performance versus M, log-normal distribution
M CPU time, sec. rκ ru r ˆχ Eκ Eu P
TT Sparse ˆχ TT Sparse TT Sparse TT
10 6 6 1.3 20 39 70 2e-4 1.7e-1 3e-4 1.5e-1 2.86e-4
15 12 92 23 27 42 381 8e-5 2e-3 3e-4 5e-4 3e-4
20 22 1e+3 67 32 50 422 6e-5 2e-3 3e-4 5e-4 2.96e-4
30 53 5e+4 137 39 50 452 6e-5 1e-1 3e-4 5.5e-2 2.78e-4
Center for Uncertainty
Quantification
tion Logo Lock-up
32 / 58
33. 4*
How does covariance length influence the ranks ?
How does covariance length influence the ranks ?
Center for Uncertainty
Quantification
tion Logo Lock-up
33 / 58
34. 4*
Performance versus cov. length, log-normal distribution
cov. CPU time, sec. rκ ru r ˆχ Eκ Eu P
length TT Sparse ˆχ TT Sparse TT Sparse TT
0.1 216 55800 0.9 70 50 1 2e-2 2e-2 1.8e-2 1.8e-2 0
0.3 317 52360 42 87 74 297 3e-3 3.5e-3 2.6e-3 2.6e-3 8e-31
0.5 195 51700 58 67 74 375 1.5e-4 2e-3 2.6e-4 3.1e-4 6e-33
1.0 57.3 55200 97 39 50 417 6.1e-5 9e-2 3.2e-4 5.6e-2 2.95e-04
1.5 32.4 49800 121 31 34 424 3.2e-5 2e-1 5e-4 1.7e-1 7.5e-04
Center for Uncertainty
Quantification
tion Logo Lock-up
34 / 58
35. 4*
How does standard deviation σ influence the ranks ?
How does the standard deviation σ influence the TT ranks ?
Center for Uncertainty
Quantification
tion Logo Lock-up
35 / 58
37. 4*
How does number of DoFs influence the ranks ?
How does number of DoFs influence the ranks ?
Center for Uncertainty
Quantification
tion Logo Lock-up
37 / 58
38. 4*
Performance versus #DoFs, log-normal distribution
#DoFs CPU time, sec rκ ru r ˆχ Eκ Eu P
TT Sparse ˆχ TT Sparse TT Sparse TT
557 6 6 1.3 20 39 71 2e-4 1.7e-1 3e-4 1.5e-1 2.86e-4
2145 9 14 1.2 20 39 76 2e-4 2e-3 3e-4 5.7e-4 2.9e-4
8417 357 171 0.8 20 40 69 1.7e-4 2e-3 3e-4 5.6e-4 2.93e-4
Center for Uncertainty
Quantification
tion Logo Lock-up
38 / 58
39. 4*
Comparison with the Monte Carlo
Comparison of the solution obtained via the (Stochastic Galerkin
+ TT) with the solution obtained via Monte Carlo (4000).
For the Monte Carlo test, we prepare the TT solution with
parameters p = 5 and M = 30.
Center for Uncertainty
Quantification
tion Logo Lock-up
39 / 58
40. 4*
Verification of the MC method (4000), log-normal distr.
Nmc TMC , sec. E¯u Evaru
P∗ EP TT results
102
0.6 9e-3 2e-1 0 ∞ Tsolve 97 sec.
103
6.2 2e-3 6e-2 0 ∞ Tˆχ 157 sec.
104
6.2·101
6e-4 7e-3 4e-4 5e-1 rκ 39
105
6.2·102
3e-4 3e-3 4e-4 5e-1 ru 50
106
6.3·103
1e-4 1e-3 5e-4 4e-1 rˆχ 432
P 6e-4
Center for Uncertainty
Quantification
tion Logo Lock-up
40 / 58
41. 4*
Part II: diffusion coefficient has beta distrib.
κ(x, ω) = B−1
5,2
1 + erf γ(x,ω)
√
2
2
+ 1,
Ba,b(z) =
1
B(a, b)
z
0
ta−1
(1 − t)b−1
dt.
Center for Uncertainty
Quantification
tion Logo Lock-up
41 / 58
42. 4*
We researched (for beta distribution)
1. Performance versus p
2. Performance versus stochastic dimension M
3. Performance versus cov. length
4. Performance versus #DoFs
5. Verification of the Monte Carlo method
Center for Uncertainty
Quantification
tion Logo Lock-up
42 / 58
43. 4*
Take to home
1. TT methods become preferable for high p, but otherwise the
full computation in a small sparse set may be incredibly fast.
This reflects well the “curse of order”, taking place for the
sparse set instead of the “curse of dimensionality” in the full
set: the cardinality of the sparse set grows exponentially with
p.
2. The TT approach scales linearly with p.
3. TT methods allow easy calculation of stochastic Galerkin
operator. With p < 10 TT storage of stoch. Galerkin operator
allows us to forget about the sparsity issues, since the number
of TT entries O(Mp2r2) is tractable.
4. Chebyshev, Laguerre, ... may be incorporated into the scheme
freely.
Center for Uncertainty
Quantification
tion Logo Lock-up
43 / 58
44. 4*
Future plans for the next article
1. Compute Sobol indices in the TT format. Which uncertain
coefficients and which PCE terms are important ?
2. Solution of this linear elliptic SPDE is a ”working horse” for
the non-linear equation and the Newton method
3. Stochastic Galerkin in TT format above can be used as a
preconditioning it is very fast!) for more complicated
non-linear problems
4. Apply to more complicated diffusion coefficients (e.g. which
are not so good splittable)
5. To create analytic u, compute analytically RHS and solve the
problem again (to avoid to use MC as a reference)
Center for Uncertainty
Quantification
tion Logo Lock-up
44 / 58
45. 4*
Approximate Bayesian Update
PART 2. Inverse Problems via approximate
Bayesian Update
Center for Uncertainty
Quantification
tion Logo Lock-up
45 / 58
46. 4*
Setting for the identification process
General idea:
We observe / measure a system, whose structure we know in
principle.
The system behaviour depends on some quantities (parameters),
which we do not know ⇒ uncertainty.
We model (uncertainty in) our knowledge in a Bayesian setting:
as a probability distribution on the parameters.
We start with what we know a priori, then perform a measurement.
This gives new information, to update our knowledge
(identification).
Update in probabilistic setting works with conditional probabilities
⇒ Bayes’s theorem.
Repeated measurements lead to better identification.
Center for Uncertainty
Quantification
tion Logo Lock-up
46 / 58
47. 4*
Mathematical setup
Consider
A(u; q) = f ⇒ u = S(f ; q),
where S is solution operator.
Operator depends on parameters q ∈ Q,
hence state u ∈ U is also function of q:
Measurement operator Y with values in Y:
y = Y (q; u) = Y (q, S(f ; q)).
Examples of measurements:
(ODE) u(t) = (x(t), y(t), z(t))T , y(t) = (x(t), y(t))T
(PDE) y(ω) = D0
u(ω, x)dx, y(ω) = D0
| grad u(ω, x)|2dx, u in
few points
Center for Uncertainty
Quantification
tion Logo Lock-up
47 / 58
48. 4*
Inverse problem
For given f , measurement y is just a function of q.
This function is usually not invertible ⇒ ill-posed problem,
measurement y does not contain enough information.
In Bayesian framework state of knowledge modelled in a
probabilistic way,
parameters q are uncertain, and assumed as random.
Bayesian setting allows updating / sharpening of information
about q when measurement is performed.
The problem of updating distribution —state of knowledge of q
becomes well-posed.
Can be applied successively, each new measurement y and
forcing f —may also be uncertain—will provide new information.
Center for Uncertainty
Quantification
tion Logo Lock-up
48 / 58
49. 4*
Conditional probability and expectation
With state u ∈ U ⊗ S a RV, the quantity to be measured
y(ω) = Y (q(ω), u(ω))) ∈ Y ⊗ S
is also uncertain, a random variable.
A new measurement z is performed, composed from the
“true” value y ∈ Y and a random error : z(ω) = y + (ω).
Classically, Bayes’s theorem gives conditional probability
P(Iq|Mz) =
P(Mz|Iq)
P(Mz)
P(Iq);
expectation with this posterior measure is conditional expectation.
Kolmogorov starts from conditional expectation E (·|Mz),
from this conditional probability via P(Iq|Mz) = E χIq |Mz .
Center for Uncertainty
Quantification
tion Logo Lock-up
49 / 58
50. 4*
IDEA of the Bayesian Update (BU)
Let Y (x, θ), θ = (θ1, ..., θM, ...), is approximated:
Y (x, θ) =
β∈Jm,p
Hβ(θ)Yβ(x),
q(x, θ) =
β∈Jm,p
Hβ(θ)qβ(x),
Yβ(x) =
1
β! Θ
Hβ(θ)Y (x, θ) P(dθ).
Take qf (ω) = q0(ω).
Linear BU: qa = qf + K · (z − y)
Non-Linear BU: qa = qf + H1 · (z − y) + (z − y)T · H2 · (z − y).
Center for Uncertainty
Quantification
tion Logo Lock-up
50 / 58
52. 4*
Open questions
Multivariate Cauchy distribution
The characteristic function ϕX(t) of the multivariate Cauchy
distribution is defined as follow:
ϕX(t) = exp i(t1, t2) · (µ1, µ2)T
−
1
2
(t1, t2)
σ2
1 0
0 σ2
2
(t1, t2)T ,
(18)
ϕX(t) ≈
R
ν=1
ϕXν,1 (t1) · ϕXν,2 (t2). (19)
Again, from the inversion theorem, the probability density of X on
R2 can be computed from ϕX(t) as follow
Center for Uncertainty
Quantification
tion Logo Lock-up
52 / 58
53. pX(y) =
1
(2π)2
R2
exp(−i y, t )ϕX(t)dt (20)
≈
1
(2π)2
R2
exp(−i(y1t1 + y2t2))
R
ν=1
ϕXν,1 (t1) · ϕXν,2 (t2)dt1dt2
(21)
≈
R
ν=1
1
(2π) R
exp(−iy1t1)ϕXν,1 (t1)dt1 ·
1
(2π) R
exp(−iy2t2)ϕXν,2 (t
(22)
≈
R
ν=1
pXν,1 (y1) · pXν,2 (y2), i.e. (23)
the probability density pX(y) is numerically splittable.
Center for Uncertainty
Quantification
tion Logo Lock-up
53 / 58
54. 4*
Elliptically contoured multivariate stable distribution
ϕX(t) = exp i(t1, t2) · (µ1, µ2)T
− (t1, t2)
σ2
1 0
0 σ2
2
(t1, t2)T
α/2
(24)
Now the question is to find a separation of
(t1, t2)
σ2
1 0
0 σ2
2
(t1, t2)T
α/2
≈
R
ν=1
φν,1(t1) · φν,2(t2), (25)
with some tensor rank R.
Center for Uncertainty
Quantification
tion Logo Lock-up
54 / 58
55. 4*
Multivariate distribution
Assume that the characteristic function ϕX(t) of some multivariate
d-dimensional distribution is approximated as follow:
ϕX(t) ≈
R
=1
d
µ=1
ϕX ,µ
(tµ). (26)
pX(y) = const
Rd
exp(−i y, t )ϕX(t)dt (27)
≈ const
Rd
exp(−i
d
j=1
yj tj )
R
=1
d
µ=1
ϕX ,µ
(tµ)dt1...dtd (28)
≈
R
=1
const
d
µ=1 R
exp(−iy t )ϕX ,µ
(tµ)dt · (29)
≈
R
=1
d
µ=1
pX ,µ
(yµ). (30)
Center for Uncertainty
Quantification
tion Logo Lock-up
55 / 58
56. 4*
Actual computation of ϕX(t)
ϕX(τβ) = E (exp(i X(θ1, ..., θm), τβ ))
= · · ·
Θ
exp(i X(θ1, ..., θm), τβ )
M
m=1
pθm (θm)dθ1...dθM,
X(ω), τβ =
α∈J
ξα
Hα(θ), τβ ≈
d
=1 α∈J
ξα
Hα(θ)tβ ,
=
α∈J
d
=1
ξα
tβ , Hα(θ) =
α∈J
ξα
, τβ Hα(θ) (31)
Now compute the exp() function from the scalar product
Center for Uncertainty
Quantification
tion Logo Lock-up
56 / 58
57. exp(i X(ω), τβ ) = exp(i
α∈J
ξα
, τβ Hα(θ)) (32)
=
α∈J
exp (i ξα
, τβ Hα(θ)) (33)
Now we apply integration:
ϕX(t) = E (exp(i X(ω), τβ ))
= · · ·
Θ α∈J
exp (i ξα
, τβ Hα(θ))
M
m=1
pθm (θm)dθ1...dθM
≈????
nq
=1
w
α∈J
exp (i ξα
, τβ Hα(θ ))
M
m=1
pθm (θm, )
Center for Uncertainty
Quantification
tion Logo Lock-up
57 / 58
58. 4*
Literature
1. Polynomial Chaos Expansion of random coefficients and the
solution of stochastic partial differential equations in the Tensor
Train format, S. Dolgov, B. N. Khoromskij, A. Litvinenko, H. G.
Matthies, 2015/3/11, arXiv:1503.03210
2. Efficient analysis of high dimensional data in tensor formats, M.
Espig, W. Hackbusch, A. Litvinenko, H.G. Matthies, E. Zander
Sparse Grids and Applications, 31-56, 40, 2013
3. Application of hierarchical matrices for computing the
Karhunen-Loeve expansion, B.N. Khoromskij, A. Litvinenko, H.G.
Matthies, Computing 84 (1-2), 49-67, 31, 2009
4. Efficient low-rank approximation of the stochastic Galerkin
matrix in tensor formats, M. Espig, W. Hackbusch, A. Litvinenko,
H.G. Matthies, P. Waehnert, Computers and Mathematics with
Applications 67 (4), 818-829
Center for Uncertainty
Quantification
tion Logo Lock-up
58 / 58