SlideShare a Scribd company logo
1 of 90
Download to read offline
Bayesian Subset Simulation

— a kriging-based subset simulation algorithm for
the estimation of small probabilities of failure —

       Ling Li, Julien Bect, Emmanuel Vazquez



                  Supelec, France

                PSAM11-ESREL12
               Helsinki, June 26, 2012
A classical problem in (probabilistic) reliability. . .       (1/2)

     ❍ Consider a system subject to uncertainties,
          ◮   aleatory and/or epistemic,
          ◮   represented by a random vector X ∼ PX
              where PX is a probability measure on X ⊂ Rd .
A classical problem in (probabilistic) reliability. . .          (1/2)

     ❍ Consider a system subject to uncertainties,
          ◮   aleatory and/or epistemic,
          ◮   represented by a random vector X ∼ PX
              where PX is a probability measure on X ⊂ Rd .


     ❍ Assume that the system fails when f (X ) > u
          ◮   f : X → R is a cost function,
          ◮   u ∈ R is the critical level.


     ❍ x → u − f (x ) is sometimes called the “limit state function”
A classical problem in (probabilistic) reliability. . .                 (2/2)

                                                           α
   ❍ Define the failure region




                                         PX
       Γ = {x ∈ X : f (x ) > u}.

                                                                              u




                                         f (x )
                                                           Γ
   ❍ The probability of failure is

      α = PX {Γ} =           1f >u dPX                         x
                         X
                                                  Figure: A 1d illustration
A classical problem in (probabilistic) reliability. . .                 (2/2)

                                                           α
   ❍ Define the failure region




                                         PX
       Γ = {x ∈ X : f (x ) > u}.

                                                                              u




                                         f (x )
                                                           Γ
   ❍ The probability of failure is

      α = PX {Γ} =           1f >u dPX                         x
                         X
                                                  Figure: A 1d illustration


   A fundamental numerical problem in reliability analysis
   How to estimate α using a computer program that can provide
   f (x ) for any given x ∈ X ?
The venerable Monte Carlo method
    ❍ The Monte Carlo (MC) estimator
                       m
                   1                                        iid
           αMC =
           ˆ                 1f (Xi )>u   with X1 , . . . , Xm ∼ PX
                   m   i=1

      has a coefficient of variation given by

                                     1−α     1
                         δ =             ≈ √    .
                                      αm     αm
The venerable Monte Carlo method
    ❍ The Monte Carlo (MC) estimator
                        m
                    1                                            iid
            αMC =
            ˆ                 1f (Xi )>u       with X1 , . . . , Xm ∼ PX
                    m   i=1

       has a coefficient of variation given by

                                      1−α     1
                          δ =             ≈ √    .
                                       αm     αm


    ❍ Computation time for a given δ ?
                                 1                         τ0
                     m ≈                   ⇒    τ MC ≈
                               δ2 α                       δ2 α
       Ex: with δ = 50%, α = 10−5 , τ0 = 5 min, τ MC ≈ 4 years.
A short and selective review of existing techniques
    ❍ The MC estimator is impractical when
         ◮   either f is expensive to evaluate (i.e., τ0 is large),
         ◮   or Γ is a rare event under PX (i.e., α is small).
A short and selective review of existing techniques
    ❍ The MC estimator is impractical when
         ◮   either f is expensive to evaluate (i.e., τ0 is large),
         ◮   or Γ is a rare event under PX (i.e., α is small).

    ❍ Approximation techniques (and related adaptive sampling
      schemes) address the first issue.
         ◮   parametric: FORM/SORM, polynomial RSM, . . .
         ◮   non-parametric: kriging (Gaussian processes), SVM, . . .
A short and selective review of existing techniques
    ❍ The MC estimator is impractical when
         ◮   either f is expensive to evaluate (i.e., τ0 is large),
         ◮   or Γ is a rare event under PX (i.e., α is small).

    ❍ Approximation techniques (and related adaptive sampling
      schemes) address the first issue.
         ◮   parametric: FORM/SORM, polynomial RSM, . . .
         ◮   non-parametric: kriging (Gaussian processes), SVM, . . .

    ❍ Variance reduction techniques (e.g., importance sampling)
      address the second issue.
         ◮   Subset simulation (Au & Beck, 2001) is especially
                                                               √
             appropriate for very small α, since δ ∝ | log α|/ m.
What if I have an expensive f and a small α ?                 (1/2)

    ❍ Some parametric approximation techniques (e.g.,
      FORM/SORM) can be still be used. . .
         ◮   strong assumption ⇒ “structural” error that cannot be
             reduced by adding more samples.
What if I have an expensive f and a small α ?                      (1/2)

    ❍ Some parametric approximation techniques (e.g.,
      FORM/SORM) can be still be used. . .
         ◮   strong assumption ⇒ “structural” error that cannot be
             reduced by adding more samples.

    ❍ Contribution of this paper: Bayesian Subset Simulation (BSS)
         ◮   Bayesian: uses a Gaussian process prior on f (kriging)
               ◮   flexibility of a non-parametric approach,
               ◮   framework to design efficient adaptive sampling schemes.
         ◮   generalizes subset simulation
               ◮   in the framework of Sequential Monte Carlo (SMC)
                   methods (Del Moral et al, 2006).
What if I have an expensive f and a small α ?                  (2/2)

    ❍ Some recent related work
        ◮   V. Dubourg, F. Deheeger and B. Sudret
            Metamodel-based importance sampling for
            structural reliability analysis. Preprint submitted to
            Probabilistic Engineering Mechanics (available on arXiv).
            ➥ use kriging + (adaptive) importance sampling

        ◮   J.-M. Bourinet, F. Deheeger and M. Lemaire
            Assessing small failure probabilities by combined
            subset simulation and Support Vector Machines,
            Structural Safety, 33:6, 343–353, 2011.
            ➥ use SVM + subset simulation
Example : deflection of a cantilever beam
    ❍ We consider a cantilever beam of length L = 6 m, with
      uniformly distributed load (Rajashekhar & Ellingwood, 1993).




                       http://en.wikipedia.org/wiki/File:Beam1svg.svg




    ❍ The maximal deflection of the beam is
                                                 3 L4 x 1
                              f (x1 , x2 ) =            3,
                                                 2 E x2

       with x1 the load per unit area and x2 the depth.

    ❍ Young’s modulus: E = 2.6 104 MPa.
Example : deflection of a cantilever beam

    ❍ We assume an imperfect knowledge of x1 and x2 :
         ◮                2
             X1 ∼ N µ1 , σ1 , µ1 = 10−3 MPa, σ1 = 0.2 µ1 ,
         ◮                2
             X2 ∼ N µ2 , σ2 , µ2 = 300 mm, σ2 = 0.1 µ2 .
         ◮   truncated independent Gaussian variables.
Example : deflection of a cantilever beam

    ❍ We assume an imperfect knowledge of x1 and x2 :
         ◮                2
             X1 ∼ N µ1 , σ1 , µ1 = 10−3 MPa, σ1 = 0.2 µ1 ,
         ◮                2
             X2 ∼ N µ2 , σ2 , µ2 = 300 mm, σ2 = 0.1 µ2 .
         ◮   truncated independent Gaussian variables.

    ❍ A failure occurs when f (X1 , X2 ) > u = L/325.

         ◮   Reference value: α ≈ 3.94 10−6 ,
         ◮   obtained by MC with m = 1010 (⇒ δ ≈ 0.5%).


    ❍ Note: our beam is thicker than the one of Rajashekhar &
      Ellingwood to make α smaller !
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
Subset simulation with p0 = 10% and m = 16 000
And now... Bayesian subset simulation !                 (1/2)
    ❍ In the previous experiment, subset simulation performed

         N = m + (1 − p0 )(T − 1)m = 88000 evaluations of f .

       where T = 6 is the number of stages.

    ❍ Idea : we can do much better with a Gaussian process prior.
And now... Bayesian subset simulation !                     (1/2)
    ❍ In the previous experiment, subset simulation performed

         N = m + (1 − p0 )(T − 1)m = 88000 evaluations of f .

       where T = 6 is the number of stages.

    ❍ Idea : we can do much better with a Gaussian process prior.

    ❍ Key idea #1 (sequential Monte Carlo)
         ◮   SS uses an expensive sequence of target densities

                                  qt ∝ 1f >ut−1 πX

             where ut is the target level at stage t.
         ◮   We replace them by the cheaper densities

                               qt ∝ Pn (f > ut−1 ) πX

             where Pn is the GP posterior given n evaluations of f .
And now... Bayesian subset simulation !                        (2/2)

    ❍ Key idea #2 (adaptive sampling)
        ◮   At each stage t, we improve our GP model around the
            next target level ut .
        ◮   Strategy: Stepwise Uncertainty Reduction (SUR)
                 (Vazquez & Piera-Martinez (2007), Vazquez & Bect (2009))

        ◮   Other strategies could be used as well. . .
                                               (e.g., Picheny et al. (2011))
And now... Bayesian subset simulation !                         (2/2)

    ❍ Key idea #2 (adaptive sampling)
         ◮   At each stage t, we improve our GP model around the
             next target level ut .
         ◮   Strategy: Stepwise Uncertainty Reduction (SUR)
                  (Vazquez & Piera-Martinez (2007), Vazquez & Bect (2009))

         ◮   Other strategies could be used as well. . .
                                                (e.g., Picheny et al. (2011))


    ❍ Miscellaneous details
         ◮   Number of evaluations per stage: chosen adaptively.
         ◮   Number of stages T , levels ut : chosen adaptively.
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
BSS with p0 = 10% and m = 16 000
Performance ?
    ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012).
         ◮   Case tests in dimensions d = 2 and d = 6.
         ◮   Comparison with plain subset simulation and the
             2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011).

             ⇒ very significant evaluation savings
               (for a comparable MSE)
Performance ?
    ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012).
         ◮   Case tests in dimensions d = 2 and d = 6.
         ◮   Comparison with plain subset simulation and the
             2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011).

             ⇒ very significant evaluation savings
               (for a comparable MSE)

    ❍ Our estimate is biased (nothing is free. . . ).
         ◮   Typically weakly biased in our experiments.
Performance ?
    ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012).
         ◮   Case tests in dimensions d = 2 and d = 6.
         ◮   Comparison with plain subset simulation and the
             2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011).

             ⇒ very significant evaluation savings
               (for a comparable MSE)

    ❍ Our estimate is biased (nothing is free. . . ).
         ◮   Typically weakly biased in our experiments.
         ◮   Two sources of bias, that can be removed
               ◮   level-adaptation bias
                         ➥ solution: two passes,
               ◮   Bayesian bias
                         ➥ solution: evaluate all points at the last stage
Closing remarks

    ❍ Estimating small probabilities of failure on expensive computer
      models is possible, using a blend of :
         ◮   advanced simulation techniques (here, SMC)
         ◮   meta-modelling (here, Gaussian process modelling)


    ❍ Benchmarking wrt state-of-the-art techniques
        ◮ work in progress
Closing remarks

    ❍ Estimating small probabilities of failure on expensive computer
      models is possible, using a blend of :
         ◮   advanced simulation techniques (here, SMC)
         ◮   meta-modelling (here, Gaussian process modelling)


    ❍ Benchmarking wrt state-of-the-art techniques
        ◮ work in progress


    ❍ Open questions
         ◮   How well do we need do know f at intermediate stages ?
         ◮   How smooth should f be for BSS to be efficient ?
         ◮   Theoretical properties ?
References


    ❍ This talk is based on the paper
         ◮ Ling Li, Julien Bect, Emmanuel Vazquez, Bayesian Subset
             Simulation : a kriging-based subset simulation algorithm for the
             estimation of small probabilities of failure, Proceedings of PSAM 11
             & ESREL 2012, June 25-29, 2012, Helsinki, Finland     [clickme]



    ❍ For more information on kriging based adaptive sampling
      strategies (a.k.a sequential design of experiments)
         ◮ Julien Bect, David Ginsbourger, Ling Li, Victor Picheny, Emmanuel
             Vazquez, Sequential design of computer experiments for the
             estimation of a probability of failure, Statistics and Computing,
             22(3):773–793, 2012.   [clickme]

More Related Content

What's hot

A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015
Junfeng Liu
 
Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)
Julyan Arbel
 
Rank awarealgs small11
Rank awarealgs small11Rank awarealgs small11
Rank awarealgs small11
Jules Esp
 
tensor-decomposition
tensor-decompositiontensor-decomposition
tensor-decomposition
Kenta Oono
 

What's hot (19)

A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015
 
Mines April 2017 Colloquium
Mines April 2017 ColloquiumMines April 2017 Colloquium
Mines April 2017 Colloquium
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Habilitation à diriger des recherches
Habilitation à diriger des recherchesHabilitation à diriger des recherches
Habilitation à diriger des recherches
 
Role of Bellman's Equation in Reinforcement Learning
Role of Bellman's Equation in Reinforcement LearningRole of Bellman's Equation in Reinforcement Learning
Role of Bellman's Equation in Reinforcement Learning
 
Cheatsheet supervised-learning
Cheatsheet supervised-learningCheatsheet supervised-learning
Cheatsheet supervised-learning
 
Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)
 
ABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space models
 
Rank awarealgs small11
Rank awarealgs small11Rank awarealgs small11
Rank awarealgs small11
 
IJSRED-V2I5P56
IJSRED-V2I5P56IJSRED-V2I5P56
IJSRED-V2I5P56
 
tensor-decomposition
tensor-decompositiontensor-decomposition
tensor-decomposition
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
 
Toxic Release And Dispersion Models
Toxic Release And Dispersion ModelsToxic Release And Dispersion Models
Toxic Release And Dispersion Models
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4
 
1 - Linear Regression
1 - Linear Regression1 - Linear Regression
1 - Linear Regression
 
Statement of stochastic programming problems
Statement of stochastic programming problemsStatement of stochastic programming problems
Statement of stochastic programming problems
 
Normal
NormalNormal
Normal
 
Gtti 10032021
Gtti 10032021Gtti 10032021
Gtti 10032021
 

Similar to Bayesian Subset Simulation

Rank awarealgs small11
Rank awarealgs small11Rank awarealgs small11
Rank awarealgs small11
Jules Esp
 
2014 vulnerability assesment of spatial network - models and solutions
2014   vulnerability assesment of spatial network - models and solutions2014   vulnerability assesment of spatial network - models and solutions
2014 vulnerability assesment of spatial network - models and solutions
Francisco Pérez
 

Similar to Bayesian Subset Simulation (20)

QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
QMC: Operator Splitting Workshop, Proximal Algorithms in Probability Spaces -...
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Coverage of Credible I...
 
Kernels and Support Vector Machines
Kernels and Support Vector  MachinesKernels and Support Vector  Machines
Kernels and Support Vector Machines
 
Tensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantificationTensor Train data format for uncertainty quantification
Tensor Train data format for uncertainty quantification
 
Concentration inequality in Machine Learning
Concentration inequality in Machine LearningConcentration inequality in Machine Learning
Concentration inequality in Machine Learning
 
Rank awarealgs small11
Rank awarealgs small11Rank awarealgs small11
Rank awarealgs small11
 
i2ml3e-chap2.pptx
i2ml3e-chap2.pptxi2ml3e-chap2.pptx
i2ml3e-chap2.pptx
 
I2ml3e chap2
I2ml3e chap2I2ml3e chap2
I2ml3e chap2
 
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
 
Monte Carlo Statistical Methods
Monte Carlo Statistical MethodsMonte Carlo Statistical Methods
Monte Carlo Statistical Methods
 
Thesis oral defense
Thesis oral defenseThesis oral defense
Thesis oral defense
 
Tensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEsTensor train to solve stochastic PDEs
Tensor train to solve stochastic PDEs
 
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
Subproblem-Tree Calibration: A Unified Approach to Max-Product Message Passin...
 
Gaussian process in machine learning
Gaussian process in machine learningGaussian process in machine learning
Gaussian process in machine learning
 
Input analysis
Input analysisInput analysis
Input analysis
 
On learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihoodOn learning statistical mixtures maximizing the complete likelihood
On learning statistical mixtures maximizing the complete likelihood
 
2014 vulnerability assesment of spatial network - models and solutions
2014   vulnerability assesment of spatial network - models and solutions2014   vulnerability assesment of spatial network - models and solutions
2014 vulnerability assesment of spatial network - models and solutions
 
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
QMC: Operator Splitting Workshop, Perturbed (accelerated) Proximal-Gradient A...
 
R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013
R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013
R exam (B) given in Paris-Dauphine, Licence Mido, Jan. 11, 2013
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
 

Recently uploaded

The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
heathfieldcps1
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
ZurliaSoop
 

Recently uploaded (20)

The basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptxThe basics of sentences session 3pptx.pptx
The basics of sentences session 3pptx.pptx
 
Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)Accessible Digital Futures project (20/03/2024)
Accessible Digital Futures project (20/03/2024)
 
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdfUGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
UGC NET Paper 1 Mathematical Reasoning & Aptitude.pdf
 
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
Jual Obat Aborsi Hongkong ( Asli No.1 ) 085657271886 Obat Penggugur Kandungan...
 
FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024FSB Advising Checklist - Orientation 2024
FSB Advising Checklist - Orientation 2024
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
Jamworks pilot and AI at Jisc (20/03/2024)
Jamworks pilot and AI at Jisc (20/03/2024)Jamworks pilot and AI at Jisc (20/03/2024)
Jamworks pilot and AI at Jisc (20/03/2024)
 
Interdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptxInterdisciplinary_Insights_Data_Collection_Methods.pptx
Interdisciplinary_Insights_Data_Collection_Methods.pptx
 
On National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan FellowsOn National Teacher Day, meet the 2024-25 Kenan Fellows
On National Teacher Day, meet the 2024-25 Kenan Fellows
 
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptxSKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
SKILL OF INTRODUCING THE LESSON MICRO SKILLS.pptx
 
Google Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptxGoogle Gemini An AI Revolution in Education.pptx
Google Gemini An AI Revolution in Education.pptx
 
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...Kodo Millet  PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
Kodo Millet PPT made by Ghanshyam bairwa college of Agriculture kumher bhara...
 
Python Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docxPython Notes for mca i year students osmania university.docx
Python Notes for mca i year students osmania university.docx
 
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
Beyond_Borders_Understanding_Anime_and_Manga_Fandom_A_Comprehensive_Audience_...
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptxHMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
HMCS Vancouver Pre-Deployment Brief - May 2024 (Web Version).pptx
 
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
2024-NATIONAL-LEARNING-CAMP-AND-OTHER.pptx
 
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdfUnit 3 Emotional Intelligence and Spiritual Intelligence.pdf
Unit 3 Emotional Intelligence and Spiritual Intelligence.pdf
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...Making communications land - Are they received and understood as intended? we...
Making communications land - Are they received and understood as intended? we...
 

Bayesian Subset Simulation

  • 1. Bayesian Subset Simulation — a kriging-based subset simulation algorithm for the estimation of small probabilities of failure — Ling Li, Julien Bect, Emmanuel Vazquez Supelec, France PSAM11-ESREL12 Helsinki, June 26, 2012
  • 2. A classical problem in (probabilistic) reliability. . . (1/2) ❍ Consider a system subject to uncertainties, ◮ aleatory and/or epistemic, ◮ represented by a random vector X ∼ PX where PX is a probability measure on X ⊂ Rd .
  • 3. A classical problem in (probabilistic) reliability. . . (1/2) ❍ Consider a system subject to uncertainties, ◮ aleatory and/or epistemic, ◮ represented by a random vector X ∼ PX where PX is a probability measure on X ⊂ Rd . ❍ Assume that the system fails when f (X ) > u ◮ f : X → R is a cost function, ◮ u ∈ R is the critical level. ❍ x → u − f (x ) is sometimes called the “limit state function”
  • 4. A classical problem in (probabilistic) reliability. . . (2/2) α ❍ Define the failure region PX Γ = {x ∈ X : f (x ) > u}. u f (x ) Γ ❍ The probability of failure is α = PX {Γ} = 1f >u dPX x X Figure: A 1d illustration
  • 5. A classical problem in (probabilistic) reliability. . . (2/2) α ❍ Define the failure region PX Γ = {x ∈ X : f (x ) > u}. u f (x ) Γ ❍ The probability of failure is α = PX {Γ} = 1f >u dPX x X Figure: A 1d illustration A fundamental numerical problem in reliability analysis How to estimate α using a computer program that can provide f (x ) for any given x ∈ X ?
  • 6. The venerable Monte Carlo method ❍ The Monte Carlo (MC) estimator m 1 iid αMC = ˆ 1f (Xi )>u with X1 , . . . , Xm ∼ PX m i=1 has a coefficient of variation given by 1−α 1 δ = ≈ √ . αm αm
  • 7. The venerable Monte Carlo method ❍ The Monte Carlo (MC) estimator m 1 iid αMC = ˆ 1f (Xi )>u with X1 , . . . , Xm ∼ PX m i=1 has a coefficient of variation given by 1−α 1 δ = ≈ √ . αm αm ❍ Computation time for a given δ ? 1 τ0 m ≈ ⇒ τ MC ≈ δ2 α δ2 α Ex: with δ = 50%, α = 10−5 , τ0 = 5 min, τ MC ≈ 4 years.
  • 8. A short and selective review of existing techniques ❍ The MC estimator is impractical when ◮ either f is expensive to evaluate (i.e., τ0 is large), ◮ or Γ is a rare event under PX (i.e., α is small).
  • 9. A short and selective review of existing techniques ❍ The MC estimator is impractical when ◮ either f is expensive to evaluate (i.e., τ0 is large), ◮ or Γ is a rare event under PX (i.e., α is small). ❍ Approximation techniques (and related adaptive sampling schemes) address the first issue. ◮ parametric: FORM/SORM, polynomial RSM, . . . ◮ non-parametric: kriging (Gaussian processes), SVM, . . .
  • 10. A short and selective review of existing techniques ❍ The MC estimator is impractical when ◮ either f is expensive to evaluate (i.e., τ0 is large), ◮ or Γ is a rare event under PX (i.e., α is small). ❍ Approximation techniques (and related adaptive sampling schemes) address the first issue. ◮ parametric: FORM/SORM, polynomial RSM, . . . ◮ non-parametric: kriging (Gaussian processes), SVM, . . . ❍ Variance reduction techniques (e.g., importance sampling) address the second issue. ◮ Subset simulation (Au & Beck, 2001) is especially √ appropriate for very small α, since δ ∝ | log α|/ m.
  • 11. What if I have an expensive f and a small α ? (1/2) ❍ Some parametric approximation techniques (e.g., FORM/SORM) can be still be used. . . ◮ strong assumption ⇒ “structural” error that cannot be reduced by adding more samples.
  • 12. What if I have an expensive f and a small α ? (1/2) ❍ Some parametric approximation techniques (e.g., FORM/SORM) can be still be used. . . ◮ strong assumption ⇒ “structural” error that cannot be reduced by adding more samples. ❍ Contribution of this paper: Bayesian Subset Simulation (BSS) ◮ Bayesian: uses a Gaussian process prior on f (kriging) ◮ flexibility of a non-parametric approach, ◮ framework to design efficient adaptive sampling schemes. ◮ generalizes subset simulation ◮ in the framework of Sequential Monte Carlo (SMC) methods (Del Moral et al, 2006).
  • 13. What if I have an expensive f and a small α ? (2/2) ❍ Some recent related work ◮ V. Dubourg, F. Deheeger and B. Sudret Metamodel-based importance sampling for structural reliability analysis. Preprint submitted to Probabilistic Engineering Mechanics (available on arXiv). ➥ use kriging + (adaptive) importance sampling ◮ J.-M. Bourinet, F. Deheeger and M. Lemaire Assessing small failure probabilities by combined subset simulation and Support Vector Machines, Structural Safety, 33:6, 343–353, 2011. ➥ use SVM + subset simulation
  • 14. Example : deflection of a cantilever beam ❍ We consider a cantilever beam of length L = 6 m, with uniformly distributed load (Rajashekhar & Ellingwood, 1993). http://en.wikipedia.org/wiki/File:Beam1svg.svg ❍ The maximal deflection of the beam is 3 L4 x 1 f (x1 , x2 ) = 3, 2 E x2 with x1 the load per unit area and x2 the depth. ❍ Young’s modulus: E = 2.6 104 MPa.
  • 15. Example : deflection of a cantilever beam ❍ We assume an imperfect knowledge of x1 and x2 : ◮ 2 X1 ∼ N µ1 , σ1 , µ1 = 10−3 MPa, σ1 = 0.2 µ1 , ◮ 2 X2 ∼ N µ2 , σ2 , µ2 = 300 mm, σ2 = 0.1 µ2 . ◮ truncated independent Gaussian variables.
  • 16. Example : deflection of a cantilever beam ❍ We assume an imperfect knowledge of x1 and x2 : ◮ 2 X1 ∼ N µ1 , σ1 , µ1 = 10−3 MPa, σ1 = 0.2 µ1 , ◮ 2 X2 ∼ N µ2 , σ2 , µ2 = 300 mm, σ2 = 0.1 µ2 . ◮ truncated independent Gaussian variables. ❍ A failure occurs when f (X1 , X2 ) > u = L/325. ◮ Reference value: α ≈ 3.94 10−6 , ◮ obtained by MC with m = 1010 (⇒ δ ≈ 0.5%). ❍ Note: our beam is thicker than the one of Rajashekhar & Ellingwood to make α smaller !
  • 17. Subset simulation with p0 = 10% and m = 16 000
  • 18. Subset simulation with p0 = 10% and m = 16 000
  • 19. Subset simulation with p0 = 10% and m = 16 000
  • 20. Subset simulation with p0 = 10% and m = 16 000
  • 21. Subset simulation with p0 = 10% and m = 16 000
  • 22. Subset simulation with p0 = 10% and m = 16 000
  • 23. Subset simulation with p0 = 10% and m = 16 000
  • 24. Subset simulation with p0 = 10% and m = 16 000
  • 25. Subset simulation with p0 = 10% and m = 16 000
  • 26. Subset simulation with p0 = 10% and m = 16 000
  • 27. Subset simulation with p0 = 10% and m = 16 000
  • 28. Subset simulation with p0 = 10% and m = 16 000
  • 29. Subset simulation with p0 = 10% and m = 16 000
  • 30. Subset simulation with p0 = 10% and m = 16 000
  • 31. Subset simulation with p0 = 10% and m = 16 000
  • 32. Subset simulation with p0 = 10% and m = 16 000
  • 33. Subset simulation with p0 = 10% and m = 16 000
  • 34. And now... Bayesian subset simulation ! (1/2) ❍ In the previous experiment, subset simulation performed N = m + (1 − p0 )(T − 1)m = 88000 evaluations of f . where T = 6 is the number of stages. ❍ Idea : we can do much better with a Gaussian process prior.
  • 35. And now... Bayesian subset simulation ! (1/2) ❍ In the previous experiment, subset simulation performed N = m + (1 − p0 )(T − 1)m = 88000 evaluations of f . where T = 6 is the number of stages. ❍ Idea : we can do much better with a Gaussian process prior. ❍ Key idea #1 (sequential Monte Carlo) ◮ SS uses an expensive sequence of target densities qt ∝ 1f >ut−1 πX where ut is the target level at stage t. ◮ We replace them by the cheaper densities qt ∝ Pn (f > ut−1 ) πX where Pn is the GP posterior given n evaluations of f .
  • 36. And now... Bayesian subset simulation ! (2/2) ❍ Key idea #2 (adaptive sampling) ◮ At each stage t, we improve our GP model around the next target level ut . ◮ Strategy: Stepwise Uncertainty Reduction (SUR) (Vazquez & Piera-Martinez (2007), Vazquez & Bect (2009)) ◮ Other strategies could be used as well. . . (e.g., Picheny et al. (2011))
  • 37. And now... Bayesian subset simulation ! (2/2) ❍ Key idea #2 (adaptive sampling) ◮ At each stage t, we improve our GP model around the next target level ut . ◮ Strategy: Stepwise Uncertainty Reduction (SUR) (Vazquez & Piera-Martinez (2007), Vazquez & Bect (2009)) ◮ Other strategies could be used as well. . . (e.g., Picheny et al. (2011)) ❍ Miscellaneous details ◮ Number of evaluations per stage: chosen adaptively. ◮ Number of stages T , levels ut : chosen adaptively.
  • 38. BSS with p0 = 10% and m = 16 000
  • 39. BSS with p0 = 10% and m = 16 000
  • 40. BSS with p0 = 10% and m = 16 000
  • 41. BSS with p0 = 10% and m = 16 000
  • 42. BSS with p0 = 10% and m = 16 000
  • 43. BSS with p0 = 10% and m = 16 000
  • 44. BSS with p0 = 10% and m = 16 000
  • 45. BSS with p0 = 10% and m = 16 000
  • 46. BSS with p0 = 10% and m = 16 000
  • 47. BSS with p0 = 10% and m = 16 000
  • 48. BSS with p0 = 10% and m = 16 000
  • 49. BSS with p0 = 10% and m = 16 000
  • 50. BSS with p0 = 10% and m = 16 000
  • 51. BSS with p0 = 10% and m = 16 000
  • 52. BSS with p0 = 10% and m = 16 000
  • 53. BSS with p0 = 10% and m = 16 000
  • 54. BSS with p0 = 10% and m = 16 000
  • 55. BSS with p0 = 10% and m = 16 000
  • 56. BSS with p0 = 10% and m = 16 000
  • 57. BSS with p0 = 10% and m = 16 000
  • 58. BSS with p0 = 10% and m = 16 000
  • 59. BSS with p0 = 10% and m = 16 000
  • 60. BSS with p0 = 10% and m = 16 000
  • 61. BSS with p0 = 10% and m = 16 000
  • 62. BSS with p0 = 10% and m = 16 000
  • 63. BSS with p0 = 10% and m = 16 000
  • 64. BSS with p0 = 10% and m = 16 000
  • 65. BSS with p0 = 10% and m = 16 000
  • 66. BSS with p0 = 10% and m = 16 000
  • 67. BSS with p0 = 10% and m = 16 000
  • 68. BSS with p0 = 10% and m = 16 000
  • 69. BSS with p0 = 10% and m = 16 000
  • 70. BSS with p0 = 10% and m = 16 000
  • 71. BSS with p0 = 10% and m = 16 000
  • 72. BSS with p0 = 10% and m = 16 000
  • 73. BSS with p0 = 10% and m = 16 000
  • 74. BSS with p0 = 10% and m = 16 000
  • 75. BSS with p0 = 10% and m = 16 000
  • 76. BSS with p0 = 10% and m = 16 000
  • 77. BSS with p0 = 10% and m = 16 000
  • 78. BSS with p0 = 10% and m = 16 000
  • 79. BSS with p0 = 10% and m = 16 000
  • 80. BSS with p0 = 10% and m = 16 000
  • 81. BSS with p0 = 10% and m = 16 000
  • 82. BSS with p0 = 10% and m = 16 000
  • 83. BSS with p0 = 10% and m = 16 000
  • 84. BSS with p0 = 10% and m = 16 000
  • 85. Performance ? ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012). ◮ Case tests in dimensions d = 2 and d = 6. ◮ Comparison with plain subset simulation and the 2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011). ⇒ very significant evaluation savings (for a comparable MSE)
  • 86. Performance ? ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012). ◮ Case tests in dimensions d = 2 and d = 6. ◮ Comparison with plain subset simulation and the 2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011). ⇒ very significant evaluation savings (for a comparable MSE) ❍ Our estimate is biased (nothing is free. . . ). ◮ Typically weakly biased in our experiments.
  • 87. Performance ? ❍ Preliminary Monte Carlo studies (PhD thesis of Ling Li, 2012). ◮ Case tests in dimensions d = 2 and d = 6. ◮ Comparison with plain subset simulation and the 2 SMART algorithm (Deheeger, 2007; Bourinet et al., 2011). ⇒ very significant evaluation savings (for a comparable MSE) ❍ Our estimate is biased (nothing is free. . . ). ◮ Typically weakly biased in our experiments. ◮ Two sources of bias, that can be removed ◮ level-adaptation bias ➥ solution: two passes, ◮ Bayesian bias ➥ solution: evaluate all points at the last stage
  • 88. Closing remarks ❍ Estimating small probabilities of failure on expensive computer models is possible, using a blend of : ◮ advanced simulation techniques (here, SMC) ◮ meta-modelling (here, Gaussian process modelling) ❍ Benchmarking wrt state-of-the-art techniques ◮ work in progress
  • 89. Closing remarks ❍ Estimating small probabilities of failure on expensive computer models is possible, using a blend of : ◮ advanced simulation techniques (here, SMC) ◮ meta-modelling (here, Gaussian process modelling) ❍ Benchmarking wrt state-of-the-art techniques ◮ work in progress ❍ Open questions ◮ How well do we need do know f at intermediate stages ? ◮ How smooth should f be for BSS to be efficient ? ◮ Theoretical properties ?
  • 90. References ❍ This talk is based on the paper ◮ Ling Li, Julien Bect, Emmanuel Vazquez, Bayesian Subset Simulation : a kriging-based subset simulation algorithm for the estimation of small probabilities of failure, Proceedings of PSAM 11 & ESREL 2012, June 25-29, 2012, Helsinki, Finland [clickme] ❍ For more information on kriging based adaptive sampling strategies (a.k.a sequential design of experiments) ◮ Julien Bect, David Ginsbourger, Ling Li, Victor Picheny, Emmanuel Vazquez, Sequential design of computer experiments for the estimation of a probability of failure, Statistics and Computing, 22(3):773–793, 2012. [clickme]