SlideShare a Scribd company logo
1 of 66
Download to read offline
DISCUSSION
                    of
Bayesian Computation via empirical likelihood

     Stefano Cabras, stefano.cabras@uc3m.es
        Universidad Carlos III de Madrid (Spain)
              Universit` di Cagliari (Italy)
                       a



                Padova, 21-Mar-2013
Summary
   ◮   Problem:
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
   ◮   we want to obtain the posterior

                             πN (θ | y ) ∝ LN (θ)π(θ).
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
   ◮   we want to obtain the posterior

                             πN (θ | y ) ∝ LN (θ)π(θ).



   ◮   BUT
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
   ◮   we want to obtain the posterior

                             πN (θ | y ) ∝ LN (θ)π(θ).



   ◮   BUT
         ◮   IF LN (θ) is not available:
               ◮   THEN all life ABC;
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
   ◮   we want to obtain the posterior

                             πN (θ | y ) ∝ LN (θ)π(θ).



   ◮   BUT
         ◮   IF LN (θ) is not available:
               ◮   THEN all life ABC;
         ◮   IF it is not even possible to simulate from f (y | θ):
Summary
   ◮   Problem:
         ◮   a statistical model f (y | θ);
         ◮   a prior π(θ) on θ;
   ◮   we want to obtain the posterior

                             πN (θ | y ) ∝ LN (θ)π(θ).



   ◮   BUT
         ◮   IF LN (θ) is not available:
               ◮   THEN all life ABC;
         ◮   IF it is not even possible to simulate from f (y | θ):
               ◮   THEN replace LN (θ) with LEL (θ)
                   (the proposed BCel procedure):

                                     π(θ|y ) ∝ LEL (θ) × π(θ).

                   .
... what remains about the f (y | θ) ?
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                              Ef (y |θ) [h(Y , θ)] = 0.
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                              Ef (y |θ) [h(Y , θ)] = 0.



     ◮   The relation between θ and obs. Y is model conditioned and
         expressed by h(Y , θ);
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                              Ef (y |θ) [h(Y , θ)] = 0.



     ◮   The relation between θ and obs. Y is model conditioned and
         expressed by h(Y , θ);

     ◮   Constraints are model driven and so there is still a timid trace
         of f (y | θ) in BCel .
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                              Ef (y |θ) [h(Y , θ)] = 0.



     ◮   The relation between θ and obs. Y is model conditioned and
         expressed by h(Y , θ);

     ◮   Constraints are model driven and so there is still a timid trace
         of f (y | θ) in BCel .
     ◮   Examples:
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                                Ef (y |θ) [h(Y , θ)] = 0.



     ◮   The relation between θ and obs. Y is model conditioned and
         expressed by h(Y , θ);

     ◮   Constraints are model driven and so there is still a timid trace
         of f (y | θ) in BCel .
     ◮   Examples:
           ◮   The coalescent model example is illuminating in suggesting the
               score of the pairwise likelihood;
... what remains about the f (y | θ) ?

     ◮   Recall that the Empirical Likelihood is defined, for iid sample,
         by means of a set of constraints:

                                Ef (y |θ) [h(Y , θ)] = 0.



     ◮   The relation between θ and obs. Y is model conditioned and
         expressed by h(Y , θ);

     ◮   Constraints are model driven and so there is still a timid trace
         of f (y | θ) in BCel .
     ◮   Examples:
           ◮   The coalescent model example is illuminating in suggesting the
               score of the pairwise likelihood;
           ◮   The residuals in GARCH models.
... a suggestion




               What if we do not even known h(·) ?
... how to elicit h(·) automatically
... how to elicit h(·) automatically
... how to elicit h(·) automatically




     ◮   Set h(Y , θ) = Y − g (θ), where

                              g (θ) = Ef (y |θ) (Y |θ),

         is the regression function of Y |θ;
... how to elicit h(·) automatically




     ◮   Set h(Y , θ) = Y − g (θ), where

                              g (θ) = Ef (y |θ) (Y |θ),

         is the regression function of Y |θ;

     ◮   g (θ) should be replaced by an estimator g (θ).
How to estimate g (θ) ?




      1
       ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras,
   Castellanos, Ruli (Ercim-2012, Oviedo).
How to estimate g (θ) ?



     ◮    Use a once forever pilot-run simulation study:      1




      1
       ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras,
   Castellanos, Ruli (Ercim-2012, Oviedo).
How to estimate g (θ) ?



     ◮    Use a once forever pilot-run simulation study:      1


    1. Consider a grid (or regular lattice) of θ made by M points:
       θ1 , . . . , θM




      1
       ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras,
   Castellanos, Ruli (Ercim-2012, Oviedo).
How to estimate g (θ) ?



     ◮    Use a once forever pilot-run simulation study:      1


    1. Consider a grid (or regular lattice) of θ made by M points:
       θ1 , . . . , θM
    2. Simulate the corresponding y1 , . . . , yM




      1
       ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras,
   Castellanos, Ruli (Ercim-2012, Oviedo).
How to estimate g (θ) ?



     ◮    Use a once forever pilot-run simulation study:       1


    1. Consider a grid (or regular lattice) of θ made by M points:
       θ1 , . . . , θM
    2. Simulate the corresponding y1 , . . . , yM
    3. Regress y1 , . . . , yM on θ 1 , . . . , θ M obtaining g (θ).




      1
       ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras,
   Castellanos, Ruli (Ercim-2012, Oviedo).
... example: y ∼ N(|θ|, 1)
   For a pilot run of M = 1000 we have g (θ) = |θ|.
                                       ˆ


                              Pilot−Run s.s.


                                 g(θ)
       10
   y

       5
       0




            −10      −5             0          5      10

                                    θ
... example: y ∼ N(|θ|, 1)
   Suppose to draw a n = 100 sample from θ = 2:


                             Histogram of y
               20
               15
   Frequency

               10
               5
               0




                    0   1     2               3   4

                                   y
... example: y ∼ N(|θ|, 1)
   The Empirical Likelihood is this
               2.5
               2.0
   Emp. Lik.

               1.5
               1.0




                     −4   −2          0   2   4

                                      θ
1st Point: Do we need necessarily have to use f (y | θ) ?
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
     ◮   How this is reflected in the BCel ?
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
     ◮   How this is reflected in the BCel ?
           ◮   For a given data y;
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
     ◮   How this is reflected in the BCel ?
           ◮   For a given data y;
           ◮   and h(Y , θ) fixed;
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
     ◮   How this is reflected in the BCel ?
           ◮   For a given data y;
           ◮   and h(Y , θ) fixed;
           ◮   the LEL (θ) is the same regardless of f (y | θ).
1st Point: Do we need necessarily have to use f (y | θ) ?




     ◮   The above data maybe drawn from a (e.g.) a Half Normal;
     ◮   How this is reflected in the BCel ?
           ◮   For a given data y;
           ◮   and h(Y , θ) fixed;
           ◮   the LEL (θ) is the same regardless of f (y | θ).

                            Can we ignore f (y | θ) ?
2nd Point: Sample free vs Simulation free
2nd Point: Sample free vs Simulation free


     ◮   The Empirical Likelihood is ”simulation free” but not ”sample
         free”, i.e.
2nd Point: Sample free vs Simulation free


     ◮   The Empirical Likelihood is ”simulation free” but not ”sample
         free”, i.e.
           ◮   LEL (θ) → LN (θ) for n → ∞,
           ◮   implying π(θ|y) → πN (θ | y ) asymptotically in n.
2nd Point: Sample free vs Simulation free


     ◮   The Empirical Likelihood is ”simulation free” but not ”sample
         free”, i.e.
           ◮   LEL (θ) → LN (θ) for n → ∞,
           ◮   implying π(θ|y) → πN (θ | y ) asymptotically in n.
     ◮   The ABC is ”sample free” but not ”simulation free”, i.e.
2nd Point: Sample free vs Simulation free


     ◮   The Empirical Likelihood is ”simulation free” but not ”sample
         free”, i.e.
           ◮   LEL (θ) → LN (θ) for n → ∞,
           ◮   implying π(θ|y) → πN (θ | y ) asymptotically in n.
     ◮   The ABC is ”sample free” but not ”simulation free”, i.e.
           ◮   π(θ|ρ(s(y), so bs) < ǫ) → πN (θ | y ) as ǫ → 0
           ◮   implying convergence in the number of simulations if s(y ) were
               sufficient.
2nd Point: Sample free vs Simulation free


     ◮   The Empirical Likelihood is ”simulation free” but not ”sample
         free”, i.e.
           ◮   LEL (θ) → LN (θ) for n → ∞,
           ◮   implying π(θ|y) → πN (θ | y ) asymptotically in n.
     ◮   The ABC is ”sample free” but not ”simulation free”, i.e.
           ◮   π(θ|ρ(s(y), so bs) < ǫ) → πN (θ | y ) as ǫ → 0
           ◮   implying convergence in the number of simulations if s(y ) were
               sufficient.

                    A quick answer recommends use BCel
                                    BUT
                  a small sample would recommend ABC ?
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003)
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003)
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
          ◮   Modified-Likelihoods:
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003)
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009)

                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003)
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009)

                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
          ◮   Quasi-Likelihoods:
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003)
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009)

                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006)
                ◮   Greco et al. (JSPI, 2008)
                ◮   Ventura et al. (JSPI, 2010)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003) : examples and coverages of C.I.
                ◮   Mengersen et al. (PNAS, 2012)

                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009)

                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006)
                ◮   Greco et al. (JSPI, 2008)
                ◮   Ventura et al. (JSPI, 2010)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003) : examples and coverages of C.I.
                ◮   Mengersen et al. (PNAS, 2012) : examples and coverages of
                    C.I.
                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009)

                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006)
                ◮   Greco et al. (JSPI, 2008)
                ◮   Ventura et al. (JSPI, 2010)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003) : examples and coverages of C.I.
                ◮   Mengersen et al. (PNAS, 2012) : examples and coverages of
                    C.I.
                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009) : second order matching
                    properties;
                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006)
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006)
                ◮   Greco et al. (JSPI, 2008)
                ◮   Ventura et al. (JSPI, 2010)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003) : examples and coverages of C.I.
                ◮   Mengersen et al. (PNAS, 2012) : examples and coverages of
                    C.I.
                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009) : second order matching
                    properties;
                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006) : examples;
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006)
                ◮   Greco et al. (JSPI, 2008)
                ◮   Ventura et al. (JSPI, 2010)
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?
    ◮   The use of pseudo-likelihoods is not new in the Bayesian
        setting:
          ◮   Empirical Likelihoods:
                ◮   Lazar (Biometrika, 2003) : examples and coverages of C.I.
                ◮   Mengersen et al. (PNAS, 2012) : examples and coverages of
                    C.I.
                ◮   ...
          ◮   Modified-Likelihoods:
                ◮   Ventura et al. (JASA, 2009) : second order matching
                    properties;
                ◮   Chang and Mukerjee (Stat. & Prob. Letters 2006) : examples;
                ◮   ...
          ◮   Quasi-Likelihoods:
                ◮   Lin (Statist. Methodol., 2006) : examples;
                ◮   Greco et al. (JSPI, 2008) : robustness properties;
                ◮   Ventura et al. (JSPI, 2010) : examples and coverages of C.I.;
                ◮   ...
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?


    ◮   Monahan & Boos (Biometrika, 1992) proposed a notion of
        validity:
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?


    ◮   Monahan & Boos (Biometrika, 1992) proposed a notion of
        validity:

     π(θ|y ) should obey the laws of probability in a fashion that is
          consistent with statements derived from Bayes’rule.
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?


    ◮   Monahan & Boos (Biometrika, 1992) proposed a notion of
        validity:

     π(θ|y ) should obey the laws of probability in a fashion that is
          consistent with statements derived from Bayes’rule.

    ◮   Very difficult!
3nd Point: How to validate a pseudo-posterior
π(θ|y ) ∝ LEL (θ) × π(θ) ?


    ◮   Monahan & Boos (Biometrika, 1992) proposed a notion of
        validity:

     π(θ|y ) should obey the laws of probability in a fashion that is
          consistent with statements derived from Bayes’rule.

    ◮   Very difficult!

     How to validate the pseudo-posterior π(θ|y ) when this is not
                              possible ?
... Last point: the ABC is still a terrific tool
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
           ◮   Statistical Journals;
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
           ◮   Statistical Journals;
           ◮   Twitter;
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
           ◮   Statistical Journals;
           ◮   Twitter;
           ◮   Xiang’s blog ( xianblog.wordpress.com )
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
           ◮   Statistical Journals;
           ◮   Twitter;
           ◮   Xiang’s blog ( xianblog.wordpress.com )
     ◮   ... it is tailored to Approximate LN (θ).
... Last point: the ABC is still a terrific tool




     ◮   ... a lot of references:
           ◮   Statistical Journals;
           ◮   Twitter;
           ◮   Xiang’s blog ( xianblog.wordpress.com )
     ◮   ... it is tailored to Approximate LN (θ).

                          Where is the A in BCel ?
Discussion cabras-robert-130323171455-phpapp02

More Related Content

What's hot

Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Christian Robert
 
Reinforcement Learning in Economics and Finance
Reinforcement Learning in Economics and FinanceReinforcement Learning in Economics and Finance
Reinforcement Learning in Economics and FinanceArthur Charpentier
 
transformations and nonparametric inference
transformations and nonparametric inferencetransformations and nonparametric inference
transformations and nonparametric inferenceArthur Charpentier
 
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier DetectionA Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier DetectionKonkuk University, Korea
 
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Christian Robert
 
Slides econometrics-2018-graduate-3
Slides econometrics-2018-graduate-3Slides econometrics-2018-graduate-3
Slides econometrics-2018-graduate-3Arthur Charpentier
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerChristian Robert
 
Econometrics, PhD Course, #1 Nonlinearities
Econometrics, PhD Course, #1 NonlinearitiesEconometrics, PhD Course, #1 Nonlinearities
Econometrics, PhD Course, #1 NonlinearitiesArthur Charpentier
 
Slides econometrics-2018-graduate-2
Slides econometrics-2018-graduate-2Slides econometrics-2018-graduate-2
Slides econometrics-2018-graduate-2Arthur Charpentier
 
Colloquium in honor of Hans Ruedi Künsch
Colloquium in honor of Hans Ruedi KünschColloquium in honor of Hans Ruedi Künsch
Colloquium in honor of Hans Ruedi KünschChristian Robert
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)NYversity
 
Slides sales-forecasting-session2-web
Slides sales-forecasting-session2-webSlides sales-forecasting-session2-web
Slides sales-forecasting-session2-webArthur Charpentier
 
Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Christian Robert
 

What's hot (20)

Slides lln-risques
Slides lln-risquesSlides lln-risques
Slides lln-risques
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
 
Side 2019, part 1
Side 2019, part 1Side 2019, part 1
Side 2019, part 1
 
Reinforcement Learning in Economics and Finance
Reinforcement Learning in Economics and FinanceReinforcement Learning in Economics and Finance
Reinforcement Learning in Economics and Finance
 
transformations and nonparametric inference
transformations and nonparametric inferencetransformations and nonparametric inference
transformations and nonparametric inference
 
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier DetectionA Tutorial of the EM-algorithm and Its Application to Outlier Detection
A Tutorial of the EM-algorithm and Its Application to Outlier Detection
 
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
Discussion of ABC talk by Francesco Pauli, Padova, March 21, 2013
 
Slides simplexe
Slides simplexeSlides simplexe
Slides simplexe
 
Slides econometrics-2018-graduate-3
Slides econometrics-2018-graduate-3Slides econometrics-2018-graduate-3
Slides econometrics-2018-graduate-3
 
Side 2019, part 2
Side 2019, part 2Side 2019, part 2
Side 2019, part 2
 
Coordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like samplerCoordinate sampler : A non-reversible Gibbs-like sampler
Coordinate sampler : A non-reversible Gibbs-like sampler
 
Side 2019 #10
Side 2019 #10Side 2019 #10
Side 2019 #10
 
Econometrics, PhD Course, #1 Nonlinearities
Econometrics, PhD Course, #1 NonlinearitiesEconometrics, PhD Course, #1 Nonlinearities
Econometrics, PhD Course, #1 Nonlinearities
 
Side 2019 #7
Side 2019 #7Side 2019 #7
Side 2019 #7
 
Slides econ-lm
Slides econ-lmSlides econ-lm
Slides econ-lm
 
Slides econometrics-2018-graduate-2
Slides econometrics-2018-graduate-2Slides econometrics-2018-graduate-2
Slides econometrics-2018-graduate-2
 
Colloquium in honor of Hans Ruedi Künsch
Colloquium in honor of Hans Ruedi KünschColloquium in honor of Hans Ruedi Künsch
Colloquium in honor of Hans Ruedi Künsch
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)
 
Slides sales-forecasting-session2-web
Slides sales-forecasting-session2-webSlides sales-forecasting-session2-web
Slides sales-forecasting-session2-web
 
Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010Mark Girolami's Read Paper 2010
Mark Girolami's Read Paper 2010
 

Viewers also liked

Model complexity
Model complexityModel complexity
Model complexityDeb Roy
 
PhD Recipe
PhD RecipePhD Recipe
PhD RecipeDeb Roy
 
Xian's abc
Xian's abcXian's abc
Xian's abcDeb Roy
 
ABC & Empirical Lkd
ABC & Empirical LkdABC & Empirical Lkd
ABC & Empirical LkdDeb Roy
 
Mcmc & lkd free II
Mcmc & lkd free IIMcmc & lkd free II
Mcmc & lkd free IIDeb Roy
 
Slides2 130201091056-phpapp01
Slides2 130201091056-phpapp01Slides2 130201091056-phpapp01
Slides2 130201091056-phpapp01Deb Roy
 
The longest ear hair in the world
The longest ear hair in the worldThe longest ear hair in the world
The longest ear hair in the worldSmay Say
 

Viewers also liked (10)

PaulaTataruVienna
PaulaTataruViennaPaulaTataruVienna
PaulaTataruVienna
 
Model complexity
Model complexityModel complexity
Model complexity
 
PhD Recipe
PhD RecipePhD Recipe
PhD Recipe
 
Xian's abc
Xian's abcXian's abc
Xian's abc
 
ABC & Empirical Lkd
ABC & Empirical LkdABC & Empirical Lkd
ABC & Empirical Lkd
 
DIC
DICDIC
DIC
 
Mcmc & lkd free II
Mcmc & lkd free IIMcmc & lkd free II
Mcmc & lkd free II
 
Slides2 130201091056-phpapp01
Slides2 130201091056-phpapp01Slides2 130201091056-phpapp01
Slides2 130201091056-phpapp01
 
Abc
AbcAbc
Abc
 
The longest ear hair in the world
The longest ear hair in the worldThe longest ear hair in the world
The longest ear hair in the world
 

Similar to Discussion cabras-robert-130323171455-phpapp02

Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionMichael Stumpf
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapterChristian Robert
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsChristian Robert
 
Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)NYversity
 
Introduction to modern Variational Inference.
Introduction to modern Variational Inference.Introduction to modern Variational Inference.
Introduction to modern Variational Inference.Tomasz Kusmierczyk
 
Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jungkyu Lee
 
Conceptual Introduction to Gaussian Processes
Conceptual Introduction to Gaussian ProcessesConceptual Introduction to Gaussian Processes
Conceptual Introduction to Gaussian ProcessesJuanPabloCarbajal3
 
P, NP and NP-Complete, Theory of NP-Completeness V2
P, NP and NP-Complete, Theory of NP-Completeness V2P, NP and NP-Complete, Theory of NP-Completeness V2
P, NP and NP-Complete, Theory of NP-Completeness V2S.Shayan Daneshvar
 
Algebras for programming languages
Algebras for programming languagesAlgebras for programming languages
Algebras for programming languagesYoshihiro Mizoguchi
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015Christian Robert
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
Cs229 notes8
Cs229 notes8Cs229 notes8
Cs229 notes8VuTran231
 
STOMA FULL SLIDE (probability of IISc bangalore)
STOMA FULL SLIDE (probability of IISc bangalore)STOMA FULL SLIDE (probability of IISc bangalore)
STOMA FULL SLIDE (probability of IISc bangalore)2010111
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
Optimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesOptimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesEmmanuel Hadoux
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...Cristiano Longo
 

Similar to Discussion cabras-robert-130323171455-phpapp02 (20)

Lecture_9.pdf
Lecture_9.pdfLecture_9.pdf
Lecture_9.pdf
 
Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model Selection
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapter
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
 
Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)
 
Introduction to modern Variational Inference.
Introduction to modern Variational Inference.Introduction to modern Variational Inference.
Introduction to modern Variational Inference.
 
Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘 Jensen's inequality, EM 알고리즘
Jensen's inequality, EM 알고리즘
 
Conceptual Introduction to Gaussian Processes
Conceptual Introduction to Gaussian ProcessesConceptual Introduction to Gaussian Processes
Conceptual Introduction to Gaussian Processes
 
P, NP and NP-Complete, Theory of NP-Completeness V2
P, NP and NP-Complete, Theory of NP-Completeness V2P, NP and NP-Complete, Theory of NP-Completeness V2
P, NP and NP-Complete, Theory of NP-Completeness V2
 
Algebras for programming languages
Algebras for programming languagesAlgebras for programming languages
Algebras for programming languages
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Cs229 notes8
Cs229 notes8Cs229 notes8
Cs229 notes8
 
STOMA FULL SLIDE (probability of IISc bangalore)
STOMA FULL SLIDE (probability of IISc bangalore)STOMA FULL SLIDE (probability of IISc bangalore)
STOMA FULL SLIDE (probability of IISc bangalore)
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Optimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processesOptimization of probabilistic argumentation with Markov processes
Optimization of probabilistic argumentation with Markov processes
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...
Herbrand-satisfiability of a Quantified Set-theoretical Fragment (Cantone, Lo...
 
Congrès SMAI 2019
Congrès SMAI 2019Congrès SMAI 2019
Congrès SMAI 2019
 

Discussion cabras-robert-130323171455-phpapp02

  • 1. DISCUSSION of Bayesian Computation via empirical likelihood Stefano Cabras, stefano.cabras@uc3m.es Universidad Carlos III de Madrid (Spain) Universit` di Cagliari (Italy) a Padova, 21-Mar-2013
  • 2. Summary ◮ Problem:
  • 3. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ;
  • 4. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ; ◮ we want to obtain the posterior πN (θ | y ) ∝ LN (θ)π(θ).
  • 5. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ; ◮ we want to obtain the posterior πN (θ | y ) ∝ LN (θ)π(θ). ◮ BUT
  • 6. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ; ◮ we want to obtain the posterior πN (θ | y ) ∝ LN (θ)π(θ). ◮ BUT ◮ IF LN (θ) is not available: ◮ THEN all life ABC;
  • 7. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ; ◮ we want to obtain the posterior πN (θ | y ) ∝ LN (θ)π(θ). ◮ BUT ◮ IF LN (θ) is not available: ◮ THEN all life ABC; ◮ IF it is not even possible to simulate from f (y | θ):
  • 8. Summary ◮ Problem: ◮ a statistical model f (y | θ); ◮ a prior π(θ) on θ; ◮ we want to obtain the posterior πN (θ | y ) ∝ LN (θ)π(θ). ◮ BUT ◮ IF LN (θ) is not available: ◮ THEN all life ABC; ◮ IF it is not even possible to simulate from f (y | θ): ◮ THEN replace LN (θ) with LEL (θ) (the proposed BCel procedure): π(θ|y ) ∝ LEL (θ) × π(θ). .
  • 9. ... what remains about the f (y | θ) ?
  • 10. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0.
  • 11. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0. ◮ The relation between θ and obs. Y is model conditioned and expressed by h(Y , θ);
  • 12. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0. ◮ The relation between θ and obs. Y is model conditioned and expressed by h(Y , θ); ◮ Constraints are model driven and so there is still a timid trace of f (y | θ) in BCel .
  • 13. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0. ◮ The relation between θ and obs. Y is model conditioned and expressed by h(Y , θ); ◮ Constraints are model driven and so there is still a timid trace of f (y | θ) in BCel . ◮ Examples:
  • 14. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0. ◮ The relation between θ and obs. Y is model conditioned and expressed by h(Y , θ); ◮ Constraints are model driven and so there is still a timid trace of f (y | θ) in BCel . ◮ Examples: ◮ The coalescent model example is illuminating in suggesting the score of the pairwise likelihood;
  • 15. ... what remains about the f (y | θ) ? ◮ Recall that the Empirical Likelihood is defined, for iid sample, by means of a set of constraints: Ef (y |θ) [h(Y , θ)] = 0. ◮ The relation between θ and obs. Y is model conditioned and expressed by h(Y , θ); ◮ Constraints are model driven and so there is still a timid trace of f (y | θ) in BCel . ◮ Examples: ◮ The coalescent model example is illuminating in suggesting the score of the pairwise likelihood; ◮ The residuals in GARCH models.
  • 16. ... a suggestion What if we do not even known h(·) ?
  • 17. ... how to elicit h(·) automatically
  • 18. ... how to elicit h(·) automatically
  • 19. ... how to elicit h(·) automatically ◮ Set h(Y , θ) = Y − g (θ), where g (θ) = Ef (y |θ) (Y |θ), is the regression function of Y |θ;
  • 20. ... how to elicit h(·) automatically ◮ Set h(Y , θ) = Y − g (θ), where g (θ) = Ef (y |θ) (Y |θ), is the regression function of Y |θ; ◮ g (θ) should be replaced by an estimator g (θ).
  • 21. How to estimate g (θ) ? 1 ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras, Castellanos, Ruli (Ercim-2012, Oviedo).
  • 22. How to estimate g (θ) ? ◮ Use a once forever pilot-run simulation study: 1 1 ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras, Castellanos, Ruli (Ercim-2012, Oviedo).
  • 23. How to estimate g (θ) ? ◮ Use a once forever pilot-run simulation study: 1 1. Consider a grid (or regular lattice) of θ made by M points: θ1 , . . . , θM 1 ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras, Castellanos, Ruli (Ercim-2012, Oviedo).
  • 24. How to estimate g (θ) ? ◮ Use a once forever pilot-run simulation study: 1 1. Consider a grid (or regular lattice) of θ made by M points: θ1 , . . . , θM 2. Simulate the corresponding y1 , . . . , yM 1 ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras, Castellanos, Ruli (Ercim-2012, Oviedo).
  • 25. How to estimate g (θ) ? ◮ Use a once forever pilot-run simulation study: 1 1. Consider a grid (or regular lattice) of θ made by M points: θ1 , . . . , θM 2. Simulate the corresponding y1 , . . . , yM 3. Regress y1 , . . . , yM on θ 1 , . . . , θ M obtaining g (θ). 1 ... similar to Fearnhead, P. and D. Prangle (JRRS-B, 2012) or Cabras, Castellanos, Ruli (Ercim-2012, Oviedo).
  • 26. ... example: y ∼ N(|θ|, 1) For a pilot run of M = 1000 we have g (θ) = |θ|. ˆ Pilot−Run s.s. g(θ) 10 y 5 0 −10 −5 0 5 10 θ
  • 27. ... example: y ∼ N(|θ|, 1) Suppose to draw a n = 100 sample from θ = 2: Histogram of y 20 15 Frequency 10 5 0 0 1 2 3 4 y
  • 28. ... example: y ∼ N(|θ|, 1) The Empirical Likelihood is this 2.5 2.0 Emp. Lik. 1.5 1.0 −4 −2 0 2 4 θ
  • 29. 1st Point: Do we need necessarily have to use f (y | θ) ?
  • 30. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal;
  • 31. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal; ◮ How this is reflected in the BCel ?
  • 32. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal; ◮ How this is reflected in the BCel ? ◮ For a given data y;
  • 33. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal; ◮ How this is reflected in the BCel ? ◮ For a given data y; ◮ and h(Y , θ) fixed;
  • 34. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal; ◮ How this is reflected in the BCel ? ◮ For a given data y; ◮ and h(Y , θ) fixed; ◮ the LEL (θ) is the same regardless of f (y | θ).
  • 35. 1st Point: Do we need necessarily have to use f (y | θ) ? ◮ The above data maybe drawn from a (e.g.) a Half Normal; ◮ How this is reflected in the BCel ? ◮ For a given data y; ◮ and h(Y , θ) fixed; ◮ the LEL (θ) is the same regardless of f (y | θ). Can we ignore f (y | θ) ?
  • 36. 2nd Point: Sample free vs Simulation free
  • 37. 2nd Point: Sample free vs Simulation free ◮ The Empirical Likelihood is ”simulation free” but not ”sample free”, i.e.
  • 38. 2nd Point: Sample free vs Simulation free ◮ The Empirical Likelihood is ”simulation free” but not ”sample free”, i.e. ◮ LEL (θ) → LN (θ) for n → ∞, ◮ implying π(θ|y) → πN (θ | y ) asymptotically in n.
  • 39. 2nd Point: Sample free vs Simulation free ◮ The Empirical Likelihood is ”simulation free” but not ”sample free”, i.e. ◮ LEL (θ) → LN (θ) for n → ∞, ◮ implying π(θ|y) → πN (θ | y ) asymptotically in n. ◮ The ABC is ”sample free” but not ”simulation free”, i.e.
  • 40. 2nd Point: Sample free vs Simulation free ◮ The Empirical Likelihood is ”simulation free” but not ”sample free”, i.e. ◮ LEL (θ) → LN (θ) for n → ∞, ◮ implying π(θ|y) → πN (θ | y ) asymptotically in n. ◮ The ABC is ”sample free” but not ”simulation free”, i.e. ◮ π(θ|ρ(s(y), so bs) < ǫ) → πN (θ | y ) as ǫ → 0 ◮ implying convergence in the number of simulations if s(y ) were sufficient.
  • 41. 2nd Point: Sample free vs Simulation free ◮ The Empirical Likelihood is ”simulation free” but not ”sample free”, i.e. ◮ LEL (θ) → LN (θ) for n → ∞, ◮ implying π(θ|y) → πN (θ | y ) asymptotically in n. ◮ The ABC is ”sample free” but not ”simulation free”, i.e. ◮ π(θ|ρ(s(y), so bs) < ǫ) → πN (θ | y ) as ǫ → 0 ◮ implying convergence in the number of simulations if s(y ) were sufficient. A quick answer recommends use BCel BUT a small sample would recommend ABC ?
  • 42. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ?
  • 43. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting:
  • 44. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods:
  • 45. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) ◮ Mengersen et al. (PNAS, 2012) ◮ ...
  • 46. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) ◮ Mengersen et al. (PNAS, 2012) ◮ ... ◮ Modified-Likelihoods:
  • 47. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) ◮ Mengersen et al. (PNAS, 2012) ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ...
  • 48. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) ◮ Mengersen et al. (PNAS, 2012) ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ... ◮ Quasi-Likelihoods:
  • 49. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) ◮ Mengersen et al. (PNAS, 2012) ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) ◮ Greco et al. (JSPI, 2008) ◮ Ventura et al. (JSPI, 2010) ◮ ...
  • 50. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) : examples and coverages of C.I. ◮ Mengersen et al. (PNAS, 2012) ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) ◮ Greco et al. (JSPI, 2008) ◮ Ventura et al. (JSPI, 2010) ◮ ...
  • 51. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) : examples and coverages of C.I. ◮ Mengersen et al. (PNAS, 2012) : examples and coverages of C.I. ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) ◮ Greco et al. (JSPI, 2008) ◮ Ventura et al. (JSPI, 2010) ◮ ...
  • 52. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) : examples and coverages of C.I. ◮ Mengersen et al. (PNAS, 2012) : examples and coverages of C.I. ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) : second order matching properties; ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) ◮ Greco et al. (JSPI, 2008) ◮ Ventura et al. (JSPI, 2010) ◮ ...
  • 53. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) : examples and coverages of C.I. ◮ Mengersen et al. (PNAS, 2012) : examples and coverages of C.I. ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) : second order matching properties; ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) : examples; ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) ◮ Greco et al. (JSPI, 2008) ◮ Ventura et al. (JSPI, 2010) ◮ ...
  • 54. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ The use of pseudo-likelihoods is not new in the Bayesian setting: ◮ Empirical Likelihoods: ◮ Lazar (Biometrika, 2003) : examples and coverages of C.I. ◮ Mengersen et al. (PNAS, 2012) : examples and coverages of C.I. ◮ ... ◮ Modified-Likelihoods: ◮ Ventura et al. (JASA, 2009) : second order matching properties; ◮ Chang and Mukerjee (Stat. & Prob. Letters 2006) : examples; ◮ ... ◮ Quasi-Likelihoods: ◮ Lin (Statist. Methodol., 2006) : examples; ◮ Greco et al. (JSPI, 2008) : robustness properties; ◮ Ventura et al. (JSPI, 2010) : examples and coverages of C.I.; ◮ ...
  • 55. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ Monahan & Boos (Biometrika, 1992) proposed a notion of validity:
  • 56. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ Monahan & Boos (Biometrika, 1992) proposed a notion of validity: π(θ|y ) should obey the laws of probability in a fashion that is consistent with statements derived from Bayes’rule.
  • 57. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ Monahan & Boos (Biometrika, 1992) proposed a notion of validity: π(θ|y ) should obey the laws of probability in a fashion that is consistent with statements derived from Bayes’rule. ◮ Very difficult!
  • 58. 3nd Point: How to validate a pseudo-posterior π(θ|y ) ∝ LEL (θ) × π(θ) ? ◮ Monahan & Boos (Biometrika, 1992) proposed a notion of validity: π(θ|y ) should obey the laws of probability in a fashion that is consistent with statements derived from Bayes’rule. ◮ Very difficult! How to validate the pseudo-posterior π(θ|y ) when this is not possible ?
  • 59. ... Last point: the ABC is still a terrific tool
  • 60. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references:
  • 61. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references: ◮ Statistical Journals;
  • 62. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references: ◮ Statistical Journals; ◮ Twitter;
  • 63. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references: ◮ Statistical Journals; ◮ Twitter; ◮ Xiang’s blog ( xianblog.wordpress.com )
  • 64. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references: ◮ Statistical Journals; ◮ Twitter; ◮ Xiang’s blog ( xianblog.wordpress.com ) ◮ ... it is tailored to Approximate LN (θ).
  • 65. ... Last point: the ABC is still a terrific tool ◮ ... a lot of references: ◮ Statistical Journals; ◮ Twitter; ◮ Xiang’s blog ( xianblog.wordpress.com ) ◮ ... it is tailored to Approximate LN (θ). Where is the A in BCel ?