SlideShare a Scribd company logo
A gentle introduction to BNP
Part I
Antonio Canale
Universit`a di Torino &
Collegio Carlo Alberto
StaTalk on BNP, 19/02/16
Introduction The Dirichlet process Nonparametric mixture models
Outline of the talk(s)
1 Why BNP? (A)
2 The Dirichlet process (A)
3 Nonparametric mixture models (A)
4 Beyond the DP (J)
5 Species sampling processes (J)
6 Completely random measures (J)
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
Why Bayesian nonparametrics (BNP)?
Why nonparametric?
• We don’t want to strictly impose any model but let the data speak;
• The idea of a true model governed by relatively few parameters is
unrealistic;
Why Bayesian?
• If we have a reasonable guess for what is the true model we want
to use this prior knowledge.
• Large support and consistency are interesting concepts related to
priors on infinite dimensional spaces (Pierpaolo’s talk in the
afternoon)
BNP is to fit a single model that can adapt its complexity to the
data.
Introduction The Dirichlet process Nonparametric mixture models
How Bayesian and nonparametric?
Define F the space of densities and let P ∈ F. A Bayesian analysis
starts with
y ∼ P
P ∼ π
where π is a measure on the space F.
Hence BNP is infinitely parametric.
Introduction The Dirichlet process Nonparametric mixture models
The Dirichlet distribution
• Start with independent Zj ∼ Ga(αj , 1), for j = 1, . . . , k (αj > 0)
• Define
πj =
Zj
k
j=1 Zj
;
• Then (π1, . . . , πk) ∼ Dir(α1, . . . , αk);
• The Dirichlet distribution is a distribution over the K-dimensional
probability simplex:
∆k = {(π1, . . . , πk) : πj > 0,
j
πj = 1}
Introduction The Dirichlet process Nonparametric mixture models
The Dirichlet distribution
• Probability density
p(π1, . . . , πk|α) =
Γ( j αj )
j Γ(αj )
j
π
αj −1
j
Introduction The Dirichlet process Nonparametric mixture models
The Dirichlet distribution in Bayesian statistics
Dirichlet distribution is conjugate to the multinomial likelihood, hence
if
π ∼ Dir(α)
y|π ∼ Multinomial(π)
p(y = j|π) = πj ,
then we have
p(π|y = j, α) = Dir(ˆα)
where ˆαj = αj + 1, ˆαi = αi for each i = j.
Introduction The Dirichlet process Nonparametric mixture models
Agglomerative property of Dirichlet distributions
• Combining entries by their sum
(π1, . . . , πk) ∼ Dir(α1, . . . , αk)
→ (π1, . . . , πi + πj , . . . , πk) ∼ Dir(α1, . . . , αi + αj , . . . , αk)
• Marginals follow Beta distributions, πj ∼ beta(αj , h=j αh).
Introduction The Dirichlet process Nonparametric mixture models
1 Introduction
2 The Dirichlet process
3 Nonparametric mixture models
Introduction The Dirichlet process Nonparametric mixture models
Ferguson (1973) definition of the Dirichlet process
Definition
• P is a random probability measure over (Y, B(Y)).
• F is the whole space of probability measures on (Y, B(Y)), so
P ∈ F.
• Let α ∈ R+ and P0 ∈ F.
• P ∼ DP(α, P0) iff for any n and any partition B1, . . . , Bn of Y
(P(B1), P(B2), . . . , P(Bn)) ∼ Dir(αP0(B1), αP0(B2), . . . , αP0(Bn))
The DP is a distribution of random probability distributions.
Introduction The Dirichlet process Nonparametric mixture models
Interpretation
If P ∼ DP(α, P0), then for any measurable A
• E(P(A)) = P0(A)
• Var(P(A)) = P0(A){1 − P0(A)}/(1 + α)
Introduction The Dirichlet process Nonparametric mixture models
Density estimation using DP priors
If yi
iid
∼ P for i = 1, . . . , n and P ∼ DP(α, P0) a priori then,
P|y ∼ DP n + α,
1
α + n
n
i=1
δyi +
α
α + n
P0
Introduction The Dirichlet process Nonparametric mixture models
Density estimation using DP priors
−6 −4 −2 0 2 4 6
0.00.20.40.60.81.0
x
−6 −4 −2 0 2 4 6
0.00.20.40.60.81.0
x
Figure: Black true density (N(1, 2)), blue base measure (N(0,1)), green
dashed ECDF, blue dashed posterior DP. First plot n = 10, second n = 50.
Introduction The Dirichlet process Nonparametric mixture models
Stick-breaking
An alternative representation of the DP is related to the so called
stick-breaking process:
Introduction The Dirichlet process Nonparametric mixture models
Stick-breaking representation of the DP
To obtain P ∼ DP(αP0):
• Draw a sequence of Beta random variables Vj
iid
∼ Beta(1, α).
• Define a sequence of weights as πj = Vj l<j (1 − Vl )
• Draw independent θ
iid
∼ P0
• Define
P =
∞
j=1
πj δθj
Introduction The Dirichlet process Nonparametric mixture models
Stochastic processes and chinese restaurants. . .
Imagine a Chinese restaurant with countably infinitely many tables,
labelled 1, 2, . . .
Customers walk in and sit down at some table. The tables are chosen
according to the following random process.
1 The first customer sits at table 1;
2 The n-th customer chooses the first unoccupied table with
probability α/(α + n − 1) and an occupied table with probability
nj /(α + n − 1), where nj is the number of people sitting at that
table.
Introduction The Dirichlet process Nonparametric mixture models
CRP or Polya urn construction of the DP
If θ
iid
∼ P0 and P ∼ DP(αP0), integrate out P and obtain
pr(θi |θ1, . . . , θi−1) =
j
nj
n + α
δθj
+
α
n + α
P0.
Obtaining that (θ1, . . . , θn) ∼ PU(αP0).
Introduction The Dirichlet process Nonparametric mixture models
Considerations
• Draw from a DP are a.s. discrete
• Unappealing if y is continuous, useful if y is discrete? (no, but
wait for my afternoon talk)
Introduction The Dirichlet process Nonparametric mixture models
Considerations
• Draw from a DP are a.s. discrete
• Unappealing if y is continuous, useful if y is discrete? (no, but
wait for my afternoon talk)
Introduction The Dirichlet process Nonparametric mixture models
Finite mixture models
Assume the following model
yi ∼ N(µSi
, σ2
Si
), pr(Si = h) = πh
with likelihood
f (y|µ, σ2
, π) =
k
j=1
πj φ(y; µj , σ2
j )
and prior
(µ, σ2
) ∼ P0, π ∼ Dir(α);
Introduction The Dirichlet process Nonparametric mixture models
FMM applications: density estimation
• With enough components, a mixture of
Gaussian can approximate any
continuous distribution.
• If the number of components equals n
we have the kernel density estimation.
0 1 2 3 4 5 6
0.00.10.20.30.40.5
geyser$duration
Probabilitydensityfunction
Introduction The Dirichlet process Nonparametric mixture models
FMM applications: model-based clustering
• Divide observations into homogeneus
clusters
• “Homogeneus” depends on what
kernel (Gaussian in previous slide)
• With Gaussian kernel, there are two
clusters in Iris dataset (truth is three!)
• See discussions in Petralia et al.
(2012), Canale and Scarpa (2015) and
Canale and De Blasi (2015)
q
q
q
q
q
q
q q
q
q
q
q
qq
q
q
q
q
qq
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
qq
q
q
qq
q
q
q
q
q
q
q
qq
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
qq
q q
q
q
q
q
q
q
q
q
q
q
q
q
q q
q
q
q
q
q
q
q q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q
q q
q
q
q
q
q
q
q
q
q
qq
q
q
q
q
q
qq q
q
q
q
q
q
q
q
q
4.5 5.0 5.5 6.0 6.5 7.0 7.5 8.0
2.02.53.03.54.0
iris$Sepal.Length
iris$Sepal.Width
Introduction The Dirichlet process Nonparametric mixture models
Infinite mixture models
• A more elegant way to write the finite mixture model is
f (y) = K(y; θ)dP(θ), P =
K
j=1
ωj δθj
,
where K(·; θ) is a general kernel (e.g. normal) parametrized by θ.
• Clearly a prior on the weights and on the parameters of the kernel
is equivalent to a prior on the finite disrete measure P.
• From FMM to IMM ⇒ P ∼ DP(αP0)!
Introduction The Dirichlet process Nonparametric mixture models
DP mixture models
• The model and prior are
y ∼ f , f (y) = K(y; θ)dP(θ), P ∼ DP(αP0).
where K(·; θ) is a general kernel (e.g. normal) parametrized by θ.
• Consider the DPM prior as a “smoothed version” of the DP prior
(just like the kernel density estimation is a smoothed version of the
histogram)
• Widely used for continuous distribution.
Introduction The Dirichlet process Nonparametric mixture models
Hyerarchical representation
Using a hyerarchical representation the mixture model can be
expressed as
yi | θi ∼ K(y; θi )
θi ∼ P
P ∼ DP(αP0).
Introduction The Dirichlet process Nonparametric mixture models
Mixture of Gaussians
• Gold standard for density estimation;
• can approximate any continuous distribution (Lo, 1984; Escobar
and West, 1995);
• large support and good frequentist properties (Ghosal et al., 1999).
The model and the prior are
f (y) = N(y; µ, τ−1
)dP(µ, τ−1
),
P ∼ DP(αP0),
where N(y; µ, τ−1) is a normal kernel having mean µ and precision τ,
P0 Normal-Gamma, for conjugacy.
Introduction The Dirichlet process Nonparametric mixture models
Mixture of Gaussians
yi | µi , τi ∼ N(µi , τ−1
i )
(µi , τi ) ∼ P
P ∼ DP(αP0).
Introduction The Dirichlet process Nonparametric mixture models
Complex data
• Mixture models can be used also when we have complex (modern)
data
• An example is functional data f1, . . . , fn
fi (t) = η(t) + it,
where η is a smooth function in t and it are random noises.
• we can model these data with
fi | ηi ∼ N(ηi , σ2
)
ηi ∼ P
P ∼ DP(αP0).

More Related Content

What's hot

An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
Christian Robert
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in Statistics
Christian Robert
 
4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar
Christian Robert
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
Christian Robert
 
Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model Selection
Michael Stumpf
 
Jsm09 talk
Jsm09 talkJsm09 talk
Jsm09 talk
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Lecture10 - Naïve Bayes
Lecture10 - Naïve BayesLecture10 - Naïve Bayes
Lecture10 - Naïve Bayes
Albert Orriols-Puig
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Martin Pelikan
 
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
Andres Mendez-Vazquez
 
Chapter 2 pertubation
Chapter 2 pertubationChapter 2 pertubation
Chapter 2 pertubation
NBER
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
comments on exponential ergodicity of the bouncy particle sampler
comments on exponential ergodicity of the bouncy particle samplercomments on exponential ergodicity of the bouncy particle sampler
comments on exponential ergodicity of the bouncy particle sampler
Christian Robert
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systems
recsysfr
 
Dictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix FactorizationDictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix Factorization
recsysfr
 
45th SIS Meeting, Padova, Italy
45th SIS Meeting, Padova, Italy45th SIS Meeting, Padova, Italy
45th SIS Meeting, Padova, Italy
Christian Robert
 
A Maximum Entropy Approach to the Loss Data Aggregation Problem
A Maximum Entropy Approach to the Loss Data Aggregation ProblemA Maximum Entropy Approach to the Loss Data Aggregation Problem
A Maximum Entropy Approach to the Loss Data Aggregation Problem
Erika G. G.
 
31 Machine Learning Unsupervised Cluster Validity
31 Machine Learning Unsupervised Cluster Validity31 Machine Learning Unsupervised Cluster Validity
31 Machine Learning Unsupervised Cluster Validity
Andres Mendez-Vazquez
 
Asynchronous Stochastic Optimization, New Analysis and Algorithms
Asynchronous Stochastic Optimization, New Analysis and AlgorithmsAsynchronous Stochastic Optimization, New Analysis and Algorithms
Asynchronous Stochastic Optimization, New Analysis and Algorithms
Fabian Pedregosa
 

What's hot (20)

An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in Statistics
 
4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar4th joint Warwick Oxford Statistics Seminar
4th joint Warwick Oxford Statistics Seminar
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
Considerate Approaches to ABC Model Selection
Considerate Approaches to ABC Model SelectionConsiderate Approaches to ABC Model Selection
Considerate Approaches to ABC Model Selection
 
Jsm09 talk
Jsm09 talkJsm09 talk
Jsm09 talk
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Lecture10 - Naïve Bayes
Lecture10 - Naïve BayesLecture10 - Naïve Bayes
Lecture10 - Naïve Bayes
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
Pairwise and Problem-Specific Distance Metrics in the Linkage Tree Genetic Al...
 
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
28 Dealing with the NP Poblems: Exponential Search and Approximation Algorithms
 
Chapter 2 pertubation
Chapter 2 pertubationChapter 2 pertubation
Chapter 2 pertubation
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
comments on exponential ergodicity of the bouncy particle sampler
comments on exponential ergodicity of the bouncy particle samplercomments on exponential ergodicity of the bouncy particle sampler
comments on exponential ergodicity of the bouncy particle sampler
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systems
 
Dictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix FactorizationDictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix Factorization
 
45th SIS Meeting, Padova, Italy
45th SIS Meeting, Padova, Italy45th SIS Meeting, Padova, Italy
45th SIS Meeting, Padova, Italy
 
A Maximum Entropy Approach to the Loss Data Aggregation Problem
A Maximum Entropy Approach to the Loss Data Aggregation ProblemA Maximum Entropy Approach to the Loss Data Aggregation Problem
A Maximum Entropy Approach to the Loss Data Aggregation Problem
 
31 Machine Learning Unsupervised Cluster Validity
31 Machine Learning Unsupervised Cluster Validity31 Machine Learning Unsupervised Cluster Validity
31 Machine Learning Unsupervised Cluster Validity
 
Asynchronous Stochastic Optimization, New Analysis and Algorithms
Asynchronous Stochastic Optimization, New Analysis and AlgorithmsAsynchronous Stochastic Optimization, New Analysis and Algorithms
Asynchronous Stochastic Optimization, New Analysis and Algorithms
 

Viewers also liked

Nested sampling
Nested samplingNested sampling
Nested sampling
Christian Robert
 
Presentation of Bassoum Abou on Stein's 1981 AoS paper
Presentation of Bassoum Abou on Stein's 1981 AoS paperPresentation of Bassoum Abou on Stein's 1981 AoS paper
Presentation of Bassoum Abou on Stein's 1981 AoS paper
Christian Robert
 
Asymptotics for discrete random measures
Asymptotics for discrete random measuresAsymptotics for discrete random measures
Asymptotics for discrete random measures
Julyan Arbel
 
Species sampling models in Bayesian Nonparametrics
Species sampling models in Bayesian NonparametricsSpecies sampling models in Bayesian Nonparametrics
Species sampling models in Bayesian Nonparametrics
Julyan Arbel
 
Bayesian Classics
Bayesian ClassicsBayesian Classics
Bayesian Classics
Julyan Arbel
 
Gelfand and Smith (1990), read by
Gelfand and Smith (1990), read byGelfand and Smith (1990), read by
Gelfand and Smith (1990), read by
Christian Robert
 
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
Christian Robert
 
Reading Birnbaum's (1962) paper, by Li Chenlu
Reading Birnbaum's (1962) paper, by Li ChenluReading Birnbaum's (1962) paper, by Li Chenlu
Reading Birnbaum's (1962) paper, by Li Chenlu
Christian Robert
 
Reading Neyman's 1933
Reading Neyman's 1933 Reading Neyman's 1933
Reading Neyman's 1933
Christian Robert
 
Testing point null hypothesis, a discussion by Amira Mziou
Testing point null hypothesis, a discussion by Amira MziouTesting point null hypothesis, a discussion by Amira Mziou
Testing point null hypothesis, a discussion by Amira Mziou
Christian Robert
 
Reading Efron's 1979 paper on bootstrap
Reading Efron's 1979 paper on bootstrapReading Efron's 1979 paper on bootstrap
Reading Efron's 1979 paper on bootstrap
Christian Robert
 
Reading the Lasso 1996 paper by Robert Tibshirani
Reading the Lasso 1996 paper by Robert TibshiraniReading the Lasso 1996 paper by Robert Tibshirani
Reading the Lasso 1996 paper by Robert Tibshirani
Christian Robert
 
slides Céline Beji
slides Céline Bejislides Céline Beji
slides Céline Beji
Christian Robert
 
Reading the Lindley-Smith 1973 paper on linear Bayes estimators
Reading the Lindley-Smith 1973 paper on linear Bayes estimatorsReading the Lindley-Smith 1973 paper on linear Bayes estimators
Reading the Lindley-Smith 1973 paper on linear Bayes estimators
Christian Robert
 

Viewers also liked (14)

Nested sampling
Nested samplingNested sampling
Nested sampling
 
Presentation of Bassoum Abou on Stein's 1981 AoS paper
Presentation of Bassoum Abou on Stein's 1981 AoS paperPresentation of Bassoum Abou on Stein's 1981 AoS paper
Presentation of Bassoum Abou on Stein's 1981 AoS paper
 
Asymptotics for discrete random measures
Asymptotics for discrete random measuresAsymptotics for discrete random measures
Asymptotics for discrete random measures
 
Species sampling models in Bayesian Nonparametrics
Species sampling models in Bayesian NonparametricsSpecies sampling models in Bayesian Nonparametrics
Species sampling models in Bayesian Nonparametrics
 
Bayesian Classics
Bayesian ClassicsBayesian Classics
Bayesian Classics
 
Gelfand and Smith (1990), read by
Gelfand and Smith (1990), read byGelfand and Smith (1990), read by
Gelfand and Smith (1990), read by
 
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
Reading Testing a point-null hypothesis, by Jiahuan Li, Feb. 25, 2013
 
Reading Birnbaum's (1962) paper, by Li Chenlu
Reading Birnbaum's (1962) paper, by Li ChenluReading Birnbaum's (1962) paper, by Li Chenlu
Reading Birnbaum's (1962) paper, by Li Chenlu
 
Reading Neyman's 1933
Reading Neyman's 1933 Reading Neyman's 1933
Reading Neyman's 1933
 
Testing point null hypothesis, a discussion by Amira Mziou
Testing point null hypothesis, a discussion by Amira MziouTesting point null hypothesis, a discussion by Amira Mziou
Testing point null hypothesis, a discussion by Amira Mziou
 
Reading Efron's 1979 paper on bootstrap
Reading Efron's 1979 paper on bootstrapReading Efron's 1979 paper on bootstrap
Reading Efron's 1979 paper on bootstrap
 
Reading the Lasso 1996 paper by Robert Tibshirani
Reading the Lasso 1996 paper by Robert TibshiraniReading the Lasso 1996 paper by Robert Tibshirani
Reading the Lasso 1996 paper by Robert Tibshirani
 
slides Céline Beji
slides Céline Bejislides Céline Beji
slides Céline Beji
 
Reading the Lindley-Smith 1973 paper on linear Bayes estimators
Reading the Lindley-Smith 1973 paper on linear Bayes estimatorsReading the Lindley-Smith 1973 paper on linear Bayes estimators
Reading the Lindley-Smith 1973 paper on linear Bayes estimators
 

Similar to A Gentle Introduction to Bayesian Nonparametrics

Chinese Restaurant Process
Chinese Restaurant ProcessChinese Restaurant Process
Chinese Restaurant Process
Mohitdeep Singh
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximations
Christian Robert
 
Explaining the Basics of Mean Field Variational Approximation for Statisticians
Explaining the Basics of Mean Field Variational Approximation for StatisticiansExplaining the Basics of Mean Field Variational Approximation for Statisticians
Explaining the Basics of Mean Field Variational Approximation for Statisticians
Wayne Lee
 
Learn from Example and Learn Probabilistic Model
Learn from Example and Learn Probabilistic ModelLearn from Example and Learn Probabilistic Model
Learn from Example and Learn Probabilistic Model
Junya Tanaka
 
Integrated modelling Cape Town
Integrated modelling Cape TownIntegrated modelling Cape Town
Integrated modelling Cape Town
Bob O'Hara
 
Discussion of Persi Diaconis' lecture at ISBA 2016
Discussion of Persi Diaconis' lecture at ISBA 2016Discussion of Persi Diaconis' lecture at ISBA 2016
Discussion of Persi Diaconis' lecture at ISBA 2016
Christian Robert
 
20230213_ComputerVision_연구.pptx
20230213_ComputerVision_연구.pptx20230213_ComputerVision_연구.pptx
20230213_ComputerVision_연구.pptx
ssuser7807522
 
Principle of Maximum Entropy
Principle of Maximum EntropyPrinciple of Maximum Entropy
Principle of Maximum Entropy
Jiawang Liu
 
ilp-nlp-slides.pdf
ilp-nlp-slides.pdfilp-nlp-slides.pdf
ilp-nlp-slides.pdf
FlorentBersani
 
Linear models for data science
Linear models for data scienceLinear models for data science
Linear models for data science
Brad Klingenberg
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet Allocation
Marco Righini
 
Building useful models for imbalanced datasets (without resampling)
Building useful models for imbalanced datasets (without resampling)Building useful models for imbalanced datasets (without resampling)
Building useful models for imbalanced datasets (without resampling)
Greg Landrum
 
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
The Statistical and Applied Mathematical Sciences Institute
 
ngboost.pptx
ngboost.pptxngboost.pptx
ngboost.pptx
Hadrian7
 
Statistics for CEGEP Biology
Statistics for CEGEP BiologyStatistics for CEGEP Biology
Statistics for CEGEP Biology
Corey Chivers
 
Lecture17 xing fei-fei
Lecture17 xing fei-feiLecture17 xing fei-fei
Lecture17 xing fei-fei
Tianlu Wang
 
Joint contrastive learning with infinite possibilities
Joint contrastive learning with infinite possibilitiesJoint contrastive learning with infinite possibilities
Joint contrastive learning with infinite possibilities
taeseon ryu
 
Yet another statistical analysis of the data of the ‘loophole free’ experime...
Yet another statistical analysis of the data of the  ‘loophole free’ experime...Yet another statistical analysis of the data of the  ‘loophole free’ experime...
Yet another statistical analysis of the data of the ‘loophole free’ experime...
Richard Gill
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
RayKim51
 
Seminaire ihp
Seminaire ihpSeminaire ihp
Seminaire ihp
Julyan Arbel
 

Similar to A Gentle Introduction to Bayesian Nonparametrics (20)

Chinese Restaurant Process
Chinese Restaurant ProcessChinese Restaurant Process
Chinese Restaurant Process
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximations
 
Explaining the Basics of Mean Field Variational Approximation for Statisticians
Explaining the Basics of Mean Field Variational Approximation for StatisticiansExplaining the Basics of Mean Field Variational Approximation for Statisticians
Explaining the Basics of Mean Field Variational Approximation for Statisticians
 
Learn from Example and Learn Probabilistic Model
Learn from Example and Learn Probabilistic ModelLearn from Example and Learn Probabilistic Model
Learn from Example and Learn Probabilistic Model
 
Integrated modelling Cape Town
Integrated modelling Cape TownIntegrated modelling Cape Town
Integrated modelling Cape Town
 
Discussion of Persi Diaconis' lecture at ISBA 2016
Discussion of Persi Diaconis' lecture at ISBA 2016Discussion of Persi Diaconis' lecture at ISBA 2016
Discussion of Persi Diaconis' lecture at ISBA 2016
 
20230213_ComputerVision_연구.pptx
20230213_ComputerVision_연구.pptx20230213_ComputerVision_연구.pptx
20230213_ComputerVision_연구.pptx
 
Principle of Maximum Entropy
Principle of Maximum EntropyPrinciple of Maximum Entropy
Principle of Maximum Entropy
 
ilp-nlp-slides.pdf
ilp-nlp-slides.pdfilp-nlp-slides.pdf
ilp-nlp-slides.pdf
 
Linear models for data science
Linear models for data scienceLinear models for data science
Linear models for data science
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet Allocation
 
Building useful models for imbalanced datasets (without resampling)
Building useful models for imbalanced datasets (without resampling)Building useful models for imbalanced datasets (without resampling)
Building useful models for imbalanced datasets (without resampling)
 
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
Deep Learning Opening Workshop - Improving Generative Models - Junier Oliva, ...
 
ngboost.pptx
ngboost.pptxngboost.pptx
ngboost.pptx
 
Statistics for CEGEP Biology
Statistics for CEGEP BiologyStatistics for CEGEP Biology
Statistics for CEGEP Biology
 
Lecture17 xing fei-fei
Lecture17 xing fei-feiLecture17 xing fei-fei
Lecture17 xing fei-fei
 
Joint contrastive learning with infinite possibilities
Joint contrastive learning with infinite possibilitiesJoint contrastive learning with infinite possibilities
Joint contrastive learning with infinite possibilities
 
Yet another statistical analysis of the data of the ‘loophole free’ experime...
Yet another statistical analysis of the data of the  ‘loophole free’ experime...Yet another statistical analysis of the data of the  ‘loophole free’ experime...
Yet another statistical analysis of the data of the ‘loophole free’ experime...
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
 
Seminaire ihp
Seminaire ihpSeminaire ihp
Seminaire ihp
 

More from Julyan Arbel

UCD_talk_nov_2020
UCD_talk_nov_2020UCD_talk_nov_2020
UCD_talk_nov_2020
Julyan Arbel
 
Bayesian neural networks increasingly sparsify their units with depth
Bayesian neural networks increasingly sparsify their units with depthBayesian neural networks increasingly sparsify their units with depth
Bayesian neural networks increasingly sparsify their units with depth
Julyan Arbel
 
Lindley smith 1972
Lindley smith 1972Lindley smith 1972
Lindley smith 1972
Julyan Arbel
 
Berger 2000
Berger 2000Berger 2000
Berger 2000
Julyan Arbel
 
Seneta 1993
Seneta 1993Seneta 1993
Seneta 1993
Julyan Arbel
 
Lehmann 1990
Lehmann 1990Lehmann 1990
Lehmann 1990
Julyan Arbel
 
Diaconis Ylvisaker 1985
Diaconis Ylvisaker 1985Diaconis Ylvisaker 1985
Diaconis Ylvisaker 1985
Julyan Arbel
 
Hastings 1970
Hastings 1970Hastings 1970
Hastings 1970
Julyan Arbel
 
Jefferys Berger 1992
Jefferys Berger 1992Jefferys Berger 1992
Jefferys Berger 1992
Julyan Arbel
 
Bayesian Classics
Bayesian ClassicsBayesian Classics
Bayesian Classics
Julyan Arbel
 
R in latex
R in latexR in latex
R in latex
Julyan Arbel
 
Arbel oviedo
Arbel oviedoArbel oviedo
Arbel oviedo
Julyan Arbel
 
Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)
Julyan Arbel
 
Causesof effects
Causesof effectsCausesof effects
Causesof effects
Julyan Arbel
 
Bayesian adaptive optimal estimation using a sieve prior
Bayesian adaptive optimal estimation using a sieve priorBayesian adaptive optimal estimation using a sieve prior
Bayesian adaptive optimal estimation using a sieve prior
Julyan Arbel
 

More from Julyan Arbel (15)

UCD_talk_nov_2020
UCD_talk_nov_2020UCD_talk_nov_2020
UCD_talk_nov_2020
 
Bayesian neural networks increasingly sparsify their units with depth
Bayesian neural networks increasingly sparsify their units with depthBayesian neural networks increasingly sparsify their units with depth
Bayesian neural networks increasingly sparsify their units with depth
 
Lindley smith 1972
Lindley smith 1972Lindley smith 1972
Lindley smith 1972
 
Berger 2000
Berger 2000Berger 2000
Berger 2000
 
Seneta 1993
Seneta 1993Seneta 1993
Seneta 1993
 
Lehmann 1990
Lehmann 1990Lehmann 1990
Lehmann 1990
 
Diaconis Ylvisaker 1985
Diaconis Ylvisaker 1985Diaconis Ylvisaker 1985
Diaconis Ylvisaker 1985
 
Hastings 1970
Hastings 1970Hastings 1970
Hastings 1970
 
Jefferys Berger 1992
Jefferys Berger 1992Jefferys Berger 1992
Jefferys Berger 1992
 
Bayesian Classics
Bayesian ClassicsBayesian Classics
Bayesian Classics
 
R in latex
R in latexR in latex
R in latex
 
Arbel oviedo
Arbel oviedoArbel oviedo
Arbel oviedo
 
Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)Poster DDP (BNP 2011 Veracruz)
Poster DDP (BNP 2011 Veracruz)
 
Causesof effects
Causesof effectsCausesof effects
Causesof effects
 
Bayesian adaptive optimal estimation using a sieve prior
Bayesian adaptive optimal estimation using a sieve priorBayesian adaptive optimal estimation using a sieve prior
Bayesian adaptive optimal estimation using a sieve prior
 

Recently uploaded

End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024
Lars Albertsson
 
一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理
aqzctr7x
 
Everything you wanted to know about LIHTC
Everything you wanted to know about LIHTCEverything you wanted to know about LIHTC
Everything you wanted to know about LIHTC
Roger Valdez
 
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
Timothy Spann
 
The Ipsos - AI - Monitor 2024 Report.pdf
The  Ipsos - AI - Monitor 2024 Report.pdfThe  Ipsos - AI - Monitor 2024 Report.pdf
The Ipsos - AI - Monitor 2024 Report.pdf
Social Samosa
 
Challenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more importantChallenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more important
Sm321
 
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
sameer shah
 
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
apvysm8
 
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdfUdemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Fernanda Palhano
 
Influence of Marketing Strategy and Market Competition on Business Plan
Influence of Marketing Strategy and Market Competition on Business PlanInfluence of Marketing Strategy and Market Competition on Business Plan
Influence of Marketing Strategy and Market Competition on Business Plan
jerlynmaetalle
 
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
74nqk8xf
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
g4dpvqap0
 
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
v7oacc3l
 
DSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelinesDSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelines
Timothy Spann
 
Palo Alto Cortex XDR presentation .......
Palo Alto Cortex XDR presentation .......Palo Alto Cortex XDR presentation .......
Palo Alto Cortex XDR presentation .......
Sachin Paul
 
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
nuttdpt
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
74nqk8xf
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
Timothy Spann
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
rwarrenll
 
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
nyfuhyz
 

Recently uploaded (20)

End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024End-to-end pipeline agility - Berlin Buzzwords 2024
End-to-end pipeline agility - Berlin Buzzwords 2024
 
一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理一比一原版(UO毕业证)渥太华大学毕业证如何办理
一比一原版(UO毕业证)渥太华大学毕业证如何办理
 
Everything you wanted to know about LIHTC
Everything you wanted to know about LIHTCEverything you wanted to know about LIHTC
Everything you wanted to know about LIHTC
 
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
06-12-2024-BudapestDataForum-BuildingReal-timePipelineswithFLaNK AIM
 
The Ipsos - AI - Monitor 2024 Report.pdf
The  Ipsos - AI - Monitor 2024 Report.pdfThe  Ipsos - AI - Monitor 2024 Report.pdf
The Ipsos - AI - Monitor 2024 Report.pdf
 
Challenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more importantChallenges of Nation Building-1.pptx with more important
Challenges of Nation Building-1.pptx with more important
 
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
STATATHON: Unleashing the Power of Statistics in a 48-Hour Knowledge Extravag...
 
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
办(uts毕业证书)悉尼科技大学毕业证学历证书原版一模一样
 
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdfUdemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
Udemy_2024_Global_Learning_Skills_Trends_Report (1).pdf
 
Influence of Marketing Strategy and Market Competition on Business Plan
Influence of Marketing Strategy and Market Competition on Business PlanInfluence of Marketing Strategy and Market Competition on Business Plan
Influence of Marketing Strategy and Market Competition on Business Plan
 
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
一比一原版(Chester毕业证书)切斯特大学毕业证如何办理
 
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
一比一原版(爱大毕业证书)爱丁堡大学毕业证如何办理
 
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
在线办理(英国UCA毕业证书)创意艺术大学毕业证在读证明一模一样
 
DSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelinesDSSML24_tspann_CodelessGenerativeAIPipelines
DSSML24_tspann_CodelessGenerativeAIPipelines
 
Palo Alto Cortex XDR presentation .......
Palo Alto Cortex XDR presentation .......Palo Alto Cortex XDR presentation .......
Palo Alto Cortex XDR presentation .......
 
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
一比一原版(UCSB文凭证书)圣芭芭拉分校毕业证如何办理
 
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
一比一原版(Coventry毕业证书)考文垂大学毕业证如何办理
 
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
06-04-2024 - NYC Tech Week - Discussion on Vector Databases, Unstructured Dat...
 
My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.My burning issue is homelessness K.C.M.O.
My burning issue is homelessness K.C.M.O.
 
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
一比一原版(UMN文凭证书)明尼苏达大学毕业证如何办理
 

A Gentle Introduction to Bayesian Nonparametrics

  • 1. A gentle introduction to BNP Part I Antonio Canale Universit`a di Torino & Collegio Carlo Alberto StaTalk on BNP, 19/02/16
  • 2. Introduction The Dirichlet process Nonparametric mixture models Outline of the talk(s) 1 Why BNP? (A) 2 The Dirichlet process (A) 3 Nonparametric mixture models (A) 4 Beyond the DP (J) 5 Species sampling processes (J) 6 Completely random measures (J)
  • 3. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 4. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 5. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 6. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 7. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 8. Introduction The Dirichlet process Nonparametric mixture models Why Bayesian nonparametrics (BNP)? Why nonparametric? • We don’t want to strictly impose any model but let the data speak; • The idea of a true model governed by relatively few parameters is unrealistic; Why Bayesian? • If we have a reasonable guess for what is the true model we want to use this prior knowledge. • Large support and consistency are interesting concepts related to priors on infinite dimensional spaces (Pierpaolo’s talk in the afternoon) BNP is to fit a single model that can adapt its complexity to the data.
  • 9. Introduction The Dirichlet process Nonparametric mixture models How Bayesian and nonparametric? Define F the space of densities and let P ∈ F. A Bayesian analysis starts with y ∼ P P ∼ π where π is a measure on the space F. Hence BNP is infinitely parametric.
  • 10. Introduction The Dirichlet process Nonparametric mixture models The Dirichlet distribution • Start with independent Zj ∼ Ga(αj , 1), for j = 1, . . . , k (αj > 0) • Define πj = Zj k j=1 Zj ; • Then (π1, . . . , πk) ∼ Dir(α1, . . . , αk); • The Dirichlet distribution is a distribution over the K-dimensional probability simplex: ∆k = {(π1, . . . , πk) : πj > 0, j πj = 1}
  • 11. Introduction The Dirichlet process Nonparametric mixture models The Dirichlet distribution • Probability density p(π1, . . . , πk|α) = Γ( j αj ) j Γ(αj ) j π αj −1 j
  • 12. Introduction The Dirichlet process Nonparametric mixture models The Dirichlet distribution in Bayesian statistics Dirichlet distribution is conjugate to the multinomial likelihood, hence if π ∼ Dir(α) y|π ∼ Multinomial(π) p(y = j|π) = πj , then we have p(π|y = j, α) = Dir(ˆα) where ˆαj = αj + 1, ˆαi = αi for each i = j.
  • 13. Introduction The Dirichlet process Nonparametric mixture models Agglomerative property of Dirichlet distributions • Combining entries by their sum (π1, . . . , πk) ∼ Dir(α1, . . . , αk) → (π1, . . . , πi + πj , . . . , πk) ∼ Dir(α1, . . . , αi + αj , . . . , αk) • Marginals follow Beta distributions, πj ∼ beta(αj , h=j αh).
  • 14. Introduction The Dirichlet process Nonparametric mixture models 1 Introduction 2 The Dirichlet process 3 Nonparametric mixture models
  • 15. Introduction The Dirichlet process Nonparametric mixture models Ferguson (1973) definition of the Dirichlet process Definition • P is a random probability measure over (Y, B(Y)). • F is the whole space of probability measures on (Y, B(Y)), so P ∈ F. • Let α ∈ R+ and P0 ∈ F. • P ∼ DP(α, P0) iff for any n and any partition B1, . . . , Bn of Y (P(B1), P(B2), . . . , P(Bn)) ∼ Dir(αP0(B1), αP0(B2), . . . , αP0(Bn)) The DP is a distribution of random probability distributions.
  • 16. Introduction The Dirichlet process Nonparametric mixture models Interpretation If P ∼ DP(α, P0), then for any measurable A • E(P(A)) = P0(A) • Var(P(A)) = P0(A){1 − P0(A)}/(1 + α)
  • 17. Introduction The Dirichlet process Nonparametric mixture models Density estimation using DP priors If yi iid ∼ P for i = 1, . . . , n and P ∼ DP(α, P0) a priori then, P|y ∼ DP n + α, 1 α + n n i=1 δyi + α α + n P0
  • 18. Introduction The Dirichlet process Nonparametric mixture models Density estimation using DP priors −6 −4 −2 0 2 4 6 0.00.20.40.60.81.0 x −6 −4 −2 0 2 4 6 0.00.20.40.60.81.0 x Figure: Black true density (N(1, 2)), blue base measure (N(0,1)), green dashed ECDF, blue dashed posterior DP. First plot n = 10, second n = 50.
  • 19. Introduction The Dirichlet process Nonparametric mixture models Stick-breaking An alternative representation of the DP is related to the so called stick-breaking process:
  • 20. Introduction The Dirichlet process Nonparametric mixture models Stick-breaking representation of the DP To obtain P ∼ DP(αP0): • Draw a sequence of Beta random variables Vj iid ∼ Beta(1, α). • Define a sequence of weights as πj = Vj l<j (1 − Vl ) • Draw independent θ iid ∼ P0 • Define P = ∞ j=1 πj δθj
  • 21. Introduction The Dirichlet process Nonparametric mixture models Stochastic processes and chinese restaurants. . . Imagine a Chinese restaurant with countably infinitely many tables, labelled 1, 2, . . . Customers walk in and sit down at some table. The tables are chosen according to the following random process. 1 The first customer sits at table 1; 2 The n-th customer chooses the first unoccupied table with probability α/(α + n − 1) and an occupied table with probability nj /(α + n − 1), where nj is the number of people sitting at that table.
  • 22. Introduction The Dirichlet process Nonparametric mixture models CRP or Polya urn construction of the DP If θ iid ∼ P0 and P ∼ DP(αP0), integrate out P and obtain pr(θi |θ1, . . . , θi−1) = j nj n + α δθj + α n + α P0. Obtaining that (θ1, . . . , θn) ∼ PU(αP0).
  • 23. Introduction The Dirichlet process Nonparametric mixture models Considerations • Draw from a DP are a.s. discrete • Unappealing if y is continuous, useful if y is discrete? (no, but wait for my afternoon talk)
  • 24. Introduction The Dirichlet process Nonparametric mixture models Considerations • Draw from a DP are a.s. discrete • Unappealing if y is continuous, useful if y is discrete? (no, but wait for my afternoon talk)
  • 25. Introduction The Dirichlet process Nonparametric mixture models Finite mixture models Assume the following model yi ∼ N(µSi , σ2 Si ), pr(Si = h) = πh with likelihood f (y|µ, σ2 , π) = k j=1 πj φ(y; µj , σ2 j ) and prior (µ, σ2 ) ∼ P0, π ∼ Dir(α);
  • 26. Introduction The Dirichlet process Nonparametric mixture models FMM applications: density estimation • With enough components, a mixture of Gaussian can approximate any continuous distribution. • If the number of components equals n we have the kernel density estimation. 0 1 2 3 4 5 6 0.00.10.20.30.40.5 geyser$duration Probabilitydensityfunction
  • 27. Introduction The Dirichlet process Nonparametric mixture models FMM applications: model-based clustering • Divide observations into homogeneus clusters • “Homogeneus” depends on what kernel (Gaussian in previous slide) • With Gaussian kernel, there are two clusters in Iris dataset (truth is three!) • See discussions in Petralia et al. (2012), Canale and Scarpa (2015) and Canale and De Blasi (2015) q q q q q q q q q q q q qq q q q q qq q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q qq q q qq q q q q q q q qq q q q q q q q q q q q q q q q qq q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q q qq q q q q q qq q q q q q q q q q 4.5 5.0 5.5 6.0 6.5 7.0 7.5 8.0 2.02.53.03.54.0 iris$Sepal.Length iris$Sepal.Width
  • 28. Introduction The Dirichlet process Nonparametric mixture models Infinite mixture models • A more elegant way to write the finite mixture model is f (y) = K(y; θ)dP(θ), P = K j=1 ωj δθj , where K(·; θ) is a general kernel (e.g. normal) parametrized by θ. • Clearly a prior on the weights and on the parameters of the kernel is equivalent to a prior on the finite disrete measure P. • From FMM to IMM ⇒ P ∼ DP(αP0)!
  • 29. Introduction The Dirichlet process Nonparametric mixture models DP mixture models • The model and prior are y ∼ f , f (y) = K(y; θ)dP(θ), P ∼ DP(αP0). where K(·; θ) is a general kernel (e.g. normal) parametrized by θ. • Consider the DPM prior as a “smoothed version” of the DP prior (just like the kernel density estimation is a smoothed version of the histogram) • Widely used for continuous distribution.
  • 30. Introduction The Dirichlet process Nonparametric mixture models Hyerarchical representation Using a hyerarchical representation the mixture model can be expressed as yi | θi ∼ K(y; θi ) θi ∼ P P ∼ DP(αP0).
  • 31. Introduction The Dirichlet process Nonparametric mixture models Mixture of Gaussians • Gold standard for density estimation; • can approximate any continuous distribution (Lo, 1984; Escobar and West, 1995); • large support and good frequentist properties (Ghosal et al., 1999). The model and the prior are f (y) = N(y; µ, τ−1 )dP(µ, τ−1 ), P ∼ DP(αP0), where N(y; µ, τ−1) is a normal kernel having mean µ and precision τ, P0 Normal-Gamma, for conjugacy.
  • 32. Introduction The Dirichlet process Nonparametric mixture models Mixture of Gaussians yi | µi , τi ∼ N(µi , τ−1 i ) (µi , τi ) ∼ P P ∼ DP(αP0).
  • 33. Introduction The Dirichlet process Nonparametric mixture models Complex data • Mixture models can be used also when we have complex (modern) data • An example is functional data f1, . . . , fn fi (t) = η(t) + it, where η is a smooth function in t and it are random noises. • we can model these data with fi | ηi ∼ N(ηi , σ2 ) ηi ∼ P P ∼ DP(αP0).