SlideShare a Scribd company logo
1 of 3
Presentation
• Maximum likelihood estimation (MLE) is a popular method for parameter estimation in both applied
probability and statistics but MLE cannot solve the problem of incomplete data or hidden data because it is
impossible to maximize likelihood function from hidden data. Expectation maximum (EM) algorithm is a
powerful mathematical tool for solving this problem if there is a relationship between hidden data and
observed data. Such hinting relationship is specified by a mapping from hidden data to observed data or by a
joint probability between hidden data and observed data (showing MLE, EM, and practical EM, hidden info
implies the hinting relationship).
• The essential ideology of EM is to maximize the expectation of likelihood function over observed data based
on the hinting relationship instead of maximizing directly the likelihood function of hidden data (showing the
full EM with proof along with two steps).
• An important application of EM is (finite) mixture model which in turn is developed towards two trends such
as infinite mixture model and semiparametric mixture model. Especially, in semiparametric mixture model,
component probabilistic density functions are not parameterized. Semiparametric mixture model is interesting
and potential for other applications where probabilistic components are not easy to be specified (showing
mixture models).
• I raise a question that whether it is possible to backward discover semiparametric EM from semiparametric
mixture model. I hope that this question will open a new trend or new extension for EM algorithm (showing
the question).
Maximum likelihood estimation (MLE), MAP
Θ = argmax
Θ
𝑙 Θ = argmax
Θ
𝑖=1
𝑁
log 𝑓 𝑋𝑖 Θ
Expectation Maximization (EM)
𝑄 Θ′ Θ =
𝜑−1 𝑌
𝑘 𝑋 𝑌, Θ log 𝑓 𝑋 Θ′ d𝑋
Practical EM
𝑄 Θ′ Θ =
𝑋
𝑓 𝑋 𝑌, Θ log 𝑓 𝑋, 𝑌 Θ′ d𝑋
Full EM
𝑄 Θ′ Θ =
𝑖=1
𝑁
𝜑−1 𝑌𝑖
𝑘 𝑋𝑖 𝑌𝑖, Θ log 𝑓 𝑋𝑖 Θ′ d𝑋𝑖
𝑄 Θ′
Θ =
𝑖=1
𝑁
𝑋
𝑓 𝑋 𝑌𝑖, Θ log 𝑓 𝑋, 𝑌𝑖 Θ′
d𝑋
Observed
data
Observed
data
Observed
data
Mixture model
Semiparametric EM ?
?
E-step: Determine Q(Θ | Θ(t))
M-step: Θ 𝑡+1
= argmax
Θ
𝑄 Θ Θ 𝑡
Require mapping
Y = φ(X)
Require PDF
f(X, Y | ϴ)
?
Tutorial on EM algorithm – Loc Nguyen (ng_phloc@yahoo.com) – http://www.locnguyen.net
Semiparametric
mixture model
Infinite
mixture model
Hidden
info
Hidden
info
Thank you for attention
3
Conditional mixture model for modeling attributed dyadic data
16/09/2021

More Related Content

More from Loc Nguyen

Adversarial Variational Autoencoders to extend and improve generative model
Adversarial Variational Autoencoders to extend and improve generative modelAdversarial Variational Autoencoders to extend and improve generative model
Adversarial Variational Autoencoders to extend and improve generative model
Loc Nguyen
 
Tutorial on Bayesian optimization
Tutorial on Bayesian optimizationTutorial on Bayesian optimization
Tutorial on Bayesian optimization
Loc Nguyen
 
Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...
Loc Nguyen
 

More from Loc Nguyen (20)

Where the dragon to fly
Where the dragon to flyWhere the dragon to fly
Where the dragon to fly
 
Adversarial Variational Autoencoders to extend and improve generative model
Adversarial Variational Autoencoders to extend and improve generative modelAdversarial Variational Autoencoders to extend and improve generative model
Adversarial Variational Autoencoders to extend and improve generative model
 
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
 
Tutorial on Bayesian optimization
Tutorial on Bayesian optimizationTutorial on Bayesian optimization
Tutorial on Bayesian optimization
 
Tutorial on Support Vector Machine
Tutorial on Support Vector MachineTutorial on Support Vector Machine
Tutorial on Support Vector Machine
 
A Proposal of Two-step Autoregressive Model
A Proposal of Two-step Autoregressive ModelA Proposal of Two-step Autoregressive Model
A Proposal of Two-step Autoregressive Model
 
Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...
 
Jagged stock investment strategy
Jagged stock investment strategyJagged stock investment strategy
Jagged stock investment strategy
 
A short study on minima distribution
A short study on minima distributionA short study on minima distribution
A short study on minima distribution
 
Tutorial on EM algorithm – Part 4
Tutorial on EM algorithm – Part 4Tutorial on EM algorithm – Part 4
Tutorial on EM algorithm – Part 4
 
Tutorial on EM algorithm – Part 3
Tutorial on EM algorithm – Part 3Tutorial on EM algorithm – Part 3
Tutorial on EM algorithm – Part 3
 
Tutorial on particle swarm optimization
Tutorial on particle swarm optimizationTutorial on particle swarm optimization
Tutorial on particle swarm optimization
 
Tutorial on EM algorithm – Part 2
Tutorial on EM algorithm – Part 2Tutorial on EM algorithm – Part 2
Tutorial on EM algorithm – Part 2
 
Tutorial on EM algorithm – Part 1
Tutorial on EM algorithm – Part 1Tutorial on EM algorithm – Part 1
Tutorial on EM algorithm – Part 1
 
Some thoughts about economics post pandemic in Southeast Asia
Some thoughts about economics post pandemic in Southeast AsiaSome thoughts about economics post pandemic in Southeast Asia
Some thoughts about economics post pandemic in Southeast Asia
 
Finite mixture model with EM algorithm
Finite mixture model with EM algorithmFinite mixture model with EM algorithm
Finite mixture model with EM algorithm
 
Some thoughts about science & technology
Some thoughts about science & technologySome thoughts about science & technology
Some thoughts about science & technology
 
Handling missing data with expectation maximization algorithm
Handling missing data with expectation maximization algorithmHandling missing data with expectation maximization algorithm
Handling missing data with expectation maximization algorithm
 
Expectation Maximization Algorithm with Combinatorial Assumption
Expectation Maximization Algorithm with Combinatorial AssumptionExpectation Maximization Algorithm with Combinatorial Assumption
Expectation Maximization Algorithm with Combinatorial Assumption
 
Triangular Learner Model
Triangular Learner ModelTriangular Learner Model
Triangular Learner Model
 

Tutorial on EM algorithm - Poster

  • 1. Presentation • Maximum likelihood estimation (MLE) is a popular method for parameter estimation in both applied probability and statistics but MLE cannot solve the problem of incomplete data or hidden data because it is impossible to maximize likelihood function from hidden data. Expectation maximum (EM) algorithm is a powerful mathematical tool for solving this problem if there is a relationship between hidden data and observed data. Such hinting relationship is specified by a mapping from hidden data to observed data or by a joint probability between hidden data and observed data (showing MLE, EM, and practical EM, hidden info implies the hinting relationship). • The essential ideology of EM is to maximize the expectation of likelihood function over observed data based on the hinting relationship instead of maximizing directly the likelihood function of hidden data (showing the full EM with proof along with two steps). • An important application of EM is (finite) mixture model which in turn is developed towards two trends such as infinite mixture model and semiparametric mixture model. Especially, in semiparametric mixture model, component probabilistic density functions are not parameterized. Semiparametric mixture model is interesting and potential for other applications where probabilistic components are not easy to be specified (showing mixture models). • I raise a question that whether it is possible to backward discover semiparametric EM from semiparametric mixture model. I hope that this question will open a new trend or new extension for EM algorithm (showing the question).
  • 2. Maximum likelihood estimation (MLE), MAP Θ = argmax Θ 𝑙 Θ = argmax Θ 𝑖=1 𝑁 log 𝑓 𝑋𝑖 Θ Expectation Maximization (EM) 𝑄 Θ′ Θ = 𝜑−1 𝑌 𝑘 𝑋 𝑌, Θ log 𝑓 𝑋 Θ′ d𝑋 Practical EM 𝑄 Θ′ Θ = 𝑋 𝑓 𝑋 𝑌, Θ log 𝑓 𝑋, 𝑌 Θ′ d𝑋 Full EM 𝑄 Θ′ Θ = 𝑖=1 𝑁 𝜑−1 𝑌𝑖 𝑘 𝑋𝑖 𝑌𝑖, Θ log 𝑓 𝑋𝑖 Θ′ d𝑋𝑖 𝑄 Θ′ Θ = 𝑖=1 𝑁 𝑋 𝑓 𝑋 𝑌𝑖, Θ log 𝑓 𝑋, 𝑌𝑖 Θ′ d𝑋 Observed data Observed data Observed data Mixture model Semiparametric EM ? ? E-step: Determine Q(Θ | Θ(t)) M-step: Θ 𝑡+1 = argmax Θ 𝑄 Θ Θ 𝑡 Require mapping Y = φ(X) Require PDF f(X, Y | ϴ) ? Tutorial on EM algorithm – Loc Nguyen (ng_phloc@yahoo.com) – http://www.locnguyen.net Semiparametric mixture model Infinite mixture model Hidden info Hidden info
  • 3. Thank you for attention 3 Conditional mixture model for modeling attributed dyadic data 16/09/2021

Editor's Notes

  1. Maximum likelihood estimation (MLE) is a popular method for parameter estimation in both applied probability and statistics but MLE cannot solve the problem of incomplete data or hidden data because it is impossible to maximize likelihood function from hidden data. Expectation maximization (EM) algorithm is a powerful mathematical tool for solving this problem if there is a relationship between hidden data and observed data. Such hinting relationship is specified by a mapping from hidden data to observed data or by a joint probability between hidden data and observed data (showing MLE, EM, and practical EM, hidden info implies the hinting relationship). The essential ideology of EM is to maximize the expectation of likelihood function over observed data based on the hinting relationship instead of maximizing directly the likelihood function of hidden data (showing the full EM with proof along with two steps). An important application of EM is (finite) mixture model which in turn is developed towards two trends such as infinite mixture model and semiparametric mixture model. Especially, in semiparametric mixture model, component probabilistic density functions are not parameterized. Semiparametric mixture model is interesting and potential for other applications where probabilistic components are not easy to be specified (showing mixture models). I raise a question that whether it is possible to backward discover semiparametric EM from semiparametric mixture model. I hope that this question will open a new trend or new extension for EM algorithm (showing the question).