SlideShare a Scribd company logo
1 of 22
Agenda
• Why there is need for EM algorithm?
• What is EM Algorithm?
• Application of EM algorithm
• Advantages and Diadvantages
• Conclusion
Lets start this with very basic question:
 I tell u three data points [1,2,x] are the draws from
normal distribution N(1,1).
Whats the best guess for x ???
Obiviously best guess for x will be the mean ie 1.
 if tell u three data points [0,1,2] are the draws
from normal distribution N(µ,1).
Whats the best guess for µ ???
Again best guess for µ will be the average ie 1.
 Why there is need for EM algo
 Now if i give u three data points [1,2,x] are the
draws from normal distribution N(µ,1).
Whats the best guess for x and µ ???
In this case we dont know the parameter that generate the
data and also don’t know the data point.
Here We don’t have two pieces of information
simultaneously.
I need one to find another . This gave us the spirit of EM Algorithm
Expectation–maximization (EM) algorithm is an iterative method
to find local maximum likelihood estimates (MLE) for latent
variables in statistical models.It is also referred to as the
latent variable model.
Latent variables
 Variables that you can’t measure ( inferred from observed)
 Unobserved variable / Hidden variables
example:- Regression model
 What is EM Algorithm
Likelihood
• Model which is most likely to have produced our data.
• Measure the goodness of the distribution for a given sample data
Lets understand likelihood with an simple example
70 kgs 85 kgs
55 kgs
µ = 70 kgs
σ = 2.5
0.12
Mathematically we say... L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12
73 kgs
µ = 73 kgs
σ = 2.5
0.21
70
Now Mathematically likelihood will be..
Distribution 2 :-
Likelihood for this data to
belong to distribution 2 is
HIGH
L(µ = 73 kgs and σ = 2.5 | person weight is 73 kgs) = 0.21
Earlier... L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12
SAME
CHANGE
Now the question comes that wheater Probablity and likelihood
are same thing or not ????
In general, The answer is YES but In Statistics Likelihood and probability both are
VERY DIFFERENT
HOW ??
PROBABLITY measure chances of event occuring and graphically the area under the fixed
distribution
Mathematically ... P(data | distribution)
LIKELIHOOD measure the goodness of the distribution of given sample data and graphically its
the Y-axis value for the fixed data point and distributions that can be moved
Mathematically ... L(distribution | data)
L(µ = 73 kgs and σ = 2.5 | person weight is 73 kgs) = 0.21
L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12
How to choose the perfect/good parameter for the given data ???
For that we need MAXIMUM LIKELIHOOD ESTIMATION (MLE)
MAXIMUM LIKELIHOOD ESTIMATES(MLE)
• Method of determining the parameters (mean, standard deviation, etc) of normally
distributed sample data.
• Method of finding the best fitting Distribution over the random sample data.And
it is done my maximizing the likelihood function
sample data
LIKELIHOOD FUNCTION FOR NORMAL DISTRIBUTION
In Summary the mean of the data is the maximum
likelihood estimate for where the center of the
normal distribution should go.
EM ALGORITHM:-
• Expectation–maximization (EM) algorithm is an
iterative method to find local maximum likelihood
estimates (MLE) for latent variables in statistical
models.
• It is used to predict values of parameters in
instances where data is missing or unobservable for
learning, and this is done until convergence of the
values occurs.
What is the convergence in EM algorithm ???
• If there are two random variables that have very less difference in their
probability, then they are known as converged.
• Simply whenever the values of two variables are matched with each other, it
is called convergence.
START
E STEP
M STEP
INITIAL
VALUES
CONVERGED?? STOP
YES
NO
WORKING
Lets talk about E step and M step.
Expectation step (E - step):
• This involves the estimation/guessing of all missing values in
the dataset.
• After E step there should not be any missing value remaining
in the dataset.
Maximization step (M - step):
• This step involves the use of estimated data in the E-step
and updating the parameters.
Repeat E-step and M-step until the convergence of the values occurs.
Now lets ge t the question we have discussed earlier
Given three data points [1,2,x] are the draws from normal
distribution N(µ,1).
Whats the best guess for x and µ ???
Using the EM algorithm
Guess x = 0 µ = 1
x = 1
x = 4/3 µ = 13/9
µ = 4/3
.
.
.
.
.
.
Converge when
µ =
𝟏 + 𝟐 + 𝒙
𝟑
= 𝑥 µ∗ = 𝑥∗ = 1.5
Converged
values
current estimate
µ
x
Application of EM algorithm
The primary aim of the EM algorithm is to estimate the
missing data in the latent variables through observed data
in datasets.
• The EM algorithm is applicable in data clustering in machine learning.
• It is used to estimate the value of the parameter in mixed models such as
the Gaussian Mixture Model.
• It can be used for discovering the values of latent variables
• It can be used as the basis of unsupervised learning of clusters
• It is often used in computer vision and NLP (Natural language processing
PROS
• It is very easy to implement the first two basic steps of the EM algorithm in
various machine learning problems, which are E-step and M- step.
• It is mostly guaranteed that likelihood will enhance after each iteration.
• It often generates a solution for the M-step in the closed form.
CONS
• It has slow convergence.
• It makes convergence to the local optima only
CONCLUSION
• In real-world applications of machine learning, the expectation-maximization (EM)
algorithm plays a significant role in determining the local maximum likelihood
estimates (MLE) for unobservable variables in statistical models.
• It is often used for the latent variables, i.e., to estimate the latent variables through
observed data in datasets. It is generally completed in two important steps, i.e., the
expectation step (E-step) and the Maximization step (M-Step), where E-step is used
to estimate the missing data in datasets, and M-step is used to update the
parameters after the complete data is generated in E-step.
• Further, the importance of the EM algorithm can be seen in various applications
such as data clustering, natural language processing (NLP), computer vision,
image reconstruction, structural engineering, etc.
THANK YOU

More Related Content

What's hot

Hypermedia messageing (UNIT 5)
Hypermedia messageing (UNIT 5)Hypermedia messageing (UNIT 5)
Hypermedia messageing (UNIT 5)nirmalbj
 
Python lab manual all the experiments are available
Python lab manual all the experiments are availablePython lab manual all the experiments are available
Python lab manual all the experiments are availableNitesh Dubey
 
Parallel algorithms
Parallel algorithmsParallel algorithms
Parallel algorithmsguest084d20
 
Two dimensional arrays
Two dimensional arraysTwo dimensional arrays
Two dimensional arraysNeeru Mittal
 
Email Security Presentation
Email Security PresentationEmail Security Presentation
Email Security PresentationYosef Gamble
 
language , grammar and automata
language , grammar and automatalanguage , grammar and automata
language , grammar and automataElakkiyaS11
 
COMPUTER GRAPHICS LAB MANUAL
COMPUTER GRAPHICS LAB MANUALCOMPUTER GRAPHICS LAB MANUAL
COMPUTER GRAPHICS LAB MANUALVivek Kumar Sinha
 
Computer graphics basic transformation
Computer graphics basic transformationComputer graphics basic transformation
Computer graphics basic transformationSelvakumar Gna
 
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) Protocol
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) ProtocolE-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) Protocol
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) ProtocolVishal Kumar
 
Computer Security Lecture 7: RSA
Computer Security Lecture 7: RSAComputer Security Lecture 7: RSA
Computer Security Lecture 7: RSAMohamed Loey
 
C programs
C programsC programs
C programsMinu S
 
Lecture 5 sending and receiving e-mail
Lecture 5   sending and receiving e-mailLecture 5   sending and receiving e-mail
Lecture 5 sending and receiving e-mailMaxwell Musonda
 

What's hot (20)

Hypermedia messageing (UNIT 5)
Hypermedia messageing (UNIT 5)Hypermedia messageing (UNIT 5)
Hypermedia messageing (UNIT 5)
 
RSA ALGORITHM
RSA ALGORITHMRSA ALGORITHM
RSA ALGORITHM
 
IP Security
IP SecurityIP Security
IP Security
 
Python lab manual all the experiments are available
Python lab manual all the experiments are availablePython lab manual all the experiments are available
Python lab manual all the experiments are available
 
Parallel algorithms
Parallel algorithmsParallel algorithms
Parallel algorithms
 
Two dimensional arrays
Two dimensional arraysTwo dimensional arrays
Two dimensional arrays
 
Recognition-of-tokens
Recognition-of-tokensRecognition-of-tokens
Recognition-of-tokens
 
Email Security Presentation
Email Security PresentationEmail Security Presentation
Email Security Presentation
 
language , grammar and automata
language , grammar and automatalanguage , grammar and automata
language , grammar and automata
 
COMPUTER GRAPHICS LAB MANUAL
COMPUTER GRAPHICS LAB MANUALCOMPUTER GRAPHICS LAB MANUAL
COMPUTER GRAPHICS LAB MANUAL
 
Lexical analysis - Compiler Design
Lexical analysis - Compiler DesignLexical analysis - Compiler Design
Lexical analysis - Compiler Design
 
Short message service
Short message serviceShort message service
Short message service
 
Computer graphics basic transformation
Computer graphics basic transformationComputer graphics basic transformation
Computer graphics basic transformation
 
Daa unit 1
Daa unit 1Daa unit 1
Daa unit 1
 
K map
K mapK map
K map
 
Spam Filtering
Spam FilteringSpam Filtering
Spam Filtering
 
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) Protocol
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) ProtocolE-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) Protocol
E-Mail Security Protocol - 1 Privacy Enhanced Mail (PEM) Protocol
 
Computer Security Lecture 7: RSA
Computer Security Lecture 7: RSAComputer Security Lecture 7: RSA
Computer Security Lecture 7: RSA
 
C programs
C programsC programs
C programs
 
Lecture 5 sending and receiving e-mail
Lecture 5   sending and receiving e-mailLecture 5   sending and receiving e-mail
Lecture 5 sending and receiving e-mail
 

Similar to Why EM Algorithm Maximizes Likelihood for Latent Variables

machine learning for engineering students
machine learning for engineering studentsmachine learning for engineering students
machine learning for engineering studentsKavitabani1
 
Sample_Subjective_Questions_Answers (1).pdf
Sample_Subjective_Questions_Answers (1).pdfSample_Subjective_Questions_Answers (1).pdf
Sample_Subjective_Questions_Answers (1).pdfAaryanArora10
 
Data Science Interview Questions | Data Science Interview Questions And Answe...
Data Science Interview Questions | Data Science Interview Questions And Answe...Data Science Interview Questions | Data Science Interview Questions And Answe...
Data Science Interview Questions | Data Science Interview Questions And Answe...Simplilearn
 
Ders 1 mean mod media st dev.pptx
Ders 1 mean mod media st dev.pptxDers 1 mean mod media st dev.pptx
Ders 1 mean mod media st dev.pptxErgin Akalpler
 
Dimd_m_004 DL.pdf
Dimd_m_004 DL.pdfDimd_m_004 DL.pdf
Dimd_m_004 DL.pdfjuan631
 
Machine learning
Machine learningMachine learning
Machine learningVikas Sinha
 
Unit 3 – AIML.pptx
Unit 3 – AIML.pptxUnit 3 – AIML.pptx
Unit 3 – AIML.pptxhiblooms
 
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...Maninda Edirisooriya
 
Simple Ensemble Learning
Simple Ensemble LearningSimple Ensemble Learning
Simple Ensemble LearningMushfiq18
 
Application of Machine Learning in Agriculture
Application of Machine  Learning in AgricultureApplication of Machine  Learning in Agriculture
Application of Machine Learning in AgricultureAman Vasisht
 
Unit III_Ch 17_Probablistic Methods.pptx
Unit III_Ch 17_Probablistic Methods.pptxUnit III_Ch 17_Probablistic Methods.pptx
Unit III_Ch 17_Probablistic Methods.pptxsmithashetty24
 

Similar to Why EM Algorithm Maximizes Likelihood for Latent Variables (20)

machine learning for engineering students
machine learning for engineering studentsmachine learning for engineering students
machine learning for engineering students
 
03 Data Mining Techniques
03 Data Mining Techniques03 Data Mining Techniques
03 Data Mining Techniques
 
Em Algorithm | Statistics
Em Algorithm | StatisticsEm Algorithm | Statistics
Em Algorithm | Statistics
 
MACHINE LEARNING.pptx
MACHINE LEARNING.pptxMACHINE LEARNING.pptx
MACHINE LEARNING.pptx
 
MLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptxMLU_DTE_Lecture_2.pptx
MLU_DTE_Lecture_2.pptx
 
Sample_Subjective_Questions_Answers (1).pdf
Sample_Subjective_Questions_Answers (1).pdfSample_Subjective_Questions_Answers (1).pdf
Sample_Subjective_Questions_Answers (1).pdf
 
Data Science Interview Questions | Data Science Interview Questions And Answe...
Data Science Interview Questions | Data Science Interview Questions And Answe...Data Science Interview Questions | Data Science Interview Questions And Answe...
Data Science Interview Questions | Data Science Interview Questions And Answe...
 
Ders 1 mean mod media st dev.pptx
Ders 1 mean mod media st dev.pptxDers 1 mean mod media st dev.pptx
Ders 1 mean mod media st dev.pptx
 
Numerical Method
Numerical Method Numerical Method
Numerical Method
 
Dimd_m_004 DL.pdf
Dimd_m_004 DL.pdfDimd_m_004 DL.pdf
Dimd_m_004 DL.pdf
 
Machine learning
Machine learningMachine learning
Machine learning
 
Unit 3 – AIML.pptx
Unit 3 – AIML.pptxUnit 3 – AIML.pptx
Unit 3 – AIML.pptx
 
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...
Lecture 7 - Bias, Variance and Regularization, a lecture in subject module St...
 
Simple Ensemble Learning
Simple Ensemble LearningSimple Ensemble Learning
Simple Ensemble Learning
 
1015 track2 abbott
1015 track2 abbott1015 track2 abbott
1015 track2 abbott
 
1030 track2 abbott
1030 track2 abbott1030 track2 abbott
1030 track2 abbott
 
E-M Algorithm
E-M AlgorithmE-M Algorithm
E-M Algorithm
 
Application of Machine Learning in Agriculture
Application of Machine  Learning in AgricultureApplication of Machine  Learning in Agriculture
Application of Machine Learning in Agriculture
 
Unit III_Ch 17_Probablistic Methods.pptx
Unit III_Ch 17_Probablistic Methods.pptxUnit III_Ch 17_Probablistic Methods.pptx
Unit III_Ch 17_Probablistic Methods.pptx
 
Moviereview prjct
Moviereview prjctMoviereview prjct
Moviereview prjct
 

Recently uploaded

Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Celine George
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesFatimaKhan178732
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxSayali Powar
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991RKavithamani
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3JemimahLaneBuaron
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdfSoniaTolstoy
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfchloefrazer622
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 

Recently uploaded (20)

Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and Actinides
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptxPOINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
POINT- BIOCHEMISTRY SEM 2 ENZYMES UNIT 5.pptx
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
Industrial Policy - 1948, 1956, 1973, 1977, 1980, 1991
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Arihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdfArihant handbook biology for class 11 .pdf
Arihant handbook biology for class 11 .pdf
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 

Why EM Algorithm Maximizes Likelihood for Latent Variables

  • 1.
  • 2. Agenda • Why there is need for EM algorithm? • What is EM Algorithm? • Application of EM algorithm • Advantages and Diadvantages • Conclusion
  • 3. Lets start this with very basic question:  I tell u three data points [1,2,x] are the draws from normal distribution N(1,1). Whats the best guess for x ??? Obiviously best guess for x will be the mean ie 1.  if tell u three data points [0,1,2] are the draws from normal distribution N(µ,1). Whats the best guess for µ ??? Again best guess for µ will be the average ie 1.  Why there is need for EM algo
  • 4.  Now if i give u three data points [1,2,x] are the draws from normal distribution N(µ,1). Whats the best guess for x and µ ??? In this case we dont know the parameter that generate the data and also don’t know the data point. Here We don’t have two pieces of information simultaneously. I need one to find another . This gave us the spirit of EM Algorithm
  • 5. Expectation–maximization (EM) algorithm is an iterative method to find local maximum likelihood estimates (MLE) for latent variables in statistical models.It is also referred to as the latent variable model. Latent variables  Variables that you can’t measure ( inferred from observed)  Unobserved variable / Hidden variables example:- Regression model  What is EM Algorithm
  • 6. Likelihood • Model which is most likely to have produced our data. • Measure the goodness of the distribution for a given sample data Lets understand likelihood with an simple example 70 kgs 85 kgs 55 kgs µ = 70 kgs σ = 2.5 0.12 Mathematically we say... L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12 73 kgs
  • 7. µ = 73 kgs σ = 2.5 0.21 70 Now Mathematically likelihood will be.. Distribution 2 :- Likelihood for this data to belong to distribution 2 is HIGH L(µ = 73 kgs and σ = 2.5 | person weight is 73 kgs) = 0.21 Earlier... L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12 SAME CHANGE
  • 8. Now the question comes that wheater Probablity and likelihood are same thing or not ???? In general, The answer is YES but In Statistics Likelihood and probability both are VERY DIFFERENT HOW ??
  • 9. PROBABLITY measure chances of event occuring and graphically the area under the fixed distribution Mathematically ... P(data | distribution) LIKELIHOOD measure the goodness of the distribution of given sample data and graphically its the Y-axis value for the fixed data point and distributions that can be moved Mathematically ... L(distribution | data) L(µ = 73 kgs and σ = 2.5 | person weight is 73 kgs) = 0.21 L(µ = 70 kgs and σ = 2.5 | person weight is 73 kgs) = 0.12 How to choose the perfect/good parameter for the given data ??? For that we need MAXIMUM LIKELIHOOD ESTIMATION (MLE)
  • 10. MAXIMUM LIKELIHOOD ESTIMATES(MLE) • Method of determining the parameters (mean, standard deviation, etc) of normally distributed sample data. • Method of finding the best fitting Distribution over the random sample data.And it is done my maximizing the likelihood function sample data
  • 11. LIKELIHOOD FUNCTION FOR NORMAL DISTRIBUTION
  • 12. In Summary the mean of the data is the maximum likelihood estimate for where the center of the normal distribution should go.
  • 13. EM ALGORITHM:- • Expectation–maximization (EM) algorithm is an iterative method to find local maximum likelihood estimates (MLE) for latent variables in statistical models. • It is used to predict values of parameters in instances where data is missing or unobservable for learning, and this is done until convergence of the values occurs. What is the convergence in EM algorithm ??? • If there are two random variables that have very less difference in their probability, then they are known as converged. • Simply whenever the values of two variables are matched with each other, it is called convergence.
  • 15. Lets talk about E step and M step. Expectation step (E - step): • This involves the estimation/guessing of all missing values in the dataset. • After E step there should not be any missing value remaining in the dataset. Maximization step (M - step): • This step involves the use of estimated data in the E-step and updating the parameters. Repeat E-step and M-step until the convergence of the values occurs.
  • 16. Now lets ge t the question we have discussed earlier Given three data points [1,2,x] are the draws from normal distribution N(µ,1). Whats the best guess for x and µ ??? Using the EM algorithm Guess x = 0 µ = 1 x = 1 x = 4/3 µ = 13/9 µ = 4/3 . . . . . .
  • 17. Converge when µ = 𝟏 + 𝟐 + 𝒙 𝟑 = 𝑥 µ∗ = 𝑥∗ = 1.5 Converged values
  • 19. Application of EM algorithm The primary aim of the EM algorithm is to estimate the missing data in the latent variables through observed data in datasets. • The EM algorithm is applicable in data clustering in machine learning. • It is used to estimate the value of the parameter in mixed models such as the Gaussian Mixture Model. • It can be used for discovering the values of latent variables • It can be used as the basis of unsupervised learning of clusters • It is often used in computer vision and NLP (Natural language processing
  • 20. PROS • It is very easy to implement the first two basic steps of the EM algorithm in various machine learning problems, which are E-step and M- step. • It is mostly guaranteed that likelihood will enhance after each iteration. • It often generates a solution for the M-step in the closed form. CONS • It has slow convergence. • It makes convergence to the local optima only
  • 21. CONCLUSION • In real-world applications of machine learning, the expectation-maximization (EM) algorithm plays a significant role in determining the local maximum likelihood estimates (MLE) for unobservable variables in statistical models. • It is often used for the latent variables, i.e., to estimate the latent variables through observed data in datasets. It is generally completed in two important steps, i.e., the expectation step (E-step) and the Maximization step (M-Step), where E-step is used to estimate the missing data in datasets, and M-step is used to update the parameters after the complete data is generated in E-step. • Further, the importance of the EM algorithm can be seen in various applications such as data clustering, natural language processing (NLP), computer vision, image reconstruction, structural engineering, etc.