SlideShare a Scribd company logo
Parametric Density Estimation using Gaussian Mixture Models




                      Parametric Density Estimation using
                           Gaussian Mixture Models
                         An Application of the EM Algorithm


          Based on tutorials by Je A. Bilmes and by Ludwig Schwardt
                       Amirkabir University of Technology  Bahman 12, 1389
Parametric Density Estimation using Gaussian Mixture Models


     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models


     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models


     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models


     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models


     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     What does EM do?




             iterative     maximization of the likelihood function
                    when data has missing values or
                    when maximization of the likelihood is dicult
                    data augmentation

             X is incomplete data

             Z = (X; Y) is the complete data set
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     EM and MLE




             maximum likelihood estimation

                    estimates density parameters
                     =) EM app: parametric density estimation
                    when density is a mixture of Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     EM and MLE




             maximum likelihood estimation

                    estimates density parameters
                     =) EM app: parametric density estimation
                    when density is a mixture of Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     EM and MLE




             maximum likelihood estimation

                    estimates density parameters
                     =) EM app: parametric density estimation
                    when density is a mixture of Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     EM and MLE




             maximum likelihood estimation

                    estimates density parameters
                     =) EM app: parametric density estimation
                    when density is a mixture of Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Introduction

     EM and MLE




             maximum likelihood estimation

                    estimates density parameters
                     =) EM app: parametric density estimation
                    when density is a mixture of Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The density estimation problem



      Denition
      Our parametric           density estimation             problem:
             X=f      xi Ng
                         i=1 , xi P R       D

             f (xi jΘ), i.i.d. assumption
                              €
             f (xi jΘ) = K=1 k p(xi jk )
                              k
             K components (given)
             Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     Constraint on mixing probabilities                               i
     Sum to unity




                                        
                                                 p(xi jk )dxi = 1
                                            RD

                                          
                                   1=              f (xi jΘ)dxi
                                              RD
                                                  K
                                                   ˆ
                                      =                   k p(xi jk )dxi
                                              RD k=1
                                            K
                                            ˆ
                                      =            k :
                                          k=1
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     Constraint on mixing probabilities                               i
     Sum to unity




                                        
                                                 p(xi jk )dxi = 1
                                            RD

                                          
                                   1=              f (xi jΘ)dxi
                                              RD
                                                  K
                                                   ˆ
                                      =                   k p(xi jk )dxi
                                              RD k=1
                                            K
                                            ˆ
                                      =            k :
                                          k=1
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The maximum likelihood problem



      Denition
      Our MLE problem:
                             N
             p(XjΘ) =           i=1   p(xi jΘ) = v(ΘjX)
             Θ£ = arg maxΘ v(ΘjX)
             the log-likelihood: log(v(ΘjX))
                                      €N             €K                    ¡
             log(v(ΘjX)) =               i=1 log        k=1   k p(xi ; k )
             dicult to optimize b/c it contains log of sum
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation



      Denition
      Our EM problem:
             Z = (X; Y)
             v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ)
             yi P 1 : : : K
             yi = k if the ith sample was generated by the kth component
             Y is a random vector
                                          €N
             log(v(ΘjX; Y)) =                i=1 log( yi p(xi jyi ))
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Continued




      Denition
      The E- and M-steps with the auxiliary function Q:
          E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ]
          M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Continued




      Denition
      The E- and M-steps with the auxiliary function Q:
          E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ]
          M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Continued




      Denition
      The E- and M-steps with the auxiliary function Q:
          E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ]
          M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     The   Q function




                  Q(Θ; Θ(t) ) = E [log(p(X; YjΘ))jX; Θ(t) ]
                                        K N
                                        ˆˆ
                                    =             log(k p(xi jk ))p(kjxi ; Θ(t) )
                                        k=1 i=1
                                         K N
                                        ˆˆ
                                    =             log(k )p(kjxi ; Θ(t) )
                                        k=1 i=1
                                        K N
                                        ˆˆ
                                    +             log(p(xi jk ))p(kjxi ; Θ(t) )
                                        k=1 i=1
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Solving for    k

                                                                  €
      We use the Lagrange multiplier to enforce                       k   k = 1.
                 @ ˆˆ                               ˆ
                   [    log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0
                @k k i                             k

                                  ˆ 1
                              p(kjxi ; Θ(t) ) +  = 0
                         i
                           k
      Summing both sides over k, we get  =  N .
                                              1 ˆ
                                    k =               p(kjxi ; Θ(t) ):
                                             N     i
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Solving for    k

                                                                  €
      We use the Lagrange multiplier to enforce                       k   k = 1.
                 @ ˆˆ                               ˆ
                   [    log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0
                @k k i                             k

                                  ˆ 1
                              p(kjxi ; Θ(t) ) +  = 0
                         i
                           k
      Summing both sides over k, we get  =  N .
                                              1 ˆ
                                    k =               p(kjxi ; Θ(t) ):
                                             N     i
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Solving for    k

                                                                  €
      We use the Lagrange multiplier to enforce                       k   k = 1.
                 @ ˆˆ                               ˆ
                   [    log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0
                @k k i                             k

                                  ˆ 1
                              p(kjxi ; Θ(t) ) +  = 0
                         i
                           k
      Summing both sides over k, we get  =  N .
                                              1 ˆ
                                    k =               p(kjxi ; Θ(t) ):
                                             N     i
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Solving for    k

                                                                  €
      We use the Lagrange multiplier to enforce                       k   k = 1.
                 @ ˆˆ                               ˆ
                   [    log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0
                @k k i                             k

                                  ˆ 1
                              p(kjxi ; Θ(t) ) +  = 0
                         i
                           k
      Summing both sides over k, we get  =  N .
                                              1 ˆ
                                    k =               p(kjxi ; Θ(t) ):
                                             N     i
Parametric Density Estimation using Gaussian Mixture Models
  Density Estimation

     The EM formulation
     Solving for    k

                                                                  €
      We use the Lagrange multiplier to enforce                       k   k = 1.
                 @ ˆˆ                               ˆ
                   [    log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0
                @k k i                             k

                                  ˆ 1
                              p(kjxi ; Θ(t) ) +  = 0
                         i
                           k
      Summing both sides over k, we get  =  N .
                                              1 ˆ
                                    k =               p(kjxi ; Θ(t) ):
                                             N     i
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     GMM
     Solving for    mk   and   Σk




      Gaussian components
                              D
      p(xi jxk ; Σk ) = (2)  2 jΣk j  2 exp(  1 (xi   mk )T Σ 1 (xi   mk ))
                                       1
                                                              2        k



                                    €
                               i        xi p(kjxi ; Θ(t) )
                         mk = €
                                        i p(kjxi ; Θ )
                                                    (t)


                                    p(kjxi ; Θ(t) )(xi   mk )(xi   mk )T
                         Σk =
                                                p(kjxi ; Θ(t) )
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     GMM
     Solving for    mk   and   Σk




      Gaussian components
                              D
      p(xi jxk ; Σk ) = (2)  2 jΣk j  2 exp(  1 (xi   mk )T Σ 1 (xi   mk ))
                                       1
                                                              2        k



                                    €
                               i        xi p(kjxi ; Θ(t) )
                         mk = €
                                        i p(kjxi ; Θ )
                                                    (t)


                                    p(kjxi ; Θ(t) )(xi   mk )(xi   mk )T
                         Σk =
                                                p(kjxi ; Θ(t) )
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Choice of Σk
     Full covariance




      Fits data best, but costly in high-dimensional space.




                                        Figure: full covariance
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Choice of Σk
     Diagonal covariance




      A tradeo between cost and quality.




                                     Figure: diagonal covariance
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Choice of Σk
     Spherical covariance,      Σk   = k I
                                        2




      Needs many components to cover data.




                                     Figure: spherical covariance

      Decision should be based on the size of training data set
          so that all parameters may be tuned
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Choice of Σk
     Spherical covariance,      Σk   = k I
                                        2




      Needs many components to cover data.




                                     Figure: spherical covariance

      Decision should be based on the size of training data set
          so that all parameters may be tuned
Parametric Density Estimation using Gaussian Mixture Models
  Gaussian Mixture Model

     Choice of Σk
     Spherical covariance,      Σk   = k I
                                        2




      Needs many components to cover data.




                                     Figure: spherical covariance

      Decision should be based on the size of training data set
          so that all parameters may be tuned
Parametric Density Estimation using Gaussian Mixture Models
  Some Results

     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Some Results

     Data

      data generated from a mixture of three bivariate Gaussians
Parametric Density Estimation using Gaussian Mixture Models
  Some Results

     PDFs

      estimated pdf contours (after 43 iterations)
Parametric Density Estimation using Gaussian Mixture Models
  Some Results

     Clusters

      xi assigned to the cluster            k corresponding to the highest p(kjxi ; Θ)
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     Outline



      1   Introduction

      2   Density Estimation

      3   Gaussian Mixture Model

      4   Some Results

      5   Other Applications of EM
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     A problem with missing data
     Life-testing experiment




             lifetime of light bulbs follows an exponential distribution with
             mean 

             two separate experiments

             N bulbs tested until failed
                    failure times recorded as         x1 ; : : : ; x   N
             M bulbs tested
                    failure times not recorded
                    only number of bulbs , that failed at time
                                                 r                         t

                    missing data are failure times 1       M  u ;:::;u
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models
  Other Applications of EM

     EM formulation
     Life-testing experiment




                                                                     €M
             log(v(jx; u)) =  N (log  + x=)  
                                                                         i   (log  + ui =)

             expected value for bulb still burning: t + 
                                                        t     k
             one that burned out:    1tee = (k) =    th(k)
                                                              ( )

                                         t =
             E-step:
              Q(; (k) ) =  (N + M ) log 
                               1 (N x + (M   r)(t +  k ) + r( k   th k ))
                               
                                                                   ( )           ( )      ( )




             M-step:         1
                          N +M (   N x + (M   r)(t + (k) ) + r((k)   th(k) ))
Parametric Density Estimation using Gaussian Mixture Models




      Thank you for your attention. Any questions?

More Related Content

What's hot

"reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli..."reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli...
Christian Robert
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard University
Christian Robert
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modules
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Nested sampling
Nested samplingNested sampling
Nested sampling
Christian Robert
 
Probability distributions for ml
Probability distributions for mlProbability distributions for ml
Probability distributions for ml
Sung Yub Kim
 
prior selection for mixture estimation
prior selection for mixture estimationprior selection for mixture estimation
prior selection for mixture estimation
Christian Robert
 
accurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannaccurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannolli0601
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
(DL hacks輪読)Bayesian Neural Network
(DL hacks輪読)Bayesian Neural Network(DL hacks輪読)Bayesian Neural Network
(DL hacks輪読)Bayesian Neural Network
Masahiro Suzuki
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimation
Christian Robert
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
Christian Robert
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
Christian Robert
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
Christian Robert
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Entity Linking via Graph-Distance Minimization
Entity Linking via Graph-Distance MinimizationEntity Linking via Graph-Distance Minimization
Entity Linking via Graph-Distance Minimization
Roi Blanco
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
The Statistical and Applied Mathematical Sciences Institute
 
Coordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like samplerCoordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like sampler
Christian Robert
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
Jae-kwang Kim
 

What's hot (20)

"reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli..."reflections on the probability space induced by moment conditions with impli...
"reflections on the probability space induced by moment conditions with impli...
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard University
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modules
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Nested sampling
Nested samplingNested sampling
Nested sampling
 
Probability distributions for ml
Probability distributions for mlProbability distributions for ml
Probability distributions for ml
 
prior selection for mixture estimation
prior selection for mixture estimationprior selection for mixture estimation
prior selection for mixture estimation
 
accurate ABC Oliver Ratmann
accurate ABC Oliver Ratmannaccurate ABC Oliver Ratmann
accurate ABC Oliver Ratmann
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
(DL hacks輪読)Bayesian Neural Network
(DL hacks輪読)Bayesian Neural Network(DL hacks輪読)Bayesian Neural Network
(DL hacks輪読)Bayesian Neural Network
 
random forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimationrandom forests for ABC model choice and parameter estimation
random forests for ABC model choice and parameter estimation
 
An overview of Bayesian testing
An overview of Bayesian testingAn overview of Bayesian testing
An overview of Bayesian testing
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Entity Linking via Graph-Distance Minimization
Entity Linking via Graph-Distance MinimizationEntity Linking via Graph-Distance Minimization
Entity Linking via Graph-Distance Minimization
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Coordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like samplerCoordinate sampler: A non-reversible Gibbs-like sampler
Coordinate sampler: A non-reversible Gibbs-like sampler
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
 

Similar to Parametric Density Estimation using Gaussian Mixture Models

Principle of Maximum Entropy
Principle of Maximum EntropyPrinciple of Maximum Entropy
Principle of Maximum Entropy
Jiawang Liu
 
Machine learning (10)
Machine learning (10)Machine learning (10)
Machine learning (10)NYversity
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
Valentin De Bortoli
 
Machine learning (8)
Machine learning (8)Machine learning (8)
Machine learning (8)NYversity
 
Cs229 notes7b
Cs229 notes7bCs229 notes7b
Cs229 notes7b
VuTran231
 
Bayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear modelsBayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear models
Caleb (Shiqiang) Jin
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling Distribution
Dexlab Analytics
 
Machine Learning
Machine LearningMachine Learning
Machine Learning
Bharat Khatri
 
Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)NYversity
 
DirichletProcessNotes
DirichletProcessNotesDirichletProcessNotes
DirichletProcessNotesAngie Shen
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)NYversity
 
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
Mamoon Ismail Khalid
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
The Statistical and Applied Mathematical Sciences Institute
 
Finite mixture model with EM algorithm
Finite mixture model with EM algorithmFinite mixture model with EM algorithm
Finite mixture model with EM algorithm
Loc Nguyen
 
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...Mengxi Jiang
 
Clustering:k-means, expect-maximization and gaussian mixture model
Clustering:k-means, expect-maximization and gaussian mixture modelClustering:k-means, expect-maximization and gaussian mixture model
Clustering:k-means, expect-maximization and gaussian mixture model
jins0618
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
RayKim51
 
Kernel estimation(ref)
Kernel estimation(ref)Kernel estimation(ref)
Kernel estimation(ref)
Zahra Amini
 

Similar to Parametric Density Estimation using Gaussian Mixture Models (20)

Principle of Maximum Entropy
Principle of Maximum EntropyPrinciple of Maximum Entropy
Principle of Maximum Entropy
 
Machine learning (10)
Machine learning (10)Machine learning (10)
Machine learning (10)
 
2008 spie gmm
2008 spie gmm2008 spie gmm
2008 spie gmm
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
Machine learning (8)
Machine learning (8)Machine learning (8)
Machine learning (8)
 
Cs229 notes7b
Cs229 notes7bCs229 notes7b
Cs229 notes7b
 
Bayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear modelsBayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear models
 
Statistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling DistributionStatistical Inference Part II: Types of Sampling Distribution
Statistical Inference Part II: Types of Sampling Distribution
 
Machine Learning
Machine LearningMachine Learning
Machine Learning
 
Machine learning (9)
Machine learning (9)Machine learning (9)
Machine learning (9)
 
DirichletProcessNotes
DirichletProcessNotesDirichletProcessNotes
DirichletProcessNotes
 
Machine learning (1)
Machine learning (1)Machine learning (1)
Machine learning (1)
 
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...Joint3DShapeMatching  - a fast approach to 3D model matching using MatchALS 3...
Joint3DShapeMatching - a fast approach to 3D model matching using MatchALS 3...
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
 
Finite mixture model with EM algorithm
Finite mixture model with EM algorithmFinite mixture model with EM algorithm
Finite mixture model with EM algorithm
 
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...
Application of Graphic LASSO in Portfolio Optimization_Yixuan Chen & Mengxi J...
 
Clustering:k-means, expect-maximization and gaussian mixture model
Clustering:k-means, expect-maximization and gaussian mixture modelClustering:k-means, expect-maximization and gaussian mixture model
Clustering:k-means, expect-maximization and gaussian mixture model
 
Multitask learning for GGM
Multitask learning for GGMMultitask learning for GGM
Multitask learning for GGM
 
Bayesian Deep Learning
Bayesian Deep LearningBayesian Deep Learning
Bayesian Deep Learning
 
Kernel estimation(ref)
Kernel estimation(ref)Kernel estimation(ref)
Kernel estimation(ref)
 

Recently uploaded

Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Paige Cruz
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
DianaGray10
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
SOFTTECHHUB
 
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
James Anderson
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
Rohit Gautam
 
Removing Uninteresting Bytes in Software Fuzzing
Removing Uninteresting Bytes in Software FuzzingRemoving Uninteresting Bytes in Software Fuzzing
Removing Uninteresting Bytes in Software Fuzzing
Aftab Hussain
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
Kari Kakkonen
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
Guy Korland
 
20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
Matthew Sinclair
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Nexer Digital
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
Adtran
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Aggregage
 
RESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for studentsRESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for students
KAMESHS29
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
Safe Software
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...
ThomasParaiso2
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
Neo4j
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
Neo4j
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Albert Hoitingh
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
Alan Dix
 

Recently uploaded (20)

Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
 
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
Alt. GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using ...
 
Large Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial ApplicationsLarge Language Model (LLM) and it’s Geospatial Applications
Large Language Model (LLM) and it’s Geospatial Applications
 
Removing Uninteresting Bytes in Software Fuzzing
Removing Uninteresting Bytes in Software FuzzingRemoving Uninteresting Bytes in Software Fuzzing
Removing Uninteresting Bytes in Software Fuzzing
 
Climate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing DaysClimate Impact of Software Testing at Nordic Testing Days
Climate Impact of Software Testing at Nordic Testing Days
 
GraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge GraphGraphRAG is All You need? LLM & Knowledge Graph
GraphRAG is All You need? LLM & Knowledge Graph
 
20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
 
Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?Elizabeth Buie - Older adults: Are we really designing for our future selves?
Elizabeth Buie - Older adults: Are we really designing for our future selves?
 
Pushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 daysPushing the limits of ePRTC: 100ns holdover for 100 days
Pushing the limits of ePRTC: 100ns holdover for 100 days
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
 
RESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for studentsRESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for students
 
Essentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FMEEssentials of Automations: The Art of Triggers and Actions in FME
Essentials of Automations: The Art of Triggers and Actions in FME
 
PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...GridMate - End to end testing is a critical piece to ensure quality and avoid...
GridMate - End to end testing is a critical piece to ensure quality and avoid...
 
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
GraphSummit Singapore | The Future of Agility: Supercharging Digital Transfor...
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
 
Epistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI supportEpistemic Interaction - tuning interfaces to provide information for AI support
Epistemic Interaction - tuning interfaces to provide information for AI support
 

Parametric Density Estimation using Gaussian Mixture Models

  • 1. Parametric Density Estimation using Gaussian Mixture Models Parametric Density Estimation using Gaussian Mixture Models An Application of the EM Algorithm Based on tutorials by Je A. Bilmes and by Ludwig Schwardt Amirkabir University of Technology Bahman 12, 1389
  • 2. Parametric Density Estimation using Gaussian Mixture Models Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 3. Parametric Density Estimation using Gaussian Mixture Models Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 4. Parametric Density Estimation using Gaussian Mixture Models Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 5. Parametric Density Estimation using Gaussian Mixture Models Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 6. Parametric Density Estimation using Gaussian Mixture Models Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 7. Parametric Density Estimation using Gaussian Mixture Models Introduction Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 8. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 9. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 10. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 11. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 12. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 13. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 14. Parametric Density Estimation using Gaussian Mixture Models Introduction What does EM do? iterative maximization of the likelihood function when data has missing values or when maximization of the likelihood is dicult data augmentation X is incomplete data Z = (X; Y) is the complete data set
  • 15. Parametric Density Estimation using Gaussian Mixture Models Introduction EM and MLE maximum likelihood estimation estimates density parameters =) EM app: parametric density estimation when density is a mixture of Gaussians
  • 16. Parametric Density Estimation using Gaussian Mixture Models Introduction EM and MLE maximum likelihood estimation estimates density parameters =) EM app: parametric density estimation when density is a mixture of Gaussians
  • 17. Parametric Density Estimation using Gaussian Mixture Models Introduction EM and MLE maximum likelihood estimation estimates density parameters =) EM app: parametric density estimation when density is a mixture of Gaussians
  • 18. Parametric Density Estimation using Gaussian Mixture Models Introduction EM and MLE maximum likelihood estimation estimates density parameters =) EM app: parametric density estimation when density is a mixture of Gaussians
  • 19. Parametric Density Estimation using Gaussian Mixture Models Introduction EM and MLE maximum likelihood estimation estimates density parameters =) EM app: parametric density estimation when density is a mixture of Gaussians
  • 20. Parametric Density Estimation using Gaussian Mixture Models Density Estimation Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 21. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 22. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 23. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 24. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 25. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 26. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The density estimation problem Denition Our parametric density estimation problem: X=f xi Ng i=1 , xi P R D f (xi jΘ), i.i.d. assumption € f (xi jΘ) = K=1 k p(xi jk ) k K components (given) Θ = (1 ; : : : ; K ; 1 ; : : : ; K )
  • 27. Parametric Density Estimation using Gaussian Mixture Models Density Estimation Constraint on mixing probabilities i Sum to unity  p(xi jk )dxi = 1 RD  1= f (xi jΘ)dxi RD  K ˆ = k p(xi jk )dxi RD k=1 K ˆ = k : k=1
  • 28. Parametric Density Estimation using Gaussian Mixture Models Density Estimation Constraint on mixing probabilities i Sum to unity  p(xi jk )dxi = 1 RD  1= f (xi jΘ)dxi RD  K ˆ = k p(xi jk )dxi RD k=1 K ˆ = k : k=1
  • 29. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 30. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 31. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 32. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 33. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 34. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The maximum likelihood problem Denition Our MLE problem: N p(XjΘ) = i=1 p(xi jΘ) = v(ΘjX) Θ£ = arg maxΘ v(ΘjX) the log-likelihood: log(v(ΘjX)) €N   €K ¡ log(v(ΘjX)) = i=1 log k=1 k p(xi ; k ) dicult to optimize b/c it contains log of sum
  • 35. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 36. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 37. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 38. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 39. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 40. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 41. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Denition Our EM problem: Z = (X; Y) v(ΘjZ) = v(ΘjX; Y) = p(X; YjΘ) yi P 1 : : : K yi = k if the ith sample was generated by the kth component Y is a random vector €N log(v(ΘjX; Y)) = i=1 log( yi p(xi jyi ))
  • 42. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Continued Denition The E- and M-steps with the auxiliary function Q: E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ] M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
  • 43. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Continued Denition The E- and M-steps with the auxiliary function Q: E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ] M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
  • 44. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Continued Denition The E- and M-steps with the auxiliary function Q: E-step: Q(Θ; Θ(i 1) ) = E [log(p(X; YjΘ))jX; Θ(i 1) ] M-step: Θ(i) = arg maxΘ Q(Θ; Θ(i 1) )
  • 45. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation The Q function Q(Θ; Θ(t) ) = E [log(p(X; YjΘ))jX; Θ(t) ] K N ˆˆ = log(k p(xi jk ))p(kjxi ; Θ(t) ) k=1 i=1 K N ˆˆ = log(k )p(kjxi ; Θ(t) ) k=1 i=1 K N ˆˆ + log(p(xi jk ))p(kjxi ; Θ(t) ) k=1 i=1
  • 46. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Solving for k € We use the Lagrange multiplier to enforce k k = 1. @ ˆˆ ˆ [ log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0 @k k i k ˆ 1 p(kjxi ; Θ(t) ) + = 0 i k Summing both sides over k, we get =  N . 1 ˆ k = p(kjxi ; Θ(t) ): N i
  • 47. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Solving for k € We use the Lagrange multiplier to enforce k k = 1. @ ˆˆ ˆ [ log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0 @k k i k ˆ 1 p(kjxi ; Θ(t) ) + = 0 i k Summing both sides over k, we get =  N . 1 ˆ k = p(kjxi ; Θ(t) ): N i
  • 48. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Solving for k € We use the Lagrange multiplier to enforce k k = 1. @ ˆˆ ˆ [ log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0 @k k i k ˆ 1 p(kjxi ; Θ(t) ) + = 0 i k Summing both sides over k, we get =  N . 1 ˆ k = p(kjxi ; Θ(t) ): N i
  • 49. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Solving for k € We use the Lagrange multiplier to enforce k k = 1. @ ˆˆ ˆ [ log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0 @k k i k ˆ 1 p(kjxi ; Θ(t) ) + = 0 i k Summing both sides over k, we get =  N . 1 ˆ k = p(kjxi ; Θ(t) ): N i
  • 50. Parametric Density Estimation using Gaussian Mixture Models Density Estimation The EM formulation Solving for k € We use the Lagrange multiplier to enforce k k = 1. @ ˆˆ ˆ [ log(k )p(kjxi ; Θ(t) ) + ( k   1)] = 0 @k k i k ˆ 1 p(kjxi ; Θ(t) ) + = 0 i k Summing both sides over k, we get =  N . 1 ˆ k = p(kjxi ; Θ(t) ): N i
  • 51. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 52. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model GMM Solving for mk and Σk Gaussian components D p(xi jxk ; Σk ) = (2)  2 jΣk j  2 exp(  1 (xi   mk )T Σ 1 (xi   mk )) 1 2 k € i xi p(kjxi ; Θ(t) ) mk = € i p(kjxi ; Θ ) (t) p(kjxi ; Θ(t) )(xi   mk )(xi   mk )T Σk = p(kjxi ; Θ(t) )
  • 53. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model GMM Solving for mk and Σk Gaussian components D p(xi jxk ; Σk ) = (2)  2 jΣk j  2 exp(  1 (xi   mk )T Σ 1 (xi   mk )) 1 2 k € i xi p(kjxi ; Θ(t) ) mk = € i p(kjxi ; Θ ) (t) p(kjxi ; Θ(t) )(xi   mk )(xi   mk )T Σk = p(kjxi ; Θ(t) )
  • 54. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Choice of Σk Full covariance Fits data best, but costly in high-dimensional space. Figure: full covariance
  • 55. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Choice of Σk Diagonal covariance A tradeo between cost and quality. Figure: diagonal covariance
  • 56. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Choice of Σk Spherical covariance, Σk = k I 2 Needs many components to cover data. Figure: spherical covariance Decision should be based on the size of training data set so that all parameters may be tuned
  • 57. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Choice of Σk Spherical covariance, Σk = k I 2 Needs many components to cover data. Figure: spherical covariance Decision should be based on the size of training data set so that all parameters may be tuned
  • 58. Parametric Density Estimation using Gaussian Mixture Models Gaussian Mixture Model Choice of Σk Spherical covariance, Σk = k I 2 Needs many components to cover data. Figure: spherical covariance Decision should be based on the size of training data set so that all parameters may be tuned
  • 59. Parametric Density Estimation using Gaussian Mixture Models Some Results Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 60. Parametric Density Estimation using Gaussian Mixture Models Some Results Data data generated from a mixture of three bivariate Gaussians
  • 61. Parametric Density Estimation using Gaussian Mixture Models Some Results PDFs estimated pdf contours (after 43 iterations)
  • 62. Parametric Density Estimation using Gaussian Mixture Models Some Results Clusters xi assigned to the cluster k corresponding to the highest p(kjxi ; Θ)
  • 63. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM Outline 1 Introduction 2 Density Estimation 3 Gaussian Mixture Model 4 Some Results 5 Other Applications of EM
  • 64. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 65. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 66. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 67. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 68. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 69. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 70. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 71. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 72. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM A problem with missing data Life-testing experiment lifetime of light bulbs follows an exponential distribution with mean two separate experiments N bulbs tested until failed failure times recorded as x1 ; : : : ; x N M bulbs tested failure times not recorded only number of bulbs , that failed at time r t missing data are failure times 1 M u ;:::;u
  • 73. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 74. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 75. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 76. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 77. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 78. Parametric Density Estimation using Gaussian Mixture Models Other Applications of EM EM formulation Life-testing experiment €M log(v(jx; u)) =  N (log + x=)   i (log + ui =) expected value for bulb still burning: t +  t k one that burned out:   1tee = (k) =   th(k) ( )   t = E-step: Q(; (k) ) =  (N + M ) log   1 (N x + (M   r)(t + k ) + r( k   th k )) ( ) ( ) ( ) M-step: 1 N +M ( N x + (M   r)(t + (k) ) + r((k)   th(k) ))
  • 79. Parametric Density Estimation using Gaussian Mixture Models Thank you for your attention. Any questions?