SlideShare a Scribd company logo
1 of 23
Download to read offline
Schedule
830-900     Introduction
900-1000    Models: small cliques
                 and special potentials
1000-1030   Tea break
1030-1200 Inference: Relaxation techniques:
            LP, Lagrangian, Dual Decomposition
1200-1230   Models: global potentials and
            global parameters + discussion
MRF with global potential
                GrabCut model [Rother et. al. ‘04]
                                                                            θF/B


 E(x,θF,θB) = ∑ Fi(θF)xi+ Bi(θB)(1-xi)           + ∑ |xi-xj|
                  i                                i,j Є N



          Fi = -log Pr(zi|θF)          Bi= -log Pr(zi|θB)

                                                     R
                                                         Background




                                                               Foreground
                                                                            G
    Image z                   Output x                 θF/B Gaussian
                                                       Mixture models


Problem: for unknown x,θF,θB the optimization is NP-hard! [Vicente et al. ‘09]
GrabCut: Iterated Graph Cuts                                         θF/B
                      [Rother et al. Siggraph ‘04]




        F                B
min E(x, θF, θB)                                                min E(x, θF, θB)
θF,θB                                                             x


  Learning of the                                           Graph cut to infer
colour distributions                                          segmentation

 Most systems with global variables work like that
 e.g. [ObjCut Kumar et. al. ‘05, PoseCut Bray et al. ’06, LayoutCRF Winn et al. ’06]
GrabCut: Iterated Graph Cuts




                       1     2     3     4

   Result       Energy after each Iteration
Colour Model




R                                         R
                               Iterated
    Background                                Background
                              graph cut

         Foreground &
         Background     G                           Foreground
                                                                 G
Optimizing over θ’s help




Input         no iteration        after convergence
           [Boykov&Jolly ‘01]       [GrabCut ‘04]




Input         after convergence
                [GrabCut ‘04]
Global optimality?



                      GrabCut       Global Optimum
                  (local optimum)   [Vicente et al. ‘09]


Is it a problem of the optimization or the model?
… first attempt to solve it
                 [Lempisky et al. ECCV ‘08]

E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj|
              pЄV                                  pq Є E

                                R                  5
                                             4
                                         3
                                     2                          wF,B
                                             6 7   8
                                 1                          G
                                         8 Gaussians
                                         whole image
Model a discrete subset:
wF= (1,1,0,1,0,0,0,0); wB = (1,0,0,0,0,0,0,1)
#solutions: wF*wB = 216

Global Optimum:
Exhaustive Search: 65.536 Graph Cuts
Branch-and-MinCut: ~ 130-500 Graph Cuts (depends on image)
Branch-and-MinCut
                                                     wF= (*,*,*,*,*,*,*,*)
                                                     wB= (*,*,*,*,*,*,*,*)
wF= (0,0,*,*,*,*,*,*)
wB= (0,*,*,*,*,*,*,*)




                                              wF= (1,1,1,1,0,0,0,0)
                                              wB= (1,0,1,1,0,1,0,0)


min BE(x,wF,wB) = min [ ∑ Fi(wF)xi+ Bi(wB)(1-xi) + ∑ wij(xi,xj) ]
x,w ,w
 F                 F B  x,w ,w

                    ≥ min [∑ min Fi(wF)xi+ min Bi(wB)(1-xi) + ∑ wij(xi,xj)]
                           x   F w           w
                                             B
Results …



   E=-618
   E = -618     E=-624 (speed-up 481)
                E = -624 (speed-up        E = -62
   GrabCut        Branch-and-MinCut
   E = -618     E = -624 (speed-up 481)   E = -62




   E = -593    E = -584 (speed-up 141)    E = -60

   E = -593
  E=-593       E = -584 (speed-up 141)
              E=-584 (speed-up 141)       E = -60
  GrabCut       Branch-and-MinCut
Object Recognition & Segmentation
       min E(x,w) with: w = Templates x Position   w
                       |w| ~ 2.000.000
Given exemplar shapes:




Test: Speed-up ~900; accuracy 98.8%
… second attempt to solve it
               [Vicente et al. ICCV ‘09]


Eliminate global color model θF,θB :
              E’(x) = min E(x,θF,θB)
                       θF,θB
Eliminate color model
      E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj|



                                                                                  K = 163

                                                                              k
        Image                discretized in bins            Image histogram
                             θF                        θB




      given x                 foreground
                                                   k
                                                            background
                                                                          k

                              distribution                  distribution
 θ Fє [0,1]K is a distributions (∑θF = 1)
                                      K
                                                       (background same)
                                       K

            Optimal θF/B given by empirical histograms: θF = nFk/nF
                                                         K


pЄV
            nF = ∑xi #fgd. pixel
                     k
            nF = ∑xi #fgd. pixel in bin k
                k                                      pЄV
                    i Є Bk
Eliminate color model
    E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj|
                    i

     E(x,θF,θB)= ∑ -nFk log θFk -nBk log θBk + ∑ wij|xi-xj|
                    k

                                  min θF,θB        (θF = nFk/nF )
                                                     K



    E’(x)= g(nF) +      ∑ hk(nk ) + ∑ wij|xi-xj|
                              F                           with nF = ∑xi, nF = ∑xi
                                                                  k
                        k                                             i Є Bk

convex g                                concave hk




            0      n/2        n   nF                 0                   max nF
                                                                               k
    Prefers “equal area” segmentation         Each color either fore- or background
How to optimize … Dual Decomposition

  E(x)= g(nF) + ∑ hk(nFk) + ∑ wij|xi-xj|
                        k
          E1(x)                      E2(x)


   min E(x) = min [ E1(x) + yTx + E2(x) – yTx ]
    x             x
              ≥ min [ E1(x’) + yTx’ ] + min [E2(x) – yTx] =: L(y)
                  x’                            x
                        Simple (no MRF)               Robust Pn Potts



Goal: - maximize concave function L(y)
        using sub-gradient                                                    E(x’)
      - no guarantees on E (NP-hard)
                                                                                L(y)

           “paramteric maxflow” gives optimal y=λ1 efficiently   [Vicente et al. ICCV ‘09]
Some results…
   Global optimum in 61% of cases (GrabCut database)
Input                 GrabCut               Global Optimum (DD)




                                            Local Optimum (DD)
Insights on the GrabCut model



                     0.3 g            0.4 g              g         1.5 g


convex g                                concave hk




            0     n/2        n   nF                  0               max nF
                                                                           k
    Prefers “equal area” segmentation         Each color either fore- or background
Relationship to Soft Pn Potts
Just different type of clustering:




    Image                One super-     another super-      GrabCut:
                         pixelization   pixelization        cluster all colors
                                                            together




                 Pairwise CRF only      robust Pn Potts
                   TextonBoost          [Kohli et al ‘08]
                [Shotton et al. ‘06]
Marginal Probability Field (MPF)
 What is the prior of a MAP-MRF solution:

Training image:                                           60% black, 40% white

MAP:                               Others less likely :
                  8
    prior(x) = 0.6 = 0.016                           prior(x) = 0.6 5 * 0.4 3 = 0.005
    MRF is a bad prior since ignores shape of the (feature) distribution !

          Introduce a global term, which controls global statistic




                                                                     [Woodford et. al. ICCV ‘09]
Marginal Probability Field (MPF)


                       0                 max       0                     max
                           True energy
                                                             MRF
Optimization done with Dual Decomposition (different ones)




                                                                   [Woodford et. al. ICCV ‘09]
Examples

               Noisy input



Ground truth                     Pairwise MRF –        Global gradient prior
                             Increase Prior strength
  In-painting:




  Segmentation:
Schedule
830-900     Introduction
900-1000    Models: small cliques
                 and special potentials
1000-1030   Tea break
1030-1200 Inference: Relaxation techniques:
            LP, Lagrangian, Dual Decomposition
1200-1230   Models: global potentials and
            global parameters + discussion
Open Questions
• Many exciting future directions
  – Exploiting latest ideas for applications (object
    recognition etc.)
  – Many other higher-order cliques:
    Topology, Grammars, etc. (this conference).

• Comparison of inference techniques needed:
  – Factor graph message passing vs. transformation vs.
    LP relaxation?

• Learning higher order Random Fields

More Related Content

What's hot

pres_IGARSS_2011_LFF_poltom.pdf
pres_IGARSS_2011_LFF_poltom.pdfpres_IGARSS_2011_LFF_poltom.pdf
pres_IGARSS_2011_LFF_poltom.pdf
grssieee
 
SPU Optimizations-part 1
SPU Optimizations-part 1SPU Optimizations-part 1
SPU Optimizations-part 1
Naughty Dog
 
SPU Optimizations - Part 2
SPU Optimizations - Part 2SPU Optimizations - Part 2
SPU Optimizations - Part 2
Naughty Dog
 
Practical Spherical Harmonics Based PRT Methods
Practical Spherical Harmonics Based PRT MethodsPractical Spherical Harmonics Based PRT Methods
Practical Spherical Harmonics Based PRT Methods
Naughty Dog
 
CVPR2010: higher order models in computer vision: Part 1, 2
CVPR2010: higher order models in computer vision: Part 1, 2CVPR2010: higher order models in computer vision: Part 1, 2
CVPR2010: higher order models in computer vision: Part 1, 2
zukun
 
Jylee probabilistic reasoning with bayesian networks
Jylee probabilistic reasoning with bayesian networksJylee probabilistic reasoning with bayesian networks
Jylee probabilistic reasoning with bayesian networks
Jungyeol
 
Tao Fayan_Iso and Full_volume rendering
Tao Fayan_Iso and Full_volume renderingTao Fayan_Iso and Full_volume rendering
Tao Fayan_Iso and Full_volume rendering
Fayan TAO
 
Spectral Learning Methods for Finite State Machines with Applications to Na...
  Spectral Learning Methods for Finite State Machines with Applications to Na...  Spectral Learning Methods for Finite State Machines with Applications to Na...
Spectral Learning Methods for Finite State Machines with Applications to Na...
LARCA UPC
 
論文紹介 Fast imagetagging
論文紹介 Fast imagetagging論文紹介 Fast imagetagging
論文紹介 Fast imagetagging
Takashi Abe
 

What's hot (17)

Lesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite IntegralsLesson 27: Evaluating Definite Integrals
Lesson 27: Evaluating Definite Integrals
 
pres_IGARSS_2011_LFF_poltom.pdf
pres_IGARSS_2011_LFF_poltom.pdfpres_IGARSS_2011_LFF_poltom.pdf
pres_IGARSS_2011_LFF_poltom.pdf
 
SPU Optimizations-part 1
SPU Optimizations-part 1SPU Optimizations-part 1
SPU Optimizations-part 1
 
SPU Optimizations - Part 2
SPU Optimizations - Part 2SPU Optimizations - Part 2
SPU Optimizations - Part 2
 
Practical Spherical Harmonics Based PRT Methods
Practical Spherical Harmonics Based PRT MethodsPractical Spherical Harmonics Based PRT Methods
Practical Spherical Harmonics Based PRT Methods
 
CVPR2010: higher order models in computer vision: Part 1, 2
CVPR2010: higher order models in computer vision: Part 1, 2CVPR2010: higher order models in computer vision: Part 1, 2
CVPR2010: higher order models in computer vision: Part 1, 2
 
The LabPQR Color Space
The LabPQR Color SpaceThe LabPQR Color Space
The LabPQR Color Space
 
確率伝播その2
確率伝播その2確率伝播その2
確率伝播その2
 
M Gumbel - SCABIO: a framework for bioinformatics algorithms in Scala
M Gumbel - SCABIO: a framework for bioinformatics algorithms in ScalaM Gumbel - SCABIO: a framework for bioinformatics algorithms in Scala
M Gumbel - SCABIO: a framework for bioinformatics algorithms in Scala
 
Jylee probabilistic reasoning with bayesian networks
Jylee probabilistic reasoning with bayesian networksJylee probabilistic reasoning with bayesian networks
Jylee probabilistic reasoning with bayesian networks
 
Tao Fayan_Iso and Full_volume rendering
Tao Fayan_Iso and Full_volume renderingTao Fayan_Iso and Full_volume rendering
Tao Fayan_Iso and Full_volume rendering
 
Introductory maths analysis chapter 17 official
Introductory maths analysis   chapter 17 officialIntroductory maths analysis   chapter 17 official
Introductory maths analysis chapter 17 official
 
Spectral Learning Methods for Finite State Machines with Applications to Na...
  Spectral Learning Methods for Finite State Machines with Applications to Na...  Spectral Learning Methods for Finite State Machines with Applications to Na...
Spectral Learning Methods for Finite State Machines with Applications to Na...
 
論文紹介 Fast imagetagging
論文紹介 Fast imagetagging論文紹介 Fast imagetagging
論文紹介 Fast imagetagging
 
Hebb network
Hebb networkHebb network
Hebb network
 
Wireless Localization: Ranging (second part)
Wireless Localization: Ranging (second part)Wireless Localization: Ranging (second part)
Wireless Localization: Ranging (second part)
 
Geometry, Topology, and all of Your Wildest Dreams Will Come True
Geometry, Topology, and all of Your Wildest Dreams Will Come TrueGeometry, Topology, and all of Your Wildest Dreams Will Come True
Geometry, Topology, and all of Your Wildest Dreams Will Come True
 

Similar to CVPR2010: higher order models in computer vision: Part 4

ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2
zukun
 
Object Detection with Discrmininatively Trained Part based Models
Object Detection with Discrmininatively Trained Part based ModelsObject Detection with Discrmininatively Trained Part based Models
Object Detection with Discrmininatively Trained Part based Models
zukun
 
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCETHE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
Frank Nielsen
 

Similar to CVPR2010: higher order models in computer vision: Part 4 (20)

Functions
FunctionsFunctions
Functions
 
Convergence of ABC methods
Convergence of ABC methodsConvergence of ABC methods
Convergence of ABC methods
 
ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2ICCV2009: MAP Inference in Discrete Models: Part 2
ICCV2009: MAP Inference in Discrete Models: Part 2
 
Fourier transform
Fourier transformFourier transform
Fourier transform
 
Formal methods 8 - category theory (last one)
Formal methods   8 - category theory (last one)Formal methods   8 - category theory (last one)
Formal methods 8 - category theory (last one)
 
Lesson 19: The Mean Value Theorem (slides)
Lesson 19: The Mean Value Theorem (slides)Lesson 19: The Mean Value Theorem (slides)
Lesson 19: The Mean Value Theorem (slides)
 
Lesson 19: The Mean Value Theorem (slides)
Lesson 19: The Mean Value Theorem (slides)Lesson 19: The Mean Value Theorem (slides)
Lesson 19: The Mean Value Theorem (slides)
 
Object Detection with Discrmininatively Trained Part based Models
Object Detection with Discrmininatively Trained Part based ModelsObject Detection with Discrmininatively Trained Part based Models
Object Detection with Discrmininatively Trained Part based Models
 
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCETHE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
THE CHORD GAP DIVERGENCE AND A GENERALIZATION OF THE BHATTACHARYYA DISTANCE
 
16 fft
16 fft16 fft
16 fft
 
Functions.ppt
Functions.pptFunctions.ppt
Functions.ppt
 
Functions.ppt
Functions.pptFunctions.ppt
Functions.ppt
 
04 1 - frequency domain filtering fundamentals
04 1 - frequency domain filtering fundamentals04 1 - frequency domain filtering fundamentals
04 1 - frequency domain filtering fundamentals
 
Ecc2
Ecc2Ecc2
Ecc2
 
Mesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh ParameterizationMesh Processing Course : Mesh Parameterization
Mesh Processing Course : Mesh Parameterization
 
image_enhancement_spatial
 image_enhancement_spatial image_enhancement_spatial
image_enhancement_spatial
 
Lesson 25: Evaluating Definite Integrals (slides
Lesson 25: Evaluating Definite Integrals (slidesLesson 25: Evaluating Definite Integrals (slides
Lesson 25: Evaluating Definite Integrals (slides
 
Evaluating definite integrals
Evaluating definite integralsEvaluating definite integrals
Evaluating definite integrals
 
Fourier transform
Fourier transformFourier transform
Fourier transform
 
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
 

More from zukun

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009
zukun
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCV
zukun
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Information
zukun
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statistics
zukun
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibration
zukun
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer vision
zukun
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluation
zukun
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-software
zukun
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptors
zukun
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectors
zukun
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-intro
zukun
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video search
zukun
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video search
zukun
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video search
zukun
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learning
zukun
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer vision
zukun
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick start
zukun
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysis
zukun
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structures
zukun
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities
zukun
 

More from zukun (20)

My lyn tutorial 2009
My lyn tutorial 2009My lyn tutorial 2009
My lyn tutorial 2009
 
ETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCVETHZ CV2012: Tutorial openCV
ETHZ CV2012: Tutorial openCV
 
ETHZ CV2012: Information
ETHZ CV2012: InformationETHZ CV2012: Information
ETHZ CV2012: Information
 
Siwei lyu: natural image statistics
Siwei lyu: natural image statisticsSiwei lyu: natural image statistics
Siwei lyu: natural image statistics
 
Lecture9 camera calibration
Lecture9 camera calibrationLecture9 camera calibration
Lecture9 camera calibration
 
Brunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer visionBrunelli 2008: template matching techniques in computer vision
Brunelli 2008: template matching techniques in computer vision
 
Modern features-part-4-evaluation
Modern features-part-4-evaluationModern features-part-4-evaluation
Modern features-part-4-evaluation
 
Modern features-part-3-software
Modern features-part-3-softwareModern features-part-3-software
Modern features-part-3-software
 
Modern features-part-2-descriptors
Modern features-part-2-descriptorsModern features-part-2-descriptors
Modern features-part-2-descriptors
 
Modern features-part-1-detectors
Modern features-part-1-detectorsModern features-part-1-detectors
Modern features-part-1-detectors
 
Modern features-part-0-intro
Modern features-part-0-introModern features-part-0-intro
Modern features-part-0-intro
 
Lecture 02 internet video search
Lecture 02 internet video searchLecture 02 internet video search
Lecture 02 internet video search
 
Lecture 01 internet video search
Lecture 01 internet video searchLecture 01 internet video search
Lecture 01 internet video search
 
Lecture 03 internet video search
Lecture 03 internet video searchLecture 03 internet video search
Lecture 03 internet video search
 
Icml2012 tutorial representation_learning
Icml2012 tutorial representation_learningIcml2012 tutorial representation_learning
Icml2012 tutorial representation_learning
 
Advances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer visionAdvances in discrete energy minimisation for computer vision
Advances in discrete energy minimisation for computer vision
 
Gephi tutorial: quick start
Gephi tutorial: quick startGephi tutorial: quick start
Gephi tutorial: quick start
 
EM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysisEM algorithm and its application in probabilistic latent semantic analysis
EM algorithm and its application in probabilistic latent semantic analysis
 
Object recognition with pictorial structures
Object recognition with pictorial structuresObject recognition with pictorial structures
Object recognition with pictorial structures
 
Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities Iccv2011 learning spatiotemporal graphs of human activities
Iccv2011 learning spatiotemporal graphs of human activities
 

Recently uploaded

The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
heathfieldcps1
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
QucHHunhnh
 
An Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdfAn Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdf
SanaAli374401
 

Recently uploaded (20)

Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1Código Creativo y Arte de Software | Unidad 1
Código Creativo y Arte de Software | Unidad 1
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
Presentation by Andreas Schleicher Tackling the School Absenteeism Crisis 30 ...
 
This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.This PowerPoint helps students to consider the concept of infinity.
This PowerPoint helps students to consider the concept of infinity.
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
SECOND SEMESTER TOPIC COVERAGE SY 2023-2024 Trends, Networks, and Critical Th...
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
An Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdfAn Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdf
 

CVPR2010: higher order models in computer vision: Part 4

  • 1. Schedule 830-900 Introduction 900-1000 Models: small cliques and special potentials 1000-1030 Tea break 1030-1200 Inference: Relaxation techniques: LP, Lagrangian, Dual Decomposition 1200-1230 Models: global potentials and global parameters + discussion
  • 2. MRF with global potential GrabCut model [Rother et. al. ‘04] θF/B E(x,θF,θB) = ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ |xi-xj| i i,j Є N Fi = -log Pr(zi|θF) Bi= -log Pr(zi|θB) R Background Foreground G Image z Output x θF/B Gaussian Mixture models Problem: for unknown x,θF,θB the optimization is NP-hard! [Vicente et al. ‘09]
  • 3. GrabCut: Iterated Graph Cuts θF/B [Rother et al. Siggraph ‘04] F B min E(x, θF, θB) min E(x, θF, θB) θF,θB x Learning of the Graph cut to infer colour distributions segmentation Most systems with global variables work like that e.g. [ObjCut Kumar et. al. ‘05, PoseCut Bray et al. ’06, LayoutCRF Winn et al. ’06]
  • 4. GrabCut: Iterated Graph Cuts 1 2 3 4 Result Energy after each Iteration
  • 5. Colour Model R R Iterated Background Background graph cut Foreground & Background G Foreground G
  • 6. Optimizing over θ’s help Input no iteration after convergence [Boykov&Jolly ‘01] [GrabCut ‘04] Input after convergence [GrabCut ‘04]
  • 7. Global optimality? GrabCut Global Optimum (local optimum) [Vicente et al. ‘09] Is it a problem of the optimization or the model?
  • 8. … first attempt to solve it [Lempisky et al. ECCV ‘08] E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj| pЄV pq Є E R 5 4 3 2 wF,B 6 7 8 1 G 8 Gaussians whole image Model a discrete subset: wF= (1,1,0,1,0,0,0,0); wB = (1,0,0,0,0,0,0,1) #solutions: wF*wB = 216 Global Optimum: Exhaustive Search: 65.536 Graph Cuts Branch-and-MinCut: ~ 130-500 Graph Cuts (depends on image)
  • 9. Branch-and-MinCut wF= (*,*,*,*,*,*,*,*) wB= (*,*,*,*,*,*,*,*) wF= (0,0,*,*,*,*,*,*) wB= (0,*,*,*,*,*,*,*) wF= (1,1,1,1,0,0,0,0) wB= (1,0,1,1,0,1,0,0) min BE(x,wF,wB) = min [ ∑ Fi(wF)xi+ Bi(wB)(1-xi) + ∑ wij(xi,xj) ] x,w ,w F F B x,w ,w ≥ min [∑ min Fi(wF)xi+ min Bi(wB)(1-xi) + ∑ wij(xi,xj)] x F w w B
  • 10. Results … E=-618 E = -618 E=-624 (speed-up 481) E = -624 (speed-up E = -62 GrabCut Branch-and-MinCut E = -618 E = -624 (speed-up 481) E = -62 E = -593 E = -584 (speed-up 141) E = -60 E = -593 E=-593 E = -584 (speed-up 141) E=-584 (speed-up 141) E = -60 GrabCut Branch-and-MinCut
  • 11. Object Recognition & Segmentation min E(x,w) with: w = Templates x Position w |w| ~ 2.000.000 Given exemplar shapes: Test: Speed-up ~900; accuracy 98.8%
  • 12. … second attempt to solve it [Vicente et al. ICCV ‘09] Eliminate global color model θF,θB : E’(x) = min E(x,θF,θB) θF,θB
  • 13. Eliminate color model E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj| K = 163 k Image discretized in bins Image histogram θF θB given x foreground k background k distribution distribution θ Fє [0,1]K is a distributions (∑θF = 1) K (background same) K Optimal θF/B given by empirical histograms: θF = nFk/nF K pЄV nF = ∑xi #fgd. pixel k nF = ∑xi #fgd. pixel in bin k k pЄV i Є Bk
  • 14. Eliminate color model E(x,θF,θB)= ∑ Fi(θF)xi+ Bi(θB)(1-xi) + ∑ wij|xi-xj| i E(x,θF,θB)= ∑ -nFk log θFk -nBk log θBk + ∑ wij|xi-xj| k min θF,θB (θF = nFk/nF ) K E’(x)= g(nF) + ∑ hk(nk ) + ∑ wij|xi-xj| F with nF = ∑xi, nF = ∑xi k k i Є Bk convex g concave hk 0 n/2 n nF 0 max nF k Prefers “equal area” segmentation Each color either fore- or background
  • 15. How to optimize … Dual Decomposition E(x)= g(nF) + ∑ hk(nFk) + ∑ wij|xi-xj| k E1(x) E2(x) min E(x) = min [ E1(x) + yTx + E2(x) – yTx ] x x ≥ min [ E1(x’) + yTx’ ] + min [E2(x) – yTx] =: L(y) x’ x Simple (no MRF) Robust Pn Potts Goal: - maximize concave function L(y) using sub-gradient E(x’) - no guarantees on E (NP-hard) L(y) “paramteric maxflow” gives optimal y=λ1 efficiently [Vicente et al. ICCV ‘09]
  • 16. Some results… Global optimum in 61% of cases (GrabCut database) Input GrabCut Global Optimum (DD) Local Optimum (DD)
  • 17. Insights on the GrabCut model 0.3 g 0.4 g g 1.5 g convex g concave hk 0 n/2 n nF 0 max nF k Prefers “equal area” segmentation Each color either fore- or background
  • 18. Relationship to Soft Pn Potts Just different type of clustering: Image One super- another super- GrabCut: pixelization pixelization cluster all colors together Pairwise CRF only robust Pn Potts TextonBoost [Kohli et al ‘08] [Shotton et al. ‘06]
  • 19. Marginal Probability Field (MPF) What is the prior of a MAP-MRF solution: Training image: 60% black, 40% white MAP: Others less likely : 8 prior(x) = 0.6 = 0.016 prior(x) = 0.6 5 * 0.4 3 = 0.005 MRF is a bad prior since ignores shape of the (feature) distribution ! Introduce a global term, which controls global statistic [Woodford et. al. ICCV ‘09]
  • 20. Marginal Probability Field (MPF) 0 max 0 max True energy MRF Optimization done with Dual Decomposition (different ones) [Woodford et. al. ICCV ‘09]
  • 21. Examples Noisy input Ground truth Pairwise MRF – Global gradient prior Increase Prior strength In-painting: Segmentation:
  • 22. Schedule 830-900 Introduction 900-1000 Models: small cliques and special potentials 1000-1030 Tea break 1030-1200 Inference: Relaxation techniques: LP, Lagrangian, Dual Decomposition 1200-1230 Models: global potentials and global parameters + discussion
  • 23. Open Questions • Many exciting future directions – Exploiting latest ideas for applications (object recognition etc.) – Many other higher-order cliques: Topology, Grammars, etc. (this conference). • Comparison of inference techniques needed: – Factor graph message passing vs. transformation vs. LP relaxation? • Learning higher order Random Fields