SlideShare a Scribd company logo
1 of 45
Download to read offline
Putting Objects in Perspective
              Derek Hoiem, Alexei A. Efros, Martial Hebert


                              presented by
                        Phongsathorn Eakamongul

                           Department of Computer Science
                            Asian Institute of Technology



                                  2009, July 2




Phongsathorn (AIT)            Putting Objects in Perspective   Short Occasion   1 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   2 / 45
Abstract




Statistic Framework for placing local object detection by modeling interdependence of
    camera viewpoint (position/orientation)
    objects identities
    surface orientations (3D scene geometry)
using single image.




       Phongsathorn (AIT)          Putting Objects in Perspective        Short Occasion   3 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   4 / 45
General Object Detection cannot be solved locally
We cannot recognize car, person, and the road in isolation




Pixel of car can be easily be interpreted as person’s shoulder, mouse, stack of books,
balcony, etc.

Psychophysical Evidence
context plays crucial role in scene understanding (i.e. Biederman 1981; Torralba 2005).
(1) car is sitting on the road
(2) right size relative to other object in the scene.


         Phongsathorn (AIT)                  Putting Objects in Perspective   Short Occasion   5 / 45
Related Work


   Heuristic Image understanding systems
        ACRONYM (Brooks 1979), VISIONS (Hanson and Riseman 1978), outdoor scene
        understanding system (Ohta and Kanade 1985), intrinsic images (Barrow and
        Tenenbaum 1978), etc.
   Small image windows (at all locations and scales) to find specific objects
        works wonderfully with face detection (Schneiderman 2004; Viola and Jones 2004) –
        since inside the face is much more important than boundary.
        unreliable for cars, pedestrians, especially at smaller scale.
   Modeling within 2D image
        modeling direct relationships between objects and other objects (Kumar and Herbert
        2003; Murphy et al. 2003), regions (He et al. 2004; Kumar and Hebert 2005; Tu et al.
        2005), scene geometries (Murphy et al. 2003; Sudderth et al. 2005)
   Modeling 3D image
        estimate rough 3D scene geometry from single image (Hoiem et al. 2005), Viewpoint
        and mean scene depth (Torralba and Sinha 2001), geometric consistency of object
        hypotheses in simple scenes using hard algebric constriants (Forsyth et al. 1994)
   Relationship between camera parameters and objects
        well calibrated camera (Jeong et al. 2001, etc.), stationary surveillance camera
        (Krahnstoever and Mendonca 2005), or both (Greienhagen et al. 2000)



     Phongsathorn (AIT)              Putting Objects in Perspective              Short Occasion   6 / 45
Overview




100 boxes sampled
(b) local object detector (Murphy et al. 2003; Dalal and Triggs 2005) at any
location/scale
(c) estimate rough surface geometry (Hoiem et al. 2005)
red, green, blue channels indicate vertical, ground, sky
(e) estimate camera viewpoint from scale of objects in image

This is like typical human eye-tracking, where subjects are asked to search for a
person in an image.

       Phongsathorn (AIT)          Putting Objects in Perspective         Short Occasion   7 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   8 / 45
Projective geometry
object’s height in image can be determined from it’s height in the world




                                                                        v0 : horizon position
                                                                        ui , vi : lower-left corordinate
   yc : camera height                                                   hi : object image height
   –                                                                    –
   yi : object height                                                   vc : center of image
                                                                        vt , vb : top and bottom of object
Assume
    all objects of interest rest on the ground plane. ( cannot find people on the roof )
    objects perpendicular to the ground

We can estimate object world height
        hi yc
yi ≈   v0 −vi

          Phongsathorn (AIT)                  Putting Objects in Perspective                     Short Occasion   9 / 45
Camera Tilt




camera tilt
                 −v
θx = 2arctan( vc 2f 0 )




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   10 / 45
Proof

Assume
         camera roll = 0 or image is rectified

        camera intrinsic parameters are typical ( skew = 0, unit aspect ratio, typical f )
                                                                                   2       3
2       3     2                     32                                           3      x
    u               f     0   uc          1          0           0         0       6 y 7
4   v  5 = 1 4 0          f   vc 5 4 0           cosθx      −sinθx        yc 5 4   6       7
            z                                                                           z 5
    1               0     0    1          0       sinθx      cosθx         0
                                                                                        1
From this,

                                                                z(fsinθx − (vc − v )cosθx − fyc )
                                                          y =                                                          (1)
                                                                      (vc − v )sinθx + fcosθx

Since y=0 at vb ,

                                                                                 fyc
                                                                z =                                                    (2)
                                                                      fsinθx − (vc − vb )cosθx

put (2) in (1)
      fyc (fsinθx )−(vc −vt )cosθx /fsinθx −(vc −vb )cosθx −fyc
y =
                         (vc −vt )sinθx +fcosθx
Consider condition that camera is on the ground,

                                                                 vc −v0
         Since θx is small; cosθx ≈ 1, sinθx ≈ θx , θx ≈
                                                                    f
                 v −vb
         y ≈ yc t        /(1 + (vc − v0 )(vc − vt )/f 2 )
                 v0 −vb

         Since tilt tend to be small; vc − v0 ≈ 0, and f > 1
                    v −vb
         y ≈ yc t
                   v0 −vb

             hi yc
so yi ≈
            v0 −vi




              Phongsathorn (AIT)                                 Putting Objects in Perspective     Short Occasion   11 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   12 / 45
Probabilistic model to describe a scene


Assume θ, o, g are conditionally independent.

                                                  From model
                                                  P(θ, o, g, eg , eo ) =
                                                       Y
                                                  P(θ)     P(oi |θ)P(eoi |oi )P(gi |oi )P(egi |gi )
                                                             i
                                                  The likelihood of the scene conditioned on
                                                  image evidence using Bayes rule

                                                  P(θ, o, g|eg , eo ) =
                                                       Y             P(oi |eoi ) P(gi |egi )
                                                  P(θ)     P(oi |θ)
                                                                       P(oi )     P(gi )
θ : viewpoint                                                i
o : object identity
g : surface geometry                              Classifier can be incorporated using
eo : object evidence                              probabilistic output P(oi |eoi )
eg : geometry evidence



       Phongsathorn (AIT)         Putting Objects in Perspective                     Short Occasion   13 / 45
Inference




   Pearl’s belief propagation algorithm (Pearl 1988) is optimal and efficient.
   inference by quantizing continuous variable v0 and yc into evenly-spaced bins ( 50
   and 100 bins, respectively )
   Their implementation make use of Bayes Net Toolbox (Murphy 2001)
   After inference, they can query, such as
        "what is expected height of this object ?"
        "what are marginal probability of cars ?"
        "What is the most probable explanation of the scene ?"
   They report results based on marginal probability from sum-product algorithm (
   this allows ROC curve to be computed )




     Phongsathorn (AIT)             Putting Objects in Perspective      Short Occasion   14 / 45
Camera Viewpoint (θ)




Consider v0 , yc to be independent a priori.

P(θ) = P(v0 )P(yc )

    v0 distribution is estimated using Simple Gaussian prior
    yc distribution is estimated using kernel density estimation .
A priori,
Most likely yc = 1.67, at eye level of typical adult male
Most likely v0 = 0.50




       Phongsathorn (AIT)            Putting Objects in Perspective   Short Occasion   15 / 45
Camera Viewpoint




(e) Viewpoint prior θ (left axis - yc , right axis - v0 ) – high variance, more informative than uniform distribution that implicitly assumed when scale is
considered irrelevant.
(g, h) local object detection
(b, c, d) Geometry estimates gi
–
Using this model, they improve estimates of
(f) Viewpoint : full
Viewpoint peak likelihood increases from 0.0037 a priori to 0.0503 after inference.
(i, j) local detection improve when viewpoint and surface geometry are considered.
At the same false positive ( cars: cyan, peds: yellow ) rate, the true detection (cars: green, peds: red) rate doubles.

Note : detections with lower confidence exist for both local and full model but not shown here.




             Phongsathorn (AIT)                                   Putting Objects in Perspective                                          Short Occasion      16 / 45
Object Identities (oi )


oi consists of
     ti : type ∈ {object, background} (i.e. pedestrian) : identity
     bboxi : boundingbox = {ui , vi , wi , hi }
           wi : width
           hi : height


Prob of object occurring at a particular location/scale
                                                    P(Ii |ti = obj, bboxi )P(oi )
P(ti = obj, bboxi |Ii ) =
                                 P(Ii |ti = obj, bboxi )P(oi ) + P(Ii |ti = obj, bboxi )(1 − P(oi ))
                                               1
                             =                      P(oi )
                                 1 + exp[−ci − log 1−P(o ) ]
                                                                 i
P(oi ) : prior
Ii : image information at ith bounding box
class-conditional log-likelihood ratio : ci = log P(Ii |ti =obj,bboxi )
                                                  P(I |t =obj,bbox ) i i       i




        Phongsathorn (AIT)                    Putting Objects in Perspective              Short Occasion   17 / 45
Object Identities (oi )




   Typically, researcher perform non-maxima suppression , assuming that high
   detection responses at neighboring positions could be due to an object at either of
   those positions (but not both).
   Applied non-maxima suppression + form a point distribution out of the
   non-maxima ( rather than discarding them )
   An object candidate is formed out of a group of closely overlapping bounding
   boxes.




      Phongsathorn (AIT)         Putting Objects in Perspective        Short Occasion   18 / 45
Object Identities (oi )




The candidate likelihood
P(ti = obj|eo ) = the likelihood of the highest confidence bounding box.
P(bboxi |ti = obj, eo ) ∝ P(ti = obj, bbox|I)


Object’s height depends on its position when given viewpoint.
P(oi |θ) ∝ p(hi |ti , vi , θ)
( due to the uniformity of P(ti , vi , wi |θ) )

From yi ≈ vhi−vi
               yc
             0
if yi is normal distribution, with parameters {µi , σi }
hi is also normal distribution with parameters { µi (vyc−vi ) ,
                                                        0                  σi (v0 −vi)
                                                                                yc
                                                                                       }




        Phongsathorn (AIT)                Putting Objects in Perspective                   Short Occasion   19 / 45
Surface Geometry (gi )

(Hoiem et al 2005) produces "confidence maps" of geometric surface for 3 classes
                                                               sky
                                                               vertical, contains 5 subclasses
                                                                        planar, facing
                                                                             left (←)
                                                                             center (↑)
                                                                             right (→)
                                                                        non-planar
                                                                             solid (O)
                                                                             porous (X)
             Figure: Geometric Label
                                                               ground
Define gi = three values, corresponding to
       whether "object surface" is visible in detection window
       whether the ground is visible just below the detection window
i.e.
       consider "car" as "planar or non-planar solid"
       consider "pedestrian" as "non-planar solid"


         Phongsathorn (AIT)            Putting Objects in Perspective                     Short Occasion   20 / 45
Surface Geometry (gi )




   Compute P(gi |oi ) and P(gi ) by counting occurrences of value of gi for both
   people and cars in training set
        If oi is background, consider P(gi |oi ) ≈ P(gi )
   Estimate P(gi |egi ) based on confidence maps.
   They found that "average geometric confidence in a window" is a "well-calibrated
   probability for the geometric value".




     Phongsathorn (AIT)                Putting Objects in Perspective   Short Occasion   21 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   22 / 45
Training
Viewpoint ( θ )




yc distribution is estimated using kernel density estimation .
      Manually labeled cars ( i.e. vans, trucks ), pedestrians in 60 outdoor images from
      LabelMe dataset (Russell et al. 2005).
      Calculate maximum likelihood estimate of camera height (yc ) based on labeled
      horizon and height distributions of cars and people in world.
      yc is estimated using "kernel density estimation" ( ksdensity in Matlab ).




         Phongsathorn (AIT)          Putting Objects in Perspective         Short Occasion   23 / 45
Training
Object Identities (oi )




       use local detector (Murphy et al. 2003)
             same local patch template features + 6 color features ( encode average L*a*b*color of
             detection window, and difference between detection window and surrounding area. )
             verified using PASCAL challange training/validation set : did not search for car shorter
             than 10% of image height ( the official entries could not detect small cars ).
             Result : average precision of 0.423 ( the best scores reported by top 3 groups: 0.613,
             0.489, 0.353 )




          Phongsathorn (AIT)              Putting Objects in Perspective             Short Occasion   24 / 45
Training
Object Identities (oi )




       logistic regression version of Adaboost classifier (Collins et al 2002) to boost
       8-node decision tree classifiers
             For car, they trained 2 views (front/back: 32x24 px, side: 40x16 px)
             For pedestrains, they trained 1 view (16x40 px)
       using full PASCAL dataset (PASCAL 2005)




          Phongsathorn (AIT)              Putting Objects in Perspective            Short Occasion   25 / 45
Training
Object Identities (oi )




       Estimate distribution of
             cars’ height ( yi ) : use data from Consumer Reports ( www.consumerreports.org )
             Result : µ = 1.59 m, σ = 0.21 m
             pedestrains’ height ( yi ) : use data from National Center for Health Statistics (
             www.cdc.edu/nchs )
             Result : µ = 1.7 m, σ = 0.085 m




          Phongsathorn (AIT)             Putting Objects in Perspective            Short Occasion   26 / 45
Training
Surface Geometry ( gi )




      Compute P(gi |oi ) by counting occurrences of value of gi for both people and cars
      in 60 training set from LabelMe
      Set P(gi ) to be uniform , because the learned values of P(gi ) over-relying on
      geometry.
      The over-reliance may be due to
            labeled image (general outdoor) being drawn from a different distribution than test set
            (streets of Boston)
            lack of modeled direct dependence between surface geometries.




         Phongsathorn (AIT)              Putting Objects in Perspective             Short Occasion   27 / 45
Evaluation




   Test Set : 422 random outdoor images from LabelMe dataset (Russell et al. 2005)
        no car or pedestrians : 60 images
        only pedestrians : 40 images
        only cars : 94 images
        both cars and pedestrians : 225 images
   Total 923 cars, 720 pedestrians
   Detect cars’ height as small as 14 px
   Detect Pedestrians’ height as small as 36 px
   To get detection confidences for each window, they reverse the process in ( in
   Object identities slide ).
   Then determine bounding boxes of objects in standard way (thresholding the
   confidences and performing non-maxima suppression)




     Phongsathorn (AIT)            Putting Objects in Perspective     Short Occasion   28 / 45
Viewpoint and Surface geometry aids Object detection




                          Figure: Object Detection Result : ROC curves




     Phongsathorn (AIT)                Putting Objects in Perspective    Short Occasion   29 / 45
About 20% of cars and pedestrians missed by the baseline are
detected for the same FP




              Figure: FP : false positive rates, and % reductions in false negatives




     Phongsathorn (AIT)                Putting Objects in Perspective              Short Occasion   30 / 45
Object and geometry improve horizon estimation



Manually labeled horizon of 100 images that contained Car+Ped.




                           Figure: Horizon estimation results


Mean/median absolute error (as percentage of image height) are shown for horizon
estimates.




      Phongsathorn (AIT)          Putting Objects in Perspective     Short Occasion   31 / 45
Horizon estimation, and object detection are more accurate when "more
object models" of the scene are known.



detect only car, only pedestrians, and both




Figure: (Horizon) : median absolute error, (First Number) : FP/image ( at 50% detection rate )
computed over all images, (Second Number) : FP/image ( at 50% detection rate ) computed over
subset of images that contain both cars and people




       Phongsathorn (AIT)             Putting Objects in Perspective           Short Occasion   32 / 45
Any local object detector can be plugging into their framework


   After converting SVM outputs to probabilities using method of (Platt 2000).
   Training 80x24 car detector; 32x96 big pedestrian detector, 16x48 small
   pedestrian detector




                     Figure: Detal-Triggs local detector (Detal and Triggs 2005)




      Phongsathorn (AIT)                 Putting Objects in Perspective            Short Occasion   33 / 45
Detection Example
baseline local detector (Murphy et al. 2003), and full model detector




                                                                         blue line = horizon estimate (
                                                                        0.5 initially )
                                                                        green box = true car, cyan box
                                                                        = false car
                                                                        red box = true ped, yellow =
                                                                        false ped
                                                                        solid line – high confidence
                                                                        detection ( 0.5 FP/Image )
                                                                        dotted line – low confidence
                                                                        detection ( 2 FP/Image )




      In most case, horizon line is correctly recovered, obj detection improves
      considerably.
      Box that make no sense (i.e. wrong scale (d), above horizon (b), in the middle of
      the ground (e)) is removed.
      Predestrians are often hallucinated (c, e) in places where they could be (but are
      not).
      (f) : a bad geometry estimate and repeated false detections along the building
      windows causes the horizon estimate to become worse and the car to be missed./ 45
          Phongsathorn (AIT)          Putting Objects in Perspective        Short Occasion 34
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   35 / 45
improve estimating v0
using detections of known objects in scene



If image does not include easily detectable known object ?
      "Edges and perspective geometry" using a lots of parallel lines (Manhattan
      Worlds) (Kosecka and Zhang 2022), (Coughlan and Yuille 2003)
            fail for less structured environments.
      using "spatial-frequency" statistics ( the "gist" ) of image to recover a general
      sense of space of the scene. (Oliva and Torralba 2006)
            They estimate rough "scene envelop" of properties such as "openess, smoothness,
            depth".
      "gist" statistics could also provide a rough estimate of "horizon position" (v0 ).
      (Torralba and Sinha 2001)
Use gist descriptor to estimate v0 instead of Simple Gaussian Prior

To obtain more precise estimate, we require
  1   large amount of training set ( image/viewpoint pairs )
  2   a way of accurately labeling this data, automatically



         Phongsathorn (AIT)                  Putting Objects in Perspective    Short Occasion   36 / 45
Automatically labeling data
to get v0 and new distribution of yi



using iterative EM-like algorithm (Lalongde te al. 2007)
      initially guess of µ and σ of people’s height (yi )
      iterates
         1   estimate "viewpoint" ( v0 ) for images that contain objects of known "size distributions" (
             yi )
         2   estimate "size distributions" ( yi ) of objects that are contained in images with known
             "viewpoints" ( v0 ).




      Figure: automatic object height estimation result, (left) picture from LabelMe (center) original
      pixel size (right) automatically estimated 3D height ( yi )




          Phongsathorn (AIT)               Putting Objects in Perspective              Short Occasion   37 / 45
Automatically labeling data
to get v0 and new distribution of yi




      Training : infer viewpoint of over 5,000 images and height of 13,000 object
      instances in 50 object classes.
      based on inferred object heights (yi ), re-estimate yi distribution of cars (
      µ = 1.51m, σ = 0.191m ), people ( µ = 1.70m, σ = 0.103m )
      consider camera viewpoint vi estimates to be reliable for 2,660 images (excluding
      images in test set) that contains at least 2 objects with known height distributions.
      calculate gist statistics (8x8 blocks at 3 scales with 8,4,4 orientations, giving 1280
      variables per gist vector).




          Phongsathorn (AIT)           Putting Objects in Perspective          Short Occasion   38 / 45
gist descriptor

     Evaluate gist nearest neighbour classifiers with various distance matrics
           nearest neighbor regression method (Navot et al. 2006)
           GRANN (Generalized Regression neural networks) – provide lowest error after tuning
           α = 0.22 on training set

                                    ˜
The horizon estimation using GRANN (v0 )
          P                            2
            i v w
v0 (x) = P 0i i ; wi = exp[ −||x−xi || ]
˜                                 2α2
              i wi
x : gist statistics
v0i : horizontal position for ith training sample


                          −˜
p(v0 |v0 ) = 2s exp[ −|v0v v0 | ] ( laplace probability density )
      ˜       1
                        s
               0
sv : scale parameter = expected error

     Result : correlation (coefficient 0.27) between error and Euclidean distance to the
     nearest neighbour in gist statistics.
                                            ˜
     Provide better estimate, by consider d ( nearest neighbour distance )
                            ˜
     fit sv = 0.022 + 0.060d, by maximum likelihood estimation over training set

        Phongsathorn (AIT)                 Putting Objects in Perspective      Short Occasion   39 / 45
Evaluation




                                                                                  ˜
          Figure: horizontal estimation using gist nearest neighbour and distance d




            Figure: mean error of horizon position estimation ( % of image height )


   Gaussian Prior
   Gaussian Prior + surface geometry, object ( using Datal-Triggs detectors )
   Gist estimation
   Gist estimation + surface geometry, object ( using Datal-Triggs detectors )
     Phongsathorn (AIT)               Putting Objects in Perspective            Short Occasion   40 / 45
Evaluation




improve detection rates ( This is measured at 1 FP/image using Dalal-Triggs object
detectors )
    cars detection 50 % to 52 %
    pedestrians 66 % to 68 %




       Phongsathorn (AIT)         Putting Objects in Perspective        Short Occasion   41 / 45
Improve estimating yc




   cannot improve significantly over prior estimate
   may because small change in yc has little impact on the global statistics.
   Therefore, re-estimate yc using larger training set




     Phongsathorn (AIT)           Putting Objects in Perspective        Short Occasion   42 / 45
Outline



1   Abstract

2   Introduction

3   Projective geometry

4   Modeling the scene

5   Training + Evaluation

6   Viewpoint by Example

7   Summary




        Phongsathorn (AIT)   Putting Objects in Perspective   Short Occasion   43 / 45
Summary




    Project estimated ground surface into overhead view, using estimated camera
    viewpoint.
    Assume cars mostly face down the same line.
What they have done : "bag of features/blackbox" classification method
  1 subtle relationship (i.e. object sizes relate through the viewpoint) can be easily
    represented.
  2 additions and extensions to the model are easy.
could be extended by modeling other scene properties ( i.e. scene category )
       Phongsathorn (AIT)           Putting Objects in Perspective         Short Occasion   44 / 45
Summary




  geometry assumptions
  not consider as conditionally independent, if two object patches are very similar.
  i.e. repeated windows in (f) causes object detection responses to be correlated.
  solution : correlation should be modeled. i.e. if two object patches are very similar,
  consider their evident jointly.
  the surface trained in this experiment is outdorr image only.




    Phongsathorn (AIT)           Putting Objects in Perspective          Short Occasion   45 / 45

More Related Content

What's hot

Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
02 probabilistic inference in graphical models
02 probabilistic inference in graphical models02 probabilistic inference in graphical models
02 probabilistic inference in graphical modelszukun
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Pierre Jacob
 
Computation of the marginal likelihood
Computation of the marginal likelihoodComputation of the marginal likelihood
Computation of the marginal likelihoodBigMC
 
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...dishii
 
Ada boost brown boost performance with noisy data
Ada boost brown boost performance with noisy dataAda boost brown boost performance with noisy data
Ada boost brown boost performance with noisy dataShadhin Rahman
 
Lesson 19: The Mean Value Theorem
Lesson 19: The Mean Value TheoremLesson 19: The Mean Value Theorem
Lesson 19: The Mean Value TheoremMatthew Leingang
 
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...Alex (Oleksiy) Varfolomiyev
 
Fast Structure From Motion in Planar Image Sequences
Fast Structure From Motion in Planar Image SequencesFast Structure From Motion in Planar Image Sequences
Fast Structure From Motion in Planar Image SequencesLuigi Bagnato
 
05 history of cv a machine learning (theory) perspective on computer vision
05  history of cv a machine learning (theory) perspective on computer vision05  history of cv a machine learning (theory) perspective on computer vision
05 history of cv a machine learning (theory) perspective on computer visionzukun
 
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...jfrchicanog
 
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)Joo-Haeng Lee
 
Portfolios and Risk Premia for the Long Run
Portfolios and Risk Premia for the Long RunPortfolios and Risk Premia for the Long Run
Portfolios and Risk Premia for the Long Runguasoni
 
Density operators
Density operatorsDensity operators
Density operatorswtyru1989
 

What's hot (17)

Fdtd
FdtdFdtd
Fdtd
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
02 probabilistic inference in graphical models
02 probabilistic inference in graphical models02 probabilistic inference in graphical models
02 probabilistic inference in graphical models
 
Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...Estimation of the score vector and observed information matrix in intractable...
Estimation of the score vector and observed information matrix in intractable...
 
BMC 2012 - Invited Talk
BMC 2012 - Invited TalkBMC 2012 - Invited Talk
BMC 2012 - Invited Talk
 
Computation of the marginal likelihood
Computation of the marginal likelihoodComputation of the marginal likelihood
Computation of the marginal likelihood
 
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
D. Ishii, K. Ueda, H. Hosobe, A. Goldsztejn: Interval-based Solving of Hybrid...
 
Ada boost brown boost performance with noisy data
Ada boost brown boost performance with noisy dataAda boost brown boost performance with noisy data
Ada boost brown boost performance with noisy data
 
Lesson 19: The Mean Value Theorem
Lesson 19: The Mean Value TheoremLesson 19: The Mean Value Theorem
Lesson 19: The Mean Value Theorem
 
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...
Optimal Finite Difference Grids for Elliptic and Parabolic PDEs with Applicat...
 
Fast Structure From Motion in Planar Image Sequences
Fast Structure From Motion in Planar Image SequencesFast Structure From Motion in Planar Image Sequences
Fast Structure From Motion in Planar Image Sequences
 
4. cft
4. cft4. cft
4. cft
 
05 history of cv a machine learning (theory) perspective on computer vision
05  history of cv a machine learning (theory) perspective on computer vision05  history of cv a machine learning (theory) perspective on computer vision
05 history of cv a machine learning (theory) perspective on computer vision
 
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
Exact Computation of the Expectation Curves of the Bit-Flip Mutation using La...
 
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)
Note on Coupled Line Cameras for Rectangle Reconstruction (ACDDE 2012)
 
Portfolios and Risk Premia for the Long Run
Portfolios and Risk Premia for the Long RunPortfolios and Risk Premia for the Long Run
Portfolios and Risk Premia for the Long Run
 
Density operators
Density operatorsDensity operators
Density operators
 

Viewers also liked

Figure 1
Figure 1Figure 1
Figure 1butest
 
Synchronization of coupled oscillators is a game
Synchronization of coupled oscillators is a gameSynchronization of coupled oscillators is a game
Synchronization of coupled oscillators is a gamemehtapgresearch
 
Figure 1
Figure 1Figure 1
Figure 1butest
 
Intro to Multitarget Tracking for CURVE
Intro to Multitarget Tracking for CURVEIntro to Multitarget Tracking for CURVE
Intro to Multitarget Tracking for CURVEchenhm
 
Kalman Filters and Adaptive Windows for Learning in Data Streams
Kalman Filters and Adaptive Windows for Learning in Data StreamsKalman Filters and Adaptive Windows for Learning in Data Streams
Kalman Filters and Adaptive Windows for Learning in Data StreamsAlbert Bifet
 

Viewers also liked (9)

Aakash riya org.
Aakash  riya org.Aakash  riya org.
Aakash riya org.
 
Resume
ResumeResume
Resume
 
Figure 1
Figure 1Figure 1
Figure 1
 
Synchronization of coupled oscillators is a game
Synchronization of coupled oscillators is a gameSynchronization of coupled oscillators is a game
Synchronization of coupled oscillators is a game
 
Figure 1
Figure 1Figure 1
Figure 1
 
Canini09a
Canini09aCanini09a
Canini09a
 
EDEM Brochure
EDEM BrochureEDEM Brochure
EDEM Brochure
 
Intro to Multitarget Tracking for CURVE
Intro to Multitarget Tracking for CURVEIntro to Multitarget Tracking for CURVE
Intro to Multitarget Tracking for CURVE
 
Kalman Filters and Adaptive Windows for Learning in Data Streams
Kalman Filters and Adaptive Windows for Learning in Data StreamsKalman Filters and Adaptive Windows for Learning in Data Streams
Kalman Filters and Adaptive Windows for Learning in Data Streams
 

Similar to Putting Objects in Perspective

TAO Fayan_X-Ray and MIP volume rendering
TAO Fayan_X-Ray and MIP volume renderingTAO Fayan_X-Ray and MIP volume rendering
TAO Fayan_X-Ray and MIP volume renderingFayan TAO
 
Journey to structure from motion
Journey to structure from motionJourney to structure from motion
Journey to structure from motionJa-Keoung Koo
 
CG OpenGL Shadows + Light + Texture -course 10
CG OpenGL Shadows + Light + Texture -course 10CG OpenGL Shadows + Light + Texture -course 10
CG OpenGL Shadows + Light + Texture -course 10fungfung Chen
 
Camera parameters
Camera parametersCamera parameters
Camera parametersTheYacine
 
3-D Visual Reconstruction: A System Perspective
3-D Visual Reconstruction: A System Perspective3-D Visual Reconstruction: A System Perspective
3-D Visual Reconstruction: A System PerspectiveGuillermo Medina Zegarra
 
generalformulationofFiniteelementofmodel
generalformulationofFiniteelementofmodelgeneralformulationofFiniteelementofmodel
generalformulationofFiniteelementofmodelPiyushDhuri1
 
generalformulation.ppt
generalformulation.pptgeneralformulation.ppt
generalformulation.pptRajuRaju183149
 
The geometry of perspective projection
The geometry of perspective projectionThe geometry of perspective projection
The geometry of perspective projectionTarun Gehlot
 
CG OpenGL surface detection+illumination+rendering models-course 9
CG OpenGL surface detection+illumination+rendering models-course 9CG OpenGL surface detection+illumination+rendering models-course 9
CG OpenGL surface detection+illumination+rendering models-course 9fungfung Chen
 
Engg. mathematics iii
Engg. mathematics iiiEngg. mathematics iii
Engg. mathematics iiimanoj302009
 
Image Processing 1
Image Processing 1Image Processing 1
Image Processing 1jainatin
 
Comparing estimation algorithms for block clustering models
Comparing estimation algorithms for block clustering modelsComparing estimation algorithms for block clustering models
Comparing estimation algorithms for block clustering modelsBigMC
 
DETECTION OF MOVING OBJECT
DETECTION OF MOVING OBJECTDETECTION OF MOVING OBJECT
DETECTION OF MOVING OBJECTAM Publications
 
Trident International Graphics Workshop 2014 2/5
Trident International Graphics Workshop 2014 2/5Trident International Graphics Workshop 2014 2/5
Trident International Graphics Workshop 2014 2/5Takao Wada
 
Lecture 7 (Digital Image Processing)
Lecture 7 (Digital Image Processing)Lecture 7 (Digital Image Processing)
Lecture 7 (Digital Image Processing)VARUN KUMAR
 

Similar to Putting Objects in Perspective (20)

TAO Fayan_X-Ray and MIP volume rendering
TAO Fayan_X-Ray and MIP volume renderingTAO Fayan_X-Ray and MIP volume rendering
TAO Fayan_X-Ray and MIP volume rendering
 
Cg
CgCg
Cg
 
Journey to structure from motion
Journey to structure from motionJourney to structure from motion
Journey to structure from motion
 
CG OpenGL Shadows + Light + Texture -course 10
CG OpenGL Shadows + Light + Texture -course 10CG OpenGL Shadows + Light + Texture -course 10
CG OpenGL Shadows + Light + Texture -course 10
 
Camera parameters
Camera parametersCamera parameters
Camera parameters
 
3-D Visual Reconstruction: A System Perspective
3-D Visual Reconstruction: A System Perspective3-D Visual Reconstruction: A System Perspective
3-D Visual Reconstruction: A System Perspective
 
generalformulationofFiniteelementofmodel
generalformulationofFiniteelementofmodelgeneralformulationofFiniteelementofmodel
generalformulationofFiniteelementofmodel
 
generalformulation.ppt
generalformulation.pptgeneralformulation.ppt
generalformulation.ppt
 
Chapter 4
Chapter 4Chapter 4
Chapter 4
 
The geometry of perspective projection
The geometry of perspective projectionThe geometry of perspective projection
The geometry of perspective projection
 
Projectile motion
Projectile motionProjectile motion
Projectile motion
 
CG OpenGL surface detection+illumination+rendering models-course 9
CG OpenGL surface detection+illumination+rendering models-course 9CG OpenGL surface detection+illumination+rendering models-course 9
CG OpenGL surface detection+illumination+rendering models-course 9
 
Unit1
Unit1Unit1
Unit1
 
Engg. mathematics iii
Engg. mathematics iiiEngg. mathematics iii
Engg. mathematics iii
 
Image Processing 1
Image Processing 1Image Processing 1
Image Processing 1
 
Comparing estimation algorithms for block clustering models
Comparing estimation algorithms for block clustering modelsComparing estimation algorithms for block clustering models
Comparing estimation algorithms for block clustering models
 
DETECTION OF MOVING OBJECT
DETECTION OF MOVING OBJECTDETECTION OF MOVING OBJECT
DETECTION OF MOVING OBJECT
 
Trident International Graphics Workshop 2014 2/5
Trident International Graphics Workshop 2014 2/5Trident International Graphics Workshop 2014 2/5
Trident International Graphics Workshop 2014 2/5
 
Lecture 7 (Digital Image Processing)
Lecture 7 (Digital Image Processing)Lecture 7 (Digital Image Processing)
Lecture 7 (Digital Image Processing)
 
6 dof.pdf
6 dof.pdf6 dof.pdf
6 dof.pdf
 

Recently uploaded

Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management SystemChristalin Nelson
 
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSGRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSJoshuaGantuangco2
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Mark Reed
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatYousafMalik24
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxAnupkumar Sharma
 
AUDIENCE THEORY -CULTIVATION THEORY - GERBNER.pptx
AUDIENCE THEORY -CULTIVATION THEORY -  GERBNER.pptxAUDIENCE THEORY -CULTIVATION THEORY -  GERBNER.pptx
AUDIENCE THEORY -CULTIVATION THEORY - GERBNER.pptxiammrhaywood
 
Food processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsFood processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsManeerUddin
 
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptx
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptxMusic 9 - 4th quarter - Vocal Music of the Romantic Period.pptx
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptxleah joy valeriano
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Celine George
 
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...JojoEDelaCruz
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfPatidar M
 
ICS2208 Lecture6 Notes for SL spaces.pdf
ICS2208 Lecture6 Notes for SL spaces.pdfICS2208 Lecture6 Notes for SL spaces.pdf
ICS2208 Lecture6 Notes for SL spaces.pdfVanessa Camilleri
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYKayeClaireEstoconing
 
ROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxVanesaIglesias10
 
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...Nguyen Thanh Tu Collection
 
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdf
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdfVirtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdf
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdfErwinPantujan2
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfJemuel Francisco
 
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxlancelewisportillo
 

Recently uploaded (20)

Transaction Management in Database Management System
Transaction Management in Database Management SystemTransaction Management in Database Management System
Transaction Management in Database Management System
 
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTSGRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
GRADE 4 - SUMMATIVE TEST QUARTER 4 ALL SUBJECTS
 
Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)Influencing policy (training slides from Fast Track Impact)
Influencing policy (training slides from Fast Track Impact)
 
Earth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice greatEarth Day Presentation wow hello nice great
Earth Day Presentation wow hello nice great
 
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptxMULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
MULTIDISCIPLINRY NATURE OF THE ENVIRONMENTAL STUDIES.pptx
 
AUDIENCE THEORY -CULTIVATION THEORY - GERBNER.pptx
AUDIENCE THEORY -CULTIVATION THEORY -  GERBNER.pptxAUDIENCE THEORY -CULTIVATION THEORY -  GERBNER.pptx
AUDIENCE THEORY -CULTIVATION THEORY - GERBNER.pptx
 
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptxLEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
LEFT_ON_C'N_ PRELIMS_EL_DORADO_2024.pptx
 
Food processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture honsFood processing presentation for bsc agriculture hons
Food processing presentation for bsc agriculture hons
 
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptx
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptxMusic 9 - 4th quarter - Vocal Music of the Romantic Period.pptx
Music 9 - 4th quarter - Vocal Music of the Romantic Period.pptx
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17
 
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...
ENG 5 Q4 WEEk 1 DAY 1 Restate sentences heard in one’s own words. Use appropr...
 
Active Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdfActive Learning Strategies (in short ALS).pdf
Active Learning Strategies (in short ALS).pdf
 
ICS2208 Lecture6 Notes for SL spaces.pdf
ICS2208 Lecture6 Notes for SL spaces.pdfICS2208 Lecture6 Notes for SL spaces.pdf
ICS2208 Lecture6 Notes for SL spaces.pdf
 
Raw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptxRaw materials used in Herbal Cosmetics.pptx
Raw materials used in Herbal Cosmetics.pptx
 
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITYISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
ISYU TUNGKOL SA SEKSWLADIDA (ISSUE ABOUT SEXUALITY
 
ROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptxROLES IN A STAGE PRODUCTION in arts.pptx
ROLES IN A STAGE PRODUCTION in arts.pptx
 
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...
HỌC TỐT TIẾNG ANH 11 THEO CHƯƠNG TRÌNH GLOBAL SUCCESS ĐÁP ÁN CHI TIẾT - CẢ NĂ...
 
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdf
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdfVirtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdf
Virtual-Orientation-on-the-Administration-of-NATG12-NATG6-and-ELLNA.pdf
 
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdfGrade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
Grade 9 Quarter 4 Dll Grade 9 Quarter 4 DLL.pdf
 
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptxQ4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
Q4-PPT-Music9_Lesson-1-Romantic-Opera.pptx
 

Putting Objects in Perspective

  • 1. Putting Objects in Perspective Derek Hoiem, Alexei A. Efros, Martial Hebert presented by Phongsathorn Eakamongul Department of Computer Science Asian Institute of Technology 2009, July 2 Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 1 / 45
  • 2. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 2 / 45
  • 3. Abstract Statistic Framework for placing local object detection by modeling interdependence of camera viewpoint (position/orientation) objects identities surface orientations (3D scene geometry) using single image. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 3 / 45
  • 4. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 4 / 45
  • 5. General Object Detection cannot be solved locally We cannot recognize car, person, and the road in isolation Pixel of car can be easily be interpreted as person’s shoulder, mouse, stack of books, balcony, etc. Psychophysical Evidence context plays crucial role in scene understanding (i.e. Biederman 1981; Torralba 2005). (1) car is sitting on the road (2) right size relative to other object in the scene. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 5 / 45
  • 6. Related Work Heuristic Image understanding systems ACRONYM (Brooks 1979), VISIONS (Hanson and Riseman 1978), outdoor scene understanding system (Ohta and Kanade 1985), intrinsic images (Barrow and Tenenbaum 1978), etc. Small image windows (at all locations and scales) to find specific objects works wonderfully with face detection (Schneiderman 2004; Viola and Jones 2004) – since inside the face is much more important than boundary. unreliable for cars, pedestrians, especially at smaller scale. Modeling within 2D image modeling direct relationships between objects and other objects (Kumar and Herbert 2003; Murphy et al. 2003), regions (He et al. 2004; Kumar and Hebert 2005; Tu et al. 2005), scene geometries (Murphy et al. 2003; Sudderth et al. 2005) Modeling 3D image estimate rough 3D scene geometry from single image (Hoiem et al. 2005), Viewpoint and mean scene depth (Torralba and Sinha 2001), geometric consistency of object hypotheses in simple scenes using hard algebric constriants (Forsyth et al. 1994) Relationship between camera parameters and objects well calibrated camera (Jeong et al. 2001, etc.), stationary surveillance camera (Krahnstoever and Mendonca 2005), or both (Greienhagen et al. 2000) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 6 / 45
  • 7. Overview 100 boxes sampled (b) local object detector (Murphy et al. 2003; Dalal and Triggs 2005) at any location/scale (c) estimate rough surface geometry (Hoiem et al. 2005) red, green, blue channels indicate vertical, ground, sky (e) estimate camera viewpoint from scale of objects in image This is like typical human eye-tracking, where subjects are asked to search for a person in an image. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 7 / 45
  • 8. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 8 / 45
  • 9. Projective geometry object’s height in image can be determined from it’s height in the world v0 : horizon position ui , vi : lower-left corordinate yc : camera height hi : object image height – – yi : object height vc : center of image vt , vb : top and bottom of object Assume all objects of interest rest on the ground plane. ( cannot find people on the roof ) objects perpendicular to the ground We can estimate object world height hi yc yi ≈ v0 −vi Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 9 / 45
  • 10. Camera Tilt camera tilt −v θx = 2arctan( vc 2f 0 ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 10 / 45
  • 11. Proof Assume camera roll = 0 or image is rectified camera intrinsic parameters are typical ( skew = 0, unit aspect ratio, typical f ) 2 3 2 3 2 32 3 x u f 0 uc 1 0 0 0 6 y 7 4 v 5 = 1 4 0 f vc 5 4 0 cosθx −sinθx yc 5 4 6 7 z z 5 1 0 0 1 0 sinθx cosθx 0 1 From this, z(fsinθx − (vc − v )cosθx − fyc ) y = (1) (vc − v )sinθx + fcosθx Since y=0 at vb , fyc z = (2) fsinθx − (vc − vb )cosθx put (2) in (1) fyc (fsinθx )−(vc −vt )cosθx /fsinθx −(vc −vb )cosθx −fyc y = (vc −vt )sinθx +fcosθx Consider condition that camera is on the ground, vc −v0 Since θx is small; cosθx ≈ 1, sinθx ≈ θx , θx ≈ f v −vb y ≈ yc t /(1 + (vc − v0 )(vc − vt )/f 2 ) v0 −vb Since tilt tend to be small; vc − v0 ≈ 0, and f > 1 v −vb y ≈ yc t v0 −vb hi yc so yi ≈ v0 −vi Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 11 / 45
  • 12. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 12 / 45
  • 13. Probabilistic model to describe a scene Assume θ, o, g are conditionally independent. From model P(θ, o, g, eg , eo ) = Y P(θ) P(oi |θ)P(eoi |oi )P(gi |oi )P(egi |gi ) i The likelihood of the scene conditioned on image evidence using Bayes rule P(θ, o, g|eg , eo ) = Y P(oi |eoi ) P(gi |egi ) P(θ) P(oi |θ) P(oi ) P(gi ) θ : viewpoint i o : object identity g : surface geometry Classifier can be incorporated using eo : object evidence probabilistic output P(oi |eoi ) eg : geometry evidence Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 13 / 45
  • 14. Inference Pearl’s belief propagation algorithm (Pearl 1988) is optimal and efficient. inference by quantizing continuous variable v0 and yc into evenly-spaced bins ( 50 and 100 bins, respectively ) Their implementation make use of Bayes Net Toolbox (Murphy 2001) After inference, they can query, such as "what is expected height of this object ?" "what are marginal probability of cars ?" "What is the most probable explanation of the scene ?" They report results based on marginal probability from sum-product algorithm ( this allows ROC curve to be computed ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 14 / 45
  • 15. Camera Viewpoint (θ) Consider v0 , yc to be independent a priori. P(θ) = P(v0 )P(yc ) v0 distribution is estimated using Simple Gaussian prior yc distribution is estimated using kernel density estimation . A priori, Most likely yc = 1.67, at eye level of typical adult male Most likely v0 = 0.50 Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 15 / 45
  • 16. Camera Viewpoint (e) Viewpoint prior θ (left axis - yc , right axis - v0 ) – high variance, more informative than uniform distribution that implicitly assumed when scale is considered irrelevant. (g, h) local object detection (b, c, d) Geometry estimates gi – Using this model, they improve estimates of (f) Viewpoint : full Viewpoint peak likelihood increases from 0.0037 a priori to 0.0503 after inference. (i, j) local detection improve when viewpoint and surface geometry are considered. At the same false positive ( cars: cyan, peds: yellow ) rate, the true detection (cars: green, peds: red) rate doubles. Note : detections with lower confidence exist for both local and full model but not shown here. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 16 / 45
  • 17. Object Identities (oi ) oi consists of ti : type ∈ {object, background} (i.e. pedestrian) : identity bboxi : boundingbox = {ui , vi , wi , hi } wi : width hi : height Prob of object occurring at a particular location/scale P(Ii |ti = obj, bboxi )P(oi ) P(ti = obj, bboxi |Ii ) = P(Ii |ti = obj, bboxi )P(oi ) + P(Ii |ti = obj, bboxi )(1 − P(oi )) 1 = P(oi ) 1 + exp[−ci − log 1−P(o ) ] i P(oi ) : prior Ii : image information at ith bounding box class-conditional log-likelihood ratio : ci = log P(Ii |ti =obj,bboxi ) P(I |t =obj,bbox ) i i i Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 17 / 45
  • 18. Object Identities (oi ) Typically, researcher perform non-maxima suppression , assuming that high detection responses at neighboring positions could be due to an object at either of those positions (but not both). Applied non-maxima suppression + form a point distribution out of the non-maxima ( rather than discarding them ) An object candidate is formed out of a group of closely overlapping bounding boxes. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 18 / 45
  • 19. Object Identities (oi ) The candidate likelihood P(ti = obj|eo ) = the likelihood of the highest confidence bounding box. P(bboxi |ti = obj, eo ) ∝ P(ti = obj, bbox|I) Object’s height depends on its position when given viewpoint. P(oi |θ) ∝ p(hi |ti , vi , θ) ( due to the uniformity of P(ti , vi , wi |θ) ) From yi ≈ vhi−vi yc 0 if yi is normal distribution, with parameters {µi , σi } hi is also normal distribution with parameters { µi (vyc−vi ) , 0 σi (v0 −vi) yc } Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 19 / 45
  • 20. Surface Geometry (gi ) (Hoiem et al 2005) produces "confidence maps" of geometric surface for 3 classes sky vertical, contains 5 subclasses planar, facing left (←) center (↑) right (→) non-planar solid (O) porous (X) Figure: Geometric Label ground Define gi = three values, corresponding to whether "object surface" is visible in detection window whether the ground is visible just below the detection window i.e. consider "car" as "planar or non-planar solid" consider "pedestrian" as "non-planar solid" Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 20 / 45
  • 21. Surface Geometry (gi ) Compute P(gi |oi ) and P(gi ) by counting occurrences of value of gi for both people and cars in training set If oi is background, consider P(gi |oi ) ≈ P(gi ) Estimate P(gi |egi ) based on confidence maps. They found that "average geometric confidence in a window" is a "well-calibrated probability for the geometric value". Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 21 / 45
  • 22. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 22 / 45
  • 23. Training Viewpoint ( θ ) yc distribution is estimated using kernel density estimation . Manually labeled cars ( i.e. vans, trucks ), pedestrians in 60 outdoor images from LabelMe dataset (Russell et al. 2005). Calculate maximum likelihood estimate of camera height (yc ) based on labeled horizon and height distributions of cars and people in world. yc is estimated using "kernel density estimation" ( ksdensity in Matlab ). Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 23 / 45
  • 24. Training Object Identities (oi ) use local detector (Murphy et al. 2003) same local patch template features + 6 color features ( encode average L*a*b*color of detection window, and difference between detection window and surrounding area. ) verified using PASCAL challange training/validation set : did not search for car shorter than 10% of image height ( the official entries could not detect small cars ). Result : average precision of 0.423 ( the best scores reported by top 3 groups: 0.613, 0.489, 0.353 ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 24 / 45
  • 25. Training Object Identities (oi ) logistic regression version of Adaboost classifier (Collins et al 2002) to boost 8-node decision tree classifiers For car, they trained 2 views (front/back: 32x24 px, side: 40x16 px) For pedestrains, they trained 1 view (16x40 px) using full PASCAL dataset (PASCAL 2005) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 25 / 45
  • 26. Training Object Identities (oi ) Estimate distribution of cars’ height ( yi ) : use data from Consumer Reports ( www.consumerreports.org ) Result : µ = 1.59 m, σ = 0.21 m pedestrains’ height ( yi ) : use data from National Center for Health Statistics ( www.cdc.edu/nchs ) Result : µ = 1.7 m, σ = 0.085 m Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 26 / 45
  • 27. Training Surface Geometry ( gi ) Compute P(gi |oi ) by counting occurrences of value of gi for both people and cars in 60 training set from LabelMe Set P(gi ) to be uniform , because the learned values of P(gi ) over-relying on geometry. The over-reliance may be due to labeled image (general outdoor) being drawn from a different distribution than test set (streets of Boston) lack of modeled direct dependence between surface geometries. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 27 / 45
  • 28. Evaluation Test Set : 422 random outdoor images from LabelMe dataset (Russell et al. 2005) no car or pedestrians : 60 images only pedestrians : 40 images only cars : 94 images both cars and pedestrians : 225 images Total 923 cars, 720 pedestrians Detect cars’ height as small as 14 px Detect Pedestrians’ height as small as 36 px To get detection confidences for each window, they reverse the process in ( in Object identities slide ). Then determine bounding boxes of objects in standard way (thresholding the confidences and performing non-maxima suppression) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 28 / 45
  • 29. Viewpoint and Surface geometry aids Object detection Figure: Object Detection Result : ROC curves Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 29 / 45
  • 30. About 20% of cars and pedestrians missed by the baseline are detected for the same FP Figure: FP : false positive rates, and % reductions in false negatives Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 30 / 45
  • 31. Object and geometry improve horizon estimation Manually labeled horizon of 100 images that contained Car+Ped. Figure: Horizon estimation results Mean/median absolute error (as percentage of image height) are shown for horizon estimates. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 31 / 45
  • 32. Horizon estimation, and object detection are more accurate when "more object models" of the scene are known. detect only car, only pedestrians, and both Figure: (Horizon) : median absolute error, (First Number) : FP/image ( at 50% detection rate ) computed over all images, (Second Number) : FP/image ( at 50% detection rate ) computed over subset of images that contain both cars and people Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 32 / 45
  • 33. Any local object detector can be plugging into their framework After converting SVM outputs to probabilities using method of (Platt 2000). Training 80x24 car detector; 32x96 big pedestrian detector, 16x48 small pedestrian detector Figure: Detal-Triggs local detector (Detal and Triggs 2005) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 33 / 45
  • 34. Detection Example baseline local detector (Murphy et al. 2003), and full model detector blue line = horizon estimate ( 0.5 initially ) green box = true car, cyan box = false car red box = true ped, yellow = false ped solid line – high confidence detection ( 0.5 FP/Image ) dotted line – low confidence detection ( 2 FP/Image ) In most case, horizon line is correctly recovered, obj detection improves considerably. Box that make no sense (i.e. wrong scale (d), above horizon (b), in the middle of the ground (e)) is removed. Predestrians are often hallucinated (c, e) in places where they could be (but are not). (f) : a bad geometry estimate and repeated false detections along the building windows causes the horizon estimate to become worse and the car to be missed./ 45 Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 34
  • 35. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 35 / 45
  • 36. improve estimating v0 using detections of known objects in scene If image does not include easily detectable known object ? "Edges and perspective geometry" using a lots of parallel lines (Manhattan Worlds) (Kosecka and Zhang 2022), (Coughlan and Yuille 2003) fail for less structured environments. using "spatial-frequency" statistics ( the "gist" ) of image to recover a general sense of space of the scene. (Oliva and Torralba 2006) They estimate rough "scene envelop" of properties such as "openess, smoothness, depth". "gist" statistics could also provide a rough estimate of "horizon position" (v0 ). (Torralba and Sinha 2001) Use gist descriptor to estimate v0 instead of Simple Gaussian Prior To obtain more precise estimate, we require 1 large amount of training set ( image/viewpoint pairs ) 2 a way of accurately labeling this data, automatically Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 36 / 45
  • 37. Automatically labeling data to get v0 and new distribution of yi using iterative EM-like algorithm (Lalongde te al. 2007) initially guess of µ and σ of people’s height (yi ) iterates 1 estimate "viewpoint" ( v0 ) for images that contain objects of known "size distributions" ( yi ) 2 estimate "size distributions" ( yi ) of objects that are contained in images with known "viewpoints" ( v0 ). Figure: automatic object height estimation result, (left) picture from LabelMe (center) original pixel size (right) automatically estimated 3D height ( yi ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 37 / 45
  • 38. Automatically labeling data to get v0 and new distribution of yi Training : infer viewpoint of over 5,000 images and height of 13,000 object instances in 50 object classes. based on inferred object heights (yi ), re-estimate yi distribution of cars ( µ = 1.51m, σ = 0.191m ), people ( µ = 1.70m, σ = 0.103m ) consider camera viewpoint vi estimates to be reliable for 2,660 images (excluding images in test set) that contains at least 2 objects with known height distributions. calculate gist statistics (8x8 blocks at 3 scales with 8,4,4 orientations, giving 1280 variables per gist vector). Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 38 / 45
  • 39. gist descriptor Evaluate gist nearest neighbour classifiers with various distance matrics nearest neighbor regression method (Navot et al. 2006) GRANN (Generalized Regression neural networks) – provide lowest error after tuning α = 0.22 on training set ˜ The horizon estimation using GRANN (v0 ) P 2 i v w v0 (x) = P 0i i ; wi = exp[ −||x−xi || ] ˜ 2α2 i wi x : gist statistics v0i : horizontal position for ith training sample −˜ p(v0 |v0 ) = 2s exp[ −|v0v v0 | ] ( laplace probability density ) ˜ 1 s 0 sv : scale parameter = expected error Result : correlation (coefficient 0.27) between error and Euclidean distance to the nearest neighbour in gist statistics. ˜ Provide better estimate, by consider d ( nearest neighbour distance ) ˜ fit sv = 0.022 + 0.060d, by maximum likelihood estimation over training set Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 39 / 45
  • 40. Evaluation ˜ Figure: horizontal estimation using gist nearest neighbour and distance d Figure: mean error of horizon position estimation ( % of image height ) Gaussian Prior Gaussian Prior + surface geometry, object ( using Datal-Triggs detectors ) Gist estimation Gist estimation + surface geometry, object ( using Datal-Triggs detectors ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 40 / 45
  • 41. Evaluation improve detection rates ( This is measured at 1 FP/image using Dalal-Triggs object detectors ) cars detection 50 % to 52 % pedestrians 66 % to 68 % Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 41 / 45
  • 42. Improve estimating yc cannot improve significantly over prior estimate may because small change in yc has little impact on the global statistics. Therefore, re-estimate yc using larger training set Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 42 / 45
  • 43. Outline 1 Abstract 2 Introduction 3 Projective geometry 4 Modeling the scene 5 Training + Evaluation 6 Viewpoint by Example 7 Summary Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 43 / 45
  • 44. Summary Project estimated ground surface into overhead view, using estimated camera viewpoint. Assume cars mostly face down the same line. What they have done : "bag of features/blackbox" classification method 1 subtle relationship (i.e. object sizes relate through the viewpoint) can be easily represented. 2 additions and extensions to the model are easy. could be extended by modeling other scene properties ( i.e. scene category ) Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 44 / 45
  • 45. Summary geometry assumptions not consider as conditionally independent, if two object patches are very similar. i.e. repeated windows in (f) causes object detection responses to be correlated. solution : correlation should be modeled. i.e. if two object patches are very similar, consider their evident jointly. the surface trained in this experiment is outdorr image only. Phongsathorn (AIT) Putting Objects in Perspective Short Occasion 45 / 45