SlideShare a Scribd company logo
1 of 79
Combining Motion Information with Appearance Information
and Utilizing of Mutual Information Encoded
Among Image Features of the Same Object
for Better Detection Performance
Zhipeng Wang
Directed by: Pro. Katsushi Ikeuchi, Dr. Masataka Kegesawa, and Dr. Shintaro Ono
Computer Vision Laboratory, The University of Tokyo
運動情報を外観情報との結びつけおよび
同一物体の画像特徴によりエンコードされた相互情報の利用
ーー検出性能の向上のため
Outline
 Background and state-of-the-art detection
methods
 Our work (goal, challenges, related work, method,
results, contribution, and conclusion)
 Section 1, 2: utilizing motion information
 Section 3, 4: utilizing mutual information
 Conclusion
 Publication list
Background
 Object detection from images:
 One of the basic perceptual skills in human
 Plays an important role in machine vision area
Two primary categories
 Sliding-window methods
 Use classifiers to answer whether each sub-
image contains a target object
 Hough transform based methods
 Infer about object status in a bottom-up manner
from detected image features
Better detection performance
 Representative methods
 Image features: HOG, sift
 Classifiers
 Machine learning methods: deep learning
 Efficient search techniques
 Branch and bound[1]
[1] Lampert, C.H.; Blaschko, M.B.; Hofmann, T.; , "Efficient Subwindow Search: A
Branch and Bound Framework for Object Localization," Pattern Analysis and
Machine Intelligence, IEEE Transactions on , vol.31, no.12, pp.2129-2142, Dec.
2009
Our work towards
better detection performance
 Combining motion information with
appearance information
 Using mutual information encoded among the
features of the same object
 By using extra information, we expect to
achieve better detection performance
Our work
 Combining motion information with
appearance information
 1. Common Fate Hough Transform
 2. Emergency Telephone Indicator Detection in
Tunnel Environment
1. Common Fate Hough
Transform
Goal and challenges
 Where and what
label is each object
in the scene
 Challenges:
 Near objects
 Similar different-
class objects
Common Fate Hough Transform
Related work
 Motion for object detection
 Background subtraction
 Optical flow[1]
 Combination of appearance and motion
information
 Mainly for tracking
[1] G. Brostow and R. Cipolla. Unsupervised bayesian detection of
independent motion in crowds. In CVPR, pages I: 594–601, 2006.
Hough transform
 On the current image:
 Detect keypoints and extract image feature
 Find the best matched codes from a trained
codebook
 Each keypoint votes for
object center and label
B. Leibe and B. Schiele. Interleaved object
categorization and segmentation.
In BMVC, pages 759–768, 2003.
Hough transform
 Discrete to continuous
 Blurring each vote
 Summing up all votes from
all object parts
Common Fate Hough Transform
The common fate principle
 One of the gestalt laws
 Elements with the same motion tend to be
perceived as one unit
Common Fate Hough Transform
Common fate Hough transform
 Motion analysis[1]
 Keypoint tracking (KLT
tracker)
 Clustering the
trajectories
[1] G. Brostow and R. Cipolla. Unsupervised bayesian detection of
independent motion in crowds. In CVPR, pages I: 594–601, 2006.
Common fate Hough transform
 The weight of each vote is related to the
support it gains from the motion group
Common fate Hough
Transform
Common fate Hough transform
Common Fate Hough
Transform
 Inference
 1. Find the highest peak in the confidence space
 2. Exclude the object parts belonging to the
highest peak
 3. If the Hough image is not empty, go to 1.
Inference
Common Fate Hough Transform
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Common Fate Hough Transform
Inference
Experimental results
 Dataset
 720×576
 401 continuous frames
 633 different-class ground truth bounding boxes
on 79 frames
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Results
Common Fate Hough Transform
Result comparison
[1] O. Barinova, V. Lempitsky, and P. Kohli. On detection of multiple object
instances using hough transforms. In CVPR, pages 2233–2240, 2010.
Contribution
 A detection method with better detection
performance than the CVPR’10 method
 A successful combination of motion and
appearance information for detection
 A successful attempt to combine human
perception rules into detection method in the
computer vision area
Common Fate Hough Transform
Conclusion
 Motion largely improves detection results by
being effectively combined with appearance
information
 The method is not efficient
  then we propose
Common Fate Hough Transform
2. Detection of Emergency
Telephone Indicators by Fusion
of Motion Information with
Appearance Information
Goal
Emergency Telephone Indicator Detection
 For vehicle positioning
 Detecting emergency telephone indicators
 Installed at known location
 Infrared cameras (far infrared) installed on vehicle’s
top
Challenges
 Challenges:
 Noisy objects
 Real-time requirement
Emergency Telephone Indicator Detection
Method
 Method: a two-step method
 Detect, verify, and cluster keypoints
 Verify keypoint clusters by appearance and
motion information
Emergency Telephone Indicator Detection
Keypoint Detection
Keypoint Verification
Keypoint Clustering
Keypoint Cluster Verification
by Appearance
Keypoint Cluster
Tracking
Keypoint Cluster Verification
by Motion
Pipeline
Pipeline
 Detect keypoints
 Averagely sampling
 Setting intensity thresholds
 160-190 (0-255)
Emergency Telephone Indicator
Pipeline
 Verify keypoints
 Intensity histogram
 Building a mixture model
using k-means
Emergency Telephone Indicator
Detection
Pipeline
 Cluster keypoints
 Build a minimum spanning tree (Euclidean
distance)
 Split the tree
Emergency Telephone Indicator
Detection
Pipeline
 Verify keypoint clusters by appearance
 Adaboost machine
 Trained from positive and negative examples
 Intensity histogram
Emergency Telephone Indicator
Detection
Pipeline
 Keypoint cluster tracking
 No occlusion
 Connect each detection
response to its nearest
trajectory (appearance,
scale, and time gap)
Emergency Telephone Indicator
Detection
Pipeline
 Keypoint cluster verification by motion
 Fit each trajectory as a straight line
 Use the significance of the fitting as criteria
Emergency Telephone Indicator
Detection
Benefits of the pipeline
 The system is real-time
 More time consuming steps dealing with
fewer instances
 Original image: 105 pixels
 Keypoint detection: 104 points
 Keypoint verification: 103 keypoints
 Keypoint clustering: 102 keypoints
  Less than 10 keypoint clusters
Emergency Telephone Indicator Detection
Experimental results
 Real-time (34 fps)
 A laptop with Intel Core2 Duo 2.8GHz
processors
Experimental results
Emergency Telephone Indicator Detection
Contribution
 A specialized real-time detection method on
cluttered data
 A successful combination of motion and
appearance information for detection
 The meaning of this research: a potential
applicable solution for vehicle positioning in
tunnel environment
Emergency Telephone Indicator Detection
Conclusion
 Motion information plays an important role in
the detecting process
Emergency Telephone Indicator Detection
Future work
 We collect new data using new camera and
plan to improve the pipeline on the new
collected data
Emergency Telephone Indicator
Detection
 Two methods utilizing motion information are
proposed
  When motion information is not available,
we then go to
Our work
 Utilizing of Mutual Information Encoded
Among Image Features of the Same Object
 3. Pyramid Match Score for Object Detection
 4. Detection of Object With In-plane Rotations
3. Pyramid Match Score for
Object Detection
Motivation
 Sliding window based methods:
 Often [1] ignore positional information of each
image feature
 Methods based on Hough transform
 Ignore overall information of each object
 Intention of this method:
 An efficient method to utilize positional information
of each image feature and overall information
gained from the whole object
[1] Lazebnik, S.; Schmid, C.; Ponce, J.; , "Beyond Bags of Features: Spatial Pyramid
Matching for Recognizing Natural Scene Categories," Computer Vision and Pattern
Recognition, 2006 IEEE Computer Society Conference on , vol.2, no., pp. 2169- 2178, 2006
Pyramid Match
 Pyramid Match is a method to find best one-
one match between two sets of points
 Here gives a 2D example
 Real data is 14D
 Appearance information:12D
 Position information: 2D
Pyramid match
 Divide the space from fine to coarse
 If two points are in the same grid under a division,
the are considered as match and excluded
Pyramid match
 The distance of two matched points
 No need to be calculated
 Assigned
 According to the size of the grid, when they are found
match
1/4 * 2 1/2 * 2 1*2
Pyramid Match Score
 Pyramid match between two sets
 Each sub-image: considered as a 14D point set
(appearance: 12, position: 2)
 A “super template” contains all 14D points of
training images
Pyramid Match Score
 Find the near best match between each sub-
image of a test image and the “super
template”
 The match score (distance) is used as the
confidence that the sub-image contains a
target object
Results on UIUC cars
Results
0
0.1
0.2
0.3
0.4
0.5
0.6
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Detectionrate
Precison
3
10
7
av
back
Comments
 The results are not good enough
 The method seems promising, and in future
we will improve the method
  This mutual information can be further
used
4 Detection of Object With
In-plane Rotations
About the related work
 Training examples of all rotation directions,
 not robust
 or very time-consuming
Related work
 Two neural networks to
deal with one sub-image
 One for estimation of the
sub-image’s rotation angle
 The sub-image is rotated so
that the potential face is
upright
 The other for decision of
whether one object exists
H. Rowley, S. Baluja, and T. Kanade. Rotation invariant neural
network-based face detection. In CVPR, pages 38–44, 1998.
Related work
 Local feature method
 Infer rotation based on the gradient direction of
sift feature
 Graph-based methods
 Consider object as graph, and change the graph
Detection of Object With In-plane Rotations
Idea
 Suppose we have a codebook trained from
objects without in-plane rotations
Detection of Object With In-plane Rotations
Idea
 With proof omitted
 For two or more votes from different
keypoints, there exists one and only one rotation
angle which will minimize the difference of
the voted centers
If all votes are good estimations, we can expect
to be a good estimation the rotation angle of the
object
Detection of Object With In-plane Rotations
To do
 Propose a method capable of detecting
objects with in-plane
 Evaluate the robustness of
Detection of Object With In-plane Rotations
Conclusion
 Motion information combined with
appearance information
 Distinguish target objects with noisy object
 Enhance detection rate
 Mutual information encoded among the
image features of the same object
 We expect to get a method capable of utilizing
positional and overall information
 We expect to get a method capable of detecting
objects with in-plane rotations
Publication list
 [1] Zhipeng Wang; Jinshi Cui; Hongbin Zha; Kegesawa, M.; Ikeuchi, K.; , "Object
detection by common fate Hough transform," Pattern Recognition (ACPR), 2011 First
Asian Conference on , vol., no., pp.613-617, 28-28 Nov. 2011
 [2] Zhipeng Wang; Kagesawa, M.; Ono, S.; Banno, A.; Ikeuchi, K.; , "Emergency light
detection in tunnel environment: An efficient method," Pattern Recognition
(ACPR), 2011 First Asian Conference on , vol., no., pp.628-632, 28-28 Nov. 2011
 [3] Zhipeng Wang; Jinshi Cui; Hongbin Zha; Masataka Kegesawa; Shintaro Ono;
Katsushi Ikeuchi;, “Detection by Motion-based Grouping of Object Parts,”
International Journal of ITS Research (submitted)
 [4] Zhipeng Wang; Masataka Kegesawa; Shintaro Ono; Atsuhiko Banno; Takeshi
Oishi; Katsushi Ikeuchi;, “Detection of Emergency Telephone Indicators Using
Infrared Cameras for Vehicle Positioning in Tunnel Environment,” ITS World
Congress 2013 (submitted)
Thank you very much!

More Related Content

What's hot

Moving object detection
Moving object detectionMoving object detection
Moving object detection
Manav Mittal
 

What's hot (20)

Object tracking
Object trackingObject tracking
Object tracking
 
Presentation of Visual Tracking
Presentation of Visual TrackingPresentation of Visual Tracking
Presentation of Visual Tracking
 
Multiple Person Tracking with Shadow Removal Using Adaptive Gaussian Mixture ...
Multiple Person Tracking with Shadow Removal Using Adaptive Gaussian Mixture ...Multiple Person Tracking with Shadow Removal Using Adaptive Gaussian Mixture ...
Multiple Person Tracking with Shadow Removal Using Adaptive Gaussian Mixture ...
 
Moving object detection in video surveillance
Moving object detection in video surveillanceMoving object detection in video surveillance
Moving object detection in video surveillance
 
IRJET - A Systematic Observation in Digital Image Forgery Detection using MATLAB
IRJET - A Systematic Observation in Digital Image Forgery Detection using MATLABIRJET - A Systematic Observation in Digital Image Forgery Detection using MATLAB
IRJET - A Systematic Observation in Digital Image Forgery Detection using MATLAB
 
Development and Comparison of Image Fusion Techniques for CT&MRI Images
Development and Comparison of Image Fusion Techniques for CT&MRI ImagesDevelopment and Comparison of Image Fusion Techniques for CT&MRI Images
Development and Comparison of Image Fusion Techniques for CT&MRI Images
 
IRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object DetectionIRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object Detection
 
Optimization of Macro Block Size for Adaptive Rood Pattern Search Block Match...
Optimization of Macro Block Size for Adaptive Rood Pattern Search Block Match...Optimization of Macro Block Size for Adaptive Rood Pattern Search Block Match...
Optimization of Macro Block Size for Adaptive Rood Pattern Search Block Match...
 
An efficient block matching algorithm for fast motion ESTIMATION USING COMBIN...
An efficient block matching algorithm for fast motion ESTIMATION USING COMBIN...An efficient block matching algorithm for fast motion ESTIMATION USING COMBIN...
An efficient block matching algorithm for fast motion ESTIMATION USING COMBIN...
 
Detection and Tracking of Moving Object: A Survey
Detection and Tracking of Moving Object: A SurveyDetection and Tracking of Moving Object: A Survey
Detection and Tracking of Moving Object: A Survey
 
Automatic identification of animal using visual and motion saliency
Automatic identification of animal using visual and motion saliencyAutomatic identification of animal using visual and motion saliency
Automatic identification of animal using visual and motion saliency
 
Fast Feature Pyramids for Object Detection
Fast Feature Pyramids for Object DetectionFast Feature Pyramids for Object Detection
Fast Feature Pyramids for Object Detection
 
Passive Image Forensic Method to Detect Resampling Forgery in Digital Images
Passive Image Forensic Method to Detect Resampling Forgery in Digital ImagesPassive Image Forensic Method to Detect Resampling Forgery in Digital Images
Passive Image Forensic Method to Detect Resampling Forgery in Digital Images
 
Moving object detection
Moving object detectionMoving object detection
Moving object detection
 
Visual Object Tracking: review
Visual Object Tracking: reviewVisual Object Tracking: review
Visual Object Tracking: review
 
TRACKING OF PARTIALLY OCCLUDED OBJECTS IN VIDEO SEQUENCES
TRACKING OF PARTIALLY OCCLUDED OBJECTS IN VIDEO SEQUENCESTRACKING OF PARTIALLY OCCLUDED OBJECTS IN VIDEO SEQUENCES
TRACKING OF PARTIALLY OCCLUDED OBJECTS IN VIDEO SEQUENCES
 
A Review on Matching For Sketch Technique
A Review on Matching For Sketch TechniqueA Review on Matching For Sketch Technique
A Review on Matching For Sketch Technique
 
Motion Human Detection & Tracking Based On Background Subtraction
Motion Human Detection & Tracking Based On Background SubtractionMotion Human Detection & Tracking Based On Background Subtraction
Motion Human Detection & Tracking Based On Background Subtraction
 
Object tracking a survey
Object tracking a surveyObject tracking a survey
Object tracking a survey
 
Human Detection and Tracking System for Automatic Video Surveillance
Human Detection and Tracking System for Automatic Video SurveillanceHuman Detection and Tracking System for Automatic Video Surveillance
Human Detection and Tracking System for Automatic Video Surveillance
 

Viewers also liked (8)

Drupal -organizaciya_razrabotki
Drupal  -organizaciya_razrabotkiDrupal  -organizaciya_razrabotki
Drupal -organizaciya_razrabotki
 
Just Sell IT Conference, Odessa, April 2014
Just Sell IT Conference, Odessa, April 2014Just Sell IT Conference, Odessa, April 2014
Just Sell IT Conference, Odessa, April 2014
 
презентация для фармацевтических и медицинских компаний
презентация для фармацевтических и медицинских компанийпрезентация для фармацевтических и медицинских компаний
презентация для фармацевтических и медицинских компаний
 
PR cases 2014 корпоративные праздники
PR cases 2014 корпоративные праздникиPR cases 2014 корпоративные праздники
PR cases 2014 корпоративные праздники
 
Velux dealers conferens
Velux dealers conferensVelux dealers conferens
Velux dealers conferens
 
Выездные мероприятия нового формата.
Выездные мероприятия нового формата.Выездные мероприятия нового формата.
Выездные мероприятия нового формата.
 
Portfolio. PR-manager.
Portfolio. PR-manager. Portfolio. PR-manager.
Portfolio. PR-manager.
 
Slide Sher Ediv Style
Slide Sher Ediv StyleSlide Sher Ediv Style
Slide Sher Ediv Style
 

Similar to Wang midterm-defence

Integrated Hidden Markov Model and Kalman Filter for Online Object Tracking
Integrated Hidden Markov Model and Kalman Filter for Online Object TrackingIntegrated Hidden Markov Model and Kalman Filter for Online Object Tracking
Integrated Hidden Markov Model and Kalman Filter for Online Object Tracking
ijsrd.com
 
20110220 computer vision_eruhimov_lecture01
20110220 computer vision_eruhimov_lecture0120110220 computer vision_eruhimov_lecture01
20110220 computer vision_eruhimov_lecture01
Computer Science Club
 
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorProposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
QUESTJOURNAL
 

Similar to Wang midterm-defence (20)

D018112429
D018112429D018112429
D018112429
 
Integrated Hidden Markov Model and Kalman Filter for Online Object Tracking
Integrated Hidden Markov Model and Kalman Filter for Online Object TrackingIntegrated Hidden Markov Model and Kalman Filter for Online Object Tracking
Integrated Hidden Markov Model and Kalman Filter for Online Object Tracking
 
When Remote Sensing Meets Artificial Intelligence
When Remote Sensing Meets Artificial IntelligenceWhen Remote Sensing Meets Artificial Intelligence
When Remote Sensing Meets Artificial Intelligence
 
Moving Object Detection And Tracking Using CNN
Moving Object Detection And Tracking Using CNNMoving Object Detection And Tracking Using CNN
Moving Object Detection And Tracking Using CNN
 
Survey on video object detection & tracking
Survey on video object detection & trackingSurvey on video object detection & tracking
Survey on video object detection & tracking
 
最近の研究情勢についていくために - Deep Learningを中心に -
最近の研究情勢についていくために - Deep Learningを中心に - 最近の研究情勢についていくために - Deep Learningを中心に -
最近の研究情勢についていくために - Deep Learningを中心に -
 
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
 
Design and implementation of video tracking system based on camera field of view
Design and implementation of video tracking system based on camera field of viewDesign and implementation of video tracking system based on camera field of view
Design and implementation of video tracking system based on camera field of view
 
A Survey on Approaches for Object Tracking
A Survey on Approaches for Object TrackingA Survey on Approaches for Object Tracking
A Survey on Approaches for Object Tracking
 
20110220 computer vision_eruhimov_lecture01
20110220 computer vision_eruhimov_lecture0120110220 computer vision_eruhimov_lecture01
20110220 computer vision_eruhimov_lecture01
 
IRJET- Application of MCNN in Object Detection
IRJET-  	  Application of MCNN in Object DetectionIRJET-  	  Application of MCNN in Object Detection
IRJET- Application of MCNN in Object Detection
 
Gesture Recognition using Principle Component Analysis & Viola-Jones Algorithm
Gesture Recognition using Principle Component Analysis &  Viola-Jones AlgorithmGesture Recognition using Principle Component Analysis &  Viola-Jones Algorithm
Gesture Recognition using Principle Component Analysis & Viola-Jones Algorithm
 
Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022Top Cited Articles in Signal & Image Processing 2021-2022
Top Cited Articles in Signal & Image Processing 2021-2022
 
G04743943
G04743943G04743943
G04743943
 
Human Re-identification with Global and Local Siamese Convolution Neural Network
Human Re-identification with Global and Local Siamese Convolution Neural NetworkHuman Re-identification with Global and Local Siamese Convolution Neural Network
Human Re-identification with Global and Local Siamese Convolution Neural Network
 
OBJECT DETECTION AND RECOGNITION: A SURVEY
OBJECT DETECTION AND RECOGNITION: A SURVEYOBJECT DETECTION AND RECOGNITION: A SURVEY
OBJECT DETECTION AND RECOGNITION: A SURVEY
 
Ear Biometrics shritosh kumar
Ear Biometrics shritosh kumarEar Biometrics shritosh kumar
Ear Biometrics shritosh kumar
 
Object Detection and Tracking AI Robot
Object Detection and Tracking AI RobotObject Detection and Tracking AI Robot
Object Detection and Tracking AI Robot
 
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorProposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
 
International Journal of Engineering Research and Development
International Journal of Engineering Research and DevelopmentInternational Journal of Engineering Research and Development
International Journal of Engineering Research and Development
 

Recently uploaded

An Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdfAn Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdf
SanaAli374401
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
ciinovamais
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
Chris Hunter
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
kauryashika82
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
heathfieldcps1
 

Recently uploaded (20)

PROCESS RECORDING FORMAT.docx
PROCESS      RECORDING        FORMAT.docxPROCESS      RECORDING        FORMAT.docx
PROCESS RECORDING FORMAT.docx
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
An Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdfAn Overview of Mutual Funds Bcom Project.pdf
An Overview of Mutual Funds Bcom Project.pdf
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
Introduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The BasicsIntroduction to Nonprofit Accounting: The Basics
Introduction to Nonprofit Accounting: The Basics
 
How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17How to Give a Domain for a Field in Odoo 17
How to Give a Domain for a Field in Odoo 17
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
Advance Mobile Application Development class 07
Advance Mobile Application Development class 07Advance Mobile Application Development class 07
Advance Mobile Application Development class 07
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
Holdier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdfHoldier Curriculum Vitae (April 2024).pdf
Holdier Curriculum Vitae (April 2024).pdf
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
Explore beautiful and ugly buildings. Mathematics helps us create beautiful d...
 
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
Ecological Succession. ( ECOSYSTEM, B. Pharmacy, 1st Year, Sem-II, Environmen...
 
Grant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy ConsultingGrant Readiness 101 TechSoup and Remy Consulting
Grant Readiness 101 TechSoup and Remy Consulting
 
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in DelhiRussian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
Russian Escort Service in Delhi 11k Hotel Foreigner Russian Call Girls in Delhi
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 

Wang midterm-defence

  • 1. Combining Motion Information with Appearance Information and Utilizing of Mutual Information Encoded Among Image Features of the Same Object for Better Detection Performance Zhipeng Wang Directed by: Pro. Katsushi Ikeuchi, Dr. Masataka Kegesawa, and Dr. Shintaro Ono Computer Vision Laboratory, The University of Tokyo 運動情報を外観情報との結びつけおよび 同一物体の画像特徴によりエンコードされた相互情報の利用 ーー検出性能の向上のため
  • 2. Outline  Background and state-of-the-art detection methods  Our work (goal, challenges, related work, method, results, contribution, and conclusion)  Section 1, 2: utilizing motion information  Section 3, 4: utilizing mutual information  Conclusion  Publication list
  • 3. Background  Object detection from images:  One of the basic perceptual skills in human  Plays an important role in machine vision area
  • 4. Two primary categories  Sliding-window methods  Use classifiers to answer whether each sub- image contains a target object  Hough transform based methods  Infer about object status in a bottom-up manner from detected image features
  • 5. Better detection performance  Representative methods  Image features: HOG, sift  Classifiers  Machine learning methods: deep learning  Efficient search techniques  Branch and bound[1] [1] Lampert, C.H.; Blaschko, M.B.; Hofmann, T.; , "Efficient Subwindow Search: A Branch and Bound Framework for Object Localization," Pattern Analysis and Machine Intelligence, IEEE Transactions on , vol.31, no.12, pp.2129-2142, Dec. 2009
  • 6. Our work towards better detection performance  Combining motion information with appearance information  Using mutual information encoded among the features of the same object  By using extra information, we expect to achieve better detection performance
  • 7. Our work  Combining motion information with appearance information  1. Common Fate Hough Transform  2. Emergency Telephone Indicator Detection in Tunnel Environment
  • 8. 1. Common Fate Hough Transform
  • 9. Goal and challenges  Where and what label is each object in the scene  Challenges:  Near objects  Similar different- class objects Common Fate Hough Transform
  • 10. Related work  Motion for object detection  Background subtraction  Optical flow[1]  Combination of appearance and motion information  Mainly for tracking [1] G. Brostow and R. Cipolla. Unsupervised bayesian detection of independent motion in crowds. In CVPR, pages I: 594–601, 2006.
  • 11. Hough transform  On the current image:  Detect keypoints and extract image feature  Find the best matched codes from a trained codebook  Each keypoint votes for object center and label B. Leibe and B. Schiele. Interleaved object categorization and segmentation. In BMVC, pages 759–768, 2003.
  • 12. Hough transform  Discrete to continuous  Blurring each vote  Summing up all votes from all object parts Common Fate Hough Transform
  • 13. The common fate principle  One of the gestalt laws  Elements with the same motion tend to be perceived as one unit Common Fate Hough Transform
  • 14. Common fate Hough transform  Motion analysis[1]  Keypoint tracking (KLT tracker)  Clustering the trajectories [1] G. Brostow and R. Cipolla. Unsupervised bayesian detection of independent motion in crowds. In CVPR, pages I: 594–601, 2006.
  • 15. Common fate Hough transform  The weight of each vote is related to the support it gains from the motion group
  • 17. Common fate Hough transform
  • 18. Common Fate Hough Transform  Inference  1. Find the highest peak in the confidence space  2. Exclude the object parts belonging to the highest peak  3. If the Hough image is not empty, go to 1.
  • 20. Common Fate Hough Transform Inference
  • 21. Common Fate Hough Transform Inference
  • 22. Common Fate Hough Transform Inference
  • 23. Common Fate Hough Transform Inference
  • 24. Common Fate Hough Transform Inference
  • 25. Common Fate Hough Transform Inference
  • 26. Common Fate Hough Transform Inference
  • 27. Experimental results  Dataset  720×576  401 continuous frames  633 different-class ground truth bounding boxes on 79 frames Common Fate Hough Transform
  • 38. Result comparison [1] O. Barinova, V. Lempitsky, and P. Kohli. On detection of multiple object instances using hough transforms. In CVPR, pages 2233–2240, 2010.
  • 39. Contribution  A detection method with better detection performance than the CVPR’10 method  A successful combination of motion and appearance information for detection  A successful attempt to combine human perception rules into detection method in the computer vision area Common Fate Hough Transform
  • 40. Conclusion  Motion largely improves detection results by being effectively combined with appearance information  The method is not efficient   then we propose Common Fate Hough Transform
  • 41. 2. Detection of Emergency Telephone Indicators by Fusion of Motion Information with Appearance Information
  • 42. Goal Emergency Telephone Indicator Detection  For vehicle positioning  Detecting emergency telephone indicators  Installed at known location  Infrared cameras (far infrared) installed on vehicle’s top
  • 43. Challenges  Challenges:  Noisy objects  Real-time requirement Emergency Telephone Indicator Detection
  • 44. Method  Method: a two-step method  Detect, verify, and cluster keypoints  Verify keypoint clusters by appearance and motion information Emergency Telephone Indicator Detection
  • 45. Keypoint Detection Keypoint Verification Keypoint Clustering Keypoint Cluster Verification by Appearance Keypoint Cluster Tracking Keypoint Cluster Verification by Motion Pipeline
  • 46. Pipeline  Detect keypoints  Averagely sampling  Setting intensity thresholds  160-190 (0-255) Emergency Telephone Indicator
  • 47. Pipeline  Verify keypoints  Intensity histogram  Building a mixture model using k-means Emergency Telephone Indicator Detection
  • 48. Pipeline  Cluster keypoints  Build a minimum spanning tree (Euclidean distance)  Split the tree Emergency Telephone Indicator Detection
  • 49. Pipeline  Verify keypoint clusters by appearance  Adaboost machine  Trained from positive and negative examples  Intensity histogram Emergency Telephone Indicator Detection
  • 50. Pipeline  Keypoint cluster tracking  No occlusion  Connect each detection response to its nearest trajectory (appearance, scale, and time gap) Emergency Telephone Indicator Detection
  • 51. Pipeline  Keypoint cluster verification by motion  Fit each trajectory as a straight line  Use the significance of the fitting as criteria Emergency Telephone Indicator Detection
  • 52. Benefits of the pipeline  The system is real-time  More time consuming steps dealing with fewer instances  Original image: 105 pixels  Keypoint detection: 104 points  Keypoint verification: 103 keypoints  Keypoint clustering: 102 keypoints   Less than 10 keypoint clusters Emergency Telephone Indicator Detection
  • 53. Experimental results  Real-time (34 fps)  A laptop with Intel Core2 Duo 2.8GHz processors
  • 55. Contribution  A specialized real-time detection method on cluttered data  A successful combination of motion and appearance information for detection  The meaning of this research: a potential applicable solution for vehicle positioning in tunnel environment Emergency Telephone Indicator Detection
  • 56. Conclusion  Motion information plays an important role in the detecting process Emergency Telephone Indicator Detection
  • 57. Future work  We collect new data using new camera and plan to improve the pipeline on the new collected data Emergency Telephone Indicator Detection
  • 58.  Two methods utilizing motion information are proposed   When motion information is not available, we then go to
  • 59. Our work  Utilizing of Mutual Information Encoded Among Image Features of the Same Object  3. Pyramid Match Score for Object Detection  4. Detection of Object With In-plane Rotations
  • 60. 3. Pyramid Match Score for Object Detection
  • 61. Motivation  Sliding window based methods:  Often [1] ignore positional information of each image feature  Methods based on Hough transform  Ignore overall information of each object  Intention of this method:  An efficient method to utilize positional information of each image feature and overall information gained from the whole object [1] Lazebnik, S.; Schmid, C.; Ponce, J.; , "Beyond Bags of Features: Spatial Pyramid Matching for Recognizing Natural Scene Categories," Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on , vol.2, no., pp. 2169- 2178, 2006
  • 62. Pyramid Match  Pyramid Match is a method to find best one- one match between two sets of points  Here gives a 2D example  Real data is 14D  Appearance information:12D  Position information: 2D
  • 63. Pyramid match  Divide the space from fine to coarse  If two points are in the same grid under a division, the are considered as match and excluded
  • 64. Pyramid match  The distance of two matched points  No need to be calculated  Assigned  According to the size of the grid, when they are found match 1/4 * 2 1/2 * 2 1*2
  • 65. Pyramid Match Score  Pyramid match between two sets  Each sub-image: considered as a 14D point set (appearance: 12, position: 2)  A “super template” contains all 14D points of training images
  • 66. Pyramid Match Score  Find the near best match between each sub- image of a test image and the “super template”  The match score (distance) is used as the confidence that the sub-image contains a target object
  • 68. Results 0 0.1 0.2 0.3 0.4 0.5 0.6 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Detectionrate Precison 3 10 7 av back
  • 69. Comments  The results are not good enough  The method seems promising, and in future we will improve the method   This mutual information can be further used
  • 70. 4 Detection of Object With In-plane Rotations
  • 71. About the related work  Training examples of all rotation directions,  not robust  or very time-consuming
  • 72. Related work  Two neural networks to deal with one sub-image  One for estimation of the sub-image’s rotation angle  The sub-image is rotated so that the potential face is upright  The other for decision of whether one object exists H. Rowley, S. Baluja, and T. Kanade. Rotation invariant neural network-based face detection. In CVPR, pages 38–44, 1998.
  • 73. Related work  Local feature method  Infer rotation based on the gradient direction of sift feature  Graph-based methods  Consider object as graph, and change the graph Detection of Object With In-plane Rotations
  • 74. Idea  Suppose we have a codebook trained from objects without in-plane rotations Detection of Object With In-plane Rotations
  • 75. Idea  With proof omitted  For two or more votes from different keypoints, there exists one and only one rotation angle which will minimize the difference of the voted centers If all votes are good estimations, we can expect to be a good estimation the rotation angle of the object Detection of Object With In-plane Rotations
  • 76. To do  Propose a method capable of detecting objects with in-plane  Evaluate the robustness of Detection of Object With In-plane Rotations
  • 77. Conclusion  Motion information combined with appearance information  Distinguish target objects with noisy object  Enhance detection rate  Mutual information encoded among the image features of the same object  We expect to get a method capable of utilizing positional and overall information  We expect to get a method capable of detecting objects with in-plane rotations
  • 78. Publication list  [1] Zhipeng Wang; Jinshi Cui; Hongbin Zha; Kegesawa, M.; Ikeuchi, K.; , "Object detection by common fate Hough transform," Pattern Recognition (ACPR), 2011 First Asian Conference on , vol., no., pp.613-617, 28-28 Nov. 2011  [2] Zhipeng Wang; Kagesawa, M.; Ono, S.; Banno, A.; Ikeuchi, K.; , "Emergency light detection in tunnel environment: An efficient method," Pattern Recognition (ACPR), 2011 First Asian Conference on , vol., no., pp.628-632, 28-28 Nov. 2011  [3] Zhipeng Wang; Jinshi Cui; Hongbin Zha; Masataka Kegesawa; Shintaro Ono; Katsushi Ikeuchi;, “Detection by Motion-based Grouping of Object Parts,” International Journal of ITS Research (submitted)  [4] Zhipeng Wang; Masataka Kegesawa; Shintaro Ono; Atsuhiko Banno; Takeshi Oishi; Katsushi Ikeuchi;, “Detection of Emergency Telephone Indicators Using Infrared Cameras for Vehicle Positioning in Tunnel Environment,” ITS World Congress 2013 (submitted)
  • 79. Thank you very much!