SlideShare a Scribd company logo
1 of 6
Download to read offline
IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 18, Issue 1, Ver. I (Jan – Feb. 2016), PP 24-29
www.iosrjournals.org
DOI: 10.9790/0661-18112429 www.iosrjournals.org 24 | Page
Competent Tracking of Moving Object Using Affine &
Illumination Insensitive Template Matching
Moiz Abdul Hussain1
, Dr. Govind Ukhandrao2
1
(Asst. Prof. ENTC Deptt. Anuradha Engineering College, Chikhli, India)
2
(Principal Sharadchandra Pawar College of Engineering, Otur, India)
Abstract : Moving object detection & tracking in real world scene is becoming significant problem in today’s
era. The extensive study in this area is motivated by potential number of applications of object tracking. In this
paper, we analyze a method for motion segmentation & tracking of nonstationary objects that uses the complete
image information using the affine flow equations. These affine flow parameters are then combined with
illumination insensitive template matching tracker to efficiently segment & track the moving objects.
Experimental results performed on various database videos shows the method effectively tracks the object
without miss detection of the target compared to other methods.
Keywords: Affine Flow, Video Surveillance, Object Tracking, Template matching.
I. Introduction
Motion segmentation & tracking has been a crucial component for many computer vision applications
such as structure from motion, objects tracking/recognition, advanced video editing and behaviour analysis. A
variety of computational models, e.g. optical flow fields [7, 8], layers [1] and boundaries [9], have been
developed to approximate the motion from a video sequence. It is significant to assess the performance of
various motion analysis algorithms to gain insight and design better models. However, little has been done for
evaluating motion analysis algorithms compared to the tremendous effort put into developing these algorithms.
The extraction of the affine parameters has fundamentally two components: the recognition of a
suitable set of spatial patches to characterize every surface in a scene; and the tracking of the patches during the
frame sequence. The tracking of the patches (2-D) and assessment of their connected affine motion parameters
is realized using weighted linear regression over an approximated optical flow field. The weights are provided
by truncated Gaussian windows, each defining the spatial extent of one of the patches being tracked. The
preference of a gaussian window function is deliberate: the group of such functions are closed under the action
of affine transformation and hence naturally represent the evolution of the patches as they warp according to the
affine motion approximation.
The main contribution of this paper is the analysis of fast version of method for obtaining affine flow
from the image sequence. These affine flow parameters are then combined with the illumination insensitive
tracker to appropriately track the nonstationary objects. The motion evaluation is based on three observations.
First, humans are experts at segmenting layers in a video sequence because human being can easily recognize
the moving objects and their relative depth relationships. Second, humans are sensitive to any differences
between two images when these two images are displayed back and forth. Third, humans have knowledge of the
smoothness and discontinuities of the motion that a moving object undergoes.
We thoroughly tested our method by conducting experiments on several videos collected for our
experimentation purpose in both indoor and outdoor surroundings. The experimental results demonstrate that the
method performs favorably well in real world surroundings. One of the advantages of our approach compared to
the other state-of-the-art methods is that it reduces the number of false detections, as pixel-level comparison can
be per-formed in regions with significant motion only. Another advantage is that the illumination insensitive
template matching eliminates the effects of illumination changes which are a major problem in visual
surveillance system.
The paper is structured as follows. Section 2 explores some of the related methods. Section 3 presents
our system methodology in detail. Section 4 describes the database collection for experimentation. Section 5
presents some experimental results & Finally, Section 6 concludes the paper.
II. Literature Review
A large number of methods are proposed for video surveillance system using human motion detection
& tracking. Bramble [10] detects people in indoor environments and uses explicit 3-D cylindrical modelling of
target objects. Haritaoglu et al. proposed a system which uses projection histograms of the single detected object
as features to classify different actions [2]. Hager and Belhumeur [11] perform photometric stereo using at least
three images of a scene under linearly independent lighting conditions.
Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching
DOI: 10.9790/0661-18112429 www.iosrjournals.org 25 | Page
In [3] Hidden Markov Models are used. However, these methods are usually much slower and
adaptation to changes of the scene is difficult. Hedvig [12] uses optical flow features for training directly.
Support vector machine is used for training. This method can get acceptable detection result in outdoor scene.
McKenna used a combination of motion, skin color and face detection to track people and their faces [6].
Karmann and Brandt [13] and Kilger [4] respectively proposed an adaptive background model based on Kalman
filtering to adapt temporal changes of weather and lighting.
Chen et al. [5] use a human attention model to create different cropped and scaled versions of images
according to the screen size. An algorithm was presented in [14] to track moving objects in the scene by
considering dynamic occlusions among moving objects; this is especially useful to track pedestrians moving in
indoor environments. Robust detection of illumination changes in outdoor video surveillance images was
effectively achieved in [15] by using an adaptive algorithm. This approach dynamically learns and updates the
background by using a segmentation algorithm, especially for sudden changes.
Since successful tracking depends a lot on precise object detection, the evaluation of object detection
algorithms within a surveillance system plays a significant part in overall performance study of the whole
system.
III. system framework
Proposed analysis is directed to build a system that is able to detect nonstationary/mobile objects in the
scene and track those using affine parameters & illumination insensitive tracker. Figure 1 below represents a
block diagram of the system.
Fig. 1: Framework for object detection & tracking
Affine flow model combined with illumination insensitive tracker overcomes the problem of
traditionally used optical flow technique like motion changes due to camera movement, dynamic background
objects (e.g. tree leave or branches), and changes in the background geometry (e.g. parked cars). For the
surveillance of outdoor scenes, sturdiness against noise and adaptivity to illumination changes are also essential.
We coupled the method of affine flow extraction with template matching algorithm with illumination intensity
minimization model which will take care of illumination changes in real images caused due to automatic
exposure alterations of the camera, adjustment of light cause irradiance, appearance of shadows or movement of
the tracked objects. Our objective is to provide a robust and improved method to find the moving objects in the
video frame as well as to track them. The proposed method is effective in reducing the number of false alarms
that may be triggered by a number of reasons. Fig.1shows the different stages involved in the proposed system.
First, image frames are taken out from video, and then, affine flow parameters are extracted from an image.
Simultaneously, illumination insensitivity hyperplane method for template matching is also applied to perfectly
track the object.
3.1 Affine Flow
The starting point is the assumption of brightness constancy, which assumes that the brightness
structures of local-time varying image regions are unchanging under motion for a short period of time.
Formally, this is defined as,
𝐼 𝑥, 𝑦, 𝑡 = 𝐼(𝑥 − 𝑢, 𝑦 − 𝑣, 𝑡 − 1) … (1)
Where (𝑥, 𝑦)𝑡
represents image position in pixel coordinates, t represents the temporal coordinate,
(𝑢, 𝑣)𝑡
represents the motion at image position (𝑥, 𝑦)𝑡
over the time 𝑡 + 1 and 𝐼(𝑥, 𝑦, 𝑡) represents the image
brightness function. Using the least-squares criteria, we seek the motion that minimizes the error 𝜖 over a region
𝑅 of the image, formally,
∈ = [ 𝑥, 𝑦, 𝑡 − 𝐼 𝑥 − 𝑢, 𝑦 − 𝑣, 𝑡 − 1 ]2
𝑥,𝑡,∈𝑅 … (2)
Video
I/P
Frame
Extraction
Affine
Parameters
Adjust Template
(βt+1, ϒt+1)
Compute µ Track Moving
Object
Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching
DOI: 10.9790/0661-18112429 www.iosrjournals.org 26 | Page
Assumint that 𝐼(𝑥, 𝑦, 𝑡) is approximately locally linear (2) is simplified by taking a Taylor series
expansion and omitting terms higher than first order.
∈ = [𝐼 𝑥, 𝑦, 𝑡 − (𝐼 𝑥, 𝑦, 𝑡) − 𝐼𝑥 𝑥, 𝑦, 𝑡 𝑢 − 𝐼𝑦 𝑥, 𝑦, 𝑡 𝑣 − 𝐼𝑡(𝑥, 𝑦, 𝑡) ]2
𝑥,𝑦,∈𝑅 … (3)
∈ = [𝐼𝑥 𝑥, 𝑦, 𝑡 𝑢 + 𝐼𝑦 𝑥, 𝑦, 𝑡)𝑣 + 𝐼𝑡 𝑥, 𝑦, 𝑡 ]2
𝑥,𝑦,∈𝑅 … (4)
Where, 𝐼𝑥 𝐼𝑦 and 𝐼𝑡 represent the partial derivatives of the intensity function with respect to the spatial
parameters 𝑥, 𝑦 and temporal parameter 𝑡, respectively. To obtain the estimate of velocity in the pure translation
case, the partial derivatives of the error in (4) are taken with respect to the translational velocity (𝑢, 𝑣) 𝑇
and set
to zero. This yields the following set of linear equations,
𝐼𝑥
2
𝐼𝑥 𝐼𝑦
𝐼𝑥 𝐼𝑦 𝐼𝑦
2
𝑢
𝑣
= −
𝐼𝑥 𝐼𝑡
𝐼𝑦 𝐼𝑡
… (5)
A minimum of two points are required for the full solution. The under constrained problem given a
single point is commonly referred to in the literature as the aperture problem. In the case where the motion is
modelled locally by an affine transformation, corresponding to a composition of rotation, dilation, shear and
translation,
(𝑢, 𝑣) 𝑇
are defined as follows,
𝑢
𝑣
=
𝑎1 𝑎2
𝑎41 𝑎5
𝑥
𝑦 +
𝑎0
𝑎3
… (6)
As in the case of translational motion, to obtain the estimates of the affine parameter (4) is differential
w.r.t the six unknown affine parameters (𝑎𝑖 𝑤𝑕𝑒𝑟𝑒 𝑖 = 0, … . , 5) and set to zero.
3.2 Template Matching
For eradicating the effects of illumination changes, a technique is to estimate illumination
compensation parameters for the current template, in order to adjust its gray-level values with respect to the
reference template. Using the linear model
𝐹𝑛𝑒𝑤 𝑥 = 𝛽𝑓 𝑥 + 𝛾 ∀ × 𝜀 𝑟 … (7)
Contrast and brightness variations are represented by illumination compensation parameters 𝛽 and 𝛾.
Least-squares minimization problem is solved when corresponding points of the initial template 𝑓(𝑟, 𝑡0) and the
current template 𝑓(𝑔(𝑟, 𝜇 (𝑡))) are known.
𝛽𝑡, 𝛾𝑡 = 𝑎𝑟𝑔𝑚𝑖𝑛 [𝛽𝑓𝑔 𝑥, 𝜇 𝑡 , 𝑡 + 𝛾 − 𝑓 𝑥, 𝑡0 ]𝑥∈𝑟 … (8)
We replace 𝑔(𝑥, 𝜇(𝑡)) with 𝑥(𝑡). Differentiating equation (8) with respect to the motion compensation
parameters yields the linear system
𝑓2 𝑥 𝑡 ,𝑡 𝑓 𝑥 𝑡 ,𝑡
𝑓 𝑥 𝑡 ,𝑡 1𝑥∈𝑟
𝛽 𝑡
𝛾 𝑡
= 𝑓2 𝑥,𝑡0 𝑓 𝑥 𝑡 ,𝑡
𝑓 𝑥 𝑡 ,𝑡0
𝑥∈𝑟 … (9)
To avoid time consuming operation of computing the matrix on the left, we exchange the reference
template with the current template and hence obtain the motion compensation parameters and for adapting the
reference template to the current template. Consequently, the matrix in equation (9) has to be computed only
once. As we still want to adapt the current template, we revert to the original illumination compensation
parameters.
3.3 Centroid Estimation
A set of two integers, 𝐶 (𝐶𝑥, 𝐶𝑦) which determines the position of the moving object in the given
scene. The centroid coordinates are calculated by:
𝐶𝑥 = 𝐶𝑥 + 𝑥 … (10)
𝐶𝑦 = 𝐶𝑦 + 𝑦 … (11)
𝑇 = 𝑇 + 1 … (12)
For each pixel, where (𝑥, 𝑦) is the current pixel location. The resulting centroid is then divided by the
total value:
𝐶𝑥 = 𝐶𝑥 / 𝑇 … (13)
𝐶𝑦 = 𝐶𝑦 / 𝑇 … (14)
This centroid positional information is then transferred to the object tracking module which is coupled
with the illumination insensitive template matching model which gives the good result for detection & tracking.
IV. Dataset
For the evaluation of our experiments we used two different data bases. First, Weizmann database
which is publically available and another is our own created database, this database contains 86 sequences
Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching
DOI: 10.9790/0661-18112429 www.iosrjournals.org 27 | Page
having five & four of classes of actions (Jump, Run, Walk, Side, Handwaving, Jogging) respectively performed
by 19 different subjects in two different conditions d1-d2.
d1: Indoor Environment;
d2: Outdoor (Playground) + illumination variations
Few samples of Indoor environment
Outdoor (Playground) + illumination variations Environment
„Running‟ „Jogging‟ „Walking‟
„Jumping‟ „Handwaving‟
“Self Database” Images
„Bend‟ „Jack‟ „Pjump‟
„Run‟ „Side‟ „Skip‟
“Weizmann Database” Images
Fig. 2: Sample images of Self & Weizmann database
V. Experimental result
This section summarizes the experimental results for motion detection & tracking system. The
experiments were performed on MatLab 7.10.0 software. The PC was equipped with Pentium corei3, 2.13 GHz
processor & 4 GB RAM Memory. The algorithm was implemented on two different datasets i.e. Weizmann
publically available & self made database. Weizmann database [180*144, Frame rate 25frames/sec] has
different actions Jump, Run, Walk, Bend, Skip and Side. All the actions were captured with steady camera. Our
own data-base [1280*720, Frame rate 29frames/sec] was captured when both camera & target are moving in
dynamic environment.
Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching
DOI: 10.9790/0661-18112429 www.iosrjournals.org 28 | Page
Proposed Model Without Illumination insensitive tracker
Fig. 3: Experimental Results on database videos
Figure 3 shows the comparison result of the proposed tracker system with the motion detection &
tracking system of affine flow without illumination insensitive template tracker. As it can be observed from the
figure 3 that, proposed model tracks the nonstationary object precisely as compared to the other state of the art
method.
The algorithm implemented has achieved highly robust & clean results in cluttered backgrounds and
illumination change. This shows the effectiveness of algorithm. Table 1 below shows the variation in the
pyramid level of affine flow calculations for the 3 number of iterations. From the value represented in the table
it can be seen that as the pyramid level increases the value of affine flow in between the pixel value decreases.
Table 1: Norm of optical flow values for 3 numbers of iterations
Iteration 1 Iteration 2 Iteration 3
Pyramid 1 318.47 311.17 309.05
Pyramid 2 191.28 190.77 189.08
Pyramid 3 104.53 101.16 99.82
Pyramid 4 47.29 43.43 41.96
Pyramid 5 23.48 21.43 21.10
Pyramid 6 10.86 9.87 9.65
Fig. 4: Error Estimation of targets tracked by proposed method & affine flow without template tracker method.
Figure 4 displays the results in terms of average tracking errors. Proposed method shows the error
maximum upto 3 pixels where as the method of the affine flow without illumination insensitive template
matching tracker have error upto 7 pixels. This error comparison shows that the proposed system is very much
precise in the motion segmentation and tracking.
VI. Conclusion
In this paper, we have presented a new proposal for object detection & tracking based using the
combination of affine flow extraction & illumination insensitive template matching model. The system uses
Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching
DOI: 10.9790/0661-18112429 www.iosrjournals.org 29 | Page
efficient technique of affine flow extraction which removes the drawback of traditionally used optical flow
technique like detecting non-stationary background objects and shadows cast by moving objects. Further
module of template matching eliminates the problem of illumination & dynamic environment conditions.
The integrated system was successfully tested on a number of sample videos containing various moving entities.
Analysis was made on its performance. Experimental results shows that this algorithm has better tracking
precision and can resolve the robust tracking problem in illumination changes & dynamic background, so it has
a very wide range of applications.
Further work will concentrate on occlusion handling & multiple object tracking problems that will
work in more complex cluttered environments. We will also focus on increasing the performance of the system
by decreasing the time duration from frame to frame processing.
References
Journal Papers:
[1] J. Y. A. Wang and E. H. Adelson, Representing moving images with layers, IEEE Trans. Image Processing, , 1994, 3(5):625 – 638.
[2] I.Haritaoglu, D.Harwood and L.S.Davis, W4: Realtime surveillance of people and their actions, IEEE Trans. on Pattern Anal. and
Mach. Intell. , August 2000, 22(8),pp.809-830.
[3] J. Kato, S. Joga, J. Rittscher, and A. Blake, An HMM-based segmentation method for traffic monitoring movies, IEEE Trans. on
PAMI, , 2002, vol. 24, no. 9, pp. 1291–1296.
[4] M. Kilger, A shadow handler in a video-based real-time traffic monitoring system, Proc. of IEEE Workshop on Applications of
Computer Vision. 1992, pp. 1060-1066.
[5] L.-Q. Chen, X. Xie, X. Fan, W.-Y. Ma, H.-J. Zhang, and H.-Q. Zhou, A visual attention model for adapting images on small
displays, Multimedia Syst., Oct. 2003. vol. 9, pp. 353–364.
Chapters in Books:
[6] McKenna, Stephen J., Tracking groups of people Computer Vision and Image Understanding 80.1 (2000): 42-56.
Proceedings Papers:
[7] B. K. P. Horn and B. G. Schunck, Determing optical flow, Artificial Intelligence, 17:185–203, 1981.
[8] B. Lucas and T. Kanade, An iterative image registration technique with an application to stereo vision, In Proc. of the Intl. Joint
Conf. on Artificial Intelligence, pages 674–679, 1981.
[9] C. Liu, W. T. Freeman, and E. H. Adelson, Analysis of contour motions, In NIPS, 2006.
[10] Isard, Michael, and John MacCormick, BraMBLe: A Bayesian multiple-blob tracker, Computer Vision, 2001. ICCV 2001.
Proceedings. Eighth IEEE International Conference on. Vol. 2. IEEE, 2001.
[11] G.D. Hager and P.N. Belhumeur, Real-time tracking of image regions with changes in geometry and illumination, Proc. IEEE
Computer Society Conference on Computer Vision and Pattern Recognition: 403 – 410, 1996.
[12] Hedvig Sidenbladh, Detecting Human Motion with Support Vector Machines, ICPR 2004.
[13] K.P. Karmann, A. Brandt, Moving object recognition using an adaptive background memory, In Cappellini, Time-varying Image
Processing and Moving Object Recognition, Elsevier, Amsterdam, The Netherlands, 1990.
[14] Marchesotti, L. Marcenaro, L. Regazzoni C. S, Tracking and counting multiple interacting pedestrian in indoor scenes, In
Proceedings of third IEEE international workshop on performance evaluation of tracking and surveillance, Copenhagen Denmark,
June 1, 2002.
[15] Regazzoni, C, Adaptive change detection approach for object detection in outdoor scenes under variable speed illumination
changes, In Proceedings of Eusipco 2000.

More Related Content

What's hot

Ijarcet vol-2-issue-3-1280-1284
Ijarcet vol-2-issue-3-1280-1284Ijarcet vol-2-issue-3-1280-1284
Ijarcet vol-2-issue-3-1280-1284
Editor IJARCET
 

What's hot (19)

A Biometric Approach to Encrypt a File with the Help of Session Key
A Biometric Approach to Encrypt a File with the Help of Session KeyA Biometric Approach to Encrypt a File with the Help of Session Key
A Biometric Approach to Encrypt a File with the Help of Session Key
 
Implementation of Object Tracking for Real Time Video
Implementation of Object Tracking for Real Time VideoImplementation of Object Tracking for Real Time Video
Implementation of Object Tracking for Real Time Video
 
Ijarcet vol-2-issue-3-1280-1284
Ijarcet vol-2-issue-3-1280-1284Ijarcet vol-2-issue-3-1280-1284
Ijarcet vol-2-issue-3-1280-1284
 
Copy-Rotate-Move Forgery Detection Based on Spatial Domain
Copy-Rotate-Move Forgery Detection Based on Spatial DomainCopy-Rotate-Move Forgery Detection Based on Spatial Domain
Copy-Rotate-Move Forgery Detection Based on Spatial Domain
 
Enhanced Optimization of Edge Detection for High Resolution Images Using Veri...
Enhanced Optimization of Edge Detection for High Resolution Images Using Veri...Enhanced Optimization of Edge Detection for High Resolution Images Using Veri...
Enhanced Optimization of Edge Detection for High Resolution Images Using Veri...
 
Automatic rectification of perspective distortion from a single image using p...
Automatic rectification of perspective distortion from a single image using p...Automatic rectification of perspective distortion from a single image using p...
Automatic rectification of perspective distortion from a single image using p...
 
V.KARTHIKEYAN PUBLISHED ARTICLE 1
V.KARTHIKEYAN PUBLISHED ARTICLE 1V.KARTHIKEYAN PUBLISHED ARTICLE 1
V.KARTHIKEYAN PUBLISHED ARTICLE 1
 
Empirical Coding for Curvature Based Linear Representation in Image Retrieval...
Empirical Coding for Curvature Based Linear Representation in Image Retrieval...Empirical Coding for Curvature Based Linear Representation in Image Retrieval...
Empirical Coding for Curvature Based Linear Representation in Image Retrieval...
 
Edge detection by using lookup table
Edge detection by using lookup tableEdge detection by using lookup table
Edge detection by using lookup table
 
Performance Evaluation of Image Edge Detection Techniques
Performance Evaluation of Image Edge Detection Techniques Performance Evaluation of Image Edge Detection Techniques
Performance Evaluation of Image Edge Detection Techniques
 
Blind detection of image manipulation @ PoliMi
Blind detection of image manipulation @ PoliMiBlind detection of image manipulation @ PoliMi
Blind detection of image manipulation @ PoliMi
 
B42020710
B42020710B42020710
B42020710
 
Gabor Filter
Gabor FilterGabor Filter
Gabor Filter
 
AUTOMATED IMAGE MOSAICING SYSTEM WITH ANALYSIS OVER VARIOUS IMAGE NOISE
AUTOMATED IMAGE MOSAICING SYSTEM WITH ANALYSIS OVER VARIOUS IMAGE NOISEAUTOMATED IMAGE MOSAICING SYSTEM WITH ANALYSIS OVER VARIOUS IMAGE NOISE
AUTOMATED IMAGE MOSAICING SYSTEM WITH ANALYSIS OVER VARIOUS IMAGE NOISE
 
Improving image resolution through the cra algorithm involved recycling proce...
Improving image resolution through the cra algorithm involved recycling proce...Improving image resolution through the cra algorithm involved recycling proce...
Improving image resolution through the cra algorithm involved recycling proce...
 
L0816166
L0816166L0816166
L0816166
 
Ijetcas14 372
Ijetcas14 372Ijetcas14 372
Ijetcas14 372
 
A NOVEL IMAGE SEGMENTATION ENHANCEMENT TECHNIQUE BASED ON ACTIVE CONTOUR AND...
A NOVEL IMAGE SEGMENTATION ENHANCEMENT TECHNIQUE BASED ON  ACTIVE CONTOUR AND...A NOVEL IMAGE SEGMENTATION ENHANCEMENT TECHNIQUE BASED ON  ACTIVE CONTOUR AND...
A NOVEL IMAGE SEGMENTATION ENHANCEMENT TECHNIQUE BASED ON ACTIVE CONTOUR AND...
 
Image Segmentation using Otsu's Method - Computer Graphics (UCS505) Project PPT
Image Segmentation using Otsu's Method - Computer Graphics (UCS505) Project PPTImage Segmentation using Otsu's Method - Computer Graphics (UCS505) Project PPT
Image Segmentation using Otsu's Method - Computer Graphics (UCS505) Project PPT
 

Viewers also liked

Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy ResinThermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
IOSR Journals
 

Viewers also liked (20)

Simulation of Signals with Field Signal Simulator
Simulation of Signals with Field Signal SimulatorSimulation of Signals with Field Signal Simulator
Simulation of Signals with Field Signal Simulator
 
I010234548
I010234548I010234548
I010234548
 
E010424043
E010424043E010424043
E010424043
 
I010626877
I010626877I010626877
I010626877
 
N18030296105
N18030296105N18030296105
N18030296105
 
M018147883
M018147883M018147883
M018147883
 
I017134347
I017134347I017134347
I017134347
 
D012522433
D012522433D012522433
D012522433
 
Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy ResinThermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
Thermal conductivity Characterization of Bamboo fiber reinforced in Epoxy Resin
 
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 compositeStudy of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
Study of the Tribological behavior of as cast Al-4.2%Cu-Al2O3 composite
 
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
Objective function to Optimize Bending Stress at Critical Section of Asymmetr...
 
M010427686
M010427686M010427686
M010427686
 
U180304130141
U180304130141U180304130141
U180304130141
 
P01213112116
P01213112116P01213112116
P01213112116
 
G0813841
G0813841G0813841
G0813841
 
A013150107
A013150107A013150107
A013150107
 
A017440109
A017440109A017440109
A017440109
 
D017532629
D017532629D017532629
D017532629
 
B018140813
B018140813B018140813
B018140813
 
H017645560
H017645560H017645560
H017645560
 

Similar to D018112429

Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorProposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
QUESTJOURNAL
 
A New Algorithm for Tracking Objects in Videos of Cluttered Scenes
A New Algorithm for Tracking Objects in Videos of Cluttered ScenesA New Algorithm for Tracking Objects in Videos of Cluttered Scenes
A New Algorithm for Tracking Objects in Videos of Cluttered Scenes
Zac Darcy
 
Schematic model for analyzing mobility and detection of multiple
Schematic model for analyzing mobility and detection of multipleSchematic model for analyzing mobility and detection of multiple
Schematic model for analyzing mobility and detection of multiple
IAEME Publication
 

Similar to D018112429 (20)

Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operatorProposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
Proposed Multi-object Tracking Algorithm Using Sobel Edge Detection operator
 
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
 
X36141145
X36141145X36141145
X36141145
 
Development of Human Tracking in Video Surveillance System for Activity Anal...
Development of Human Tracking in Video Surveillance System  for Activity Anal...Development of Human Tracking in Video Surveillance System  for Activity Anal...
Development of Human Tracking in Video Surveillance System for Activity Anal...
 
Survey on video object detection & tracking
Survey on video object detection & trackingSurvey on video object detection & tracking
Survey on video object detection & tracking
 
Motion Object Detection Using BGS Technique
Motion Object Detection Using BGS TechniqueMotion Object Detection Using BGS Technique
Motion Object Detection Using BGS Technique
 
Motion Object Detection Using BGS Technique
Motion Object Detection Using BGS TechniqueMotion Object Detection Using BGS Technique
Motion Object Detection Using BGS Technique
 
Motion Human Detection & Tracking Based On Background Subtraction
Motion Human Detection & Tracking Based On Background SubtractionMotion Human Detection & Tracking Based On Background Subtraction
Motion Human Detection & Tracking Based On Background Subtraction
 
Moving object detection using background subtraction algorithm using simulink
Moving object detection using background subtraction algorithm using simulinkMoving object detection using background subtraction algorithm using simulink
Moving object detection using background subtraction algorithm using simulink
 
Vehicle Tracking Using Kalman Filter and Features
Vehicle Tracking Using Kalman Filter and FeaturesVehicle Tracking Using Kalman Filter and Features
Vehicle Tracking Using Kalman Filter and Features
 
K-Means Clustering in Moving Objects Extraction with Selective Background
K-Means Clustering in Moving Objects Extraction with Selective BackgroundK-Means Clustering in Moving Objects Extraction with Selective Background
K-Means Clustering in Moving Objects Extraction with Selective Background
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET-  	  Behavior Analysis from Videos using Motion based Feature ExtractionIRJET-  	  Behavior Analysis from Videos using Motion based Feature Extraction
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
 
A New Algorithm for Tracking Objects in Videos of Cluttered Scenes
A New Algorithm for Tracking Objects in Videos of Cluttered ScenesA New Algorithm for Tracking Objects in Videos of Cluttered Scenes
A New Algorithm for Tracking Objects in Videos of Cluttered Scenes
 
IRJET-Vision Based Occupant Detection in Unattended Vehicle
IRJET-Vision Based Occupant Detection in Unattended VehicleIRJET-Vision Based Occupant Detection in Unattended Vehicle
IRJET-Vision Based Occupant Detection in Unattended Vehicle
 
Schematic model for analyzing mobility and detection of multiple
Schematic model for analyzing mobility and detection of multipleSchematic model for analyzing mobility and detection of multiple
Schematic model for analyzing mobility and detection of multiple
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
New Approach for Detecting and Tracking a Moving Object
New Approach for Detecting and Tracking a Moving Object New Approach for Detecting and Tracking a Moving Object
New Approach for Detecting and Tracking a Moving Object
 
Analysis of Human Behavior Based On Centroid and Treading Track
Analysis of Human Behavior Based On Centroid and Treading  TrackAnalysis of Human Behavior Based On Centroid and Treading  Track
Analysis of Human Behavior Based On Centroid and Treading Track
 
A Robust Method for Moving Object Detection Using Modified Statistical Mean M...
A Robust Method for Moving Object Detection Using Modified Statistical Mean M...A Robust Method for Moving Object Detection Using Modified Statistical Mean M...
A Robust Method for Moving Object Detection Using Modified Statistical Mean M...
 
[IJET V2I5P5] Authors: CHETANA M, SHIVA MURTHY. G
[IJET V2I5P5] Authors: CHETANA M, SHIVA MURTHY. G [IJET V2I5P5] Authors: CHETANA M, SHIVA MURTHY. G
[IJET V2I5P5] Authors: CHETANA M, SHIVA MURTHY. G
 

More from IOSR Journals

More from IOSR Journals (20)

A011140104
A011140104A011140104
A011140104
 
M0111397100
M0111397100M0111397100
M0111397100
 
L011138596
L011138596L011138596
L011138596
 
K011138084
K011138084K011138084
K011138084
 
J011137479
J011137479J011137479
J011137479
 
I011136673
I011136673I011136673
I011136673
 
G011134454
G011134454G011134454
G011134454
 
H011135565
H011135565H011135565
H011135565
 
F011134043
F011134043F011134043
F011134043
 
E011133639
E011133639E011133639
E011133639
 
D011132635
D011132635D011132635
D011132635
 
C011131925
C011131925C011131925
C011131925
 
B011130918
B011130918B011130918
B011130918
 
A011130108
A011130108A011130108
A011130108
 
I011125160
I011125160I011125160
I011125160
 
H011124050
H011124050H011124050
H011124050
 
G011123539
G011123539G011123539
G011123539
 
F011123134
F011123134F011123134
F011123134
 
E011122530
E011122530E011122530
E011122530
 
D011121524
D011121524D011121524
D011121524
 

Recently uploaded

Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Victor Rentea
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Victor Rentea
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
?#DUbAI#??##{{(☎️+971_581248768%)**%*]'#abortion pills for sale in dubai@
 

Recently uploaded (20)

Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
Apidays New York 2024 - APIs in 2030: The Risk of Technological Sleepwalk by ...
 
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ..."I see eyes in my soup": How Delivery Hero implemented the safety system for ...
"I see eyes in my soup": How Delivery Hero implemented the safety system for ...
 
Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..Understanding the FAA Part 107 License ..
Understanding the FAA Part 107 License ..
 
WSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering DevelopersWSO2's API Vision: Unifying Control, Empowering Developers
WSO2's API Vision: Unifying Control, Empowering Developers
 
ICT role in 21st century education and its challenges
ICT role in 21st century education and its challengesICT role in 21st century education and its challenges
ICT role in 21st century education and its challenges
 
MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024MINDCTI Revenue Release Quarter One 2024
MINDCTI Revenue Release Quarter One 2024
 
FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024FWD Group - Insurer Innovation Award 2024
FWD Group - Insurer Innovation Award 2024
 
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
Web Form Automation for Bonterra Impact Management (fka Social Solutions Apri...
 
Vector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptxVector Search -An Introduction in Oracle Database 23ai.pptx
Vector Search -An Introduction in Oracle Database 23ai.pptx
 
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data DiscoveryTrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
TrustArc Webinar - Unlock the Power of AI-Driven Data Discovery
 
Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)Introduction to Multilingual Retrieval Augmented Generation (RAG)
Introduction to Multilingual Retrieval Augmented Generation (RAG)
 
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot TakeoffStrategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
Strategize a Smooth Tenant-to-tenant Migration and Copilot Takeoff
 
Artificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : UncertaintyArtificial Intelligence Chap.5 : Uncertainty
Artificial Intelligence Chap.5 : Uncertainty
 
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024Finding Java's Hidden Performance Traps @ DevoxxUK 2024
Finding Java's Hidden Performance Traps @ DevoxxUK 2024
 
MS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectorsMS Copilot expands with MS Graph connectors
MS Copilot expands with MS Graph connectors
 
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
Modular Monolith - a Practical Alternative to Microservices @ Devoxx UK 2024
 
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin WoodPolkadot JAM Slides - Token2049 - By Dr. Gavin Wood
Polkadot JAM Slides - Token2049 - By Dr. Gavin Wood
 
AWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of TerraformAWS Community Day CPH - Three problems of Terraform
AWS Community Day CPH - Three problems of Terraform
 
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
+971581248768>> SAFE AND ORIGINAL ABORTION PILLS FOR SALE IN DUBAI AND ABUDHA...
 
Exploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with MilvusExploring Multimodal Embeddings with Milvus
Exploring Multimodal Embeddings with Milvus
 

D018112429

  • 1. IOSR Journal of Computer Engineering (IOSR-JCE) e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 18, Issue 1, Ver. I (Jan – Feb. 2016), PP 24-29 www.iosrjournals.org DOI: 10.9790/0661-18112429 www.iosrjournals.org 24 | Page Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching Moiz Abdul Hussain1 , Dr. Govind Ukhandrao2 1 (Asst. Prof. ENTC Deptt. Anuradha Engineering College, Chikhli, India) 2 (Principal Sharadchandra Pawar College of Engineering, Otur, India) Abstract : Moving object detection & tracking in real world scene is becoming significant problem in today’s era. The extensive study in this area is motivated by potential number of applications of object tracking. In this paper, we analyze a method for motion segmentation & tracking of nonstationary objects that uses the complete image information using the affine flow equations. These affine flow parameters are then combined with illumination insensitive template matching tracker to efficiently segment & track the moving objects. Experimental results performed on various database videos shows the method effectively tracks the object without miss detection of the target compared to other methods. Keywords: Affine Flow, Video Surveillance, Object Tracking, Template matching. I. Introduction Motion segmentation & tracking has been a crucial component for many computer vision applications such as structure from motion, objects tracking/recognition, advanced video editing and behaviour analysis. A variety of computational models, e.g. optical flow fields [7, 8], layers [1] and boundaries [9], have been developed to approximate the motion from a video sequence. It is significant to assess the performance of various motion analysis algorithms to gain insight and design better models. However, little has been done for evaluating motion analysis algorithms compared to the tremendous effort put into developing these algorithms. The extraction of the affine parameters has fundamentally two components: the recognition of a suitable set of spatial patches to characterize every surface in a scene; and the tracking of the patches during the frame sequence. The tracking of the patches (2-D) and assessment of their connected affine motion parameters is realized using weighted linear regression over an approximated optical flow field. The weights are provided by truncated Gaussian windows, each defining the spatial extent of one of the patches being tracked. The preference of a gaussian window function is deliberate: the group of such functions are closed under the action of affine transformation and hence naturally represent the evolution of the patches as they warp according to the affine motion approximation. The main contribution of this paper is the analysis of fast version of method for obtaining affine flow from the image sequence. These affine flow parameters are then combined with the illumination insensitive tracker to appropriately track the nonstationary objects. The motion evaluation is based on three observations. First, humans are experts at segmenting layers in a video sequence because human being can easily recognize the moving objects and their relative depth relationships. Second, humans are sensitive to any differences between two images when these two images are displayed back and forth. Third, humans have knowledge of the smoothness and discontinuities of the motion that a moving object undergoes. We thoroughly tested our method by conducting experiments on several videos collected for our experimentation purpose in both indoor and outdoor surroundings. The experimental results demonstrate that the method performs favorably well in real world surroundings. One of the advantages of our approach compared to the other state-of-the-art methods is that it reduces the number of false detections, as pixel-level comparison can be per-formed in regions with significant motion only. Another advantage is that the illumination insensitive template matching eliminates the effects of illumination changes which are a major problem in visual surveillance system. The paper is structured as follows. Section 2 explores some of the related methods. Section 3 presents our system methodology in detail. Section 4 describes the database collection for experimentation. Section 5 presents some experimental results & Finally, Section 6 concludes the paper. II. Literature Review A large number of methods are proposed for video surveillance system using human motion detection & tracking. Bramble [10] detects people in indoor environments and uses explicit 3-D cylindrical modelling of target objects. Haritaoglu et al. proposed a system which uses projection histograms of the single detected object as features to classify different actions [2]. Hager and Belhumeur [11] perform photometric stereo using at least three images of a scene under linearly independent lighting conditions.
  • 2. Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching DOI: 10.9790/0661-18112429 www.iosrjournals.org 25 | Page In [3] Hidden Markov Models are used. However, these methods are usually much slower and adaptation to changes of the scene is difficult. Hedvig [12] uses optical flow features for training directly. Support vector machine is used for training. This method can get acceptable detection result in outdoor scene. McKenna used a combination of motion, skin color and face detection to track people and their faces [6]. Karmann and Brandt [13] and Kilger [4] respectively proposed an adaptive background model based on Kalman filtering to adapt temporal changes of weather and lighting. Chen et al. [5] use a human attention model to create different cropped and scaled versions of images according to the screen size. An algorithm was presented in [14] to track moving objects in the scene by considering dynamic occlusions among moving objects; this is especially useful to track pedestrians moving in indoor environments. Robust detection of illumination changes in outdoor video surveillance images was effectively achieved in [15] by using an adaptive algorithm. This approach dynamically learns and updates the background by using a segmentation algorithm, especially for sudden changes. Since successful tracking depends a lot on precise object detection, the evaluation of object detection algorithms within a surveillance system plays a significant part in overall performance study of the whole system. III. system framework Proposed analysis is directed to build a system that is able to detect nonstationary/mobile objects in the scene and track those using affine parameters & illumination insensitive tracker. Figure 1 below represents a block diagram of the system. Fig. 1: Framework for object detection & tracking Affine flow model combined with illumination insensitive tracker overcomes the problem of traditionally used optical flow technique like motion changes due to camera movement, dynamic background objects (e.g. tree leave or branches), and changes in the background geometry (e.g. parked cars). For the surveillance of outdoor scenes, sturdiness against noise and adaptivity to illumination changes are also essential. We coupled the method of affine flow extraction with template matching algorithm with illumination intensity minimization model which will take care of illumination changes in real images caused due to automatic exposure alterations of the camera, adjustment of light cause irradiance, appearance of shadows or movement of the tracked objects. Our objective is to provide a robust and improved method to find the moving objects in the video frame as well as to track them. The proposed method is effective in reducing the number of false alarms that may be triggered by a number of reasons. Fig.1shows the different stages involved in the proposed system. First, image frames are taken out from video, and then, affine flow parameters are extracted from an image. Simultaneously, illumination insensitivity hyperplane method for template matching is also applied to perfectly track the object. 3.1 Affine Flow The starting point is the assumption of brightness constancy, which assumes that the brightness structures of local-time varying image regions are unchanging under motion for a short period of time. Formally, this is defined as, 𝐼 𝑥, 𝑦, 𝑡 = 𝐼(𝑥 − 𝑢, 𝑦 − 𝑣, 𝑡 − 1) … (1) Where (𝑥, 𝑦)𝑡 represents image position in pixel coordinates, t represents the temporal coordinate, (𝑢, 𝑣)𝑡 represents the motion at image position (𝑥, 𝑦)𝑡 over the time 𝑡 + 1 and 𝐼(𝑥, 𝑦, 𝑡) represents the image brightness function. Using the least-squares criteria, we seek the motion that minimizes the error 𝜖 over a region 𝑅 of the image, formally, ∈ = [ 𝑥, 𝑦, 𝑡 − 𝐼 𝑥 − 𝑢, 𝑦 − 𝑣, 𝑡 − 1 ]2 𝑥,𝑡,∈𝑅 … (2) Video I/P Frame Extraction Affine Parameters Adjust Template (βt+1, ϒt+1) Compute µ Track Moving Object
  • 3. Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching DOI: 10.9790/0661-18112429 www.iosrjournals.org 26 | Page Assumint that 𝐼(𝑥, 𝑦, 𝑡) is approximately locally linear (2) is simplified by taking a Taylor series expansion and omitting terms higher than first order. ∈ = [𝐼 𝑥, 𝑦, 𝑡 − (𝐼 𝑥, 𝑦, 𝑡) − 𝐼𝑥 𝑥, 𝑦, 𝑡 𝑢 − 𝐼𝑦 𝑥, 𝑦, 𝑡 𝑣 − 𝐼𝑡(𝑥, 𝑦, 𝑡) ]2 𝑥,𝑦,∈𝑅 … (3) ∈ = [𝐼𝑥 𝑥, 𝑦, 𝑡 𝑢 + 𝐼𝑦 𝑥, 𝑦, 𝑡)𝑣 + 𝐼𝑡 𝑥, 𝑦, 𝑡 ]2 𝑥,𝑦,∈𝑅 … (4) Where, 𝐼𝑥 𝐼𝑦 and 𝐼𝑡 represent the partial derivatives of the intensity function with respect to the spatial parameters 𝑥, 𝑦 and temporal parameter 𝑡, respectively. To obtain the estimate of velocity in the pure translation case, the partial derivatives of the error in (4) are taken with respect to the translational velocity (𝑢, 𝑣) 𝑇 and set to zero. This yields the following set of linear equations, 𝐼𝑥 2 𝐼𝑥 𝐼𝑦 𝐼𝑥 𝐼𝑦 𝐼𝑦 2 𝑢 𝑣 = − 𝐼𝑥 𝐼𝑡 𝐼𝑦 𝐼𝑡 … (5) A minimum of two points are required for the full solution. The under constrained problem given a single point is commonly referred to in the literature as the aperture problem. In the case where the motion is modelled locally by an affine transformation, corresponding to a composition of rotation, dilation, shear and translation, (𝑢, 𝑣) 𝑇 are defined as follows, 𝑢 𝑣 = 𝑎1 𝑎2 𝑎41 𝑎5 𝑥 𝑦 + 𝑎0 𝑎3 … (6) As in the case of translational motion, to obtain the estimates of the affine parameter (4) is differential w.r.t the six unknown affine parameters (𝑎𝑖 𝑤𝑕𝑒𝑟𝑒 𝑖 = 0, … . , 5) and set to zero. 3.2 Template Matching For eradicating the effects of illumination changes, a technique is to estimate illumination compensation parameters for the current template, in order to adjust its gray-level values with respect to the reference template. Using the linear model 𝐹𝑛𝑒𝑤 𝑥 = 𝛽𝑓 𝑥 + 𝛾 ∀ × 𝜀 𝑟 … (7) Contrast and brightness variations are represented by illumination compensation parameters 𝛽 and 𝛾. Least-squares minimization problem is solved when corresponding points of the initial template 𝑓(𝑟, 𝑡0) and the current template 𝑓(𝑔(𝑟, 𝜇 (𝑡))) are known. 𝛽𝑡, 𝛾𝑡 = 𝑎𝑟𝑔𝑚𝑖𝑛 [𝛽𝑓𝑔 𝑥, 𝜇 𝑡 , 𝑡 + 𝛾 − 𝑓 𝑥, 𝑡0 ]𝑥∈𝑟 … (8) We replace 𝑔(𝑥, 𝜇(𝑡)) with 𝑥(𝑡). Differentiating equation (8) with respect to the motion compensation parameters yields the linear system 𝑓2 𝑥 𝑡 ,𝑡 𝑓 𝑥 𝑡 ,𝑡 𝑓 𝑥 𝑡 ,𝑡 1𝑥∈𝑟 𝛽 𝑡 𝛾 𝑡 = 𝑓2 𝑥,𝑡0 𝑓 𝑥 𝑡 ,𝑡 𝑓 𝑥 𝑡 ,𝑡0 𝑥∈𝑟 … (9) To avoid time consuming operation of computing the matrix on the left, we exchange the reference template with the current template and hence obtain the motion compensation parameters and for adapting the reference template to the current template. Consequently, the matrix in equation (9) has to be computed only once. As we still want to adapt the current template, we revert to the original illumination compensation parameters. 3.3 Centroid Estimation A set of two integers, 𝐶 (𝐶𝑥, 𝐶𝑦) which determines the position of the moving object in the given scene. The centroid coordinates are calculated by: 𝐶𝑥 = 𝐶𝑥 + 𝑥 … (10) 𝐶𝑦 = 𝐶𝑦 + 𝑦 … (11) 𝑇 = 𝑇 + 1 … (12) For each pixel, where (𝑥, 𝑦) is the current pixel location. The resulting centroid is then divided by the total value: 𝐶𝑥 = 𝐶𝑥 / 𝑇 … (13) 𝐶𝑦 = 𝐶𝑦 / 𝑇 … (14) This centroid positional information is then transferred to the object tracking module which is coupled with the illumination insensitive template matching model which gives the good result for detection & tracking. IV. Dataset For the evaluation of our experiments we used two different data bases. First, Weizmann database which is publically available and another is our own created database, this database contains 86 sequences
  • 4. Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching DOI: 10.9790/0661-18112429 www.iosrjournals.org 27 | Page having five & four of classes of actions (Jump, Run, Walk, Side, Handwaving, Jogging) respectively performed by 19 different subjects in two different conditions d1-d2. d1: Indoor Environment; d2: Outdoor (Playground) + illumination variations Few samples of Indoor environment Outdoor (Playground) + illumination variations Environment „Running‟ „Jogging‟ „Walking‟ „Jumping‟ „Handwaving‟ “Self Database” Images „Bend‟ „Jack‟ „Pjump‟ „Run‟ „Side‟ „Skip‟ “Weizmann Database” Images Fig. 2: Sample images of Self & Weizmann database V. Experimental result This section summarizes the experimental results for motion detection & tracking system. The experiments were performed on MatLab 7.10.0 software. The PC was equipped with Pentium corei3, 2.13 GHz processor & 4 GB RAM Memory. The algorithm was implemented on two different datasets i.e. Weizmann publically available & self made database. Weizmann database [180*144, Frame rate 25frames/sec] has different actions Jump, Run, Walk, Bend, Skip and Side. All the actions were captured with steady camera. Our own data-base [1280*720, Frame rate 29frames/sec] was captured when both camera & target are moving in dynamic environment.
  • 5. Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching DOI: 10.9790/0661-18112429 www.iosrjournals.org 28 | Page Proposed Model Without Illumination insensitive tracker Fig. 3: Experimental Results on database videos Figure 3 shows the comparison result of the proposed tracker system with the motion detection & tracking system of affine flow without illumination insensitive template tracker. As it can be observed from the figure 3 that, proposed model tracks the nonstationary object precisely as compared to the other state of the art method. The algorithm implemented has achieved highly robust & clean results in cluttered backgrounds and illumination change. This shows the effectiveness of algorithm. Table 1 below shows the variation in the pyramid level of affine flow calculations for the 3 number of iterations. From the value represented in the table it can be seen that as the pyramid level increases the value of affine flow in between the pixel value decreases. Table 1: Norm of optical flow values for 3 numbers of iterations Iteration 1 Iteration 2 Iteration 3 Pyramid 1 318.47 311.17 309.05 Pyramid 2 191.28 190.77 189.08 Pyramid 3 104.53 101.16 99.82 Pyramid 4 47.29 43.43 41.96 Pyramid 5 23.48 21.43 21.10 Pyramid 6 10.86 9.87 9.65 Fig. 4: Error Estimation of targets tracked by proposed method & affine flow without template tracker method. Figure 4 displays the results in terms of average tracking errors. Proposed method shows the error maximum upto 3 pixels where as the method of the affine flow without illumination insensitive template matching tracker have error upto 7 pixels. This error comparison shows that the proposed system is very much precise in the motion segmentation and tracking. VI. Conclusion In this paper, we have presented a new proposal for object detection & tracking based using the combination of affine flow extraction & illumination insensitive template matching model. The system uses
  • 6. Competent Tracking of Moving Object Using Affine & Illumination Insensitive Template Matching DOI: 10.9790/0661-18112429 www.iosrjournals.org 29 | Page efficient technique of affine flow extraction which removes the drawback of traditionally used optical flow technique like detecting non-stationary background objects and shadows cast by moving objects. Further module of template matching eliminates the problem of illumination & dynamic environment conditions. The integrated system was successfully tested on a number of sample videos containing various moving entities. Analysis was made on its performance. Experimental results shows that this algorithm has better tracking precision and can resolve the robust tracking problem in illumination changes & dynamic background, so it has a very wide range of applications. Further work will concentrate on occlusion handling & multiple object tracking problems that will work in more complex cluttered environments. We will also focus on increasing the performance of the system by decreasing the time duration from frame to frame processing. References Journal Papers: [1] J. Y. A. Wang and E. H. Adelson, Representing moving images with layers, IEEE Trans. Image Processing, , 1994, 3(5):625 – 638. [2] I.Haritaoglu, D.Harwood and L.S.Davis, W4: Realtime surveillance of people and their actions, IEEE Trans. on Pattern Anal. and Mach. Intell. , August 2000, 22(8),pp.809-830. [3] J. Kato, S. Joga, J. Rittscher, and A. Blake, An HMM-based segmentation method for traffic monitoring movies, IEEE Trans. on PAMI, , 2002, vol. 24, no. 9, pp. 1291–1296. [4] M. Kilger, A shadow handler in a video-based real-time traffic monitoring system, Proc. of IEEE Workshop on Applications of Computer Vision. 1992, pp. 1060-1066. [5] L.-Q. Chen, X. Xie, X. Fan, W.-Y. Ma, H.-J. Zhang, and H.-Q. Zhou, A visual attention model for adapting images on small displays, Multimedia Syst., Oct. 2003. vol. 9, pp. 353–364. Chapters in Books: [6] McKenna, Stephen J., Tracking groups of people Computer Vision and Image Understanding 80.1 (2000): 42-56. Proceedings Papers: [7] B. K. P. Horn and B. G. Schunck, Determing optical flow, Artificial Intelligence, 17:185–203, 1981. [8] B. Lucas and T. Kanade, An iterative image registration technique with an application to stereo vision, In Proc. of the Intl. Joint Conf. on Artificial Intelligence, pages 674–679, 1981. [9] C. Liu, W. T. Freeman, and E. H. Adelson, Analysis of contour motions, In NIPS, 2006. [10] Isard, Michael, and John MacCormick, BraMBLe: A Bayesian multiple-blob tracker, Computer Vision, 2001. ICCV 2001. Proceedings. Eighth IEEE International Conference on. Vol. 2. IEEE, 2001. [11] G.D. Hager and P.N. Belhumeur, Real-time tracking of image regions with changes in geometry and illumination, Proc. IEEE Computer Society Conference on Computer Vision and Pattern Recognition: 403 – 410, 1996. [12] Hedvig Sidenbladh, Detecting Human Motion with Support Vector Machines, ICPR 2004. [13] K.P. Karmann, A. Brandt, Moving object recognition using an adaptive background memory, In Cappellini, Time-varying Image Processing and Moving Object Recognition, Elsevier, Amsterdam, The Netherlands, 1990. [14] Marchesotti, L. Marcenaro, L. Regazzoni C. S, Tracking and counting multiple interacting pedestrian in indoor scenes, In Proceedings of third IEEE international workshop on performance evaluation of tracking and surveillance, Copenhagen Denmark, June 1, 2002. [15] Regazzoni, C, Adaptive change detection approach for object detection in outdoor scenes under variable speed illumination changes, In Proceedings of Eusipco 2000.