SlideShare a Scribd company logo
1 of 7
Download to read offline
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.1864-1870

RESEARCH ARTICLE

www.ijera.com

OPEN ACCESS

A Novel Algorithm To Detect Changes In Surveillance Video
Haitham Kareem Ali (**), Yasir Khalil Ibrahim(*),
(*) Faculty of Engineering, Jarash University, Jarash, Jordan.
(**)Slemani Polytechnic University, Slemani College of Technical Engineering, Communication & Electronics
Department, Slemani, Kurdistan Region , Iraq.
Abstract
Tracking changes in an acquired surveillance video is essentially impractical task for the human due to the
complexity of the monitoring process, specifically for a long-time period. Finding changes in a sequence of images
captured by surveillance video system is a complicated task for the human and it is complicated by fitting it to a
classification model. Given a sequence of images, we are concerned with automatically detecting the occurrence of
changes in the given sequence of images. In this paper, the problem of detecting changes in digital images captured
by surveillance video system is explored. A novel algorithm, involving correlation coefficients and eigenstructurebased Mosaab distance measure, is suggested. The proposed algorithm is a mobile-based application, that can be
used to process pixel-based data (i.e., sequence of images) at camera locations and send only real-time numeric
results to client's mobile device(s) . Therefore, the aim of proposed algorithm is to analyize and eliminate the
amount of pixel-based data to be sent via the network. Consequently, the received real time numeric results can be
visualized using an application software installed in client's mobile device. The main function of the mobile-based
application software is to illustrate the received results as a real-time graph. The proposed algorithm shows
effectiveness in eliminating the flow of raw pixel-based data in the network.
Key words:
Mosaab-distance measure, correlation coefficient, surveillance video system.

I.

INTRODUCTION

Normally, the surveillance video system
produces huge amount of data that requires analysis.
Usually, the main analysis task, which is monitoring
changes in a sequence of images (i.e., real time
analysis), is tackled by human. Tracking changes in
the acquired video is essentially impractical task for
the human since it is a long-term monitoring process.
In any scene, there is a number of changes that take
place over the time. We assume that the input images
of a scene are captured by surveillance video system
with timing setup as a discrete variable. An entity can
be either an object or human (i.e., individual), in this
context, if an entity is not a human, then it is
considered as an object. An activity is defined as an
action that can happen to an entity at a particular point
in time. When an activity is performed on an entity it
becomes an event. Thus, implicitly, we are looking
for the occurrences of an event or events in a
particular scene. For example, remove is an activity,
when it is performed on a book (i.e., entity) it
become an event, which is: the book is removed
from the table (i.e., removed from a particular
scene). The state of any scene is considered as a
function of time. During any period of time, a scene
is said to be static if there is zero event occurred,
otherwise a scene is said to be dynamic. Finding
changes sequentially in a sequence of images
captured by surveillance video system is
complicated task for the human and it is
complicated by fitting it to a classification model.
Our work focuses exclusively on the following:
www.ijera.com

Given a sequence of images captured for a particular
scene, we are concerned with autodetecting the
occurrence of changes in the given sequence of images
and the change occurs when one activity or a set of
activates are performed on an entity or a set of entities
in a particular scene, under the assumptions: (i) lack of
information about those activities (i.e., no prior
knowledge or insufficient labeled datasets that can be
used for training) and (ii) the sets of activities and
entities are previously unknown. Moreover, the
detection possess must be performed locally at camera
locations, to eliminate the amount of pixel-based data to
be sent via the network to mobile device(s) of a client.
In this study, we adding another task to the
surveillance system, which is the analysis of a
sequence of images captured by surveillance system
installed in highly sensitive areas (e.g., offices of
V.I.P.s). The advantage behind adding this task is to
detect activities (i.e., changes) in a sequence of images
locally at camera locations and instead of transmitting
huge amount of pixel-based data to the mobile
device(s) of a client via the network, the system sends
only the activity detection results. In this way, we can
eliminate the amount of pixel-based data to be
transmitted from camera locations to the mobile
device(s) of a client via the network. Sending only few
real numbers (i.e., the detection results) to the mobile
device (s) of a client via the network will minimize the
potential threats that result from sending pixel-based
data to the mobile device(s) of a client via the
network.

1864 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05
In this paper, we briefly survey existing
approaches to the problem, outline the method we have
implemented and present our results and ideas for future
work. The structure of this study is summarized as
follows. In Section 2, which it consists of three
subsections, we present materials and methods. In
subsection 2.1, we present related work. In subsection
2.2, we present the statement of the problem. In
subsection 2.2, we present the propose algorithm. In
Section 3, we present the experimentation and results.
Finally, conclusions are presented in Section 4.

II.

MATERIALS AND METHODS

2.1 Related work: The main stages of any pattern
recognition system are: (i) feature extraction stage and
(ii) classification stage. There two main types of
classification: Either (i) supervised classification or (ii)
unsupervised classification Jain (2000). Extracting
information from data is an essential step for pattern
recognition systems and the validity of the extracted
information has an impact on the classification stage.
Machine detection, thus, can be used to validate the
human detection in order to increase accuracy.
Replacing eye-based detection by machine-based
detection will often be the first step in building an
auto-detection system that has the capability to
analyze a sequence of images captured by a
surveillance system.
As we mentioned in the previous section, there
are two components that arise in this context: (i) a set of
activities and (ii) a set of entities. The composition of
any dynamic scene is result from the interaction
between those two components (i.e., these two sets). An
event is an activity that happens to an entity or set of
entities at a particular point in time. An event is
categorized as either normal event or abnormal event
(i.e., anomaly). An extensive work has been done in the
area of analyzing video sequences. However, there
exists a few challenges that must be considered, for
example: Detecting activities in surveillance video
sequences at camera locations and send only numeric
results as a real-time chart to the mobile device(s) of a
client.
Detecting activities in surveillance video
sequences is well known problem in several research
areas and it is under the focus of different research
communities such as: Pattern Recognition, Computer
Vision, Artificial Intelligence and Soft Computing,
Pattern Analysis and Machine Intelligence and
Computational Intelligence and Data Mining.
Precisely, activities in a scene are classified as either
(i) normal activates or (i) abnormal activities. They are
two approaches that can be used to solve this problem:
Either (i) supervised classification or (ii) unsupervised
classification. The data availability for the considered
activities is the key point in deciding which approach
should be used to solve this problem. Since we are

www.ijera.com

www.ijera.com

classifying each activity into two classes (i.e., either
change class or no-change class), then a distancebased classifiers are required. Activities are related to
an entity (e.g., human(s), object(s)) and either
performed by one entity or by a group of entities (e.g.,
motion of one person or group of people in a scene,
moving object from one position to another).
We just presented the contributions achieved
by researchers in the field in the recent years,
however, the goal of this paper, which we stated in the
previous section (i.e., paper statement) is differ from
the goals of the contributions that we just reviewed in
this section. In this study, the main challenge to
overcome is to detect changes defined by a set of
activities, those activities are either associated with an
entity or a set of entities (e.g., object(s) or human (s))
in a scene, under the following assumptions: (i) lack of
information about those activities (i.e., no prior
knowledge or insufficient labeled datasets that can be
used for training), (iii) the sets of activities and entities
are previously unknown and (iii) to eliminate the
amount of pixel-based data to be transmitted from
camera locations to the mobile device(s) of a client via
the network. In this section, we presented the related
work. In the next section, we present the statement of
the problem.
2.2 The statement of the problem: The problem
under investigation is a well known problem in image
analysis area. It is the problem of detecting changes in
a sequence of images acquired by a surveillance
system. Mathematically speaking, suppose that we
have a surveillance system that produces a huge
amount of data, where the data are represented by
images. Without loss of generality, suppose that the
system produces images captured for a highly
sensitive site and the problems that arise in this
context is to monitor those images and track the
changes by human efficiently for a long-term period
and send results to mobile device(s) of a client. The
solution for this problem can be achieved by replacing
the human-based monitoring by machine-based
monitoring and thus a pattern recognition system is
required since the changes in those images are
considered as patterns. The core of any pattern
recognition system is the recognizer or the algorithm
that can used to recognize those changes over the time.
The time is discrete since those images captured at
discrete points in time: t1, t2, t3, ..., tk. Hence, we
assume that we have a sequence of k images: Image1,
Image2,..., Imagek captured at t1, t2, t3,..., tk
respectively. The data under consideration are images
or pixel-based data and mapping these images from
the data space into the real-valued feature space is an
essential step to reduce the mount of input data to be
processed by the pattern recognition system in order to
detect changes in these images sequentially.

1865 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05

www.ijera.com

Fig. 1: The statement of the problem

Fig. 2: The proposed decomposition process performed on two successive images captured by a surveillance
system
Instead of detecting changes in these images
embedded in the data space, which can be achieved
by measuring the distance between successive images
in the sequence, detecting changes is performed in the
feature space by measuring the distance between the
representation of successive images embedded in the
feature space (Fig. 1).
In this section, we presented the statement of
the problem. In the next section, we present the
proposed algorithm.
2.3 The proposed algorithm: In this paper, as we
mentioned, our goal is to detect changes in a
sequence of images Image1, Image2,..., Imagek of size
m × n captured by surveillance system(Fig 2). Those
www.ijera.com

changes are defined by a set of activities, those
activities are either associated with an entity or a set
of entities (e.g., object(s) or human(s)) in a scene,
under the following assumptions: (i) lack of
information about those activities (i.e., no prior
knowledge or insufficient labeled datasets that can be
used for training) and (ii) the sets of activities and
entities are previously unknown. As we know,
algorithms for monitoring changes in a sequence of
images is mainly based on global features of images
(e.g., global correlation between successive images)
often yield many cases of false detection. In addition,
the performance of threshold value-based algorithms
is based on the accuracy of estimating the threshold
values and changing those parameters (i.e., threshold
1866 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05
values) to decrease the rate of false detection may
increases the rate of false detection.
For the above reasons, we build our
algorithm (feature extractor and distance-based
classifier) by consider the following concepts:(i)
statistical-based local features (e.g., the correlation
coefficient between partitions of the image), (ii)
eigenstructure based Mosaab distance measure (it is
an inverse-free ℜp×p-based distance measure that can
works with singular matrices and it require less
computation compared with Mahalanobis distance),
(iii) high dimensional-based representation (e.g., a
representation embedded in ℜp×p). Precisely, the
proposed classifier is a distance-based classifier
which uses the statistical-based local features to
perform the required task.
As we mentioned earlier in this paper, the
data under consideration are images, in fact, it is
pixel-based data, which is a huge data. To reduce the
amount of data to be processed, it is necessary to map
those images from the data space (i.e., pixel-based) to
the real-value feature space (i.e., real-valued based)
Thus, we have to define a new representation to those
images under consideration. The representation must
be embedded in ℜp×p since eigenstructure-based
Mosaab distance measure is considered. The working
mechanism of the proposed algorithm can be
summarized follows. First, we partitioned each image
into four same size sub-images. Then, the correlation
coefficients between those sub-images are computed
(i.e., 6 correlation coefficients).
Let (i)  corr (Imagei, j Imageir ) (where, j = 1,
jr
2,..., 4 and r = j + 1, 2,..., 4) be the correlation
coefficient between the image partitions (i.e., subimages) Imageij and Imageir of Imagei. For each
Imagei (i = 1, 2,…,k), the algorithm computes
(i)
(i)
(i)
(i)
(i)
11 , 12 , 13 , 14 , (i) ,and 34 , where those coefficient
23
are real values between 0 and 1.
Daoud and Kremer, (Daoud and Kremer,
2009; Daoud, 2010; Daoud and Kremer, 2006)
proposed a new measure that is defined in terms of
variations to capture the deviations between two sets
of biosequences or between a pair of sequences. The
metric is built using the objective function:
(1)
Maximizing the above objective function
leads to the following metric:
(2)
where, 1 is the largest eigenvector associated with 1,
the largest eigenvalue of the matrix (Ωi-Ωj) and:
Var(  'X(k) )  E(  'X(k) X(k) ' )
  'E(X(k) X(k) ' ) 
  ' i  for k  i, j

www.ijera.com

(3)

www.ijera.com

Definition 1: Mosaab distance measure (Daoud and
Kremer, 2009; Daoud, 2010; Daoud and Kremer, 2006)
Dij (1):ℜp×p×ℜp×p ⇒ ℜ+ is a variance-covariance
structure-based distance measure that measures the
distance between two variance-covariance structures
represented by the variance-covariance matrices Ωi and
Ωj.
The proposed distance measure can measure
the deviation in variations of the same feature vector
being measured on two datasets (e.g. objects, images,
sub-images, sequences, set of sequences etc...). The
deviation in variations is explained by the eigenstructure of the difference between two variance
covariance structures, where each variance covariance
structure represents a dataset(e.g. object, image, subimage, sequence, set of sequence etc...). The eigenstructure is defined in terms of the largest eigen-value
(i.e. max, see equations 1-2) of the difference between
two variance-covariance structures.
It is clear that Mosaab distance measure is a
variance-covariance structure-based distance measure
and it is a matrix-based distance measure (i.e.,
ℜp×p×ℜp×p ⇒ ℜ+). Thus the representation of images
must be embedded in ℜp×p (i.e., feature space). Since
for each Imagei (i = 1, 2,…. k), the representation
(i)  (where j = 1, 2,..., 4 and r = j + 1, 2,..., 4),
jr
thus, for each Imagei (i = 1, 2,…, k) we construct a
diagonal
matrix
(i)
(i)
(i)
(i)
i  diag(11 , 12 , 13 , 14 , (i) , (i) and (i) 66 ) . Since
23
24
34
the matrix Σi is neither variance-covariance matrix
nor correlation matrix, we redefine Mosaab distance
measure as follows.
Definition 2: The modified form of Mosaab distance
measure i (,): ℜp×p×ℜp×p ⇒ ℜ+ is defined as:
'
i (i , (i 1) ) | 1 (i  (i 1) )1 || 1 | 0

(4)

A distance measure that measures the
distance between two matrices Σi and Σ(i+1) that
represent Imagei and Image(i+1) respectively, where 1
is the eigenvector associated with the largest
eigenvalues 1 of the matrix (Σi-Σ(i+1)) and:
(i)
(i)
(i)
i  diag (11 , 12 ,..., 34 )
(i)
11
0

(i)
0 12

 M M

0
 0


L
L
O
K

0 

0 
M

(i)
34 


(5)

Finally, the proposed algorithm (see
Algorithm 1) produces a sequence of eigenvalues 1,
2,..., (k-1) that represent the sequential changes in
Image1, Image2,..., Imagek respectively. In this
section, we presented a novel algorithm to detect the
changes in a sequence of images captured by a
surveillance system. In the next section, we present
the experimentation and results.
1867 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05
III.

RESULTS and DISCUSSION

In this section, we used a sequence of 100
images captured by a surveillance system to evaluate
the performance of the proposed algorithm and a
discussion is given below. A sample output from
Matlab code, which shows the partition process, is
given in Fig. 3.
The results of the proposed algorithm are
shown in Fig. 8. In Fig. 8, the horizontal axis
represents the index of the distance value i and the
vertical axis represents the value of the modified
Mosaab distance measure. The solid points of the red
curve represent the distance values measured in the
feature space between successive images in the
sequence of 100 images. There are few peak values
clearly appeared at i = 3, 8, 12, 31, 36, 56, 57, 84, 86
and 89. The distance values are: 3 = 0.371, 8 =
0.8618, 12 = 0.3167, 31 = 0.3199, 36 = 0.3084, 56
= 0.3494, 57 = 0.4578, 84 = 0.3572, 86 = 0.5412,
and 89 = 0.3488.
Algorithm 1: Detecting Changes Sequentially in a
Sequence of Images
input : Given a sequence of images Image1,
Image2,...,
Imagek of size m × n captured by surveillance system.
output: A sequence of eigenvalues 1, 2,...,  (k-1)
that represent the sequential changes in Image1,
Image2,...,
Imagek.
1 for i ← 1 to k do
2 Partition Imagei into four same size sub-images
(i.e. four disjoint partitions). Let Imagei1, Imagei2,
Imagei3 and Imagei4 be the four partitions of Imagei
respectively.
Partition Imagei+1 into four same size sub-images
(i.e. four disjoint partitions). Let Imagei1, Image(i+1)2,
Image(i+1)3 and Image(i+1)4 be the four partitions of
Imagei+1 respectively. Let corr(,) be the correlation
coefficient.
3 for j ← 1 to 4 do
4 for r ← j + 1 to 4 do
5 Compute corr(Imageij ; Imageir)
6 Compute corr(Image(i+1)j ; Image(i+1)r)
end
end
9 Construct the diagonal matrix
Σi = diag( corr(Imagei1; Imagei2),
corr (Imagei1;Imagei3),
corr(Imagei1, Imagei4),
corr (Imagei2; Imagei3),
corr(Imagei2, Imagei4),
corr (Imagei3; Imagei4))
10 Construct the diagonal matrix
Σ(i+1) = diag(corr(Image(i+1)1; Image(i+1)2),
corr(Image(i+1)1; Image(i+1)3),
corr(Image(i+1)1; Image(i+1)4),
corr(Image(i+1)2; Image(i+1)3),
corr(Image(i+1)2; Image(i+1)4),
corr(Image(i+1)3; Image(i+1)4))
www.ijera.com

γ

www.ijera.com

11 Compute λi = Abs(γ' (Σi - Σ(i+1)) γ), where
is the largest eigenvector of the matrix
(Σi-Σ(i+1)).

end
For example, the distance value 3 = 0.371 is
estimated from comparing the images 3 and 4, and
the distance value 8 = 0.8618 is estimated from
comparing the images 8 and 9 (see Figure 4).
Moreover, the distance value 84 = 0.3572 is
estimated from comparing the images 84 and 85, and
the distance value 89 = 0.3488 is estimated from
comparing the images 89 and 90 (see Fig. 5). From
Figures 4 and 5 , our conclusions (i.e. human eyes)
coincided with the results of the proposed algorithm,
but for a long-term period, it is harder for the human
eyes to recognize such changes (i.e., in case of longterm monitoring process). However, there are other
cases where the change is unrecognizable by the
human eyes due to the speed of change or due to the
nature of change (e.g. tiny changes), but it is easily
can be recognized by the proposed algorithm. Those
cases are illustrated in Figure 6. Consider the four
successive images: 85, 86, 87 and 88. The discussions
of those cases are given below.

Fig. 3:

A sample output from matlab code shows
the partition process

Fig. 4: Peak values 3 = 0.371 (the corresponding
images: 3 and 4) 8 = 0.8618 (the corresponding
images: 8 and 9)
1868 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05

Fig. 5: Peak values 84 = 0.3572 (the corresponding
images: 84 and 85) and 89 = 0.3488 (the
corresponding images: 89 and 90)

Fig. 7:

www.ijera.com

Two images represent: (i) Light switched
ON, and (ii) Light switched OFF

Case 1: The main difference between image 85 and
image 86 is the position of object1. Object1 moved
from one position to another. This change is
efficiently detected by the proposed algorithm and the
distance value is 85 = 0.1835 (Fig. 6).
Case 2: The main differences between image 86 and
image 87 are: (i) object1 removed, (ii) object2 moved
from one position to another, (iiii) object3 removed and
(iv) objects4 added to the scene (Fig. 6). Those changes
are efficiently detected by the proposed algorithm and
the distance value is 86 = 0.5412. Since the number of
changes occurred in this case is more than the number
of changes occurred in case 1, then the distance value
86 is greater than the distance value 85.
Fig. 6:

Four successive images captured by the
surveillance system

Fig. 8: The results of the proposed algorithm used for testing a sequence of 100 images captured by surveillance
system
Case 3: There is no main difference between image
87 and image 88, thus, the distance between the two
images has the value 87 = 0.0866, which is a very
small value compared with distance values 85 and
86 (Fig. 6). There are two points arise here: (i) the
quality of the digital camera used in this experiment
www.ijera.com

and (ii) the small hidden changes that are
undetectable by human eyes which they may cause
this problem (i.e., the distance value is not equal to
0).
There is another problem that we should
mention in this section, it is the visibility problem. The
1869 | P a g e
Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application
ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05

www.ijera.com

problem is illustrated in Fig. 7. When the lights are
switched off the visibility is almost zero. However, this
case is perfectly detected by the proposed algorithm as
a change in lights intensity, but probably the scene
details remain unchanged. In this case, advanced
equipments are needed to handle such problem, which
is beyond the scope of this paper. Finally, in this
section, we presented the experimentation and results
of the proposed algorithm. In the next section,
conclusions are presented.

IV.

CONCLUSION

While the results of the proposed algorithm
were generally quite good, there are several potential
areas for future improvement: instead of performing
4-subimages decomposition process on images, one
may perform 8-subimages decomposition process on
images which might be beneficial. In particular, we
concluded that the proposed algorithm is useful for
detecting changes in a sequence of images locally at
camera location and send real-time numeric results to
mobile device(s). In other words, the proposed
algorithm showed effectiveness in eliminating the
flow of raw pixel-based data in the network and
minimizing the potential threats that result from
sending highly sensitive pixel-based data to the
mobile device(s).

REFERENCES
[1]

[2]

[3]

[4]

Daoud and Kremer 2006. Daoud, M. and
S.C. Kremer. Neural and statistical
classification to families of bio-sequences.
International Joint Conference on Neural
Networks, 2006 (IJCNN ’06), pp: 699-704,
2006.http://dx.doi.org/10.1109/IJCNN.2006.
246752
Daoud and Kremer 2009. Daoud, M. and S.C.
Kremer. A new distance distribution
paradigm to detect the variability of the
influenza-A virus in high dimensional
spaces, Applications of Machine Learning in
Bioinformatics
Workshop,
IEEE
International Conference on Bioinformatics
& Biomedicine (IEEE BIBM 09), November
1-4, 2009, Washington D.C.
Daoud, 2010. Daoud, M.. A new variancecovariance structure-based statistical pattern
recognition system for solving the sequenceset proximity problem under the homologyfree assumption. PhD thesis, 2010, ISBN:
9780494678268.
Jain, 2000. Jain, Duin and Mao Anil K.
Jain, Robert P. W. Duin and Jianchang Mao.
Statistical pattern recognition: A review.
IEEE Transactions Pattern Analysis
Machine
Intell.,
22:
4-37,
2000.
http://doi.ieeecomputersociety.org/10.1109/3
4.824819

www.ijera.com

1870 | P a g e

More Related Content

What's hot

IRJET- Surveillance of Object Motion Detection and Caution System using B...
IRJET-  	  Surveillance of Object Motion Detection and Caution System using B...IRJET-  	  Surveillance of Object Motion Detection and Caution System using B...
IRJET- Surveillance of Object Motion Detection and Caution System using B...IRJET Journal
 
An Approach for Copy-Move Attack Detection and Transformation Recovery
An Approach for Copy-Move Attack Detection and Transformation RecoveryAn Approach for Copy-Move Attack Detection and Transformation Recovery
An Approach for Copy-Move Attack Detection and Transformation RecoveryIRJET Journal
 
IRJET- Criminal Recognization in CCTV Surveillance Video
IRJET-  	  Criminal Recognization in CCTV Surveillance VideoIRJET-  	  Criminal Recognization in CCTV Surveillance Video
IRJET- Criminal Recognization in CCTV Surveillance VideoIRJET Journal
 
Framework for Contextual Outlier Identification using Multivariate Analysis a...
Framework for Contextual Outlier Identification using Multivariate Analysis a...Framework for Contextual Outlier Identification using Multivariate Analysis a...
Framework for Contextual Outlier Identification using Multivariate Analysis a...IJECEIAES
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueIRJET Journal
 
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...ijtsrd
 
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)csandit
 
Motion detection system
Motion detection systemMotion detection system
Motion detection systemWritingHubUK
 
Smart web cam motion detection
Smart web cam motion detectionSmart web cam motion detection
Smart web cam motion detectionAnu Mathew
 
Deep-learning based single object tracker for night surveillance
Deep-learning based single object tracker for night surveillance  Deep-learning based single object tracker for night surveillance
Deep-learning based single object tracker for night surveillance IJECEIAES
 
IRJET- A Review on Moving Object Detection in Video Forensics
IRJET- A Review on Moving Object Detection in Video Forensics IRJET- A Review on Moving Object Detection in Video Forensics
IRJET- A Review on Moving Object Detection in Video Forensics IRJET Journal
 
IRJET - An Robust and Dynamic Fire Detection Method using Convolutional N...
IRJET -  	  An Robust and Dynamic Fire Detection Method using Convolutional N...IRJET -  	  An Robust and Dynamic Fire Detection Method using Convolutional N...
IRJET - An Robust and Dynamic Fire Detection Method using Convolutional N...IRJET Journal
 
IRJET- Anomaly Detection System in CCTV Derived Videos
IRJET- Anomaly Detection System in CCTV Derived VideosIRJET- Anomaly Detection System in CCTV Derived Videos
IRJET- Anomaly Detection System in CCTV Derived VideosIRJET Journal
 
IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2
 IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2 IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2
IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2IRJET Journal
 
Real Time Detection of Moving Object Based on Fpga
Real Time Detection of Moving Object Based on FpgaReal Time Detection of Moving Object Based on Fpga
Real Time Detection of Moving Object Based on Fpgaiosrjce
 
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET-  	  Convenience Improvement for Graphical Interface using Gesture Dete...IRJET-  	  Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...IRJET Journal
 
IRJET- Abandoned Object Detection System – A Review
IRJET- Abandoned Object Detection System – A ReviewIRJET- Abandoned Object Detection System – A Review
IRJET- Abandoned Object Detection System – A ReviewIRJET Journal
 
IRJET- Survey on Detection of Crime
IRJET-  	  Survey on Detection of CrimeIRJET-  	  Survey on Detection of Crime
IRJET- Survey on Detection of CrimeIRJET Journal
 
Event Detection Using Background Subtraction For Surveillance Systems
Event Detection Using Background Subtraction For Surveillance SystemsEvent Detection Using Background Subtraction For Surveillance Systems
Event Detection Using Background Subtraction For Surveillance SystemsIRJET Journal
 

What's hot (19)

IRJET- Surveillance of Object Motion Detection and Caution System using B...
IRJET-  	  Surveillance of Object Motion Detection and Caution System using B...IRJET-  	  Surveillance of Object Motion Detection and Caution System using B...
IRJET- Surveillance of Object Motion Detection and Caution System using B...
 
An Approach for Copy-Move Attack Detection and Transformation Recovery
An Approach for Copy-Move Attack Detection and Transformation RecoveryAn Approach for Copy-Move Attack Detection and Transformation Recovery
An Approach for Copy-Move Attack Detection and Transformation Recovery
 
IRJET- Criminal Recognization in CCTV Surveillance Video
IRJET-  	  Criminal Recognization in CCTV Surveillance VideoIRJET-  	  Criminal Recognization in CCTV Surveillance Video
IRJET- Criminal Recognization in CCTV Surveillance Video
 
Framework for Contextual Outlier Identification using Multivariate Analysis a...
Framework for Contextual Outlier Identification using Multivariate Analysis a...Framework for Contextual Outlier Identification using Multivariate Analysis a...
Framework for Contextual Outlier Identification using Multivariate Analysis a...
 
Human Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision TechniqueHuman Motion Detection in Video Surveillance using Computer Vision Technique
Human Motion Detection in Video Surveillance using Computer Vision Technique
 
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
Discovering Anomalies Based on Saliency Detection and Segmentation in Surveil...
 
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
AUTOMATIC THEFT SECURITY SYSTEM (SMART SURVEILLANCE CAMERA)
 
Motion detection system
Motion detection systemMotion detection system
Motion detection system
 
Smart web cam motion detection
Smart web cam motion detectionSmart web cam motion detection
Smart web cam motion detection
 
Deep-learning based single object tracker for night surveillance
Deep-learning based single object tracker for night surveillance  Deep-learning based single object tracker for night surveillance
Deep-learning based single object tracker for night surveillance
 
IRJET- A Review on Moving Object Detection in Video Forensics
IRJET- A Review on Moving Object Detection in Video Forensics IRJET- A Review on Moving Object Detection in Video Forensics
IRJET- A Review on Moving Object Detection in Video Forensics
 
IRJET - An Robust and Dynamic Fire Detection Method using Convolutional N...
IRJET -  	  An Robust and Dynamic Fire Detection Method using Convolutional N...IRJET -  	  An Robust and Dynamic Fire Detection Method using Convolutional N...
IRJET - An Robust and Dynamic Fire Detection Method using Convolutional N...
 
IRJET- Anomaly Detection System in CCTV Derived Videos
IRJET- Anomaly Detection System in CCTV Derived VideosIRJET- Anomaly Detection System in CCTV Derived Videos
IRJET- Anomaly Detection System in CCTV Derived Videos
 
IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2
 IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2 IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2
IRJET - Explicit Content Detection using Faster R-CNN and SSD Mobilenet V2
 
Real Time Detection of Moving Object Based on Fpga
Real Time Detection of Moving Object Based on FpgaReal Time Detection of Moving Object Based on Fpga
Real Time Detection of Moving Object Based on Fpga
 
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET-  	  Convenience Improvement for Graphical Interface using Gesture Dete...IRJET-  	  Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...
 
IRJET- Abandoned Object Detection System – A Review
IRJET- Abandoned Object Detection System – A ReviewIRJET- Abandoned Object Detection System – A Review
IRJET- Abandoned Object Detection System – A Review
 
IRJET- Survey on Detection of Crime
IRJET-  	  Survey on Detection of CrimeIRJET-  	  Survey on Detection of Crime
IRJET- Survey on Detection of Crime
 
Event Detection Using Background Subtraction For Surveillance Systems
Event Detection Using Background Subtraction For Surveillance SystemsEvent Detection Using Background Subtraction For Surveillance Systems
Event Detection Using Background Subtraction For Surveillance Systems
 

Viewers also liked (20)

Ku3518591863
Ku3518591863Ku3518591863
Ku3518591863
 
Kr3518331839
Kr3518331839Kr3518331839
Kr3518331839
 
La3518941898
La3518941898La3518941898
La3518941898
 
Lc3519051910
Lc3519051910Lc3519051910
Lc3519051910
 
Ks3518401849
Ks3518401849Ks3518401849
Ks3518401849
 
Lg3519251930
Lg3519251930Lg3519251930
Lg3519251930
 
Lh3519311940
Lh3519311940Lh3519311940
Lh3519311940
 
Ld35191119013
Ld35191119013Ld35191119013
Ld35191119013
 
Ky3518821886
Ky3518821886Ky3518821886
Ky3518821886
 
Kp3518241828
Kp3518241828Kp3518241828
Kp3518241828
 
Kw3518711873
Kw3518711873Kw3518711873
Kw3518711873
 
Kn3518141819
Kn3518141819Kn3518141819
Kn3518141819
 
Kz3518871893
Kz3518871893Kz3518871893
Kz3518871893
 
Ko3518201823
Ko3518201823Ko3518201823
Ko3518201823
 
Km3518071813
Km3518071813Km3518071813
Km3518071813
 
Kx3518741881
Kx3518741881Kx3518741881
Kx3518741881
 
Kt3518501858
Kt3518501858Kt3518501858
Kt3518501858
 
आम आदमी
आम आदमीआम आदमी
आम आदमी
 
Building a house in st john (2)
Building a house in st john (2)Building a house in st john (2)
Building a house in st john (2)
 
Algorithm Class at KPHB C, C++, Computer Science Training,IT Training,CS Trai...
Algorithm Class at KPHB C, C++, Computer Science Training,IT Training,CS Trai...Algorithm Class at KPHB C, C++, Computer Science Training,IT Training,CS Trai...
Algorithm Class at KPHB C, C++, Computer Science Training,IT Training,CS Trai...
 

Similar to Kv3518641870

A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN
A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN
A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN csandit
 
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN cscpconf
 
Development of Human Tracking in Video Surveillance System for Activity Anal...
Development of Human Tracking in Video Surveillance System  for Activity Anal...Development of Human Tracking in Video Surveillance System  for Activity Anal...
Development of Human Tracking in Video Surveillance System for Activity Anal...IOSR Journals
 
Abnormal activity detection in surveillance video scenes
Abnormal activity detection in surveillance video scenesAbnormal activity detection in surveillance video scenes
Abnormal activity detection in surveillance video scenesTELKOMNIKA JOURNAL
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET-  	  Behavior Analysis from Videos using Motion based Feature ExtractionIRJET-  	  Behavior Analysis from Videos using Motion based Feature Extraction
IRJET- Behavior Analysis from Videos using Motion based Feature ExtractionIRJET Journal
 
Survey on video object detection & tracking
Survey on video object detection & trackingSurvey on video object detection & tracking
Survey on video object detection & trackingijctet
 
Intelligent Video Surveillance System using Deep Learning
Intelligent Video Surveillance System using Deep LearningIntelligent Video Surveillance System using Deep Learning
Intelligent Video Surveillance System using Deep LearningIRJET Journal
 
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...sipij
 
IRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object DetectionIRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object DetectionIRJET Journal
 
Abandoned Object Detection Based on Statistics for Labeled Regions
Abandoned Object Detection Based on Statistics for Labeled RegionsAbandoned Object Detection Based on Statistics for Labeled Regions
Abandoned Object Detection Based on Statistics for Labeled RegionsIRJET Journal
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity RecognitionIRJET Journal
 
A Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesA Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesCSCJournals
 
IRJET- Moving Object Detection using Foreground Detection for Video Surveil...
IRJET- 	 Moving Object Detection using Foreground Detection for Video Surveil...IRJET- 	 Moving Object Detection using Foreground Detection for Video Surveil...
IRJET- Moving Object Detection using Foreground Detection for Video Surveil...IRJET Journal
 
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...cscpconf
 
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...Surveillance Video Based Robust Detection and Notification of Real Time Suspi...
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...csandit
 
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...CSCJournals
 
Automated Surveillance System and Data Communication
Automated Surveillance System and Data CommunicationAutomated Surveillance System and Data Communication
Automated Surveillance System and Data CommunicationIOSR Journals
 

Similar to Kv3518641870 (20)

A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN
A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN
A NOVEL BACKGROUND SUBTRACTION ALGORITHM FOR PERSON TRACKING BASED ON K-NN
 
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN
A Novel Background Subtraction Algorithm for Person Tracking Based on K-NN
 
F011113741
F011113741F011113741
F011113741
 
Development of Human Tracking in Video Surveillance System for Activity Anal...
Development of Human Tracking in Video Surveillance System  for Activity Anal...Development of Human Tracking in Video Surveillance System  for Activity Anal...
Development of Human Tracking in Video Surveillance System for Activity Anal...
 
Abnormal activity detection in surveillance video scenes
Abnormal activity detection in surveillance video scenesAbnormal activity detection in surveillance video scenes
Abnormal activity detection in surveillance video scenes
 
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
[IJET-V1I3P20] Authors:Prof. D.S.Patil, Miss. R.B.Khanderay, Prof.Teena Padvi.
 
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
IRJET-  	  Behavior Analysis from Videos using Motion based Feature ExtractionIRJET-  	  Behavior Analysis from Videos using Motion based Feature Extraction
IRJET- Behavior Analysis from Videos using Motion based Feature Extraction
 
Survey on video object detection & tracking
Survey on video object detection & trackingSurvey on video object detection & tracking
Survey on video object detection & tracking
 
Intelligent Video Surveillance System using Deep Learning
Intelligent Video Surveillance System using Deep LearningIntelligent Video Surveillance System using Deep Learning
Intelligent Video Surveillance System using Deep Learning
 
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...
A NOVEL METHOD FOR PERSON TRACKING BASED K-NN : COMPARISON WITH SIFT AND MEAN...
 
IRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object DetectionIRJET- Comparative Analysis of Video Processing Object Detection
IRJET- Comparative Analysis of Video Processing Object Detection
 
Abandoned Object Detection Based on Statistics for Labeled Regions
Abandoned Object Detection Based on Statistics for Labeled RegionsAbandoned Object Detection Based on Statistics for Labeled Regions
Abandoned Object Detection Based on Statistics for Labeled Regions
 
Human Activity Recognition
Human Activity RecognitionHuman Activity Recognition
Human Activity Recognition
 
A Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal FeaturesA Framework for Human Action Detection via Extraction of Multimodal Features
A Framework for Human Action Detection via Extraction of Multimodal Features
 
Q180305116119
Q180305116119Q180305116119
Q180305116119
 
IRJET- Moving Object Detection using Foreground Detection for Video Surveil...
IRJET- 	 Moving Object Detection using Foreground Detection for Video Surveil...IRJET- 	 Moving Object Detection using Foreground Detection for Video Surveil...
IRJET- Moving Object Detection using Foreground Detection for Video Surveil...
 
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...
SURVEILLANCE VIDEO BASED ROBUST DETECTION AND NOTIFICATION OF REAL TIME SUSPI...
 
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...Surveillance Video Based Robust Detection and Notification of Real Time Suspi...
Surveillance Video Based Robust Detection and Notification of Real Time Suspi...
 
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
Java Implementation based Heterogeneous Video Sequence Automated Surveillance...
 
Automated Surveillance System and Data Communication
Automated Surveillance System and Data CommunicationAutomated Surveillance System and Data Communication
Automated Surveillance System and Data Communication
 

Recently uploaded

A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)Gabriella Davis
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhisoniya singh
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonAnna Loughnan Colquhoun
 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024Results
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Drew Madelung
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking MenDelhi Call girls
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreternaman860154
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerThousandEyes
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024Rafal Los
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxOnBoard
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonetsnaman860154
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Servicegiselly40
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountPuma Security, LLC
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGSujit Pal
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking MenDelhi Call girls
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Paola De la Torre
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slidespraypatel2
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 

Recently uploaded (20)

A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)A Domino Admins Adventures (Engage 2024)
A Domino Admins Adventures (Engage 2024)
 
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | DelhiFULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
FULL ENJOY 🔝 8264348440 🔝 Call Girls in Diplomatic Enclave | Delhi
 
Data Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt RobisonData Cloud, More than a CDP by Matt Robison
Data Cloud, More than a CDP by Matt Robison
 
A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024A Call to Action for Generative AI in 2024
A Call to Action for Generative AI in 2024
 
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
Strategies for Unlocking Knowledge Management in Microsoft 365 in the Copilot...
 
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
08448380779 Call Girls In Diplomatic Enclave Women Seeking Men
 
Presentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreterPresentation on how to chat with PDF using ChatGPT code interpreter
Presentation on how to chat with PDF using ChatGPT code interpreter
 
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
Neo4j - How KGs are shaping the future of Generative AI at AWS Summit London ...
 
How to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected WorkerHow to Troubleshoot Apps for the Modern Connected Worker
How to Troubleshoot Apps for the Modern Connected Worker
 
The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024The 7 Things I Know About Cyber Security After 25 Years | April 2024
The 7 Things I Know About Cyber Security After 25 Years | April 2024
 
Maximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptxMaximizing Board Effectiveness 2024 Webinar.pptx
Maximizing Board Effectiveness 2024 Webinar.pptx
 
How to convert PDF to text with Nanonets
How to convert PDF to text with NanonetsHow to convert PDF to text with Nanonets
How to convert PDF to text with Nanonets
 
CNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of ServiceCNv6 Instructor Chapter 6 Quality of Service
CNv6 Instructor Chapter 6 Quality of Service
 
Breaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path MountBreaking the Kubernetes Kill Chain: Host Path Mount
Breaking the Kubernetes Kill Chain: Host Path Mount
 
Google AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAGGoogle AI Hackathon: LLM based Evaluator for RAG
Google AI Hackathon: LLM based Evaluator for RAG
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men08448380779 Call Girls In Civil Lines Women Seeking Men
08448380779 Call Girls In Civil Lines Women Seeking Men
 
Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101Salesforce Community Group Quito, Salesforce 101
Salesforce Community Group Quito, Salesforce 101
 
Slack Application Development 101 Slides
Slack Application Development 101 SlidesSlack Application Development 101 Slides
Slack Application Development 101 Slides
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 

Kv3518641870

  • 1. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.1864-1870 RESEARCH ARTICLE www.ijera.com OPEN ACCESS A Novel Algorithm To Detect Changes In Surveillance Video Haitham Kareem Ali (**), Yasir Khalil Ibrahim(*), (*) Faculty of Engineering, Jarash University, Jarash, Jordan. (**)Slemani Polytechnic University, Slemani College of Technical Engineering, Communication & Electronics Department, Slemani, Kurdistan Region , Iraq. Abstract Tracking changes in an acquired surveillance video is essentially impractical task for the human due to the complexity of the monitoring process, specifically for a long-time period. Finding changes in a sequence of images captured by surveillance video system is a complicated task for the human and it is complicated by fitting it to a classification model. Given a sequence of images, we are concerned with automatically detecting the occurrence of changes in the given sequence of images. In this paper, the problem of detecting changes in digital images captured by surveillance video system is explored. A novel algorithm, involving correlation coefficients and eigenstructurebased Mosaab distance measure, is suggested. The proposed algorithm is a mobile-based application, that can be used to process pixel-based data (i.e., sequence of images) at camera locations and send only real-time numeric results to client's mobile device(s) . Therefore, the aim of proposed algorithm is to analyize and eliminate the amount of pixel-based data to be sent via the network. Consequently, the received real time numeric results can be visualized using an application software installed in client's mobile device. The main function of the mobile-based application software is to illustrate the received results as a real-time graph. The proposed algorithm shows effectiveness in eliminating the flow of raw pixel-based data in the network. Key words: Mosaab-distance measure, correlation coefficient, surveillance video system. I. INTRODUCTION Normally, the surveillance video system produces huge amount of data that requires analysis. Usually, the main analysis task, which is monitoring changes in a sequence of images (i.e., real time analysis), is tackled by human. Tracking changes in the acquired video is essentially impractical task for the human since it is a long-term monitoring process. In any scene, there is a number of changes that take place over the time. We assume that the input images of a scene are captured by surveillance video system with timing setup as a discrete variable. An entity can be either an object or human (i.e., individual), in this context, if an entity is not a human, then it is considered as an object. An activity is defined as an action that can happen to an entity at a particular point in time. When an activity is performed on an entity it becomes an event. Thus, implicitly, we are looking for the occurrences of an event or events in a particular scene. For example, remove is an activity, when it is performed on a book (i.e., entity) it become an event, which is: the book is removed from the table (i.e., removed from a particular scene). The state of any scene is considered as a function of time. During any period of time, a scene is said to be static if there is zero event occurred, otherwise a scene is said to be dynamic. Finding changes sequentially in a sequence of images captured by surveillance video system is complicated task for the human and it is complicated by fitting it to a classification model. Our work focuses exclusively on the following: www.ijera.com Given a sequence of images captured for a particular scene, we are concerned with autodetecting the occurrence of changes in the given sequence of images and the change occurs when one activity or a set of activates are performed on an entity or a set of entities in a particular scene, under the assumptions: (i) lack of information about those activities (i.e., no prior knowledge or insufficient labeled datasets that can be used for training) and (ii) the sets of activities and entities are previously unknown. Moreover, the detection possess must be performed locally at camera locations, to eliminate the amount of pixel-based data to be sent via the network to mobile device(s) of a client. In this study, we adding another task to the surveillance system, which is the analysis of a sequence of images captured by surveillance system installed in highly sensitive areas (e.g., offices of V.I.P.s). The advantage behind adding this task is to detect activities (i.e., changes) in a sequence of images locally at camera locations and instead of transmitting huge amount of pixel-based data to the mobile device(s) of a client via the network, the system sends only the activity detection results. In this way, we can eliminate the amount of pixel-based data to be transmitted from camera locations to the mobile device(s) of a client via the network. Sending only few real numbers (i.e., the detection results) to the mobile device (s) of a client via the network will minimize the potential threats that result from sending pixel-based data to the mobile device(s) of a client via the network. 1864 | P a g e
  • 2. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 In this paper, we briefly survey existing approaches to the problem, outline the method we have implemented and present our results and ideas for future work. The structure of this study is summarized as follows. In Section 2, which it consists of three subsections, we present materials and methods. In subsection 2.1, we present related work. In subsection 2.2, we present the statement of the problem. In subsection 2.2, we present the propose algorithm. In Section 3, we present the experimentation and results. Finally, conclusions are presented in Section 4. II. MATERIALS AND METHODS 2.1 Related work: The main stages of any pattern recognition system are: (i) feature extraction stage and (ii) classification stage. There two main types of classification: Either (i) supervised classification or (ii) unsupervised classification Jain (2000). Extracting information from data is an essential step for pattern recognition systems and the validity of the extracted information has an impact on the classification stage. Machine detection, thus, can be used to validate the human detection in order to increase accuracy. Replacing eye-based detection by machine-based detection will often be the first step in building an auto-detection system that has the capability to analyze a sequence of images captured by a surveillance system. As we mentioned in the previous section, there are two components that arise in this context: (i) a set of activities and (ii) a set of entities. The composition of any dynamic scene is result from the interaction between those two components (i.e., these two sets). An event is an activity that happens to an entity or set of entities at a particular point in time. An event is categorized as either normal event or abnormal event (i.e., anomaly). An extensive work has been done in the area of analyzing video sequences. However, there exists a few challenges that must be considered, for example: Detecting activities in surveillance video sequences at camera locations and send only numeric results as a real-time chart to the mobile device(s) of a client. Detecting activities in surveillance video sequences is well known problem in several research areas and it is under the focus of different research communities such as: Pattern Recognition, Computer Vision, Artificial Intelligence and Soft Computing, Pattern Analysis and Machine Intelligence and Computational Intelligence and Data Mining. Precisely, activities in a scene are classified as either (i) normal activates or (i) abnormal activities. They are two approaches that can be used to solve this problem: Either (i) supervised classification or (ii) unsupervised classification. The data availability for the considered activities is the key point in deciding which approach should be used to solve this problem. Since we are www.ijera.com www.ijera.com classifying each activity into two classes (i.e., either change class or no-change class), then a distancebased classifiers are required. Activities are related to an entity (e.g., human(s), object(s)) and either performed by one entity or by a group of entities (e.g., motion of one person or group of people in a scene, moving object from one position to another). We just presented the contributions achieved by researchers in the field in the recent years, however, the goal of this paper, which we stated in the previous section (i.e., paper statement) is differ from the goals of the contributions that we just reviewed in this section. In this study, the main challenge to overcome is to detect changes defined by a set of activities, those activities are either associated with an entity or a set of entities (e.g., object(s) or human (s)) in a scene, under the following assumptions: (i) lack of information about those activities (i.e., no prior knowledge or insufficient labeled datasets that can be used for training), (iii) the sets of activities and entities are previously unknown and (iii) to eliminate the amount of pixel-based data to be transmitted from camera locations to the mobile device(s) of a client via the network. In this section, we presented the related work. In the next section, we present the statement of the problem. 2.2 The statement of the problem: The problem under investigation is a well known problem in image analysis area. It is the problem of detecting changes in a sequence of images acquired by a surveillance system. Mathematically speaking, suppose that we have a surveillance system that produces a huge amount of data, where the data are represented by images. Without loss of generality, suppose that the system produces images captured for a highly sensitive site and the problems that arise in this context is to monitor those images and track the changes by human efficiently for a long-term period and send results to mobile device(s) of a client. The solution for this problem can be achieved by replacing the human-based monitoring by machine-based monitoring and thus a pattern recognition system is required since the changes in those images are considered as patterns. The core of any pattern recognition system is the recognizer or the algorithm that can used to recognize those changes over the time. The time is discrete since those images captured at discrete points in time: t1, t2, t3, ..., tk. Hence, we assume that we have a sequence of k images: Image1, Image2,..., Imagek captured at t1, t2, t3,..., tk respectively. The data under consideration are images or pixel-based data and mapping these images from the data space into the real-valued feature space is an essential step to reduce the mount of input data to be processed by the pattern recognition system in order to detect changes in these images sequentially. 1865 | P a g e
  • 3. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 www.ijera.com Fig. 1: The statement of the problem Fig. 2: The proposed decomposition process performed on two successive images captured by a surveillance system Instead of detecting changes in these images embedded in the data space, which can be achieved by measuring the distance between successive images in the sequence, detecting changes is performed in the feature space by measuring the distance between the representation of successive images embedded in the feature space (Fig. 1). In this section, we presented the statement of the problem. In the next section, we present the proposed algorithm. 2.3 The proposed algorithm: In this paper, as we mentioned, our goal is to detect changes in a sequence of images Image1, Image2,..., Imagek of size m × n captured by surveillance system(Fig 2). Those www.ijera.com changes are defined by a set of activities, those activities are either associated with an entity or a set of entities (e.g., object(s) or human(s)) in a scene, under the following assumptions: (i) lack of information about those activities (i.e., no prior knowledge or insufficient labeled datasets that can be used for training) and (ii) the sets of activities and entities are previously unknown. As we know, algorithms for monitoring changes in a sequence of images is mainly based on global features of images (e.g., global correlation between successive images) often yield many cases of false detection. In addition, the performance of threshold value-based algorithms is based on the accuracy of estimating the threshold values and changing those parameters (i.e., threshold 1866 | P a g e
  • 4. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 values) to decrease the rate of false detection may increases the rate of false detection. For the above reasons, we build our algorithm (feature extractor and distance-based classifier) by consider the following concepts:(i) statistical-based local features (e.g., the correlation coefficient between partitions of the image), (ii) eigenstructure based Mosaab distance measure (it is an inverse-free ℜp×p-based distance measure that can works with singular matrices and it require less computation compared with Mahalanobis distance), (iii) high dimensional-based representation (e.g., a representation embedded in ℜp×p). Precisely, the proposed classifier is a distance-based classifier which uses the statistical-based local features to perform the required task. As we mentioned earlier in this paper, the data under consideration are images, in fact, it is pixel-based data, which is a huge data. To reduce the amount of data to be processed, it is necessary to map those images from the data space (i.e., pixel-based) to the real-value feature space (i.e., real-valued based) Thus, we have to define a new representation to those images under consideration. The representation must be embedded in ℜp×p since eigenstructure-based Mosaab distance measure is considered. The working mechanism of the proposed algorithm can be summarized follows. First, we partitioned each image into four same size sub-images. Then, the correlation coefficients between those sub-images are computed (i.e., 6 correlation coefficients). Let (i)  corr (Imagei, j Imageir ) (where, j = 1, jr 2,..., 4 and r = j + 1, 2,..., 4) be the correlation coefficient between the image partitions (i.e., subimages) Imageij and Imageir of Imagei. For each Imagei (i = 1, 2,…,k), the algorithm computes (i) (i) (i) (i) (i) 11 , 12 , 13 , 14 , (i) ,and 34 , where those coefficient 23 are real values between 0 and 1. Daoud and Kremer, (Daoud and Kremer, 2009; Daoud, 2010; Daoud and Kremer, 2006) proposed a new measure that is defined in terms of variations to capture the deviations between two sets of biosequences or between a pair of sequences. The metric is built using the objective function: (1) Maximizing the above objective function leads to the following metric: (2) where, 1 is the largest eigenvector associated with 1, the largest eigenvalue of the matrix (Ωi-Ωj) and: Var(  'X(k) )  E(  'X(k) X(k) ' )   'E(X(k) X(k) ' )    ' i  for k  i, j www.ijera.com (3) www.ijera.com Definition 1: Mosaab distance measure (Daoud and Kremer, 2009; Daoud, 2010; Daoud and Kremer, 2006) Dij (1):ℜp×p×ℜp×p ⇒ ℜ+ is a variance-covariance structure-based distance measure that measures the distance between two variance-covariance structures represented by the variance-covariance matrices Ωi and Ωj. The proposed distance measure can measure the deviation in variations of the same feature vector being measured on two datasets (e.g. objects, images, sub-images, sequences, set of sequences etc...). The deviation in variations is explained by the eigenstructure of the difference between two variance covariance structures, where each variance covariance structure represents a dataset(e.g. object, image, subimage, sequence, set of sequence etc...). The eigenstructure is defined in terms of the largest eigen-value (i.e. max, see equations 1-2) of the difference between two variance-covariance structures. It is clear that Mosaab distance measure is a variance-covariance structure-based distance measure and it is a matrix-based distance measure (i.e., ℜp×p×ℜp×p ⇒ ℜ+). Thus the representation of images must be embedded in ℜp×p (i.e., feature space). Since for each Imagei (i = 1, 2,…. k), the representation (i)  (where j = 1, 2,..., 4 and r = j + 1, 2,..., 4), jr thus, for each Imagei (i = 1, 2,…, k) we construct a diagonal matrix (i) (i) (i) (i) i  diag(11 , 12 , 13 , 14 , (i) , (i) and (i) 66 ) . Since 23 24 34 the matrix Σi is neither variance-covariance matrix nor correlation matrix, we redefine Mosaab distance measure as follows. Definition 2: The modified form of Mosaab distance measure i (,): ℜp×p×ℜp×p ⇒ ℜ+ is defined as: ' i (i , (i 1) ) | 1 (i  (i 1) )1 || 1 | 0 (4) A distance measure that measures the distance between two matrices Σi and Σ(i+1) that represent Imagei and Image(i+1) respectively, where 1 is the eigenvector associated with the largest eigenvalues 1 of the matrix (Σi-Σ(i+1)) and: (i) (i) (i) i  diag (11 , 12 ,..., 34 ) (i) 11 0  (i) 0 12   M M  0  0  L L O K 0   0  M  (i) 34   (5) Finally, the proposed algorithm (see Algorithm 1) produces a sequence of eigenvalues 1, 2,..., (k-1) that represent the sequential changes in Image1, Image2,..., Imagek respectively. In this section, we presented a novel algorithm to detect the changes in a sequence of images captured by a surveillance system. In the next section, we present the experimentation and results. 1867 | P a g e
  • 5. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 III. RESULTS and DISCUSSION In this section, we used a sequence of 100 images captured by a surveillance system to evaluate the performance of the proposed algorithm and a discussion is given below. A sample output from Matlab code, which shows the partition process, is given in Fig. 3. The results of the proposed algorithm are shown in Fig. 8. In Fig. 8, the horizontal axis represents the index of the distance value i and the vertical axis represents the value of the modified Mosaab distance measure. The solid points of the red curve represent the distance values measured in the feature space between successive images in the sequence of 100 images. There are few peak values clearly appeared at i = 3, 8, 12, 31, 36, 56, 57, 84, 86 and 89. The distance values are: 3 = 0.371, 8 = 0.8618, 12 = 0.3167, 31 = 0.3199, 36 = 0.3084, 56 = 0.3494, 57 = 0.4578, 84 = 0.3572, 86 = 0.5412, and 89 = 0.3488. Algorithm 1: Detecting Changes Sequentially in a Sequence of Images input : Given a sequence of images Image1, Image2,..., Imagek of size m × n captured by surveillance system. output: A sequence of eigenvalues 1, 2,...,  (k-1) that represent the sequential changes in Image1, Image2,..., Imagek. 1 for i ← 1 to k do 2 Partition Imagei into four same size sub-images (i.e. four disjoint partitions). Let Imagei1, Imagei2, Imagei3 and Imagei4 be the four partitions of Imagei respectively. Partition Imagei+1 into four same size sub-images (i.e. four disjoint partitions). Let Imagei1, Image(i+1)2, Image(i+1)3 and Image(i+1)4 be the four partitions of Imagei+1 respectively. Let corr(,) be the correlation coefficient. 3 for j ← 1 to 4 do 4 for r ← j + 1 to 4 do 5 Compute corr(Imageij ; Imageir) 6 Compute corr(Image(i+1)j ; Image(i+1)r) end end 9 Construct the diagonal matrix Σi = diag( corr(Imagei1; Imagei2), corr (Imagei1;Imagei3), corr(Imagei1, Imagei4), corr (Imagei2; Imagei3), corr(Imagei2, Imagei4), corr (Imagei3; Imagei4)) 10 Construct the diagonal matrix Σ(i+1) = diag(corr(Image(i+1)1; Image(i+1)2), corr(Image(i+1)1; Image(i+1)3), corr(Image(i+1)1; Image(i+1)4), corr(Image(i+1)2; Image(i+1)3), corr(Image(i+1)2; Image(i+1)4), corr(Image(i+1)3; Image(i+1)4)) www.ijera.com γ www.ijera.com 11 Compute λi = Abs(γ' (Σi - Σ(i+1)) γ), where is the largest eigenvector of the matrix (Σi-Σ(i+1)). end For example, the distance value 3 = 0.371 is estimated from comparing the images 3 and 4, and the distance value 8 = 0.8618 is estimated from comparing the images 8 and 9 (see Figure 4). Moreover, the distance value 84 = 0.3572 is estimated from comparing the images 84 and 85, and the distance value 89 = 0.3488 is estimated from comparing the images 89 and 90 (see Fig. 5). From Figures 4 and 5 , our conclusions (i.e. human eyes) coincided with the results of the proposed algorithm, but for a long-term period, it is harder for the human eyes to recognize such changes (i.e., in case of longterm monitoring process). However, there are other cases where the change is unrecognizable by the human eyes due to the speed of change or due to the nature of change (e.g. tiny changes), but it is easily can be recognized by the proposed algorithm. Those cases are illustrated in Figure 6. Consider the four successive images: 85, 86, 87 and 88. The discussions of those cases are given below. Fig. 3: A sample output from matlab code shows the partition process Fig. 4: Peak values 3 = 0.371 (the corresponding images: 3 and 4) 8 = 0.8618 (the corresponding images: 8 and 9) 1868 | P a g e
  • 6. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 Fig. 5: Peak values 84 = 0.3572 (the corresponding images: 84 and 85) and 89 = 0.3488 (the corresponding images: 89 and 90) Fig. 7: www.ijera.com Two images represent: (i) Light switched ON, and (ii) Light switched OFF Case 1: The main difference between image 85 and image 86 is the position of object1. Object1 moved from one position to another. This change is efficiently detected by the proposed algorithm and the distance value is 85 = 0.1835 (Fig. 6). Case 2: The main differences between image 86 and image 87 are: (i) object1 removed, (ii) object2 moved from one position to another, (iiii) object3 removed and (iv) objects4 added to the scene (Fig. 6). Those changes are efficiently detected by the proposed algorithm and the distance value is 86 = 0.5412. Since the number of changes occurred in this case is more than the number of changes occurred in case 1, then the distance value 86 is greater than the distance value 85. Fig. 6: Four successive images captured by the surveillance system Fig. 8: The results of the proposed algorithm used for testing a sequence of 100 images captured by surveillance system Case 3: There is no main difference between image 87 and image 88, thus, the distance between the two images has the value 87 = 0.0866, which is a very small value compared with distance values 85 and 86 (Fig. 6). There are two points arise here: (i) the quality of the digital camera used in this experiment www.ijera.com and (ii) the small hidden changes that are undetectable by human eyes which they may cause this problem (i.e., the distance value is not equal to 0). There is another problem that we should mention in this section, it is the visibility problem. The 1869 | P a g e
  • 7. Yasir Khalil Ibrahim et al Int. Journal of Engineering Research and Application ISSN : 2248-9622, Vol. 3, Issue 5, Sep-Oct 2013, pp.01-05 www.ijera.com problem is illustrated in Fig. 7. When the lights are switched off the visibility is almost zero. However, this case is perfectly detected by the proposed algorithm as a change in lights intensity, but probably the scene details remain unchanged. In this case, advanced equipments are needed to handle such problem, which is beyond the scope of this paper. Finally, in this section, we presented the experimentation and results of the proposed algorithm. In the next section, conclusions are presented. IV. CONCLUSION While the results of the proposed algorithm were generally quite good, there are several potential areas for future improvement: instead of performing 4-subimages decomposition process on images, one may perform 8-subimages decomposition process on images which might be beneficial. In particular, we concluded that the proposed algorithm is useful for detecting changes in a sequence of images locally at camera location and send real-time numeric results to mobile device(s). In other words, the proposed algorithm showed effectiveness in eliminating the flow of raw pixel-based data in the network and minimizing the potential threats that result from sending highly sensitive pixel-based data to the mobile device(s). REFERENCES [1] [2] [3] [4] Daoud and Kremer 2006. Daoud, M. and S.C. Kremer. Neural and statistical classification to families of bio-sequences. International Joint Conference on Neural Networks, 2006 (IJCNN ’06), pp: 699-704, 2006.http://dx.doi.org/10.1109/IJCNN.2006. 246752 Daoud and Kremer 2009. Daoud, M. and S.C. Kremer. A new distance distribution paradigm to detect the variability of the influenza-A virus in high dimensional spaces, Applications of Machine Learning in Bioinformatics Workshop, IEEE International Conference on Bioinformatics & Biomedicine (IEEE BIBM 09), November 1-4, 2009, Washington D.C. Daoud, 2010. Daoud, M.. A new variancecovariance structure-based statistical pattern recognition system for solving the sequenceset proximity problem under the homologyfree assumption. PhD thesis, 2010, ISBN: 9780494678268. Jain, 2000. Jain, Duin and Mao Anil K. Jain, Robert P. W. Duin and Jianchang Mao. Statistical pattern recognition: A review. IEEE Transactions Pattern Analysis Machine Intell., 22: 4-37, 2000. http://doi.ieeecomputersociety.org/10.1109/3 4.824819 www.ijera.com 1870 | P a g e