SlideShare a Scribd company logo
Online Weighted One-Class Ensemble for Feature Selection in
Background/Foreground Separation
Caroline Silva, Thierry Bouwmans, Carl Fr´elicot, Micka¨el Coustaty
December 6th, 2016 - Mexico
Summary
2/29
1 Introduction
2 Proposed Work
3 Experimental Results
4 Conclusion and Future Works
Introduction
3/29
Background Subtraction
Ini$alize	
	Backgrounwd	Model	
Background	Model	
Maintenance	
Foreground	
Detec$on	
Model	
Input	Stream	 Output	Masks	
Figure: Block diagram of the background subtraction process.
Introduction
4/29
Background Subtraction
Background subtraction is a widely used approach for detecting moving objects from static
cameras.
Figure: Moving object tracking results.
Introduction
5/29
Why is it interesting today
X1
X2
X3
X4X5
Figure: Given a complex scene with different regions X1, X2, X3, X4 and X5, each region can be characterized
by different features such as: texture, color, texture-color, motion and edge.
Common methods
Same feature is used globally for the whole scene.
Key challenges
It requires a deep knowledge of the scene.
It is possible to automatically select the most relevant features
Introduction
6/29
Feature Selection
Traditionally, feature selection methods can be categorized into three main groups:
Wrapper-based methods
• employ a classification algorithm as a “black box” for selecting a set of relevant features.
Embedded methods
• the feature selection is incorporated as part of the classification algorithm, such as in
the decision trees or in neural networks.
Filter-based methods
• evaluate the relevance of the features based on a statistical measure estimated directly
from the data.
Ensemble-based approaches for feature selection
Summary
7/29
1 Introduction
2 Proposed Work
3 Experimental Results
4 Conclusion and Future Works
Proposed Work
8/29
N
u
m
b
e
r
o
f
fra
m
e
s
200
150
Frames
100
50
0
Feature Importance
Motion
Gradient
Color+Texture
Features
Texture
Color
0.6
1
0.8
0.4
0
0.2
Importance
Figure: Illustration of a complex scene (left) and its features
importance over time. The bar-graph (right) shows the feature
importance variations for a certain region of the scene along
time. The color, texture, cor-texture, gradients and motion
features are represented by pink, blue, pale orange, green and
dark orange color, respectively.
We propose an online weighted ensemble
of one-class SVMs (Support Vector
Machines) for foreground-background
separation.
It automatically selects the best features
for different regions of the image, and the
more relevant features are used for
foreground segmentation.
9/29
ω1	
ω3	
Ψ1	
Ψ2	
Ψ3	
Final	
predic.on	
	Features	set	(p)	
weighted	random	
subspace	
p*1	⊂	p	
P*2	⊂	p	
p*3	⊂	p	
build	M	weak		
classifiers	
ω2	
		
p*M	⊂	p	
Ψ1	
ᵦ1	
		
Ψ2	
ΨL	
select	and	combine	the	L	best	
base	classifiers	
Adap,ve	Importance	
Computa,on	(ᵦ)	
ground	truth	data	
heuris,c	model	
update	
ΨM	
ωM	
ᵦ2	
		
ᵦL	
		
Figure: Brief overview of the proposed framework.
Proposed Work
10/29
An Online Weighted One-Class Ensemble for Feature Selection
A. Generate multiple base models
N
u
m
b
e
r
o
f
fra
m
e
s
This above approach increases the diversity of base
classifiers since different weights of each random sub-
space are taken to distinguish the decision boundaries
computed by the classifiers.
Algorithm 1 Generate multiple base background models
1: Require: IWOC-SVM training procedure, training sequence X, subspace dimension p∗, number of base classifiers M, weight
distribution δ(x)
2: k ← 1
3: repeat
4: Sk ← SelectRandomSubspace(X,p∗)
5: Train k-th IWOC-SVM on Sk with respect to weights w ∼ δ(x)
6: k ← k+1
7: until k > M
8: Output: trained IWOC-SVM base classifiers Ψ = {Ψ1,Ψ2,...,ΨM }
Proposed Work
11/29
How to Choose the Classifier?
Bicego and Figueiredo, 2009 proposed a Weighted version that allow to use weights W = {w1,...,wN } in [0,1]
for the data.
	
	
	
	
	
	
													
	
	
R	
										
	
	
	
	
	
	
	
	
	
													
	
	
R	
										
	
	
	
R	
a	
wiξi	
target	
outlier	
xi	
Weighted One-Class SVM (WOC-SVM)
Minimizing the hypersphere volume implies the
minimization of R2
. To prevent the classifier from
over-fitting with noisy data, slack variables ξi are
introduced to allow some target points (respectively
outliers) outside (respectively inside) the hypersphere.
Therefore the problem is to minimize:
Θ(a,R) = R2
+C
N
∑
i=1
wi ξi (1)
where C is a user-defined parameter that controls the
trade-off between the volume and the number of
target points rejected. The larger C, the less outliers
in the hypersphere.
The BS task requires adjust the learned model to the scene variations over time
12/29
Why update the background model?
Proposed Work
13/29
An Incremental Weighted One-Class SVM (IWOC-SVM)
We propose an IWOC-SVM which is closely related to the procedure proposed by Tax and Laskov (2003).
Given new samples Z1 = {x1,x2,...,xs} and its
respective weights not learned by the IWOC-SVM.
Karush-Kuhn-Tucker (KKT) conditions:
αi = 0 ⇒ ||xi −a||2
< R2
(2)
0 < αi < C ⇒ ||xi −a||2
= R2
(3)
αi = C ⇒ ||xi −a||2
> R2
(4)
The mathematical model can be defined as:
R −θ ≤ ||x −a||≤ R (5)
where θ ∈ [0,R] is relative to the distribution of
previous training set.
	
	
	
	
	
	
													
	
	
R	
										
	
	
	
	
	
	
	
	
	
													
	
	
R	
										
	
	
	
R	
a	
wiξi	
target	
outlier	
xi	
Algorithm 2 Incremental Weighted One-Class SVM
1: Require: Previous training set Z0, newly added training set Z1 and its respective weights
2: Train IWOC-SVM classifier on Z0, then split Z0 = SV0 ∪NSV0
3: Input new samples Z1. Put samples that violate KKT conditions in ZV
1 . If ZV
1 = /0, then goto 2.
4: Put samples from NSV0 that satisfy Eq. (5) into NSVS
0 .
5: Set Z0 = SV0 ∪NSVS
0 ∪ZV
1 and train IWOC-SVM classifier on Z0.
6: Output: IWOC-SVM classifier Ω and the new training set Z0.
.
Proposed Work
14/29
B. Selection of the best base classifiers
Let λcorrect
k (respectively λwrong
k ) be the number of times a pixel was correctly (respectively uncorrectly) classified
by the k-th (k = 1,...,M) base classifier from given ground truth data. Then, the corresponding error is given by:
errork =
λwrong
k
λcorrect
k +λwrong
k
(6)
Note that only the base classifiers that have the smallest errors are combined and used to differentiate the
moving objects from the background model in the scene.
15/29
The Algorithm 3 was adapted from the algorithm proposed by (Wozniak, 2013)
Algorithm 3 Adaptive Importance (AI) computation
1: Require: Strong classifier H, validation set (t1,y1),...,(tN ,yN ) where ti ∈ T, yi ∈ Y = 0,1 for background and foreground
examples respectively, set of L best base classifiers Ψ = {Ψ1,Ψ2,...,ΨL}, learning rate parameter γ
2: Initialize all L best classifiers with importance: βl = 1
3: repeat
4: Classify ti using the strong classifier H according to Eq. (9)
5: for l = 1 : L do
6: Checks response of Ψl and calculates their errorl according to Eq. (6)
7: For each best classifier, Ψl , update the importance βl = βl (i −1)+
Pa(Ψl )−Pa(Ψl (i−1))
(N+γ)
8: where Pa(Ψl ) = 1 −errorl according to Eq.(6).
9: end for
10: until i < N
11: Normalize the importance β of each L best classifier.
12: Output: New importance assigned to the best classifiers β = {β1,β2,...,βL}
Proposed Work
16/29
C. Background Detection
Given an incoming pixel x to be classified, one can define a support function associated to the target class ω for
each of the L best base classifiers: ∀l = 1,...,L
Fl (x,ω) =
1
s1
exp(−d(x,a)/s2) (7)
where d(x,a) is a distance metric from x to the center a of the target class ω, s1 is a normalization factor and s2
is a scale parameter. Each Fl (x,ω) is then compared to a threshold t1 to obtain the positive or negative class
labels:∀l = 1,...,L
cl (x,ω) =
1 if Fl (x,ω) t1
−1 otherwise
(8)
Comparing the weighted sum of theses L class labels as in (Tax and Duin, 2001) to another threshold t2 allows
to define the strong classifier for x as follows:
H(x) =
1 if 1
L ∑L
l=1 βl cl (x,ω) t2
0 otherwise
(9)
A pixel x is classified as a background pixel if H(x) = 0.
Proposed Work
17/29
D. Heuristic approach for Background Model Maintenance
The Small Votes Instance Selection (SVIS) introduced by Guo and Boukir (Guo and Boukir, 2015) consists of an
unsupervised ensemble margin that combines the first c(1) and second most voted class c(2) labels under the
learned model. Let vc(1)
and vc(2)
denote the relative number of votes. Then the margin, taking value in [0,1] is:
m(x) =
vc(1)
−vc(2)
L
(10)
where L represents the number of best base classifiers in the ensemble. The first smallest margin instances are
selected as support vector candidates. The strong model is updated by the first smallest margin instances. This
procedure is presented in the Algorithm 4.
This heuristic significantly reduces the IWOC-SVM training task complexity while maintaining the accuracy of the
IWOC-SVM classification.
Proposed Work
18/29
Algorithm 4 Heuristic approach for model maintenance
1: Require: strong classifier H, test set Z = {z1,z2,...,zt }, weight distribution δ(z), user defined parameter time, user defined
parameter η.
2: i ← 1
3: repeat
4: if H(zi ) = 1 (background) then
5: Compute the margin m(zi ) by Eq. (10).
6: end if
7: if time is reached then
8: Order all the test instances according to their margin values, in ascending order.
9: The η smallest margin instances are selected as support vectors.
10: H(x) is updated using Z1 and its weight w ∼ δ(x).
11: end if
12: i ← i +1
13: until i > t
Summary
19/29
1 Introduction
2 Proposed Work
3 Experimental Results
4 Conclusion and Future Works
Experimental Results
20/29
Dataset
MSVS dataset (Benezeth et al., 2014) which consists of a set of 5 video sequences containing 7
multispectral bands and color video sequence (RGB).
Parameter Settings
IWOC-SVM with RBF (Radial Basis Function) kernel as a base classifier with C = 1.
The pool of classifiers was homogeneous and consisted of 10 base classifiers of the same type.
The classification threshold t1 was set to 0.9 and t2 to 0.5 for combining the best one-class classifiers.
Video sequences were resized to 160 ×120
The p* = 5 for the random subspace dimension from the original p = 26-dimensional.
Features: Color features (R,G,B, H,S,V and gray-scale), texture feature (XCS-LBP), color-texture
(OC-LBP), edge feature (gradient orientation and magnitude), motion feature (optical flow) and
multispectral bands (7 spectral narrow bands).
Experimental Results
21/29
MSVS dataset
Figure: Background subtraction results using the MSVS dataset – (a) original frame, (b) ground truth and (c) proposed method.
The true positives (TP) pixels are in white, true negatives (TN) pixels in black, false positives (FP) pixels in red and false negatives
(FN) pixels in green.
Experimental Results
22/29
Table: Performance of the different methods using the MSVS data set.
Videos Method Precision Recall F-score
Scene
01
MD (RGB)(Benezeth et al., 2014) 0.6536 0.6376 0.6536
MD (MSB)(Benezeth et al., 2014) 0.7850 0.8377 0.8105
Pooling (MSB)(Benezeth et al., 2014) 0.7475 0.8568 0.7984
Proposed 0.8500 0.9580 0.9008
Scene
02
MD (RGB)(Benezeth et al., 2014) 0.8346 0.9100 0.8707
MD (MSB)(Benezeth et al., 2014) 0.8549 0.9281 0.8900
Pooling (MSB)(Benezeth et al., 2014) 0.8639 0.8997 0.8815
Proposed 0.8277 0.8245 0.8727
Scene
03
MD (RGB)(Benezeth et al., 2014) 0.7494 0.5967 0.6644
MD (MSB)(Benezeth et al., 2014) 0.7533 0.6332 0.6889
Pooling (MSB)(Benezeth et al., 2014) 0.8809 0.5134 0.6487
Proposed 0.9056 0.9953 0.9483
Scene
04
MD(RGB)(Benezeth et al., 2014) 0.8402 0.7929 0.8158
MD (MSB)(Benezeth et al., 2014) 0.8430 0.8226 0.8327
Pooling (MSB)(Benezeth et al., 2014) 0.8146 0.8654 0.8392
Proposed 0.9534 0.8374 0.8997
Scene
05
MD (RGB)(Benezeth et al., 2014) 0.7359 0.7626 0.7490
MD (MSB)(Benezeth et al., 2014) 0.7341 0.8149 0.7724
Pooling (MSB)(Benezeth et al., 2014) 0.7373 0.8066 0.8066
Proposed 0.7316 0.8392 0.8400
*MD = Mahalanobis distance
Experimental Results
23/29
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Importance
Gray
Red
Green
Blue
Hue
Saturation
Value
XCS−LBP
OCLBPRR
OCLBPGG
CLBPBB
OCLBPRG
OCLBPRB
OCLBPGB
GradientX
GradientY
GradientMagnitude
GradientDirection
OpticalFlow
MS1
MS2
MS3
MS4
MS5
MS6
MS7
Scene 01
most (+): Gradient Direction with medium contrib. multispectral features less (-): OCLBP-GB
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Importance
Gray
Red
Green
Blue
Hue
Saturation
Value
XCS−LBP
OCLBPRR
OCLBPGG
CLBPBB
OCLBPRG
OCLBPRB
OCLBPGB
GradientX
GradientY
GradientMagnitude
GradientDirection
OpticalFlow
MS1
MS2
MS3
MS4
MS5
MS6
MS7
most (+): MS1,MS2 and MS6 with Color, Gradient X features less (-): XCS-LBP and MS4
Experimental Results
24/29
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Importance
Gray
Red
Green
Blue
Hue
Saturation
Value
XCS−LBP
OCLBPRR
OCLBPGG
CLBPBB
OCLBPRG
OCLBPRB
OCLBPGB
GradientX
GradientY
GradientMagnitude
GradientDirection
OpticalFlow
MS1
MS2
MS3
MS4
MS5
MS6
MS7
most (+): OCLBP-GG,RR with medium contrib. of other OCLBP channels and gradient features less (-): Hue, Optical flow and
multispectral features
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Importance
Gray
Red
Green
Blue
Hue
Saturation
Value
XCS−LBP
OCLBPRR
OCLBPGG
CLBPBB
OCLBPRG
OCLBPRB
OCLBPGB
GradientX
GradientY
GradientMagnitude
GradientDirection
OpticalFlow
MS1
MS2
MS3
MS4
MS5
MS6
MS7
most (+): OCLBP-BB,RR,RG and GG with medium contrib. of gradient features less (-):
Multispectral and color features
Experimental Results
25/29
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
1
Importance
Gray
Red
Green
Blue
Hue
Saturation
Value
XCS−LBP
OCLBPRR
OCLBPGG
CLBPBB
OCLBPRG
OCLBPRB
OCLBPGB
GradientX
GradientY
GradientMagnitude
GradientDirection
OpticalFlow
MS1
MS2
MS3
MS4
MS5
MS6
MS7
most (+): OCLBP-RR with high contrib. of other OCLBP channels and multispectral features less (-): Gradient Magnitude
Summary
26/29
1 Introduction
2 Proposed Work
3 Experimental Results
4 Conclusion and Future Works
Conclusion and Future Works
27/29
Conclusion
An incremental version of the WOC-SVM algorithm, called Incremental Weighted One-Class Support
Vector Machine (IWOC-SVM).
An online weighted version of random subspace (OW-RS) to increase the diversity of the classifiers pool.
A mechanism called Adaptive Importance Calculation (AIC) to suitably update the relative importance of
each feature over time.
A heuristic approach for IWOC-SVM model updating to improve speed.
Future Works
A superpixel segmentation strategy to improve the segmentation performance, increasing the
computational efficiency of our ensemble.
An unsupervised mechanism to suitably update the importance of each feature discarding insignificantly
features over time without ground-truth data.
28/29
THANK YOU!!!
References
29/29
Benezeth, Y., Sidibe, D., and Thomas, J. B. (2014). Background subtraction with multispectral video sequences. In IEEE International Conference on Robotics and
Automation (ICRA).
Guo, L. and Boukir, S. (2015). Fast data selection for SVM training using ensemble margin. Pattern Recognition Letters (PRL), 51.
Tax, D. and Duin, R. (2001). Combining one-class classifiers. In Proceedings of the Second International Workshop on Multiple Classifier Systems (MCS), pages
299–308.
Tax, D. and Laskov, P. (2003). Online SVM learning: from classification to data description and back. In IEEE Workshop on Neural Network for Signal Processing
(NNSP), pages 499–508.
Wozniak, M. (2013). Hybrid classifiers: methods of data, knowledge, and classifier combination. Springer Publishing Company, Incorporated.

More Related Content

What's hot

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
The Statistical and Applied Mathematical Sciences Institute
 
Svm vs ls svm
Svm vs ls svmSvm vs ls svm
Svm vs ls svm
Pulipaka Sai Ravi Teja
 
Semi-Supervised Regression using Cluster Ensemble
Semi-Supervised Regression using Cluster EnsembleSemi-Supervised Regression using Cluster Ensemble
Semi-Supervised Regression using Cluster Ensemble
Alexander Litvinenko
 
Time series Forecasting using svm
Time series Forecasting using  svmTime series Forecasting using  svm
Time series Forecasting using svm
Institute of Technology Telkom
 
MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4
arogozhnikov
 
MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1
arogozhnikov
 
ICPR 2012
ICPR 2012ICPR 2012
ICPR 2012
BOUWMANS Thierry
 
Additive model and boosting tree
Additive model and boosting treeAdditive model and boosting tree
Additive model and boosting tree
Dong Guo
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
The Statistical and Applied Mathematical Sciences Institute
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Universitat Politècnica de Catalunya
 
MLHEP 2015: Introductory Lecture #2
MLHEP 2015: Introductory Lecture #2MLHEP 2015: Introductory Lecture #2
MLHEP 2015: Introductory Lecture #2
arogozhnikov
 
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Universitat Politècnica de Catalunya
 
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector MachinesDongseo University
 
Svm V SVC
Svm V SVCSvm V SVC
Svm V SVC
AMR koura
 
20 k-means, k-center, k-meoids and variations
20 k-means, k-center, k-meoids and variations20 k-means, k-center, k-meoids and variations
20 k-means, k-center, k-meoids and variations
Andres Mendez-Vazquez
 
Support Vector Machines for Classification
Support Vector Machines for ClassificationSupport Vector Machines for Classification
Support Vector Machines for Classification
Prakash Pimpale
 
Detailed Description on Cross Entropy Loss Function
Detailed Description on Cross Entropy Loss FunctionDetailed Description on Cross Entropy Loss Function
Detailed Description on Cross Entropy Loss Function
범준 김
 
A BA-based algorithm for parameter optimization of support vector machine
A BA-based algorithm for parameter optimization of support vector machineA BA-based algorithm for parameter optimization of support vector machine
A BA-based algorithm for parameter optimization of support vector machine
Aboul Ella Hassanien
 
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen BoydH2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
Sri Ambati
 

What's hot (20)

Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
Program on Quasi-Monte Carlo and High-Dimensional Sampling Methods for Applie...
 
Svm vs ls svm
Svm vs ls svmSvm vs ls svm
Svm vs ls svm
 
Semi-Supervised Regression using Cluster Ensemble
Semi-Supervised Regression using Cluster EnsembleSemi-Supervised Regression using Cluster Ensemble
Semi-Supervised Regression using Cluster Ensemble
 
Time series Forecasting using svm
Time series Forecasting using  svmTime series Forecasting using  svm
Time series Forecasting using svm
 
MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4MLHEP 2015: Introductory Lecture #4
MLHEP 2015: Introductory Lecture #4
 
MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1MLHEP 2015: Introductory Lecture #1
MLHEP 2015: Introductory Lecture #1
 
ICPR 2012
ICPR 2012ICPR 2012
ICPR 2012
 
Additive model and boosting tree
Additive model and boosting treeAdditive model and boosting tree
Additive model and boosting tree
 
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
QMC Opening Workshop, High Accuracy Algorithms for Interpolating and Integrat...
 
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
 
MLHEP 2015: Introductory Lecture #2
MLHEP 2015: Introductory Lecture #2MLHEP 2015: Introductory Lecture #2
MLHEP 2015: Introductory Lecture #2
 
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018Backpropagation - Elisa Sayrol - UPC Barcelona 2018
Backpropagation - Elisa Sayrol - UPC Barcelona 2018
 
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
2013-1 Machine Learning Lecture 05 - Andrew Moore - Support Vector Machines
 
Svm V SVC
Svm V SVCSvm V SVC
Svm V SVC
 
20 k-means, k-center, k-meoids and variations
20 k-means, k-center, k-meoids and variations20 k-means, k-center, k-meoids and variations
20 k-means, k-center, k-meoids and variations
 
Support Vector Machines for Classification
Support Vector Machines for ClassificationSupport Vector Machines for Classification
Support Vector Machines for Classification
 
Detailed Description on Cross Entropy Loss Function
Detailed Description on Cross Entropy Loss FunctionDetailed Description on Cross Entropy Loss Function
Detailed Description on Cross Entropy Loss Function
 
A BA-based algorithm for parameter optimization of support vector machine
A BA-based algorithm for parameter optimization of support vector machineA BA-based algorithm for parameter optimization of support vector machine
A BA-based algorithm for parameter optimization of support vector machine
 
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen BoydH2O World - Consensus Optimization and Machine Learning - Stephen Boyd
H2O World - Consensus Optimization and Machine Learning - Stephen Boyd
 

Similar to ICPR 2016

2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…
2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…
2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…Dongseo University
 
Implementing Minimum Error Rate Classifier
Implementing Minimum Error Rate ClassifierImplementing Minimum Error Rate Classifier
Implementing Minimum Error Rate Classifier
Dipesh Shome
 
Fast Object Recognition from 3D Depth Data with Extreme Learning Machine
Fast Object Recognition from 3D Depth Data with Extreme Learning MachineFast Object Recognition from 3D Depth Data with Extreme Learning Machine
Fast Object Recognition from 3D Depth Data with Extreme Learning Machine
Soma Boubou
 
Clustering lect
Clustering lectClustering lect
Clustering lect
Shadi Nabil Albarqouni
 
The world of loss function
The world of loss functionThe world of loss function
The world of loss function
홍배 김
 
Anomaly detection using deep one class classifier
Anomaly detection using deep one class classifierAnomaly detection using deep one class classifier
Anomaly detection using deep one class classifier
홍배 김
 
Learning Convolutional Neural Networks for Graphs
Learning Convolutional Neural Networks for GraphsLearning Convolutional Neural Networks for Graphs
Learning Convolutional Neural Networks for Graphs
pione30
 
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHMADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
Wireilla
 
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHMADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ijfls
 
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdfCD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
RajJain516913
 
SVM (2).ppt
SVM (2).pptSVM (2).ppt
SVM (2).ppt
NoorUlHaq47
 
Dominance-Based Pareto-Surrogate for Multi-Objective Optimization
Dominance-Based Pareto-Surrogate for Multi-Objective OptimizationDominance-Based Pareto-Surrogate for Multi-Objective Optimization
Dominance-Based Pareto-Surrogate for Multi-Objective OptimizationIlya Loshchilov
 
Cheatsheet supervised-learning
Cheatsheet supervised-learningCheatsheet supervised-learning
Cheatsheet supervised-learning
Steve Nouri
 
ENBIS 2018 presentation on Deep k-Means
ENBIS 2018 presentation on Deep k-MeansENBIS 2018 presentation on Deep k-Means
ENBIS 2018 presentation on Deep k-Means
tthonet
 
4.Support Vector Machines.ppt machine learning and development
4.Support Vector Machines.ppt machine learning and development4.Support Vector Machines.ppt machine learning and development
4.Support Vector Machines.ppt machine learning and development
PriyankaRamavath3
 
Introduction to Big Data Science
Introduction to Big Data ScienceIntroduction to Big Data Science
Introduction to Big Data Science
Albert Bifet
 
Lecture7 cross validation
Lecture7 cross validationLecture7 cross validation
Lecture7 cross validationStéphane Canu
 
Chapter 11. Cluster Analysis Advanced Methods.ppt
Chapter 11. Cluster Analysis Advanced Methods.pptChapter 11. Cluster Analysis Advanced Methods.ppt
Chapter 11. Cluster Analysis Advanced Methods.ppt
Subrata Kumer Paul
 

Similar to ICPR 2016 (20)

2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…
2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…
2013-1 Machine Learning Lecture 06 - Artur Ferreira - A Survey on Boosting…
 
Implementing Minimum Error Rate Classifier
Implementing Minimum Error Rate ClassifierImplementing Minimum Error Rate Classifier
Implementing Minimum Error Rate Classifier
 
Fast Object Recognition from 3D Depth Data with Extreme Learning Machine
Fast Object Recognition from 3D Depth Data with Extreme Learning MachineFast Object Recognition from 3D Depth Data with Extreme Learning Machine
Fast Object Recognition from 3D Depth Data with Extreme Learning Machine
 
Clustering lect
Clustering lectClustering lect
Clustering lect
 
The world of loss function
The world of loss functionThe world of loss function
The world of loss function
 
Anomaly detection using deep one class classifier
Anomaly detection using deep one class classifierAnomaly detection using deep one class classifier
Anomaly detection using deep one class classifier
 
Learning Convolutional Neural Networks for Graphs
Learning Convolutional Neural Networks for GraphsLearning Convolutional Neural Networks for Graphs
Learning Convolutional Neural Networks for Graphs
 
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHMADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
 
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHMADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
ADAPTIVE FUZZY KERNEL CLUSTERING ALGORITHM
 
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdfCD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
CD504 CGM_Lab Manual_004e08d3838702ed11fc6d03cc82f7be.pdf
 
SVM (2).ppt
SVM (2).pptSVM (2).ppt
SVM (2).ppt
 
xldb-2015
xldb-2015xldb-2015
xldb-2015
 
Dominance-Based Pareto-Surrogate for Multi-Objective Optimization
Dominance-Based Pareto-Surrogate for Multi-Objective OptimizationDominance-Based Pareto-Surrogate for Multi-Objective Optimization
Dominance-Based Pareto-Surrogate for Multi-Objective Optimization
 
Cheatsheet supervised-learning
Cheatsheet supervised-learningCheatsheet supervised-learning
Cheatsheet supervised-learning
 
ENBIS 2018 presentation on Deep k-Means
ENBIS 2018 presentation on Deep k-MeansENBIS 2018 presentation on Deep k-Means
ENBIS 2018 presentation on Deep k-Means
 
4.Support Vector Machines.ppt machine learning and development
4.Support Vector Machines.ppt machine learning and development4.Support Vector Machines.ppt machine learning and development
4.Support Vector Machines.ppt machine learning and development
 
SVM.ppt
SVM.pptSVM.ppt
SVM.ppt
 
Introduction to Big Data Science
Introduction to Big Data ScienceIntroduction to Big Data Science
Introduction to Big Data Science
 
Lecture7 cross validation
Lecture7 cross validationLecture7 cross validation
Lecture7 cross validation
 
Chapter 11. Cluster Analysis Advanced Methods.ppt
Chapter 11. Cluster Analysis Advanced Methods.pptChapter 11. Cluster Analysis Advanced Methods.ppt
Chapter 11. Cluster Analysis Advanced Methods.ppt
 

Recently uploaded

PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
ControlCase
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
Product School
 
UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3
DianaGray10
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
James Anderson
 
When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...
Elena Simperl
 
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Ramesh Iyer
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
Elena Simperl
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
Jemma Hussein Allen
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
Safe Software
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Albert Hoitingh
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
91mobiles
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
DanBrown980551
 
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
Product School
 
Assuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyesAssuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyes
ThousandEyes
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
KatiaHIMEUR1
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
OnBoard
 
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Jeffrey Haguewood
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
Paul Groth
 
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Product School
 

Recently uploaded (20)

PCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase TeamPCI PIN Basics Webinar from the Controlcase Team
PCI PIN Basics Webinar from the Controlcase Team
 
How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...How world-class product teams are winning in the AI era by CEO and Founder, P...
How world-class product teams are winning in the AI era by CEO and Founder, P...
 
UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3UiPath Test Automation using UiPath Test Suite series, part 3
UiPath Test Automation using UiPath Test Suite series, part 3
 
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...
 
When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...When stars align: studies in data quality, knowledge graphs, and machine lear...
When stars align: studies in data quality, knowledge graphs, and machine lear...
 
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...
 
Knowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and backKnowledge engineering: from people to machines and back
Knowledge engineering: from people to machines and back
 
The Future of Platform Engineering
The Future of Platform EngineeringThe Future of Platform Engineering
The Future of Platform Engineering
 
FIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdfFIDO Alliance Osaka Seminar: Overview.pdf
FIDO Alliance Osaka Seminar: Overview.pdf
 
Essentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with ParametersEssentials of Automations: Optimizing FME Workflows with Parameters
Essentials of Automations: Optimizing FME Workflows with Parameters
 
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
Encryption in Microsoft 365 - ExpertsLive Netherlands 2024
 
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdfSmart TV Buyer Insights Survey 2024 by 91mobiles.pdf
Smart TV Buyer Insights Survey 2024 by 91mobiles.pdf
 
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...
 
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
From Siloed Products to Connected Ecosystem: Building a Sustainable and Scala...
 
Assuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyesAssuring Contact Center Experiences for Your Customers With ThousandEyes
Assuring Contact Center Experiences for Your Customers With ThousandEyes
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
 
Leading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdfLeading Change strategies and insights for effective change management pdf 1.pdf
Leading Change strategies and insights for effective change management pdf 1.pdf
 
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...
 
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMsTo Graph or Not to Graph Knowledge Graph Architectures and LLMs
To Graph or Not to Graph Knowledge Graph Architectures and LLMs
 
Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...Mission to Decommission: Importance of Decommissioning Products to Increase E...
Mission to Decommission: Importance of Decommissioning Products to Increase E...
 

ICPR 2016

  • 1. Online Weighted One-Class Ensemble for Feature Selection in Background/Foreground Separation Caroline Silva, Thierry Bouwmans, Carl Fr´elicot, Micka¨el Coustaty December 6th, 2016 - Mexico
  • 2. Summary 2/29 1 Introduction 2 Proposed Work 3 Experimental Results 4 Conclusion and Future Works
  • 4. Introduction 4/29 Background Subtraction Background subtraction is a widely used approach for detecting moving objects from static cameras. Figure: Moving object tracking results.
  • 5. Introduction 5/29 Why is it interesting today X1 X2 X3 X4X5 Figure: Given a complex scene with different regions X1, X2, X3, X4 and X5, each region can be characterized by different features such as: texture, color, texture-color, motion and edge. Common methods Same feature is used globally for the whole scene. Key challenges It requires a deep knowledge of the scene. It is possible to automatically select the most relevant features
  • 6. Introduction 6/29 Feature Selection Traditionally, feature selection methods can be categorized into three main groups: Wrapper-based methods • employ a classification algorithm as a “black box” for selecting a set of relevant features. Embedded methods • the feature selection is incorporated as part of the classification algorithm, such as in the decision trees or in neural networks. Filter-based methods • evaluate the relevance of the features based on a statistical measure estimated directly from the data. Ensemble-based approaches for feature selection
  • 7. Summary 7/29 1 Introduction 2 Proposed Work 3 Experimental Results 4 Conclusion and Future Works
  • 8. Proposed Work 8/29 N u m b e r o f fra m e s 200 150 Frames 100 50 0 Feature Importance Motion Gradient Color+Texture Features Texture Color 0.6 1 0.8 0.4 0 0.2 Importance Figure: Illustration of a complex scene (left) and its features importance over time. The bar-graph (right) shows the feature importance variations for a certain region of the scene along time. The color, texture, cor-texture, gradients and motion features are represented by pink, blue, pale orange, green and dark orange color, respectively. We propose an online weighted ensemble of one-class SVMs (Support Vector Machines) for foreground-background separation. It automatically selects the best features for different regions of the image, and the more relevant features are used for foreground segmentation.
  • 10. Proposed Work 10/29 An Online Weighted One-Class Ensemble for Feature Selection A. Generate multiple base models N u m b e r o f fra m e s This above approach increases the diversity of base classifiers since different weights of each random sub- space are taken to distinguish the decision boundaries computed by the classifiers. Algorithm 1 Generate multiple base background models 1: Require: IWOC-SVM training procedure, training sequence X, subspace dimension p∗, number of base classifiers M, weight distribution δ(x) 2: k ← 1 3: repeat 4: Sk ← SelectRandomSubspace(X,p∗) 5: Train k-th IWOC-SVM on Sk with respect to weights w ∼ δ(x) 6: k ← k+1 7: until k > M 8: Output: trained IWOC-SVM base classifiers Ψ = {Ψ1,Ψ2,...,ΨM }
  • 11. Proposed Work 11/29 How to Choose the Classifier? Bicego and Figueiredo, 2009 proposed a Weighted version that allow to use weights W = {w1,...,wN } in [0,1] for the data. R R R a wiξi target outlier xi Weighted One-Class SVM (WOC-SVM) Minimizing the hypersphere volume implies the minimization of R2 . To prevent the classifier from over-fitting with noisy data, slack variables ξi are introduced to allow some target points (respectively outliers) outside (respectively inside) the hypersphere. Therefore the problem is to minimize: Θ(a,R) = R2 +C N ∑ i=1 wi ξi (1) where C is a user-defined parameter that controls the trade-off between the volume and the number of target points rejected. The larger C, the less outliers in the hypersphere. The BS task requires adjust the learned model to the scene variations over time
  • 12. 12/29 Why update the background model?
  • 13. Proposed Work 13/29 An Incremental Weighted One-Class SVM (IWOC-SVM) We propose an IWOC-SVM which is closely related to the procedure proposed by Tax and Laskov (2003). Given new samples Z1 = {x1,x2,...,xs} and its respective weights not learned by the IWOC-SVM. Karush-Kuhn-Tucker (KKT) conditions: αi = 0 ⇒ ||xi −a||2 < R2 (2) 0 < αi < C ⇒ ||xi −a||2 = R2 (3) αi = C ⇒ ||xi −a||2 > R2 (4) The mathematical model can be defined as: R −θ ≤ ||x −a||≤ R (5) where θ ∈ [0,R] is relative to the distribution of previous training set. R R R a wiξi target outlier xi Algorithm 2 Incremental Weighted One-Class SVM 1: Require: Previous training set Z0, newly added training set Z1 and its respective weights 2: Train IWOC-SVM classifier on Z0, then split Z0 = SV0 ∪NSV0 3: Input new samples Z1. Put samples that violate KKT conditions in ZV 1 . If ZV 1 = /0, then goto 2. 4: Put samples from NSV0 that satisfy Eq. (5) into NSVS 0 . 5: Set Z0 = SV0 ∪NSVS 0 ∪ZV 1 and train IWOC-SVM classifier on Z0. 6: Output: IWOC-SVM classifier Ω and the new training set Z0. .
  • 14. Proposed Work 14/29 B. Selection of the best base classifiers Let λcorrect k (respectively λwrong k ) be the number of times a pixel was correctly (respectively uncorrectly) classified by the k-th (k = 1,...,M) base classifier from given ground truth data. Then, the corresponding error is given by: errork = λwrong k λcorrect k +λwrong k (6) Note that only the base classifiers that have the smallest errors are combined and used to differentiate the moving objects from the background model in the scene.
  • 15. 15/29 The Algorithm 3 was adapted from the algorithm proposed by (Wozniak, 2013) Algorithm 3 Adaptive Importance (AI) computation 1: Require: Strong classifier H, validation set (t1,y1),...,(tN ,yN ) where ti ∈ T, yi ∈ Y = 0,1 for background and foreground examples respectively, set of L best base classifiers Ψ = {Ψ1,Ψ2,...,ΨL}, learning rate parameter γ 2: Initialize all L best classifiers with importance: βl = 1 3: repeat 4: Classify ti using the strong classifier H according to Eq. (9) 5: for l = 1 : L do 6: Checks response of Ψl and calculates their errorl according to Eq. (6) 7: For each best classifier, Ψl , update the importance βl = βl (i −1)+ Pa(Ψl )−Pa(Ψl (i−1)) (N+γ) 8: where Pa(Ψl ) = 1 −errorl according to Eq.(6). 9: end for 10: until i < N 11: Normalize the importance β of each L best classifier. 12: Output: New importance assigned to the best classifiers β = {β1,β2,...,βL}
  • 16. Proposed Work 16/29 C. Background Detection Given an incoming pixel x to be classified, one can define a support function associated to the target class ω for each of the L best base classifiers: ∀l = 1,...,L Fl (x,ω) = 1 s1 exp(−d(x,a)/s2) (7) where d(x,a) is a distance metric from x to the center a of the target class ω, s1 is a normalization factor and s2 is a scale parameter. Each Fl (x,ω) is then compared to a threshold t1 to obtain the positive or negative class labels:∀l = 1,...,L cl (x,ω) = 1 if Fl (x,ω) t1 −1 otherwise (8) Comparing the weighted sum of theses L class labels as in (Tax and Duin, 2001) to another threshold t2 allows to define the strong classifier for x as follows: H(x) = 1 if 1 L ∑L l=1 βl cl (x,ω) t2 0 otherwise (9) A pixel x is classified as a background pixel if H(x) = 0.
  • 17. Proposed Work 17/29 D. Heuristic approach for Background Model Maintenance The Small Votes Instance Selection (SVIS) introduced by Guo and Boukir (Guo and Boukir, 2015) consists of an unsupervised ensemble margin that combines the first c(1) and second most voted class c(2) labels under the learned model. Let vc(1) and vc(2) denote the relative number of votes. Then the margin, taking value in [0,1] is: m(x) = vc(1) −vc(2) L (10) where L represents the number of best base classifiers in the ensemble. The first smallest margin instances are selected as support vector candidates. The strong model is updated by the first smallest margin instances. This procedure is presented in the Algorithm 4. This heuristic significantly reduces the IWOC-SVM training task complexity while maintaining the accuracy of the IWOC-SVM classification.
  • 18. Proposed Work 18/29 Algorithm 4 Heuristic approach for model maintenance 1: Require: strong classifier H, test set Z = {z1,z2,...,zt }, weight distribution δ(z), user defined parameter time, user defined parameter η. 2: i ← 1 3: repeat 4: if H(zi ) = 1 (background) then 5: Compute the margin m(zi ) by Eq. (10). 6: end if 7: if time is reached then 8: Order all the test instances according to their margin values, in ascending order. 9: The η smallest margin instances are selected as support vectors. 10: H(x) is updated using Z1 and its weight w ∼ δ(x). 11: end if 12: i ← i +1 13: until i > t
  • 19. Summary 19/29 1 Introduction 2 Proposed Work 3 Experimental Results 4 Conclusion and Future Works
  • 20. Experimental Results 20/29 Dataset MSVS dataset (Benezeth et al., 2014) which consists of a set of 5 video sequences containing 7 multispectral bands and color video sequence (RGB). Parameter Settings IWOC-SVM with RBF (Radial Basis Function) kernel as a base classifier with C = 1. The pool of classifiers was homogeneous and consisted of 10 base classifiers of the same type. The classification threshold t1 was set to 0.9 and t2 to 0.5 for combining the best one-class classifiers. Video sequences were resized to 160 ×120 The p* = 5 for the random subspace dimension from the original p = 26-dimensional. Features: Color features (R,G,B, H,S,V and gray-scale), texture feature (XCS-LBP), color-texture (OC-LBP), edge feature (gradient orientation and magnitude), motion feature (optical flow) and multispectral bands (7 spectral narrow bands).
  • 21. Experimental Results 21/29 MSVS dataset Figure: Background subtraction results using the MSVS dataset – (a) original frame, (b) ground truth and (c) proposed method. The true positives (TP) pixels are in white, true negatives (TN) pixels in black, false positives (FP) pixels in red and false negatives (FN) pixels in green.
  • 22. Experimental Results 22/29 Table: Performance of the different methods using the MSVS data set. Videos Method Precision Recall F-score Scene 01 MD (RGB)(Benezeth et al., 2014) 0.6536 0.6376 0.6536 MD (MSB)(Benezeth et al., 2014) 0.7850 0.8377 0.8105 Pooling (MSB)(Benezeth et al., 2014) 0.7475 0.8568 0.7984 Proposed 0.8500 0.9580 0.9008 Scene 02 MD (RGB)(Benezeth et al., 2014) 0.8346 0.9100 0.8707 MD (MSB)(Benezeth et al., 2014) 0.8549 0.9281 0.8900 Pooling (MSB)(Benezeth et al., 2014) 0.8639 0.8997 0.8815 Proposed 0.8277 0.8245 0.8727 Scene 03 MD (RGB)(Benezeth et al., 2014) 0.7494 0.5967 0.6644 MD (MSB)(Benezeth et al., 2014) 0.7533 0.6332 0.6889 Pooling (MSB)(Benezeth et al., 2014) 0.8809 0.5134 0.6487 Proposed 0.9056 0.9953 0.9483 Scene 04 MD(RGB)(Benezeth et al., 2014) 0.8402 0.7929 0.8158 MD (MSB)(Benezeth et al., 2014) 0.8430 0.8226 0.8327 Pooling (MSB)(Benezeth et al., 2014) 0.8146 0.8654 0.8392 Proposed 0.9534 0.8374 0.8997 Scene 05 MD (RGB)(Benezeth et al., 2014) 0.7359 0.7626 0.7490 MD (MSB)(Benezeth et al., 2014) 0.7341 0.8149 0.7724 Pooling (MSB)(Benezeth et al., 2014) 0.7373 0.8066 0.8066 Proposed 0.7316 0.8392 0.8400 *MD = Mahalanobis distance
  • 23. Experimental Results 23/29 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Importance Gray Red Green Blue Hue Saturation Value XCS−LBP OCLBPRR OCLBPGG CLBPBB OCLBPRG OCLBPRB OCLBPGB GradientX GradientY GradientMagnitude GradientDirection OpticalFlow MS1 MS2 MS3 MS4 MS5 MS6 MS7 Scene 01 most (+): Gradient Direction with medium contrib. multispectral features less (-): OCLBP-GB 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Importance Gray Red Green Blue Hue Saturation Value XCS−LBP OCLBPRR OCLBPGG CLBPBB OCLBPRG OCLBPRB OCLBPGB GradientX GradientY GradientMagnitude GradientDirection OpticalFlow MS1 MS2 MS3 MS4 MS5 MS6 MS7 most (+): MS1,MS2 and MS6 with Color, Gradient X features less (-): XCS-LBP and MS4
  • 24. Experimental Results 24/29 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Importance Gray Red Green Blue Hue Saturation Value XCS−LBP OCLBPRR OCLBPGG CLBPBB OCLBPRG OCLBPRB OCLBPGB GradientX GradientY GradientMagnitude GradientDirection OpticalFlow MS1 MS2 MS3 MS4 MS5 MS6 MS7 most (+): OCLBP-GG,RR with medium contrib. of other OCLBP channels and gradient features less (-): Hue, Optical flow and multispectral features 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Importance Gray Red Green Blue Hue Saturation Value XCS−LBP OCLBPRR OCLBPGG CLBPBB OCLBPRG OCLBPRB OCLBPGB GradientX GradientY GradientMagnitude GradientDirection OpticalFlow MS1 MS2 MS3 MS4 MS5 MS6 MS7 most (+): OCLBP-BB,RR,RG and GG with medium contrib. of gradient features less (-): Multispectral and color features
  • 26. Summary 26/29 1 Introduction 2 Proposed Work 3 Experimental Results 4 Conclusion and Future Works
  • 27. Conclusion and Future Works 27/29 Conclusion An incremental version of the WOC-SVM algorithm, called Incremental Weighted One-Class Support Vector Machine (IWOC-SVM). An online weighted version of random subspace (OW-RS) to increase the diversity of the classifiers pool. A mechanism called Adaptive Importance Calculation (AIC) to suitably update the relative importance of each feature over time. A heuristic approach for IWOC-SVM model updating to improve speed. Future Works A superpixel segmentation strategy to improve the segmentation performance, increasing the computational efficiency of our ensemble. An unsupervised mechanism to suitably update the importance of each feature discarding insignificantly features over time without ground-truth data.
  • 29. References 29/29 Benezeth, Y., Sidibe, D., and Thomas, J. B. (2014). Background subtraction with multispectral video sequences. In IEEE International Conference on Robotics and Automation (ICRA). Guo, L. and Boukir, S. (2015). Fast data selection for SVM training using ensemble margin. Pattern Recognition Letters (PRL), 51. Tax, D. and Duin, R. (2001). Combining one-class classifiers. In Proceedings of the Second International Workshop on Multiple Classifier Systems (MCS), pages 299–308. Tax, D. and Laskov, P. (2003). Online SVM learning: from classification to data description and back. In IEEE Workshop on Neural Network for Signal Processing (NNSP), pages 499–508. Wozniak, M. (2013). Hybrid classifiers: methods of data, knowledge, and classifier combination. Springer Publishing Company, Incorporated.