SlideShare a Scribd company logo
1 of 51
Download to read offline
Incremental and Multi-feature
Tensor Subspace Learning
applied for Background
Modeling and Subtraction
Andrews Sobral
Ph.D. Student, Computer Vision
Lab. L3I – Univ. de La Rochelle, France
Summary
▪ Context
▪ Introduction to tensors
▪ Tensor decomposition methods
▪ Incremental tensor learning
▪ Incremental SVD
▪ Related Works
▪ Proposed method
▪ Conclusions
Context
Background Subtraction
Initialize
Background
Model
frame model
Foreground
Detection
Background
Model
Maintenance
Background Subtraction
Basic methods, mean and variance over time
Fuzzy based methods
Statistical methods
Non-parametric methods
Neural and neuro-fuzzy methods
Matrix and Tensor Factorization methods:
Eigenspace-based methods (PCA / SVD)
-> RPCA, LRR, NMF - based methods
-> NTF, Tensor - based methods
https://github.com/andrewssobral/bgslibrary
LRSLibrary (MatLab)
BGSLibrary (C++)
https://github.com/andrewssobral/lrslibrary
Introduction to tensors
Background Subtraction
Initialize
Background
Model
frame model
Foreground
Detection
Background
Model
Maintenance
Introduction to tensors
▪ Tensors are simply mathematical objects that can be used to describe physical properties.
In fact tensors are merely a generalization of scalars, vectors and matrices; a scalar is a
zero rank tensor, a vector is a first rank tensor and a matrix is the second rank tensor.
Introduction to tensors
▪ Subarrays, tubes and slices of a 3rd order tensor.
Introduction to tensors
▪ Matricization of a 3rd order tensor.
1
10
19
28
37
46
51
1
9
17
25
33
41
48
1
9
17
25
33
41
48
kj
i
1
10
19
28
37
46
51
1
9
17
25
33
41
48
1
9
17
25
33
41
48
kj
i
1
10
19
28
37
46
51
1
9
17
25
33
41
48
1
9
17
25
33
41
48
kj
i
1
10
19
28
37
46
51
1
9
17
25
33
41
48
1
9
17
25
33
41
48
kj
i
1
10
19
28
37
46
51
1
9
17
25
33
41
48
1
9
17
25
33
41
48
kj
i
Frontal Vertical Horizontal
Horizontal, vertical and frontal slices from a 3rd order tensor
Introduction to tensors
▪ Unfolding a 3rd order tensor.
Introduction to tensors
▪ Tensor transposition
▪ While there is only one way transpose a matrix there are an exponential number of ways to
transpose an order-n tensor.
▪ The 3rd order case:
Tensor decomposition
methods
Tensor decomposition methods
▪ Approaches:
▪ Tucker / HOSVD
▪ CANDECOMP-PARAFAC (CP)
▪ Hierarchical Tucker (HT)
▪ Tensor-Train decomposition (TT)
▪ NTF (Non-negative Tensor Factorization)
▪ NTD (Non-negative Tucker Decomposition)
▪ NCP (Non-negative CP Decomposition)
References:
Tensor Decompositions and Applications (Kolda and Bader, 2008)
Tucker / HoSVD
CP
▪ The CP model is a special case of the Tucker model, where the core tensor is
superdiagonal and the number of components in the factor matrices is the same.
Solving by ALS (alternating least squares) framework
Tensor decomposition methods
▪ Softwares
▪ There are several MATLAB toolboxes available for dealing with tensors in CP and Tucker
decomposition, including the Tensor Toolbox, the N-way toolbox, the PLS Toolbox, and the
Tensorlab. The TT-Toolbox provides MATLAB classes covering tensors in TT and QTT
decomposition, as well as linear operators. There is also a Python implementation of the TT-toolbox
called ttpy. The htucker toolbox provides a MATLAB class representing a tensor in HT
decomposition.
Incremental tensor
learning
Incremental SVD
Singular Value Decomposition
▪ Formally, the singular value decomposition of an m×n real or complex matrix M is a
factorization of the form:
▪ where U is a m×m real or complex unitary matrix, Σ is an m×n rectangular diagonal matrix with
nonnegative real numbers on the diagonal, and V* (the conjugate transpose of V, or simply the
transpose of V if V is real) is an n×n real or complex unitary matrix. The diagonal entries Σi,i of Σ
are known as the singular values of M. The m columns of U and the n columns of V are called
the left-singular vectors and right-singular vectors of M, respectively.
-3 -2 -1 0 1 2 3
-8
-6
-4
-2
0
2
4
6
8
10
X Z
-3 -2 -1 0 1 2 3
-8
-6
-4
-2
0
2
4
6
8
10
Y Z Z
10 20 30 40
5
10
15
20
25
30
35
40
45
-8
-6
-4
-2
0
2
4
6
-4
-2
0
2
4
-4
-2
0
2
4
-10
-5
0
5
10
ORIGINAL
-4
-2
0
2
4
-4
-2
0
2
4
-10
-5
0
5
10
Z =  1
-4
-2
0
2
4
-4
-2
0
2
4
-10
-5
0
5
10
Z =  1 +  2
Incremental SVD
▪ Problem:
▪ The matrix factorization step in SVD is computationally very expensive.
▪ Solution:
▪ Have a small pre-computed SVD model, and build upon this model incrementally using inexpensive
techniques.
▪ Businger (1970) and Bunch and Nielsen (1978) are the first authors who have proposed to update
SVD sequentially with the arrival of more samples, i.e. appending/removing a row/column.
▪ Subsequently various approaches have been proposed to update the SVD more efficiently and
supporting new operations.
References:
Businger, P.A. Updating a singular value decomposition. 1970
Bunch, J.R.; Nielsen, C.P. Updating the singular value decomposition. 1978
Incremental SVD
F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5)
[U, S, V] = SVD( [ F(t), …, F(t+2) ] ) [U’, S’, V’] = iSVD([F(t+3), …, F(t+5)], [U,S,V])
F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5)
[U, S, V] = SVD( [ F(t), …, F(t+1) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V])
F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5)
[U, S, V] = SVD( [ F(t), …, F(t+2) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V])
F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5)
[U, S, V] = SVD( [ F(t), …, F(t+1) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V])
Incremental SVD
▪ Updating operation proposed by Sarwar et al. (2002):
References:
Incremental Singular Value Decomposition Algorithms for Highly Scalable Recommender Systems (Sarwar et al., 2002)
Incremental SVD
▪ Operations proposed by Matthew Brand (2006):
References:
Fast low-rank modifications of the thin singular value decomposition (Matthew Brand, 2006)
Incremental SVD
▪ Operations proposed by Melenchon and Martinez (2007):
References:
Efficiently Downdating, Composing and Splitting Singular Value Decompositions Preserving the Mean Information (Melenchón and Martínez, 2007)
Incremental SVD algorithms in Matlab
By Christopher Baker (Baker et al., 2012)
http://www.math.fsu.edu/~cbaker/IncPACK/
[Up,Sp,Vp] = SEQKL(A,k,t,[U S V])
Original version only supports the Updating operation
Added exponential forgetting factor to support Downdating operation
By David Ross (Ross et al., 2007)
http://www.cs.toronto.edu/~dross/ivt/
[U, D, mu, n] = sklm(data, U0, D0, mu0, n0, ff, K)
Supports mean-update, updating and downdating
By David Wingate (Matthew Brand, 2006)
http://web.mit.edu/~wingated/www/index.html
[Up,Sp,Vp] = svd_update(U,S,V,A,B,force_orth)
update the SVD to be [X + A'*B]=Up*Sp*Vp' (a general matrix update).
[Up,Sp,Vp] = addblock_svd_update(U,S,V,A,force_orth)
update the SVD to be [X A] = Up*Sp*Vp' (add columns [ie, new data points])
size of Vp increases
A must be square matrix
[Up,Sp,Vp] = rank_one_svd_update(U,S,V,a,b,force_orth)
update the SVD to be [X + a*b'] = Up*Sp*Vp' (that is, a general rank-one update.
This can be used to add columns, zero columns, change columns, recenter the
matrix, etc. ).
Incremental tensor learning
Incremental tensor learning
▪ Proposed methods by Sun et al. (2008)
▪ Dynamic Tensor Analysis (DTA)
▪ Streaming Tensor Analysis (STA)
▪ Window-based Tensor Analysis (WTA)
References:
Incremental tensor analysis: Theory and applications (Sun et al, 2008)
Incremental tensor learning
▪ Dynamic Tensor Analysis (DTA)
▪ [T, C] = DTA(Xnew, R, C, alpha)
▪ Approximately updates tensor PCA, according to new tensor Xnew, old variance matrices in C and he
specified dimensions in vector R. The input Xnew is a tensor. The result returned in T is a tucker tensor and C
is the cell array of new variance matrices.
Incremental tensor learning
▪ Streaming Tensor Analysis (STA)
▪ [Tnew, S] = STA(Xnew, R, T, alpha, samplingPercent)
▪ Approximately updates tensor PCA, according to new tensor Xnew, old tucker decomposition T and he
specified dimensions in vector R. The input Xnew is a tensor or sptensor. The result returned in T is a new
tucker tensor for Xnew, S has the energy vector along each mode.
Incremental tensor learning
▪ Window-based Tensor Analysis (WTA)
▪ [T, C] = WTA(Xnew, R, Xold, overlap, type, C)
▪ Compute window-based tensor decomposition, according to Xnew (Xold) the new (old) tensor window,
overlap is the number of overlapping tensors and C variance matrices except for the time mode of previous
tensor window and the specified dimensions in vector R, type can be 'tucker' (default) or 'parafac'. The input
Xnew (Xold) is a tensor,sptensor, where the first mode is time. The result returned in T is a tucker or kruskal
tensor depending on type and C is the cell array of new variance matrices.
Incremental tensor learning
▪ Proposed methods by Hu et al. (2011)
▪ Incremental rank-(R1,R2,R3) tensor-based subspace learning
▪ IRTSA-GBM (grayscale)
▪ IRTSA-CBM (color)
References:
Incremental Tensor Subspace Learning and Its Applications to Foreground Segmentation and Tracking (Hu et al., 2011)
Apply iSVD
SKLM
(Ross et al., 2007)
Incremental tensor learning
▪ IRTSA architecture
Incremental rank-(R1,R2,R3) tensor-based subspace learning (Hu et al., 2011)
streaming
video data
Background Modeling
Exponential moving average
LK
J(t+1)
new sub-frame
tensor
Â(t)
sub-tensor
APPLY STANDARD RANK-R SVD
UNFOLDING MODE-1,2,3
Set of N background images
For first N frames
low-rank
sub-tensor
model
B(t+1)
New background
sub-frame
Â(t+1)
B(t+1)
drop last frame
APPLY INCREMENTAL SVD
UNFOLDING MODE-1,2,3
updated sub-tensor
foreground mask
For the next
frames
For the next background
sub-frame
Â(t+1)
updated sub-tensor
UPDATE
SKLM
(Ross et al., 2007)
{bg|fg} = P( J[t+1] | U[1], U[2], V[3] )
Foreground Detection
Background Model
Initialization
Background Model
Maintenance
Proposed method
Proposed method
Performs feature extraction
in the sliding block
Build or update the
tensor model
Store the last N frames
in a sliding block
streaming
video data
Performs the iHoSVD to build
or update the low-rank model
foreground mask
Performs the
Foreground Detection
 
low-rank
model
values
pixels
features
 
remove the last frame
from sliding block
add first frame coming
from video stream
 
…
(a) (b) (c) (d) (e)
Proposed method
▪ Sliding block
remove the last frame from sliding block
add first frame coming from video
stream
 
Proposed method
▪ Building Tensor Model
▪ A total of 8 features are extracted:
▪ 1) red channel,
▪ 2) green channel,
▪ 3) blue channel,
▪ 4) gray-scale,
▪ 5) local binary patterns (LBP),
▪ 6) spatial gradients in horizontal direction,
▪ 7) spatial gradients in vertical direction, and
▪ 8) spatial gradients magnitude. values
pixels
features
tensor model  
…
Proposed method
▪ Incremental Higher-order Singular Value Decomposition
Fixed parameters:
• r1 = 1
• r2 = 8
• r3 = 2
• t1 = t2 = t3 = 0.01
It creates a low-rank tensor
model Lt with the dominant
singular subspaces of the
tensor model Tt.
Proposed method
▪ Foreground detection
For each new frame a weighted combination of similarity
measures is performed.
- Feature’s set extracted from input frame.
- Feature’s set extracted from low-rank model.
- Similarity measure for the feature n.
- Weighted combination of similarity measures.
The weights are chosen empirically:
w1 = w2 = w3 = w6 = w7 = w8 = 0.125, w_4 = 0.225, w_5 = 0.025
Threshold t = 0.5
Proposed method
▪ Architecture
Feature Extraction
+
iHoSVD
Initialize
Background Model
frame model
Foreground
Detection
Background Model
Maintenance
 
 
Low Rank
Tensor Model
Weighted
combination of
similarity
measures
 
x1
x3
x3
w1
w2
w3
yΣ ⁄
Proposed method
▪ Experimental results on BMC data set
Proposed method
▪ Experimental results on BMC data set
Proposed method
▪ Experimental results on BMC data set
Conclusions
▪ Experimental results shows that the proposed method achieves interesting results for
background subtraction task.
▪ However, additional features can be added, enabling a more robust model of the
background scene.
▪ In addition, the proposed foreground detection approach can be changed to automatically
selects the best features allowing an accurate foreground detection.
▪ Further research consists to improve the speed of low-rank decomposition for real-time
applications.
▪ Additional supports for dynamic backgrounds might be interesting for real and complex
scenes.
Incremental and Multi-feature Tensor Subspace Learning applied for Background Modeling and Subtraction

More Related Content

What's hot

Differential analyses of structures in HiC data
Differential analyses of structures in HiC dataDifferential analyses of structures in HiC data
Differential analyses of structures in HiC datatuxette
 
Convolutional Neural Network (CNN) presentation from theory to code in Theano
Convolutional Neural Network (CNN) presentation from theory to code in TheanoConvolutional Neural Network (CNN) presentation from theory to code in Theano
Convolutional Neural Network (CNN) presentation from theory to code in TheanoSeongwon Hwang
 
Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)Tatsuya Yokota
 
From RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphsFrom RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphstuxette
 
Convolutional networks and graph networks through kernels
Convolutional networks and graph networks through kernelsConvolutional networks and graph networks through kernels
Convolutional networks and graph networks through kernelstuxette
 
Robust Low-rank and Sparse Decomposition for Moving Object Detection
Robust Low-rank and Sparse Decomposition for Moving Object DetectionRobust Low-rank and Sparse Decomposition for Moving Object Detection
Robust Low-rank and Sparse Decomposition for Moving Object DetectionActiveEon
 
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...ijcsit
 
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural NetworksMasahiro Suzuki
 
Dictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix FactorizationDictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix Factorizationrecsysfr
 
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...Masahiro Suzuki
 
Investigating the 3D structure of the genome with Hi-C data analysis
Investigating the 3D structure of the genome with Hi-C data analysisInvestigating the 3D structure of the genome with Hi-C data analysis
Investigating the 3D structure of the genome with Hi-C data analysistuxette
 
Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Mostafa G. M. Mostafa
 
17 Machine Learning Radial Basis Functions
17 Machine Learning Radial Basis Functions17 Machine Learning Radial Basis Functions
17 Machine Learning Radial Basis FunctionsAndres Mendez-Vazquez
 
Tracking the tracker: Time Series Analysis in Python from First Principles
Tracking the tracker: Time Series Analysis in Python from First PrinciplesTracking the tracker: Time Series Analysis in Python from First Principles
Tracking the tracker: Time Series Analysis in Python from First Principleskenluck2001
 
Lecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural NetworksLecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural NetworksSang Jun Lee
 
A short and naive introduction to using network in prediction models
A short and naive introduction to using network in prediction modelsA short and naive introduction to using network in prediction models
A short and naive introduction to using network in prediction modelstuxette
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systemsrecsysfr
 

What's hot (20)

Polynomial Matrix Decompositions
Polynomial Matrix DecompositionsPolynomial Matrix Decompositions
Polynomial Matrix Decompositions
 
Differential analyses of structures in HiC data
Differential analyses of structures in HiC dataDifferential analyses of structures in HiC data
Differential analyses of structures in HiC data
 
Convolutional Neural Network (CNN) presentation from theory to code in Theano
Convolutional Neural Network (CNN) presentation from theory to code in TheanoConvolutional Neural Network (CNN) presentation from theory to code in Theano
Convolutional Neural Network (CNN) presentation from theory to code in Theano
 
Estimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample SetsEstimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample Sets
 
Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)Linked CP Tensor Decomposition (presented by ICONIP2012)
Linked CP Tensor Decomposition (presented by ICONIP2012)
 
From RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphsFrom RNN to neural networks for cyclic undirected graphs
From RNN to neural networks for cyclic undirected graphs
 
Convolutional networks and graph networks through kernels
Convolutional networks and graph networks through kernelsConvolutional networks and graph networks through kernels
Convolutional networks and graph networks through kernels
 
Robust Low-rank and Sparse Decomposition for Moving Object Detection
Robust Low-rank and Sparse Decomposition for Moving Object DetectionRobust Low-rank and Sparse Decomposition for Moving Object Detection
Robust Low-rank and Sparse Decomposition for Moving Object Detection
 
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...
COMPARISON OF WAVELET NETWORK AND LOGISTIC REGRESSION IN PREDICTING ENTERPRIS...
 
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
(DL輪読)Variational Dropout Sparsifies Deep Neural Networks
 
Dictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix FactorizationDictionary Learning for Massive Matrix Factorization
Dictionary Learning for Massive Matrix Factorization
 
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...
(DL hacks輪読) How to Train Deep Variational Autoencoders and Probabilistic Lad...
 
Investigating the 3D structure of the genome with Hi-C data analysis
Investigating the 3D structure of the genome with Hi-C data analysisInvestigating the 3D structure of the genome with Hi-C data analysis
Investigating the 3D structure of the genome with Hi-C data analysis
 
Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)Neural Networks: Radial Bases Functions (RBF)
Neural Networks: Radial Bases Functions (RBF)
 
17 Machine Learning Radial Basis Functions
17 Machine Learning Radial Basis Functions17 Machine Learning Radial Basis Functions
17 Machine Learning Radial Basis Functions
 
Tracking the tracker: Time Series Analysis in Python from First Principles
Tracking the tracker: Time Series Analysis in Python from First PrinciplesTracking the tracker: Time Series Analysis in Python from First Principles
Tracking the tracker: Time Series Analysis in Python from First Principles
 
Lecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural NetworksLecture 6: Convolutional Neural Networks
Lecture 6: Convolutional Neural Networks
 
Kaggle kenneth
Kaggle kennethKaggle kenneth
Kaggle kenneth
 
A short and naive introduction to using network in prediction models
A short and naive introduction to using network in prediction modelsA short and naive introduction to using network in prediction models
A short and naive introduction to using network in prediction models
 
Data-Driven Recommender Systems
Data-Driven Recommender SystemsData-Driven Recommender Systems
Data-Driven Recommender Systems
 

Similar to Incremental and Multi-feature Tensor Subspace Learning applied for Background Modeling and Subtraction

TENSOR DECOMPOSITION WITH PYTHON
TENSOR DECOMPOSITION WITH PYTHONTENSOR DECOMPOSITION WITH PYTHON
TENSOR DECOMPOSITION WITH PYTHONAndré Panisson
 
Sequential and parallel algorithm to find maximum flow on extended mixed netw...
Sequential and parallel algorithm to find maximum flow on extended mixed netw...Sequential and parallel algorithm to find maximum flow on extended mixed netw...
Sequential and parallel algorithm to find maximum flow on extended mixed netw...csandit
 
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...cscpconf
 
TensorFlow for IITians
TensorFlow for IITiansTensorFlow for IITians
TensorFlow for IITiansAshish Bansal
 
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQL
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQLModeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQL
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQLKostis Kyzirakos
 
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...SYRTO Project
 
Computing the Nucleon Spin from Lattice QCD
Computing the Nucleon Spin from Lattice QCDComputing the Nucleon Spin from Lattice QCD
Computing the Nucleon Spin from Lattice QCDChristos Kallidonis
 
DSP_FFT_150525.pptx
DSP_FFT_150525.pptxDSP_FFT_150525.pptx
DSP_FFT_150525.pptxHamedNassar5
 
Applied machine learning for search engine relevance 3
Applied machine learning for search engine relevance 3Applied machine learning for search engine relevance 3
Applied machine learning for search engine relevance 3Charles Martin
 
DSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptxDSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptxHamedNassar5
 
An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...IJMIT JOURNAL
 
An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...IJMIT JOURNAL
 
International Journal of Managing Information Technology (IJMIT)
International Journal of Managing Information Technology (IJMIT)International Journal of Managing Information Technology (IJMIT)
International Journal of Managing Information Technology (IJMIT)IJMIT JOURNAL
 
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...Nick Pruehs
 

Similar to Incremental and Multi-feature Tensor Subspace Learning applied for Background Modeling and Subtraction (20)

TENSOR DECOMPOSITION WITH PYTHON
TENSOR DECOMPOSITION WITH PYTHONTENSOR DECOMPOSITION WITH PYTHON
TENSOR DECOMPOSITION WITH PYTHON
 
Sequential and parallel algorithm to find maximum flow on extended mixed netw...
Sequential and parallel algorithm to find maximum flow on extended mixed netw...Sequential and parallel algorithm to find maximum flow on extended mixed netw...
Sequential and parallel algorithm to find maximum flow on extended mixed netw...
 
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...
SEQUENTIAL AND PARALLEL ALGORITHM TO FIND MAXIMUM FLOW ON EXTENDED MIXED NETW...
 
TensorFlow for IITians
TensorFlow for IITiansTensorFlow for IITians
TensorFlow for IITians
 
Ijetcas14 605
Ijetcas14 605Ijetcas14 605
Ijetcas14 605
 
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQL
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQLModeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQL
Modeling and Querying Metadata in the Semantic Sensor Web: stRDF and stSPARQL
 
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
Spillover Dynamics for Systemic Risk Measurement Using Spatial Financial Time...
 
06 recurrent neural_networks
06 recurrent neural_networks06 recurrent neural_networks
06 recurrent neural_networks
 
Computing the Nucleon Spin from Lattice QCD
Computing the Nucleon Spin from Lattice QCDComputing the Nucleon Spin from Lattice QCD
Computing the Nucleon Spin from Lattice QCD
 
Data Analysis Assignment Help
Data Analysis Assignment HelpData Analysis Assignment Help
Data Analysis Assignment Help
 
Triggering patterns of topology changes in dynamic attributed graphs
Triggering patterns of topology changes in dynamic attributed graphsTriggering patterns of topology changes in dynamic attributed graphs
Triggering patterns of topology changes in dynamic attributed graphs
 
Congrès SMAI 2019
Congrès SMAI 2019Congrès SMAI 2019
Congrès SMAI 2019
 
DSP_FFT_150525.pptx
DSP_FFT_150525.pptxDSP_FFT_150525.pptx
DSP_FFT_150525.pptx
 
Applied machine learning for search engine relevance 3
Applied machine learning for search engine relevance 3Applied machine learning for search engine relevance 3
Applied machine learning for search engine relevance 3
 
DSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptxDSP_DiscSignals_LinearS_150417.pptx
DSP_DiscSignals_LinearS_150417.pptx
 
An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...
 
Neural ODE
Neural ODENeural ODE
Neural ODE
 
An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...An improved spfa algorithm for single source shortest path problem using forw...
An improved spfa algorithm for single source shortest path problem using forw...
 
International Journal of Managing Information Technology (IJMIT)
International Journal of Managing Information Technology (IJMIT)International Journal of Managing Information Technology (IJMIT)
International Journal of Managing Information Technology (IJMIT)
 
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...
Implementation of Thorup's Linear Time Algorithm for Undirected Single Source...
 

More from ActiveEon

ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIAL
ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIALENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIAL
ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIALActiveEon
 
Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)ActiveEon
 
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...ActiveEon
 
Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...ActiveEon
 
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...ActiveEon
 
SPPRA'2013 Paper Presentation
SPPRA'2013 Paper PresentationSPPRA'2013 Paper Presentation
SPPRA'2013 Paper PresentationActiveEon
 
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...ActiveEon
 

More from ActiveEon (7)

ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIAL
ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIALENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIAL
ENGENHARIA DE COMPUTAÇÃO E INTELIGÊNCIA ARTIFICIAL
 
Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)Machine Learning for Dummies (without mathematics)
Machine Learning for Dummies (without mathematics)
 
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...
PhD Thesis Defense Presentation: Robust Low-rank and Sparse Decomposition for...
 
Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...Comparison of Matrix Completion Algorithms for Background Initialization in V...
Comparison of Matrix Completion Algorithms for Background Initialization in V...
 
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...
Double-constrained RPCA based on Saliency Maps for Foreground Detection in Au...
 
SPPRA'2013 Paper Presentation
SPPRA'2013 Paper PresentationSPPRA'2013 Paper Presentation
SPPRA'2013 Paper Presentation
 
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...
Classificação Automática do Estado do Trânsito Utilizando Propriedades Holíst...
 

Recently uploaded

Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPirithiRaju
 
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...Sérgio Sacani
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticssakshisoni2385
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfSumit Kumar yadav
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTSérgio Sacani
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bSérgio Sacani
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsSumit Kumar yadav
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.Nitya salvi
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Sérgio Sacani
 
DIFFERENCE IN BACK CROSS AND TEST CROSS
DIFFERENCE IN  BACK CROSS AND TEST CROSSDIFFERENCE IN  BACK CROSS AND TEST CROSS
DIFFERENCE IN BACK CROSS AND TEST CROSSLeenakshiTyagi
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxgindu3009
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​kaibalyasahoo82800
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfSumit Kumar yadav
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoSérgio Sacani
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxAArockiyaNisha
 
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service 🪡
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service  🪡CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service  🪡
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service 🪡anilsa9823
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfSumit Kumar yadav
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PPRINCE C P
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)Areesha Ahmad
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksSérgio Sacani
 

Recently uploaded (20)

Pests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdfPests of mustard_Identification_Management_Dr.UPR.pdf
Pests of mustard_Identification_Management_Dr.UPR.pdf
 
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
PossibleEoarcheanRecordsoftheGeomagneticFieldPreservedintheIsuaSupracrustalBe...
 
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceuticsPulmonary drug delivery system M.pharm -2nd sem P'ceutics
Pulmonary drug delivery system M.pharm -2nd sem P'ceutics
 
Botany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdfBotany 4th semester file By Sumit Kumar yadav.pdf
Botany 4th semester file By Sumit Kumar yadav.pdf
 
Disentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOSTDisentangling the origin of chemical differences using GHOST
Disentangling the origin of chemical differences using GHOST
 
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43bNightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
Nightside clouds and disequilibrium chemistry on the hot Jupiter WASP-43b
 
Botany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questionsBotany krishna series 2nd semester Only Mcq type questions
Botany krishna series 2nd semester Only Mcq type questions
 
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
❤Jammu Kashmir Call Girls 8617697112 Personal Whatsapp Number 💦✅.
 
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
Discovery of an Accretion Streamer and a Slow Wide-angle Outflow around FUOri...
 
DIFFERENCE IN BACK CROSS AND TEST CROSS
DIFFERENCE IN  BACK CROSS AND TEST CROSSDIFFERENCE IN  BACK CROSS AND TEST CROSS
DIFFERENCE IN BACK CROSS AND TEST CROSS
 
Presentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptxPresentation Vikram Lander by Vedansh Gupta.pptx
Presentation Vikram Lander by Vedansh Gupta.pptx
 
Nanoparticles synthesis and characterization​ ​
Nanoparticles synthesis and characterization​  ​Nanoparticles synthesis and characterization​  ​
Nanoparticles synthesis and characterization​ ​
 
Chemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdfChemistry 4th semester series (krishna).pdf
Chemistry 4th semester series (krishna).pdf
 
Isotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on IoIsotopic evidence of long-lived volcanism on Io
Isotopic evidence of long-lived volcanism on Io
 
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptxPhysiochemical properties of nanomaterials and its nanotoxicity.pptx
Physiochemical properties of nanomaterials and its nanotoxicity.pptx
 
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service 🪡
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service  🪡CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service  🪡
CALL ON ➥8923113531 🔝Call Girls Kesar Bagh Lucknow best Night Fun service 🪡
 
Botany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdfBotany 4th semester series (krishna).pdf
Botany 4th semester series (krishna).pdf
 
VIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C PVIRUSES structure and classification ppt by Dr.Prince C P
VIRUSES structure and classification ppt by Dr.Prince C P
 
GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)GBSN - Biochemistry (Unit 1)
GBSN - Biochemistry (Unit 1)
 
Formation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disksFormation of low mass protostars and their circumstellar disks
Formation of low mass protostars and their circumstellar disks
 

Incremental and Multi-feature Tensor Subspace Learning applied for Background Modeling and Subtraction

  • 1. Incremental and Multi-feature Tensor Subspace Learning applied for Background Modeling and Subtraction Andrews Sobral Ph.D. Student, Computer Vision Lab. L3I – Univ. de La Rochelle, France
  • 2. Summary ▪ Context ▪ Introduction to tensors ▪ Tensor decomposition methods ▪ Incremental tensor learning ▪ Incremental SVD ▪ Related Works ▪ Proposed method ▪ Conclusions
  • 5. Background Subtraction Basic methods, mean and variance over time Fuzzy based methods Statistical methods Non-parametric methods Neural and neuro-fuzzy methods Matrix and Tensor Factorization methods: Eigenspace-based methods (PCA / SVD) -> RPCA, LRR, NMF - based methods -> NTF, Tensor - based methods https://github.com/andrewssobral/bgslibrary LRSLibrary (MatLab) BGSLibrary (C++) https://github.com/andrewssobral/lrslibrary
  • 8. Introduction to tensors ▪ Tensors are simply mathematical objects that can be used to describe physical properties. In fact tensors are merely a generalization of scalars, vectors and matrices; a scalar is a zero rank tensor, a vector is a first rank tensor and a matrix is the second rank tensor.
  • 9. Introduction to tensors ▪ Subarrays, tubes and slices of a 3rd order tensor.
  • 10. Introduction to tensors ▪ Matricization of a 3rd order tensor.
  • 12. Introduction to tensors ▪ Unfolding a 3rd order tensor.
  • 13. Introduction to tensors ▪ Tensor transposition ▪ While there is only one way transpose a matrix there are an exponential number of ways to transpose an order-n tensor. ▪ The 3rd order case:
  • 15. Tensor decomposition methods ▪ Approaches: ▪ Tucker / HOSVD ▪ CANDECOMP-PARAFAC (CP) ▪ Hierarchical Tucker (HT) ▪ Tensor-Train decomposition (TT) ▪ NTF (Non-negative Tensor Factorization) ▪ NTD (Non-negative Tucker Decomposition) ▪ NCP (Non-negative CP Decomposition) References: Tensor Decompositions and Applications (Kolda and Bader, 2008)
  • 16.
  • 18. CP ▪ The CP model is a special case of the Tucker model, where the core tensor is superdiagonal and the number of components in the factor matrices is the same. Solving by ALS (alternating least squares) framework
  • 19.
  • 20.
  • 21. Tensor decomposition methods ▪ Softwares ▪ There are several MATLAB toolboxes available for dealing with tensors in CP and Tucker decomposition, including the Tensor Toolbox, the N-way toolbox, the PLS Toolbox, and the Tensorlab. The TT-Toolbox provides MATLAB classes covering tensors in TT and QTT decomposition, as well as linear operators. There is also a Python implementation of the TT-toolbox called ttpy. The htucker toolbox provides a MATLAB class representing a tensor in HT decomposition.
  • 24. Singular Value Decomposition ▪ Formally, the singular value decomposition of an m×n real or complex matrix M is a factorization of the form: ▪ where U is a m×m real or complex unitary matrix, Σ is an m×n rectangular diagonal matrix with nonnegative real numbers on the diagonal, and V* (the conjugate transpose of V, or simply the transpose of V if V is real) is an n×n real or complex unitary matrix. The diagonal entries Σi,i of Σ are known as the singular values of M. The m columns of U and the n columns of V are called the left-singular vectors and right-singular vectors of M, respectively.
  • 25. -3 -2 -1 0 1 2 3 -8 -6 -4 -2 0 2 4 6 8 10 X Z -3 -2 -1 0 1 2 3 -8 -6 -4 -2 0 2 4 6 8 10 Y Z Z 10 20 30 40 5 10 15 20 25 30 35 40 45 -8 -6 -4 -2 0 2 4 6 -4 -2 0 2 4 -4 -2 0 2 4 -10 -5 0 5 10 ORIGINAL -4 -2 0 2 4 -4 -2 0 2 4 -10 -5 0 5 10 Z =  1 -4 -2 0 2 4 -4 -2 0 2 4 -10 -5 0 5 10 Z =  1 +  2
  • 26. Incremental SVD ▪ Problem: ▪ The matrix factorization step in SVD is computationally very expensive. ▪ Solution: ▪ Have a small pre-computed SVD model, and build upon this model incrementally using inexpensive techniques. ▪ Businger (1970) and Bunch and Nielsen (1978) are the first authors who have proposed to update SVD sequentially with the arrival of more samples, i.e. appending/removing a row/column. ▪ Subsequently various approaches have been proposed to update the SVD more efficiently and supporting new operations. References: Businger, P.A. Updating a singular value decomposition. 1970 Bunch, J.R.; Nielsen, C.P. Updating the singular value decomposition. 1978
  • 27. Incremental SVD F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5) [U, S, V] = SVD( [ F(t), …, F(t+2) ] ) [U’, S’, V’] = iSVD([F(t+3), …, F(t+5)], [U,S,V]) F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5) [U, S, V] = SVD( [ F(t), …, F(t+1) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V]) F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5) [U, S, V] = SVD( [ F(t), …, F(t+2) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V]) F(t) F(t+1) F(t+2) F(t+3) F(t+4) F(t+5) [U, S, V] = SVD( [ F(t), …, F(t+1) ] ) [U’, S’, V’] = iSVD([F(t+1), …, F(t+3)], [U,S,V])
  • 28. Incremental SVD ▪ Updating operation proposed by Sarwar et al. (2002): References: Incremental Singular Value Decomposition Algorithms for Highly Scalable Recommender Systems (Sarwar et al., 2002)
  • 29. Incremental SVD ▪ Operations proposed by Matthew Brand (2006): References: Fast low-rank modifications of the thin singular value decomposition (Matthew Brand, 2006)
  • 30. Incremental SVD ▪ Operations proposed by Melenchon and Martinez (2007): References: Efficiently Downdating, Composing and Splitting Singular Value Decompositions Preserving the Mean Information (Melenchón and Martínez, 2007)
  • 31. Incremental SVD algorithms in Matlab By Christopher Baker (Baker et al., 2012) http://www.math.fsu.edu/~cbaker/IncPACK/ [Up,Sp,Vp] = SEQKL(A,k,t,[U S V]) Original version only supports the Updating operation Added exponential forgetting factor to support Downdating operation By David Ross (Ross et al., 2007) http://www.cs.toronto.edu/~dross/ivt/ [U, D, mu, n] = sklm(data, U0, D0, mu0, n0, ff, K) Supports mean-update, updating and downdating By David Wingate (Matthew Brand, 2006) http://web.mit.edu/~wingated/www/index.html [Up,Sp,Vp] = svd_update(U,S,V,A,B,force_orth) update the SVD to be [X + A'*B]=Up*Sp*Vp' (a general matrix update). [Up,Sp,Vp] = addblock_svd_update(U,S,V,A,force_orth) update the SVD to be [X A] = Up*Sp*Vp' (add columns [ie, new data points]) size of Vp increases A must be square matrix [Up,Sp,Vp] = rank_one_svd_update(U,S,V,a,b,force_orth) update the SVD to be [X + a*b'] = Up*Sp*Vp' (that is, a general rank-one update. This can be used to add columns, zero columns, change columns, recenter the matrix, etc. ).
  • 33. Incremental tensor learning ▪ Proposed methods by Sun et al. (2008) ▪ Dynamic Tensor Analysis (DTA) ▪ Streaming Tensor Analysis (STA) ▪ Window-based Tensor Analysis (WTA) References: Incremental tensor analysis: Theory and applications (Sun et al, 2008)
  • 34. Incremental tensor learning ▪ Dynamic Tensor Analysis (DTA) ▪ [T, C] = DTA(Xnew, R, C, alpha) ▪ Approximately updates tensor PCA, according to new tensor Xnew, old variance matrices in C and he specified dimensions in vector R. The input Xnew is a tensor. The result returned in T is a tucker tensor and C is the cell array of new variance matrices.
  • 35. Incremental tensor learning ▪ Streaming Tensor Analysis (STA) ▪ [Tnew, S] = STA(Xnew, R, T, alpha, samplingPercent) ▪ Approximately updates tensor PCA, according to new tensor Xnew, old tucker decomposition T and he specified dimensions in vector R. The input Xnew is a tensor or sptensor. The result returned in T is a new tucker tensor for Xnew, S has the energy vector along each mode.
  • 36. Incremental tensor learning ▪ Window-based Tensor Analysis (WTA) ▪ [T, C] = WTA(Xnew, R, Xold, overlap, type, C) ▪ Compute window-based tensor decomposition, according to Xnew (Xold) the new (old) tensor window, overlap is the number of overlapping tensors and C variance matrices except for the time mode of previous tensor window and the specified dimensions in vector R, type can be 'tucker' (default) or 'parafac'. The input Xnew (Xold) is a tensor,sptensor, where the first mode is time. The result returned in T is a tucker or kruskal tensor depending on type and C is the cell array of new variance matrices.
  • 37. Incremental tensor learning ▪ Proposed methods by Hu et al. (2011) ▪ Incremental rank-(R1,R2,R3) tensor-based subspace learning ▪ IRTSA-GBM (grayscale) ▪ IRTSA-CBM (color) References: Incremental Tensor Subspace Learning and Its Applications to Foreground Segmentation and Tracking (Hu et al., 2011) Apply iSVD SKLM (Ross et al., 2007)
  • 38. Incremental tensor learning ▪ IRTSA architecture
  • 39. Incremental rank-(R1,R2,R3) tensor-based subspace learning (Hu et al., 2011) streaming video data Background Modeling Exponential moving average LK J(t+1) new sub-frame tensor Â(t) sub-tensor APPLY STANDARD RANK-R SVD UNFOLDING MODE-1,2,3 Set of N background images For first N frames low-rank sub-tensor model B(t+1) New background sub-frame Â(t+1) B(t+1) drop last frame APPLY INCREMENTAL SVD UNFOLDING MODE-1,2,3 updated sub-tensor foreground mask For the next frames For the next background sub-frame Â(t+1) updated sub-tensor UPDATE SKLM (Ross et al., 2007) {bg|fg} = P( J[t+1] | U[1], U[2], V[3] ) Foreground Detection Background Model Initialization Background Model Maintenance
  • 41. Proposed method Performs feature extraction in the sliding block Build or update the tensor model Store the last N frames in a sliding block streaming video data Performs the iHoSVD to build or update the low-rank model foreground mask Performs the Foreground Detection   low-rank model values pixels features   remove the last frame from sliding block add first frame coming from video stream   … (a) (b) (c) (d) (e)
  • 42. Proposed method ▪ Sliding block remove the last frame from sliding block add first frame coming from video stream  
  • 43. Proposed method ▪ Building Tensor Model ▪ A total of 8 features are extracted: ▪ 1) red channel, ▪ 2) green channel, ▪ 3) blue channel, ▪ 4) gray-scale, ▪ 5) local binary patterns (LBP), ▪ 6) spatial gradients in horizontal direction, ▪ 7) spatial gradients in vertical direction, and ▪ 8) spatial gradients magnitude. values pixels features tensor model   …
  • 44. Proposed method ▪ Incremental Higher-order Singular Value Decomposition Fixed parameters: • r1 = 1 • r2 = 8 • r3 = 2 • t1 = t2 = t3 = 0.01 It creates a low-rank tensor model Lt with the dominant singular subspaces of the tensor model Tt.
  • 45. Proposed method ▪ Foreground detection For each new frame a weighted combination of similarity measures is performed. - Feature’s set extracted from input frame. - Feature’s set extracted from low-rank model. - Similarity measure for the feature n. - Weighted combination of similarity measures. The weights are chosen empirically: w1 = w2 = w3 = w6 = w7 = w8 = 0.125, w_4 = 0.225, w_5 = 0.025 Threshold t = 0.5
  • 46. Proposed method ▪ Architecture Feature Extraction + iHoSVD Initialize Background Model frame model Foreground Detection Background Model Maintenance     Low Rank Tensor Model Weighted combination of similarity measures   x1 x3 x3 w1 w2 w3 yΣ ⁄
  • 47. Proposed method ▪ Experimental results on BMC data set
  • 48. Proposed method ▪ Experimental results on BMC data set
  • 49. Proposed method ▪ Experimental results on BMC data set
  • 50. Conclusions ▪ Experimental results shows that the proposed method achieves interesting results for background subtraction task. ▪ However, additional features can be added, enabling a more robust model of the background scene. ▪ In addition, the proposed foreground detection approach can be changed to automatically selects the best features allowing an accurate foreground detection. ▪ Further research consists to improve the speed of low-rank decomposition for real-time applications. ▪ Additional supports for dynamic backgrounds might be interesting for real and complex scenes.