Six Myths about Ontologies: The Basics of Formal Ontology
Â
Kernel Entropy Component Analysis in Remote Sensing Data Clustering.pdf
1. Kernel Entropy Component Analysis
in Remote Sensing Data Clustering
Luis GĂłmez-Chova1 Robert Jenssen2 Gustavo Camps-Valls1
1 Image Processing Laboratory (IPL), Universitat de ValĂšncia, Spain.
luis.gomez-chova@uv.es , http://www.valencia.edu/chovago
2 Department of Physics and Technology, University of TromsĂž, Norway.
robert.jenssen@uit.no , http://www.phys.uit.no/âŒrobertj
IGARSS 2011 â Vancouver, Canada
*
IPL
Image Processing Laboratory
2. Intro ECA KECA Clustering Results Conclusions
Outline
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 1/26
3. Intro ECA KECA Clustering Results Conclusions
Motivation
Feature Extraction
Feature selection/extraction essential before classiïŹcation or regression
to discard redundant or noisy components
to reduce the dimensionality of the data
Create a subset of new features by combinations of the existing ones
Linear Feature Extraction
OïŹer Interpretability ⌠knowledge discovery
PCA: projections maximizing the data set variance
PLS: projections maximally aligned with the labels
ICA: non-orthogonal projections with maximal independent axes
Fail when data distributions are curved
Nonlinear feature relations
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 2/26
4. Intro ECA KECA Clustering Results Conclusions
Objectives
Objectives
Kernel-based non-linear data-transformation
Captures the data higher order statistics
Extracts features suited for clustering
Method
Kernel Entropy Component Analysis (KECA) [Jenssen, 2010]
Based on Information Theory:
Maximally preserves entropy of the input data
Angular clustering maximizes cluster divergence
Out-of-sample extension to deal with test data
Experiments
Cloud screening from ENVISAT/MERIS multispectral images
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 3/26
5. Intro ECA KECA Clustering Results Conclusions
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 4/26
6. Intro ECA KECA Clustering Results Conclusions
Information-Theoretic Learning
Entropy Concept
Entropy of a probability density function (pdf) is a measure of information
Entropy â Shape
of the pdf
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 5/26
7. Intro ECA KECA Clustering Results Conclusions
Information-Theoretic Learning
Divergence Concept
The entropy concept can be extended to obtain a measure of dissimilarity
between distributions
ââ
Divergence â Distance
between pdfs
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 6/26
12. Intro ECA KECA Clustering Results Conclusions
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 11/26
13. Intro ECA KECA Clustering Results Conclusions
Kernel Principal Component Analysis (KPCA)
Principal Component Analysis (PCA)
Find projections of X = [x1 , . . . , xN ] maximizing the variance of data XU
PCA: maximize: Trace{(XU) (XU)} = Trace{U Cxx U}
subject to: U U=I
Including Lagrange multipliers λ, this is equivalent to the eigenproblem
Cxx ui = λi ui â Cxx U = UD
ui are the eigenvectors of Cxx and they are orthonormal, ui uj = 0
PCA
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 12/26
14. Intro ECA KECA Clustering Results Conclusions
Kernel Principal Component Analysis (KPCA)
Kernel Principal Component Analysis (KPCA)
Find projections maximizing variance of mapped data [Ï(x1 ), . . . , Ï(xN )]
KPCA: maximize: Tr{(ΊU) (ΊU)} = Tr{U Ί ΊU}
subject to: U U=I
The covariance matrix Ί Ί and projection matrix U are dH à dH !!!
KPCA through kernel trick
Apply the representerâs theorem: U = Ί A where A = [α1 , . . . , αN ]
KPCA: maximize: Tr{A ΊΊ ΊΊ A} = Tr{A KKA}
subject to: U U = A ΊΊ A = A KA = I
Including Lagrange multipliers λ, this is equivalent to the eigenproblem
KKαi = λi Kαi â Kαi = λi αi
Now matrix A is N Ă N !!! (eigendecomposition of K = EDE = AA )
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 13/26
16. Intro ECA KECA Clustering Results Conclusions
Kernel ECA Transformation
KECA example
Original PCA KPCA KECA
KECA reveals cluster structure â underlying labels of the data
Nonlinearly related clusters in X â diïŹerent angular directions in H
An angular clustering based on the kernel features Ίeca seems reasonable
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 15/26
17. Intro ECA KECA Clustering Results Conclusions
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 16/26
18. Intro ECA KECA Clustering Results Conclusions
KECA Spectral Clustering
Cauchy-Schwarz divergence
The Cauchy-Schwarz divergence between the pdf of two clusters is
R
pi (x)pj (x)d x
DCS (pi , pj ) = â log(VCS (pi , pj )) = â log qR R
pi (x)d x pj2 (x)d x
2
Measuring dissimilarity in a probability space is a complex issue
1
Ï(xt ):
P
Entropy interpretation in the kernel space â mean vector ” = N
Z
1 1
V (p) = p 2 (x)dx = 2 1 K1 = 2 1 ΊΊ 1 = ” ” = ” 2
Ë Ë
N N
”i ”j
Diverg. via Parzen windowing â VCS (pi , pj ) =
Ë
”i ”j
= cos â (”i , ”j )
KECA Spectral Clustering
Angular clustering of Ίeca maximizes the CS divergence between clusters:
k
X
J(C1 , . . . , Ck ) = Ni cos â (Ïeca (x), ”i )
i =1
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 17/26
19. Intro ECA KECA Clustering Results Conclusions
KECA Spectral Clustering
KECA Spectral Clustering Algorithm
1 Obtain Ίeca by Kernel ECA
2 Initialize means ”i , i = 1, . . . , k
3 For all training samples assign a cluster
xt â Ci maximizing cos â (Ïeca (xt ), ”i )
4 Update mean vectors ”i CS
5 Repeat steps 3 and 4 until convergence
py
tro
En
Intuition
A kernel feature space data point Ïeca (xt ) is assigned to the cluster represented
by the closest mean vector ”i in terms of angular distance
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 18/26
20. Intro ECA KECA Clustering Results Conclusions
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 19/26
21. Intro ECA KECA Clustering Results Conclusions
Experimental results: Data material
Cloud masking from ENVISAT/MERIS multispectral images
Pixel-wise binary decisions about the presence/absence of clouds
MERIS images taken over Spain and France
Input samples with 13 spectral bands and 6 physically inspired features
Barrax (BR-2003-07-14) Barrax (BR-2004-07-14) France (FR-2005-03-19)
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 20/26
22. Intro ECA KECA Clustering Results Conclusions
Experimental results: Numerical comparison
Experimental setup
KECA compared with k-means, KPCA + k-means, and Kernel k-means
Number of clusters ïŹxed to k = 2 (cloud-free and cloudy areas)
Number of KPCA and KECA features ïŹxed to m = 2 (stress diïŹerences)
RBF-kernel width parameter is selected by gird-search for all methods
Numerical results
Validation results on 10000 pixels per image manually labeled
Kappa statistic results over 10 realizations for all images
BR-2003-07-14 BR-2004-07-14 FR-2005-03-19
1 0.8 0.6
0.5
0.9
0.7
Estimated Îș statistic
Estimated Îș statistic
Estimated Îș statistic
0.4
0.8
0.6 0.3
KECA
0.7 KPCA
0.2 Kernel k-means
0.5 k-means
0.6
0.1
0.5 0.4 0
200 400 600 800 1000 200 400 600 800 1000 200 400 600 800 1000
#Samples #Samples #Samples
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 21/26
23. Intro ECA KECA Clustering Results Conclusions
Experimental results: Numerical comparison
Average numerical results
0.8
0.7
Estimated Îș statistic
KECA
0.6 KPCA
Kernel k-means
k-means
0.5
0.4
200 400 600 800 1000
#Samples
KECA outperforms k-means (+25%) and Kk-means and KPCA (+15%)
In general, the number of training samples positively aïŹect the results
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 22/26
25. Intro ECA KECA Clustering Results Conclusions
1 Introduction
2 Entropy Component Analysis
3 Kernel Entropy Component Analysis (KECA)
4 KECA Spectral Clustering
5 Experimental Results
6 Conclusions and Open questions
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 24/26
26. Intro ECA KECA Clustering Results Conclusions
Conclusions and open questions
Conclusions
Kernel entropy component analysis for clustering remote sensing data
Nonlinear features preserving entropy of the input data
Angular clustering reveals structure in terms of clusters divergence
Out-of-sample extension for test data â mandatory in remote sensing
Good results on cloud screening from MERIS images
KECA code is available at http://www.phys.uit.no/âŒrobertj/
Simple feature extraction toolbox (SIMFEAT) soon at http://isp.uv.es
Open questions and Future work
Pre-images of transformed data in the input space
Learn kernel parameters in an automatic way
Test KECA in more remote sensing applications
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 25/26
27. Intro ECA KECA Clustering Results Conclusions
Kernel Entropy Component Analysis
in Remote Sensing Data Clustering
Luis GĂłmez-Chova1 Robert Jenssen2 Gustavo Camps-Valls1
1 Image Processing Laboratory (IPL), Universitat de ValĂšncia, Spain.
luis.gomez-chova@uv.es , http://www.valencia.edu/chovago
2 Department of Physics and Technology, University of TromsĂž, Norway.
robert.jenssen@uit.no , http://www.phys.uit.no/âŒrobertj
IGARSS 2011 â Vancouver, Canada
*
IPL
Image Processing Laboratory
L. GĂłmez-Chova et al. Kernel Entropy Component Analysis IGARSS 2011 â Vancouver 26/26