SlideShare a Scribd company logo
Unsupervised learning in Hyperspectral Classifiers using
Hidden Markov Models
Vikram Jayaram and Bryan Usevitch
Dept. of Electrical and Computer Engineering
The University of Texas at El Paso
500 W. University Ave., El Paso, TX 79968-0523
ABSTRACT
Hyperspectral data represents a mixture of several component spectra from many classifiable sources. The
knowledge of the contributions of the underlying sources to the recorded spectra is valuable in many remote
sensing applications. Traditional Hyperspectral classification and segmentation algorithms have used Markov
random field (MRF) based estimation in recent investigations. Although, this method reflects plausible local,
spatial correlation in the true scene, it is limited to using supervised learning schemes for parameter estimation.
Besides, the expectation-maximization (EM) for the hidden MRF is considerably more difficult to realize due
to the absence of a closed form formulation. In this paper, we propose a hidden Markov model (HMM) based
probability density function (PDF) classifier for reduced dimensional feature space. Our approach uses an
unsupervised learning scheme for maximum-likelihood (ML) parameter estimation that combines both model
selection and estimation in a single algorithm. The proposed method accurately models and synthesizes the
approximate observations of the true data in a reduced dimensional feature space.
Keywords: Unsupervised Learning, hidden Markov model, expectation-maximization, maximum-likelihood.

1. INTRODUCTION
Hyperspectral images exploit the fact that each material radiates a different amount of electromagnetic energy
throughout the spectra. This unique characteristic of the material is commonly known as the spectral signature
and we can read this signature from the images obtained by airborne or spaceborne-based detectors. The
bandwidth of these sensors ranges from the visible region (0.4-0.7 μm) through the near infrared (about 2.4 μm)
in hundreds of narrow contiguous bands about 10 nm wide.1
Classification of Hyperspectral imagery (HSI) data is a challenging problem for two main reasons. First,
limited spatial resolution of HSI sensors and/or the distance of the observed scene, the images invariably contain
pixels composed of several materials. It is desirable to resolve the contributions of the constituents from the
observed image without relying on high spatial resolution images. Remote sensing cameras have been designed to
capture a wide spectral range motivating the use of post-processing techniques to distinguish materials via their
spectral signatures. Secondly, available training data for most pattern recognition problems in HSI processing is
severely inadequate. Under the framework of statistical classifiers, Hughes2 was able to demonstrate the impact
of this problem on a theoretical basis. Concerning the second problem, feature extraction and optimal band
selection are the methods most commonly used for finding useful features in high-dimensional data.3, 4 On the
other hand reduced dimensionality algorithms suffer from theoretical loss of performance. This performance loss
occurs due to reduction of data to features, and further approximating the theoretical features to PDFs. Figure
1 hypothetically illustrates the different types of feature representation. However, it is beneficial to understand
the trade-off between need to retain as much information (increased feature space), and the need to obtain better
PDF estimates (reduced feature dimensionality) for a wide array of HSI applications.
In statistical pattern recognition, finite mixtures allow a probabilistic approach to unsupervised learning
(clustering).5, 6 Our problem of interest is to introduce a new unsupervised algorithm for learning a finite
Further author information: (Send correspondence to Vikram Jayaram)
V. Jayaram: E-mail: vjayaram@miners.utep.edu, Telephone: 1 915 747 5869
Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XV,
edited by Sylvia S. Shen, Paul E. Lewis, Proc. of SPIE Vol. 7334, 73340F · © 2009 SPIE
CCC code: 0277-786X/09/$18 · doi: 10.1117/12.820325
Proc. of SPIE Vol. 7334 73340F-1
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
mixture model from the multivariate HSI data using an HMM. This HMM approach estimates the proportion of
each HSI class present in a finite mixture model by incorporating both the estimation step and model selection
in a single algorithm. The model selection step that was previously introduced in7 automatically assigns mixture
components for a GM. Our technique utilizes a reduced dimensional feature space to model and synthesize the
approximate observations of the true HSI data. In order to define the relevance of using finite mixture model for
HSI, let us consider a random variable X, the finite mixture models decompose a PDF f (x) into sum of K class
PDFs. A general density function f (x) is considered K class probability density functions. A general density
function f (x) is considered semiparametric, since it may be decomposed into K components. Let fk (x) denote
the k th class PDF. The finite mixture model with K components expands as
K

ak fk (x),

f (x) =

(1)

k=1

where ak denotes the proportion of the k th class. The proportion ak may be interpreted as the prior probability of
observing a sample from class k. Furthermore, the prior probabilities ak for each distribution must be nonnegative
and sum-to-one, or
(2)
ak ≥ 0 f or k = 1, · · ·, K,
where

K

ak = 1.

(3)

k=1

Since, the underlying probability densities of the mixture are initially unknown, one must estimate the
densities from samples of each class iteratively. Thus, we formally extend the PDF based classification approach
to the analysis of HSI data (dependent data). In our approach we adapt a stationary Markovian model which
is a powerful stochastic model that can closely approximate many naturally occurring phenomena. One such
famous phenomena is approximation of human speech.8 While a very powerful stochastic model, a single HMM
cannot easily act as a good classifier between a wide variety of signal classes. Instead, it is best to design them
specifically for each signal type and feature type.
The rest of the paper is organized as follows. In Section II, we mention some examples of previous work in
literature that is related to our experiments. In Section III, we review some of the basics of HMM formulation,
problem of mixture learning and density estimation. Section IV describes briefly about minimum noise fraction
(MNF) transform. Section V reports experimental results, and Section VI ends the paper by presenting some
concluding remarks.

2. EARLIER WORK
Few investigations have introduced HMM in HSI processing in recent times. In Du et. al.,9 hidden Markov
model information divergence (HMMID) was introduced as a discriminatory measure among target spectra.
Comparison were made to deterministic distance metrics such as the spectral angle mapper (SAM) and minimum
Euclidean distance (MED). More recently, Bali et. al.10 shows the problem of joint segmentation of hyperspectral
images in the Bayesian framework. This approach based on a HMM of the images with common segmentation, or
equivalently with common hidden classification label variables which is modeled by a Potts Markov Random Field.
In a related work, Li et. al.11 proposed a two dimensional HMM for image classification. This method provided a
structured way to incorporate context information into classification. All the above mentioned approaches come
under the domain of supervised techniques. The modest development of unsupervised classification techniques
in the HSI regime has been the primary source of motivation for the proposed work.
Multidimensional data such as the HSI can be modeled by a multidimensional Gaussian mixture (GM).12
Normally, a GM in the form of the PDF for z ∈ RP is given by
L

αi N (z, μi , Σi )

p(z) =
i=1

Proc. of SPIE Vol. 7334 73340F-2
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
z

255

Feature space (3D)

Feature Vector

Scatter plot (20)

Figure 1. An illustration of Image and Feature space representation.

where
N (z, μi , Σi ) =

−1
1
1
e{− 2 (z−µi ) Σi (z−µi )} .
(2π)P/2 |Σi |1/2

Here L is the number of mixture components and P the number of spectral channels (bands). The GM parameters
are denoted by λ = {αi , μi , Σi }. The parameters of the GM are estimated using maximum likelihood by means
of the EM algorithm. In7 we show the structural learning of a GM that is employed to model and classify HSI
data. This methodology utilizes a fast and automatic assignment of mixture components to model PDFs. Later
on, we employ the same mechanism to estimate parameters and further model state PDFs of an HMM.
Consider a data that consists of K samples of dimension P , it is not necessary or even desirable to group all
the data together in to a single KXP -dimensional sample. In the simplest case, all K samples are independent
and we may regard them as samples of the same RV. For most practical cases, they are not independent. The
Markovian principle assumes consecutive samples are statistically independent when conditioned on knowing the
samples that preceded it. This leads to an elegant solution of HMM which employs a set of M PDFs of dimension
P . The HMM regards each of the K samples as having originated from one of the M possible states and there
is a distinct probability that the underlying model “jumps” from one state to another. In our approach, the
HMMs uses GM to model each state PDFs.5 We have focused on an unsupervised learning algorithm for ML
parameter estimation which in turn is used as a reduced dimensional PDF based classifier.

3. UNSUPERVISED LEARNING OF MARKOV SOURCES
In this section, we mention general formulation of HMM, re-estimation of HMM parameters, observed PDFs and
GM parameters. Let us begin following the notational approach of Rabiner,8 consider there are T observation
times. At each time 1 ≤ t ≤ T , there is a discrete state variable qt which takes one of N values qt ∈ {S1 , S2 , · ·
·, SN }. According to the Markovian assumption, the probability distribution of qt+1 depends only on the value
of qt . This is described compactly as a state transition probability matrix A whose elements aij represents the
probability that qt+1 equals j given that qt equals i. The initial state probabilities are denoted πi , the probability
that q1 equals Si . It is hidden Markov model because the states qt are hidden from view; that is we cannot
observe them. But, we can observe the random data Ot which is generated according to a PDF dependent on
the state at time t as illustrated in Figure 2. We denote the PDF of Ot under state j as bj (Ot ). The complete
set of model parameters that define the HMM are ∧ = {πj , aij , bj }.
The EM also known as the Baum-Welch algorithm calculates new estimates ∧ given an observation sequence O =
O1 , O2 , O3 , · · ·OT and a previous estimate of ∧. The algorithm is composed of two parts: the forward/backward
procedure and the re-estimation of parameters.
Using Gaussian Mixtures for bj (Ot )
We model the PDFs bj (Ot ) as GM,

Proc. of SPIE Vol. 7334 73340F-3
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
S

I
A

T

i.0 
2
30

0

0

0

a
0

0

0

p

p

14 0

0

43

40

S

V
I

I

I

Hidden

SthS

p (z)

4'

2:

1

I

p (z)

p (z)

2

p (z)

3

4

1

p (z)

5

Observer ® ®
Figure 2. A hidden Markov model. The observer makes his observations whose PDF depends on the state.

M

cjm N (O, μjm , Ujm ),

bj (O) =

1≤j≤N

m=1

where

N (O, μjm , Ujm ) =

1
P

2π 2 | Ujm |

− 1 (O−µjm )
2

e

U−1 (O−µjm )
jm

and P is the dimension of O. We will refer to these GM parameters collectively as bj

,
{cjm , μjm , Ujm }.

Forward/Backward Procedure
We wish to compute the probability of observation sequence O = O1 , O2 , ···, OT given the model ∧ = {πj , aij , bj }.
The forward procedure for p(O|∧) is
• Initialization:

1≤i≤N

αi = πi bi (O1 ),
• Induction:

N

αt+1 (j) =

1≤t≤T −1

αt (i)aij bj (Ot+1 ),
i=1

• Termination:

N

p(O|∧) =

αT (i)
i=1

The backward procedure is
• Initialization:
βt (i) = 1

Proc. of SPIE Vol. 7334 73340F-4
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms

1≤j≤N
• Induction:

N

βt (i) =

t = T − 1, T − 2, · · ··, 1

aij bj (Ot+1 )βt+1 (j),

1≤i≤N

j=1

Re-estimation of HMM parameters
The re-estimation procedure calculates new estimates of ∧ given the observation sequence O = O1 , O2 , O3 , ···OT .
We first define
ξt (i, j) =

N
i=1

αt (i)aij bj (Ot+1 )βt+1 (j)
N
j=1

αt (i)aij bj (Ot+1 )βt+1 (j)

and
N

γt (i) =

ξt (i, j).
j=1

The updated state priors are
πi = γ1 (i).
The updated state transition matrix is
T −1
t=1 ξt (i, j)
.
T −1
t=1 γt (i)

aij =
Re-estimation of Observation PDF’s13

In order to update the observation PDF’s, it is necessary to maximize
T

Qj =

wtj log bj (Ot )
t=1

over the PDF bj , where
αt (j)βt (j)
.
N
i=1 αt (i)βt (i)

wt,j =

This is a “weighted” likelihood (ML) procedure since if wtj = cj , the results are strict ML estimates. The weights
wtj are interpreted as the probability that the Markov chain is in state j at time t.
Re-estimation of Gaussian Mixture Parameters
In our experiments, bj (O) are modeled as GM by simply determining the weighted ML estimates of the GM
parameters. This would require iterating to convergence at each step. A more global approach is possible if the
mixture components assignments are regarded as “missing data”.13 The result is that the quantity
T

M

Qj =

γt (j, m) log bj (Ot )
t=1 m=1

is maximized, where

Proc. of SPIE Vol. 7334 73340F-5
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
cjm N (Ot , μjm , Ujm )
M
k=1 cjk N (Ot , μjm , Ujm )

γt (j, m) = wt,j

.

Here, the weights γt (j, m) are interpreted as the probability that the Markov chain is in state j and the
observation is from mixture component m at time t. The resulting update equations for cjm , μjm , and Ujm are
computed as follows:

cjm =
ˆ

T
t=1

T
t=1

γt (j, m)
M
l=1

γt (j, l)

.

The above expression is similar to re-estimation of GM.5 This means that the algorithms designed for GM
are applicable for updating the state PDFs of the HMM. Therefore,
T
t=1 γt (j, m)Ot
T
t=1 γt (j, m)

μjm =
ˆ

ˆ
Ujm =

T
t=1

γt (j, m)(Ot − μjm )(Ot − μjm )
T
t=1

γt (j, m)

.

4. MINIMUM NOISE FRACTION TRANSFORM
Before we begin the our section on experiments, we shall define minimum noise fraction (MNF) transform
since we use them to obtain a 2D feature plot of the true data as shown in Figure 3 (right). The MNF
transformation is a highly useful spectral processing tool in HSI analysis.14 It is used to determine the inherent
dimensionality of image data, to segregate noise in the data, and to reduce the computational requirements
for subsequent processing. This transform is essentially two cascaded principal components transformations.
The first transformation, based on an estimated noise covariance matrix, decorrelates and rescales the noise
in the data. This first step results in transformed data in which the noise has unit variance and no band-toband correlations. The second step is a standard principal components transformation of the noise-whitened
data. For the purposes of further spectral processing, the inherent dimensionality of the data is determined by
examination of the final eigenvalues and the associated images. The data space can be divided into two parts:
one part associated with large eigenvalues and coherent eigenimages, and a complementary part with near-unity
eigenvalues and noise-dominated images. By using only the coherent portions, the noise is separated from the
data, thus the image bands get ranked based on signal to noise ratios (SNR).

5. EXPERIMENTS
The remote sensing data sets that we have used in our experiments comes from an Airborne Visible/Infrared
Imaging Spectrometer (AVIRIS) sensor image. AVIRIS is a unique optical sensor that delivers calibrated images
of the upwelling spectral radiance in 224 contiguous spectral bands with wavelengths corresponding to 0.4-2.5
μm. AVIRIS is flown all across the US, Canada and Europe. Figure 3 shows data sets used in our experiments
that belong to a Indian Pine scene in northwest Indiana. The spatial bands of this scene are of size 169 X 169
pixels. Since, HSI imagery is highly correlated in the spectral direction using MNF rotation is an obvious choice
for decorrelation among the bands. This also results in a 2D “scatter” plot of the first two MNF components
of the data as shown in Figure 3. The scatter plots used in the paper are similar to marginalized PDF on
any 2D plane. Marginalization could be easily depicted for visualizing state PDFs of HMM. To illustrate this
visualization scheme, let z = [z1 , z2 , z3 , z4 ]. For example, to visualize on the (z2 , z4 ) plane, we would need to
compute
p(z2 , z4 ) =

z1

z3

p(z1 , z2 , z3 , z4 )dz1 dz3 .

Proc. of SPIE Vol. 7334 73340F-6
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
I
Figure 3. (Left) Composite image of Indian Pine scene. (Right) Composite image of MNF transformed scene.

This utility is very useful when visualizing high-dimensional PDF. On basis of an initial analysis based on
Iso-data and K-means unsupervised classifiers, it was found that the scene consisted of 3 prominent mixture
classes. Therefore, we begin the training by considering a tri-state (corresponding to the 3 mixture classes
identified)uniform state transition matrix A and prior probability π to initialize the HMM parameters. The
PDF of the feature vector in each state is approximated by Gaussian mixtures. The automatic learning and
initialization of the Gaussian mixtures are explicitly dealt in our earlier work.7 The algorithm outputs the total
log likelihood at each iteration.
Training an HMM is an iterative process that seeks to maximize the probability that the HMM accounts for
the example sequences. However, there are chances of running into a “local maximum” problem; the model,
though converged to some locally optimal choice of parameters, is not guaranteed to be the best possible model.
In an attempt to avoid this pitfall, we use a simulated annealing procedure along side of training. This step is
performed by expanding the covariance matrices of the PDF estimates and by pushing the state transition matrix
and prior state probabilities closer to “uniform”. We attempt to find a “bad” stationary point by re-running
the above sequence until one is found. The PDF plots of the three state PDF’s after convergence are shown in
Figures 5, 6 and 7. In our experiments, (both model and synthesis stage) we use Viterbi algorithm8 to estimate
the most likely state sequence. A few outliers are also observed in one or more state PDFs. Now that we have
mixtures modeled by their corresponding state PDFs, we would like to test the model by generating synthetic
observations. In Figure 8 we were able to synthesize 100 observation. We clearly notice that the synthetic
observations closely approximate the true data observations. This result is also exemplified in Figure 9 when we
compare true states of the data with the estimated states of the synthetic observations. Similarly, in Figures 10
and 11 we show instances that compare 300 and 600 synthetic observations to the true data. These comparisons
show that the underlying mixture density were adequately modeled using HMM.

6. CONCLUSIONS
In this paper, we proposed the use of hidden Markov model that uses structural learning for approximating
underlying mixture densities. Algorithm test were carried out using real Hyperspectral data consisting of a
scene from Indian pines of NW Indiana. In our experiments, we utilized only the first two components of MNF
transformed bands to ensure feature learning in reduced representation of the data. We show that mixture
learning for multivariate Gaussians is very similar to learning HMM parameters. In fact, unsupervised learning
of GM parameters for each class are seamlessly integrated to model the state PDFs of a HMM in a single
algorithm. This technique could be applied to any type of parameter mixture model that utilizes EM algorithm.
Our experiments exemplifies that the proposed method models and well synthesizes the observations of the
HSI data in a reduced dimensional feature space. This technique can considered a new paradigm of reduced
dimensional classifier in processing HSI data.

Proc. of SPIE Vol. 7334 73340F-7
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
40

7000

6000

30
7000

20

5000

4000

5000

10

4000

MNF

Band

2

6000

3000
3000

0

2000
2000
1000

−10
0

−10

0
−10

MNF Band 1

20

10
20

−20
−20

1000

10
0
30

30

0

MNF Band 2

−10

0

10

20

30

40

MNF Band 1

Figure 4. (Left) 2D scatter plot of MNF transformed Band 1 Vs. Band 2. (Right) 2D Histogram of MNF bands 1 and 2.

State 1

MNF2

40
20
0
−20
−30

−20

−10

0

10
MNF1

20

30

40

50

−20

−10

0

10
MNF1

20

30

40

50

MNF2

40
20
0
−20
−30

Figure 5. (Top) 2D scatter plot of true data. (Bottom) PDF of State 1 after convergence.

Proc. of SPIE Vol. 7334 73340F-8
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
State 2

MNF2

40
20
0
−20
−30

−20

−10

0

10
MNF1

20

30

40

50

−20

−10

0

10
MNF1

20

30

40

50

MNF2

40
20
0
−20
−30

Figure 6. (Top) 2D scatter plot of true data. (Bottom) PDF of State 2 after convergence.

State 3

MNF2

40
20
0
−20
−30

−20

−10

0

10
MNF1

20

30

40

50

−20

−10

0

10
MNF1

20

30

40

50

MNF2

40
20
0
−20
−30

Figure 7. (Top) 2D scatter plot of true data. (Bottom) PDF of State 3 after convergence.

Proc. of SPIE Vol. 7334 73340F-9
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
40

30
Original Samples
Synthetic Samples
MNF Band 2

20

10

0

−10

−20
−20

−10

0

10
MNF Band 1

20

30

40

Figure 8. Comparison of true data Vs. 100 synthetic observations.

3
2.8

True States
Estimates States

2.6
2.4

States

2.2
2
1.8
1.6
1.4
1.2
1
0

10

20

30

40

50

60

70

80

90

100

No.of Samples

Figure 9. Comparison of true states vs. estimated states from synthetic observation.

Proc. of SPIE Vol. 7334 73340F-10
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
40

30
Original Samples
Synthetic Samples

MNF Band 2

20

10

0

−10

−20
−20

−10

0

10

20

30

40

MNF Band 1

Figure 10. Comparison of true data Vs. 300 synthetic observations.

40

30
Original Samples
Synthetic Samples

MNF Band 2

20

10

0

−10

−20
−20

−10

0

10

20

30

MNF Band 1

Figure 11. Comparison of true data Vs. 600 synthetic observations.

Proc. of SPIE Vol. 7334 73340F-11
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms

40
ACKNOWLEDGMENTS
We would like to thank department of Geological Sciences at UTEP for providing access to the ENVI software
and LARS, Purdue University for making the HSI data15 available. This work was supported by NASA Earth
System Science (ESS) doctoral fellowship at the University of Texas at El Paso.

REFERENCES
[1] Schott, J. R., [Remote Sensing: The Image Chain Approach], Oxford University Press.
[2] Hughes, G. F., “On the mean accuracy of statistical pattern recognizers,” IEEE Transactions on Information
Theory 14, 55–63 (1968).
[3] Shaw, G. and Manolakis, D., “Signal processing for hyperspectral image exploitation,” IEEE Signal Processing Magazine 19, 12–16 (2002).
[4] Keshava, N., “Distance metrics & band selection in hyperspectral processing with applications to material
identification and spectral libraries,” IEEE Transactions on Geoscience and Remote Sensing 42, No. 7,
1552–1565 (July 2004).
[5] McLachlan, G. and Peel, D., [Finite Mixture Models], Wiley Series in Probability and Statistics, New York,
NY, second ed. (2000).
[6] Figueiredo, M. A. T. and Jain, A. K., “Unsupervised learning of finite mixture models,” IEEE Transactions
on Pattern Analysis and Machine Intelligence 24, 381–396 (2002).
[7] Jayaram, V. and Usevitch, B., “Dynamic mixing kernels in gaussian mixture classifier for hyperspectral
classification,” in [Mathematics of Data/Image Pattern Recognition, Compression, and Encryption with
Applications XI, Proceedings of the SPIE], 70750L–70750L–8 (2008).
[8] Rabiner, L. R., “A tutorial on hidden Markov models and selected applications in speech recognition,” in
[Proceedings of the IEEE], 257–286 (1989).
[9] Du, Q. and Chang, C.-I., “A hidden markov model approach to spectral analysis for hyperspectral imagery,”
Optical Engineering 40, No. 10, 2277–2284 (2001).
[10] Bali, N. and Mohammad-Djafari, A., “Bayesian approach with hidden Markov modeling and Mean Field
Approximation for Hyperspectral data analysis,” IEEE Transactions on Image Processing 17, No. 2, 217–
225 (2008).
[11] J. Li, A. N. and Gray, R. M., “Image classification by a two-dimensional hidden Markov model,” IEEE
Trans. Signal Processing 48, 517–533 (2000).
[12] Marden, D. B. and Manolakis, D. G., “Modeling hyperspectral imaging data,” in [Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery IX. Edited by Shen, Sylvia S.; Lewis, Paul
E. Proceedings of the SPIE], 253–262 (2003).
[13] Huang, B. H., “Maximum likelihood estimation for mixture multivariate stochastic observations of markov
chains,” in [AT&T Technical Journal], 1235–1249 (1985).
[14] A. A. Green, M. Berman, P. S. and Craig, M. D., “A transformation for ordering multispectral data in
terms of image quality with implications for noise removal,” IEEE Transcations on Geoscience and Remote
Sensing 26, 65–74 (1988).
[15] Landgrebe, D., “AVIRIS derived Northwest Indianas Indian Pines 1992 Hyperspectral dataset,”
http://dynamo.ecn.purdue.edu/ biehl/MultiSpec/documentation.html. .

Proc. of SPIE Vol. 7334 73340F-12
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms

More Related Content

What's hot

Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
RSIS International
 
F017533540
F017533540F017533540
F017533540
IOSR Journals
 
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
ijiert bestjournal
 
Multiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image SearchMultiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image Search
1crore projects
 
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence InformationLatent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
csandit
 
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
IJCSIS Research Publications
 
A survey on methods and applications of meta-learning with GNNs
A survey on methods and applications of meta-learning with GNNsA survey on methods and applications of meta-learning with GNNs
A survey on methods and applications of meta-learning with GNNs
Shreya Goyal
 
M.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing ProjectsM.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing Projects
Vijay Karan
 
M.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing ProjectsM.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing Projects
Vijay Karan
 
One Graduate Paper
One Graduate PaperOne Graduate Paper
One Graduate Paper
Colleen Farrelly
 
algorithms
algorithmsalgorithms
algorithms
DikshaGupta535173
 
M.E Computer Science Image Processing Projects
M.E Computer Science Image Processing ProjectsM.E Computer Science Image Processing Projects
M.E Computer Science Image Processing Projects
Vijay Karan
 
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
IRJET Journal
 
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
CSCJournals
 
Morse-Smale Regression for Risk Modeling
Morse-Smale Regression for Risk ModelingMorse-Smale Regression for Risk Modeling
Morse-Smale Regression for Risk Modeling
Colleen Farrelly
 
ME Synopsis
ME SynopsisME Synopsis
ME Synopsis
Poonam Debnath
 
Kernel based similarity estimation and real time tracking of moving
Kernel based similarity estimation and real time tracking of movingKernel based similarity estimation and real time tracking of moving
Kernel based similarity estimation and real time tracking of moving
IAEME Publication
 
A0310112
A0310112A0310112
A0310112
iosrjournals
 
Improved wolf algorithm on document images detection using optimum mean techn...
Improved wolf algorithm on document images detection using optimum mean techn...Improved wolf algorithm on document images detection using optimum mean techn...
Improved wolf algorithm on document images detection using optimum mean techn...
journalBEEI
 
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
IJAEMSJORNAL
 

What's hot (20)

Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...
 
F017533540
F017533540F017533540
F017533540
 
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...
 
Multiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image SearchMultiview Alignment Hashing for Efficient Image Search
Multiview Alignment Hashing for Efficient Image Search
 
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence InformationLatent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Information
 
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
A Kernel Approach for Semi-Supervised Clustering Framework for High Dimension...
 
A survey on methods and applications of meta-learning with GNNs
A survey on methods and applications of meta-learning with GNNsA survey on methods and applications of meta-learning with GNNs
A survey on methods and applications of meta-learning with GNNs
 
M.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing ProjectsM.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing Projects
 
M.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing ProjectsM.Phil Computer Science Image Processing Projects
M.Phil Computer Science Image Processing Projects
 
One Graduate Paper
One Graduate PaperOne Graduate Paper
One Graduate Paper
 
algorithms
algorithmsalgorithms
algorithms
 
M.E Computer Science Image Processing Projects
M.E Computer Science Image Processing ProjectsM.E Computer Science Image Processing Projects
M.E Computer Science Image Processing Projects
 
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...
 
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...
 
Morse-Smale Regression for Risk Modeling
Morse-Smale Regression for Risk ModelingMorse-Smale Regression for Risk Modeling
Morse-Smale Regression for Risk Modeling
 
ME Synopsis
ME SynopsisME Synopsis
ME Synopsis
 
Kernel based similarity estimation and real time tracking of moving
Kernel based similarity estimation and real time tracking of movingKernel based similarity estimation and real time tracking of moving
Kernel based similarity estimation and real time tracking of moving
 
A0310112
A0310112A0310112
A0310112
 
Improved wolf algorithm on document images detection using optimum mean techn...
Improved wolf algorithm on document images detection using optimum mean techn...Improved wolf algorithm on document images detection using optimum mean techn...
Improved wolf algorithm on document images detection using optimum mean techn...
 
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...
 

Viewers also liked

2009 asilomar
2009 asilomar2009 asilomar
A Rapid Location Independent Full Tensor Gravity Algorithm
A Rapid Location Independent Full Tensor Gravity AlgorithmA Rapid Location Independent Full Tensor Gravity Algorithm
A Rapid Location Independent Full Tensor Gravity Algorithm
Pioneer Natural Resources
 
A comparison of classification techniques for seismic facies recognition
A comparison of classification techniques for seismic facies recognitionA comparison of classification techniques for seismic facies recognition
A comparison of classification techniques for seismic facies recognition
Pioneer Natural Resources
 
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
Pioneer Natural Resources
 
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
Pioneer Natural Resources
 
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
Pioneer Natural Resources
 
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
Pioneer Natural Resources
 
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
Pioneer Natural Resources
 
Active learning algorithms in seismic facies classification
Active learning algorithms in seismic facies classificationActive learning algorithms in seismic facies classification
Active learning algorithms in seismic facies classification
Pioneer Natural Resources
 

Viewers also liked (9)

2009 asilomar
2009 asilomar2009 asilomar
2009 asilomar
 
A Rapid Location Independent Full Tensor Gravity Algorithm
A Rapid Location Independent Full Tensor Gravity AlgorithmA Rapid Location Independent Full Tensor Gravity Algorithm
A Rapid Location Independent Full Tensor Gravity Algorithm
 
A comparison of classification techniques for seismic facies recognition
A comparison of classification techniques for seismic facies recognitionA comparison of classification techniques for seismic facies recognition
A comparison of classification techniques for seismic facies recognition
 
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
Receiver deghosting method to mitigate F-­K transform artifacts: A non-­windo...
 
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...
 
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...
 
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
3D Gravity Modeling of Osage County Oklahoma for 3D Gravity Interpretation
 
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...
 
Active learning algorithms in seismic facies classification
Active learning algorithms in seismic facies classificationActive learning algorithms in seismic facies classification
Active learning algorithms in seismic facies classification
 

Similar to 2009 spie hmm

Iee egold2010 presentazione_finale_veracini
Iee egold2010 presentazione_finale_veraciniIee egold2010 presentazione_finale_veracini
Iee egold2010 presentazione_finale_veracini
grssieee
 
Ijetcas14 327
Ijetcas14 327Ijetcas14 327
Ijetcas14 327
Iasir Journals
 
IRJET-Multimodal Image Classification through Band and K-Means Clustering
IRJET-Multimodal Image Classification through Band and K-Means ClusteringIRJET-Multimodal Image Classification through Band and K-Means Clustering
IRJET-Multimodal Image Classification through Band and K-Means Clustering
IRJET Journal
 
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERWITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
kevig
 
Predicting electricity consumption using hidden parameters
Predicting electricity consumption using hidden parametersPredicting electricity consumption using hidden parameters
Predicting electricity consumption using hidden parameters
IJLT EMAS
 
2008 spie gmm
2008 spie gmm2008 spie gmm
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...
Q UANTUM  C LUSTERING -B ASED  F EATURE SUBSET  S ELECTION FOR MAMMOGRAPHIC I...Q UANTUM  C LUSTERING -B ASED  F EATURE SUBSET  S ELECTION FOR MAMMOGRAPHIC I...
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...
ijcsit
 
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
IOSR Journals
 
Research Proposal
Research ProposalResearch Proposal
Research Proposal
Komlan Atitey
 
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSOA Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
rahulmonikasharma
 
Oversampling technique in student performance classification from engineering...
Oversampling technique in student performance classification from engineering...Oversampling technique in student performance classification from engineering...
Oversampling technique in student performance classification from engineering...
IJECEIAES
 
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
ijsrd.com
 
Hybrid features selection method using random forest and meerkat clan algorithm
Hybrid features selection method using random forest and meerkat clan algorithmHybrid features selection method using random forest and meerkat clan algorithm
Hybrid features selection method using random forest and meerkat clan algorithm
TELKOMNIKA JOURNAL
 
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
Konstantinos Demertzis
 
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
ijaia
 
M.Phil Computer Science Remote Sensing Projects
M.Phil Computer Science Remote Sensing ProjectsM.Phil Computer Science Remote Sensing Projects
M.Phil Computer Science Remote Sensing Projects
Vijay Karan
 
M phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projectsM phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projects
Vijay Karan
 
M phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projectsM phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projects
Vijay Karan
 
A Neural Network Approach to Identify Hyperspectral Image Content
A Neural Network Approach to Identify Hyperspectral  Image Content A Neural Network Approach to Identify Hyperspectral  Image Content
A Neural Network Approach to Identify Hyperspectral Image Content
IJECEIAES
 
Equirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
Equirs: Explicitly Query Understanding Information Retrieval System Based on HmmEquirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
Equirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
International Journal of Engineering Inventions www.ijeijournal.com
 

Similar to 2009 spie hmm (20)

Iee egold2010 presentazione_finale_veracini
Iee egold2010 presentazione_finale_veraciniIee egold2010 presentazione_finale_veracini
Iee egold2010 presentazione_finale_veracini
 
Ijetcas14 327
Ijetcas14 327Ijetcas14 327
Ijetcas14 327
 
IRJET-Multimodal Image Classification through Band and K-Means Clustering
IRJET-Multimodal Image Classification through Band and K-Means ClusteringIRJET-Multimodal Image Classification through Band and K-Means Clustering
IRJET-Multimodal Image Classification through Band and K-Means Clustering
 
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERWITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSER
 
Predicting electricity consumption using hidden parameters
Predicting electricity consumption using hidden parametersPredicting electricity consumption using hidden parameters
Predicting electricity consumption using hidden parameters
 
2008 spie gmm
2008 spie gmm2008 spie gmm
2008 spie gmm
 
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...
Q UANTUM  C LUSTERING -B ASED  F EATURE SUBSET  S ELECTION FOR MAMMOGRAPHIC I...Q UANTUM  C LUSTERING -B ASED  F EATURE SUBSET  S ELECTION FOR MAMMOGRAPHIC I...
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...
 
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...
 
Research Proposal
Research ProposalResearch Proposal
Research Proposal
 
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSOA Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSO
 
Oversampling technique in student performance classification from engineering...
Oversampling technique in student performance classification from engineering...Oversampling technique in student performance classification from engineering...
Oversampling technique in student performance classification from engineering...
 
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...
 
Hybrid features selection method using random forest and meerkat clan algorithm
Hybrid features selection method using random forest and meerkat clan algorithmHybrid features selection method using random forest and meerkat clan algorithm
Hybrid features selection method using random forest and meerkat clan algorithm
 
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...
 
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...
 
M.Phil Computer Science Remote Sensing Projects
M.Phil Computer Science Remote Sensing ProjectsM.Phil Computer Science Remote Sensing Projects
M.Phil Computer Science Remote Sensing Projects
 
M phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projectsM phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projects
 
M phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projectsM phil-computer-science-remote-sensing-projects
M phil-computer-science-remote-sensing-projects
 
A Neural Network Approach to Identify Hyperspectral Image Content
A Neural Network Approach to Identify Hyperspectral  Image Content A Neural Network Approach to Identify Hyperspectral  Image Content
A Neural Network Approach to Identify Hyperspectral Image Content
 
Equirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
Equirs: Explicitly Query Understanding Information Retrieval System Based on HmmEquirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
Equirs: Explicitly Query Understanding Information Retrieval System Based on Hmm
 

More from Pioneer Natural Resources

Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
Pioneer Natural Resources
 
Interpretation Special-Section: Insights into digital oilfield data using ar...
 Interpretation Special-Section: Insights into digital oilfield data using ar... Interpretation Special-Section: Insights into digital oilfield data using ar...
Interpretation Special-Section: Insights into digital oilfield data using ar...
Pioneer Natural Resources
 
Machine learning for Seismic Data Analysis
Machine learning for Seismic Data AnalysisMachine learning for Seismic Data Analysis
Machine learning for Seismic Data Analysis
Pioneer Natural Resources
 
An approach to offer management: maximizing sales with fare products and anci...
An approach to offer management: maximizing sales with fare products and anci...An approach to offer management: maximizing sales with fare products and anci...
An approach to offer management: maximizing sales with fare products and anci...
Pioneer Natural Resources
 
2007 asprs
2007 asprs2007 asprs
2006 ssiai
2006 ssiai2006 ssiai
Detection and Classification in Hyperspectral Images using Rate Distortion an...
Detection and Classification in Hyperspectral Images using Rate Distortion an...Detection and Classification in Hyperspectral Images using Rate Distortion an...
Detection and Classification in Hyperspectral Images using Rate Distortion an...
Pioneer Natural Resources
 

More from Pioneer Natural Resources (7)

Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
Hydraulic Fracturing Stimulation Monitoring with Distributed Fiber Optic Sens...
 
Interpretation Special-Section: Insights into digital oilfield data using ar...
 Interpretation Special-Section: Insights into digital oilfield data using ar... Interpretation Special-Section: Insights into digital oilfield data using ar...
Interpretation Special-Section: Insights into digital oilfield data using ar...
 
Machine learning for Seismic Data Analysis
Machine learning for Seismic Data AnalysisMachine learning for Seismic Data Analysis
Machine learning for Seismic Data Analysis
 
An approach to offer management: maximizing sales with fare products and anci...
An approach to offer management: maximizing sales with fare products and anci...An approach to offer management: maximizing sales with fare products and anci...
An approach to offer management: maximizing sales with fare products and anci...
 
2007 asprs
2007 asprs2007 asprs
2007 asprs
 
2006 ssiai
2006 ssiai2006 ssiai
2006 ssiai
 
Detection and Classification in Hyperspectral Images using Rate Distortion an...
Detection and Classification in Hyperspectral Images using Rate Distortion an...Detection and Classification in Hyperspectral Images using Rate Distortion an...
Detection and Classification in Hyperspectral Images using Rate Distortion an...
 

Recently uploaded

GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
Neo4j
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
mikeeftimakis1
 
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
Neo4j
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
KatiaHIMEUR1
 
National Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practicesNational Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practices
Quotidiano Piemontese
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
Matthew Sinclair
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
Ana-Maria Mihalceanu
 
20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
Matthew Sinclair
 
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
Edge AI and Vision Alliance
 
How to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For FlutterHow to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For Flutter
Daiki Mogmet Ito
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
Neo4j
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
Alpen-Adria-Universität
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
SOFTTECHHUB
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
DianaGray10
 
RESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for studentsRESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for students
KAMESHS29
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Paige Cruz
 
“I’m still / I’m still / Chaining from the Block”
“I’m still / I’m still / Chaining from the Block”“I’m still / I’m still / Chaining from the Block”
“I’m still / I’m still / Chaining from the Block”
Claudio Di Ciccio
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Aggregage
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
Neo4j
 
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
SOFTTECHHUB
 

Recently uploaded (20)

GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
GraphSummit Singapore | Neo4j Product Vision & Roadmap - Q2 2024
 
Introduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - CybersecurityIntroduction to CHERI technology - Cybersecurity
Introduction to CHERI technology - Cybersecurity
 
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
GraphSummit Singapore | Graphing Success: Revolutionising Organisational Stru...
 
Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !Securing your Kubernetes cluster_ a step-by-step guide to success !
Securing your Kubernetes cluster_ a step-by-step guide to success !
 
National Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practicesNational Security Agency - NSA mobile device best practices
National Security Agency - NSA mobile device best practices
 
20240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 202420240609 QFM020 Irresponsible AI Reading List May 2024
20240609 QFM020 Irresponsible AI Reading List May 2024
 
Monitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR EventsMonitoring Java Application Security with JDK Tools and JFR Events
Monitoring Java Application Security with JDK Tools and JFR Events
 
20240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 202420240605 QFM017 Machine Intelligence Reading List May 2024
20240605 QFM017 Machine Intelligence Reading List May 2024
 
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
“Building and Scaling AI Applications with the Nx AI Manager,” a Presentation...
 
How to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For FlutterHow to use Firebase Data Connect For Flutter
How to use Firebase Data Connect For Flutter
 
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024GraphSummit Singapore | The Art of the  Possible with Graph - Q2 2024
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024
 
Video Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the FutureVideo Streaming: Then, Now, and in the Future
Video Streaming: Then, Now, and in the Future
 
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...
 
Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1Communications Mining Series - Zero to Hero - Session 1
Communications Mining Series - Zero to Hero - Session 1
 
RESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for studentsRESUME BUILDER APPLICATION Project for students
RESUME BUILDER APPLICATION Project for students
 
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfObservability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdf
 
“I’m still / I’m still / Chaining from the Block”
“I’m still / I’m still / Chaining from the Block”“I’m still / I’m still / Chaining from the Block”
“I’m still / I’m still / Chaining from the Block”
 
Generative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionGenerative AI Deep Dive: Advancing from Proof of Concept to Production
Generative AI Deep Dive: Advancing from Proof of Concept to Production
 
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
GraphSummit Singapore | Enhancing Changi Airport Group's Passenger Experience...
 
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!
 

2009 spie hmm

  • 1. Unsupervised learning in Hyperspectral Classifiers using Hidden Markov Models Vikram Jayaram and Bryan Usevitch Dept. of Electrical and Computer Engineering The University of Texas at El Paso 500 W. University Ave., El Paso, TX 79968-0523 ABSTRACT Hyperspectral data represents a mixture of several component spectra from many classifiable sources. The knowledge of the contributions of the underlying sources to the recorded spectra is valuable in many remote sensing applications. Traditional Hyperspectral classification and segmentation algorithms have used Markov random field (MRF) based estimation in recent investigations. Although, this method reflects plausible local, spatial correlation in the true scene, it is limited to using supervised learning schemes for parameter estimation. Besides, the expectation-maximization (EM) for the hidden MRF is considerably more difficult to realize due to the absence of a closed form formulation. In this paper, we propose a hidden Markov model (HMM) based probability density function (PDF) classifier for reduced dimensional feature space. Our approach uses an unsupervised learning scheme for maximum-likelihood (ML) parameter estimation that combines both model selection and estimation in a single algorithm. The proposed method accurately models and synthesizes the approximate observations of the true data in a reduced dimensional feature space. Keywords: Unsupervised Learning, hidden Markov model, expectation-maximization, maximum-likelihood. 1. INTRODUCTION Hyperspectral images exploit the fact that each material radiates a different amount of electromagnetic energy throughout the spectra. This unique characteristic of the material is commonly known as the spectral signature and we can read this signature from the images obtained by airborne or spaceborne-based detectors. The bandwidth of these sensors ranges from the visible region (0.4-0.7 μm) through the near infrared (about 2.4 μm) in hundreds of narrow contiguous bands about 10 nm wide.1 Classification of Hyperspectral imagery (HSI) data is a challenging problem for two main reasons. First, limited spatial resolution of HSI sensors and/or the distance of the observed scene, the images invariably contain pixels composed of several materials. It is desirable to resolve the contributions of the constituents from the observed image without relying on high spatial resolution images. Remote sensing cameras have been designed to capture a wide spectral range motivating the use of post-processing techniques to distinguish materials via their spectral signatures. Secondly, available training data for most pattern recognition problems in HSI processing is severely inadequate. Under the framework of statistical classifiers, Hughes2 was able to demonstrate the impact of this problem on a theoretical basis. Concerning the second problem, feature extraction and optimal band selection are the methods most commonly used for finding useful features in high-dimensional data.3, 4 On the other hand reduced dimensionality algorithms suffer from theoretical loss of performance. This performance loss occurs due to reduction of data to features, and further approximating the theoretical features to PDFs. Figure 1 hypothetically illustrates the different types of feature representation. However, it is beneficial to understand the trade-off between need to retain as much information (increased feature space), and the need to obtain better PDF estimates (reduced feature dimensionality) for a wide array of HSI applications. In statistical pattern recognition, finite mixtures allow a probabilistic approach to unsupervised learning (clustering).5, 6 Our problem of interest is to introduce a new unsupervised algorithm for learning a finite Further author information: (Send correspondence to Vikram Jayaram) V. Jayaram: E-mail: vjayaram@miners.utep.edu, Telephone: 1 915 747 5869 Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery XV, edited by Sylvia S. Shen, Paul E. Lewis, Proc. of SPIE Vol. 7334, 73340F · © 2009 SPIE CCC code: 0277-786X/09/$18 · doi: 10.1117/12.820325 Proc. of SPIE Vol. 7334 73340F-1 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 2. mixture model from the multivariate HSI data using an HMM. This HMM approach estimates the proportion of each HSI class present in a finite mixture model by incorporating both the estimation step and model selection in a single algorithm. The model selection step that was previously introduced in7 automatically assigns mixture components for a GM. Our technique utilizes a reduced dimensional feature space to model and synthesize the approximate observations of the true HSI data. In order to define the relevance of using finite mixture model for HSI, let us consider a random variable X, the finite mixture models decompose a PDF f (x) into sum of K class PDFs. A general density function f (x) is considered K class probability density functions. A general density function f (x) is considered semiparametric, since it may be decomposed into K components. Let fk (x) denote the k th class PDF. The finite mixture model with K components expands as K ak fk (x), f (x) = (1) k=1 where ak denotes the proportion of the k th class. The proportion ak may be interpreted as the prior probability of observing a sample from class k. Furthermore, the prior probabilities ak for each distribution must be nonnegative and sum-to-one, or (2) ak ≥ 0 f or k = 1, · · ·, K, where K ak = 1. (3) k=1 Since, the underlying probability densities of the mixture are initially unknown, one must estimate the densities from samples of each class iteratively. Thus, we formally extend the PDF based classification approach to the analysis of HSI data (dependent data). In our approach we adapt a stationary Markovian model which is a powerful stochastic model that can closely approximate many naturally occurring phenomena. One such famous phenomena is approximation of human speech.8 While a very powerful stochastic model, a single HMM cannot easily act as a good classifier between a wide variety of signal classes. Instead, it is best to design them specifically for each signal type and feature type. The rest of the paper is organized as follows. In Section II, we mention some examples of previous work in literature that is related to our experiments. In Section III, we review some of the basics of HMM formulation, problem of mixture learning and density estimation. Section IV describes briefly about minimum noise fraction (MNF) transform. Section V reports experimental results, and Section VI ends the paper by presenting some concluding remarks. 2. EARLIER WORK Few investigations have introduced HMM in HSI processing in recent times. In Du et. al.,9 hidden Markov model information divergence (HMMID) was introduced as a discriminatory measure among target spectra. Comparison were made to deterministic distance metrics such as the spectral angle mapper (SAM) and minimum Euclidean distance (MED). More recently, Bali et. al.10 shows the problem of joint segmentation of hyperspectral images in the Bayesian framework. This approach based on a HMM of the images with common segmentation, or equivalently with common hidden classification label variables which is modeled by a Potts Markov Random Field. In a related work, Li et. al.11 proposed a two dimensional HMM for image classification. This method provided a structured way to incorporate context information into classification. All the above mentioned approaches come under the domain of supervised techniques. The modest development of unsupervised classification techniques in the HSI regime has been the primary source of motivation for the proposed work. Multidimensional data such as the HSI can be modeled by a multidimensional Gaussian mixture (GM).12 Normally, a GM in the form of the PDF for z ∈ RP is given by L αi N (z, μi , Σi ) p(z) = i=1 Proc. of SPIE Vol. 7334 73340F-2 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 3. z 255 Feature space (3D) Feature Vector Scatter plot (20) Figure 1. An illustration of Image and Feature space representation. where N (z, μi , Σi ) = −1 1 1 e{− 2 (z−µi ) Σi (z−µi )} . (2π)P/2 |Σi |1/2 Here L is the number of mixture components and P the number of spectral channels (bands). The GM parameters are denoted by λ = {αi , μi , Σi }. The parameters of the GM are estimated using maximum likelihood by means of the EM algorithm. In7 we show the structural learning of a GM that is employed to model and classify HSI data. This methodology utilizes a fast and automatic assignment of mixture components to model PDFs. Later on, we employ the same mechanism to estimate parameters and further model state PDFs of an HMM. Consider a data that consists of K samples of dimension P , it is not necessary or even desirable to group all the data together in to a single KXP -dimensional sample. In the simplest case, all K samples are independent and we may regard them as samples of the same RV. For most practical cases, they are not independent. The Markovian principle assumes consecutive samples are statistically independent when conditioned on knowing the samples that preceded it. This leads to an elegant solution of HMM which employs a set of M PDFs of dimension P . The HMM regards each of the K samples as having originated from one of the M possible states and there is a distinct probability that the underlying model “jumps” from one state to another. In our approach, the HMMs uses GM to model each state PDFs.5 We have focused on an unsupervised learning algorithm for ML parameter estimation which in turn is used as a reduced dimensional PDF based classifier. 3. UNSUPERVISED LEARNING OF MARKOV SOURCES In this section, we mention general formulation of HMM, re-estimation of HMM parameters, observed PDFs and GM parameters. Let us begin following the notational approach of Rabiner,8 consider there are T observation times. At each time 1 ≤ t ≤ T , there is a discrete state variable qt which takes one of N values qt ∈ {S1 , S2 , · · ·, SN }. According to the Markovian assumption, the probability distribution of qt+1 depends only on the value of qt . This is described compactly as a state transition probability matrix A whose elements aij represents the probability that qt+1 equals j given that qt equals i. The initial state probabilities are denoted πi , the probability that q1 equals Si . It is hidden Markov model because the states qt are hidden from view; that is we cannot observe them. But, we can observe the random data Ot which is generated according to a PDF dependent on the state at time t as illustrated in Figure 2. We denote the PDF of Ot under state j as bj (Ot ). The complete set of model parameters that define the HMM are ∧ = {πj , aij , bj }. The EM also known as the Baum-Welch algorithm calculates new estimates ∧ given an observation sequence O = O1 , O2 , O3 , · · ·OT and a previous estimate of ∧. The algorithm is composed of two parts: the forward/backward procedure and the re-estimation of parameters. Using Gaussian Mixtures for bj (Ot ) We model the PDFs bj (Ot ) as GM, Proc. of SPIE Vol. 7334 73340F-3 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 4. S I A T i.0 2 30 0 0 0 a 0 0 0 p p 14 0 0 43 40 S V I I I Hidden SthS p (z) 4' 2: 1 I p (z) p (z) 2 p (z) 3 4 1 p (z) 5 Observer ® ® Figure 2. A hidden Markov model. The observer makes his observations whose PDF depends on the state. M cjm N (O, μjm , Ujm ), bj (O) = 1≤j≤N m=1 where N (O, μjm , Ujm ) = 1 P 2π 2 | Ujm | − 1 (O−µjm ) 2 e U−1 (O−µjm ) jm and P is the dimension of O. We will refer to these GM parameters collectively as bj , {cjm , μjm , Ujm }. Forward/Backward Procedure We wish to compute the probability of observation sequence O = O1 , O2 , ···, OT given the model ∧ = {πj , aij , bj }. The forward procedure for p(O|∧) is • Initialization: 1≤i≤N αi = πi bi (O1 ), • Induction: N αt+1 (j) = 1≤t≤T −1 αt (i)aij bj (Ot+1 ), i=1 • Termination: N p(O|∧) = αT (i) i=1 The backward procedure is • Initialization: βt (i) = 1 Proc. of SPIE Vol. 7334 73340F-4 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms 1≤j≤N
  • 5. • Induction: N βt (i) = t = T − 1, T − 2, · · ··, 1 aij bj (Ot+1 )βt+1 (j), 1≤i≤N j=1 Re-estimation of HMM parameters The re-estimation procedure calculates new estimates of ∧ given the observation sequence O = O1 , O2 , O3 , ···OT . We first define ξt (i, j) = N i=1 αt (i)aij bj (Ot+1 )βt+1 (j) N j=1 αt (i)aij bj (Ot+1 )βt+1 (j) and N γt (i) = ξt (i, j). j=1 The updated state priors are πi = γ1 (i). The updated state transition matrix is T −1 t=1 ξt (i, j) . T −1 t=1 γt (i) aij = Re-estimation of Observation PDF’s13 In order to update the observation PDF’s, it is necessary to maximize T Qj = wtj log bj (Ot ) t=1 over the PDF bj , where αt (j)βt (j) . N i=1 αt (i)βt (i) wt,j = This is a “weighted” likelihood (ML) procedure since if wtj = cj , the results are strict ML estimates. The weights wtj are interpreted as the probability that the Markov chain is in state j at time t. Re-estimation of Gaussian Mixture Parameters In our experiments, bj (O) are modeled as GM by simply determining the weighted ML estimates of the GM parameters. This would require iterating to convergence at each step. A more global approach is possible if the mixture components assignments are regarded as “missing data”.13 The result is that the quantity T M Qj = γt (j, m) log bj (Ot ) t=1 m=1 is maximized, where Proc. of SPIE Vol. 7334 73340F-5 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 6. cjm N (Ot , μjm , Ujm ) M k=1 cjk N (Ot , μjm , Ujm ) γt (j, m) = wt,j . Here, the weights γt (j, m) are interpreted as the probability that the Markov chain is in state j and the observation is from mixture component m at time t. The resulting update equations for cjm , μjm , and Ujm are computed as follows: cjm = ˆ T t=1 T t=1 γt (j, m) M l=1 γt (j, l) . The above expression is similar to re-estimation of GM.5 This means that the algorithms designed for GM are applicable for updating the state PDFs of the HMM. Therefore, T t=1 γt (j, m)Ot T t=1 γt (j, m) μjm = ˆ ˆ Ujm = T t=1 γt (j, m)(Ot − μjm )(Ot − μjm ) T t=1 γt (j, m) . 4. MINIMUM NOISE FRACTION TRANSFORM Before we begin the our section on experiments, we shall define minimum noise fraction (MNF) transform since we use them to obtain a 2D feature plot of the true data as shown in Figure 3 (right). The MNF transformation is a highly useful spectral processing tool in HSI analysis.14 It is used to determine the inherent dimensionality of image data, to segregate noise in the data, and to reduce the computational requirements for subsequent processing. This transform is essentially two cascaded principal components transformations. The first transformation, based on an estimated noise covariance matrix, decorrelates and rescales the noise in the data. This first step results in transformed data in which the noise has unit variance and no band-toband correlations. The second step is a standard principal components transformation of the noise-whitened data. For the purposes of further spectral processing, the inherent dimensionality of the data is determined by examination of the final eigenvalues and the associated images. The data space can be divided into two parts: one part associated with large eigenvalues and coherent eigenimages, and a complementary part with near-unity eigenvalues and noise-dominated images. By using only the coherent portions, the noise is separated from the data, thus the image bands get ranked based on signal to noise ratios (SNR). 5. EXPERIMENTS The remote sensing data sets that we have used in our experiments comes from an Airborne Visible/Infrared Imaging Spectrometer (AVIRIS) sensor image. AVIRIS is a unique optical sensor that delivers calibrated images of the upwelling spectral radiance in 224 contiguous spectral bands with wavelengths corresponding to 0.4-2.5 μm. AVIRIS is flown all across the US, Canada and Europe. Figure 3 shows data sets used in our experiments that belong to a Indian Pine scene in northwest Indiana. The spatial bands of this scene are of size 169 X 169 pixels. Since, HSI imagery is highly correlated in the spectral direction using MNF rotation is an obvious choice for decorrelation among the bands. This also results in a 2D “scatter” plot of the first two MNF components of the data as shown in Figure 3. The scatter plots used in the paper are similar to marginalized PDF on any 2D plane. Marginalization could be easily depicted for visualizing state PDFs of HMM. To illustrate this visualization scheme, let z = [z1 , z2 , z3 , z4 ]. For example, to visualize on the (z2 , z4 ) plane, we would need to compute p(z2 , z4 ) = z1 z3 p(z1 , z2 , z3 , z4 )dz1 dz3 . Proc. of SPIE Vol. 7334 73340F-6 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 7. I Figure 3. (Left) Composite image of Indian Pine scene. (Right) Composite image of MNF transformed scene. This utility is very useful when visualizing high-dimensional PDF. On basis of an initial analysis based on Iso-data and K-means unsupervised classifiers, it was found that the scene consisted of 3 prominent mixture classes. Therefore, we begin the training by considering a tri-state (corresponding to the 3 mixture classes identified)uniform state transition matrix A and prior probability π to initialize the HMM parameters. The PDF of the feature vector in each state is approximated by Gaussian mixtures. The automatic learning and initialization of the Gaussian mixtures are explicitly dealt in our earlier work.7 The algorithm outputs the total log likelihood at each iteration. Training an HMM is an iterative process that seeks to maximize the probability that the HMM accounts for the example sequences. However, there are chances of running into a “local maximum” problem; the model, though converged to some locally optimal choice of parameters, is not guaranteed to be the best possible model. In an attempt to avoid this pitfall, we use a simulated annealing procedure along side of training. This step is performed by expanding the covariance matrices of the PDF estimates and by pushing the state transition matrix and prior state probabilities closer to “uniform”. We attempt to find a “bad” stationary point by re-running the above sequence until one is found. The PDF plots of the three state PDF’s after convergence are shown in Figures 5, 6 and 7. In our experiments, (both model and synthesis stage) we use Viterbi algorithm8 to estimate the most likely state sequence. A few outliers are also observed in one or more state PDFs. Now that we have mixtures modeled by their corresponding state PDFs, we would like to test the model by generating synthetic observations. In Figure 8 we were able to synthesize 100 observation. We clearly notice that the synthetic observations closely approximate the true data observations. This result is also exemplified in Figure 9 when we compare true states of the data with the estimated states of the synthetic observations. Similarly, in Figures 10 and 11 we show instances that compare 300 and 600 synthetic observations to the true data. These comparisons show that the underlying mixture density were adequately modeled using HMM. 6. CONCLUSIONS In this paper, we proposed the use of hidden Markov model that uses structural learning for approximating underlying mixture densities. Algorithm test were carried out using real Hyperspectral data consisting of a scene from Indian pines of NW Indiana. In our experiments, we utilized only the first two components of MNF transformed bands to ensure feature learning in reduced representation of the data. We show that mixture learning for multivariate Gaussians is very similar to learning HMM parameters. In fact, unsupervised learning of GM parameters for each class are seamlessly integrated to model the state PDFs of a HMM in a single algorithm. This technique could be applied to any type of parameter mixture model that utilizes EM algorithm. Our experiments exemplifies that the proposed method models and well synthesizes the observations of the HSI data in a reduced dimensional feature space. This technique can considered a new paradigm of reduced dimensional classifier in processing HSI data. Proc. of SPIE Vol. 7334 73340F-7 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 8. 40 7000 6000 30 7000 20 5000 4000 5000 10 4000 MNF Band 2 6000 3000 3000 0 2000 2000 1000 −10 0 −10 0 −10 MNF Band 1 20 10 20 −20 −20 1000 10 0 30 30 0 MNF Band 2 −10 0 10 20 30 40 MNF Band 1 Figure 4. (Left) 2D scatter plot of MNF transformed Band 1 Vs. Band 2. (Right) 2D Histogram of MNF bands 1 and 2. State 1 MNF2 40 20 0 −20 −30 −20 −10 0 10 MNF1 20 30 40 50 −20 −10 0 10 MNF1 20 30 40 50 MNF2 40 20 0 −20 −30 Figure 5. (Top) 2D scatter plot of true data. (Bottom) PDF of State 1 after convergence. Proc. of SPIE Vol. 7334 73340F-8 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 9. State 2 MNF2 40 20 0 −20 −30 −20 −10 0 10 MNF1 20 30 40 50 −20 −10 0 10 MNF1 20 30 40 50 MNF2 40 20 0 −20 −30 Figure 6. (Top) 2D scatter plot of true data. (Bottom) PDF of State 2 after convergence. State 3 MNF2 40 20 0 −20 −30 −20 −10 0 10 MNF1 20 30 40 50 −20 −10 0 10 MNF1 20 30 40 50 MNF2 40 20 0 −20 −30 Figure 7. (Top) 2D scatter plot of true data. (Bottom) PDF of State 3 after convergence. Proc. of SPIE Vol. 7334 73340F-9 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 10. 40 30 Original Samples Synthetic Samples MNF Band 2 20 10 0 −10 −20 −20 −10 0 10 MNF Band 1 20 30 40 Figure 8. Comparison of true data Vs. 100 synthetic observations. 3 2.8 True States Estimates States 2.6 2.4 States 2.2 2 1.8 1.6 1.4 1.2 1 0 10 20 30 40 50 60 70 80 90 100 No.of Samples Figure 9. Comparison of true states vs. estimated states from synthetic observation. Proc. of SPIE Vol. 7334 73340F-10 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
  • 11. 40 30 Original Samples Synthetic Samples MNF Band 2 20 10 0 −10 −20 −20 −10 0 10 20 30 40 MNF Band 1 Figure 10. Comparison of true data Vs. 300 synthetic observations. 40 30 Original Samples Synthetic Samples MNF Band 2 20 10 0 −10 −20 −20 −10 0 10 20 30 MNF Band 1 Figure 11. Comparison of true data Vs. 600 synthetic observations. Proc. of SPIE Vol. 7334 73340F-11 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms 40
  • 12. ACKNOWLEDGMENTS We would like to thank department of Geological Sciences at UTEP for providing access to the ENVI software and LARS, Purdue University for making the HSI data15 available. This work was supported by NASA Earth System Science (ESS) doctoral fellowship at the University of Texas at El Paso. REFERENCES [1] Schott, J. R., [Remote Sensing: The Image Chain Approach], Oxford University Press. [2] Hughes, G. F., “On the mean accuracy of statistical pattern recognizers,” IEEE Transactions on Information Theory 14, 55–63 (1968). [3] Shaw, G. and Manolakis, D., “Signal processing for hyperspectral image exploitation,” IEEE Signal Processing Magazine 19, 12–16 (2002). [4] Keshava, N., “Distance metrics & band selection in hyperspectral processing with applications to material identification and spectral libraries,” IEEE Transactions on Geoscience and Remote Sensing 42, No. 7, 1552–1565 (July 2004). [5] McLachlan, G. and Peel, D., [Finite Mixture Models], Wiley Series in Probability and Statistics, New York, NY, second ed. (2000). [6] Figueiredo, M. A. T. and Jain, A. K., “Unsupervised learning of finite mixture models,” IEEE Transactions on Pattern Analysis and Machine Intelligence 24, 381–396 (2002). [7] Jayaram, V. and Usevitch, B., “Dynamic mixing kernels in gaussian mixture classifier for hyperspectral classification,” in [Mathematics of Data/Image Pattern Recognition, Compression, and Encryption with Applications XI, Proceedings of the SPIE], 70750L–70750L–8 (2008). [8] Rabiner, L. R., “A tutorial on hidden Markov models and selected applications in speech recognition,” in [Proceedings of the IEEE], 257–286 (1989). [9] Du, Q. and Chang, C.-I., “A hidden markov model approach to spectral analysis for hyperspectral imagery,” Optical Engineering 40, No. 10, 2277–2284 (2001). [10] Bali, N. and Mohammad-Djafari, A., “Bayesian approach with hidden Markov modeling and Mean Field Approximation for Hyperspectral data analysis,” IEEE Transactions on Image Processing 17, No. 2, 217– 225 (2008). [11] J. Li, A. N. and Gray, R. M., “Image classification by a two-dimensional hidden Markov model,” IEEE Trans. Signal Processing 48, 517–533 (2000). [12] Marden, D. B. and Manolakis, D. G., “Modeling hyperspectral imaging data,” in [Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery IX. Edited by Shen, Sylvia S.; Lewis, Paul E. Proceedings of the SPIE], 253–262 (2003). [13] Huang, B. H., “Maximum likelihood estimation for mixture multivariate stochastic observations of markov chains,” in [AT&T Technical Journal], 1235–1249 (1985). [14] A. A. Green, M. Berman, P. S. and Craig, M. D., “A transformation for ordering multispectral data in terms of image quality with implications for noise removal,” IEEE Transcations on Geoscience and Remote Sensing 26, 65–74 (1988). [15] Landgrebe, D., “AVIRIS derived Northwest Indianas Indian Pines 1992 Hyperspectral dataset,” http://dynamo.ecn.purdue.edu/ biehl/MultiSpec/documentation.html. . Proc. of SPIE Vol. 7334 73340F-12 Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms