This document discusses using hidden Markov models (HMMs) for unsupervised learning in hyperspectral image classification. It proposes an HMM-based probability density function classifier that models hyperspectral data using a reduced feature space. The approach uses an unsupervised learning scheme for maximum likelihood parameter estimation, combining both model selection and estimation. This HMM method can accurately model and synthesize approximate observations of true hyperspectral data in a reduced feature space without relying on supervised learning.
A Novel Approach to Mathematical Concepts in Data Miningijdmtaiir
-This paper describes three different fundamental
mathematical programming approaches that are relevant to
data mining. They are: Feature Selection, Clustering and
Robust Representation. This paper comprises of two clustering
algorithms such as K-mean algorithm and K-median
algorithms. Clustering is illustrated by the unsupervised
learning of patterns and clusters that may exist in a given
databases and useful tool for Knowledge Discovery in
Database (KDD). The results of k-median algorithm are used
to collecting the blood cancer patient from a medical database.
K-mean clustering is a data mining/machine learning algorithm
used to cluster observations into groups of related observations
without any prior knowledge of those relationships. The kmean algorithm is one of the simplest clustering techniques
and it is commonly used in medical imaging, biometrics and
related fields.
An Efficient Clustering Method for Aggregation on Data FragmentsIJMER
Clustering is an important step in the process of data analysis with applications to numerous fields. Clustering ensembles, has emerged as a powerful technique for combining different clustering results to obtain a quality cluster. Existing clustering aggregation algorithms are applied directly to large number of data points. The algorithms are inefficient if the number of data points is large. This project defines an efficient approach for clustering aggregation based on data fragments. In fragment-based approach, a data fragment is any subset of the data. To increase the efficiency of the proposed approach, the clustering aggregation can be performed directly on data fragments under comparison measure and normalized mutual information measures for clustering aggregation, enhanced clustering aggregation algorithms are described. To show the minimal computational complexity. (Agglomerative, Furthest, and Local Search); nevertheless, which increases the accuracy.
In this video from the 2015 HPC User Forum in Broomfield, Barry Bolding from Cray presents: HPC + D + A = HPDA?
"The flexible, multi-use Cray Urika-XA extreme analytics platform addresses perhaps the most critical obstacle in data analytics today — limitation. Analytics problems are getting more varied and complex but the available solution technologies have significant constraints. Traditional analytics appliances lock you into a single approach and building a custom solution in-house is so difficult and time consuming that the business value derived from analytics fails to materialize. In contrast, the Urika-XA platform is open, high performing and cost effective, serving a wide range of analytics tools with varying computing demands in a single environment. Pre-integrated with the Hadoop and Spark frameworks, the Urika-XA system combines the benefits of a turnkey analytics appliance with a flexible, open platform that you can modify for future analytics workloads. This single-platform consolidation of workloads reduces your analytics footprint and total cost of ownership."
Learn more: http://www.cray.com/products/analytics/urika-xa
Watch the video presentation: http://wp.me/p3RLEV-3yR
Sign up for our insideBIGDATA Newsletter: http://insidebigdata.com/newsletter
TOPIC EXTRACTION OF CRAWLED DOCUMENTS COLLECTION USING CORRELATED TOPIC MODEL...ijnlc
The tremendous increase in the amount of available research documents impels researchers to propose topic models to extract the latent semantic themes of a documents collection. However, how to extract the hidden topics of the documents collection has become a crucial task for many topic model applications. Moreover, conventional topic modeling approaches suffer from the scalability problem when the size of documents collection increases. In this paper, the Correlated Topic Model with variational ExpectationMaximization algorithm is implemented in MapReduce framework to solve the scalability problem. The proposed approach utilizes the dataset crawled from the public digital library. In addition, the full-texts of the crawled documents are analysed to enhance the accuracy of MapReduce CTM. The experiments are conducted to demonstrate the performance of the proposed algorithm. From the evaluation, the proposed approach has a comparable performance in terms of topic coherences with LDA implemented in MapReduce framework.
A Combined Approach for Feature Subset Selection and Size Reduction for High ...IJERA Editor
selection of relevant feature from a given set of feature is one of the important issues in the field of
data mining as well as classification. In general the dataset may contain a number of features however it is not
necessary that the whole set features are important for particular analysis of decision making because the
features may share the common information‟s and can also be completely irrelevant to the undergoing
processing. This generally happen because of improper selection of features during the dataset formation or
because of improper information availability about the observed system. However in both cases the data will
contain the features that will just increase the processing burden which may ultimately cause the improper
outcome when used for analysis. Because of these reasons some kind of methods are required to detect and
remove these features hence in this paper we are presenting an efficient approach for not just removing the
unimportant features but also the size of complete dataset size. The proposed algorithm utilizes the information
theory to detect the information gain from each feature and minimum span tree to group the similar features
with that the fuzzy c-means clustering is used to remove the similar entries from the dataset. Finally the
algorithm is tested with SVM classifier using 35 publicly available real-world high-dimensional dataset and the
results shows that the presented algorithm not only reduces the feature set and data lengths but also improves the
performances of the classifier.
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERijnlc
This document presents an adaptive log file parser that uses semantics and hidden Markov models. It first clusters log file lines based on semantics to limit unstructured text. It then builds a hidden Markov model to represent parsing patterns, with log entries as states and extracted values as emissions. When applied to a new system, it adapts the model's transition and emission probabilities to fit the new data. The approach achieves over 99.99% accuracy when trained on one system and applied to another with slightly different log patterns.
Textual Data Partitioning with Relationship and Discriminative AnalysisEditor IJMTER
Data partitioning methods are used to partition the data values with similarity. Similarity
measures are used to estimate transaction relationships. Hierarchical clustering model produces tree
structured results. Partitioned clustering produces results in grid format. Text documents are
unstructured data values with high dimensional attributes. Document clustering group ups unlabeled text
documents into meaningful clusters. Traditional clustering methods require cluster count (K) for the
document grouping process. Clustering accuracy degrades drastically with reference to the unsuitable
cluster count.
Textual data elements are divided into two types’ discriminative words and nondiscriminative
words. Only discriminative words are useful for grouping documents. The involvement of
nondiscriminative words confuses the clustering process and leads to poor clustering solution in return.
A variation inference algorithm is used to infer the document collection structure and partition of
document words at the same time. Dirichlet Process Mixture (DPM) model is used to partition
documents. DPM clustering model uses both the data likelihood and the clustering property of the
Dirichlet Process (DP). Dirichlet Process Mixture Model for Feature Partition (DPMFP) is used to
discover the latent cluster structure based on the DPM model. DPMFP clustering is performed without
requiring the number of clusters as input.
Document labels are used to estimate the discriminative word identification process. Concept
relationships are analyzed with Ontology support. Semantic weight model is used for the document
similarity analysis. The system improves the scalability with the support of labels and concept relations
for dimensionality reduction process.
Extended pso algorithm for improvement problems k means clustering algorithmIJMIT JOURNAL
The clustering is a without monitoring process and one of the most common data mining techniques. The
purpose of clustering is grouping similar data together in a group, so were most similar to each other in a
cluster and the difference with most other instances in the cluster are. In this paper we focus on clustering
partition k-means, due to ease of implementation and high-speed performance of large data sets, After 30
year it is still very popular among the developed clustering algorithm and then for improvement problem of
placing of k-means algorithm in local optimal, we pose extended PSO algorithm, that its name is ECPSO.
Our new algorithm is able to be cause of exit from local optimal and with high percent produce the
problem’s optimal answer. The probe of results show that mooted algorithm have better performance
regards as other clustering algorithms specially in two index, the carefulness of clustering and the quality
of clustering.
A Novel Approach to Mathematical Concepts in Data Miningijdmtaiir
-This paper describes three different fundamental
mathematical programming approaches that are relevant to
data mining. They are: Feature Selection, Clustering and
Robust Representation. This paper comprises of two clustering
algorithms such as K-mean algorithm and K-median
algorithms. Clustering is illustrated by the unsupervised
learning of patterns and clusters that may exist in a given
databases and useful tool for Knowledge Discovery in
Database (KDD). The results of k-median algorithm are used
to collecting the blood cancer patient from a medical database.
K-mean clustering is a data mining/machine learning algorithm
used to cluster observations into groups of related observations
without any prior knowledge of those relationships. The kmean algorithm is one of the simplest clustering techniques
and it is commonly used in medical imaging, biometrics and
related fields.
An Efficient Clustering Method for Aggregation on Data FragmentsIJMER
Clustering is an important step in the process of data analysis with applications to numerous fields. Clustering ensembles, has emerged as a powerful technique for combining different clustering results to obtain a quality cluster. Existing clustering aggregation algorithms are applied directly to large number of data points. The algorithms are inefficient if the number of data points is large. This project defines an efficient approach for clustering aggregation based on data fragments. In fragment-based approach, a data fragment is any subset of the data. To increase the efficiency of the proposed approach, the clustering aggregation can be performed directly on data fragments under comparison measure and normalized mutual information measures for clustering aggregation, enhanced clustering aggregation algorithms are described. To show the minimal computational complexity. (Agglomerative, Furthest, and Local Search); nevertheless, which increases the accuracy.
In this video from the 2015 HPC User Forum in Broomfield, Barry Bolding from Cray presents: HPC + D + A = HPDA?
"The flexible, multi-use Cray Urika-XA extreme analytics platform addresses perhaps the most critical obstacle in data analytics today — limitation. Analytics problems are getting more varied and complex but the available solution technologies have significant constraints. Traditional analytics appliances lock you into a single approach and building a custom solution in-house is so difficult and time consuming that the business value derived from analytics fails to materialize. In contrast, the Urika-XA platform is open, high performing and cost effective, serving a wide range of analytics tools with varying computing demands in a single environment. Pre-integrated with the Hadoop and Spark frameworks, the Urika-XA system combines the benefits of a turnkey analytics appliance with a flexible, open platform that you can modify for future analytics workloads. This single-platform consolidation of workloads reduces your analytics footprint and total cost of ownership."
Learn more: http://www.cray.com/products/analytics/urika-xa
Watch the video presentation: http://wp.me/p3RLEV-3yR
Sign up for our insideBIGDATA Newsletter: http://insidebigdata.com/newsletter
TOPIC EXTRACTION OF CRAWLED DOCUMENTS COLLECTION USING CORRELATED TOPIC MODEL...ijnlc
The tremendous increase in the amount of available research documents impels researchers to propose topic models to extract the latent semantic themes of a documents collection. However, how to extract the hidden topics of the documents collection has become a crucial task for many topic model applications. Moreover, conventional topic modeling approaches suffer from the scalability problem when the size of documents collection increases. In this paper, the Correlated Topic Model with variational ExpectationMaximization algorithm is implemented in MapReduce framework to solve the scalability problem. The proposed approach utilizes the dataset crawled from the public digital library. In addition, the full-texts of the crawled documents are analysed to enhance the accuracy of MapReduce CTM. The experiments are conducted to demonstrate the performance of the proposed algorithm. From the evaluation, the proposed approach has a comparable performance in terms of topic coherences with LDA implemented in MapReduce framework.
A Combined Approach for Feature Subset Selection and Size Reduction for High ...IJERA Editor
selection of relevant feature from a given set of feature is one of the important issues in the field of
data mining as well as classification. In general the dataset may contain a number of features however it is not
necessary that the whole set features are important for particular analysis of decision making because the
features may share the common information‟s and can also be completely irrelevant to the undergoing
processing. This generally happen because of improper selection of features during the dataset formation or
because of improper information availability about the observed system. However in both cases the data will
contain the features that will just increase the processing burden which may ultimately cause the improper
outcome when used for analysis. Because of these reasons some kind of methods are required to detect and
remove these features hence in this paper we are presenting an efficient approach for not just removing the
unimportant features but also the size of complete dataset size. The proposed algorithm utilizes the information
theory to detect the information gain from each feature and minimum span tree to group the similar features
with that the fuzzy c-means clustering is used to remove the similar entries from the dataset. Finally the
algorithm is tested with SVM classifier using 35 publicly available real-world high-dimensional dataset and the
results shows that the presented algorithm not only reduces the feature set and data lengths but also improves the
performances of the classifier.
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERijnlc
This document presents an adaptive log file parser that uses semantics and hidden Markov models. It first clusters log file lines based on semantics to limit unstructured text. It then builds a hidden Markov model to represent parsing patterns, with log entries as states and extracted values as emissions. When applied to a new system, it adapts the model's transition and emission probabilities to fit the new data. The approach achieves over 99.99% accuracy when trained on one system and applied to another with slightly different log patterns.
Textual Data Partitioning with Relationship and Discriminative AnalysisEditor IJMTER
Data partitioning methods are used to partition the data values with similarity. Similarity
measures are used to estimate transaction relationships. Hierarchical clustering model produces tree
structured results. Partitioned clustering produces results in grid format. Text documents are
unstructured data values with high dimensional attributes. Document clustering group ups unlabeled text
documents into meaningful clusters. Traditional clustering methods require cluster count (K) for the
document grouping process. Clustering accuracy degrades drastically with reference to the unsuitable
cluster count.
Textual data elements are divided into two types’ discriminative words and nondiscriminative
words. Only discriminative words are useful for grouping documents. The involvement of
nondiscriminative words confuses the clustering process and leads to poor clustering solution in return.
A variation inference algorithm is used to infer the document collection structure and partition of
document words at the same time. Dirichlet Process Mixture (DPM) model is used to partition
documents. DPM clustering model uses both the data likelihood and the clustering property of the
Dirichlet Process (DP). Dirichlet Process Mixture Model for Feature Partition (DPMFP) is used to
discover the latent cluster structure based on the DPM model. DPMFP clustering is performed without
requiring the number of clusters as input.
Document labels are used to estimate the discriminative word identification process. Concept
relationships are analyzed with Ontology support. Semantic weight model is used for the document
similarity analysis. The system improves the scalability with the support of labels and concept relations
for dimensionality reduction process.
Extended pso algorithm for improvement problems k means clustering algorithmIJMIT JOURNAL
The clustering is a without monitoring process and one of the most common data mining techniques. The
purpose of clustering is grouping similar data together in a group, so were most similar to each other in a
cluster and the difference with most other instances in the cluster are. In this paper we focus on clustering
partition k-means, due to ease of implementation and high-speed performance of large data sets, After 30
year it is still very popular among the developed clustering algorithm and then for improvement problem of
placing of k-means algorithm in local optimal, we pose extended PSO algorithm, that its name is ECPSO.
Our new algorithm is able to be cause of exit from local optimal and with high percent produce the
problem’s optimal answer. The probe of results show that mooted algorithm have better performance
regards as other clustering algorithms specially in two index, the carefulness of clustering and the quality
of clustering.
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...RSIS International
The biggest challenge in today’s world is a searching of
images in a large database. For searching of an image a
technique which can be termed as Hashing is used. Already there
are many hashing techniques are present for retrieval of an
image in a large databank. The hashing technique can be done
on images by considering the high dimensional descriptor of an
image but in the existing hashing techniques single dimensional
descriptor is used from this the performance of the probability of
distribution of an image search will not achieve as expected. And
the drawback is that giving the query input in a texture format
leads to the limitations of image search that is firstly due to the
limited keyword and the second is Annotation approach by
human is ambiguous and incomplete.
To overcome from these drawbacks a new technique has been
proposed named as Multiview Alignment Hashing technique in
which it keeps the high dimensional feature descriptor data as
well probability of distribution of an images in a database. Along
with the Multiview feature descriptor another technique can be
used that is Nonnegetive Matrix Factorization (NMF). NMF is a
popular technique used in data mining in which clustering of a
data will be takes place by considering only the non- negative
matrix value.
This document summarizes research on improving image classification results using neural networks. It compares common image classification methods like support vector machines (SVM) and K-nearest neighbors (KNN). It then evaluates the performance of multilayer perceptron (MLP) neural networks and radial basis function (RBF) neural networks on image classification. The document tests various configurations of MLP and RBF networks on a dataset containing 2310 images across 7 classes. It finds that a MLP network with two hidden layers of 10 neurons each achieves the best results, with an average accuracy of 98.84%. This is significantly higher than the 84.47% average accuracy of RBF networks and outperforms KNN classification as well. The research concludes that neural
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...ijiert bestjournal
In Natural Scene Image,Text detection is important tasks which are used for many content based image analysis. A maximally stable external region based method is us ed for scene detection .This MSER based method incl udes stages character candidate extraction,text candida te construction,text candidate elimination & text candidate classification. Main limitations of this method are how to detect highly blurred text in low resolutio n natural scene images. The current technology not focuses on any t ext extraction method. In proposed system a Conditi onal Random field (CRF) model is used to assign candidat e component as one of the two classes (text& Non Te xt) by Considering both unary component properties and bin ary contextual component relationship. For this pur pose we are using connected component analysis method. The proposed system also performs a text extraction usi ng OCR
Multiview Alignment Hashing for Efficient Image Search1crore projects
The document describes a new method called Multiview Alignment Hashing (MAH) for efficient image search. MAH aims to fuse multiple image feature representations while preserving the high-dimensional joint probability distribution of the data and obtaining orthogonal bases. It does this by formulating an objective function that is optimized using an alternate optimization procedure. This finds low-dimensional matrix factorizations via a technique called Regularized Kernel Nonnegative Matrix Factorization. After optimization, binary hash codes are obtained for images that can be used for efficient similarity search. The method is evaluated on several image datasets and is shown to outperform other state-of-the-art multiview hashing techniques.
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Informationcsandit
The document describes a novel word sense disambiguation method using global co-occurrence information and non-negative matrix factorization. It proposes using global co-occurrence frequencies between words in a large corpus rather than local context windows to address data sparseness issues. An experiment compares the method to two baselines and shows it achieves the highest and most stable precision for word sense disambiguation.
Clustering of high dimensionality data which can be seen in almost all fields these days is becoming
very tedious process. The key disadvantage of high dimensional data which we can pen down is curse
of dimensionality. As the magnitude of datasets grows the data points become sparse and density of
area becomes less making it difficult to cluster that data which further reduces the performance of
traditional algorithms used for clustering. Semi-supervised clustering algorithms aim to improve
clustering results using limited supervision. The supervision is generally given as pair wise
constraints; such constraints are natural for graphs, yet most semi-supervised clustering algorithms are
designed for data represented as vectors [2]. In this paper, we unify vector-based and graph-based
approaches. We first show that a recently-proposed objective function for semi-supervised clustering
based on Hidden Markov Random Fields, with squared Euclidean distance and a certain class of
constraint penalty functions, can be expressed as a special case of the global kernel k-means objective
[3]. A recent theoretical connection between global kernel k-means and several graph clustering
objectives enables us to perform semi-supervised clustering of data. In particular, some methods have
been proposed for semi supervised clustering based on pair wise similarity or dissimilarity
information. In this paper, we propose a kernel approach for semi supervised clustering and present in
detail two special cases of this kernel approach.
A survey on methods and applications of meta-learning with GNNsShreya Goyal
This survey paper has provided a comprehensive review of works that are a combination of graph neural networks (GNNs) and meta-learning. They have also provided a thorough review, summary of methods, and applications in these categories. The application of meta-learning to GNNs is a growing and exciting field; many graph problems will benefit immensely from the combination of the two approaches.
This document describes a novel graph embedding procedure based on simplicial complexes for graph classification tasks. Simplicial complexes are mathematical objects that can capture multi-way relationships in data beyond pairwise relationships. The proposed approach uses simplicial complexes to extract meaningful substructures from graphs, clusters these substructures to form an alphabet, and then embeds each graph as a symbolic histogram over the alphabet. This moves the problem into a metric space where standard machine learning algorithms can be applied. The approach is tested on 30 graph classification benchmarks and two protein analysis applications to demonstrate its effectiveness.
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...IRJET Journal
This document proposes a new oversampling technique called Centroid Oversampling to address imbalanced class distributions in customer churn prediction problems. It summarizes existing oversampling methods like SMOTE and introduces Centroid Oversampling, which generates synthetic samples by calculating the centroid of the three nearest data points rather than oversampling outliers. Experimental results on three telecom datasets show Centroid Oversampling achieves better accuracy, recall, and F-measure than SMOTE when used with a KNN classifier, particularly on datasets with high imbalance.
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...CSCJournals
Biclusters are required to analyzing gene expression patterns of genes comparing rows in expression profiles and analyzing expression profiles of samples by comparing columns in gene expression matrix. In the process of biclustering we need to cluster genes and samples. The algorithm presented in this paper is based upon the two-way clustering approach in which the genes and samples are clustered using parallel fuzzy C-means clustering using message passing interface, we call it MFCM. MFCM applied for clustering on genes and samples which maximize membership function values of the data set. It is a parallelized rework of a parallel fuzzy two-way clustering algorithm for microarray gene expression data [9], to study the efficiency and parallelization improvement of the algorithm. The algorithm uses gene entropy measure to filter the clustered data to find biclusters. The method is able to get highly correlated biclusters of the gene expression dataset.
Presentation summarizes main content of Farrelly, C. M. (2017). Extensions of Morse-Smale Regression with Application to Actuarial Science. arXiv preprint arXiv:1708.05712.
Paper was accepted December 2017 by Casualty Actuarial Society.
1) The document discusses mining data streams using an improved version of McDiarmid's bound. It aims to enhance the bounds obtained by McDiarmid's tree algorithm and improve processing efficiency.
2) Traditional data mining techniques cannot be directly applied to data streams due to their continuous, rapid arrival. The document proposes using Gaussian approximations to McDiarmid's bounds to reduce the size of training samples needed for split criteria selection.
3) It describes Hoeffding's inequality, which is commonly used but not sufficient for data streams. The document argues that McDiarmid's inequality, used appropriately, provides a more efficient technique for high-speed, time-changing data streams.
Kernel based similarity estimation and real time tracking of movingIAEME Publication
This document discusses kernel-based mean shift algorithm for real-time object tracking. It presents the following:
1) The algorithm uses kernel density estimation to calculate the similarity between a target model and candidate windows, using the Bhattacharyya coefficient. 2) It can successfully track objects moving uniformly at slow speeds but struggles with fast or non-uniform motion, or changes in scale. 3) The algorithm was tested on video streams and could track objects moving slowly but failed for fast or irregular motion. Adaptive target windows are needed to handle changes in scale.
The document provides a literature review of different clustering techniques. It begins by defining clustering and its applications. It then categorizes and describes several clustering methods including hierarchical (BIRCH, CURE, CHAMELEON), partitioning (k-means, k-medoids), density-based (DBSCAN, OPTICS, DENCLUE), grid-based (CLIQUE, STING, MAFIA), and model-based (RBMN, SOM) methods. For each method, it discusses the algorithm, advantages, disadvantages and time complexity. The document aims to provide an overview of various clustering techniques for classification and comparison.
Improved wolf algorithm on document images detection using optimum mean techn...journalBEEI
Detection text from handwriting in historical documents provides high-level features for the challenging problem of handwriting recognition. Such handwriting often contains noise, faint or incomplete strokes, strokes with gaps, and competing lines when embedded in a table or form, making it unsuitable for local line following algorithms or associated binarization schemes. In this paper, a proposed method based on the optimum threshold value and namely as the Optimum Mean method was presented. Besides, Wolf method unsuccessful in order to detect the thin text in the non-uniform input image. However, the proposed method was suggested to overcome the Wolf method problem by suggesting a maximum threshold value using optimum mean. Based on the calculation, the proposed method obtained a higher F-measure (74.53), PSNR (14.77) and lowest NRM (0.11) compared to the Wolf method. In conclusion, the proposed method successful and effective to solve the wolf problem by producing a high-quality output image.
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...IJAEMSJORNAL
The current trend in the industry is to analyze large data sets and apply data mining, machine learning techniques to identify a pattern. But the challenges with huge data sets are the high dimensions associated with it. Sometimes in data analytics applications, large amounts of data produce worse performance. Also, most of the data mining algorithms are implemented column wise and too many columns restrict the performance and make it slower. Therefore, dimensionality reduction is an important step in data analysis. Dimensionality reduction is a technique that converts high dimensional data into much lower dimension, such that maximum variance is explained within the first few dimensions. This paper focuses on multivariate statistical and artificial neural networks techniques for data reduction. Each method has a different rationale to preserve the relationship between input parameters during analysis. Principal Component Analysis which is a multivariate technique and Self Organising Map a neural network technique is presented in this paper. Also, a hierarchical clustering approach has been applied to the reduced data set. A case study of Air quality measurement has been considered to evaluate the performance of the proposed techniques.
1. The document discusses using statistical learning methods like Gaussian mixture models (GMM) and dynamic component allocation (DCA) for hyperspectral image classification.
2. GMM represents pixel spectra as mixtures of Gaussian distributions. DCA is an algorithm that dynamically adds and removes Gaussian components to better characterize the data during training.
3. The document outlines how DCA works, including merging similar Gaussian modes, splitting modes with high kurtosis, and pruning insignificant modes. These techniques aim to learn the appropriate number of mixture components from the data.
Efficient Image Retrieval by Multi-view Alignment Technique with Non Negative...RSIS International
The biggest challenge in today’s world is a searching of
images in a large database. For searching of an image a
technique which can be termed as Hashing is used. Already there
are many hashing techniques are present for retrieval of an
image in a large databank. The hashing technique can be done
on images by considering the high dimensional descriptor of an
image but in the existing hashing techniques single dimensional
descriptor is used from this the performance of the probability of
distribution of an image search will not achieve as expected. And
the drawback is that giving the query input in a texture format
leads to the limitations of image search that is firstly due to the
limited keyword and the second is Annotation approach by
human is ambiguous and incomplete.
To overcome from these drawbacks a new technique has been
proposed named as Multiview Alignment Hashing technique in
which it keeps the high dimensional feature descriptor data as
well probability of distribution of an images in a database. Along
with the Multiview feature descriptor another technique can be
used that is Nonnegetive Matrix Factorization (NMF). NMF is a
popular technique used in data mining in which clustering of a
data will be takes place by considering only the non- negative
matrix value.
This document summarizes research on improving image classification results using neural networks. It compares common image classification methods like support vector machines (SVM) and K-nearest neighbors (KNN). It then evaluates the performance of multilayer perceptron (MLP) neural networks and radial basis function (RBF) neural networks on image classification. The document tests various configurations of MLP and RBF networks on a dataset containing 2310 images across 7 classes. It finds that a MLP network with two hidden layers of 10 neurons each achieves the best results, with an average accuracy of 98.84%. This is significantly higher than the 84.47% average accuracy of RBF networks and outperforms KNN classification as well. The research concludes that neural
ROBUST TEXT DETECTION AND EXTRACTION IN NATURAL SCENE IMAGES USING CONDITIONA...ijiert bestjournal
In Natural Scene Image,Text detection is important tasks which are used for many content based image analysis. A maximally stable external region based method is us ed for scene detection .This MSER based method incl udes stages character candidate extraction,text candida te construction,text candidate elimination & text candidate classification. Main limitations of this method are how to detect highly blurred text in low resolutio n natural scene images. The current technology not focuses on any t ext extraction method. In proposed system a Conditi onal Random field (CRF) model is used to assign candidat e component as one of the two classes (text& Non Te xt) by Considering both unary component properties and bin ary contextual component relationship. For this pur pose we are using connected component analysis method. The proposed system also performs a text extraction usi ng OCR
Multiview Alignment Hashing for Efficient Image Search1crore projects
The document describes a new method called Multiview Alignment Hashing (MAH) for efficient image search. MAH aims to fuse multiple image feature representations while preserving the high-dimensional joint probability distribution of the data and obtaining orthogonal bases. It does this by formulating an objective function that is optimized using an alternate optimization procedure. This finds low-dimensional matrix factorizations via a technique called Regularized Kernel Nonnegative Matrix Factorization. After optimization, binary hash codes are obtained for images that can be used for efficient similarity search. The method is evaluated on several image datasets and is shown to outperform other state-of-the-art multiview hashing techniques.
Latent Semantic Word Sense Disambiguation Using Global Co-Occurrence Informationcsandit
The document describes a novel word sense disambiguation method using global co-occurrence information and non-negative matrix factorization. It proposes using global co-occurrence frequencies between words in a large corpus rather than local context windows to address data sparseness issues. An experiment compares the method to two baselines and shows it achieves the highest and most stable precision for word sense disambiguation.
Clustering of high dimensionality data which can be seen in almost all fields these days is becoming
very tedious process. The key disadvantage of high dimensional data which we can pen down is curse
of dimensionality. As the magnitude of datasets grows the data points become sparse and density of
area becomes less making it difficult to cluster that data which further reduces the performance of
traditional algorithms used for clustering. Semi-supervised clustering algorithms aim to improve
clustering results using limited supervision. The supervision is generally given as pair wise
constraints; such constraints are natural for graphs, yet most semi-supervised clustering algorithms are
designed for data represented as vectors [2]. In this paper, we unify vector-based and graph-based
approaches. We first show that a recently-proposed objective function for semi-supervised clustering
based on Hidden Markov Random Fields, with squared Euclidean distance and a certain class of
constraint penalty functions, can be expressed as a special case of the global kernel k-means objective
[3]. A recent theoretical connection between global kernel k-means and several graph clustering
objectives enables us to perform semi-supervised clustering of data. In particular, some methods have
been proposed for semi supervised clustering based on pair wise similarity or dissimilarity
information. In this paper, we propose a kernel approach for semi supervised clustering and present in
detail two special cases of this kernel approach.
A survey on methods and applications of meta-learning with GNNsShreya Goyal
This survey paper has provided a comprehensive review of works that are a combination of graph neural networks (GNNs) and meta-learning. They have also provided a thorough review, summary of methods, and applications in these categories. The application of meta-learning to GNNs is a growing and exciting field; many graph problems will benefit immensely from the combination of the two approaches.
This document describes a novel graph embedding procedure based on simplicial complexes for graph classification tasks. Simplicial complexes are mathematical objects that can capture multi-way relationships in data beyond pairwise relationships. The proposed approach uses simplicial complexes to extract meaningful substructures from graphs, clusters these substructures to form an alphabet, and then embeds each graph as a symbolic histogram over the alphabet. This moves the problem into a metric space where standard machine learning algorithms can be applied. The approach is tested on 30 graph classification benchmarks and two protein analysis applications to demonstrate its effectiveness.
IRJET- Predicting Customers Churn in Telecom Industry using Centroid Oversamp...IRJET Journal
This document proposes a new oversampling technique called Centroid Oversampling to address imbalanced class distributions in customer churn prediction problems. It summarizes existing oversampling methods like SMOTE and introduces Centroid Oversampling, which generates synthetic samples by calculating the centroid of the three nearest data points rather than oversampling outliers. Experimental results on three telecom datasets show Centroid Oversampling achieves better accuracy, recall, and F-measure than SMOTE when used with a KNN classifier, particularly on datasets with high imbalance.
Biclustering using Parallel Fuzzy Approach for Analysis of Microarray Gene Ex...CSCJournals
Biclusters are required to analyzing gene expression patterns of genes comparing rows in expression profiles and analyzing expression profiles of samples by comparing columns in gene expression matrix. In the process of biclustering we need to cluster genes and samples. The algorithm presented in this paper is based upon the two-way clustering approach in which the genes and samples are clustered using parallel fuzzy C-means clustering using message passing interface, we call it MFCM. MFCM applied for clustering on genes and samples which maximize membership function values of the data set. It is a parallelized rework of a parallel fuzzy two-way clustering algorithm for microarray gene expression data [9], to study the efficiency and parallelization improvement of the algorithm. The algorithm uses gene entropy measure to filter the clustered data to find biclusters. The method is able to get highly correlated biclusters of the gene expression dataset.
Presentation summarizes main content of Farrelly, C. M. (2017). Extensions of Morse-Smale Regression with Application to Actuarial Science. arXiv preprint arXiv:1708.05712.
Paper was accepted December 2017 by Casualty Actuarial Society.
1) The document discusses mining data streams using an improved version of McDiarmid's bound. It aims to enhance the bounds obtained by McDiarmid's tree algorithm and improve processing efficiency.
2) Traditional data mining techniques cannot be directly applied to data streams due to their continuous, rapid arrival. The document proposes using Gaussian approximations to McDiarmid's bounds to reduce the size of training samples needed for split criteria selection.
3) It describes Hoeffding's inequality, which is commonly used but not sufficient for data streams. The document argues that McDiarmid's inequality, used appropriately, provides a more efficient technique for high-speed, time-changing data streams.
Kernel based similarity estimation and real time tracking of movingIAEME Publication
This document discusses kernel-based mean shift algorithm for real-time object tracking. It presents the following:
1) The algorithm uses kernel density estimation to calculate the similarity between a target model and candidate windows, using the Bhattacharyya coefficient. 2) It can successfully track objects moving uniformly at slow speeds but struggles with fast or non-uniform motion, or changes in scale. 3) The algorithm was tested on video streams and could track objects moving slowly but failed for fast or irregular motion. Adaptive target windows are needed to handle changes in scale.
The document provides a literature review of different clustering techniques. It begins by defining clustering and its applications. It then categorizes and describes several clustering methods including hierarchical (BIRCH, CURE, CHAMELEON), partitioning (k-means, k-medoids), density-based (DBSCAN, OPTICS, DENCLUE), grid-based (CLIQUE, STING, MAFIA), and model-based (RBMN, SOM) methods. For each method, it discusses the algorithm, advantages, disadvantages and time complexity. The document aims to provide an overview of various clustering techniques for classification and comparison.
Improved wolf algorithm on document images detection using optimum mean techn...journalBEEI
Detection text from handwriting in historical documents provides high-level features for the challenging problem of handwriting recognition. Such handwriting often contains noise, faint or incomplete strokes, strokes with gaps, and competing lines when embedded in a table or form, making it unsuitable for local line following algorithms or associated binarization schemes. In this paper, a proposed method based on the optimum threshold value and namely as the Optimum Mean method was presented. Besides, Wolf method unsuccessful in order to detect the thin text in the non-uniform input image. However, the proposed method was suggested to overcome the Wolf method problem by suggesting a maximum threshold value using optimum mean. Based on the calculation, the proposed method obtained a higher F-measure (74.53), PSNR (14.77) and lowest NRM (0.11) compared to the Wolf method. In conclusion, the proposed method successful and effective to solve the wolf problem by producing a high-quality output image.
Influence over the Dimensionality Reduction and Clustering for Air Quality Me...IJAEMSJORNAL
The current trend in the industry is to analyze large data sets and apply data mining, machine learning techniques to identify a pattern. But the challenges with huge data sets are the high dimensions associated with it. Sometimes in data analytics applications, large amounts of data produce worse performance. Also, most of the data mining algorithms are implemented column wise and too many columns restrict the performance and make it slower. Therefore, dimensionality reduction is an important step in data analysis. Dimensionality reduction is a technique that converts high dimensional data into much lower dimension, such that maximum variance is explained within the first few dimensions. This paper focuses on multivariate statistical and artificial neural networks techniques for data reduction. Each method has a different rationale to preserve the relationship between input parameters during analysis. Principal Component Analysis which is a multivariate technique and Self Organising Map a neural network technique is presented in this paper. Also, a hierarchical clustering approach has been applied to the reduced data set. A case study of Air quality measurement has been considered to evaluate the performance of the proposed techniques.
1. The document discusses using statistical learning methods like Gaussian mixture models (GMM) and dynamic component allocation (DCA) for hyperspectral image classification.
2. GMM represents pixel spectra as mixtures of Gaussian distributions. DCA is an algorithm that dynamically adds and removes Gaussian components to better characterize the data during training.
3. The document outlines how DCA works, including merging similar Gaussian modes, splitting modes with high kurtosis, and pruning insignificant modes. These techniques aim to learn the appropriate number of mixture components from the data.
During the past decade, the size of 3D seismic data volumes and the number of seismic attributes have increased
to the extent that it is difficult, if not impossible, for interpreters to examine every seismic line and time
slice. To address this problem, several seismic facies classification algorithms including k-means, self-organizing
maps, generative topographic mapping, support vector machines, Gaussian mixture models, and artificial neural
networks have been successfully used to extract features of geologic interest from multiple volumes. Although
well documented in the literature, the terminology and complexity of these algorithms may bewilder the average
seismic interpreter, and few papers have applied these competing methods to the same data volume. We have
reviewed six commonly used algorithms and applied them to a single 3D seismic data volume acquired over the
Canterbury Basin, offshore New Zealand, where one of the main objectives was to differentiate the architectural
elements of a turbidite system. Not surprisingly, the most important parameter in this analysis was the choice of
the correct input attributes, which in turn depended on careful pattern recognition by the interpreter. We found
that supervised learning methods provided accurate estimates of the desired seismic facies, whereas unsupervised
learning methods also highlighted features that might otherwise be overlooked.
Receiver deghosting method to mitigate F-K transform artifacts: A non-windo...Pioneer Natural Resources
In this study, we implemented and tested a new processing- based broadband solution for mitigating F-K transform arti- facts for receiver deghosting in a marine environment. The F- K transform has traditionally been used for flat cable (constant depth) deghosting and often times tailored to meet the slanted (variable depth) cable criteria. Recently, the usage of τ − p do- main deterministic deghost operator has been more prominent with slant cable deghosting. Irrespective of the type of trans- form or deghost operator used, a windowed process is essential due to the time and offset varying character of the ghost. This use of a windowed process usually results in poor reconstruc- tion of deghosted signals and artifacts beyond the control of the transform(s) itself. The windowing in time and offset produces edgy effects which can be clearly seen in the difference plots. Our method, using a non-windowing approach, demonstrates a better representation of the deghosted signals without the arti- facts caused by the boundary of the windows. This method has also been well-tested for both the flat and slant cable receiver deghosting workflows in synthetic and field data examples.
OPTIMIZED RATE ALLOCATION OF HYPERSPECTRAL IMAGES IN COMPRESSED DOMAIN USING ...Pioneer Natural Resources
This document discusses optimizing the rate allocation of hyperspectral images compressed using JPEG2000. It presents a mixed model for bit allocation that combines high and low bit rate models. This mixed model and an optimal rate allocation approach based on minimizing mean squared error under a rate constraint provide lower reconstruction errors than traditional approaches. Computational tests on hyperspectral data show the discrete wavelet transform allows for faster processing and less memory usage compared to the Karhunen-Loeve transform.
Directional Analysis and Filtering for Dust Storm detection in NOAA-AVHRR Ima...Pioneer Natural Resources
This document proposes techniques for detecting dust storms and determining their direction of transport using NOAA-AVHRR satellite imagery. It introduces a two-part approach using image processing algorithms: 1) A visualization technique uses filters, edge detectors and classification to locate dust sources. 2) An automation technique performs power spectrum analysis to detect dust storm direction by analyzing texture orientation in image blocks. The goal is to automatically detect and track dust storms for applications like hazard monitoring.
New exploration challenges and current research demands
3D gravity modeling with 3D geology interpretations. In
the near future, multi-parameter and multi-dimensional
interpretations represent the observed and expected in situ
geology, geophysical, and petro-physical data that will be
used for join multi-parameter, multi-dimensional
inversions. We present an initial 3D gravity model of
Osage County in northeastern Oklahoma, where there is a
greater than 40 mGal, 100 km diameter semi-circular
gravity anomaly that cannot be effectively removed by
traditional gravity processing techniques.
Distance Metric Based Multi-Attribute Seismic Facies Classification to Identi...Pioneer Natural Resources
Conventional reservoirs benefit from a long scientific history that correlates successful plays to seismic measurements through depositional, tectonic, and digenetic models. Unconventional reservoirs are less well understood, however benefit from significantly denser well control. Thus, allowing us to establish statistical rather than model-based correlations between seismic data, geology, and successful completion strategies. One of the more commonly encountered correlation techniques is based on computer assisted pattern recognition. The pattern recognition techniques have found their niche in a plethora of applications ranging from flagging suspicious credit card purchase patterns to rewarding repeating online buying patterns. Classification of a given seismic response as having a “good” or “bad” pattern requires a “distance metric”. Distance metric “learning” uses past experiences (well performance) as training data to develop a distance metric. Alternative distance metrics have demonstrated significant value in the identification and classification of repeated or anomalous behaviors in public health, security, and marketing. In this paper we examine the value of three of these alternative distance metrics of 3D seismic attributes to the identification of sweet spots in a Barnett Shale play.
We illustrate unsupervised and supervised learning algorithms that accurately classify the lithological variations in the 3D seismic data. We demonstrate blind source separation techniques such as the principal components (PCA) and noise adjusted principal
components in conjunction with Kohonen Self organizing maps to produce superior unsupervised classification maps.
Further, we utilize the PCA space training in Maximum likelihood (ML) supervised classification. Results demonstrate that the ML supervised classification produces an improved classification of the facies in the 3D seismic dataset from the Anadarko basin in central Oklahoma.
The document proposes an anomaly detection scheme for hyperspectral images based on a non-Gaussian mixture model using a Student's t-distribution. It estimates the background probability density function using a Bayesian approach that models each pixel as a mixture of Student's t distributions. The anomaly detection strategy then applies a generalized likelihood ratio test. Experimental results on real hyperspectral data show the proposed Bayesian Student's t mixture model can reliably estimate the background distribution and effectively detect anomalous objects, outperforming a Gaussian mixture model approach.
This document presents a novel fuzzy k-nearest neighbor equality (FK-NNE) algorithm for classifying masses in mammograms as benign or malignant. The algorithm assigns membership values to different classes based on distances to k-nearest neighbors. It achieved 94.46% sensitivity, 96.81% specificity, and 96.52% accuracy, outperforming k-nearest neighbors, fuzzy k-nearest neighbors, and k-nearest neighbor equality algorithms. The algorithm considers relative importance of neighbors and assigns partial membership to classes, addressing issues with insufficient knowledge faced by other techniques. Experimental results demonstrated FK-NNE had the best performance with an area under the ROC curve of 0.9734, indicating high diagnostic accuracy.
IRJET-Multimodal Image Classification through Band and K-Means ClusteringIRJET Journal
This document proposes a bilayer graph-based learning framework for multimodal image classification using limited labeled pixels. It constructs a simple graph in the first layer where each vertex is a pixel and edge weights encode pixel similarity. Unsupervised learning estimates grouping relations among pixels. These relations form a hypergraph in the second layer, on which semisupervised learning classifies pixels to address challenges of complex relationships and limited labels in multimodal images. The framework effectively exploits the underlying data structure.
WITH SEMANTICS AND HIDDEN MARKOV MODELS TO AN ADAPTIVE LOG FILE PARSERkevig
We aim to model an adaptive log file parser. As the content of log files often evolves over time, we
established a dynamic statistical model which learns and adapts processing and parsing rules. First, we
limit the amount of unstructured text by clustering based on semantics of log file lines. Next, we only take
the most relevant cluster into account and focus only on those frequent patterns which lead to the desired
output table similar to Vaarandi [10]. Furthermore, we transform the found frequent patterns and the
output stating the parsed table into a Hidden Markov Model (HMM). We use this HMM as a specific,
however, flexible representation of a pattern for log file parsing to maintain high quality output. After
training our model on one system type and applying it to a different system with slightly different log file
patterns, we achieve an accuracy over 99.99%.
Predicting electricity consumption using hidden parametersIJLT EMAS
data mining technique to forecast power demand of a
biological region based on the metrological conditions. The value
forecast analytical data mining technique is implement with the
Hidden Marko Model. The morals of the factor such as heat,
clamminess and municipal celebration on which influence
operation depends and the everyday utilization morals compose
the data. Data mining operation are perform on this
chronological data to form a forecast model which is able of
predict every day utilization provide the meteorological
parameter. The steps of information detection of data process are
implemented. The data is preprocessed and fed to HMM for
guidance it. The educated HMM network is used to predict the
electricity demand for the given meteorological conditions.
This document describes a new method for training Gaussian mixture classifiers for hyperspectral image classification. The method uses dynamic pruning, splitting, and merging of Gaussian mixture kernels to automatically determine the appropriate number of components during training. This "structural learning" approach is employed to model and classify hyperspectral imagery data. Experimental results on AVIRIS hyperspectral data sets suggest this approach is a potential alternative to traditional Gaussian mixture modeling and classification using expectation-maximization.
Q UANTUM C LUSTERING -B ASED F EATURE SUBSET S ELECTION FOR MAMMOGRAPHIC I...ijcsit
In this paper, we present an algorithm for feature selection. This algorithm labeled QC-FS: Quantum
Clustering for Feature Selection performs the selection in two steps. Partitioning the original features
space in order to group similar features is performed using the Quantum Clustering algorithm. Then the
selection of a representative for each cluster is carried out. It uses similarity measures such as correlation
coefficient (CC) and the mutual information (MI). The feature which maximizes this information is chosen
by the algorithm
Implementation of Fuzzy Logic for the High-Resolution Remote Sensing Images w...IOSR Journals
This document describes an implementation of fuzzy logic for high-resolution remote sensing image classification with improved accuracy. It discusses using an object-based approach with fuzzy rules to classify urban land covers in a satellite image. The approach involves image segmentation using k-means clustering or ISODATA clustering. Features are then extracted from the image objects and fuzzy logic is applied to classify the objects based on membership functions. The method was tested on different sensor and resolution images in MATLAB and showed improved classification accuracy over other techniques, achieving lower entropy in results. Future work planned includes designing an unsupervised classification model combining k-means clustering and fuzzy-based object orientation.
This PhD research proposal discusses using Bayesian inference methods for multi-target tracking in big data settings. The researcher proposes developing new stochastic MCMC algorithms that can scale to billions of data points by using small subsets of data in each iteration. This would make Bayesian methods computationally feasible for big data. The proposal outlines reviewing relevant literature, developing the theoretical foundations, and empirically validating new algorithms like sequential Monte Carlo on real-world problems to analyze text and user preferences at large scale.
A Survey on: Hyper Spectral Image Segmentation and Classification Using FODPSOrahulmonikasharma
The Spatial analysis of image sensed and captured from a satellite provides less accurate information about a remote location. Hence analyzing spectral becomes essential. Hyper spectral images are one of the remotely sensed images, they are superior to multispectral images in providing spectral information. Detection of target is one of the significant requirements in many are assuc has military, agriculture etc. This paper gives the analysis of hyper spectral image segmentation using fuzzy C-Mean (FCM)clustering technique with FODPSO classifier algorithm. The 2D adaptive log filter is proposed to denoise the sensed and captured hyper spectral image in order to remove the speckle noise.
Oversampling technique in student performance classification from engineering...IJECEIAES
This document discusses various oversampling techniques for dealing with imbalanced data in student performance classification. It compares SMOTE, Borderline-SMOTE, SVMSMOTE, and ADASYN oversampling combined with MLP, gradient boosting, AdaBoost, and random forest classifiers. The results show that Borderline-SMOTE gave the best performance for predicting the minority (low performance) class according to several evaluation metrics. SVMSMOTE also performed well overall, particularly for recall, F1-measure, and AUC. Gradient boosting provided high and consistent precision, recall, F1-measure, and AUC across the different oversampling methods.
A Novel Penalized and Compensated Constraints Based Modified Fuzzy Possibilis...ijsrd.com
A cluster is a group of objects which are similar to each other within a cluster and are dissimilar to the objects of other clusters. The similarity is typically calculated on the basis of distance between two objects or clusters. Two or more objects present inside a cluster and only if those objects are close to each other based on the distance between them.The major objective of clustering is to discover collection of comparable objects based on similarity metric. Fuzzy Possibilistic C-Means (FPCM) is the effective clustering algorithm available to cluster unlabeled data that produces both membership and typicality values during clustering process. In this approach, the efficiency of the Fuzzy Possibilistic C-means clustering approach is enhanced by using the penalized and compensated constraints based FPCM (PCFPCM). The proposed PCFPCM approach differ from the conventional clustering techniques by imposing the possibilistic reasoning strategy on fuzzy clustering with penalized and compensated constraints for updating the grades of membership and typicality. The performance of the proposed approaches is evaluated on the University of California, Irvine (UCI) machine repository datasets such as Iris, Wine, Lung Cancer and Lymphograma. The parameters used for the evaluation is Clustering accuracy, Mean Squared Error (MSE), Execution Time and Convergence behavior.
Hybrid features selection method using random forest and meerkat clan algorithmTELKOMNIKA JOURNAL
In the majority of gene expression investigations, selecting relevant genes for sample classification is considered a frequent challenge, with researchers attempting to discover the minimum feasible number of genes while yet achieving excellent predictive performance. Various gene selection methods employ univariate (gene-by-gene) gene relevance rankings as well as arbitrary thresholds for selecting the number of genes, are only applicable to 2-class problems and use gene selection ranking criteria unrelated to the algorithm of classification. A modified random forest (MRF) algorithm depending on the meerkat clan algorithm (MCA) is provided in this work.
It is one of the swarm intelligence algorithms and one of the most significant machine learning approaches in the decision tree. MCA is used to choose characteristics for the RF algorithm. In information systems, databases, and other applications, feature selection imputation is critical. The proposed algorithm was applied to three different databases, where the experimental results for accuracy and time proved the superiority of the proposed algorithm over the original algorithm.
GeoAI: A Model-Agnostic Meta-Ensemble Zero-Shot Learning Method for Hyperspec...Konstantinos Demertzis
The document discusses a new meta-ensemble zero-shot learning method called MAME-ZsL for hyperspectral image analysis and classification. MAME-ZsL overcomes the difficulties of traditional deep learning methods that require large labeled datasets and long training times. It reduces computational costs, avoids overfitting, and achieves high classification accuracy even when testing classes were not present during training. The method is a novel optimization-based meta-ensemble architecture that facilitates learning representations from limited labeled examples to enable one-shot and zero-shot learning.
ON THE PREDICTION ACCURACIES OF THREE MOST KNOWN REGULARIZERS : RIDGE REGRESS...ijaia
The document compares the prediction accuracies of ridge regression, lasso regression, and elastic net regularization methods on 13 datasets. It finds that the prediction accuracy depends heavily on the nature of the datasets. When using cross-validation, the lasso method worked better than ridge regression and elastic net on two datasets. When using BIC scoring, BIC produced better predictions than cross-validation for 6 datasets, especially favoring ridge regression on highly correlated datasets. Overall, ridge regression, lasso, and elastic net tended to perform similarly except on two datasets where ridge regression outperformed the others.
A Neural Network Approach to Identify Hyperspectral Image Content IJECEIAES
A Hyperspectral is the imaging technique that contains very large dimension data with the hundreds of channels. Meanwhile, the Hyperspectral Images (HISs) delivers the complete knowledge of imaging; therefore applying a classification algorithm is very important tool for practical uses. The HSIs are always having a large number of correlated and redundant feature, which causes the decrement in the classification accuracy; moreover, the features redundancy come up with some extra burden of computation that without adding any beneficial information to the classification accuracy. In this study, an unsupervised based Band Selection Algorithm (BSA) is considered with the Linear Projection (LP) that depends upon the metric-band similarities. Afterwards Monogenetic Binary Feature (MBF) has consider to perform the „texture analysis‟ of the HSI, where three operational component represents the monogenetic signal such as; phase, amplitude and orientation. In post processing classification stage, feature-mapping function can provide important information, which help to adopt the Kernel based Neural Network (KNN) to optimize the generalization ability. However, an alternative method of multiclass application can be adopt through KNN, if we consider the multi-output nodes instead of taking single-output node.
This document presents a new model called EQUIRS (Explicitly Query Understanding Information Retrieval System) based on Hidden Markov Models (HMM) to improve natural language processing for text query information retrieval. The proposed EQUIRS system is compared to previous fuzzy clustering methods. Experimental results on a dataset of 900 files across 5 categories show that EQUIRS has higher accuracy than fuzzy clustering, as measured by precision, recall, F-measure, though it has longer training and searching times. The document concludes that EQUIRS is an effective approach for information retrieval based on HMM.
Hydraulic fracturing stimulation designs are moving towards tighter spaced clusters, longer stage length, and more proppant volumes. However, effectively evaluating the hydraulic fracturing stimulation efficiency remains a challenge. Distributed fiber optic sensing, which includes DAS and DTS, can continuously monitor the hydraulic fracturing stimulation downhole and be compared with other monitoring technology such as microseismic.
The DAS and DTS data, when integrated with the microseismic, highlight processes relevant to the completion design and allow for a better understanding and interpretation of each dataset.
This paper outlines a workflow to improve processing and interpretation of DAS and DTS data. In addition,
an estimate of the slurry distribution can be made. These methods will be demonstrated for a horizontal
Wolfcamp well in the Permian Basin. Here we compare key aspects of the microseismic, DAS, and DTS
results in several fracture stages to understand the downhole geomechanical processes. In order to interpret
the DTS data a thermal model is developed (using DTS data) to simulate the temperature behavior after
pumping has ceased. A slurry distribution is obtained by matching the simulated temperature with the
measured temperature from DTS. In addition, the DAS data signal is studied in the frequency domain and
the dominant frequencies are identified that are mostly related to fluid flow and to reduce the background
noise. This time frequency analysis enhances the ability to monitor and optimize well treatments.
After reducing the background noise, the acoustic intensity is correlated to the slurry distribution. The fluid
distribution data from DAS and DTS are compared with the microseismic and near field strain to better
understand the completion processes. We utilized fiber optic microseismic to better understand and
compare it to conventional microseismic.
Finally, we highlight the dynamics of strain and microseismic signature as fluid moves from an offset well
completion into the prior stimulated fiber well to better understand the reservoir and far field effects of the
completion.
Interpretation Special-Section: Insights into digital oilfield data using ar...Pioneer Natural Resources
Invitation to contribute to our special issue titled “Insights to the Digital Oil Field data using Artificial Intelligence & Big Data Analytics”. The scope of this special issue is to further bridge the gap between the geophysical interpretation and well planning (drilling, completions and production).
The technology communities both in industry and academia are utilizing advanced signal processing / machine learning algorithms, high compute / Big Data architectures at scale to develop these practical solutions. Your contributions to advanced algorithms in signal processing/machine learning, subsurface imaging and interpretation will be a good fit for this issue.
We hope you will find this participation both rewarding and worthwhile for our industry and to the Interpretation community in general.
Dear Colleagues,
Call for papers for another Machine Learning special issue of SEG/AAPG Journal of Interpretation focusing on the Seismic Data Analysis has been announced.
We look forward to your contribution.
Vikram Jayaram
Special Section Editor
Interpretation
An approach to offer management: maximizing sales with fare products and anci...Pioneer Natural Resources
With the growth in ancillary sales, an area of increasing importance for airlines is the concept of offer management, which entails the creation of dynamic, custom, personalized offers consisting of a flight itinerary and ancillary products offered by an airline. This practice-oriented, overview paper provides an end-to-end, future-oriented framework for determining the composition of optimal base fare and ancillary bundles by customer trip purpose segment followed by 1:1 personalization to maximize total sales. Our focus in this paper is primarily on the proposed offer management framework and its sub-components.
The document discusses using discrete wavelet transform (DWT) and principal component analysis (PCA) as decorrelating transforms for hyperspectral image classification under JPEG2000 compression. It compares the classification performance of DWT and PCA when applying lossless compression and two JPEG2000 scalability options: color and quality. Color scalability decompresses a subset of bands, while quality scalability assigns more bits to important bands. The DWT provides similar classification to PCA but is faster and does not require additional files. Reordering bands by variance before color decompression improved DWT classification results compared to using the initial DWT band order.
The document discusses optimizing rate allocation for compressing hyperspectral images using JPEG2000. It proposes using the discrete wavelet transform instead of the Karhunen-Loeve transform for decorrelation due to lower computational complexity. A mixed model is used for rate distortion optimal bit allocation instead of experimentally obtained rate distortion curves. Comparisons show the mixed model approach results in lower mean squared error than traditional bit allocation schemes, while having lower implementation complexity than prior methods.
Detection and Classification in Hyperspectral Images using Rate Distortion an...Pioneer Natural Resources
This document summarizes an experiment that compares two methods of bit allocation for compressing hyperspectral imagery using JPEG2000: 1) the traditional high bit rate quantizer approach and 2) the rate distortion optimal (RDO) approach. The experiment shows that both methods perform well at relatively low bit rates, achieving over 96% classification accuracy. However, at very low bit rates, the RDO approach outperforms the high bit rate quantizer approach, achieving 90% accuracy at 0.0375 bpppb compared to less than 90% for the high bit rate method. The RDO approach also achieves lower mean squared error than the high bit rate quantizer approach.
Maruthi Prithivirajan, Head of ASEAN & IN Solution Architecture, Neo4j
Get an inside look at the latest Neo4j innovations that enable relationship-driven intelligence at scale. Learn more about the newest cloud integrations and product enhancements that make Neo4j an essential choice for developers building apps with interconnected data and generative AI.
Sudheer Mechineni, Head of Application Frameworks, Standard Chartered Bank
Discover how Standard Chartered Bank harnessed the power of Neo4j to transform complex data access challenges into a dynamic, scalable graph database solution. This keynote will cover their journey from initial adoption to deploying a fully automated, enterprise-grade causal cluster, highlighting key strategies for modelling organisational changes and ensuring robust disaster recovery. Learn how these innovations have not only enhanced Standard Chartered Bank’s data infrastructure but also positioned them as pioneers in the banking sector’s adoption of graph technology.
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
Threats to mobile devices are more prevalent and increasing in scope and complexity. Users of mobile devices desire to take full advantage of the features
available on those devices, but many of the features provide convenience and capability but sacrifice security. This best practices guide outlines steps the users can take to better protect personal devices and information.
For the full video of this presentation, please visit: https://www.edge-ai-vision.com/2024/06/building-and-scaling-ai-applications-with-the-nx-ai-manager-a-presentation-from-network-optix/
Robin van Emden, Senior Director of Data Science at Network Optix, presents the “Building and Scaling AI Applications with the Nx AI Manager,” tutorial at the May 2024 Embedded Vision Summit.
In this presentation, van Emden covers the basics of scaling edge AI solutions using the Nx tool kit. He emphasizes the process of developing AI models and deploying them globally. He also showcases the conversion of AI models and the creation of effective edge AI pipelines, with a focus on pre-processing, model conversion, selecting the appropriate inference engine for the target hardware and post-processing.
van Emden shows how Nx can simplify the developer’s life and facilitate a rapid transition from concept to production-ready applications.He provides valuable insights into developing scalable and efficient edge AI solutions, with a strong focus on practical implementation.
GraphSummit Singapore | The Art of the Possible with Graph - Q2 2024Neo4j
Neha Bajwa, Vice President of Product Marketing, Neo4j
Join us as we explore breakthrough innovations enabled by interconnected data and AI. Discover firsthand how organizations use relationships in data to uncover contextual insights and solve our most pressing challenges – from optimizing supply chains, detecting fraud, and improving customer experiences to accelerating drug discoveries.
In his public lecture, Christian Timmerer provides insights into the fascinating history of video streaming, starting from its humble beginnings before YouTube to the groundbreaking technologies that now dominate platforms like Netflix and ORF ON. Timmerer also presents provocative contributions of his own that have significantly influenced the industry. He concludes by looking at future challenges and invites the audience to join in a discussion.
Why You Should Replace Windows 11 with Nitrux Linux 3.5.0 for enhanced perfor...SOFTTECHHUB
The choice of an operating system plays a pivotal role in shaping our computing experience. For decades, Microsoft's Windows has dominated the market, offering a familiar and widely adopted platform for personal and professional use. However, as technological advancements continue to push the boundaries of innovation, alternative operating systems have emerged, challenging the status quo and offering users a fresh perspective on computing.
One such alternative that has garnered significant attention and acclaim is Nitrux Linux 3.5.0, a sleek, powerful, and user-friendly Linux distribution that promises to redefine the way we interact with our devices. With its focus on performance, security, and customization, Nitrux Linux presents a compelling case for those seeking to break free from the constraints of proprietary software and embrace the freedom and flexibility of open-source computing.
Communications Mining Series - Zero to Hero - Session 1DianaGray10
This session provides introduction to UiPath Communication Mining, importance and platform overview. You will acquire a good understand of the phases in Communication Mining as we go over the platform with you. Topics covered:
• Communication Mining Overview
• Why is it important?
• How can it help today’s business and the benefits
• Phases in Communication Mining
• Demo on Platform overview
• Q/A
Observability Concepts EVERY Developer Should Know -- DeveloperWeek Europe.pdfPaige Cruz
Monitoring and observability aren’t traditionally found in software curriculums and many of us cobble this knowledge together from whatever vendor or ecosystem we were first introduced to and whatever is a part of your current company’s observability stack.
While the dev and ops silo continues to crumble….many organizations still relegate monitoring & observability as the purview of ops, infra and SRE teams. This is a mistake - achieving a highly observable system requires collaboration up and down the stack.
I, a former op, would like to extend an invitation to all application developers to join the observability party will share these foundational concepts to build on:
“An Outlook of the Ongoing and Future Relationship between Blockchain Technologies and Process-aware Information Systems.” Invited talk at the joint workshop on Blockchain for Information Systems (BC4IS) and Blockchain for Trusted Data Sharing (B4TDS), co-located with with the 36th International Conference on Advanced Information Systems Engineering (CAiSE), 3 June 2024, Limassol, Cyprus.
Generative AI Deep Dive: Advancing from Proof of Concept to ProductionAggregage
Join Maher Hanafi, VP of Engineering at Betterworks, in this new session where he'll share a practical framework to transform Gen AI prototypes into impactful products! He'll delve into the complexities of data collection and management, model selection and optimization, and ensuring security, scalability, and responsible use.
Dr. Sean Tan, Head of Data Science, Changi Airport Group
Discover how Changi Airport Group (CAG) leverages graph technologies and generative AI to revolutionize their search capabilities. This session delves into the unique search needs of CAG’s diverse passengers and customers, showcasing how graph data structures enhance the accuracy and relevance of AI-generated search results, mitigating the risk of “hallucinations” and improving the overall customer journey.
Goodbye Windows 11: Make Way for Nitrux Linux 3.5.0!SOFTTECHHUB
As the digital landscape continually evolves, operating systems play a critical role in shaping user experiences and productivity. The launch of Nitrux Linux 3.5.0 marks a significant milestone, offering a robust alternative to traditional systems such as Windows 11. This article delves into the essence of Nitrux Linux 3.5.0, exploring its unique features, advantages, and how it stands as a compelling choice for both casual users and tech enthusiasts.
2. mixture model from the multivariate HSI data using an HMM. This HMM approach estimates the proportion of
each HSI class present in a finite mixture model by incorporating both the estimation step and model selection
in a single algorithm. The model selection step that was previously introduced in7 automatically assigns mixture
components for a GM. Our technique utilizes a reduced dimensional feature space to model and synthesize the
approximate observations of the true HSI data. In order to define the relevance of using finite mixture model for
HSI, let us consider a random variable X, the finite mixture models decompose a PDF f (x) into sum of K class
PDFs. A general density function f (x) is considered K class probability density functions. A general density
function f (x) is considered semiparametric, since it may be decomposed into K components. Let fk (x) denote
the k th class PDF. The finite mixture model with K components expands as
K
ak fk (x),
f (x) =
(1)
k=1
where ak denotes the proportion of the k th class. The proportion ak may be interpreted as the prior probability of
observing a sample from class k. Furthermore, the prior probabilities ak for each distribution must be nonnegative
and sum-to-one, or
(2)
ak ≥ 0 f or k = 1, · · ·, K,
where
K
ak = 1.
(3)
k=1
Since, the underlying probability densities of the mixture are initially unknown, one must estimate the
densities from samples of each class iteratively. Thus, we formally extend the PDF based classification approach
to the analysis of HSI data (dependent data). In our approach we adapt a stationary Markovian model which
is a powerful stochastic model that can closely approximate many naturally occurring phenomena. One such
famous phenomena is approximation of human speech.8 While a very powerful stochastic model, a single HMM
cannot easily act as a good classifier between a wide variety of signal classes. Instead, it is best to design them
specifically for each signal type and feature type.
The rest of the paper is organized as follows. In Section II, we mention some examples of previous work in
literature that is related to our experiments. In Section III, we review some of the basics of HMM formulation,
problem of mixture learning and density estimation. Section IV describes briefly about minimum noise fraction
(MNF) transform. Section V reports experimental results, and Section VI ends the paper by presenting some
concluding remarks.
2. EARLIER WORK
Few investigations have introduced HMM in HSI processing in recent times. In Du et. al.,9 hidden Markov
model information divergence (HMMID) was introduced as a discriminatory measure among target spectra.
Comparison were made to deterministic distance metrics such as the spectral angle mapper (SAM) and minimum
Euclidean distance (MED). More recently, Bali et. al.10 shows the problem of joint segmentation of hyperspectral
images in the Bayesian framework. This approach based on a HMM of the images with common segmentation, or
equivalently with common hidden classification label variables which is modeled by a Potts Markov Random Field.
In a related work, Li et. al.11 proposed a two dimensional HMM for image classification. This method provided a
structured way to incorporate context information into classification. All the above mentioned approaches come
under the domain of supervised techniques. The modest development of unsupervised classification techniques
in the HSI regime has been the primary source of motivation for the proposed work.
Multidimensional data such as the HSI can be modeled by a multidimensional Gaussian mixture (GM).12
Normally, a GM in the form of the PDF for z ∈ RP is given by
L
αi N (z, μi , Σi )
p(z) =
i=1
Proc. of SPIE Vol. 7334 73340F-2
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
3. z
255
Feature space (3D)
Feature Vector
Scatter plot (20)
Figure 1. An illustration of Image and Feature space representation.
where
N (z, μi , Σi ) =
−1
1
1
e{− 2 (z−µi ) Σi (z−µi )} .
(2π)P/2 |Σi |1/2
Here L is the number of mixture components and P the number of spectral channels (bands). The GM parameters
are denoted by λ = {αi , μi , Σi }. The parameters of the GM are estimated using maximum likelihood by means
of the EM algorithm. In7 we show the structural learning of a GM that is employed to model and classify HSI
data. This methodology utilizes a fast and automatic assignment of mixture components to model PDFs. Later
on, we employ the same mechanism to estimate parameters and further model state PDFs of an HMM.
Consider a data that consists of K samples of dimension P , it is not necessary or even desirable to group all
the data together in to a single KXP -dimensional sample. In the simplest case, all K samples are independent
and we may regard them as samples of the same RV. For most practical cases, they are not independent. The
Markovian principle assumes consecutive samples are statistically independent when conditioned on knowing the
samples that preceded it. This leads to an elegant solution of HMM which employs a set of M PDFs of dimension
P . The HMM regards each of the K samples as having originated from one of the M possible states and there
is a distinct probability that the underlying model “jumps” from one state to another. In our approach, the
HMMs uses GM to model each state PDFs.5 We have focused on an unsupervised learning algorithm for ML
parameter estimation which in turn is used as a reduced dimensional PDF based classifier.
3. UNSUPERVISED LEARNING OF MARKOV SOURCES
In this section, we mention general formulation of HMM, re-estimation of HMM parameters, observed PDFs and
GM parameters. Let us begin following the notational approach of Rabiner,8 consider there are T observation
times. At each time 1 ≤ t ≤ T , there is a discrete state variable qt which takes one of N values qt ∈ {S1 , S2 , · ·
·, SN }. According to the Markovian assumption, the probability distribution of qt+1 depends only on the value
of qt . This is described compactly as a state transition probability matrix A whose elements aij represents the
probability that qt+1 equals j given that qt equals i. The initial state probabilities are denoted πi , the probability
that q1 equals Si . It is hidden Markov model because the states qt are hidden from view; that is we cannot
observe them. But, we can observe the random data Ot which is generated according to a PDF dependent on
the state at time t as illustrated in Figure 2. We denote the PDF of Ot under state j as bj (Ot ). The complete
set of model parameters that define the HMM are ∧ = {πj , aij , bj }.
The EM also known as the Baum-Welch algorithm calculates new estimates ∧ given an observation sequence O =
O1 , O2 , O3 , · · ·OT and a previous estimate of ∧. The algorithm is composed of two parts: the forward/backward
procedure and the re-estimation of parameters.
Using Gaussian Mixtures for bj (Ot )
We model the PDFs bj (Ot ) as GM,
Proc. of SPIE Vol. 7334 73340F-3
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
4. S
I
A
T
i.0
2
30
0
0
0
a
0
0
0
p
p
14 0
0
43
40
S
V
I
I
I
Hidden
SthS
p (z)
4'
2:
1
I
p (z)
p (z)
2
p (z)
3
4
1
p (z)
5
Observer ® ®
Figure 2. A hidden Markov model. The observer makes his observations whose PDF depends on the state.
M
cjm N (O, μjm , Ujm ),
bj (O) =
1≤j≤N
m=1
where
N (O, μjm , Ujm ) =
1
P
2π 2 | Ujm |
− 1 (O−µjm )
2
e
U−1 (O−µjm )
jm
and P is the dimension of O. We will refer to these GM parameters collectively as bj
,
{cjm , μjm , Ujm }.
Forward/Backward Procedure
We wish to compute the probability of observation sequence O = O1 , O2 , ···, OT given the model ∧ = {πj , aij , bj }.
The forward procedure for p(O|∧) is
• Initialization:
1≤i≤N
αi = πi bi (O1 ),
• Induction:
N
αt+1 (j) =
1≤t≤T −1
αt (i)aij bj (Ot+1 ),
i=1
• Termination:
N
p(O|∧) =
αT (i)
i=1
The backward procedure is
• Initialization:
βt (i) = 1
Proc. of SPIE Vol. 7334 73340F-4
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
1≤j≤N
5. • Induction:
N
βt (i) =
t = T − 1, T − 2, · · ··, 1
aij bj (Ot+1 )βt+1 (j),
1≤i≤N
j=1
Re-estimation of HMM parameters
The re-estimation procedure calculates new estimates of ∧ given the observation sequence O = O1 , O2 , O3 , ···OT .
We first define
ξt (i, j) =
N
i=1
αt (i)aij bj (Ot+1 )βt+1 (j)
N
j=1
αt (i)aij bj (Ot+1 )βt+1 (j)
and
N
γt (i) =
ξt (i, j).
j=1
The updated state priors are
πi = γ1 (i).
The updated state transition matrix is
T −1
t=1 ξt (i, j)
.
T −1
t=1 γt (i)
aij =
Re-estimation of Observation PDF’s13
In order to update the observation PDF’s, it is necessary to maximize
T
Qj =
wtj log bj (Ot )
t=1
over the PDF bj , where
αt (j)βt (j)
.
N
i=1 αt (i)βt (i)
wt,j =
This is a “weighted” likelihood (ML) procedure since if wtj = cj , the results are strict ML estimates. The weights
wtj are interpreted as the probability that the Markov chain is in state j at time t.
Re-estimation of Gaussian Mixture Parameters
In our experiments, bj (O) are modeled as GM by simply determining the weighted ML estimates of the GM
parameters. This would require iterating to convergence at each step. A more global approach is possible if the
mixture components assignments are regarded as “missing data”.13 The result is that the quantity
T
M
Qj =
γt (j, m) log bj (Ot )
t=1 m=1
is maximized, where
Proc. of SPIE Vol. 7334 73340F-5
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
6. cjm N (Ot , μjm , Ujm )
M
k=1 cjk N (Ot , μjm , Ujm )
γt (j, m) = wt,j
.
Here, the weights γt (j, m) are interpreted as the probability that the Markov chain is in state j and the
observation is from mixture component m at time t. The resulting update equations for cjm , μjm , and Ujm are
computed as follows:
cjm =
ˆ
T
t=1
T
t=1
γt (j, m)
M
l=1
γt (j, l)
.
The above expression is similar to re-estimation of GM.5 This means that the algorithms designed for GM
are applicable for updating the state PDFs of the HMM. Therefore,
T
t=1 γt (j, m)Ot
T
t=1 γt (j, m)
μjm =
ˆ
ˆ
Ujm =
T
t=1
γt (j, m)(Ot − μjm )(Ot − μjm )
T
t=1
γt (j, m)
.
4. MINIMUM NOISE FRACTION TRANSFORM
Before we begin the our section on experiments, we shall define minimum noise fraction (MNF) transform
since we use them to obtain a 2D feature plot of the true data as shown in Figure 3 (right). The MNF
transformation is a highly useful spectral processing tool in HSI analysis.14 It is used to determine the inherent
dimensionality of image data, to segregate noise in the data, and to reduce the computational requirements
for subsequent processing. This transform is essentially two cascaded principal components transformations.
The first transformation, based on an estimated noise covariance matrix, decorrelates and rescales the noise
in the data. This first step results in transformed data in which the noise has unit variance and no band-toband correlations. The second step is a standard principal components transformation of the noise-whitened
data. For the purposes of further spectral processing, the inherent dimensionality of the data is determined by
examination of the final eigenvalues and the associated images. The data space can be divided into two parts:
one part associated with large eigenvalues and coherent eigenimages, and a complementary part with near-unity
eigenvalues and noise-dominated images. By using only the coherent portions, the noise is separated from the
data, thus the image bands get ranked based on signal to noise ratios (SNR).
5. EXPERIMENTS
The remote sensing data sets that we have used in our experiments comes from an Airborne Visible/Infrared
Imaging Spectrometer (AVIRIS) sensor image. AVIRIS is a unique optical sensor that delivers calibrated images
of the upwelling spectral radiance in 224 contiguous spectral bands with wavelengths corresponding to 0.4-2.5
μm. AVIRIS is flown all across the US, Canada and Europe. Figure 3 shows data sets used in our experiments
that belong to a Indian Pine scene in northwest Indiana. The spatial bands of this scene are of size 169 X 169
pixels. Since, HSI imagery is highly correlated in the spectral direction using MNF rotation is an obvious choice
for decorrelation among the bands. This also results in a 2D “scatter” plot of the first two MNF components
of the data as shown in Figure 3. The scatter plots used in the paper are similar to marginalized PDF on
any 2D plane. Marginalization could be easily depicted for visualizing state PDFs of HMM. To illustrate this
visualization scheme, let z = [z1 , z2 , z3 , z4 ]. For example, to visualize on the (z2 , z4 ) plane, we would need to
compute
p(z2 , z4 ) =
z1
z3
p(z1 , z2 , z3 , z4 )dz1 dz3 .
Proc. of SPIE Vol. 7334 73340F-6
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
7. I
Figure 3. (Left) Composite image of Indian Pine scene. (Right) Composite image of MNF transformed scene.
This utility is very useful when visualizing high-dimensional PDF. On basis of an initial analysis based on
Iso-data and K-means unsupervised classifiers, it was found that the scene consisted of 3 prominent mixture
classes. Therefore, we begin the training by considering a tri-state (corresponding to the 3 mixture classes
identified)uniform state transition matrix A and prior probability π to initialize the HMM parameters. The
PDF of the feature vector in each state is approximated by Gaussian mixtures. The automatic learning and
initialization of the Gaussian mixtures are explicitly dealt in our earlier work.7 The algorithm outputs the total
log likelihood at each iteration.
Training an HMM is an iterative process that seeks to maximize the probability that the HMM accounts for
the example sequences. However, there are chances of running into a “local maximum” problem; the model,
though converged to some locally optimal choice of parameters, is not guaranteed to be the best possible model.
In an attempt to avoid this pitfall, we use a simulated annealing procedure along side of training. This step is
performed by expanding the covariance matrices of the PDF estimates and by pushing the state transition matrix
and prior state probabilities closer to “uniform”. We attempt to find a “bad” stationary point by re-running
the above sequence until one is found. The PDF plots of the three state PDF’s after convergence are shown in
Figures 5, 6 and 7. In our experiments, (both model and synthesis stage) we use Viterbi algorithm8 to estimate
the most likely state sequence. A few outliers are also observed in one or more state PDFs. Now that we have
mixtures modeled by their corresponding state PDFs, we would like to test the model by generating synthetic
observations. In Figure 8 we were able to synthesize 100 observation. We clearly notice that the synthetic
observations closely approximate the true data observations. This result is also exemplified in Figure 9 when we
compare true states of the data with the estimated states of the synthetic observations. Similarly, in Figures 10
and 11 we show instances that compare 300 and 600 synthetic observations to the true data. These comparisons
show that the underlying mixture density were adequately modeled using HMM.
6. CONCLUSIONS
In this paper, we proposed the use of hidden Markov model that uses structural learning for approximating
underlying mixture densities. Algorithm test were carried out using real Hyperspectral data consisting of a
scene from Indian pines of NW Indiana. In our experiments, we utilized only the first two components of MNF
transformed bands to ensure feature learning in reduced representation of the data. We show that mixture
learning for multivariate Gaussians is very similar to learning HMM parameters. In fact, unsupervised learning
of GM parameters for each class are seamlessly integrated to model the state PDFs of a HMM in a single
algorithm. This technique could be applied to any type of parameter mixture model that utilizes EM algorithm.
Our experiments exemplifies that the proposed method models and well synthesizes the observations of the
HSI data in a reduced dimensional feature space. This technique can considered a new paradigm of reduced
dimensional classifier in processing HSI data.
Proc. of SPIE Vol. 7334 73340F-7
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
8. 40
7000
6000
30
7000
20
5000
4000
5000
10
4000
MNF
Band
2
6000
3000
3000
0
2000
2000
1000
−10
0
−10
0
−10
MNF Band 1
20
10
20
−20
−20
1000
10
0
30
30
0
MNF Band 2
−10
0
10
20
30
40
MNF Band 1
Figure 4. (Left) 2D scatter plot of MNF transformed Band 1 Vs. Band 2. (Right) 2D Histogram of MNF bands 1 and 2.
State 1
MNF2
40
20
0
−20
−30
−20
−10
0
10
MNF1
20
30
40
50
−20
−10
0
10
MNF1
20
30
40
50
MNF2
40
20
0
−20
−30
Figure 5. (Top) 2D scatter plot of true data. (Bottom) PDF of State 1 after convergence.
Proc. of SPIE Vol. 7334 73340F-8
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
9. State 2
MNF2
40
20
0
−20
−30
−20
−10
0
10
MNF1
20
30
40
50
−20
−10
0
10
MNF1
20
30
40
50
MNF2
40
20
0
−20
−30
Figure 6. (Top) 2D scatter plot of true data. (Bottom) PDF of State 2 after convergence.
State 3
MNF2
40
20
0
−20
−30
−20
−10
0
10
MNF1
20
30
40
50
−20
−10
0
10
MNF1
20
30
40
50
MNF2
40
20
0
−20
−30
Figure 7. (Top) 2D scatter plot of true data. (Bottom) PDF of State 3 after convergence.
Proc. of SPIE Vol. 7334 73340F-9
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
10. 40
30
Original Samples
Synthetic Samples
MNF Band 2
20
10
0
−10
−20
−20
−10
0
10
MNF Band 1
20
30
40
Figure 8. Comparison of true data Vs. 100 synthetic observations.
3
2.8
True States
Estimates States
2.6
2.4
States
2.2
2
1.8
1.6
1.4
1.2
1
0
10
20
30
40
50
60
70
80
90
100
No.of Samples
Figure 9. Comparison of true states vs. estimated states from synthetic observation.
Proc. of SPIE Vol. 7334 73340F-10
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
11. 40
30
Original Samples
Synthetic Samples
MNF Band 2
20
10
0
−10
−20
−20
−10
0
10
20
30
40
MNF Band 1
Figure 10. Comparison of true data Vs. 300 synthetic observations.
40
30
Original Samples
Synthetic Samples
MNF Band 2
20
10
0
−10
−20
−20
−10
0
10
20
30
MNF Band 1
Figure 11. Comparison of true data Vs. 600 synthetic observations.
Proc. of SPIE Vol. 7334 73340F-11
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms
40
12. ACKNOWLEDGMENTS
We would like to thank department of Geological Sciences at UTEP for providing access to the ENVI software
and LARS, Purdue University for making the HSI data15 available. This work was supported by NASA Earth
System Science (ESS) doctoral fellowship at the University of Texas at El Paso.
REFERENCES
[1] Schott, J. R., [Remote Sensing: The Image Chain Approach], Oxford University Press.
[2] Hughes, G. F., “On the mean accuracy of statistical pattern recognizers,” IEEE Transactions on Information
Theory 14, 55–63 (1968).
[3] Shaw, G. and Manolakis, D., “Signal processing for hyperspectral image exploitation,” IEEE Signal Processing Magazine 19, 12–16 (2002).
[4] Keshava, N., “Distance metrics & band selection in hyperspectral processing with applications to material
identification and spectral libraries,” IEEE Transactions on Geoscience and Remote Sensing 42, No. 7,
1552–1565 (July 2004).
[5] McLachlan, G. and Peel, D., [Finite Mixture Models], Wiley Series in Probability and Statistics, New York,
NY, second ed. (2000).
[6] Figueiredo, M. A. T. and Jain, A. K., “Unsupervised learning of finite mixture models,” IEEE Transactions
on Pattern Analysis and Machine Intelligence 24, 381–396 (2002).
[7] Jayaram, V. and Usevitch, B., “Dynamic mixing kernels in gaussian mixture classifier for hyperspectral
classification,” in [Mathematics of Data/Image Pattern Recognition, Compression, and Encryption with
Applications XI, Proceedings of the SPIE], 70750L–70750L–8 (2008).
[8] Rabiner, L. R., “A tutorial on hidden Markov models and selected applications in speech recognition,” in
[Proceedings of the IEEE], 257–286 (1989).
[9] Du, Q. and Chang, C.-I., “A hidden markov model approach to spectral analysis for hyperspectral imagery,”
Optical Engineering 40, No. 10, 2277–2284 (2001).
[10] Bali, N. and Mohammad-Djafari, A., “Bayesian approach with hidden Markov modeling and Mean Field
Approximation for Hyperspectral data analysis,” IEEE Transactions on Image Processing 17, No. 2, 217–
225 (2008).
[11] J. Li, A. N. and Gray, R. M., “Image classification by a two-dimensional hidden Markov model,” IEEE
Trans. Signal Processing 48, 517–533 (2000).
[12] Marden, D. B. and Manolakis, D. G., “Modeling hyperspectral imaging data,” in [Algorithms and Technologies for Multispectral, Hyperspectral, and Ultraspectral Imagery IX. Edited by Shen, Sylvia S.; Lewis, Paul
E. Proceedings of the SPIE], 253–262 (2003).
[13] Huang, B. H., “Maximum likelihood estimation for mixture multivariate stochastic observations of markov
chains,” in [AT&T Technical Journal], 1235–1249 (1985).
[14] A. A. Green, M. Berman, P. S. and Craig, M. D., “A transformation for ordering multispectral data in
terms of image quality with implications for noise removal,” IEEE Transcations on Geoscience and Remote
Sensing 26, 65–74 (1988).
[15] Landgrebe, D., “AVIRIS derived Northwest Indianas Indian Pines 1992 Hyperspectral dataset,”
http://dynamo.ecn.purdue.edu/ biehl/MultiSpec/documentation.html. .
Proc. of SPIE Vol. 7334 73340F-12
Downloaded From: http://spiedigitallibrary.org/ on 09/04/2013 Terms of Use: http://spiedl.org/terms