This paper proposes a clustering algorithm based on the Self Organizing Map (SOM) method. To find the optimal number of clusters, our algorithm uses the Davies Bouldin index which has not been used previously in the multi-SOM. The proposed algorithm is compared to three clustering methods based on five databases. Results show that our algorithm is as performing as concurrent methods.
This document presents an improved multi-SOM clustering algorithm that uses the Davies-Bouldin index to determine the optimal number of clusters. The multi-SOM algorithm iteratively clusters an initial self-organizing map (SOM) grid using the DB index at each level until the index reaches its minimum value, indicating the best number of clusters. Experimental results on five datasets show the proposed algorithm performs as well as or better than k-means, BIRCH, and a previous multi-SOM algorithm in determining the correct number of clusters.
Kernel based similarity estimation and real time tracking of movingIAEME Publication
This document discusses kernel-based mean shift algorithm for real-time object tracking. It presents the following:
1) The algorithm uses kernel density estimation to calculate the similarity between a target model and candidate windows, using the Bhattacharyya coefficient. 2) It can successfully track objects moving uniformly at slow speeds but struggles with fast or non-uniform motion, or changes in scale. 3) The algorithm was tested on video streams and could track objects moving slowly but failed for fast or irregular motion. Adaptive target windows are needed to handle changes in scale.
This document presents a scalable method for image classification using sparse coding and dictionary learning. It proposes parallelizing the computation of image similarity for faster recognition. Specifically, it distributes the task of measuring similarity between images among multiple cores in a cluster. Experimental results on a face recognition dataset show nearly linear speedup when balancing the dataset size and number of nodes. Reconstruction errors are used as a similarity measure, with dictionaries learned using K-SVD for each image. The proposed parallel method distributes this similarity computation process to achieve faster image classification.
Online Multi-Person Tracking Using Variance Magnitude of Image colors and Sol...Pourya Jafarzadeh
The document describes a multi-object tracking method that formulates tracking as a Short Minimum Clique Problem (SMCP). It uses three consecutive frames divided into three clusters, where each clique between clusters represents a tracklet (partial trajectory) of a person. Edges between clusters are weighted based on color histogram similarity and eigenvalue similarity of bounding boxes. Occlusion handling is performed by saving color histograms of occluded people in a buffer and comparing them to newly detected people. The method was evaluated on challenging datasets and shown to achieve promising results compared to state-of-the-art methods.
Textual information in images constitutes a very rich source of high-level semantics for retrieval and indexing. In this paper, a new approach is proposed using Cellular Automata (CA) which strives towards identifying scene text on natural images. Initially, a binary edge map is calculated. Then, taking advantage of the CA flexibility, the transition rules are changing and are applied in four consecutive steps resulting in four time steps CA evolution. Finally, a post-processing technique based on edge projection analysis is employed for high density edge images concerning the elimination of possible false positives. Evaluation results indicate considerable performance gains without sacrificing text detection accuracy.
In recent machine learning community, there is a trend of constructing a linear logarithm version of
nonlinear version through the ‘kernel method’ for example kernel principal component analysis, kernel
fisher discriminant analysis, support Vector Machines (SVMs), and the current kernel clustering
algorithms. Typically, in unsupervised methods of clustering algorithms utilizing kernel method, a
nonlinear mapping is operated initially in order to map the data into a much higher space feature, and then
clustering is executed. A hitch of these kernel clustering algorithms is that the clustering prototype resides
in increased features specs of dimensions and therefore lack intuitive and clear descriptions without
utilizing added approximation of projection from the specs to the data as executed in the literature
presented. This paper aims to utilize the ‘kernel method’, a novel clustering algorithm, founded on the
conventional fuzzy clustering algorithm (FCM) is anticipated and known as kernel fuzzy c-means algorithm
(KFCM). This method embraces a novel kernel-induced metric in the space of data in order to interchange
the novel Euclidean matric norm in cluster prototype and fuzzy clustering algorithm still reside in the space
of data so that the results of clustering could be interpreted and reformulated in the spaces which are
original. This property is used for clustering incomplete data. Execution on supposed data illustrate that
KFCM has improved performance of clustering and stout as compare to other transformations of FCM for
clustering incomplete data.
Self-Organizing Maps (SOM) are a type of neural network that can be used for clustering and visualizing complex, high-dimensional data. SOM reduces dimensionality while preserving topological relationships. It arranges nodes on a grid such that similar input vectors are mapped to nearby nodes. During training, the best matching node and its neighbors are adjusted to better match the input. This results in a 2D map where similar data clusters together. For example, a SOM was used to cluster countries based on quality of life indicators, grouping those with similar living standards. SOM can be useful for applications like data mining, pattern recognition, and more.
An efficient technique for color image classification based on lower feature ...Alexander Decker
This document discusses an efficient technique for color image classification using support vector machines with radial basis functions (SVM-RBF). It presents SVM-RBF as an improvement over other classification methods like SVM with ant colony optimization (SVM-ACO) and directed acyclic graph (SVM-DAG). The paper tests the different classifiers on 600 images across 3 classes, finding SVM-RBF achieved the highest precision and recall rates, with precision of 92.3-94% and recall of 84.8-91%. It concludes SVM-RBF more effectively reduces noise and the semantic gap to enhance image classification performance compared to the other methods.
This document presents an improved multi-SOM clustering algorithm that uses the Davies-Bouldin index to determine the optimal number of clusters. The multi-SOM algorithm iteratively clusters an initial self-organizing map (SOM) grid using the DB index at each level until the index reaches its minimum value, indicating the best number of clusters. Experimental results on five datasets show the proposed algorithm performs as well as or better than k-means, BIRCH, and a previous multi-SOM algorithm in determining the correct number of clusters.
Kernel based similarity estimation and real time tracking of movingIAEME Publication
This document discusses kernel-based mean shift algorithm for real-time object tracking. It presents the following:
1) The algorithm uses kernel density estimation to calculate the similarity between a target model and candidate windows, using the Bhattacharyya coefficient. 2) It can successfully track objects moving uniformly at slow speeds but struggles with fast or non-uniform motion, or changes in scale. 3) The algorithm was tested on video streams and could track objects moving slowly but failed for fast or irregular motion. Adaptive target windows are needed to handle changes in scale.
This document presents a scalable method for image classification using sparse coding and dictionary learning. It proposes parallelizing the computation of image similarity for faster recognition. Specifically, it distributes the task of measuring similarity between images among multiple cores in a cluster. Experimental results on a face recognition dataset show nearly linear speedup when balancing the dataset size and number of nodes. Reconstruction errors are used as a similarity measure, with dictionaries learned using K-SVD for each image. The proposed parallel method distributes this similarity computation process to achieve faster image classification.
Online Multi-Person Tracking Using Variance Magnitude of Image colors and Sol...Pourya Jafarzadeh
The document describes a multi-object tracking method that formulates tracking as a Short Minimum Clique Problem (SMCP). It uses three consecutive frames divided into three clusters, where each clique between clusters represents a tracklet (partial trajectory) of a person. Edges between clusters are weighted based on color histogram similarity and eigenvalue similarity of bounding boxes. Occlusion handling is performed by saving color histograms of occluded people in a buffer and comparing them to newly detected people. The method was evaluated on challenging datasets and shown to achieve promising results compared to state-of-the-art methods.
Textual information in images constitutes a very rich source of high-level semantics for retrieval and indexing. In this paper, a new approach is proposed using Cellular Automata (CA) which strives towards identifying scene text on natural images. Initially, a binary edge map is calculated. Then, taking advantage of the CA flexibility, the transition rules are changing and are applied in four consecutive steps resulting in four time steps CA evolution. Finally, a post-processing technique based on edge projection analysis is employed for high density edge images concerning the elimination of possible false positives. Evaluation results indicate considerable performance gains without sacrificing text detection accuracy.
In recent machine learning community, there is a trend of constructing a linear logarithm version of
nonlinear version through the ‘kernel method’ for example kernel principal component analysis, kernel
fisher discriminant analysis, support Vector Machines (SVMs), and the current kernel clustering
algorithms. Typically, in unsupervised methods of clustering algorithms utilizing kernel method, a
nonlinear mapping is operated initially in order to map the data into a much higher space feature, and then
clustering is executed. A hitch of these kernel clustering algorithms is that the clustering prototype resides
in increased features specs of dimensions and therefore lack intuitive and clear descriptions without
utilizing added approximation of projection from the specs to the data as executed in the literature
presented. This paper aims to utilize the ‘kernel method’, a novel clustering algorithm, founded on the
conventional fuzzy clustering algorithm (FCM) is anticipated and known as kernel fuzzy c-means algorithm
(KFCM). This method embraces a novel kernel-induced metric in the space of data in order to interchange
the novel Euclidean matric norm in cluster prototype and fuzzy clustering algorithm still reside in the space
of data so that the results of clustering could be interpreted and reformulated in the spaces which are
original. This property is used for clustering incomplete data. Execution on supposed data illustrate that
KFCM has improved performance of clustering and stout as compare to other transformations of FCM for
clustering incomplete data.
Self-Organizing Maps (SOM) are a type of neural network that can be used for clustering and visualizing complex, high-dimensional data. SOM reduces dimensionality while preserving topological relationships. It arranges nodes on a grid such that similar input vectors are mapped to nearby nodes. During training, the best matching node and its neighbors are adjusted to better match the input. This results in a 2D map where similar data clusters together. For example, a SOM was used to cluster countries based on quality of life indicators, grouping those with similar living standards. SOM can be useful for applications like data mining, pattern recognition, and more.
An efficient technique for color image classification based on lower feature ...Alexander Decker
This document discusses an efficient technique for color image classification using support vector machines with radial basis functions (SVM-RBF). It presents SVM-RBF as an improvement over other classification methods like SVM with ant colony optimization (SVM-ACO) and directed acyclic graph (SVM-DAG). The paper tests the different classifiers on 600 images across 3 classes, finding SVM-RBF achieved the highest precision and recall rates, with precision of 92.3-94% and recall of 84.8-91%. It concludes SVM-RBF more effectively reduces noise and the semantic gap to enhance image classification performance compared to the other methods.
Handwritten and Machine Printed Text Separation in Document Images using the ...Konstantinos Zagoris
In a number of types of documents, ranging from forms to archive documents and books with annotations, machine printed and handwritten text may be present in the same document image, giving rise to significant issues within a digitisation and recognition pipeline. It is therefore necessary to separate the two types of text before applying different recognition methodologies to each. In this paper, a new approach is proposed which strives towards identifying and separating handwritten from machine printed text using the Bag of Visual Words paradigm (BoVW). Initially, blocks of interest are detected in the document image. For each block, a descriptor is calculated based on the BoVW. The final characterization of the blocks as Handwritten,Machine Printed or Noise is made by a Support Vector Machine classifier. The promising performance of the proposed approach is shown by using a consistent evaluation methodology which couples meaningful measures along with a new dataset.
A novel secure image steganography method based on chaos theory in spatial do...ijsptm
This paper presents a novel approach of building a secure data hiding technique in digital images. The
image steganography technique takes the advantage of limited power of human visual system (HVS). It uses
image as cover media for embedding secret message. The most important requirement for a steganographic
algorithm is to be imperceptible while maximizing the size of the payload. In this paper a method is
proposed to encrypt the secret bits of the message based on chaos theory before embedding into the cover
image. A 3-3-2 LSB insertion method has been used for image steganography. Experimental results show a
substantial improvement in the Peak Signal to Noise Ratio (PSNR) and Image Fidelity (IF) value of the
proposed technique over the base technique of 3-3-2 LSB insertion.
A new block cipher for image encryption based on multi chaotic systemsTELKOMNIKA JOURNAL
In this paper, a new algorithm for image encryption is proposed based on three chaotic systems which are Chen system,logistic map and two-dimensional (2D) Arnold cat map. First, a permutation scheme is applied to the image, and then shuffled image is partitioned into blocks of pixels. For each block, Chen system is employed for confusion and then logistic map is employed for generating subsititution-box (S-box) to substitute image blocks. The S-box is dynamic, where it is shuffled for each image block using permutation operation. Then, 2D Arnold cat map is used for providing diffusion, after that XORing the result using Chen system to obtain the encrypted image.The high security of proposed algorithm is experimented using histograms, unified average changing intensity (UACI), number of pixels change rate (NPCR), entropy, correlation and keyspace analyses.
Spectroscopy or hyperspectral imaging consists in the acquisition, analysis, and extraction of the spectral information measured on a specific region or object using an airborne or satellite device. Hyperspectral imaging has become an active field of research recently. One way of analysing such data is through clustering. However, due to the high dimensionality of the data and the small distance between the different material signatures, clustering such a data is a challenging task.In this paper, we empirically compared five clustering techniques in different hyperspectral data sets. The considered clustering techniques are K-means, K-medoids, fuzzy Cmeans, hierarchical, and density-based spatial clustering of applications with noise. Four data sets are used to achieve this purpose which is Botswana, Kennedy space centre, Pavia, and Pavia University. Beside the accuracy, we adopted four more similarity measures: Rand statistics, Jaccard coefficient, Fowlkes-Mallows index, and Hubert index. According to accuracy, we found that fuzzy C-means clustering is doing better on Botswana and Pavia data sets, K-means and K-medoids are giving better results on Kennedy space centre data set, and for Pavia University the hierarchical clustering is better
Big data Clustering Algorithms And StrategiesFarzad Nozarian
The document discusses various algorithms for big data clustering. It begins by covering preprocessing techniques such as data reduction. It then covers hierarchical, prototype-based, density-based, grid-based, and scalability clustering algorithms. Specific algorithms discussed include K-means, K-medoids, PAM, CLARA/CLARANS, DBSCAN, OPTICS, MR-DBSCAN, DBCURE, and hierarchical algorithms like PINK and l-SL. The document emphasizes techniques for scaling these algorithms to large datasets, including partitioning, sampling, approximation strategies, and MapReduce implementations.
The document provides an overview of self-organizing maps (SOM). It defines SOM as an unsupervised learning technique that reduces the dimensions of data through the use of self-organizing neural networks. SOM is based on competitive learning where the closest neural network unit to the input vector (the best matching unit or BMU) is identified and adjusted along with neighboring units. The algorithm involves initializing weight vectors, presenting input vectors, identifying the BMU, and updating weights of the BMU and neighboring units. SOM can be used for applications like dimensionality reduction, clustering, and visualization.
This document compares the k-means and grid density clustering algorithms. K-means partitions data into k clusters based on minimizing distances between points and cluster centroids. It works well with numerical data but can be affected by outliers. Grid density determines dense grids based on neighbor densities and can handle different shaped and multi-density clusters without knowing the number of clusters beforehand. It has advantages over k-means in that it can handle categorical data, noise and arbitrary shaped clusters.
This document summarizes a research paper that proposes a new density-based clustering technique called Triangle-Density Based Clustering Technique (TDCT) to efficiently cluster large spatial datasets. TDCT uses a polygon approach where the number of data points inside each triangle of a polygon is calculated to determine triangle densities. Triangle densities are used to identify clusters based on a density confidence threshold. The technique aims to identify clusters of arbitrary shapes and densities while minimizing computational costs. Experimental results demonstrate the technique's superiority in terms of cluster quality and complexity compared to other density-based clustering algorithms.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Hiding text in speech signal using K-means, LSB techniques and chaotic maps IJECEIAES
In this paper, a new technique that hides a secret text inside a speech signal without any apparent noise is presented. The technique for encoding the secret text is through first scrambling the text using Chaotic Map, then encoding the scraped text using the Zaslavsky map, and finally hiding the text by breaking the speech signal into blocks and using only half of each block with the LSB, K-means algorithms. The measures (SNR, PSNR, Correlation, SSIM, and MSE) are used on various speech files (“.WAV”), and various secret texts. We observed that the suggested technique offers high security (SNR, PSNR, Correlation, and SSIM) of an encrypted text with low error (MSE). This indicates that the noise level in the speech signal is very low and the speech purity is high, so the suggested method is effective for embedding encrypted text into speech files.
Intrusion Detection Model using Self Organizing Maps.Tushar Shinde
Proposed Model:
[I] Pre-processing of server logs:
Our web-site server log file analyser performs the following steps when provided with a log file:
1) It scans the entries in the log files to help identify unique visitor’s sessions.
2) For each identified sessions, the analyser has to examine its key matching features to generate the session’s dimensional feature-vector representation.
[II] Session identification:
In this process of dividing a web-site server access log enters into sessions. Session identification is performed by:
1) Grouping all HTTP requests on web-sites that originate from the same IP address that matches the visitor and also are described by the same user-agent strings.
2) By applying a timeout approach to divide into unique sessions to avoid any mishaps.
[III] Dataset labelling:
labels each feature-vector as belonging to one of the following four categories:
1. Human visitor’s normal Known.
2. well-behaved web-site attackers.
3. malicious attackers.
4. unknown visitors unidentified.
Thus, allow a better understanding of the cluster’s nature and significance results can be generated.
Techniques:
SOM Algorithm, NNtool, MATLAB, WEKA toolkit, KDD Data-set.
Self Organizing Feature Map(SOM), Topographic Product, Cascade 2 AlgorithmChenghao Jin
This document discusses several algorithms used in the SCPSNSP method for day-ahead electricity load and price forecasting, including Self Organizing Feature Map (SOM), Topographic Product, and Cascade 2. SOM is used for clustering electricity time series data into patterns. The Topographic Product measures how well the feature map represents the input data. Cascade 2, a neural network training algorithm, is used for next symbol prediction based on the clustered pattern sequences.
Architecture neural network deep optimizing based on self organizing feature ...journalBEEI
Forward neural network (FNN) execution relying on the algorithm of training and architecture selection. Different parameters using for nip out the architecture of FNN such as the connections number among strata, neurons hidden number in each strata hidden and hidden strata number. Feature architectural combinations exponential could be uncontrollable manually so specific architecture can be design automatically by using special algorithm which build system with ability generalization better. Determination of architecture FNN can be done by using the algorithm of optimization numerous. In this paper methodology new proposes achievement where FNN neurons respective with hidden layers estimation work where in this work collect algorithm training self organizing feature map (SOFM) with advantages to explain how the best architectural selected automatically by SOFM from criteria error testing based on architecture populated. Different size of dataset benchmark of 4 classifications tested for approach proposed.
Segmentation - based Historical Handwritten Word Spotting using document-spec...Konstantinos Zagoris
Many word spotting strategies for the modern documents are not directly applicable to historical handwritten documents due to writing styles variety and intense degradation. In this paper, a new method that permits effective word spotting in handwritten documents is presented that relies upon document-specific local features which take into account texture information around representative keypoints. Experimental work on two historical handwritten datasets using standard evaluation measures shows the improved performance achieved by the proposed methodology.
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...cscpconf
In this paper, Design and Implementation of Binary Neural Network Learning with Fuzzy
Clustering (DIBNNFC), is proposed to classify semisupervised data, it is based on the
concept of binary neural network and geometrical expansion. Parameters are updated
according to the geometrical location of the training samples in the input space, and each
sample in the training set is learned only once. It’s a semisupervised based approach, the
training samples are semi-labelled i.e. for some samples, labels are known and for some
samples data labels are not known. The method starts with classification, which is done by
using the concept of ETL algorithm. In classification process various classes are formed.
These classes classify samples in to two classes after that considers each class as a region and calculates the average of the entire region separately. This average is centres of the region which is used for the purpose of clustering by using FCM algorithm. Once clustering process over labelling of semi supervised data is done, then whole samples would be classify by (DIBNNFC). The method proposes here is exhaustively tested with different benchmark datasets and it is found that, on increasing value of training parameters number of hidden neurons and training time both are getting decrease. The result reported, using real character recognition data set and result will compare with existing semi-supervised classifier, the proposed approach learned with semi-supervised leads to higher classification accuracy.
Chaos Image Encryption Methods: A Survey StudyjournalBEEI
This document surveys various chaos encryption techniques for encrypting image data. It begins by explaining why traditional encryption techniques are unsuitable for images and how chaos encryption provides an effective solution. It then provides background on chaos theory and describes the general process of chaos-based image encryption. The document proceeds to summarize several specific chaos encryption algorithms proposed in other papers, evaluating aspects like key space, correlation coefficient, and resistance to attacks. It concludes that chaos encryption is an effective method for secure image encryption and multiple techniques can be combined to further increase security.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Improving search time for contentment based image retrieval via, LSH, MTRee, ...IOSR Journals
This document proposes a new index structure called LSH-LUBMTree to improve search time for content-based image retrieval using the Earth Mover's Distance metric. LSH-LUBMTree combines Locality Sensitive Hashing (LSH) and the LUBMTree index. Images hashed to the same bucket via LSH are then stored in the LUBMTree to reduce false positives and accelerate search time. Experimental results show LSH-LUBMTree performs better than standard LSH in terms of search time by leveraging advantages of both LSH and LUBMTree indexing.
CONTENT BASED VIDEO CATEGORIZATION USING RELATIONAL CLUSTERING WITH LOCAL SCA...ijcsit
This paper introduces a novel approach for efficient video categorization. It relies on two main
components. The first one is a new relational clustering technique that identifies video key frames by
learning cluster dependent Gaussian kernels. The proposed algorithm, called clustering and Local Scale
Learning algorithm (LSL) learns the underlying cluster dependent dissimilarity measure while finding
compact clusters in the given dataset. The learned measure is a Gaussian dissimilarity function defined
with respect to each cluster. We minimize one objective function to optimize the optimal partition and the
cluster dependent parameter. This optimization is done iteratively by dynamically updating the partition
and the local measure. The kernel learning task exploits the unlabeled data and reciprocally, the
categorization task takes advantages of the local learned kernel. The second component of the proposed
video categorization system consists in discovering the video categories in an unsupervised manner using
the proposed LSL. We illustrate the clustering performance of LSL on synthetic 2D datasets and on high
dimensional real data. Also, we assess the proposed video categorization system using a real video
collection and LSL algorithm.
This document discusses using a learning automata approach to predict target locations in wireless sensor networks to reduce energy consumption and improve tracking accuracy. It proposes a learning automata based method that uses a target's movement history to predict its next location. Related works on target tracking techniques like tree-based, cluster-based, and prediction-based methods are summarized. Learning automata concepts are introduced. Simulation results are said to show the proposed method improves energy efficiency, reduces missed targets, and decreases transmitted packets compared to other methods.
Comparison Between Clustering Algorithms for Microarray Data AnalysisIOSR Journals
Currently, there are two techniques used for large-scale gene-expression profiling; microarray and
RNA-Sequence (RNA-Seq).This paper is intended to study and compare different clustering algorithms that used
in microarray data analysis. Microarray is a DNA molecules array which allows multiple hybridization
experiments to be carried out simultaneously and trace expression levels of thousands of genes. It is a highthroughput
technology for gene expression analysis and becomes an effective tool for biomedical research.
Microarray analysis aims to interpret the data produced from experiments on DNA, RNA, and protein
microarrays, which enable researchers to investigate the expression state of a large number of genes. Data
clustering represents the first and main process in microarray data analysis. The k-means, fuzzy c-mean, selforganizing
map, and hierarchical clustering algorithms are under investigation in this paper. These algorithms
are compared based on their clustering model.
A Density Based Clustering Technique For Large Spatial Data Using Polygon App...IOSR Journals
This document presents a density-based clustering technique called TDCT (Triangle-density based clustering technique) for efficiently clustering large spatial datasets. The technique uses a polygon approach where the number of data points inside each triangle of a polygon is calculated. If the ratio of point densities between two neighboring triangles exceeds a threshold, the triangles are merged into the same cluster. The technique is capable of identifying clusters of arbitrary shapes and densities. Experimental results demonstrate the technique has superior cluster quality and complexity compared to other methods.
Handwritten and Machine Printed Text Separation in Document Images using the ...Konstantinos Zagoris
In a number of types of documents, ranging from forms to archive documents and books with annotations, machine printed and handwritten text may be present in the same document image, giving rise to significant issues within a digitisation and recognition pipeline. It is therefore necessary to separate the two types of text before applying different recognition methodologies to each. In this paper, a new approach is proposed which strives towards identifying and separating handwritten from machine printed text using the Bag of Visual Words paradigm (BoVW). Initially, blocks of interest are detected in the document image. For each block, a descriptor is calculated based on the BoVW. The final characterization of the blocks as Handwritten,Machine Printed or Noise is made by a Support Vector Machine classifier. The promising performance of the proposed approach is shown by using a consistent evaluation methodology which couples meaningful measures along with a new dataset.
A novel secure image steganography method based on chaos theory in spatial do...ijsptm
This paper presents a novel approach of building a secure data hiding technique in digital images. The
image steganography technique takes the advantage of limited power of human visual system (HVS). It uses
image as cover media for embedding secret message. The most important requirement for a steganographic
algorithm is to be imperceptible while maximizing the size of the payload. In this paper a method is
proposed to encrypt the secret bits of the message based on chaos theory before embedding into the cover
image. A 3-3-2 LSB insertion method has been used for image steganography. Experimental results show a
substantial improvement in the Peak Signal to Noise Ratio (PSNR) and Image Fidelity (IF) value of the
proposed technique over the base technique of 3-3-2 LSB insertion.
A new block cipher for image encryption based on multi chaotic systemsTELKOMNIKA JOURNAL
In this paper, a new algorithm for image encryption is proposed based on three chaotic systems which are Chen system,logistic map and two-dimensional (2D) Arnold cat map. First, a permutation scheme is applied to the image, and then shuffled image is partitioned into blocks of pixels. For each block, Chen system is employed for confusion and then logistic map is employed for generating subsititution-box (S-box) to substitute image blocks. The S-box is dynamic, where it is shuffled for each image block using permutation operation. Then, 2D Arnold cat map is used for providing diffusion, after that XORing the result using Chen system to obtain the encrypted image.The high security of proposed algorithm is experimented using histograms, unified average changing intensity (UACI), number of pixels change rate (NPCR), entropy, correlation and keyspace analyses.
Spectroscopy or hyperspectral imaging consists in the acquisition, analysis, and extraction of the spectral information measured on a specific region or object using an airborne or satellite device. Hyperspectral imaging has become an active field of research recently. One way of analysing such data is through clustering. However, due to the high dimensionality of the data and the small distance between the different material signatures, clustering such a data is a challenging task.In this paper, we empirically compared five clustering techniques in different hyperspectral data sets. The considered clustering techniques are K-means, K-medoids, fuzzy Cmeans, hierarchical, and density-based spatial clustering of applications with noise. Four data sets are used to achieve this purpose which is Botswana, Kennedy space centre, Pavia, and Pavia University. Beside the accuracy, we adopted four more similarity measures: Rand statistics, Jaccard coefficient, Fowlkes-Mallows index, and Hubert index. According to accuracy, we found that fuzzy C-means clustering is doing better on Botswana and Pavia data sets, K-means and K-medoids are giving better results on Kennedy space centre data set, and for Pavia University the hierarchical clustering is better
Big data Clustering Algorithms And StrategiesFarzad Nozarian
The document discusses various algorithms for big data clustering. It begins by covering preprocessing techniques such as data reduction. It then covers hierarchical, prototype-based, density-based, grid-based, and scalability clustering algorithms. Specific algorithms discussed include K-means, K-medoids, PAM, CLARA/CLARANS, DBSCAN, OPTICS, MR-DBSCAN, DBCURE, and hierarchical algorithms like PINK and l-SL. The document emphasizes techniques for scaling these algorithms to large datasets, including partitioning, sampling, approximation strategies, and MapReduce implementations.
The document provides an overview of self-organizing maps (SOM). It defines SOM as an unsupervised learning technique that reduces the dimensions of data through the use of self-organizing neural networks. SOM is based on competitive learning where the closest neural network unit to the input vector (the best matching unit or BMU) is identified and adjusted along with neighboring units. The algorithm involves initializing weight vectors, presenting input vectors, identifying the BMU, and updating weights of the BMU and neighboring units. SOM can be used for applications like dimensionality reduction, clustering, and visualization.
This document compares the k-means and grid density clustering algorithms. K-means partitions data into k clusters based on minimizing distances between points and cluster centroids. It works well with numerical data but can be affected by outliers. Grid density determines dense grids based on neighbor densities and can handle different shaped and multi-density clusters without knowing the number of clusters beforehand. It has advantages over k-means in that it can handle categorical data, noise and arbitrary shaped clusters.
This document summarizes a research paper that proposes a new density-based clustering technique called Triangle-Density Based Clustering Technique (TDCT) to efficiently cluster large spatial datasets. TDCT uses a polygon approach where the number of data points inside each triangle of a polygon is calculated to determine triangle densities. Triangle densities are used to identify clusters based on a density confidence threshold. The technique aims to identify clusters of arbitrary shapes and densities while minimizing computational costs. Experimental results demonstrate the technique's superiority in terms of cluster quality and complexity compared to other density-based clustering algorithms.
IJERA (International journal of Engineering Research and Applications) is International online, ... peer reviewed journal. For more detail or submit your article, please visit www.ijera.com
Hiding text in speech signal using K-means, LSB techniques and chaotic maps IJECEIAES
In this paper, a new technique that hides a secret text inside a speech signal without any apparent noise is presented. The technique for encoding the secret text is through first scrambling the text using Chaotic Map, then encoding the scraped text using the Zaslavsky map, and finally hiding the text by breaking the speech signal into blocks and using only half of each block with the LSB, K-means algorithms. The measures (SNR, PSNR, Correlation, SSIM, and MSE) are used on various speech files (“.WAV”), and various secret texts. We observed that the suggested technique offers high security (SNR, PSNR, Correlation, and SSIM) of an encrypted text with low error (MSE). This indicates that the noise level in the speech signal is very low and the speech purity is high, so the suggested method is effective for embedding encrypted text into speech files.
Intrusion Detection Model using Self Organizing Maps.Tushar Shinde
Proposed Model:
[I] Pre-processing of server logs:
Our web-site server log file analyser performs the following steps when provided with a log file:
1) It scans the entries in the log files to help identify unique visitor’s sessions.
2) For each identified sessions, the analyser has to examine its key matching features to generate the session’s dimensional feature-vector representation.
[II] Session identification:
In this process of dividing a web-site server access log enters into sessions. Session identification is performed by:
1) Grouping all HTTP requests on web-sites that originate from the same IP address that matches the visitor and also are described by the same user-agent strings.
2) By applying a timeout approach to divide into unique sessions to avoid any mishaps.
[III] Dataset labelling:
labels each feature-vector as belonging to one of the following four categories:
1. Human visitor’s normal Known.
2. well-behaved web-site attackers.
3. malicious attackers.
4. unknown visitors unidentified.
Thus, allow a better understanding of the cluster’s nature and significance results can be generated.
Techniques:
SOM Algorithm, NNtool, MATLAB, WEKA toolkit, KDD Data-set.
Self Organizing Feature Map(SOM), Topographic Product, Cascade 2 AlgorithmChenghao Jin
This document discusses several algorithms used in the SCPSNSP method for day-ahead electricity load and price forecasting, including Self Organizing Feature Map (SOM), Topographic Product, and Cascade 2. SOM is used for clustering electricity time series data into patterns. The Topographic Product measures how well the feature map represents the input data. Cascade 2, a neural network training algorithm, is used for next symbol prediction based on the clustered pattern sequences.
Architecture neural network deep optimizing based on self organizing feature ...journalBEEI
Forward neural network (FNN) execution relying on the algorithm of training and architecture selection. Different parameters using for nip out the architecture of FNN such as the connections number among strata, neurons hidden number in each strata hidden and hidden strata number. Feature architectural combinations exponential could be uncontrollable manually so specific architecture can be design automatically by using special algorithm which build system with ability generalization better. Determination of architecture FNN can be done by using the algorithm of optimization numerous. In this paper methodology new proposes achievement where FNN neurons respective with hidden layers estimation work where in this work collect algorithm training self organizing feature map (SOFM) with advantages to explain how the best architectural selected automatically by SOFM from criteria error testing based on architecture populated. Different size of dataset benchmark of 4 classifications tested for approach proposed.
Segmentation - based Historical Handwritten Word Spotting using document-spec...Konstantinos Zagoris
Many word spotting strategies for the modern documents are not directly applicable to historical handwritten documents due to writing styles variety and intense degradation. In this paper, a new method that permits effective word spotting in handwritten documents is presented that relies upon document-specific local features which take into account texture information around representative keypoints. Experimental work on two historical handwritten datasets using standard evaluation measures shows the improved performance achieved by the proposed methodology.
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...cscpconf
In this paper, Design and Implementation of Binary Neural Network Learning with Fuzzy
Clustering (DIBNNFC), is proposed to classify semisupervised data, it is based on the
concept of binary neural network and geometrical expansion. Parameters are updated
according to the geometrical location of the training samples in the input space, and each
sample in the training set is learned only once. It’s a semisupervised based approach, the
training samples are semi-labelled i.e. for some samples, labels are known and for some
samples data labels are not known. The method starts with classification, which is done by
using the concept of ETL algorithm. In classification process various classes are formed.
These classes classify samples in to two classes after that considers each class as a region and calculates the average of the entire region separately. This average is centres of the region which is used for the purpose of clustering by using FCM algorithm. Once clustering process over labelling of semi supervised data is done, then whole samples would be classify by (DIBNNFC). The method proposes here is exhaustively tested with different benchmark datasets and it is found that, on increasing value of training parameters number of hidden neurons and training time both are getting decrease. The result reported, using real character recognition data set and result will compare with existing semi-supervised classifier, the proposed approach learned with semi-supervised leads to higher classification accuracy.
Chaos Image Encryption Methods: A Survey StudyjournalBEEI
This document surveys various chaos encryption techniques for encrypting image data. It begins by explaining why traditional encryption techniques are unsuitable for images and how chaos encryption provides an effective solution. It then provides background on chaos theory and describes the general process of chaos-based image encryption. The document proceeds to summarize several specific chaos encryption algorithms proposed in other papers, evaluating aspects like key space, correlation coefficient, and resistance to attacks. It concludes that chaos encryption is an effective method for secure image encryption and multiple techniques can be combined to further increase security.
International Journal of Engineering and Science Invention (IJESI) is an international journal intended for professionals and researchers in all fields of computer science and electronics. IJESI publishes research articles and reviews within the whole field Engineering Science and Technology, new teaching methods, assessment, validation and the impact of new technologies and it will continue to provide information on the latest trends and developments in this ever-expanding subject. The publications of papers are selected through double peer reviewed to ensure originality, relevance, and readability. The articles published in our journal can be accessed online.
Improving search time for contentment based image retrieval via, LSH, MTRee, ...IOSR Journals
This document proposes a new index structure called LSH-LUBMTree to improve search time for content-based image retrieval using the Earth Mover's Distance metric. LSH-LUBMTree combines Locality Sensitive Hashing (LSH) and the LUBMTree index. Images hashed to the same bucket via LSH are then stored in the LUBMTree to reduce false positives and accelerate search time. Experimental results show LSH-LUBMTree performs better than standard LSH in terms of search time by leveraging advantages of both LSH and LUBMTree indexing.
CONTENT BASED VIDEO CATEGORIZATION USING RELATIONAL CLUSTERING WITH LOCAL SCA...ijcsit
This paper introduces a novel approach for efficient video categorization. It relies on two main
components. The first one is a new relational clustering technique that identifies video key frames by
learning cluster dependent Gaussian kernels. The proposed algorithm, called clustering and Local Scale
Learning algorithm (LSL) learns the underlying cluster dependent dissimilarity measure while finding
compact clusters in the given dataset. The learned measure is a Gaussian dissimilarity function defined
with respect to each cluster. We minimize one objective function to optimize the optimal partition and the
cluster dependent parameter. This optimization is done iteratively by dynamically updating the partition
and the local measure. The kernel learning task exploits the unlabeled data and reciprocally, the
categorization task takes advantages of the local learned kernel. The second component of the proposed
video categorization system consists in discovering the video categories in an unsupervised manner using
the proposed LSL. We illustrate the clustering performance of LSL on synthetic 2D datasets and on high
dimensional real data. Also, we assess the proposed video categorization system using a real video
collection and LSL algorithm.
This document discusses using a learning automata approach to predict target locations in wireless sensor networks to reduce energy consumption and improve tracking accuracy. It proposes a learning automata based method that uses a target's movement history to predict its next location. Related works on target tracking techniques like tree-based, cluster-based, and prediction-based methods are summarized. Learning automata concepts are introduced. Simulation results are said to show the proposed method improves energy efficiency, reduces missed targets, and decreases transmitted packets compared to other methods.
Comparison Between Clustering Algorithms for Microarray Data AnalysisIOSR Journals
Currently, there are two techniques used for large-scale gene-expression profiling; microarray and
RNA-Sequence (RNA-Seq).This paper is intended to study and compare different clustering algorithms that used
in microarray data analysis. Microarray is a DNA molecules array which allows multiple hybridization
experiments to be carried out simultaneously and trace expression levels of thousands of genes. It is a highthroughput
technology for gene expression analysis and becomes an effective tool for biomedical research.
Microarray analysis aims to interpret the data produced from experiments on DNA, RNA, and protein
microarrays, which enable researchers to investigate the expression state of a large number of genes. Data
clustering represents the first and main process in microarray data analysis. The k-means, fuzzy c-mean, selforganizing
map, and hierarchical clustering algorithms are under investigation in this paper. These algorithms
are compared based on their clustering model.
A Density Based Clustering Technique For Large Spatial Data Using Polygon App...IOSR Journals
This document presents a density-based clustering technique called TDCT (Triangle-density based clustering technique) for efficiently clustering large spatial datasets. The technique uses a polygon approach where the number of data points inside each triangle of a polygon is calculated. If the ratio of point densities between two neighboring triangles exceeds a threshold, the triangles are merged into the same cluster. The technique is capable of identifying clusters of arbitrary shapes and densities. Experimental results demonstrate the technique has superior cluster quality and complexity compared to other methods.
Vertex covering has important applications for wireless sensor networks such as monitoring link failures,
facility location, clustering, and data aggregation. In this study, we designed three algorithms for
constructing vertex cover in wireless sensor networks. The first algorithm, which is an adaption of the
Parnas & Ron’s algorithm, is a greedy approach that finds a vertex cover by using the degrees of the
nodes. The second algorithm finds a vertex cover from graph matching where Hoepman’s weighted
matching algorithm is used. The third algorithm firstly forms a breadth-first search tree and then
constructs a vertex cover by selecting nodes with predefined levels from breadth-first tree. We show the
operation of the designed algorithms, analyze them, and provide the simulation results in the TOSSIM
environment. Finally we have implemented, compared and assessed all these approaches. The transmitted
message count of the first algorithm is smallest among other algorithms where the third algorithm has
turned out to be presenting the best results in vertex cover approximation ratio.
An Efficient Clustering Method for Aggregation on Data FragmentsIJMER
Clustering is an important step in the process of data analysis with applications to numerous fields. Clustering ensembles, has emerged as a powerful technique for combining different clustering results to obtain a quality cluster. Existing clustering aggregation algorithms are applied directly to large number of data points. The algorithms are inefficient if the number of data points is large. This project defines an efficient approach for clustering aggregation based on data fragments. In fragment-based approach, a data fragment is any subset of the data. To increase the efficiency of the proposed approach, the clustering aggregation can be performed directly on data fragments under comparison measure and normalized mutual information measures for clustering aggregation, enhanced clustering aggregation algorithms are described. To show the minimal computational complexity. (Agglomerative, Furthest, and Local Search); nevertheless, which increases the accuracy.
Juha vesanto esa alhoniemi 2000:clustering of the somArchiLab 7
This document summarizes a research paper that proposes clustering the self-organizing map (SOM) as a way to analyze cluster structure in data. The paper discusses:
1) Using a two-stage process where data is first mapped to prototypes using a SOM, then the SOM prototypes are clustered, reducing computational load compared to direct data clustering.
2) Different clustering algorithms that can be applied to the SOM prototypes, including hierarchical and partitive (k-means) methods.
3) The benefits of clustering the SOM include noise reduction and being able to cluster large datasets more efficiently.
The document proposes a strategy for clustering distributed databases using self-organizing maps (SOM) and K-means algorithms. The strategy applies SOM locally to each distributed data set to obtain representative subsets, then combines the results and applies SOM and K-means globally. Specifically, it performs local SOM clustering, sends representative data to a central site, applies SOM again on the combined data, then uses K-means on the unified map to produce the final clustering result.
Clustering Using Shared Reference Points Algorithm Based On a Sound Data ModelWaqas Tariq
A novel clustering algorithm CSHARP is presented for the purpose of finding clusters of arbitrary shapes and arbitrary densities in high dimensional feature spaces. It can be considered as a variation of the Shared Nearest Neighbor algorithm (SNN), in which each sample data point votes for the points in its k-nearest neighborhood. Sets of points sharing a common mutual nearest neighbor are considered as dense regions/ blocks. These blocks are the seeds from which clusters may grow up. Therefore, CSHARP is not a point-to-point clustering algorithm. Rather, it is a block-to-block clustering technique. Much of its advantages come from these facts: Noise points and outliers correspond to blocks of small sizes, and homogeneous blocks highly overlap. This technique is not prone to merge clusters of different densities or different homogeneity. The algorithm has been applied to a variety of low and high dimensional data sets with superior results over existing techniques such as DBScan, K-means, Chameleon, Mitosis and Spectral Clustering. The quality of its results as well as its time complexity, rank it at the front of these techniques.
This document compares the k-means and grid density clustering algorithms. It summarizes that grid density clustering determines dense grids based on the densities of neighboring grids, and is able to handle different shaped clusters in multi-density environments. The grid density algorithm does not require distance computation and is not dependent on the number of clusters being known in advance like k-means. The document concludes that grid density clustering is better than k-means clustering as it can handle noise and outliers, find arbitrary shaped clusters, and has lower time complexity.
A PSO-Based Subtractive Data Clustering AlgorithmIJORCS
There is a tremendous proliferation in the amount of information available on the largest shared information source, the World Wide Web. Fast and high-quality clustering algorithms play an important role in helping users to effectively navigate, summarize, and organize the information. Recent studies have shown that partitional clustering algorithms such as the k-means algorithm are the most popular algorithms for clustering large datasets. The major problem with partitional clustering algorithms is that they are sensitive to the selection of the initial partitions and are prone to premature converge to local optima. Subtractive clustering is a fast, one-pass algorithm for estimating the number of clusters and cluster centers for any given set of data. The cluster estimates can be used to initialize iterative optimization-based clustering methods and model identification methods. In this paper, we present a hybrid Particle Swarm Optimization, Subtractive + (PSO) clustering algorithm that performs fast clustering. For comparison purpose, we applied the Subtractive + (PSO) clustering algorithm, PSO, and the Subtractive clustering algorithms on three different datasets. The results illustrate that the Subtractive + (PSO) clustering algorithm can generate the most compact clustering results as compared to other algorithms.
The document is a research paper that proposes and compares several clustering algorithms for remote sensing data:
1) DBSCAN, a density-based clustering algorithm that groups together densely populated areas.
2) OPTICS, an improved version of DBSCAN that handles varying cluster densities better.
3) Grid-based clustering that divides data into a grid for faster processing time.
4) Hybrid approaches like Grid-DBSCAN and Grid-OPTICS that combine grid-based clustering with DBSCAN and OPTICS to reduce computational complexity.
The paper evaluates and compares the accuracy and runtime of these algorithms on remote sensing image data.
A frame work for clustering time evolving dataiaemedu
The document proposes a framework for clustering time-evolving categorical data using a sliding window technique. It uses an existing clustering algorithm (Node Importance Representative) and a Drifting Concept Detection algorithm to detect changes in cluster distributions between the current and previous data windows. If a threshold difference in clusters is exceeded, reclustering is performed on the new window. Otherwise, the new clusters are added to the previous results. The framework aims to improve on prior work by handling drifting concepts in categorical time-series data.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
This document summarizes several cluster head selection techniques for mobile ad-hoc networks (MANETs). It begins with an introduction to MANETs and clustering in MANETs. It then surveys 12 different cluster head selection techniques: Lowest ID, Highest Degree, K-hop Connectivity ID, Mobility Based D-hop, Adaptive Cluster Load Balance, Least Cluster Change, Load Balancing, Power-aware Dominant Set, Weighted Approach, Max-Min D-cluster Formation, and Mobility Based Cluster Formation. It provides a brief description of each technique and analyzes their merits and demerits. Finally, it concludes that different techniques select the cluster head based on various parameters like node ID,
A Survey Paper on Cluster Head Selection Techniques for Mobile Ad-Hoc NetworkIOSR Journals
This document summarizes several cluster head selection techniques for mobile ad-hoc networks (MANETs). It discusses techniques that select the cluster head based on attributes like node ID, degree of connectivity, mobility, load balancing, and power consumption. Some techniques aim to improve stability and reduce overhead by minimizing cluster changes. Each technique has advantages like simplicity or load balancing, and disadvantages like additional messaging or inability to eliminate ties between nodes. The survey provides a comparison of the techniques on their selection criteria and merits and demerits.
This document discusses using artificial neural networks for network intrusion detection. Specifically, it proposes a hybrid classification model that uses entropy-based feature selection to reduce the dataset, followed by four neural network techniques (RBFN, SOM, SMO, PART) for classification. It provides details on each neural network technique and the overall methodology, which uses 10-fold cross validation to evaluate performance based on standard criteria. The goal is to build an efficient intrusion detection system with low false alarms and high detection rates.
Ameliorate the performance using soft computing approaches in wireless networksIJECEIAES
Wireless sensor networks are an innovative and rapidly advanced network occupying the broad spectrum of wireless networks. It works on the principle of “use with less expense, effort and with more comfort.” In these networks, routing provides efficient and effective data transmission between different sources to access points using the clustering technique. This work addresses the low-energy adaptive clustering hierarchy (LEACH) protocol’s main backdrop of choosing head nodes based on a random value. In this, the soft computing methods are used, namely the fuzzy approach, to overcome this barrier in LEACH. Our approach’s primary goal is to extend the network lifetime with efficient energy consumption and by choosing the appropriate head node in each cluster based on the fuzzy parameters. The proposed clustering algorithm focused on two fuzzy inference structures, namely Mamdani and Sugeno fuzzy logic models in two scenarios, respectively. We compared our approach with four existing works, the conventional LEACH, LEACH using the fuzzy method, multicriteria cluster head delegation, and fuzzy-based energy efficient clustering approach (FEECA) in wireless sensor network. The proposed scenario based fuzzy LEACH protocol approaches are better than the four existing methods regarding stability, network survivability, and energy consumption.
A fuzzy clustering algorithm for high dimensional streaming dataAlexander Decker
This document summarizes a research paper that proposes a new dimension-reduced weighted fuzzy clustering algorithm (sWFCM-HD) for high-dimensional streaming data. The algorithm can cluster datasets that have both high dimensionality and a streaming (continuously arriving) nature. It combines previous work on clustering algorithms for streaming data and high-dimensional data. The paper introduces the algorithm and compares it experimentally to show improvements in memory usage and runtime over other approaches for these types of datasets.
A generic algorithm to determine connected dominating sets for mobile ad hoc ...csandit
The high-level contributions of this paper are: (1) A generic algorithm to determine connected
dominating sets (CDS) for mobile ad hoc networks and its use to find CDS of different
categories: maximum density-based (MaxD-CDS), node ID-based (ID-CDS) and stability-based
(minimum velocity-based, MinV-CDS); (2) Performance comparison of the above three
categories of CDS algorithms with respect to two categories of mobility models: random node
mobility models (Random Waypoint model) and the grid-based vehicular ad hoc network
(VANET) mobility models (City Section and Manhattan mobility models), with respect to the
CDS Node Size and Lifetime. The performance of a CDS is observed to depend on the criteria
used to form the CDS (i.e., the node selection criteria of the underlying algorithm) and the
mobility model driving the topological changes. For each category of CDS, we identify the
mobility model under which one can simultaneously maximize the lifetime and node size, with
minimal tradeoff. For the two VANET mobility models, we also evaluate the impact of the grid
block length on the CDS lifetime and node size.
A GENERIC ALGORITHM TO DETERMINE CONNECTED DOMINATING SETS FOR MOBILE AD HOC ...csandit
This document summarizes a research paper that presents a generic algorithm for determining connected dominating sets (CDS) in mobile ad hoc networks. The algorithm can be used to find three types of CDS: maximum density-based, node ID-based, and stability-based. The performance of these three CDS algorithms is evaluated under different mobility models and metrics like CDS node size and lifetime. Key findings include that the performance of a CDS depends on the criteria used to form it and the mobility model, and certain combinations of CDS algorithm and mobility model can maximize lifetime and node size with minimal tradeoff.
This study Examines the Effectiveness of Talent Procurement through the Imple...DharmaBanothu
In the world with high technology and fast
forward mindset recruiters are walking/showing interest
towards E-Recruitment. Present most of the HRs of
many companies are choosing E-Recruitment as the best
choice for recruitment. E-Recruitment is being done
through many online platforms like Linkedin, Naukri,
Instagram , Facebook etc. Now with high technology E-
Recruitment has gone through next level by using
Artificial Intelligence too.
Key Words : Talent Management, Talent Acquisition , E-
Recruitment , Artificial Intelligence Introduction
Effectiveness of Talent Acquisition through E-
Recruitment in this topic we will discuss about 4important
and interlinked topics which are
A high-Speed Communication System is based on the Design of a Bi-NoC Router, ...DharmaBanothu
The Network on Chip (NoC) has emerged as an effective
solution for intercommunication infrastructure within System on
Chip (SoC) designs, overcoming the limitations of traditional
methods that face significant bottlenecks. However, the complexity
of NoC design presents numerous challenges related to
performance metrics such as scalability, latency, power
consumption, and signal integrity. This project addresses the
issues within the router's memory unit and proposes an enhanced
memory structure. To achieve efficient data transfer, FIFO buffers
are implemented in distributed RAM and virtual channels for
FPGA-based NoC. The project introduces advanced FIFO-based
memory units within the NoC router, assessing their performance
in a Bi-directional NoC (Bi-NoC) configuration. The primary
objective is to reduce the router's workload while enhancing the
FIFO internal structure. To further improve data transfer speed,
a Bi-NoC with a self-configurable intercommunication channel is
suggested. Simulation and synthesis results demonstrate
guaranteed throughput, predictable latency, and equitable
network access, showing significant improvement over previous
designs
Sri Guru Hargobind Ji - Bandi Chor Guru.pdfBalvir Singh
Sri Guru Hargobind Ji (19 June 1595 - 3 March 1644) is revered as the Sixth Nanak.
• On 25 May 1606 Guru Arjan nominated his son Sri Hargobind Ji as his successor. Shortly
afterwards, Guru Arjan was arrested, tortured and killed by order of the Mogul Emperor
Jahangir.
• Guru Hargobind's succession ceremony took place on 24 June 1606. He was barely
eleven years old when he became 6th Guru.
• As ordered by Guru Arjan Dev Ji, he put on two swords, one indicated his spiritual
authority (PIRI) and the other, his temporal authority (MIRI). He thus for the first time
initiated military tradition in the Sikh faith to resist religious persecution, protect
people’s freedom and independence to practice religion by choice. He transformed
Sikhs to be Saints and Soldier.
• He had a long tenure as Guru, lasting 37 years, 9 months and 3 days
Impartiality as per ISO /IEC 17025:2017 StandardMuhammadJazib15
This document provides basic guidelines for imparitallity requirement of ISO 17025. It defines in detial how it is met and wiudhwdih jdhsjdhwudjwkdbjwkdddddddddddkkkkkkkkkkkkkkkkkkkkkkkwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwioiiiiiiiiiiiii uwwwwwwwwwwwwwwwwhe wiqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq gbbbbbbbbbbbbb owdjjjjjjjjjjjjjjjjjjjj widhi owqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq uwdhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhwqiiiiiiiiiiiiiiiiiiiiiiiiiiiiw0pooooojjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjjj whhhhhhhhhhh wheeeeeeee wihieiiiiii wihe
e qqqqqqqqqqeuwiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiqw dddddddddd cccccccccccccccv s w c r
cdf cb bicbsad ishd d qwkbdwiur e wetwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwww w
dddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddddfffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffffw
uuuuhhhhhhhhhhhhhhhhhhhhhhhhe qiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiiii iqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqq eeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeee qqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqccccccccccccccccccccccccccccccccccccccccccccccccccccccccccccc ccccccccccccccccccccccccccccccccccc bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbu uuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuum
m
m mmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmm m i
g i dijsd sjdnsjd ndjajsdnnsa adjdnawddddddddddddd uw
We have designed & manufacture the Lubi Valves LBF series type of Butterfly Valves for General Utility Water applications as well as for HVAC applications.
Null Bangalore | Pentesters Approach to AWS IAMDivyanshu
#Abstract:
- Learn more about the real-world methods for auditing AWS IAM (Identity and Access Management) as a pentester. So let us proceed with a brief discussion of IAM as well as some typical misconfigurations and their potential exploits in order to reinforce the understanding of IAM security best practices.
- Gain actionable insights into AWS IAM policies and roles, using hands on approach.
#Prerequisites:
- Basic understanding of AWS services and architecture
- Familiarity with cloud security concepts
- Experience using the AWS Management Console or AWS CLI.
- For hands on lab create account on [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
# Scenario Covered:
- Basics of IAM in AWS
- Implementing IAM Policies with Least Privilege to Manage S3 Bucket
- Objective: Create an S3 bucket with least privilege IAM policy and validate access.
- Steps:
- Create S3 bucket.
- Attach least privilege policy to IAM user.
- Validate access.
- Exploiting IAM PassRole Misconfiguration
-Allows a user to pass a specific IAM role to an AWS service (ec2), typically used for service access delegation. Then exploit PassRole Misconfiguration granting unauthorized access to sensitive resources.
- Objective: Demonstrate how a PassRole misconfiguration can grant unauthorized access.
- Steps:
- Allow user to pass IAM role to EC2.
- Exploit misconfiguration for unauthorized access.
- Access sensitive resources.
- Exploiting IAM AssumeRole Misconfiguration with Overly Permissive Role
- An overly permissive IAM role configuration can lead to privilege escalation by creating a role with administrative privileges and allow a user to assume this role.
- Objective: Show how overly permissive IAM roles can lead to privilege escalation.
- Steps:
- Create role with administrative privileges.
- Allow user to assume the role.
- Perform administrative actions.
- Differentiation between PassRole vs AssumeRole
Try at [killercoda.com](https://killercoda.com/cloudsecurity-scenario/)
Supermarket Management System Project Report.pdfKamal Acharya
Supermarket management is a stand-alone J2EE using Eclipse Juno program.
This project contains all the necessary required information about maintaining
the supermarket billing system.
The core idea of this project to minimize the paper work and centralize the
data. Here all the communication is taken in secure manner. That is, in this
application the information will be stored in client itself. For further security the
data base is stored in the back-end oracle and so no intruders can access it.
Tools & Techniques for Commissioning and Maintaining PV Systems W-Animations ...Transcat
Join us for this solutions-based webinar on the tools and techniques for commissioning and maintaining PV Systems. In this session, we'll review the process of building and maintaining a solar array, starting with installation and commissioning, then reviewing operations and maintenance of the system. This course will review insulation resistance testing, I-V curve testing, earth-bond continuity, ground resistance testing, performance tests, visual inspections, ground and arc fault testing procedures, and power quality analysis.
Fluke Solar Application Specialist Will White is presenting on this engaging topic:
Will has worked in the renewable energy industry since 2005, first as an installer for a small east coast solar integrator before adding sales, design, and project management to his skillset. In 2022, Will joined Fluke as a solar application specialist, where he supports their renewable energy testing equipment like IV-curve tracers, electrical meters, and thermal imaging cameras. Experienced in wind power, solar thermal, energy storage, and all scales of PV, Will has primarily focused on residential and small commercial systems. He is passionate about implementing high-quality, code-compliant installation techniques.
Height and depth gauge linear metrology.pdfq30122000
Height gauges may also be used to measure the height of an object by using the underside of the scriber as the datum. The datum may be permanently fixed or the height gauge may have provision to adjust the scale, this is done by sliding the scale vertically along the body of the height gauge by turning a fine feed screw at the top of the gauge; then with the scriber set to the same level as the base, the scale can be matched to it. This adjustment allows different scribers or probes to be used, as well as adjusting for any errors in a damaged or resharpened probe.
Open Channel Flow: fluid flow with a free surfaceIndrajeet sahu
Open Channel Flow: This topic focuses on fluid flow with a free surface, such as in rivers, canals, and drainage ditches. Key concepts include the classification of flow types (steady vs. unsteady, uniform vs. non-uniform), hydraulic radius, flow resistance, Manning's equation, critical flow conditions, and energy and momentum principles. It also covers flow measurement techniques, gradually varied flow analysis, and the design of open channels. Understanding these principles is vital for effective water resource management and engineering applications.
1. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
DOI : 10.5121/ijnsa.2013.5414 181
AN IMPROVED MULTI-SOM ALGORITHM
Imen Khanchouch1
, Khaddouja Boujenfa2
and Mohamed Limam3
1
LARODEC ISG, University of Tunis
kh.imen88@gmail.com
2
LARODEC ISG, University of Tunis
khadouja.Boujenfa@isg.rnu.tn
3
LARODEC, ISG University of Tunis, Dhofar University, Oman
Mohamed.limam@isg.rnu.tn
ABSTRACT
This paper proposes a clustering algorithm based on the Self Organizing Map (SOM) method. To find the
optimal number of clusters, our algorithm uses the Davies Bouldin index which has not been used
previously in the multi-SOM. The proposed algorithm is compared to three clustering methods based on
five databases. Results show that our algorithm is as performing as concurrent methods.
KEYWORDS
Clustering, SOM, multi-SOM, DB index.
1. INTRODUCTION
Clustering is an unsupervised learning technique aiming to obtain homogeneous partitions of
objects while promoting the heterogeneity between partitions.In the literature there are many
clustering categories such as hierarchical [13], partition-based [5], density-based [1] and neuronal
networks (NN) [6].
Hierarchical methods aim to build a hierarchy of clusters with many levels. There are two types
of hierarchical clustering approaches: the agglomerative methods (bottom-up) and the divisive
methods (Top-down).Agglomerative methods start by many data objects taken as clusters and are
successively joined two by two until obtaining a single partition containing all the objects.
However, divisive methods begin with a sample of data as one cluster and successively get N
divided clusters as objects. Hierarchical methods are time consuming in the presence of large
amount of data. Consequently, the resulting dendrogram is very large and may include incorrect
information.
Partitioning methods divide the data set into disjoint partitions where each partition represents a
cluster. Clusters are formed to optimize an objective partitioning criterion, often called a
similarity function, such as distance. Each cluster is represented by a centroid or a representative
cluster. Partitioning methods suffer from the sensibility of initialization. Thus, inappropriate
initialization may lead to bad results. However, they are faster than hierarchical methods.
Density-based clustering methods aim to discover clusters with different shapes. They are based
on the assumption that regions with high density constitute clusters, which are separated by
regions with low density. They are based on the concept of cloud of points with higher density
where the neighborhoods of a point are defined by a threshold of distance or number of nearest
neighbors.
2. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
182
NN are complex systems with interconnected neurons. Unlike hierarchical and partitioning
clustering methods, NN can handle a large number of high dimensional data. Neural Gas is an
artificial NN proposed by [11] which is based on feature vectors to find optimal representations
for the input data. The algorithm’s name refers to the dynamicity of the feature vectors during the
adaptation process.
Based on competitive learning, SOM [6] is the most commonly used NN method. In the training
process, the nodes compete to be the most similar to the input vector node. Euclidean distance is
commonly used to measure distances between input vectors and output nodes’ weights. The node
with the minimum distance is the winner, also known as the Best Matching Unit (BMU). This
latter, is a SOM unit having the closest weight to the current input vector after calculating the
Euclidean distance from each existing weight vector to the chosen input record. Therefore, the
neighbors of the BMU on the map are determined and adjusted. The main function of SOM is to
map the input data from a high dimensional space to a lower dimensional one. It is appropriate for
visualization of high-dimensional data allowing a reduction of data and its complexity. However,
SOM map is insufficient to define the boundaries of each cluster since there is no clear separation
of data items. Thus, extracting partitions from SOM grid is a crucial task. Also, SOM initializes
the topology and the size of the grid where the choice of the size is very sensitive to the
generalization of the method. Hence, we look for an extension of the multi-SOM to overcome
these shortcomings and give the optimal number of clusters without any initialization.
This paper is structured as follows. Section 2 describes different clustering approaches. Section 3
details the multi-SOM approach and the proposed algorithm. Experimental results on real datasets
are given in Section 4. Finally, a conclusion and some future works are given in Section 5.
2. THE MULTI-SOM APPROACH
The multi-SOM method was firstly introduced by [8] for scientific and technical information
analysis specifically for patenting transgenic plant to improve the resistance of plants to pathogen
agents. They proposed an extension of SOM, called multi-SOM, to introduce the notion of
viewpoints into the information analysis with its multiple map visualization and dynamicity. A
viewpoint is defined as a partition of the analyst reasoning. The objects in a partition could be
homogenous or heterogeneous and not necessary similar. However objects in a cluster are similar
and homogenous where a criterion of similarity is inevitably used. Each map in multi-SOM
represents a viewpoint and the information in each map is represented by nodes (classes) and
logical areas (group of classes).
[7] applied multi-SOM on an iconographic database. Iconographic is the collected representation
illustrating a subject which can be an image or a document text. Then, multi-SOM model is
applied in the domain of patent analysis in [10] and [9], where a patent is an official document
conferring a right. The experiments use a database of one thousand patents about oil engineering
technology and indicate the efficiency of viewpoint oriented analysis, where selected viewpoints
correspond to: uses, advantages, patentees and titles subfields of patents.
[12] applied multi-SOM to a zoo data set from the UCI repository to illustrate the technique
combining multiple SOMs which visualizes the different feature maps of the zoo data with color
coded clusters superimposed. The multi-SOM algorithm supplies good map coverage with a
minimal topological defects but it does not facilitate the integration of new data dimensions.
[4] applied the Multi-SOM algorithm for macrophage gene expression analysis. Their proposed
algorithm overcomes some weaknesses of clustering methods which are the cluster number
3. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
183
estimation in partitioning methods and the delimitation of partitions from the output grid of SOM
algorithm. The idea of [3] consists on obtaining compact and well separated clusters using an
evaluation criterion namely Dynamic Validity Index (DVI). The DVI metric is derived from
compactness and separation properties. Thus compactness and separation are two criteria to
evaluate clustering quality and to select the optimal clustering layer. Compactness is assessed by
the intra-distance variability which should be minimized and separation is assessed by the inter-
distance between two clusters which should be maximized. The DVI metric is given by
( ) ( )
{ }
k
InterRatio
k
IntraRatio
DVI K
k
+
= = ..
1
min
Where k denotes the number of activated nodes on the layer and is a modulating parameter.
Figure 1. Architecture of the proposed multi-SOM algorithm
There are other evaluation criteria such as DVI. [2] used the DB index to measure cluster quality
in Multi-Layer Growing SOM algorithm (GSOM) for expression data analysis. The DB index
aims to define the compactness and how well separated are the clusters. GSOM is a model of NN
algorithm which belongs to the hierarchical agglomerative clustering (HAC) algorithms. It is
based on SOM approach but it starts with a minimum number of nodes which is usually four
nodes and grows with a new node at every iteration.
We have chosen to use the DB index because it belongs to the internal criteria which is based on
the compactness and separation of the clusters and well used in many works but not in the multi-
SOM algorithm. The DB index is given by
( ) ( )
( )
+
= ∑
=
≠
j
i
j
i
c
i
j
i
c
c
d
X
d
X
d
c
DB
,
max
1
1
Where c defines the number of clusters, i and j denote the clusters, d(Xi) and d(Xj) are distances
between all objects in clusters i and j to their respective cluster centroids, and d(ci, cj) is the
distance between centroids. Smaller values of DB index indicate better clustering quality.
4. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
184
3. The proposed multi-SOM algorithm
The proposed algorithm (Algorithm1) aims to find the optimal clusters using the DB index as an
evaluation criterion.
Given a dataset, the multi-SOM algorithm first uses the SOM approach to cluster the dataset and
to generate the first SOM grid, namely SOM1 and then the first SOM grid is iteratively clustered
as shown in figure 1. The grid height (Hs) and the grid width (Ws) of SOM1 are given by the user.
The multi-SOM algorithm uses the Batch map function proposed by [3] which is a version of
SOM Kohonen algorithm but it is faster than SOM in the training process. Then, the SOM map is
clustered iteratively from one level to another where the DB index is computed at each level. The
size of the grid decreases at each level until the optimal number of clusters is reached.
In [4], the training process stops when one neuron is reached. However, our proposed algorithm
stops when the DB index gets its minimum value. At this value the optimal number of clusters is
given. Consequently, the proposed algorithm uses less computation time than the one proposed by
[4].
The time complexity of DB is O (DB) = O(C) from the formula of DB index, where C is the
number of clusters. This time complexity is less than the time complexity of the DVI, where O
(DVI) = O (InterRatio + IntraRatio). This integrates many operations to compute the intra- and
inter-distance. Thus, the computation of DVI values at each grid requires more memory space and
time than that of the DB index.
The different steps of the algorithm are as follow:
s: the SOM layer current number
Hs : the SOMs grid height
Ws : the SOMs grid width
Is : the input of SOMs
max_it: the maximum number of iterations for training SOM grids
Algorithm1: multi-SOM
Input: W1, H1, I1, max_it
Output: Optimal cluster number, data partitions
Begin
• Step1: Clustering data by SOM
s = 1;
Batch SOM (W1, H1,I1,max_it) ;
Compute DB index;
s = s+1;
• Step2: Clustering of the SOM and cluster delimitation
HS=HS-1;
WS=WS-1;
repeat
Batch SOM (Ws, Hs, Is, max_it);
Compute DB index on each SOM grid;
s=s+1;
until(DBs<DBs+1);
Return (Data partitions, Optimal cluster number);
End
5. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
185
4. EXPERIMENTAL RESULTS
This section gives the results of the proposed multi-SOM algorithm on five public datasets. Our
algorithm is compared with a partitioning clustering method, namely k-means ([5]), a hierarchical
method, namely BIRCH ([1]) and the algorithm proposed by [4]. The different data sets used in
this work are extracted from the UCI machine learning repositories which are: iris, Pima Indians
diabetes, wine, breast cancer and seeds.
Figure 2. Variation of the number of partitions and the values of DB index with Iris
Figure 3. Variation of the number of partitions and the values of DB index with Wine
Figure 2 and 3 illustrate the variation of the DB index and the number of partitions for Iris and
Wine datasets. We note that the DB index decreases gradually as the number of partitions
decreases from one grid to another to obtain better clustering results until it reaches its lowest
value of 0.161 and 0.202 for Iris and Wine datasets respectively. These values are relative to the
optimal number of clusters which is 3 on Iris and Wine datasets
6. International Journal of Network Security & Its Applications (IJNSA), Vol.5, No.4, July 2013
186
TABLE 1. Evaluation of the proposed multi-SOM algorithm
Class number k-means Birch Ghouila et al.(2008) Multi-SOM
Iris 3 2 3 3 3
Pima 2 2 3 2 2
Breast cancer 2 2 2 2 2
Wine 3 5 4 3 3
Seeds 3 4 2 3 3
For all datasets, results given in Table (1) show that the proposed algorithm is as performing as
that of [4]. On each dataset, the two algorithms succeeded in yielding the real number of clusters.
However, k-means method determines the exact number of clusters only on Pima and breast
datasets. On Iris, Breast and Seeds datasets, Birch algorithm generates the real number of clusters.
With Wine dataset, the number of generated clusters given by the proposed multi-SOM algorithm
is better than those given by k-means and Birch methods.
5. CONCLUSIONS
In this paper, a multi-SOM algorithm based on the DB index to determine the optimal number of
clusters is proposed. In fact, the minimum value generated by DB index refers to the optimal
cluster number. We have shown that our algorithm takes less iterations steps than that of [4].
Thus, the complexity of the proposed multi-SOM algorithm is better than the complexity of [4]
algorithm. As a future work, we will investigate other validity indices and adapt our algorithm to
fuzzy clustering.
REFERENCES
[1] Ester, M., Kriegel, H.-P., Sander, J., and Xu, X, (1996). A density-based algorithm for discovering clusters
in large spatial databases with noise. In Proceedings of 2nd International Conference on KDD,Portland,
Oregon, pp 226–231.
[2] Fonseka, A. and Alahakoon, D, (2010). Exploratory data analysis with multi-layer growing self-organizing
maps. IEEE Xplore, pp 132–137.
[3] Fort J.C. Letrémy P. Cottrell M. (2002) , Advantages and Drawbacks of the Batch Kohonen Algorithm,
Proc. ESANN 2002, Bruges, 2002, M.Verleysen Ed., Editions D Facto, Bruxelles p. 223-230.
[4] Ghouila, A., Yahia, S. B., Malouche, D., Jmel, H., Laouini, D., Z.Guerfali, F., and Abdelhak, S.(2008).
Application of multisom clustering approach to macrophage gene expression analysis. Infection,Genetics
and Evolution, Vol 9,pp 328–329.
[5] J.MacQueen (1967). Some methods for classification and analysis of multivariate observations.
InProceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Vol 1, pp 281–
289.
[6] Kohonen, T. (1981). Automatic formation of topological maps of patterns in a self-organising system. In
Proceedings of the 2SCIA, Scand. Conference on Image Analysis, pp 214–220.
[7] Lamirel, J. C. (2002). Multisom: a multimap extension of the som model. Application to information
discovery in an iconographic context.3:pp 1790–1795.
[8] Lamirel, J.-C., Polanco, X., and Francois, C. (2001). Using artificial neural networks for mapping of
science and technology: A multi self-organizing maps approach. Scientometrics, 51:pp 267–292.
[9] Lamirel, J. C. and Shehabi, S. (2006). Multisom: a multimap extension of the som model. Application to
information discovery in an iconographic context. IEEE CONFERENCE PUBLICATIONS, pp 42–54.
[10] Lamirel, J. C., Shehabi, S., Hoffmann, and Francois, C. (2003). Intelligent patent analysis through the use
of a neural network: experiment of multi-viewpoint analysis with the multisom model. pp 7–23.
[11] Martinetz, T. and Schulten, K. (1991). A neural-gas network learns topologies. In Artificial Neural
Networks.North-Holland, Amsterdam, pp 397–402.
[12] Smith, T. (2009). Adapting to increasing data availability using multi-layered self-organizing maps.IEEE
Computer Society, pp 108–113.
[13] Zhang, T., Ramakrishna, R., and Livny, M. (1996). Birch: An efficient data clustering method for very
large databases. pp 103–114.