SlideShare a Scribd company logo
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
Optimal Feature Selection from VMware ESXi 5.1 
Feature Set 
Amartya Hatua 
Bangalore, India 
Abstract 
A study of VMware ESXi 5.1 server has been carried out to find the optimal set of parameters which 
suggest usage of different resources of the server. Feature selection algorithms have been used to extract 
the optimum set of parameters of the data obtained from VMware ESXi 5.1 server using esxtop command. 
Multiple virtual machines (VMs) are running in the mentioned server. K-means algorithm is used for 
clustering the VMs. The goodness of each cluster is determined by Davies Bouldin index and Dunn index 
respectively. The best cluster is further identified by the determined indices. The features of the best cluster 
are considered into a set of optimal parameters. 
Index Terms 
Clustering, Feature selection, K-means clustering algorithm, VMware ESXi 5.1 
1.INTRODUCTION 
Feature selection or dimensionality reduction problem for supervised or unsupervised inductive 
learning basically generates a set of candidate features. These set of features can be defined mainly 
in three different approaches, a) the number or size of features is specified and they optimize an 
evaluation measure. b) The subset of features must satisfy some predefined restriction on 
evaluation measure. c) The subset with the best size and value measured by evaluation measure. 
Improvement of inductive learning is the prime objective of feature selection. The improvement 
can be measured in terms of learning speed, ease of the presentation or ability of generalization. To 
improve the results of inducer, the associated key factors are diminishing the volume of storage, 
reduction of the noise generated by irrelevant or redundant features and elimination of useless 
knowledge. 
The motivation behind the work comes from selection procedure of VMs during VM migration 
process or load balancing of VMware ESXi 5.1 server. If a server is overloaded with VMs then 
some of the VMs should be migrated to other sever. The selection of VMs happens depending on 
the load of different resources. For every resource there are many parameters or features are 
present in VMware ESXi 5.1 server. The set of all parameters and their values can be observed 
using ESXTOP command. Selection of the optimal set of parameters has been carried out in this 
study. Higher value for particular VM for a set of parameters indicates high usage or load of the 
particular resource. Depending on the types of load of different VMs; the selection of VMs can be 
done, which are eligible for migration. 
DOI:10.5121/ijccms.2014.3301 1
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
In the following sections Feature selection methods, Clustering methods, Cluster analysis methods, 
Experiment methodology, Result and Conclusion are elaborately discussed. 
2 
II.FEATURE SELECTION METHODS 
A feature selection algorithm (FSA) is a computational solution that is defined by a certain degree 
of relevance. However the relevance of a feature always depends on the objective of the inductive 
learning algorithm which uses the reduce set of feature. For induction an irrelevant feature is not 
useful at the same time all relevant features are not required always [1]. Some of the feature 
selection methods are: Wrapper method, Filter methods, Rough set methods and Schemata search 
method. Feature selection using similarity, Principal Component Analysis method, Filter-Wrapper 
Hybrid method. Again different types filtering method are also available for feature selection, like 
FOCUS algorithm, RELIEF algorithm, Las Vegas algorithm, Markov Blanket filtering, Decision 
Tree method and Genetic algorithms [2], [3], [4]. 
A.Correlation-based Feature Selection method 
Correlation-based Feature Selection (CFS) [5] comes under the category of filter algorithm. In this 
type of algorithm the subset of features are chosen based on ranks. The rank is determined based 
on a correlation based heuristic evaluation function. The evaluation function selects features which 
are highly correlated with the class and at the same time uncorrelated with each other. The 
evaluation function selects features which are highly correlated with the class and at the same time 
uncorrelated with each other. Selection of feature happens mainly based on the scale to which it 
predicts classes in the instance space not predicted by some other features. CFSs feature subset 
evaluation function is repeated here (with slightly modified notation) for ease of reference: 
is representing the heuristic merit of a feature subset S containing k features. is denoting 
the mean feature-class correlation (f ∈ s), and is stands for average feature-feature inter 
correlation. Numerator of the equation is giving an idea regarding the prediction of the class. The 
denominator is representing redundancy among the features. CFS uses best heuristic search 
strategy. If continuous five times fully expanded subsets do not give a better result the process 
stops; in other words that is the stopping condition of the process. 
B.Wrapper method 
The wrapper methodology, popularized by Kohavi and John (1997), provides an efficient way to 
determine the required subset of features, regardless of the machine learning algorithm. Here the 
machine learning algorithm is considered as a perfect black box. In practice, one needs to define: 
(i) the procedure to find the best subset among all possible variable subsets; (ii) the procedure to 
evaluate the prediction performance of a learning machine, which helps to guide the search and 
reaches the stopping condition; and (iii) the selection of predictor. Wrapper method performs a 
search in the space of all possible subsets. 
Use of machine learning algorithms makes wrapper method remarkably universal and simple. On 
the other hand when the number of features are very large then this method required an exhaustive 
search for O( ) number of subsets. Where n is the number of features. The complexity of the 
algorithm is O( ). So the algorithms is not efficient when n is very large.
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
3 
C.Filter Algorithm 
Filter algorithms determine the subset of optimal features by measuring the intrinsic properties of 
the data. The algorithm calculates a relevance score for each feature. Features having low 
relevance score are removed. Afterwards, the subset of relevant features is used as input in 
different machine learning algorithms. 
RELIEF Algorithm: The RELIEF assigns different weight to every feature depending on the 
relevance of the particular feature on the given context. RELIEF algorithm follows the filter 
method to determine the minimal set of features. It samples instances from the training data 
randomly and finds the relevance value based on near-hit and near-miss [6]. 
D.Rough Set Method 
In Velayutham [7] Rough Set Theory (RST) is used as a tool to discover data dependencies and to 
find the subset of relevant features. 
The rough set is approximation of set by two concepts known as lower and upper approximations. 
The domain objects are represented by the lower approximation. These domain objects belong to 
the subset of interest with a certainty. On the other hand the upper approximation describes the 
objects which are possibly belonging to subset. Approximations are made with respect to a 
particular subset of features or parameters. 
Considering an information system as I = (U, A d). In the information system U is the universe 
with a non-empty set of finite objects. On the other hand A represents nonempty finite set of 
condition attributes, and d represents the decision attribute, the corresponding function is 
, where represents set of values of a. If then there is an associated 
equivalence relation: 
The partition of U generated by is denoted by U/P. If (x, y) IND(P) holds good, then it is 
difficult to find a clearly distinguish x and y are by attributes from P. The equivalence classes of 
the P in discernibility relation are denoted [x]p. Considering XU the P-lower approximation PX of 
set X and P-upper approximation PX of set X can be defined as: 
PX = {X ∈ U|[X]p X } 
X = {X ∈ U|[X]p X } 
Let P,Q A be equivalence relations over U , then positive region defined as: 
PX. The positive region of the partition U/Q with respect to P, is 
the set of all objects of U that can be certainly classified to blocks of the partition . U/Q by means 
of P. Q depends on P in a degree denoted 
Where 
If k=1, Q depends totally on P, if depends partially on P, if k=0 then Q does not 
depend on P.
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
The Unsupervised Quick Reduct (USQR) algorithm endeavors to calculate a reduct without fully 
generating all possible subsets. In Velayutham [8] the algorithm starts off with an empty set. In 
each iteration one attribute is added in the set. Those attributes are selected depending on the rough 
set dependency metric. This procedure continues until the value reaches its maximum value for 
that dataset. The mean dependency of each attribute is calculated and best candidate is chosen: 
4 
The data set is considered as consistent if the process ends at dependency value 1; otherwise it is 
inconsistent. 
A.Chi-square feature selection 
Chi-square method calculates the lack of information between a feature and a category. Chi-square 
distribution is used for comparison with one degree of freedom to judge extremeness [9]. It is 
defined as: 
Here feature selection metrics (four known measures and two proposed variants), which are 
functions of the following four dependency tuples: 
( ): presence of t and membership in . 
( ): presence of t and non-membership in . 
( ): absence of and membership in . 
( ): absence of t and non-membership in . 
III.CLUSTERING METHOD 
K-means clustering [9] is a method of cluster creation, where the main objective is to partition n 
observations into K groups/clusters in each instance belongs to the cluster with nearest mean. This 
results in a partitioning of the data space into Voronoi cells. Mathematically Voronoi diagram 
helps to divide a space into a number of regions. Depending on some predefined conditions set of 
point is taken in consideration they are known as seeds, sites, or generators. Around each site 
several points which are closet to that site form a region. The regions are called Voronoi cells. The 
most popular algorithm uses an iterative technique to reach the stopping point. It is commonly 
famous as K-means algorithm. In computer science community it is also known as Lloyd’s 
algorithm. 
IV.CLUSTER ANALYSIS METHOD 
The cluster analysis methods provide the measure, using which the goodness clusters can be 
determined. The cluster analysis methods has been used are Davies Bouldin index and 
Dunn index (J. C. Dunn 1974).
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
5 
A.DaviesBouldin index 
The Davies Bouldin index can be calculated by the following formula: 
In the above equation n represents the number of clusters. Cx stands for centroid of the cluster x. 
is representing the average distance between each of the point in the cluster and the centroid. 
is denoting the distance between two centroid of two clusters. 
The clustering algorithm produces a collection of clusters having smallest Davies Bouldin index 
is considered to be the best algorithm. Clusters with smallest Davies Bouldin index suggests low 
intra-cluster distances and high intra-cluster similarity. 
B.Dunn index 
The main objective of Dunn index is to find dense and well-separated clusters. It is defined as the 
ratio between the minimal inter- cluster distance to maximal intra-cluster distance. The following 
formula is used to calculate Dunn index for each cluster partition. 
If i and j are two clusters then distance between these two cluster is represented by d(i,j). For a 
cluster K, d(k) represents 
intra-cluster distance. The distance between centroids of the clusters can be considered as inter-cluster 
distance represented by: d(i,j). Similarly intra-cluster distance d(k) can be measured as 
distance between any two elements of cluster K. As clusters with high intra-cluster similarity and 
low inter-cluster similarity are the desirable that is why cluster with high Dunn index are 
preferable. 
V.EXPERIMENT METHODOLOGY 
The steps of the experiment are following: 
Environment setup: Firstly in a physical server VMware ESXi 5.1 has been installed and twenty 
six virtual machines (VM) have been installed in that. 
Load selection and run the load: There are four different types (CPU intensive, Network 
intensive, Memory intensive and Disk intensive) of load have been run. Each of CPU, 
Network and Disk intensive loads has been run in seven different VMs, whereas Network 
intensive load has been run 
in rest of the five VMs. 
Collection of result set: In third step data has been collected from the server using ESXTOP reply 
mode. Performance of different factor has been collected like CPU, Memory, Disk, Network and 
Power.
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
Feature selection: In the result set there are many parameters in this phase the relevant parameters 
are selected using feature selection algorithms. 
6 
Use K-means algorithm: By using the clustering algorithms clusters has been created using the 
subset of features resulting from the feature selection algorithms. 
Comparison: In this step the clusters has been compared based on some cluster comparison 
method. 
TABLE I. LIST OF FEATURES FOR DIFFERENT RESOURCES 
Resources Parameters 
CPU 
%USED, %RUN, %SYS, %WAIT, %VMWAIT, %RDY, 
%IDLE, %OVRLP, %CSTP, %MLMTD, %SWPWT, 
SWTCH/s, MIG/s 
Memory 
SWCUR, SWTGT, SWR/s, SWW/s, LLSWR/s, LLSWW/s, 
CPTDR, CPTTGT, ZERO, SHRD,SHRDSVD, COWH, 
NHN, 
NMIG, NRMEM, NLMEM, N_L, GST_ND0, 
OVD_ND0,GST_ ND1, OVD_ND1, OVHDUM, OVHD, 
OVHDMAX, CMTTGT, CMTTGT, CMTCHRG, 
CMTPPS, CACHESZ, CACHUSD, ZIP/s, UNZIP/s, 
MEMSZ, GRANT, SZTGT, TCHD, TCHD W, ACTV, 
ACTVS, ACTVF, ACTVN, MCTLSZ, MCTLTGT, 
MCTLMAX 
Disk 
CMDS/s, READS/s, WRITES/s, MBREAD/s, 
MBWRTN/s, LAT/rd, LAT/wr 
Network 
PKTTX/s, MbTX/s, PKTRX/s, MbRX/s, %DRPTX, 
%DRPRX, ACTN/s, PKTTXMUL/s, PKTRXMUL/s, 
PKTTXBRD/s, PKTRXBRD/s 
Power 
%USED, %UTIL, %C0, %C1, %C2, %C3, %T0, %T1, 
%T2, %T3, %T4, %T5, %T6,%T7 
VI.RESULTS 
A.Features selected by using different feature selection algorithms 
There are five different feature selection algorithms have been used: CFS method, RELIEF 
method, Chi Square Method, Wrapper method, Rough set method. The feature selection methods 
have been used on the data generated from the experiment. The methods select a subset of relevant 
features depending on some measuring metrics defined by the method. 
TABLE I .LIST OF SELECTED FEATURES FOR DIFFERENT RESOURCES 
Resource CFS RELIEF Chi 
Square 
Wrapper Rough 
Set 
CPU %USED, 
%IDLE, 
MIG/s 
%USED, 
%RUN, 
%SYS, 
%VMWAI 
T,%RDY, 
%IDLE, 
%OVRLP, 
%SWPWT 
MIG/s 
%USED, 
%RUN, 
%SYS, 
%WAIT, 
%VMWAI 
T,%RDY, 
%IDLE, 
%OVRLP, 
%CSTP, 
%MLMT, 
%USED, 
%RUN, 
%SYS, 
%WAIT, 
%VMWA 
IT,%RD, 
%IDLE, 
%OVRLP, 
%CSTP, 
%MLMD, 
%USD 
%SYS, 
MIG/s
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
7 
%SWPW, 
SWTCH/s, 
MIG/s 
%SWPT, 
SWTCH/, 
MIG/s 
Disk CMDS/s, 
LAT/r22 
2d 
CMDS/s, 
READS/s, 
WRITES/, 
MBREAD/ 
s,LAT/rd, 
LAT/wr 
CDMS/s, 
READS/s, 
WRITES/, 
MBREAD/ 
s,MBWRT 
N/s,LAT/r 
d,LAT/wr 
CMDS/s, 
MBREAD 
/s,LAT/rd 
CMDS, 
LAT/rd, 
LAT/wr 
Network PKTRXs, 
MbRX/s 
PKTTX/s, 
MbTX/s, 
PKTRX/s, 
MbRX/s, 
ACTN/s, 
PKTRXM 
UL/ 
PKTTX/s, 
MbTX/s, 
PKTRX/s, 
MbRX/s 
PKTTX/s, 
MbTX/s, 
PKTRX/s, 
ACTN/s, 
PKTRXM 
UL/ 
PKTR/, 
ACTN/ 
s, 
PKTRX 
,MUL/ 
s,PKTR 
XBRD/ 
s 
Memory SWCUR, 
COWH, 
OVHD, 
GRANT, 
SZTGT 
SZTGT, 
GRANT, 
MCTLMA 
X, 
TCHD W, 
MEMSZ, 
TCHD, 
SHRDSD, 
ACTVS, 
MCTLSZ, 
SHRD, 
ZERO, 
SWCUR, 
MCTLTT, 
ACTVF, 
ACTV, 
CMTTGT, 
ACTVN, 
OVHDU, 
OVHDM, 
OVHD, 
COWH, 
CACHUS 
D,NLME 
M,NHN, 
NRMEM, 
N L, 
CMTPPS, 
CACHESZ 
SWCUR, 
ZERO, 
SHRD, 
SHRDSV 
D, 
COWH, 
NHN, 
NRMEM, 
NLMEM, 
N L, 
OVHDUM 
,OVHD, 
OVHDMA 
X, 
CMTTGT, 
CACHESZ 
,CACHUS 
D,MEMSZ 
,GRANT, 
SZTGT, 
TCHD, 
TCHD W, 
ACTV, 
ACTVS, 
ACTVF, 
ACTVN, 
MCTLMA 
X 
GRANT, 
SZTGT, 
SWCUR, 
SHRDSV 
D,SHRD, 
ZERO, 
ACTVS, 
TCHD W, 
ACTVF, 
TCHD, 
ACTV, 
CMTTGT, 
ACTVN, 
OVHD, 
MEMSZ, 
MCTLM 
AX, 
MCTLTG 
T, 
MCTLSZ, 
OVHDM 
AX, 
OVHDU 
M, 
COWH, 
CACHUS 
D 
SHRDS 
VD, 
GRAN 
T 
Power %USED %USED, 
%C1, 
%USED 
%USED, 
%C0, 
%C1 
%USED, 
%C0, 
%C1 
%USE 
D,%UT 
IL,%C1 
B.Clustering using K-means algorithm 
After getting the selected features K-means clustering algorithm has been applied on each of the 
set. Where each setconsists of all the selected features of all resources by a particular method of 
feature selection. Here number of clusteris four. To find the goodness of the clusters, the analysis 
has been done by the help of Davies Bouldin index and Dunn index. 
Using these two different indices the goodness of the clustering algorithm has been found out. The 
cluster having maximum Davies Bouldin index and minimum Dunn index is considered to be the 
best cluster and the feature selection algorithm related to that cluster is considered to be the best 
algorithm among these algorithms using the data set. From this result it can be concluded that the 
parameters selected are optimum parameters from which we can have a clear idea about types of
International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 
load running in VMware ESXi 5.1 server. Here is the table where Davies Bouldin index and Dunn 
index for different clusters have been given. Where clusters found using the parameters selected by 
the respective feature selection algorithms. 
8 
TABLE I .DAVIES BOULDIN INDEX AND DUNN INDEX FOR CLUSTERS. 
Feature selection 
technique 
Davies Bouldin 
index 
Dunn index 
CFS 9.9122e+004 1.4196e-006 
RELIEF 2.2679e+00 1.0874e-009 
ChiSquare Method 2.9018e+007 2.442-009 
Wrapper 2.1199e+007 1.0862e-009 
Rough set 3.9079e+005 4.9040e-006 
From the result it can be observed that the best result in both of the case is given by the clusters 
formed by the features selected by the CFS algorithm. 
VII.CONCLUSION 
So from the result it can be concluded that the set of parameters selected by the CFS is the best set 
of parameters of VMware ESXi 5.1. So from that set of parameters we can decide the type of load 
in VMware ESXi 5.1. At the same time VMs having high values of those parameters suggest the 
higher usage of those resources by the VM, which can be helpful for load balancing of Virtual 
Machine. 
REFERENCES 
[1] How Useful is Relevance? R.A. Caruana and D. Freitag, Technical report, Fall94 AAAI Symposium 
on Relevance, New Orleans, 1994, AAAI. 
[2] Selection of Relevant Features and Examples in Machine Learning, A.L.Blum and P.Langley, R. 
Greiner and D. Subramanian, eds., Artificial Intelligence on Relevance, volume 97, pages 245271. 
Artificial Intelligence, 1997. 
[3] An Evaluation of Feature Selection Methods and their Application to Computer Security, J.Doak, An 
Evaluation of Feature Selection Methods and their Application to Computer Security, 245271, 1992, 
Technical Report CSE9218, Davis, CA: University of California, Department of Computer Science. 
[4] Feature Selection for Knowledge Discovery and Data Mining, H.Liu and H.Motoda, 1998, Kluwer 
Academic Publishers, London, GB.[5]Mark A.Hall, Feature Selection for Discrete and Numeric Class 
Machine Learning. 
[6] D. W. Yijun Sun, A relief based feature extraction algorithm, Journal of the Royal Statistical Society. 
Series B (Methodological) Vol. 39, No. 1(1977), pp. 1-38, Tech. Rep., 2011. 
[7] A Novel Entropy Based Unsupervised Feature Selection Algorithm Using Rough Set Theory, K. 
Thangavel and C.Velayutham, Science And Management (ICAESM -2012) March 30, 31, 2012, 2012, 
IEEE International Conference On Advances In Engineering. 
[8] Feature selection for text categorization on imbalanced data, M. A. Hall,Sigkdd Explorations ,Volume 
6, Issue 1 - Page 80, Tech. Rep., 2003.K. Elissa, “Title of paper if known,” unpublished. 
[9] Extensions to the k-Means Algorithm for Clustering Large Data Setswith Categorical Values, 
ZHEXUE HUANG, P1: SUD Data Mining and Knowledge Discovery KL657-03-Huang 12:59 Data 
Mining and Knowledge Discovery 2, 283304 (1998) 1998, October 27, 1998, Kluwer Academic 
Publishers. Manufactured in The Netherlands.

More Related Content

What's hot

An introduction to variable and feature selection
An introduction to variable and feature selectionAn introduction to variable and feature selection
An introduction to variable and feature selection
Marco Meoni
 
Fuzzy inference systems
Fuzzy inference systemsFuzzy inference systems
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...
IRJET -  	  Movie Genre Prediction from Plot Summaries by Comparing Various C...IRJET -  	  Movie Genre Prediction from Plot Summaries by Comparing Various C...
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...
IRJET Journal
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
IEEEFINALYEARPROJECTS
 
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ijcsit
 
Classification using L1-Penalized Logistic Regression
Classification using L1-Penalized Logistic RegressionClassification using L1-Penalized Logistic Regression
Classification using L1-Penalized Logistic Regression
Setia Pramana
 
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
IAEME Publication
 
Feature selection and microarray data
Feature selection and microarray dataFeature selection and microarray data
Feature selection and microarray data
Gianluca Bontempi
 
Hc3413121317
Hc3413121317Hc3413121317
Hc3413121317
IJERA Editor
 
Numerical analysis genetic algorithms
Numerical analysis  genetic algorithmsNumerical analysis  genetic algorithms
Numerical analysis genetic algorithms
SHAMJITH KM
 
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
Xin-She Yang
 
Ica 2013021816274759
Ica 2013021816274759Ica 2013021816274759
Ica 2013021816274759
Valentino Selayan
 
New feature selection based on kernel
New feature selection based on kernelNew feature selection based on kernel
New feature selection based on kernel
journalBEEI
 
Implicit and explicit sequence control with exception handling
Implicit and explicit sequence control with exception handlingImplicit and explicit sequence control with exception handling
Implicit and explicit sequence control with exception handling
VIKASH MAINANWAL
 
Feature selection using modified particle swarm optimisation for face recogni...
Feature selection using modified particle swarm optimisation for face recogni...Feature selection using modified particle swarm optimisation for face recogni...
Feature selection using modified particle swarm optimisation for face recogni...
eSAT Journals
 
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATIONAN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
ecij
 

What's hot (16)

An introduction to variable and feature selection
An introduction to variable and feature selectionAn introduction to variable and feature selection
An introduction to variable and feature selection
 
Fuzzy inference systems
Fuzzy inference systemsFuzzy inference systems
Fuzzy inference systems
 
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...
IRJET -  	  Movie Genre Prediction from Plot Summaries by Comparing Various C...IRJET -  	  Movie Genre Prediction from Plot Summaries by Comparing Various C...
IRJET - Movie Genre Prediction from Plot Summaries by Comparing Various C...
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
ON FEATURE SELECTION ALGORITHMS AND FEATURE SELECTION STABILITY MEASURES: A C...
 
Classification using L1-Penalized Logistic Regression
Classification using L1-Penalized Logistic RegressionClassification using L1-Penalized Logistic Regression
Classification using L1-Penalized Logistic Regression
 
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
PARTICLE SWARM OPTIMIZATION FOR MULTIDIMENSIONAL CLUSTERING OF NATURAL LANGUA...
 
Feature selection and microarray data
Feature selection and microarray dataFeature selection and microarray data
Feature selection and microarray data
 
Hc3413121317
Hc3413121317Hc3413121317
Hc3413121317
 
Numerical analysis genetic algorithms
Numerical analysis  genetic algorithmsNumerical analysis  genetic algorithms
Numerical analysis genetic algorithms
 
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
Applications and Analysis of Bio-Inspired Eagle Strategy for Engineering Opti...
 
Ica 2013021816274759
Ica 2013021816274759Ica 2013021816274759
Ica 2013021816274759
 
New feature selection based on kernel
New feature selection based on kernelNew feature selection based on kernel
New feature selection based on kernel
 
Implicit and explicit sequence control with exception handling
Implicit and explicit sequence control with exception handlingImplicit and explicit sequence control with exception handling
Implicit and explicit sequence control with exception handling
 
Feature selection using modified particle swarm optimisation for face recogni...
Feature selection using modified particle swarm optimisation for face recogni...Feature selection using modified particle swarm optimisation for face recogni...
Feature selection using modified particle swarm optimisation for face recogni...
 
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATIONAN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
AN ENVIRONMENT FOR NON-DUPLICATE TEST GENERATION FOR WEB BASED APPLICATION
 

Viewers also liked

OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
ijccmsjournal
 
Iss3
Iss3Iss3
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
ijccmsjournal
 
FED GOV CON - Grant Writing & SBIR's
FED GOV CON -  Grant Writing & SBIR'sFED GOV CON -  Grant Writing & SBIR's
FED GOV CON - Grant Writing & SBIR's
JSchaus & Associates
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
ijccmsjournal
 
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern FoundryFED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
JSchaus & Associates
 
Synchronization and inverse synchronization of some different dimensional dis...
Synchronization and inverse synchronization of some different dimensional dis...Synchronization and inverse synchronization of some different dimensional dis...
Synchronization and inverse synchronization of some different dimensional dis...
ijccmsjournal
 

Viewers also liked (7)

OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
OPTIMIZATION OF ARTIFICIAL NEURAL NETWORK MODEL FOR IMPROVEMENT OF ARTIFICIAL...
 
Iss3
Iss3Iss3
Iss3
 
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
Synchronization and Inverse Synchronization of Some Different Dimensional Dis...
 
FED GOV CON - Grant Writing & SBIR's
FED GOV CON -  Grant Writing & SBIR'sFED GOV CON -  Grant Writing & SBIR's
FED GOV CON - Grant Writing & SBIR's
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
 
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern FoundryFED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
FED GOV CON - Govt Contracting GSA SCHEDULES - Eastern Foundry
 
Synchronization and inverse synchronization of some different dimensional dis...
Synchronization and inverse synchronization of some different dimensional dis...Synchronization and inverse synchronization of some different dimensional dis...
Synchronization and inverse synchronization of some different dimensional dis...
 

Similar to Optimal feature selection from v mware esxi 5.1 feature set

Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
ijccmsjournal
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1  Feature SetOptimal Feature Selection from VMware ESXi 5.1  Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
ijccmsjournal
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
ijccmsjournal
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
IJERA Editor
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
AIRCC Publishing Corporation
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 On Feature Selection Algorithms and Feature Selection Stability Measures : A... On Feature Selection Algorithms and Feature Selection Stability Measures : A...
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
AIRCC Publishing Corporation
 
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
IEEEMEMTECHSTUDENTSPROJECTS
 
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
IEEEFINALYEARSTUDENTPROJECTS
 
A Threshold Fuzzy Entropy Based Feature Selection: Comparative Study
A Threshold Fuzzy Entropy Based Feature Selection:  Comparative StudyA Threshold Fuzzy Entropy Based Feature Selection:  Comparative Study
A Threshold Fuzzy Entropy Based Feature Selection: Comparative Study
IJMER
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
JPINFOTECH JAYAPRAKASH
 
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
IEEEGLOBALSOFTTECHNOLOGIES
 
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
IEEEGLOBALSOFTTECHNOLOGIES
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
IEEEFINALYEARPROJECTS
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
IEEEGLOBALSOFTTECHNOLOGIES
 
Machine learning Mind Map
Machine learning Mind MapMachine learning Mind Map
Machine learning Mind Map
Ashish Patel
 
Iaetsd an enhanced feature selection for
Iaetsd an enhanced feature selection forIaetsd an enhanced feature selection for
Iaetsd an enhanced feature selection for
Iaetsd Iaetsd
 
Performance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning AlgorithmsPerformance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning Algorithms
Dinusha Dilanka
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
IEEEGLOBALSOFTTECHNOLOGIES
 
Machine Learning.pdf
Machine Learning.pdfMachine Learning.pdf
Machine Learning.pdf
BeyaNasr1
 
Evaluation of a hybrid method for constructing multiple SVM kernels
Evaluation of a hybrid method for constructing multiple SVM kernelsEvaluation of a hybrid method for constructing multiple SVM kernels
Evaluation of a hybrid method for constructing multiple SVM kernels
infopapers
 

Similar to Optimal feature selection from v mware esxi 5.1 feature set (20)

Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1  Feature SetOptimal Feature Selection from VMware ESXi 5.1  Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
 
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature SetOptimal Feature Selection from VMware ESXi 5.1 Feature Set
Optimal Feature Selection from VMware ESXi 5.1 Feature Set
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
On Feature Selection Algorithms and Feature Selection Stability Measures : A ...
 
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 On Feature Selection Algorithms and Feature Selection Stability Measures : A... On Feature Selection Algorithms and Feature Selection Stability Measures : A...
On Feature Selection Algorithms and Feature Selection Stability Measures : A...
 
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
2014 IEEE JAVA DATA MINING PROJECT A fast clustering based feature subset sel...
 
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
IEEE 2014 JAVA DATA MINING PROJECTS A fast clustering based feature subset se...
 
A Threshold Fuzzy Entropy Based Feature Selection: Comparative Study
A Threshold Fuzzy Entropy Based Feature Selection:  Comparative StudyA Threshold Fuzzy Entropy Based Feature Selection:  Comparative Study
A Threshold Fuzzy Entropy Based Feature Selection: Comparative Study
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
JAVA 2013 IEEE DATAMINING PROJECT A fast clustering based feature subset sele...
 
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
JAVA 2013 IEEE PROJECT A fast clustering based feature subset selection algor...
 
A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...A fast clustering based feature subset selection algorithm for high-dimension...
A fast clustering based feature subset selection algorithm for high-dimension...
 
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
JAVA 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subset ...
 
Machine learning Mind Map
Machine learning Mind MapMachine learning Mind Map
Machine learning Mind Map
 
Iaetsd an enhanced feature selection for
Iaetsd an enhanced feature selection forIaetsd an enhanced feature selection for
Iaetsd an enhanced feature selection for
 
Performance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning AlgorithmsPerformance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning Algorithms
 
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
DOTNET 2013 IEEE CLOUDCOMPUTING PROJECT A fast clustering based feature subse...
 
Machine Learning.pdf
Machine Learning.pdfMachine Learning.pdf
Machine Learning.pdf
 
Evaluation of a hybrid method for constructing multiple SVM kernels
Evaluation of a hybrid method for constructing multiple SVM kernelsEvaluation of a hybrid method for constructing multiple SVM kernels
Evaluation of a hybrid method for constructing multiple SVM kernels
 

Recently uploaded

Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Soumen Santra
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
obonagu
 
Series of visio cisco devices Cisco_Icons.ppt
Series of visio cisco devices Cisco_Icons.pptSeries of visio cisco devices Cisco_Icons.ppt
Series of visio cisco devices Cisco_Icons.ppt
PauloRodrigues104553
 
New techniques for characterising damage in rock slopes.pdf
New techniques for characterising damage in rock slopes.pdfNew techniques for characterising damage in rock slopes.pdf
New techniques for characterising damage in rock slopes.pdf
wisnuprabawa3
 
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesHarnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Christina Lin
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
Madan Karki
 
Literature Review Basics and Understanding Reference Management.pptx
Literature Review Basics and Understanding Reference Management.pptxLiterature Review Basics and Understanding Reference Management.pptx
Literature Review Basics and Understanding Reference Management.pptx
Dr Ramhari Poudyal
 
introduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdfintroduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdf
ravindarpurohit26
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
Ratnakar Mikkili
 
ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024
Rahul
 
Generative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of contentGenerative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of content
Hitesh Mohapatra
 
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
awadeshbabu
 
Low power architecture of logic gates using adiabatic techniques
Low power architecture of logic gates using adiabatic techniquesLow power architecture of logic gates using adiabatic techniques
Low power architecture of logic gates using adiabatic techniques
nooriasukmaningtyas
 
DfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributionsDfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributions
gestioneergodomus
 
PPT on GRP pipes manufacturing and testing
PPT on GRP pipes manufacturing and testingPPT on GRP pipes manufacturing and testing
PPT on GRP pipes manufacturing and testing
anoopmanoharan2
 
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
obonagu
 
Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
Aditya Rajan Patra
 
Swimming pool mechanical components design.pptx
Swimming pool  mechanical components design.pptxSwimming pool  mechanical components design.pptx
Swimming pool mechanical components design.pptx
yokeleetan1
 
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
thanhdowork
 
Modelagem de um CSTR com reação endotermica.pdf
Modelagem de um CSTR com reação endotermica.pdfModelagem de um CSTR com reação endotermica.pdf
Modelagem de um CSTR com reação endotermica.pdf
camseq
 

Recently uploaded (20)

Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTSHeap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
Heap Sort (SS).ppt FOR ENGINEERING GRADUATES, BCA, MCA, MTECH, BSC STUDENTS
 
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
在线办理(ANU毕业证书)澳洲国立大学毕业证录取通知书一模一样
 
Series of visio cisco devices Cisco_Icons.ppt
Series of visio cisco devices Cisco_Icons.pptSeries of visio cisco devices Cisco_Icons.ppt
Series of visio cisco devices Cisco_Icons.ppt
 
New techniques for characterising damage in rock slopes.pdf
New techniques for characterising damage in rock slopes.pdfNew techniques for characterising damage in rock slopes.pdf
New techniques for characterising damage in rock slopes.pdf
 
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming PipelinesHarnessing WebAssembly for Real-time Stateless Streaming Pipelines
Harnessing WebAssembly for Real-time Stateless Streaming Pipelines
 
spirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptxspirit beverages ppt without graphics.pptx
spirit beverages ppt without graphics.pptx
 
Literature Review Basics and Understanding Reference Management.pptx
Literature Review Basics and Understanding Reference Management.pptxLiterature Review Basics and Understanding Reference Management.pptx
Literature Review Basics and Understanding Reference Management.pptx
 
introduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdfintroduction to solar energy for engineering.pdf
introduction to solar energy for engineering.pdf
 
Exception Handling notes in java exception
Exception Handling notes in java exceptionException Handling notes in java exception
Exception Handling notes in java exception
 
ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024ACEP Magazine edition 4th launched on 05.06.2024
ACEP Magazine edition 4th launched on 05.06.2024
 
Generative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of contentGenerative AI leverages algorithms to create various forms of content
Generative AI leverages algorithms to create various forms of content
 
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
[JPP-1] - (JEE 3.0) - Kinematics 1D - 14th May..pdf
 
Low power architecture of logic gates using adiabatic techniques
Low power architecture of logic gates using adiabatic techniquesLow power architecture of logic gates using adiabatic techniques
Low power architecture of logic gates using adiabatic techniques
 
DfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributionsDfMAy 2024 - key insights and contributions
DfMAy 2024 - key insights and contributions
 
PPT on GRP pipes manufacturing and testing
PPT on GRP pipes manufacturing and testingPPT on GRP pipes manufacturing and testing
PPT on GRP pipes manufacturing and testing
 
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
原版制作(unimelb毕业证书)墨尔本大学毕业证Offer一模一样
 
Recycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part IIIRecycled Concrete Aggregate in Construction Part III
Recycled Concrete Aggregate in Construction Part III
 
Swimming pool mechanical components design.pptx
Swimming pool  mechanical components design.pptxSwimming pool  mechanical components design.pptx
Swimming pool mechanical components design.pptx
 
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
RAT: Retrieval Augmented Thoughts Elicit Context-Aware Reasoning in Long-Hori...
 
Modelagem de um CSTR com reação endotermica.pdf
Modelagem de um CSTR com reação endotermica.pdfModelagem de um CSTR com reação endotermica.pdf
Modelagem de um CSTR com reação endotermica.pdf
 

Optimal feature selection from v mware esxi 5.1 feature set

  • 1. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 Optimal Feature Selection from VMware ESXi 5.1 Feature Set Amartya Hatua Bangalore, India Abstract A study of VMware ESXi 5.1 server has been carried out to find the optimal set of parameters which suggest usage of different resources of the server. Feature selection algorithms have been used to extract the optimum set of parameters of the data obtained from VMware ESXi 5.1 server using esxtop command. Multiple virtual machines (VMs) are running in the mentioned server. K-means algorithm is used for clustering the VMs. The goodness of each cluster is determined by Davies Bouldin index and Dunn index respectively. The best cluster is further identified by the determined indices. The features of the best cluster are considered into a set of optimal parameters. Index Terms Clustering, Feature selection, K-means clustering algorithm, VMware ESXi 5.1 1.INTRODUCTION Feature selection or dimensionality reduction problem for supervised or unsupervised inductive learning basically generates a set of candidate features. These set of features can be defined mainly in three different approaches, a) the number or size of features is specified and they optimize an evaluation measure. b) The subset of features must satisfy some predefined restriction on evaluation measure. c) The subset with the best size and value measured by evaluation measure. Improvement of inductive learning is the prime objective of feature selection. The improvement can be measured in terms of learning speed, ease of the presentation or ability of generalization. To improve the results of inducer, the associated key factors are diminishing the volume of storage, reduction of the noise generated by irrelevant or redundant features and elimination of useless knowledge. The motivation behind the work comes from selection procedure of VMs during VM migration process or load balancing of VMware ESXi 5.1 server. If a server is overloaded with VMs then some of the VMs should be migrated to other sever. The selection of VMs happens depending on the load of different resources. For every resource there are many parameters or features are present in VMware ESXi 5.1 server. The set of all parameters and their values can be observed using ESXTOP command. Selection of the optimal set of parameters has been carried out in this study. Higher value for particular VM for a set of parameters indicates high usage or load of the particular resource. Depending on the types of load of different VMs; the selection of VMs can be done, which are eligible for migration. DOI:10.5121/ijccms.2014.3301 1
  • 2. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 In the following sections Feature selection methods, Clustering methods, Cluster analysis methods, Experiment methodology, Result and Conclusion are elaborately discussed. 2 II.FEATURE SELECTION METHODS A feature selection algorithm (FSA) is a computational solution that is defined by a certain degree of relevance. However the relevance of a feature always depends on the objective of the inductive learning algorithm which uses the reduce set of feature. For induction an irrelevant feature is not useful at the same time all relevant features are not required always [1]. Some of the feature selection methods are: Wrapper method, Filter methods, Rough set methods and Schemata search method. Feature selection using similarity, Principal Component Analysis method, Filter-Wrapper Hybrid method. Again different types filtering method are also available for feature selection, like FOCUS algorithm, RELIEF algorithm, Las Vegas algorithm, Markov Blanket filtering, Decision Tree method and Genetic algorithms [2], [3], [4]. A.Correlation-based Feature Selection method Correlation-based Feature Selection (CFS) [5] comes under the category of filter algorithm. In this type of algorithm the subset of features are chosen based on ranks. The rank is determined based on a correlation based heuristic evaluation function. The evaluation function selects features which are highly correlated with the class and at the same time uncorrelated with each other. The evaluation function selects features which are highly correlated with the class and at the same time uncorrelated with each other. Selection of feature happens mainly based on the scale to which it predicts classes in the instance space not predicted by some other features. CFSs feature subset evaluation function is repeated here (with slightly modified notation) for ease of reference: is representing the heuristic merit of a feature subset S containing k features. is denoting the mean feature-class correlation (f ∈ s), and is stands for average feature-feature inter correlation. Numerator of the equation is giving an idea regarding the prediction of the class. The denominator is representing redundancy among the features. CFS uses best heuristic search strategy. If continuous five times fully expanded subsets do not give a better result the process stops; in other words that is the stopping condition of the process. B.Wrapper method The wrapper methodology, popularized by Kohavi and John (1997), provides an efficient way to determine the required subset of features, regardless of the machine learning algorithm. Here the machine learning algorithm is considered as a perfect black box. In practice, one needs to define: (i) the procedure to find the best subset among all possible variable subsets; (ii) the procedure to evaluate the prediction performance of a learning machine, which helps to guide the search and reaches the stopping condition; and (iii) the selection of predictor. Wrapper method performs a search in the space of all possible subsets. Use of machine learning algorithms makes wrapper method remarkably universal and simple. On the other hand when the number of features are very large then this method required an exhaustive search for O( ) number of subsets. Where n is the number of features. The complexity of the algorithm is O( ). So the algorithms is not efficient when n is very large.
  • 3. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 3 C.Filter Algorithm Filter algorithms determine the subset of optimal features by measuring the intrinsic properties of the data. The algorithm calculates a relevance score for each feature. Features having low relevance score are removed. Afterwards, the subset of relevant features is used as input in different machine learning algorithms. RELIEF Algorithm: The RELIEF assigns different weight to every feature depending on the relevance of the particular feature on the given context. RELIEF algorithm follows the filter method to determine the minimal set of features. It samples instances from the training data randomly and finds the relevance value based on near-hit and near-miss [6]. D.Rough Set Method In Velayutham [7] Rough Set Theory (RST) is used as a tool to discover data dependencies and to find the subset of relevant features. The rough set is approximation of set by two concepts known as lower and upper approximations. The domain objects are represented by the lower approximation. These domain objects belong to the subset of interest with a certainty. On the other hand the upper approximation describes the objects which are possibly belonging to subset. Approximations are made with respect to a particular subset of features or parameters. Considering an information system as I = (U, A d). In the information system U is the universe with a non-empty set of finite objects. On the other hand A represents nonempty finite set of condition attributes, and d represents the decision attribute, the corresponding function is , where represents set of values of a. If then there is an associated equivalence relation: The partition of U generated by is denoted by U/P. If (x, y) IND(P) holds good, then it is difficult to find a clearly distinguish x and y are by attributes from P. The equivalence classes of the P in discernibility relation are denoted [x]p. Considering XU the P-lower approximation PX of set X and P-upper approximation PX of set X can be defined as: PX = {X ∈ U|[X]p X } X = {X ∈ U|[X]p X } Let P,Q A be equivalence relations over U , then positive region defined as: PX. The positive region of the partition U/Q with respect to P, is the set of all objects of U that can be certainly classified to blocks of the partition . U/Q by means of P. Q depends on P in a degree denoted Where If k=1, Q depends totally on P, if depends partially on P, if k=0 then Q does not depend on P.
  • 4. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 The Unsupervised Quick Reduct (USQR) algorithm endeavors to calculate a reduct without fully generating all possible subsets. In Velayutham [8] the algorithm starts off with an empty set. In each iteration one attribute is added in the set. Those attributes are selected depending on the rough set dependency metric. This procedure continues until the value reaches its maximum value for that dataset. The mean dependency of each attribute is calculated and best candidate is chosen: 4 The data set is considered as consistent if the process ends at dependency value 1; otherwise it is inconsistent. A.Chi-square feature selection Chi-square method calculates the lack of information between a feature and a category. Chi-square distribution is used for comparison with one degree of freedom to judge extremeness [9]. It is defined as: Here feature selection metrics (four known measures and two proposed variants), which are functions of the following four dependency tuples: ( ): presence of t and membership in . ( ): presence of t and non-membership in . ( ): absence of and membership in . ( ): absence of t and non-membership in . III.CLUSTERING METHOD K-means clustering [9] is a method of cluster creation, where the main objective is to partition n observations into K groups/clusters in each instance belongs to the cluster with nearest mean. This results in a partitioning of the data space into Voronoi cells. Mathematically Voronoi diagram helps to divide a space into a number of regions. Depending on some predefined conditions set of point is taken in consideration they are known as seeds, sites, or generators. Around each site several points which are closet to that site form a region. The regions are called Voronoi cells. The most popular algorithm uses an iterative technique to reach the stopping point. It is commonly famous as K-means algorithm. In computer science community it is also known as Lloyd’s algorithm. IV.CLUSTER ANALYSIS METHOD The cluster analysis methods provide the measure, using which the goodness clusters can be determined. The cluster analysis methods has been used are Davies Bouldin index and Dunn index (J. C. Dunn 1974).
  • 5. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 5 A.DaviesBouldin index The Davies Bouldin index can be calculated by the following formula: In the above equation n represents the number of clusters. Cx stands for centroid of the cluster x. is representing the average distance between each of the point in the cluster and the centroid. is denoting the distance between two centroid of two clusters. The clustering algorithm produces a collection of clusters having smallest Davies Bouldin index is considered to be the best algorithm. Clusters with smallest Davies Bouldin index suggests low intra-cluster distances and high intra-cluster similarity. B.Dunn index The main objective of Dunn index is to find dense and well-separated clusters. It is defined as the ratio between the minimal inter- cluster distance to maximal intra-cluster distance. The following formula is used to calculate Dunn index for each cluster partition. If i and j are two clusters then distance between these two cluster is represented by d(i,j). For a cluster K, d(k) represents intra-cluster distance. The distance between centroids of the clusters can be considered as inter-cluster distance represented by: d(i,j). Similarly intra-cluster distance d(k) can be measured as distance between any two elements of cluster K. As clusters with high intra-cluster similarity and low inter-cluster similarity are the desirable that is why cluster with high Dunn index are preferable. V.EXPERIMENT METHODOLOGY The steps of the experiment are following: Environment setup: Firstly in a physical server VMware ESXi 5.1 has been installed and twenty six virtual machines (VM) have been installed in that. Load selection and run the load: There are four different types (CPU intensive, Network intensive, Memory intensive and Disk intensive) of load have been run. Each of CPU, Network and Disk intensive loads has been run in seven different VMs, whereas Network intensive load has been run in rest of the five VMs. Collection of result set: In third step data has been collected from the server using ESXTOP reply mode. Performance of different factor has been collected like CPU, Memory, Disk, Network and Power.
  • 6. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 Feature selection: In the result set there are many parameters in this phase the relevant parameters are selected using feature selection algorithms. 6 Use K-means algorithm: By using the clustering algorithms clusters has been created using the subset of features resulting from the feature selection algorithms. Comparison: In this step the clusters has been compared based on some cluster comparison method. TABLE I. LIST OF FEATURES FOR DIFFERENT RESOURCES Resources Parameters CPU %USED, %RUN, %SYS, %WAIT, %VMWAIT, %RDY, %IDLE, %OVRLP, %CSTP, %MLMTD, %SWPWT, SWTCH/s, MIG/s Memory SWCUR, SWTGT, SWR/s, SWW/s, LLSWR/s, LLSWW/s, CPTDR, CPTTGT, ZERO, SHRD,SHRDSVD, COWH, NHN, NMIG, NRMEM, NLMEM, N_L, GST_ND0, OVD_ND0,GST_ ND1, OVD_ND1, OVHDUM, OVHD, OVHDMAX, CMTTGT, CMTTGT, CMTCHRG, CMTPPS, CACHESZ, CACHUSD, ZIP/s, UNZIP/s, MEMSZ, GRANT, SZTGT, TCHD, TCHD W, ACTV, ACTVS, ACTVF, ACTVN, MCTLSZ, MCTLTGT, MCTLMAX Disk CMDS/s, READS/s, WRITES/s, MBREAD/s, MBWRTN/s, LAT/rd, LAT/wr Network PKTTX/s, MbTX/s, PKTRX/s, MbRX/s, %DRPTX, %DRPRX, ACTN/s, PKTTXMUL/s, PKTRXMUL/s, PKTTXBRD/s, PKTRXBRD/s Power %USED, %UTIL, %C0, %C1, %C2, %C3, %T0, %T1, %T2, %T3, %T4, %T5, %T6,%T7 VI.RESULTS A.Features selected by using different feature selection algorithms There are five different feature selection algorithms have been used: CFS method, RELIEF method, Chi Square Method, Wrapper method, Rough set method. The feature selection methods have been used on the data generated from the experiment. The methods select a subset of relevant features depending on some measuring metrics defined by the method. TABLE I .LIST OF SELECTED FEATURES FOR DIFFERENT RESOURCES Resource CFS RELIEF Chi Square Wrapper Rough Set CPU %USED, %IDLE, MIG/s %USED, %RUN, %SYS, %VMWAI T,%RDY, %IDLE, %OVRLP, %SWPWT MIG/s %USED, %RUN, %SYS, %WAIT, %VMWAI T,%RDY, %IDLE, %OVRLP, %CSTP, %MLMT, %USED, %RUN, %SYS, %WAIT, %VMWA IT,%RD, %IDLE, %OVRLP, %CSTP, %MLMD, %USD %SYS, MIG/s
  • 7. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 7 %SWPW, SWTCH/s, MIG/s %SWPT, SWTCH/, MIG/s Disk CMDS/s, LAT/r22 2d CMDS/s, READS/s, WRITES/, MBREAD/ s,LAT/rd, LAT/wr CDMS/s, READS/s, WRITES/, MBREAD/ s,MBWRT N/s,LAT/r d,LAT/wr CMDS/s, MBREAD /s,LAT/rd CMDS, LAT/rd, LAT/wr Network PKTRXs, MbRX/s PKTTX/s, MbTX/s, PKTRX/s, MbRX/s, ACTN/s, PKTRXM UL/ PKTTX/s, MbTX/s, PKTRX/s, MbRX/s PKTTX/s, MbTX/s, PKTRX/s, ACTN/s, PKTRXM UL/ PKTR/, ACTN/ s, PKTRX ,MUL/ s,PKTR XBRD/ s Memory SWCUR, COWH, OVHD, GRANT, SZTGT SZTGT, GRANT, MCTLMA X, TCHD W, MEMSZ, TCHD, SHRDSD, ACTVS, MCTLSZ, SHRD, ZERO, SWCUR, MCTLTT, ACTVF, ACTV, CMTTGT, ACTVN, OVHDU, OVHDM, OVHD, COWH, CACHUS D,NLME M,NHN, NRMEM, N L, CMTPPS, CACHESZ SWCUR, ZERO, SHRD, SHRDSV D, COWH, NHN, NRMEM, NLMEM, N L, OVHDUM ,OVHD, OVHDMA X, CMTTGT, CACHESZ ,CACHUS D,MEMSZ ,GRANT, SZTGT, TCHD, TCHD W, ACTV, ACTVS, ACTVF, ACTVN, MCTLMA X GRANT, SZTGT, SWCUR, SHRDSV D,SHRD, ZERO, ACTVS, TCHD W, ACTVF, TCHD, ACTV, CMTTGT, ACTVN, OVHD, MEMSZ, MCTLM AX, MCTLTG T, MCTLSZ, OVHDM AX, OVHDU M, COWH, CACHUS D SHRDS VD, GRAN T Power %USED %USED, %C1, %USED %USED, %C0, %C1 %USED, %C0, %C1 %USE D,%UT IL,%C1 B.Clustering using K-means algorithm After getting the selected features K-means clustering algorithm has been applied on each of the set. Where each setconsists of all the selected features of all resources by a particular method of feature selection. Here number of clusteris four. To find the goodness of the clusters, the analysis has been done by the help of Davies Bouldin index and Dunn index. Using these two different indices the goodness of the clustering algorithm has been found out. The cluster having maximum Davies Bouldin index and minimum Dunn index is considered to be the best cluster and the feature selection algorithm related to that cluster is considered to be the best algorithm among these algorithms using the data set. From this result it can be concluded that the parameters selected are optimum parameters from which we can have a clear idea about types of
  • 8. International Journal of Chaos, Control, Modelling and Simulation (IJCCMS) Vol.3, No.3, September 2014 load running in VMware ESXi 5.1 server. Here is the table where Davies Bouldin index and Dunn index for different clusters have been given. Where clusters found using the parameters selected by the respective feature selection algorithms. 8 TABLE I .DAVIES BOULDIN INDEX AND DUNN INDEX FOR CLUSTERS. Feature selection technique Davies Bouldin index Dunn index CFS 9.9122e+004 1.4196e-006 RELIEF 2.2679e+00 1.0874e-009 ChiSquare Method 2.9018e+007 2.442-009 Wrapper 2.1199e+007 1.0862e-009 Rough set 3.9079e+005 4.9040e-006 From the result it can be observed that the best result in both of the case is given by the clusters formed by the features selected by the CFS algorithm. VII.CONCLUSION So from the result it can be concluded that the set of parameters selected by the CFS is the best set of parameters of VMware ESXi 5.1. So from that set of parameters we can decide the type of load in VMware ESXi 5.1. At the same time VMs having high values of those parameters suggest the higher usage of those resources by the VM, which can be helpful for load balancing of Virtual Machine. REFERENCES [1] How Useful is Relevance? R.A. Caruana and D. Freitag, Technical report, Fall94 AAAI Symposium on Relevance, New Orleans, 1994, AAAI. [2] Selection of Relevant Features and Examples in Machine Learning, A.L.Blum and P.Langley, R. Greiner and D. Subramanian, eds., Artificial Intelligence on Relevance, volume 97, pages 245271. Artificial Intelligence, 1997. [3] An Evaluation of Feature Selection Methods and their Application to Computer Security, J.Doak, An Evaluation of Feature Selection Methods and their Application to Computer Security, 245271, 1992, Technical Report CSE9218, Davis, CA: University of California, Department of Computer Science. [4] Feature Selection for Knowledge Discovery and Data Mining, H.Liu and H.Motoda, 1998, Kluwer Academic Publishers, London, GB.[5]Mark A.Hall, Feature Selection for Discrete and Numeric Class Machine Learning. [6] D. W. Yijun Sun, A relief based feature extraction algorithm, Journal of the Royal Statistical Society. Series B (Methodological) Vol. 39, No. 1(1977), pp. 1-38, Tech. Rep., 2011. [7] A Novel Entropy Based Unsupervised Feature Selection Algorithm Using Rough Set Theory, K. Thangavel and C.Velayutham, Science And Management (ICAESM -2012) March 30, 31, 2012, 2012, IEEE International Conference On Advances In Engineering. [8] Feature selection for text categorization on imbalanced data, M. A. Hall,Sigkdd Explorations ,Volume 6, Issue 1 - Page 80, Tech. Rep., 2003.K. Elissa, “Title of paper if known,” unpublished. [9] Extensions to the k-Means Algorithm for Clustering Large Data Setswith Categorical Values, ZHEXUE HUANG, P1: SUD Data Mining and Knowledge Discovery KL657-03-Huang 12:59 Data Mining and Knowledge Discovery 2, 283304 (1998) 1998, October 27, 1998, Kluwer Academic Publishers. Manufactured in The Netherlands.