SlideShare a Scribd company logo
1 of 9
Download to read offline
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
322
A NEW METHOD TO REMOVE DEPENDENCE
OF FUZZY C-MEANS CLUSTERING
TECHNIQUE ON RANDOM INITIALIZATION
Samarjit Das1
, Hemanta K. Baruah2
1
Department of Computer Science &IT, 2
Vice-Chancellor
1
Cotton College, Assam, India
2
Bodoland University, Assam, India
1
ssaimm@rediffmail.com, 2
hemanta_bh@yahoo.com
ABSTARCT:
Fuzzy clustering techniques deal the situations where there is a possibility of belonging a single object to
more than one cluster. Although Fuzzy C-Means clustering technique of Bezdek is widely studied and
applied, its performance is highly dependent on the randomly initialized membership values of the
objects used for choosing the initial centroids. This paper proposes a modified method to remove the
effect of random initialization from Fuzzy C-Means clustering technique and to improve the overall
performance of it. In our proposed method we have used the algorithm of Yuan et al to determine the
initial centroids. These initial centroids are then used in the conventional Fuzzy C- Means clustering
technique of Bezdek to obtain the final clusters. We have tried to compare the performance of our
proposed method with that of conventional Fuzzy C-means clustering technique of Bezdek by using
Partition Coefficient and Clustering Entropy as validity indices.
Keywords: Fuzzy C-Means clustering technique, Random initialization, Partition Coefficient, Clustering
Entropy.
1. INTRODUCTION
In conventional hard clustering techniques a large dataset is partitioned into some smaller clusters where an
object either belongs completely to a particular cluster or does not belong to it at all. With the advent of the
concept of Fuzzy Set Theory (FST) of Zadeh (1965) which particularly deals the situations pertaining to non-
probabilistic uncertainty, the conventional hard clustering techniques have unlocked a new way of clustering,
known as fuzzy clustering, where a single object may belong exactly to one cluster or partially to more than one
cluster depending on the membership value of that object. Baruah (2011a, 2011b) has proved that the
membership value of a fuzzy number can be expressed as a difference between the membership function and a
reference function and therefore the fuzzy membership value and the fuzzy membership function for the
complement of a fuzzy set are not the same. In the literature the Fuzzy C-Means (FCM) clustering technique of
Bezdek (1981) has been found to be very popular among the research community. Derrig and
Ostaszewski (1995) have applied the FCM of Bezdek (1981) in their research work where they have explained a
method of pattern recognition for risk and claim classification. Das and Baruah (2013a) have shown the
application of the FCM of Bezdek (1981) on vehicular pollution, through which they have discussed the
importance of application of a fuzzy clustering technique on a dataset describing vehicular pollution, instead of
a hard clustering technique. Das and Baruah (2013b) have applied the FCM of Bezdek (1981) and Gustafson
and Kessel (GK) clustering technique of Gustafson and Kessel (1979) on the same dataset to make a comparison
between these two clustering techniques and found that the overall performance of FCM is better than that of
GK. Although it is evident in the literature that the FCM performs better as compared to other fuzzy clustering
techniques, the performance of FCM is highly dependent on the randomly initialized membership values of the
objects used for selecting the initial centroids. Yuan et al. (2004) proposed a systematic method for finding the
initial centroids where there is no scope of randomness and therefore the centroids obtained by this method are
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
323
found to be consistent. In our proposed work we use these centroids thus obtained as the initial centroids in
FCM of Bezdek (1981) to remove the effect of random initialization from FCM and also to improve the overall
performance of it. Using Partition Coefficient (PC) and Clustering Entropy (CE) as validity indices (see
Bezdek (1981) and Bensaid et al.(1996)) we have tried to make a comparison of the performances of these two
clustering techniques.
In section-2 we provide the steps of the algorithms used in our present work. Through section-3 we describe
our proposed model. The results and analysis of our present work have been placed in section-4. Finally we put
the conclusion in section-5.
2. ALGORITHMS
The basic task of a clustering technique is to divide n patterns, where n is a natural number , represented by
vectors in a p-dimensional Euclidean space, into c, 2≤ c <n , categorically homogeneous subsets which are
called clusters. Let the data set be X= {x1, x2, ……….., xn }, where xk ={ xk1, xk2, ……….., xkp },
k= 1,2,3,……..,n. Each xk is called a feature vector and xkj where j=1,2,…..p is the jth
feature of the kth
feature
vector. A partition of the dataset X into clusters is described by the membership functions of the elements of the
cluster. Let S1, S2,…….,Sc denote the clusters with corresponding membership functions
1Sµ , 2Sµ , …, cSµ . A c x n matrix containing the membership values of the objects in the clusters
Ũ = nkcikx ,......,2,1,,.....,2,1S )]([ i ==µ is a fuzzy c- partition if it satisfies the following conditions
∑=
=
c
i
ks xi
1
1)(µ for each k= 1,2,….,n. (1)
∑=
≤≤
n
k
ks nxi
1
)(0 µ for each i= 1,2,….,c. (2)
Condition (1) says that each feature vector xk has its total membership value 1 divided among all clusters.
Condition (2) states that the sum of membership degrees of feature vectors in a given cluster does not exceed the
total number of feature vectors. In our proposed model we have used the algorithm of Yuan et al. (2004) as a
preprocessor to the FCM algorithm of Bezdek (1981). In sections 2.1 and 2.2 we provide the steps of FCM
algorithm of Bezdek (1981) and the algorithm of Yuan et al. (2004) respectively.
2.1. FCM Algorithm of Bezdek
Step 1: Choose the number of clusters, c, 2≤c<n, where n is the total number of feature vectors. Choose m,
1≤ m <α. Define the vector norm || ||(generally defined by the Euclidean distance) i.e.
|||| ik vx − = ∑=
−
p
j
ijkj vx
1
2
)( (3)
where kjx is the jth
feature of the kth
feature vector, for k=1,2,……,n; j=1,2,….,p and ijv , j-dimensional
centre of the ith
cluster for i=1,2,……,c; j=1,2,….,p; n, p and c denote the total number of feature vector , no. of
features in each feature vector and total number of clusters respectively.
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
324
Choose the initial fuzzy partition (by putting some random values)
nkciks xi ≤≤≤≤= 1,1
)0((0)
)]([U µ
Choose a parameter ∈>0 (this will tell us when to stop the iteration). Set the iteration counting parameter l
equal to 0.
Step 2: Calculate the fuzzy cluster centers ci
l
iv ,.....,2,1
)(
}{ = given by the following formula
∑
∑
=
=
= n
k
m
k
l
s
n
k
k
m
k
l
s
l
i
x
xx
v
i
i
1
)(
1
)(
)(
))((
))((
µ
µ
(4)
for i = 1, 2 , ….. c; k= 1, 2, , …..n.
Step 3: Calculate the new partition matrix (i.e. membership matrix)
nkcik
l
s
l
xU i ≤≤≤≤
++
= 1,1
)1()1(
)]([µ ,
where
∑=
−
+
−
−
= c
j
m
l
jk
l
ik
k
l
s
vx
vx
xi
1
1
2
)(
)(
)1(
)
||||
||||
(
1
)(µ (5)
for i=1,2,……..,c and k=1,2,……..,n.
If ,
)(l
ik vx = formula (5) cannot be used. In this case the membership function is
iifk
ciiifkk
l
s xi
=
=≠
+
= 1
,....,2,1,0
)1(
{)(µ
Step 4: Calculate ∆ = |||| )()1( ll
UU −+
(6)
If ∆ >∈, repeat steps 2, 3 and 4. Otherwise, stop at some iteration count
*
l .
2.2. Algorithm of Yuan et al.
Step 1: Set m =1;
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
325
Step 2: Compute the distance between each data point and all other data points in the set X;
Step 3:Find the closet pair of data points from the set X and form a data point set Am (1≤m≤c, c is the number of
clusters) which contains these two data points, delete these two data points from the set X;
Step 4: Find the data point in X that is the closet to the data point set Am, add it to Am and delete it from X;
Step 5: Repeat step 4 until the number of data points in Am reaches 0.75*(n/c); (where .75 is a multiplication
factor (MF))
Step 6: If m<c, then m = m+1, find another pair of data points from X between which the distance is shortest,
form another data point set Am and delete them from X, go to step 4;
Step 7: For each data point set Am (1≤m≤c) find the arithmetic mean of the vectors of data points in Am, these
means will be the initial centroids.
3. OUR PRESENT WORK
As the initial centroids in FCM of Bezdek (1981) are obtained based on the randomly initialized membership
values of the objects, therefore the final clusters thus obtained are also not fixed. In other words there is
significant variation in the performance of FCM clustering technique while executed different times. In the
algorithm of Yuan et al. (2004) a systematic method is used to find the initial centroids where there is no
random initialization. In our proposed model we use these initial centroids thus obtained as inputs of FCM of
Bezdek (1981). In this way we have tried to remove the effect of random initialization from FCM clustering
technique and also to improve its overall performance. We explain our proposed model with a flowchart (see
Fig.1). We have applied both FCM clustering technique and our method ten (10) times each on the same dataset
(see Table1) and tried to make a comparison of the performances of these two clustering techniques. We have
used two validity measures of Bezdek (1981) and Bensaid et al. (1996) and the number of iterations to obtain
the performances of these two clustering techniques. The mathematical formulae of these two validity measures
have been given in the following.
(a) Partition Coefficient (PC): measures the overlapping between clusters.
∑∑= =
=
c
i
n
j
ij
n
cPC
1
2
1
)(
1
)( µ
(b) Clustering Entropy (CE): measures the fuzziness of the cluster partition
)log(
1
)(
1 1
∑∑= =
=
c
i
ij
n
j
ij
n
cCE µµ
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
326
The proposed model of our present work has been given in the following.
Figure 1. Flowchart of our proposed model.
The dataset of our present work consists of fifty(50) Feature Vectors (FV) each of which is of dimension
three(03) namely Intelligent Quotient (IQ) , Achievement Motivation(AM) and Social Adjustment (SA). The
numerical values of our dataset have been given in the following table.
Find the initial centroids by the
algorithm of Yuan et al.
Read Dataset
Start
Take these centroids as the input of
FCM to calculate the membership
values of objects by FCM of Bezdek.
if(∆ >∈)
Update the membership
values by FCM of Bezdek.
Update the centroids by FCM
of Bezdek.
Stop
Yes No
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
327
Table 1. Data set of individual differences of fifty (50) feature vectors
with dimension (feature) three (03).
FV IQ AM SA FV IQ AM SA
1 91 18 55 26 110 18 55
2 85 16 40 27 100 16 40
3 120 19 74 28 100 18 75
4 90 18 75 29 70 14 30
5 92 17 74 30 105 17 55
6 82 17 55 31 79 14 35
7 95 19 75 32 80 15 34
8 89 18 74 33 125 20 75
9 96 19 75 34 100 19 75
10 90 17 55 35 125 19 85
11 97 16 54 36 80 18 60
12 125 21 74 37 85 18 70
13 100 19 75 38 145 25 90
14 90 17 54 39 80 18 74
15 100 18 84 40 92 17 55
16 95 19 75 41 120 18 70
17 130 23 85 42 145 30 80
18 130 19 75 43 95 18 50
19 90 17 55 44 80 16 36
20 91 17 56 45 90 17 55
21 140 22 82 46 115 23 84
22 92 18 75 47 100 18 80
23 101 18 55 48 80 14 35
24 85 16 54 49 105 19 75
25 97 19 54 50 120 21 74
4. RESULTS AND ANALYSIS
In this section we provide the results and analysis of our present work. Before making a comparison of the
performance of FCM of Bezdek (1981) with that of our proposed model we have tried to optimize the
performance level of our proposed model by taking the best choice of the multiplication factor (MF) ( see step 5
of the algorithm of Yuan et al. (2004) in section 2.2). Fig. 2 shows that the performance of our proposed model
is optimized when MF=.65.
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
328
Figure 2. Performance levels of our proposed model with different values of MF.
In Fig. 3 we see that the value of the validity index PC of FCM varies significantly in ten (10) different
executions. It is also seen in Fig. 3 that with the best choice of MF (when MF=.65) our proposed model results
consistent and better performance (i.e. with higher values of PC) than FCM. A similar result is reflected in
Fig.4 with the validity index CE. That is, our proposed model shows consistent and better performance (i.e.
lower values of CE) in contrast to inconsistent performance of FCM in ten(10) different executions.
Figure 3. Measures of the validity index PC of FCM and that (optimized value only) of
our proposed model in ten (10) different executions.
Figure 4. Measures of the validity index CE of FCM and that (optimized value only) of
our proposed model in ten (10) different executions.
Fig. 5 shows that the average value of the validity index PC (for different values of MF) of our proposed
model is more than that of FCM. This means that our model performs better than FCM even though the best
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
M
F
=.1
5
M
F
=.2
5
M
F
=.3
5
M
F
=.4
5
M
F
=.5
5
M
F
=.6
5
M
F
=.7
5
M
F
=.8
5
M
F
=.9
5
M
F
=1
.0
A
V
G
.
MULTIFICATION FACTOR(MF)
VALIDITYINDICES
PC CE
0.65
0.66
0.67
0.68
0.69
0.7
0.71
0.72
0.73
0.74
0.75
0.76
RUN1 RUN2 RUN3 RUN4 RUN5 RUN6 RUN7 RUN8 RUN9 RUN10
DIFFERENT EXECUTIONS
VALIDITYINDEXPC
PC FCM PC PROPOSED MODEL
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
RUN1 RUN2 RUN3 RUN4 RUN5 RUN6 RUN7 RUN8 RUN9 RUN10 AVG.
DIFFERENT EXECUTIONS
VALIDITYINDEXCE
CE FCM CE PROPOSED MODEL
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
329
choice of MF in our proposed model is not taken. In Fig. 6 we see that the average performance of our proposed
model is better ( i.e. the value of CE is less) than that of FCM.
Figure 5. Average value of the validity index PC of FCM and that (for different values of MF) of
our proposed model in ten (10) different executions.
Figure 6. Average value of the validity index CE of FCM and that (for different values of MF) of
our proposed model in ten (10) different executions.
In Fig. 7 we see that the average number of iterations (for different values of MF) of our proposed model is
less than that of FCM in ten(10) different executions.
0.7088
0.71731
0.704
0.706
0.708
0.71
0.712
0.714
0.716
0.718
AVERAGE
VALUE OF PC
FCM PROPOSED
MODEL
Series1
0.60056
0.58525
0.575
0.58
0.585
0.59
0.595
0.6
0.605
AVERAGE
VALUE OF CE
FCM PROPOSED
MODEL
Series1
E-ISSN: 2321–9637
Volume 2, Issue 1, January 2014
International Journal of Research in Advent Technology
Available Online at: http://www.ijrat.org
330
Figure 7. Average number of iterations of FCM and that (for different values of MF) of
our proposed model in ten (10) different executions.
Thus we see in the results that with respect to the two validity indices (i.e. PC and CE) and the number of
iterations our proposed model has a consistent and better performance.
5. CONCLUSIONS
Although FCM clustering technique is very popular among the research community, the major disadvantage of
it is that its performance is very inconsistent due to the randomly initialized membership values of the feature
vectors for selecting the initial centroids. Our proposed model which uses the algorithm of Yuan et al. as a
preprocessor of FCM of Bezdek, can remove this inconsistency of FCM due to randomness by giving consistent
and better performance. Although the average performance level of our proposed model is higher than that of
FCM, it is advisable to optimize the performance level of our model with the best choice of the multiplication
factor.
References
[1] Baruah, H.K.( 2011a): Towards forming a field of fuzzy sets. International Journal of Energy, Information and Communications, 2(1),
pp. 16-20.
[2] Baruah, H.K.(2011b): The theory of fuzzy sets: beliefs and realities. International Journal of Energy, Information and
Communications, 2( 2),pp. 1-22.
[3] Bensaid, A.M.; Hall, L.O.; Bezdek, J.C.(1996): Validity- guided (re) clustering with applications to image segmentation. IEEE Trans.
on Fuzzy Object, 2(2), pp.112-123.
[4] Bezdek, J.C.(1981). Pattern recognition with fuzzy objective function algorithms, Plenum Press, New York.
[5] Das, S.; Baruah, H. K.(2013a): Application of Fuzzy C-Means Clustering Technique in Vehicular Pollution. Journal of Process
Management – New Technologies,1( 3), pp.96-107.
[6] Das, S.; Baruah, H. K.(2013b): A comparison of two fuzzy clustering techniques . Journal of Process Management – New
Technologies,1( 4), pp.1-15.
[7] Derrig, R. A.; Ostaszewski, K. M.(1995): Fuzzy techniques of pattern recognition in risk and claim classification. Journal of Risk and
Insurance, 62(3), pp.447-482.
[8] Gustafson, D.E.; Kessel ,W.C. (1979):Fuzzy clustering with a fuzzy covariance matrix, Proc. IEEE CDC, San Diego, CA, USA,
pp.761–766.
[9] Yuan, F. ; Meng, Z.H.; Zhang, H.X.; Dong, C.R.(2004): A new algorithm to get the initial centroids, Proc. of the 3rd International
Conference on Machine Learning and Cybernetics, pp. 26-29.
[10] Zadeh, L. A.( 1965): Fuzzy sets. Information and Control, 8(3), pp. 338-353.
11
7
0
2
4
6
8
10
12
AVERAGE
VALUE OF
ITERATIONS
FCM PROPOSED MODEL
Series1

More Related Content

What's hot

The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...Waqas Tariq
 
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...ijcsa
 
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...ijistjournal
 
Segmentation and Classification of MRI Brain Tumor
Segmentation and Classification of MRI Brain TumorSegmentation and Classification of MRI Brain Tumor
Segmentation and Classification of MRI Brain TumorIRJET Journal
 
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...paperpublications3
 
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...IJCSEA Journal
 
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATION
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATIONFUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATION
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATIONijcsit
 
Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)sipij
 
2 achuthan c_pankaj--23-39
2 achuthan c_pankaj--23-392 achuthan c_pankaj--23-39
2 achuthan c_pankaj--23-39Alexander Decker
 
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASEBINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASEIJCSEA Journal
 
Kandemir Inferring Object Relevance From Gaze In Dynamic Scenes
Kandemir Inferring Object Relevance From Gaze In Dynamic ScenesKandemir Inferring Object Relevance From Gaze In Dynamic Scenes
Kandemir Inferring Object Relevance From Gaze In Dynamic ScenesKalle
 
Shift Invarient and Eigen Feature Based Image Fusion
Shift Invarient and Eigen Feature Based Image Fusion Shift Invarient and Eigen Feature Based Image Fusion
Shift Invarient and Eigen Feature Based Image Fusion ijcisjournal
 
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATION
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATIONMAGNETIC RESONANCE BRAIN IMAGE SEGMENTATION
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATIONVLSICS Design
 
An Efficient Multiplierless Transform algorithm for Video Coding
An Efficient Multiplierless Transform algorithm for Video CodingAn Efficient Multiplierless Transform algorithm for Video Coding
An Efficient Multiplierless Transform algorithm for Video CodingCSCJournals
 

What's hot (18)

Fuzzy c-means
Fuzzy c-meansFuzzy c-means
Fuzzy c-means
 
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
 
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...
Partition Sort Revisited: Reconfirming the Robustness in Average Case and Muc...
 
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...
DESPECKLING OF SAR IMAGES BY OPTIMIZING AVERAGED POWER SPECTRAL VALUE IN CURV...
 
Segmentation and Classification of MRI Brain Tumor
Segmentation and Classification of MRI Brain TumorSegmentation and Classification of MRI Brain Tumor
Segmentation and Classification of MRI Brain Tumor
 
A280108
A280108A280108
A280108
 
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...
An Approach for Iris Recognition Based On Singular Value Decomposition and Hi...
 
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...
PARTITION SORT REVISITED: RECONFIRMING THE ROBUSTNESS IN AVERAGE CASE AND MUC...
 
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATION
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATIONFUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATION
FUZZY IMAGE SEGMENTATION USING VALIDITY INDEXES CORRELATION
 
Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)Blind Image Seperation Using Forward Difference Method (FDM)
Blind Image Seperation Using Forward Difference Method (FDM)
 
2 achuthan c_pankaj--23-39
2 achuthan c_pankaj--23-392 achuthan c_pankaj--23-39
2 achuthan c_pankaj--23-39
 
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASEBINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
BINARY TREE SORT IS MORE ROBUST THAN QUICK SORT IN AVERAGE CASE
 
I046850
I046850I046850
I046850
 
Nn
Nn Nn
Nn
 
Kandemir Inferring Object Relevance From Gaze In Dynamic Scenes
Kandemir Inferring Object Relevance From Gaze In Dynamic ScenesKandemir Inferring Object Relevance From Gaze In Dynamic Scenes
Kandemir Inferring Object Relevance From Gaze In Dynamic Scenes
 
Shift Invarient and Eigen Feature Based Image Fusion
Shift Invarient and Eigen Feature Based Image Fusion Shift Invarient and Eigen Feature Based Image Fusion
Shift Invarient and Eigen Feature Based Image Fusion
 
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATION
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATIONMAGNETIC RESONANCE BRAIN IMAGE SEGMENTATION
MAGNETIC RESONANCE BRAIN IMAGE SEGMENTATION
 
An Efficient Multiplierless Transform algorithm for Video Coding
An Efficient Multiplierless Transform algorithm for Video CodingAn Efficient Multiplierless Transform algorithm for Video Coding
An Efficient Multiplierless Transform algorithm for Video Coding
 

Viewers also liked

Paper id 252014138
Paper id 252014138Paper id 252014138
Paper id 252014138IJRAT
 
Paper id 41201605
Paper id 41201605Paper id 41201605
Paper id 41201605IJRAT
 
Paper id 27201444
Paper id 27201444Paper id 27201444
Paper id 27201444IJRAT
 
Paper id 24201441
Paper id 24201441Paper id 24201441
Paper id 24201441IJRAT
 
Paper id 36201528
Paper id 36201528Paper id 36201528
Paper id 36201528IJRAT
 
Paper id 35201501
Paper id 35201501Paper id 35201501
Paper id 35201501IJRAT
 
Paper id 37201535
Paper id 37201535Paper id 37201535
Paper id 37201535IJRAT
 
Paper id 28201425
Paper id 28201425Paper id 28201425
Paper id 28201425IJRAT
 
Paper id 252014103
Paper id 252014103Paper id 252014103
Paper id 252014103IJRAT
 
Paper id 41201624
Paper id 41201624Paper id 41201624
Paper id 41201624IJRAT
 
Paper id 41201610
Paper id 41201610Paper id 41201610
Paper id 41201610IJRAT
 
Paper id 36201521
Paper id 36201521Paper id 36201521
Paper id 36201521IJRAT
 
Paper id 252014146
Paper id 252014146Paper id 252014146
Paper id 252014146IJRAT
 
Paper id 312201514
Paper id 312201514Paper id 312201514
Paper id 312201514IJRAT
 
Paper id 212014109
Paper id 212014109Paper id 212014109
Paper id 212014109IJRAT
 
Paper id 212014121
Paper id 212014121Paper id 212014121
Paper id 212014121IJRAT
 
Paper id 42201607
Paper id 42201607Paper id 42201607
Paper id 42201607IJRAT
 
Paper id 212014100
Paper id 212014100Paper id 212014100
Paper id 212014100IJRAT
 
Paper id 312201526
Paper id 312201526Paper id 312201526
Paper id 312201526IJRAT
 
Paper id 21201462
Paper id 21201462Paper id 21201462
Paper id 21201462IJRAT
 

Viewers also liked (20)

Paper id 252014138
Paper id 252014138Paper id 252014138
Paper id 252014138
 
Paper id 41201605
Paper id 41201605Paper id 41201605
Paper id 41201605
 
Paper id 27201444
Paper id 27201444Paper id 27201444
Paper id 27201444
 
Paper id 24201441
Paper id 24201441Paper id 24201441
Paper id 24201441
 
Paper id 36201528
Paper id 36201528Paper id 36201528
Paper id 36201528
 
Paper id 35201501
Paper id 35201501Paper id 35201501
Paper id 35201501
 
Paper id 37201535
Paper id 37201535Paper id 37201535
Paper id 37201535
 
Paper id 28201425
Paper id 28201425Paper id 28201425
Paper id 28201425
 
Paper id 252014103
Paper id 252014103Paper id 252014103
Paper id 252014103
 
Paper id 41201624
Paper id 41201624Paper id 41201624
Paper id 41201624
 
Paper id 41201610
Paper id 41201610Paper id 41201610
Paper id 41201610
 
Paper id 36201521
Paper id 36201521Paper id 36201521
Paper id 36201521
 
Paper id 252014146
Paper id 252014146Paper id 252014146
Paper id 252014146
 
Paper id 312201514
Paper id 312201514Paper id 312201514
Paper id 312201514
 
Paper id 212014109
Paper id 212014109Paper id 212014109
Paper id 212014109
 
Paper id 212014121
Paper id 212014121Paper id 212014121
Paper id 212014121
 
Paper id 42201607
Paper id 42201607Paper id 42201607
Paper id 42201607
 
Paper id 212014100
Paper id 212014100Paper id 212014100
Paper id 212014100
 
Paper id 312201526
Paper id 312201526Paper id 312201526
Paper id 312201526
 
Paper id 21201462
Paper id 21201462Paper id 21201462
Paper id 21201462
 

Similar to Paper id 21201488

The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...CSCJournals
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)theijes
 
Optimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering AlgorithmOptimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering AlgorithmIJERA Editor
 
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...ijsc
 
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...ijsc
 
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...IJCNCJournal
 
Fuzzy clustering1
Fuzzy clustering1Fuzzy clustering1
Fuzzy clustering1abc
 
Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemnooriasukmaningtyas
 
New Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionNew Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionIJERA Editor
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
 
Number of sources estimation using a hybrid algorithm for smart antenna
Number of sources estimation using a hybrid algorithm for  smart antennaNumber of sources estimation using a hybrid algorithm for  smart antenna
Number of sources estimation using a hybrid algorithm for smart antennaIJECEIAES
 
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...cscpconf
 
An Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution AlgorithmAn Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution AlgorithmIOSR Journals
 
Artificial bee colony with fcm for data clustering
Artificial bee colony with fcm for data clusteringArtificial bee colony with fcm for data clustering
Artificial bee colony with fcm for data clusteringAlie Banyuripan
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelIJECEIAES
 
Fuzzy c means_realestate_application
Fuzzy c means_realestate_applicationFuzzy c means_realestate_application
Fuzzy c means_realestate_applicationCemal Ardil
 
Comparison Between Clustering Algorithms for Microarray Data Analysis
Comparison Between Clustering Algorithms for Microarray Data AnalysisComparison Between Clustering Algorithms for Microarray Data Analysis
Comparison Between Clustering Algorithms for Microarray Data AnalysisIOSR Journals
 
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...IJERA Editor
 
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM ijcisjournal
 

Similar to Paper id 21201488 (20)

The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
Optimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering AlgorithmOptimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering Algorithm
 
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
INVERSIONOF MAGNETIC ANOMALIES DUE TO 2-D CYLINDRICAL STRUCTURES –BY AN ARTIF...
 
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...
Inversion of Magnetic Anomalies Due to 2-D Cylindrical Structures – By an Art...
 
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...
CONSTRUCTING A FUZZY NETWORK INTRUSION CLASSIFIER BASED ON DIFFERENTIAL EVOLU...
 
Fuzzy clustering1
Fuzzy clustering1Fuzzy clustering1
Fuzzy clustering1
 
Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problem
 
New Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionNew Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral Recognition
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
 
Number of sources estimation using a hybrid algorithm for smart antenna
Number of sources estimation using a hybrid algorithm for  smart antennaNumber of sources estimation using a hybrid algorithm for  smart antenna
Number of sources estimation using a hybrid algorithm for smart antenna
 
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...
Robust Adaptive Threshold Algorithm based on Kernel Fuzzy Clustering on Image...
 
An Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution AlgorithmAn Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution Algorithm
 
H010223640
H010223640H010223640
H010223640
 
Artificial bee colony with fcm for data clustering
Artificial bee colony with fcm for data clusteringArtificial bee colony with fcm for data clustering
Artificial bee colony with fcm for data clustering
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
 
Fuzzy c means_realestate_application
Fuzzy c means_realestate_applicationFuzzy c means_realestate_application
Fuzzy c means_realestate_application
 
Comparison Between Clustering Algorithms for Microarray Data Analysis
Comparison Between Clustering Algorithms for Microarray Data AnalysisComparison Between Clustering Algorithms for Microarray Data Analysis
Comparison Between Clustering Algorithms for Microarray Data Analysis
 
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...
Comparision of Clustering Algorithms usingNeural Network Classifier for Satel...
 
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM
Colour Image Segmentation Using Soft Rough Fuzzy-C-Means and Multi Class SVM
 

More from IJRAT

96202108
9620210896202108
96202108IJRAT
 
97202107
9720210797202107
97202107IJRAT
 
93202101
9320210193202101
93202101IJRAT
 
92202102
9220210292202102
92202102IJRAT
 
91202104
9120210491202104
91202104IJRAT
 
87202003
8720200387202003
87202003IJRAT
 
87202001
8720200187202001
87202001IJRAT
 
86202013
8620201386202013
86202013IJRAT
 
86202008
8620200886202008
86202008IJRAT
 
86202005
8620200586202005
86202005IJRAT
 
86202004
8620200486202004
86202004IJRAT
 
85202026
8520202685202026
85202026IJRAT
 
711201940
711201940711201940
711201940IJRAT
 
711201939
711201939711201939
711201939IJRAT
 
711201935
711201935711201935
711201935IJRAT
 
711201927
711201927711201927
711201927IJRAT
 
711201905
711201905711201905
711201905IJRAT
 
710201947
710201947710201947
710201947IJRAT
 
712201907
712201907712201907
712201907IJRAT
 
712201903
712201903712201903
712201903IJRAT
 

More from IJRAT (20)

96202108
9620210896202108
96202108
 
97202107
9720210797202107
97202107
 
93202101
9320210193202101
93202101
 
92202102
9220210292202102
92202102
 
91202104
9120210491202104
91202104
 
87202003
8720200387202003
87202003
 
87202001
8720200187202001
87202001
 
86202013
8620201386202013
86202013
 
86202008
8620200886202008
86202008
 
86202005
8620200586202005
86202005
 
86202004
8620200486202004
86202004
 
85202026
8520202685202026
85202026
 
711201940
711201940711201940
711201940
 
711201939
711201939711201939
711201939
 
711201935
711201935711201935
711201935
 
711201927
711201927711201927
711201927
 
711201905
711201905711201905
711201905
 
710201947
710201947710201947
710201947
 
712201907
712201907712201907
712201907
 
712201903
712201903712201903
712201903
 

Paper id 21201488

  • 1. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 322 A NEW METHOD TO REMOVE DEPENDENCE OF FUZZY C-MEANS CLUSTERING TECHNIQUE ON RANDOM INITIALIZATION Samarjit Das1 , Hemanta K. Baruah2 1 Department of Computer Science &IT, 2 Vice-Chancellor 1 Cotton College, Assam, India 2 Bodoland University, Assam, India 1 ssaimm@rediffmail.com, 2 hemanta_bh@yahoo.com ABSTARCT: Fuzzy clustering techniques deal the situations where there is a possibility of belonging a single object to more than one cluster. Although Fuzzy C-Means clustering technique of Bezdek is widely studied and applied, its performance is highly dependent on the randomly initialized membership values of the objects used for choosing the initial centroids. This paper proposes a modified method to remove the effect of random initialization from Fuzzy C-Means clustering technique and to improve the overall performance of it. In our proposed method we have used the algorithm of Yuan et al to determine the initial centroids. These initial centroids are then used in the conventional Fuzzy C- Means clustering technique of Bezdek to obtain the final clusters. We have tried to compare the performance of our proposed method with that of conventional Fuzzy C-means clustering technique of Bezdek by using Partition Coefficient and Clustering Entropy as validity indices. Keywords: Fuzzy C-Means clustering technique, Random initialization, Partition Coefficient, Clustering Entropy. 1. INTRODUCTION In conventional hard clustering techniques a large dataset is partitioned into some smaller clusters where an object either belongs completely to a particular cluster or does not belong to it at all. With the advent of the concept of Fuzzy Set Theory (FST) of Zadeh (1965) which particularly deals the situations pertaining to non- probabilistic uncertainty, the conventional hard clustering techniques have unlocked a new way of clustering, known as fuzzy clustering, where a single object may belong exactly to one cluster or partially to more than one cluster depending on the membership value of that object. Baruah (2011a, 2011b) has proved that the membership value of a fuzzy number can be expressed as a difference between the membership function and a reference function and therefore the fuzzy membership value and the fuzzy membership function for the complement of a fuzzy set are not the same. In the literature the Fuzzy C-Means (FCM) clustering technique of Bezdek (1981) has been found to be very popular among the research community. Derrig and Ostaszewski (1995) have applied the FCM of Bezdek (1981) in their research work where they have explained a method of pattern recognition for risk and claim classification. Das and Baruah (2013a) have shown the application of the FCM of Bezdek (1981) on vehicular pollution, through which they have discussed the importance of application of a fuzzy clustering technique on a dataset describing vehicular pollution, instead of a hard clustering technique. Das and Baruah (2013b) have applied the FCM of Bezdek (1981) and Gustafson and Kessel (GK) clustering technique of Gustafson and Kessel (1979) on the same dataset to make a comparison between these two clustering techniques and found that the overall performance of FCM is better than that of GK. Although it is evident in the literature that the FCM performs better as compared to other fuzzy clustering techniques, the performance of FCM is highly dependent on the randomly initialized membership values of the objects used for selecting the initial centroids. Yuan et al. (2004) proposed a systematic method for finding the initial centroids where there is no scope of randomness and therefore the centroids obtained by this method are
  • 2. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 323 found to be consistent. In our proposed work we use these centroids thus obtained as the initial centroids in FCM of Bezdek (1981) to remove the effect of random initialization from FCM and also to improve the overall performance of it. Using Partition Coefficient (PC) and Clustering Entropy (CE) as validity indices (see Bezdek (1981) and Bensaid et al.(1996)) we have tried to make a comparison of the performances of these two clustering techniques. In section-2 we provide the steps of the algorithms used in our present work. Through section-3 we describe our proposed model. The results and analysis of our present work have been placed in section-4. Finally we put the conclusion in section-5. 2. ALGORITHMS The basic task of a clustering technique is to divide n patterns, where n is a natural number , represented by vectors in a p-dimensional Euclidean space, into c, 2≤ c <n , categorically homogeneous subsets which are called clusters. Let the data set be X= {x1, x2, ……….., xn }, where xk ={ xk1, xk2, ……….., xkp }, k= 1,2,3,……..,n. Each xk is called a feature vector and xkj where j=1,2,…..p is the jth feature of the kth feature vector. A partition of the dataset X into clusters is described by the membership functions of the elements of the cluster. Let S1, S2,…….,Sc denote the clusters with corresponding membership functions 1Sµ , 2Sµ , …, cSµ . A c x n matrix containing the membership values of the objects in the clusters Ũ = nkcikx ,......,2,1,,.....,2,1S )]([ i ==µ is a fuzzy c- partition if it satisfies the following conditions ∑= = c i ks xi 1 1)(µ for each k= 1,2,….,n. (1) ∑= ≤≤ n k ks nxi 1 )(0 µ for each i= 1,2,….,c. (2) Condition (1) says that each feature vector xk has its total membership value 1 divided among all clusters. Condition (2) states that the sum of membership degrees of feature vectors in a given cluster does not exceed the total number of feature vectors. In our proposed model we have used the algorithm of Yuan et al. (2004) as a preprocessor to the FCM algorithm of Bezdek (1981). In sections 2.1 and 2.2 we provide the steps of FCM algorithm of Bezdek (1981) and the algorithm of Yuan et al. (2004) respectively. 2.1. FCM Algorithm of Bezdek Step 1: Choose the number of clusters, c, 2≤c<n, where n is the total number of feature vectors. Choose m, 1≤ m <α. Define the vector norm || ||(generally defined by the Euclidean distance) i.e. |||| ik vx − = ∑= − p j ijkj vx 1 2 )( (3) where kjx is the jth feature of the kth feature vector, for k=1,2,……,n; j=1,2,….,p and ijv , j-dimensional centre of the ith cluster for i=1,2,……,c; j=1,2,….,p; n, p and c denote the total number of feature vector , no. of features in each feature vector and total number of clusters respectively.
  • 3. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 324 Choose the initial fuzzy partition (by putting some random values) nkciks xi ≤≤≤≤= 1,1 )0((0) )]([U µ Choose a parameter ∈>0 (this will tell us when to stop the iteration). Set the iteration counting parameter l equal to 0. Step 2: Calculate the fuzzy cluster centers ci l iv ,.....,2,1 )( }{ = given by the following formula ∑ ∑ = = = n k m k l s n k k m k l s l i x xx v i i 1 )( 1 )( )( ))(( ))(( µ µ (4) for i = 1, 2 , ….. c; k= 1, 2, , …..n. Step 3: Calculate the new partition matrix (i.e. membership matrix) nkcik l s l xU i ≤≤≤≤ ++ = 1,1 )1()1( )]([µ , where ∑= − + − − = c j m l jk l ik k l s vx vx xi 1 1 2 )( )( )1( ) |||| |||| ( 1 )(µ (5) for i=1,2,……..,c and k=1,2,……..,n. If , )(l ik vx = formula (5) cannot be used. In this case the membership function is iifk ciiifkk l s xi = =≠ + = 1 ,....,2,1,0 )1( {)(µ Step 4: Calculate ∆ = |||| )()1( ll UU −+ (6) If ∆ >∈, repeat steps 2, 3 and 4. Otherwise, stop at some iteration count * l . 2.2. Algorithm of Yuan et al. Step 1: Set m =1;
  • 4. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 325 Step 2: Compute the distance between each data point and all other data points in the set X; Step 3:Find the closet pair of data points from the set X and form a data point set Am (1≤m≤c, c is the number of clusters) which contains these two data points, delete these two data points from the set X; Step 4: Find the data point in X that is the closet to the data point set Am, add it to Am and delete it from X; Step 5: Repeat step 4 until the number of data points in Am reaches 0.75*(n/c); (where .75 is a multiplication factor (MF)) Step 6: If m<c, then m = m+1, find another pair of data points from X between which the distance is shortest, form another data point set Am and delete them from X, go to step 4; Step 7: For each data point set Am (1≤m≤c) find the arithmetic mean of the vectors of data points in Am, these means will be the initial centroids. 3. OUR PRESENT WORK As the initial centroids in FCM of Bezdek (1981) are obtained based on the randomly initialized membership values of the objects, therefore the final clusters thus obtained are also not fixed. In other words there is significant variation in the performance of FCM clustering technique while executed different times. In the algorithm of Yuan et al. (2004) a systematic method is used to find the initial centroids where there is no random initialization. In our proposed model we use these initial centroids thus obtained as inputs of FCM of Bezdek (1981). In this way we have tried to remove the effect of random initialization from FCM clustering technique and also to improve its overall performance. We explain our proposed model with a flowchart (see Fig.1). We have applied both FCM clustering technique and our method ten (10) times each on the same dataset (see Table1) and tried to make a comparison of the performances of these two clustering techniques. We have used two validity measures of Bezdek (1981) and Bensaid et al. (1996) and the number of iterations to obtain the performances of these two clustering techniques. The mathematical formulae of these two validity measures have been given in the following. (a) Partition Coefficient (PC): measures the overlapping between clusters. ∑∑= = = c i n j ij n cPC 1 2 1 )( 1 )( µ (b) Clustering Entropy (CE): measures the fuzziness of the cluster partition )log( 1 )( 1 1 ∑∑= = = c i ij n j ij n cCE µµ
  • 5. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 326 The proposed model of our present work has been given in the following. Figure 1. Flowchart of our proposed model. The dataset of our present work consists of fifty(50) Feature Vectors (FV) each of which is of dimension three(03) namely Intelligent Quotient (IQ) , Achievement Motivation(AM) and Social Adjustment (SA). The numerical values of our dataset have been given in the following table. Find the initial centroids by the algorithm of Yuan et al. Read Dataset Start Take these centroids as the input of FCM to calculate the membership values of objects by FCM of Bezdek. if(∆ >∈) Update the membership values by FCM of Bezdek. Update the centroids by FCM of Bezdek. Stop Yes No
  • 6. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 327 Table 1. Data set of individual differences of fifty (50) feature vectors with dimension (feature) three (03). FV IQ AM SA FV IQ AM SA 1 91 18 55 26 110 18 55 2 85 16 40 27 100 16 40 3 120 19 74 28 100 18 75 4 90 18 75 29 70 14 30 5 92 17 74 30 105 17 55 6 82 17 55 31 79 14 35 7 95 19 75 32 80 15 34 8 89 18 74 33 125 20 75 9 96 19 75 34 100 19 75 10 90 17 55 35 125 19 85 11 97 16 54 36 80 18 60 12 125 21 74 37 85 18 70 13 100 19 75 38 145 25 90 14 90 17 54 39 80 18 74 15 100 18 84 40 92 17 55 16 95 19 75 41 120 18 70 17 130 23 85 42 145 30 80 18 130 19 75 43 95 18 50 19 90 17 55 44 80 16 36 20 91 17 56 45 90 17 55 21 140 22 82 46 115 23 84 22 92 18 75 47 100 18 80 23 101 18 55 48 80 14 35 24 85 16 54 49 105 19 75 25 97 19 54 50 120 21 74 4. RESULTS AND ANALYSIS In this section we provide the results and analysis of our present work. Before making a comparison of the performance of FCM of Bezdek (1981) with that of our proposed model we have tried to optimize the performance level of our proposed model by taking the best choice of the multiplication factor (MF) ( see step 5 of the algorithm of Yuan et al. (2004) in section 2.2). Fig. 2 shows that the performance of our proposed model is optimized when MF=.65.
  • 7. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 328 Figure 2. Performance levels of our proposed model with different values of MF. In Fig. 3 we see that the value of the validity index PC of FCM varies significantly in ten (10) different executions. It is also seen in Fig. 3 that with the best choice of MF (when MF=.65) our proposed model results consistent and better performance (i.e. with higher values of PC) than FCM. A similar result is reflected in Fig.4 with the validity index CE. That is, our proposed model shows consistent and better performance (i.e. lower values of CE) in contrast to inconsistent performance of FCM in ten(10) different executions. Figure 3. Measures of the validity index PC of FCM and that (optimized value only) of our proposed model in ten (10) different executions. Figure 4. Measures of the validity index CE of FCM and that (optimized value only) of our proposed model in ten (10) different executions. Fig. 5 shows that the average value of the validity index PC (for different values of MF) of our proposed model is more than that of FCM. This means that our model performs better than FCM even though the best 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 M F =.1 5 M F =.2 5 M F =.3 5 M F =.4 5 M F =.5 5 M F =.6 5 M F =.7 5 M F =.8 5 M F =.9 5 M F =1 .0 A V G . MULTIFICATION FACTOR(MF) VALIDITYINDICES PC CE 0.65 0.66 0.67 0.68 0.69 0.7 0.71 0.72 0.73 0.74 0.75 0.76 RUN1 RUN2 RUN3 RUN4 RUN5 RUN6 RUN7 RUN8 RUN9 RUN10 DIFFERENT EXECUTIONS VALIDITYINDEXPC PC FCM PC PROPOSED MODEL 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 RUN1 RUN2 RUN3 RUN4 RUN5 RUN6 RUN7 RUN8 RUN9 RUN10 AVG. DIFFERENT EXECUTIONS VALIDITYINDEXCE CE FCM CE PROPOSED MODEL
  • 8. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 329 choice of MF in our proposed model is not taken. In Fig. 6 we see that the average performance of our proposed model is better ( i.e. the value of CE is less) than that of FCM. Figure 5. Average value of the validity index PC of FCM and that (for different values of MF) of our proposed model in ten (10) different executions. Figure 6. Average value of the validity index CE of FCM and that (for different values of MF) of our proposed model in ten (10) different executions. In Fig. 7 we see that the average number of iterations (for different values of MF) of our proposed model is less than that of FCM in ten(10) different executions. 0.7088 0.71731 0.704 0.706 0.708 0.71 0.712 0.714 0.716 0.718 AVERAGE VALUE OF PC FCM PROPOSED MODEL Series1 0.60056 0.58525 0.575 0.58 0.585 0.59 0.595 0.6 0.605 AVERAGE VALUE OF CE FCM PROPOSED MODEL Series1
  • 9. E-ISSN: 2321–9637 Volume 2, Issue 1, January 2014 International Journal of Research in Advent Technology Available Online at: http://www.ijrat.org 330 Figure 7. Average number of iterations of FCM and that (for different values of MF) of our proposed model in ten (10) different executions. Thus we see in the results that with respect to the two validity indices (i.e. PC and CE) and the number of iterations our proposed model has a consistent and better performance. 5. CONCLUSIONS Although FCM clustering technique is very popular among the research community, the major disadvantage of it is that its performance is very inconsistent due to the randomly initialized membership values of the feature vectors for selecting the initial centroids. Our proposed model which uses the algorithm of Yuan et al. as a preprocessor of FCM of Bezdek, can remove this inconsistency of FCM due to randomness by giving consistent and better performance. Although the average performance level of our proposed model is higher than that of FCM, it is advisable to optimize the performance level of our model with the best choice of the multiplication factor. References [1] Baruah, H.K.( 2011a): Towards forming a field of fuzzy sets. International Journal of Energy, Information and Communications, 2(1), pp. 16-20. [2] Baruah, H.K.(2011b): The theory of fuzzy sets: beliefs and realities. International Journal of Energy, Information and Communications, 2( 2),pp. 1-22. [3] Bensaid, A.M.; Hall, L.O.; Bezdek, J.C.(1996): Validity- guided (re) clustering with applications to image segmentation. IEEE Trans. on Fuzzy Object, 2(2), pp.112-123. [4] Bezdek, J.C.(1981). Pattern recognition with fuzzy objective function algorithms, Plenum Press, New York. [5] Das, S.; Baruah, H. K.(2013a): Application of Fuzzy C-Means Clustering Technique in Vehicular Pollution. Journal of Process Management – New Technologies,1( 3), pp.96-107. [6] Das, S.; Baruah, H. K.(2013b): A comparison of two fuzzy clustering techniques . Journal of Process Management – New Technologies,1( 4), pp.1-15. [7] Derrig, R. A.; Ostaszewski, K. M.(1995): Fuzzy techniques of pattern recognition in risk and claim classification. Journal of Risk and Insurance, 62(3), pp.447-482. [8] Gustafson, D.E.; Kessel ,W.C. (1979):Fuzzy clustering with a fuzzy covariance matrix, Proc. IEEE CDC, San Diego, CA, USA, pp.761–766. [9] Yuan, F. ; Meng, Z.H.; Zhang, H.X.; Dong, C.R.(2004): A new algorithm to get the initial centroids, Proc. of the 3rd International Conference on Machine Learning and Cybernetics, pp. 26-29. [10] Zadeh, L. A.( 1965): Fuzzy sets. Information and Control, 8(3), pp. 338-353. 11 7 0 2 4 6 8 10 12 AVERAGE VALUE OF ITERATIONS FCM PROPOSED MODEL Series1