4. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
5. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
• rules with conf. and target label (How?)
6. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
• rules with conf. and target label (How?)
• combination optimization problem (6 ) n
7. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
• rules with conf. and target label (How?)
• combination optimization problem (6 ) n
• Use SA to find a set of fuzzy rules (Contribution)
8. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
• rules with conf. and target label (How?)
• combination optimization problem (6 ) n
• Use SA to find a set of fuzzy rules (Contribution)
• authors said
9. Goal and Contribution
• Construct a fuzzy classifier (Goal)
• map the attrs. to predefined fuzzy sets
• rules with conf. and target label (How?)
• combination optimization problem (6 ) n
• Use SA to find a set of fuzzy rules (Contribution)
• authors said
10. The used antecedent
fuzzy sets
1.0
Membership
1.small
2.medium small
S MS M ML L
3.medium
4.medium large
Attribute Value
0.0 1.0 5.large
6.don’t care
Membership
1.0 if x1 is small and
x2 is medium and
DC
x3 is don’t care
Encode: 136
Attribute Value
0.0 1.0
11. Determination of Cj and CFj
1. calculate the compatibility of each training pattern xp with the rule Rj
µ j (x p ) = µ j1 (x p1 ) ×L × µ jn (x pn ), p = 1,K , m
2. for each class, calculate the relative sum of compatibility grades
of the training patterns in class h with the rule Rj
∑
βClass h (R j ) = µ j (x p ) / N Class h
x p ∈Class h
3. Find class hj hat
{ } if 0 or conflict, set Cj be φ
βClass h j (R j ) = max βClass1 (R j ),L , βClass c (R j )
)
4. if Cj = φ, set CFj of rule Rj to 0. Otherwise
c
( ) where β = ∑ βClass h (R j ) / (c − 1)
CFj = βClass h (R j ) − β / ∑ βClass h (R j )
)
)
j
h=1 h≠ h j
5. classify the sample xp with rule set S
{ } reject if µ j (x p ) = 0 ∀R j ∈S
µ j* (x p ) ⋅ CFj* = max µ j (x p ) ⋅ CFj R j ∈S
13. Structure of the goal classifier
Classifier #1
Set of rules
for class #1
Classifier #2
Set of rules
Decision
for class #2 Detected
Test
. Fusion Class
Dataset
.
.
Classifier #c
Set of rules
for class #c
14. Procedure of SAFCS
T = Tmax
Scurrent = Sinit
Sbest = Scurrent
EFcurrent = NNCP(Scurrent)
EFbest = NNCP(Sbest)
While (T ≥ Tmin)
For i = 1 to k
Call Metropolis(Scurrent, EFcurrent, Sbest, EFbest,T)
Time = Time + k
k=β×k
T = α ×T
Return(Sbest)
15. Procedure of SAFCS
Mb
1 −ΔEFb
T = Tmax # M = Mg + Mb ∑ ΔEFb Tmax
ΔEFb = =
Mb ln(Pinit )
Scurrent = Sinit # Ninit i=1
Sbest = Scurrent N
EFcurrent = NNCP(Scurrent) #NNCP(S)= m− ∑ NCP(Rj )
EFbest = NNCP(Sbest) j=1
While (T ≥ Tmin) # Tmin = 0.01
For i = 1 to k # k is num of calls of metropolis
Call Metropolis(Scurrent, EFcurrent, Sbest, EFbest,T)
Time = Time + k # Time is the spend time so far
k=β×k # β is a constant (set to 1)
T = α ×T # α is the cooling rate (set to 0.9)
Return(Sbest)
16. Metropolis Procedure
Snew = Perturb (Scurrent) # generate new S
EFnew = NNCP(Snew)
ΔEF = EFnew - EFcurrent
IF (ΔEF < 0), Then # better rule set
Scurrent = Snew
IF EFnew < EFbest , Then # better than best
Sbest = Snew
ELSEIF (rand(0,1) < exp(-ΔEF/T)), Then # accept, too
Scurrent = Snew
17. Perturbation(3 func.)
1. Modify
• select a rule from S randomly
• modified one or more antecedent of it
• if the consequent is equal, then replace;
otherwise, repeated
2. Delete f itnessmax (SClass h ) − f itness(R)
P (R) =
select with, f itnessmax (SClass h ) − f itnessmin (SClass h )
3. Create
the same as modify, but add
(NB:change more linguistic values than“Modify”,
they said for jump)
19. Parameters
Parameters Values
Initial set of rule size (Ninit) 50
Initial temperature (Tmax) 100
Final temperature (Tmin) 0.01
Cooling rate (α) 0.90
# Iteration at each temperature (k) 40
Iteration increment rate (β) 1
Estimate: 88 × 40 = 3520 iterations (keep in mind)
21. Dataset (UCI)
1829
H. Mohamadi et al. / Pattern Recognition 41 (2008) 1824 – 1833
Table 1
Features of the data sets used in computational experiments
Name #Instance #Attribute #Real. #Nominal #Class Dev. cla. (%) Mag. cla. (%) Min. cla. (%)
bswd 625 4 4 – 3 18.03 46.08 7.84
cra 690 15 6 9 2 5.51 55.51 44.49
ion 351 34 34 – 2 14.10 64.10 35.90
iris 150 4 4 – 3 – 33.33 33.33
lab 57 16 8 8 2 14.91 64.91 35.09
pima 768 8 8 – 2 15.10 65.10 34.90
wave 5000 40 40 – 3 0.36 33.84 33.06
wine 178 13 13 – 3 5.28 39.89 26.97
Dev.cla., deviation of class distribution; Mag. Cla, percentage of majority class instances; Min. Cla, percentage of minority class instances.
IBk [31] is the nearest neighbor classifier technique. It uses
Table 2
Parameters specification in computer simulations for the SAFCS the whole training set as the core of the classifier and Euclidean
distance to select the k nearest instances. The class prediction
Parameter Value
10-fold cross validation provided by the system is the majority class in these k examples.
Initial set of rules size (Ninit ) 50
Here, k is set equal to 3.
Initial temperature (Tmax ) 100
Naïve Bayes [32] is a very simple Bayesian network approach
Final temperature (Tmin ) 0.01
that assumes that the predictive attributes are conditionally
Cooling rate ( ) 0.90
independent given the class and also that no hidden or latent
# Iteration at each temperature (k) 40
Iteration increment rate ( ) 1 attributes influence the prediction process. These assumptions