SlideShare a Scribd company logo
1 of 9
Download to read offline
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
         Comparison of Hybrid PSO-SA Algorithm and Genetic

                              Algorithm for Classification

                       S. G. Sanjeevi1* A. Naga Nikhila2 Thaseem Khan3 G. Sumathi4
    1.    Associate Professor, Dept. of Comp. Science & Engg., National Institute of Technology, Warangal,
          A.P., India
    2.    Dept. of Comp. Science & Engg., National Institute of Technology, Warangal, A.P., India
    3.    Dept. of Comp. Science & Engg.,National Institute of Technology, Warangal, A.P., India
    4.    Dept. of Comp. Science & Engg., National Institute of Technology, Warangal, A.P., India
  * E-mail of the corresponding author: sgsanjeevi@yahoo.com

Abstract
In this work, we propose and present a Hybrid particle swarm optimization-Simulated annealing algorithm
and compare it with a Genetic algorithm for training respectively neural networks of identical architectures.
These neural networks were then tested on a classification task. In particle swarm optimization, behavior of
a particle is influenced by the experiential knowledge of the particle as well as socially exchanged
information. Particle swarm optimization follows a parallel search strategy. In simulated annealing uphill
moves are made in the search space in a stochastic fashion in addition to the downhill moves. Simulated
annealing therefore has better scope of escaping local minima and reach a global minimum in the search
space. Thus simulated annealing gives a selective randomness to the search. Genetic algorithm performs
parallel and randomized search. The goal of training the neural network is to minimize the sum of the
squares of the error between the target and observed output values for all the training samples and to deliver
good test performance on the test inputs. We compared the performance of the neural networks of identical
architectures trained by the Hybrid particle swarm optimization-simulated annealing and Genetic
algorithm respectively on a classification task and noted the results obtained. Neural network trained by
Hybrid particle swarm optimization-simulated annealing has given better results compared to the neural
network trained by the Genetic algorithm in the tests conducted by us.

Keywords: Classification, Hybrid particle swarm optimization-Simulated annealing, Simulated
Annealing, Genetic algorithm, Neural Network etc.


1. Introduction
Classification is an important activity of machine learning. Various algorithms are conventionally used for
classification task namely, Decision tree learning using ID3[1], Concept learning using Candidate
elimination [2], Neural networks [3], Naïve Bayes classifier [4] are some of the traditional methods used
for classification. Ever since back propagation algorithm was invented and popularized by [3], [5] and [6]
neural networks were actively used for classification. However, since back-propagation method follows hill
climbing approach, it is susceptible to occurrence of local minima. Hence we examine the use of alternative
methods for training neural networks. We examine the use of i) Hybrid particle swarm optimization-
simulated annealing algorithm ii) Genetic algorithm to train the neural networks. We study and compare
the performance of the neural networks trained by these algorithms on a classification task.


2. Architecture of Neural Network
Neural network designed for the classification task has the following architecture. It has four input units,
three hidden units and three output units in the input layer, hidden layer and output layer respectively.
Sigmoid activation functions were used with hidden and output units. Figure 1 shows the architectural
diagram of the neural network. Neurons are connected in the feed forward fashion as shown. Neural

                                                     37
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
Network has 12 weights between input and hidden layer and 9 weights between hidden and output layer. So
the total number of weights in the network are 21.




                                 Input      Hidden        Output


                                Figure 1 Architecture of Neural Network used

2.1 .Iris data
Neural Network shown in figure 1 is used to perform classification task on IRIS data. The data was taken
from the Univ. of California, Irvine (UCI), Machine learning repository. Iris data consists of 150 input-
output vector pairs. Each input vector consists of a 4 tuple having four attribute values corresponding to the
four input attributes respectively. Based on the input vector, output vector gives class to which it belongs.
Each output vector is a 3 tuple and will have a ‘1’ in first, second or third positions and zeros in rest two
positions, thereby indicating the class to which the input vector being considered belongs. Hence, we use 1-
of-n encoding on the output side for denoting the class value. The data of 150 input-output pairs is divided
randomly into two parts to create the training set and test set respectively. Data from training set is used to
train the neural network and data from the test set is used for test purposes. Few samples of IRIS data are
shown in table 1.
                                         Table 1.Sample Of Iris Data Used.


           S. No.     Attr. 1    Attr. 2        Attr. 3      Attr. 4   Class 1   Class 2   Class 3
           1          0.224      0.624          0.067        0.043     1         0         0
           2          0.749      0.502          0.627        0.541     0         1         0
           3          0.557      0.541          0.847        1         0         0         1
           4          0.11       0.502          0.051        0.043     1         0         0
           5          0.722      0.459          0.663        0.584     0         1         0




3. Genetic Algorithm to Train the Neural Network
Error back propagation algorithm is conventionally used to train the neural networks. Error-Back
propagation algorithm uses gradient descent search which is based on the concept of hill climbing. Main
disadvantage of neural network using back propagation algorithm is that since it uses hill climbing
approach it can get stuck at local minima.
Hence, here we explore the usage of alternative algorithms instead of conventional backpropagation
algorithm to optimize the performance of training a neural network and compare them.
The objective function required to be minimized for training the neural network is the error function E.
We define the error function E as follows. The error function E is given by

                                                          38
Computer Engineering and Intelligent Systems                                                   www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012




where x is a specific training example and X is the set of all training examples,      denotes the target
output for the kth output neuron corresponding to training sample x,      denotes the observed output for
the kth output neuron corresponding to training sample x. Hence, error function computed is a measure of
the sum of the squares of the error between target and observed output values for all the output neurons,
across all the training samples.
Genetic algorithms (GAs) were proposed and enunciated by John Holland at the University of Michigan
and popularized by David Goldberg [9, 10]. They have the characteristics of parallel search being
undertaken through potential solutions with a random exploration of the search space. They use the
evolutionary operators of selection, crossover and mutation on the population of the potential solutions.
Unlike backpropagation, GAs are stochastic and hence do not get into the problem of local minimum.
Genetic algorithms can be used for optimization. In the problem we have to optimize the error function
defined in equation (1) by minimizing it. Hence, the objective function we use is the error function in
equation (1) which needs to be minimized. To do this we define the fitness function f as A – E where A is a
positive number which is appropriately selected to be sufficiently large so that A – E is positive during the
search task undertaken through GA. Task of minimizing the error function is converted to maximizing the
value of A – E which is the fitness function f used with the genetic algorithm during the training of the
neural network.
We are using the genetic algorithm for the purpose of efficiently training the neural network. The idea of
efficiently training the neural network is achieved if the neural network predicts with higher accuracy the
test outputs corresponding to the test inputs given to the network. GAs use set of binary strings called
chromosomes to form a population of binary strings. Each chromosome which is a binary string is a
potential solution to the network. Hence, each chromosome needs to encode the set of weights present in a
neural network. The neural network shown in figure 1 has 21 weights in it. Hence each member of the
population (chromosome) needs to encode these complete set of 21 weights so that we can evaluate the
fitness function defined above. Below we describe the encoding of the weights for the neural network
shown in figure 1 for each chromosome.


3.1.Encoding of the Weights in the Neural Network
The Neural Network in figure 1 has 12 weights between input and hidden layer and 9 weights between
hidden and output layer. So the total number of weights in the network are 21. We coded these weights in
binary form. Each weight was assumed to vary between +12.75 and -12.75 with a precision of 0.05. This is
because weights learned by a typical neural network will be small positive or negative real numbers. Any
real number between -12.75 and +12.75 with a precision of 0.05 can be represented by a 9-bit binary
string. Thus, -12.75 was represented with a binary string ‘000000000’ and +12.75 was represented with
‘111111111’. For example, number 0.1 is represented by ‘100000010’. Each of the twenty one weights
present in the neural network was represented by a separate 9 bit binary string and the set of all the weights
present in the neural network is represented by their concatenation having 189 bits binary string. These 189
bits represent one possible assignment of values to the weights in the neural network.
In the initial population a set of 50 randomly generated binary strings were created, each binary string
having 189 bits. Different binary strings in population represent different potential solutions to the problem
of training the neural network. Each binary string represents one possible assignment of weights to the
neural network.
3.2. Genetic Algorithm
Here we describe the Genetic algorithm used for training the neural network.
    1.   A set of 50 binary strings, each string having 189 bits were randomly created to form the initial


                                                     39
Computer Engineering and Intelligent Systems                                                    www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
         population.
    2.   Evaluate the fitness function f for each of the binary strings in the present population P.
    3.   Elitism: Two of the binary strings with highest fitness values from the population are carried over
         to the new population being created Pnew.
    4.   Selection: The selection operation selects the parent strings from the present population for the
         crossover operation. This is done using the selection probability pi = fi /Σ fi . The roulette wheel
         concept is used to make the selection of binary strings. The probability pi is determined as the
         ratio of fitness of a given binary string with the sum of fitness values for all the 50 binary strings
         in the population P.
    5.   Crossover: Pairs of chromosomes were selected from population P using selection probability
         defined above and a single point crossover is performed on the pair to form two offspring.
         Selection probabilities encourage the selection of high fitness individuals to be selected as a pair.
         Crossover point was chosen at random and the two strings are interchanged at the point. This
         process was repeated for each of the 24 pairs of strings selected. Selection of each pair was done
         using selection probability calculations as explained above. Strings generated after crossover are
         kept in the new population being created Pnew.
    6.   Mutation: Mutation involves flipping a randomly selected bit. Mutation was done with a
         probability of 5% on the strings in Pnew . In the string selected for mutation, a randomly selected bit
         is flipped.
    7.   On completing above step the new population Pnew is formed. Make the new population created
         Pnew as the present population P for the next iteration of GA ie. Set P ←Pnew. Repeat steps 2 to 7
         for sufficient number of iterations. Steps 2 to 7 were repeated for 15000 iterations in our work.
         The string with the highest fitness among the population is returned as the solution and gives the
         assignment of weights to the neural network after training through the Genetic Algorithm.
4. Hybrid PSO - SA Algorithm
In particle swarm optimization [11, 12] a swarm of particles are flown through a multidimensional search
space. Position of each particle represents a potential solution. Position of each particle is changed by
adding a velocity vector to it. Velocity vector is influenced by the experiential knowledge of the particle as
well as socially exchanged information. The experiential knowledge of a particle A describes the distance
of the particle A from its own best position since the particle A’s first time step. This best position of a
particle is referred to as the personal best of the particle. The global best position in a swarm at a time t is
the best position found in the swarm of particles at the time t. The socially exchanged information of a
particle A describes the distance of a particle A from the global best position in the swarm at time t. The
experiential knowledge and socially exchanged information are also referred to as cognitive and social
components respectively. We propose and present here the hybrid PSO-SA algorithm in table 2.



               Table 2. Hybrid Particle Swarm Optimization-Simulated Annealing Algorithm


           Create a nx dimensional swarm of ns particles;
           repeat
              for each particle i = 1, . . .,ns do
           // yi denotes the personal best position of the particle i so far
                    // set the personal best position


               if f (xi) <f (yi) then


                                                        40
Computer Engineering and Intelligent Systems                                               www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
                yi = xi ;
             end


                  // ŷ denotes the global best of the swarm so far
                 // set the global best position
            if f (yi) <f ( ŷ ) then
            ŷ = yi ;
               end
             end


            for each particle i = 1, . . ., ns do
            update the velocity vi of particle i using equation (2);


         vi(t+1) = vi(t) + c1 r1(t)[yi(t) – xi(t) ] + c2 r2(t)[ŷ(t) – xi(t)] (2)


           // where yi denotes the personal best position of the particle i
           // and ŷ denotes the global best position of the swarm
           // and xi denotes the present position vector of particle i.


               update the position using equation (3);
                       bi(t) = xi (t) //storing present position
                       xi(t+1) = xi(t) + vi (t+1)           (3)


                     // applying simulated annealing
                       compute ∆E = (f (xi(t) - f(xi(t+1))         (4)
         // ∆E = (E value for the previous network before weight change) - (E value for the
         present network with changed configuration of weights).
          • if ∆E is positive then
         // new value of E is smaller and therefore better than previous value. Then
         // accept the new position and make it the current position
            else accept the new position xi(t+1) (even though it has a higher E value) with a
         probability p defined by
                                p=                    (5)

         • Revise the temperature T as per schedule defined below.
               The temperature schedule followed is defined as follows:
               The starting temperature is taken as T =1050;
               Current temp = T/log(iterations+1);
               Current temperature was changed after every 100 iterations using above formula.
         end



                                                         41
Computer Engineering and Intelligent Systems                                                     www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
            until stopping condition is true ;


In equation (2) of table 2, vi ,yi and ŷ are vectors of nx dimensions. vij denotes scalar component of vi in
dimension j. vij is calculated as shown in equation 6 where r1j and r2j are random values in the range [0,1]
and c1 and c2 are learning factors chosen as c1 = c2 = 2.
      vij(t+1) = vij(t) + c1 r1j(t)[yij(t) – xij(t) ] + c2 r2j(t)[ŷj(t) – xij) (t)]   (6)


4.1. Implementation details of Hybrid PSO-SA algorithm for training a neural network
We describe here the implementation details of hybrid PSO-SA algorithm for training the neural
network. There are 21 weights in the neural network shown in figure 1.
Hybrid PSO-SA algorithm combines particle swarm optimization algorithm with simulated annealing
approach in the context of training a neural network. The swarm is initialized with a population of 50
particles. Each particle has 21 weights. Each weight value corresponds to a position in a particular
dimension of the particle. Since there are 21 weights for each particle, there are therefore 21 dimensions.
Hence, position vector of each particle corresponds to a 21 dimensional weight vector. Position (weight) in
each dimension is modified by adding velocity value to it in that dimension. Fitness function
corresponding to a position xi , is f (xi) and it denotes the distance of position xi from the final solution. In
the hybrid PSO-SA algorithm this fitness function f (xi) is defined as error function E(w). Error function
E(w) is a function of the weights in the neural network as defined in equation (1). In the optimization
problem of training neural networks fitness function which is error function needs to be minimized.
Each particle’s velocity vector is updated by considering the personal best position of the particle, global
best position of the entire swarm and the present position vector of the particle as shown in equation (2) of
table 2. Velocity of a particle i in dimension j is calculated as shown in equation (6).
Hybrid PSO-SA algorithm combines pso algorithm with simulated annealing approach. Each of the 50
particles in the swarm is associated with 21 weights present in the neural network. The error function E(w)
which is a function of the weights in the neural network as defined in equation (1) is treated as the fitness
function. Error E(w) ( fitness function) needs to be minimized. For each of the 50 particles in the swarm,
the solution (position) given by the particle is accepted if the change in error function ∆E as defined in
equation 4 is positive, since this indicates error function is reducing in the present iteration compared to the
previous iteration value. If ∆Eis not positive then new position is accepted with a probability p given by
formula in equation (5). This is implemented by generating a random number between 0 and 1. If the
number generated is lesser than p then the new position is accepted, else the previous position of particle is
retained without changing. Each particle’s personal best position and the global best position of the swarm
are updated after each iteration. Hybrid PSO-SA algorithm was run for 15000 iterations. After 15000
iterations global best position gbest of the swarm is returned as the solution. Hence, stopping criterion for
the algorithm was chosen as completion of 15000 iterations.
Hybrid PSO-SA algorithm combines parallel search approach of PSO and selective random search and
global search properties of simulated annealing [7, 8] and hence combines the advantages of both the
approaches.


5. Experiments and Results
We have trained the neural network with architecture shown in figure 1 with genetic algorithm described in
section 3 on training set taken from the IRIS data. Training set was a subset of samples chosen randomly
from the IRIS data. Remaining samples from IRIS data were included in the test set. Performance was
observed on the test data set for predicting the class of each sample.

                   Table 3. Results Of Testing For Neural Network Using Genetic Algorithm




                                                                42
Computer Engineering and Intelligent Systems                                                  www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
   Sl. No               Samples        in   Samples in Test   Correct                 Misclassifications
                        Training Set        Set               classifications

   1                    100                 50                44                      6

   2                    95                  55                45                      10

   3                    85                  65                53                      12

   4                    75                  75                61                      14


We have chosen similarly, training sets of varying sizes from the IRIS data and included each time the
samples which were not selected for training set into test set. Neural network was trained by each of the
training sets and tested the performance of the network on corresponding test sets. Training was done for
15000 iterations with each of the training sets using the genetic algorithm. Results are shown in table 3.
We have also trained the neural network with architecture shown in figure 1 with Hybrid PSO-SA algorithm
described in section 4, with each of the training sets chosen above in table 3 and tested the performance of
the network on corresponding test sets. Training was done using Hybrid PSO-SA algorithm for 15000
iterations with each of the training sets. The results are shown in table 4.




               Table 4. Results Of Testing For Neural Network Using Hybrid PSO-SA Algorithm
       Sl.No    Samples in Training    Samples in Test   Correct                  Misclassifications
                Set                    Set               Classifications

       1        100                    50                47                       3

       2        95                     55                51                       4


       3        85                     65                61                       4

       4        75                     75                70                       5



Two neural networks with same architecture as shown in figure 1 were used for training and testing with
the two algorithms of Hybrid PSO-SA algorithm and Genetic algorithm respectively. Training was
performed for same number of 15000 iterations on each neural network. Same training and test sets were
used for comparison of neural networks performance with both the algorithms. Results of experiments and
testing point out that neural network trained with Hybrid PSO-SA algorithm gives better performance over
neural network trained with the Genetic algorithm across the training and test sets used.
6. Conclusion
Our objective was to compare the performance of feed-forward neural network trained with Hybrid PSO-
SA algorithm with the neural network trained by Genetic algorithm. We have trained the neural networks
with identical architectures with the Hybrid PSO-SA algorithm and the Genetic algorithm respectively. The
task we have tested using neural networks trained separately using these two algorithms is the IRIS data
classification. We found that neural network trained with Hybrid PSO-SA algorithm has given better

                                                    43
Computer Engineering and Intelligent Systems                                                 www.iiste.org
ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online)
Vol 3, No.2, 2012
classification performance among the two. Neural network trained with Hybrid PSO-SA algorithm
combines parallel search approach of PSO and selective random search and global search properties of
simulated annealing and hence combines the advantages of both the approaches. GA also uses parallel and
randomized search. Hence we compared these two global search approaches. Hybrid PSO-SA algorithm has
given better results for training a neural network than the Genetic algorithm.


References
A Quinlan, J.R.(1986). Induction of decision trees. Machine Learning, 1 (1), 81-106.
Mitchell, T.M.(1977). Version spaces: A candidate elimination approach to rule learning. Fifth International
Joint Conference on AI (pp.305-310). Cambridge, MA: MIT Press.
Rumelhart, D.E., & McClelland, J.L.(1986). Parallel distributed processing: exploration in the
microstructure of cognition(Vols.1 & 2). Cambridge, MA : MIT Press.
Duda, R.O., & Hart, P.E.(1973). Pattern classification and scene analysis New York: John Wiley & Sons.
Werbos, P. (1975). Beyond regression: New tools for prediction and analysis in the behavioral sciences
(Ph.D. dissertation). Harvard University.
Parker, D.(1985). Learning logic (MIT Technical Report TR-47).MIT Center for Research in
Computational Economics and Management Science.
Kirkpatrick, S., Gelatt,Jr., C.D., and M.P. Vecchi 1983. Optimization by simulated annealing.Science
220(4598).
Russel, S., &Norvig, P. (1995). Artificial intelligence: A modern approach. Englewood Cliffs, NJ:Prentice-
Hall.
Mitchell, T.M., 1997.Machine Learning, New York: McGraw-Hill.
K Goldberg, David E. (1989). Genetic Algorithms in Search Optimization and Machine Learning. Addison
Wesley. p. 41.ISBN 0201157675.
Kennedy, J., Eberhart, R., 1995. “Particle swarm optimization”, IEEE International Conference on Neural
Networks, Perth, WA , (pp.1942 – 1948), Australia.
Engelbrecht, A.P., 2007, Computational Intelligence, England: John Wiley & Sons.


Sriram G. Sanjeevi has acquired B.E. (Electronics and Communication Engineering) from Osmania
University, Hyderabad, India in 1981, M.Tech (Computer Science and Engineering) from M.I.T. Manipal,
Mangalore University, India in 1991 and Ph.D (Computer Science and Engineering) from IIT, Bombay,
India in 2009. He is currently Associate Professor & Head of Computer Science and Engineering
Department, National Institute of Technology, Warangal, India. His research interests are Neural networks,
Machine learning and Soft computing.

A. N. Nikhila has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in
2011. Her research interests are Neural networks, Machine learning.

Thaseem Khan has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in
2011. His research interests are Neural networks, Machine learning.

G. Sumathi has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in 2011.
Her research interests are Neural networks, Machine learning.




                                                    44
International Journals Call for Paper
The IISTE, a U.S. publisher, is currently hosting the academic journals listed below. The peer review process of the following journals
usually takes LESS THAN 14 business days and IISTE usually publishes a qualified article within 30 days. Authors should
send their full paper to the following email address. More information can be found in the IISTE website : www.iiste.org

Business, Economics, Finance and Management               PAPER SUBMISSION EMAIL
European Journal of Business and Management               EJBM@iiste.org
Research Journal of Finance and Accounting                RJFA@iiste.org
Journal of Economics and Sustainable Development          JESD@iiste.org
Information and Knowledge Management                      IKM@iiste.org
Developing Country Studies                                DCS@iiste.org
Industrial Engineering Letters                            IEL@iiste.org


Physical Sciences, Mathematics and Chemistry              PAPER SUBMISSION EMAIL
Journal of Natural Sciences Research                      JNSR@iiste.org
Chemistry and Materials Research                          CMR@iiste.org
Mathematical Theory and Modeling                          MTM@iiste.org
Advances in Physics Theories and Applications             APTA@iiste.org
Chemical and Process Engineering Research                 CPER@iiste.org


Engineering, Technology and Systems                       PAPER SUBMISSION EMAIL
Computer Engineering and Intelligent Systems              CEIS@iiste.org
Innovative Systems Design and Engineering                 ISDE@iiste.org
Journal of Energy Technologies and Policy                 JETP@iiste.org
Information and Knowledge Management                      IKM@iiste.org
Control Theory and Informatics                            CTI@iiste.org
Journal of Information Engineering and Applications       JIEA@iiste.org
Industrial Engineering Letters                            IEL@iiste.org
Network and Complex Systems                               NCS@iiste.org


Environment, Civil, Materials Sciences                    PAPER SUBMISSION EMAIL
Journal of Environment and Earth Science                  JEES@iiste.org
Civil and Environmental Research                          CER@iiste.org
Journal of Natural Sciences Research                      JNSR@iiste.org
Civil and Environmental Research                          CER@iiste.org


Life Science, Food and Medical Sciences                   PAPER SUBMISSION EMAIL
Journal of Natural Sciences Research                      JNSR@iiste.org
Journal of Biology, Agriculture and Healthcare            JBAH@iiste.org
Food Science and Quality Management                       FSQM@iiste.org
Chemistry and Materials Research                          CMR@iiste.org


Education, and other Social Sciences                      PAPER SUBMISSION EMAIL
Journal of Education and Practice                         JEP@iiste.org
Journal of Law, Policy and Globalization                  JLPG@iiste.org                       Global knowledge sharing:
New Media and Mass Communication                          NMMC@iiste.org                       EBSCO, Index Copernicus, Ulrich's
Journal of Energy Technologies and Policy                 JETP@iiste.org                       Periodicals Directory, JournalTOCS, PKP
Historical Research Letter                                HRL@iiste.org                        Open Archives Harvester, Bielefeld
                                                                                               Academic Search Engine, Elektronische
Public Policy and Administration Research                 PPAR@iiste.org                       Zeitschriftenbibliothek EZB, Open J-Gate,
International Affairs and Global Strategy                 IAGS@iiste.org                       OCLC WorldCat, Universe Digtial Library ,
Research on Humanities and Social Sciences                RHSS@iiste.org                       NewJour, Google Scholar.

Developing Country Studies                                DCS@iiste.org                        IISTE is member of CrossRef. All journals
Arts and Design Studies                                   ADS@iiste.org                        have high IC Impact Factor Values (ICV).

More Related Content

What's hot

MATLAB Code + Description : Very Simple Automatic English Optical Character R...
MATLAB Code + Description : Very Simple Automatic English Optical Character R...MATLAB Code + Description : Very Simple Automatic English Optical Character R...
MATLAB Code + Description : Very Simple Automatic English Optical Character R...Ahmed Gad
 
Unsupervised Feature Learning
Unsupervised Feature LearningUnsupervised Feature Learning
Unsupervised Feature LearningAmgad Muhammad
 
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14Daniel Lewis
 
Predictive Text Embedding using LINE
Predictive Text Embedding using LINEPredictive Text Embedding using LINE
Predictive Text Embedding using LINENishant Prateek
 
IRJET- Image Classification – Cat and Dog Images
IRJET- Image Classification – Cat and Dog ImagesIRJET- Image Classification – Cat and Dog Images
IRJET- Image Classification – Cat and Dog ImagesIRJET Journal
 
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...CSCJournals
 
A survey research summary on neural networks
A survey research summary on neural networksA survey research summary on neural networks
A survey research summary on neural networkseSAT Publishing House
 
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable Hardware
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable HardwareGenetic Algorithm Processor for Image Noise Filtering Using Evolvable Hardware
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable HardwareCSCJournals
 
Survey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsSurvey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsIRJET Journal
 
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...IJERA Editor
 
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLING
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGLOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLING
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGijaia
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...cscpconf
 
Improving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docImproving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docbutest
 
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural Network
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural NetworkTargeted Visual Content Recognition Using Multi-Layer Perceptron Neural Network
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural Networkijceronline
 
Performance Analysis of Various Activation Functions in Generalized MLP Archi...
Performance Analysis of Various Activation Functions in Generalized MLP Archi...Performance Analysis of Various Activation Functions in Generalized MLP Archi...
Performance Analysis of Various Activation Functions in Generalized MLP Archi...Waqas Tariq
 
Data Science - Part XVII - Deep Learning & Image Processing
Data Science - Part XVII - Deep Learning & Image ProcessingData Science - Part XVII - Deep Learning & Image Processing
Data Science - Part XVII - Deep Learning & Image ProcessingDerek Kane
 

What's hot (18)

MATLAB Code + Description : Very Simple Automatic English Optical Character R...
MATLAB Code + Description : Very Simple Automatic English Optical Character R...MATLAB Code + Description : Very Simple Automatic English Optical Character R...
MATLAB Code + Description : Very Simple Automatic English Optical Character R...
 
Unsupervised Feature Learning
Unsupervised Feature LearningUnsupervised Feature Learning
Unsupervised Feature Learning
 
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
Piotr Mirowski - Review Autoencoders (Deep Learning) - CIUUK14
 
Predictive Text Embedding using LINE
Predictive Text Embedding using LINEPredictive Text Embedding using LINE
Predictive Text Embedding using LINE
 
IRJET- Image Classification – Cat and Dog Images
IRJET- Image Classification – Cat and Dog ImagesIRJET- Image Classification – Cat and Dog Images
IRJET- Image Classification – Cat and Dog Images
 
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...
Classification of Electroencephalograph (EEG) Signals Using Quantum Neural Ne...
 
A survey research summary on neural networks
A survey research summary on neural networksA survey research summary on neural networks
A survey research summary on neural networks
 
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable Hardware
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable HardwareGenetic Algorithm Processor for Image Noise Filtering Using Evolvable Hardware
Genetic Algorithm Processor for Image Noise Filtering Using Evolvable Hardware
 
Survey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique AlgorithmsSurvey on Artificial Neural Network Learning Technique Algorithms
Survey on Artificial Neural Network Learning Technique Algorithms
 
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
Optimization of Number of Neurons in the Hidden Layer in Feed Forward Neural ...
 
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLING
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLINGLOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLING
LOG MESSAGE ANOMALY DETECTION WITH OVERSAMPLING
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
 
E44082429
E44082429E44082429
E44082429
 
Improving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..docImproving Classifier Accuracy using Unlabeled Data..doc
Improving Classifier Accuracy using Unlabeled Data..doc
 
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural Network
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural NetworkTargeted Visual Content Recognition Using Multi-Layer Perceptron Neural Network
Targeted Visual Content Recognition Using Multi-Layer Perceptron Neural Network
 
Hs3613611366
Hs3613611366Hs3613611366
Hs3613611366
 
Performance Analysis of Various Activation Functions in Generalized MLP Archi...
Performance Analysis of Various Activation Functions in Generalized MLP Archi...Performance Analysis of Various Activation Functions in Generalized MLP Archi...
Performance Analysis of Various Activation Functions in Generalized MLP Archi...
 
Data Science - Part XVII - Deep Learning & Image Processing
Data Science - Part XVII - Deep Learning & Image ProcessingData Science - Part XVII - Deep Learning & Image Processing
Data Science - Part XVII - Deep Learning & Image Processing
 

Viewers also liked

Testing the TurboLAMP stack with Joomla!
Testing the TurboLAMP stack with Joomla!Testing the TurboLAMP stack with Joomla!
Testing the TurboLAMP stack with Joomla!JoomlaHost.it
 
"Aprende a diseñar (un poco) en 13 minutos" por @martuishere
"Aprende a diseñar (un poco) en 13 minutos" por @martuishere"Aprende a diseñar (un poco) en 13 minutos" por @martuishere
"Aprende a diseñar (un poco) en 13 minutos" por @martuisherewebcat
 
Elastic Web Mining
Elastic Web MiningElastic Web Mining
Elastic Web MiningKen Krugler
 
Get Invited to Speak at Conferences
Get Invited to Speak at ConferencesGet Invited to Speak at Conferences
Get Invited to Speak at Conferencessarahmilstein
 
Mobile Strategies Summit Boston
Mobile Strategies Summit Boston   Mobile Strategies Summit Boston
Mobile Strategies Summit Boston Michael Ricci
 
Google for Entrepreneurs
Google for EntrepreneursGoogle for Entrepreneurs
Google for EntrepreneursLaura Lorek
 
Tu cuenta de twitter en piloto automático
Tu cuenta de twitter en piloto automáticoTu cuenta de twitter en piloto automático
Tu cuenta de twitter en piloto automáticoCésar Torrell
 
eMapMail - The Mind Mapping solution for complex e-mail requirements
eMapMail - The Mind Mapping solution for complex e-mail requirementseMapMail - The Mind Mapping solution for complex e-mail requirements
eMapMail - The Mind Mapping solution for complex e-mail requirementsJosé M. Guerrero
 
Cloud Computing por German Realpe
Cloud Computing por German RealpeCloud Computing por German Realpe
Cloud Computing por German RealpeInterlat
 
Rafael Bagmanov «Scala in a wild enterprise»
Rafael Bagmanov «Scala in a wild enterprise»Rafael Bagmanov «Scala in a wild enterprise»
Rafael Bagmanov «Scala in a wild enterprise»e-Legion
 

Viewers also liked (11)

Testing the TurboLAMP stack with Joomla!
Testing the TurboLAMP stack with Joomla!Testing the TurboLAMP stack with Joomla!
Testing the TurboLAMP stack with Joomla!
 
"Aprende a diseñar (un poco) en 13 minutos" por @martuishere
"Aprende a diseñar (un poco) en 13 minutos" por @martuishere"Aprende a diseñar (un poco) en 13 minutos" por @martuishere
"Aprende a diseñar (un poco) en 13 minutos" por @martuishere
 
Elastic Web Mining
Elastic Web MiningElastic Web Mining
Elastic Web Mining
 
Get Invited to Speak at Conferences
Get Invited to Speak at ConferencesGet Invited to Speak at Conferences
Get Invited to Speak at Conferences
 
Mobile Strategies Summit Boston
Mobile Strategies Summit Boston   Mobile Strategies Summit Boston
Mobile Strategies Summit Boston
 
Google for Entrepreneurs
Google for EntrepreneursGoogle for Entrepreneurs
Google for Entrepreneurs
 
Facebook for Art
Facebook for ArtFacebook for Art
Facebook for Art
 
Tu cuenta de twitter en piloto automático
Tu cuenta de twitter en piloto automáticoTu cuenta de twitter en piloto automático
Tu cuenta de twitter en piloto automático
 
eMapMail - The Mind Mapping solution for complex e-mail requirements
eMapMail - The Mind Mapping solution for complex e-mail requirementseMapMail - The Mind Mapping solution for complex e-mail requirements
eMapMail - The Mind Mapping solution for complex e-mail requirements
 
Cloud Computing por German Realpe
Cloud Computing por German RealpeCloud Computing por German Realpe
Cloud Computing por German Realpe
 
Rafael Bagmanov «Scala in a wild enterprise»
Rafael Bagmanov «Scala in a wild enterprise»Rafael Bagmanov «Scala in a wild enterprise»
Rafael Bagmanov «Scala in a wild enterprise»
 

Similar to 11.comparison of hybrid pso sa algorithm and genetic algorithm for classification

Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsaciijournal
 
Neural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabNeural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabijcses
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithmsaciijournal
 
Classification Of Iris Plant Using Feedforward Neural Network
Classification Of Iris Plant Using Feedforward Neural NetworkClassification Of Iris Plant Using Feedforward Neural Network
Classification Of Iris Plant Using Feedforward Neural Networkirjes
 
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptxssuser67281d
 
NeuralProcessingofGeneralPurposeApproximatePrograms
NeuralProcessingofGeneralPurposeApproximateProgramsNeuralProcessingofGeneralPurposeApproximatePrograms
NeuralProcessingofGeneralPurposeApproximateProgramsMohid Nabil
 
Towards neuralprocessingofgeneralpurposeapproximateprograms
Towards neuralprocessingofgeneralpurposeapproximateprogramsTowards neuralprocessingofgeneralpurposeapproximateprograms
Towards neuralprocessingofgeneralpurposeapproximateprogramsParidha Saxena
 
Modeling of neural image compression using gradient decent technology
Modeling of neural image compression using gradient decent technologyModeling of neural image compression using gradient decent technology
Modeling of neural image compression using gradient decent technologytheijes
 
Feed forward neural network for sine
Feed forward neural network for sineFeed forward neural network for sine
Feed forward neural network for sineijcsa
 
Open CV Implementation of Object Recognition Using Artificial Neural Networks
Open CV Implementation of Object Recognition Using Artificial Neural NetworksOpen CV Implementation of Object Recognition Using Artificial Neural Networks
Open CV Implementation of Object Recognition Using Artificial Neural Networksijceronline
 
Predicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesPredicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesVarad Meru
 
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...IJMER
 
Final Semester Project
Final Semester ProjectFinal Semester Project
Final Semester ProjectDebraj Paul
 
Neural basics
Neural basicsNeural basics
Neural basicscoursesub
 
CCC-Bicluster Analysis for Time Series Gene Expression Data
CCC-Bicluster Analysis for Time Series Gene Expression DataCCC-Bicluster Analysis for Time Series Gene Expression Data
CCC-Bicluster Analysis for Time Series Gene Expression DataIRJET Journal
 
Fuzzy Logic Final Report
Fuzzy Logic Final ReportFuzzy Logic Final Report
Fuzzy Logic Final ReportShikhar Agarwal
 
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...cscpconf
 
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...IRJET Journal
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
 
Artifical Neural Network and its applications
Artifical Neural Network and its applicationsArtifical Neural Network and its applications
Artifical Neural Network and its applicationsSangeeta Tiwari
 

Similar to 11.comparison of hybrid pso sa algorithm and genetic algorithm for classification (20)

Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
 
Neural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlabNeural network based numerical digits recognization using nnt in matlab
Neural network based numerical digits recognization using nnt in matlab
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
 
Classification Of Iris Plant Using Feedforward Neural Network
Classification Of Iris Plant Using Feedforward Neural NetworkClassification Of Iris Plant Using Feedforward Neural Network
Classification Of Iris Plant Using Feedforward Neural Network
 
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx
2. NEURAL NETWORKS USING GENETIC ALGORITHMS.pptx
 
NeuralProcessingofGeneralPurposeApproximatePrograms
NeuralProcessingofGeneralPurposeApproximateProgramsNeuralProcessingofGeneralPurposeApproximatePrograms
NeuralProcessingofGeneralPurposeApproximatePrograms
 
Towards neuralprocessingofgeneralpurposeapproximateprograms
Towards neuralprocessingofgeneralpurposeapproximateprogramsTowards neuralprocessingofgeneralpurposeapproximateprograms
Towards neuralprocessingofgeneralpurposeapproximateprograms
 
Modeling of neural image compression using gradient decent technology
Modeling of neural image compression using gradient decent technologyModeling of neural image compression using gradient decent technology
Modeling of neural image compression using gradient decent technology
 
Feed forward neural network for sine
Feed forward neural network for sineFeed forward neural network for sine
Feed forward neural network for sine
 
Open CV Implementation of Object Recognition Using Artificial Neural Networks
Open CV Implementation of Object Recognition Using Artificial Neural NetworksOpen CV Implementation of Object Recognition Using Artificial Neural Networks
Open CV Implementation of Object Recognition Using Artificial Neural Networks
 
Predicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensemblesPredicting rainfall using ensemble of ensembles
Predicting rainfall using ensemble of ensembles
 
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
Artificial Neural Networks (ANNS) For Prediction of California Bearing Ratio ...
 
Final Semester Project
Final Semester ProjectFinal Semester Project
Final Semester Project
 
Neural basics
Neural basicsNeural basics
Neural basics
 
CCC-Bicluster Analysis for Time Series Gene Expression Data
CCC-Bicluster Analysis for Time Series Gene Expression DataCCC-Bicluster Analysis for Time Series Gene Expression Data
CCC-Bicluster Analysis for Time Series Gene Expression Data
 
Fuzzy Logic Final Report
Fuzzy Logic Final ReportFuzzy Logic Final Report
Fuzzy Logic Final Report
 
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...
Employing Neocognitron Neural Network Base Ensemble Classifiers To Enhance Ef...
 
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
IRJET-Performance Enhancement in Machine Learning System using Hybrid Bee Col...
 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
 
Artifical Neural Network and its applications
Artifical Neural Network and its applicationsArtifical Neural Network and its applications
Artifical Neural Network and its applications
 

More from Alexander Decker

Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...Alexander Decker
 
A validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale inA validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale inAlexander Decker
 
A usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websitesA usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websitesAlexander Decker
 
A universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banksA universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banksAlexander Decker
 
A unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized dA unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized dAlexander Decker
 
A trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistanceA trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistanceAlexander Decker
 
A transformational generative approach towards understanding al-istifham
A transformational  generative approach towards understanding al-istifhamA transformational  generative approach towards understanding al-istifham
A transformational generative approach towards understanding al-istifhamAlexander Decker
 
A time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibiaA time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibiaAlexander Decker
 
A therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school childrenA therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school childrenAlexander Decker
 
A theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banksA theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banksAlexander Decker
 
A systematic evaluation of link budget for
A systematic evaluation of link budget forA systematic evaluation of link budget for
A systematic evaluation of link budget forAlexander Decker
 
A synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjabA synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjabAlexander Decker
 
A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...Alexander Decker
 
A survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incrementalA survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incrementalAlexander Decker
 
A survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniquesA survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniquesAlexander Decker
 
A survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo dbA survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo dbAlexander Decker
 
A survey on challenges to the media cloud
A survey on challenges to the media cloudA survey on challenges to the media cloud
A survey on challenges to the media cloudAlexander Decker
 
A survey of provenance leveraged
A survey of provenance leveragedA survey of provenance leveraged
A survey of provenance leveragedAlexander Decker
 
A survey of private equity investments in kenya
A survey of private equity investments in kenyaA survey of private equity investments in kenya
A survey of private equity investments in kenyaAlexander Decker
 
A study to measures the financial health of
A study to measures the financial health ofA study to measures the financial health of
A study to measures the financial health ofAlexander Decker
 

More from Alexander Decker (20)

Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...Abnormalities of hormones and inflammatory cytokines in women affected with p...
Abnormalities of hormones and inflammatory cytokines in women affected with p...
 
A validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale inA validation of the adverse childhood experiences scale in
A validation of the adverse childhood experiences scale in
 
A usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websitesA usability evaluation framework for b2 c e commerce websites
A usability evaluation framework for b2 c e commerce websites
 
A universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banksA universal model for managing the marketing executives in nigerian banks
A universal model for managing the marketing executives in nigerian banks
 
A unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized dA unique common fixed point theorems in generalized d
A unique common fixed point theorems in generalized d
 
A trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistanceA trends of salmonella and antibiotic resistance
A trends of salmonella and antibiotic resistance
 
A transformational generative approach towards understanding al-istifham
A transformational  generative approach towards understanding al-istifhamA transformational  generative approach towards understanding al-istifham
A transformational generative approach towards understanding al-istifham
 
A time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibiaA time series analysis of the determinants of savings in namibia
A time series analysis of the determinants of savings in namibia
 
A therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school childrenA therapy for physical and mental fitness of school children
A therapy for physical and mental fitness of school children
 
A theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banksA theory of efficiency for managing the marketing executives in nigerian banks
A theory of efficiency for managing the marketing executives in nigerian banks
 
A systematic evaluation of link budget for
A systematic evaluation of link budget forA systematic evaluation of link budget for
A systematic evaluation of link budget for
 
A synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjabA synthetic review of contraceptive supplies in punjab
A synthetic review of contraceptive supplies in punjab
 
A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...A synthesis of taylor’s and fayol’s management approaches for managing market...
A synthesis of taylor’s and fayol’s management approaches for managing market...
 
A survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incrementalA survey paper on sequence pattern mining with incremental
A survey paper on sequence pattern mining with incremental
 
A survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniquesA survey on live virtual machine migrations and its techniques
A survey on live virtual machine migrations and its techniques
 
A survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo dbA survey on data mining and analysis in hadoop and mongo db
A survey on data mining and analysis in hadoop and mongo db
 
A survey on challenges to the media cloud
A survey on challenges to the media cloudA survey on challenges to the media cloud
A survey on challenges to the media cloud
 
A survey of provenance leveraged
A survey of provenance leveragedA survey of provenance leveraged
A survey of provenance leveraged
 
A survey of private equity investments in kenya
A survey of private equity investments in kenyaA survey of private equity investments in kenya
A survey of private equity investments in kenya
 
A study to measures the financial health of
A study to measures the financial health ofA study to measures the financial health of
A study to measures the financial health of
 

Recently uploaded

Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machinePadma Pradeep
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitecturePixlogix Infotech
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...shyamraj55
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Alan Dix
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDGMarianaLemus7
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationRidwan Fadjar
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptxLBM Solutions
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024Scott Keck-Warren
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...Fwdays
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphNeo4j
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Scott Keck-Warren
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsRizwan Syed
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsMemoori
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubKalema Edgar
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationSlibray Presentation
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsMark Billinghurst
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Mattias Andersson
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024BookNet Canada
 

Recently uploaded (20)

Install Stable Diffusion in windows machine
Install Stable Diffusion in windows machineInstall Stable Diffusion in windows machine
Install Stable Diffusion in windows machine
 
Understanding the Laravel MVC Architecture
Understanding the Laravel MVC ArchitectureUnderstanding the Laravel MVC Architecture
Understanding the Laravel MVC Architecture
 
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
Automating Business Process via MuleSoft Composer | Bangalore MuleSoft Meetup...
 
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...Swan(sea) Song – personal research during my six years at Swansea ... and bey...
Swan(sea) Song – personal research during my six years at Swansea ... and bey...
 
APIForce Zurich 5 April Automation LPDG
APIForce Zurich 5 April  Automation LPDGAPIForce Zurich 5 April  Automation LPDG
APIForce Zurich 5 April Automation LPDG
 
My Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 PresentationMy Hashitalk Indonesia April 2024 Presentation
My Hashitalk Indonesia April 2024 Presentation
 
Key Features Of Token Development (1).pptx
Key  Features Of Token  Development (1).pptxKey  Features Of Token  Development (1).pptx
Key Features Of Token Development (1).pptx
 
SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024SQL Database Design For Developers at php[tek] 2024
SQL Database Design For Developers at php[tek] 2024
 
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks..."LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
"LLMs for Python Engineers: Advanced Data Analysis and Semantic Kernel",Oleks...
 
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge GraphSIEMENS: RAPUNZEL – A Tale About Knowledge Graph
SIEMENS: RAPUNZEL – A Tale About Knowledge Graph
 
Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024Advanced Test Driven-Development @ php[tek] 2024
Advanced Test Driven-Development @ php[tek] 2024
 
Scanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL CertsScanning the Internet for External Cloud Exposures via SSL Certs
Scanning the Internet for External Cloud Exposures via SSL Certs
 
AI as an Interface for Commercial Buildings
AI as an Interface for Commercial BuildingsAI as an Interface for Commercial Buildings
AI as an Interface for Commercial Buildings
 
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptxE-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
E-Vehicle_Hacking_by_Parul Sharma_null_owasp.pptx
 
Unleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding ClubUnleash Your Potential - Namagunga Girls Coding Club
Unleash Your Potential - Namagunga Girls Coding Club
 
Connect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck PresentationConnect Wave/ connectwave Pitch Deck Presentation
Connect Wave/ connectwave Pitch Deck Presentation
 
The transition to renewables in India.pdf
The transition to renewables in India.pdfThe transition to renewables in India.pdf
The transition to renewables in India.pdf
 
Human Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR SystemsHuman Factors of XR: Using Human Factors to Design XR Systems
Human Factors of XR: Using Human Factors to Design XR Systems
 
Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?Are Multi-Cloud and Serverless Good or Bad?
Are Multi-Cloud and Serverless Good or Bad?
 
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
#StandardsGoals for 2024: What’s new for BISAC - Tech Forum 2024
 

11.comparison of hybrid pso sa algorithm and genetic algorithm for classification

  • 1. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 Comparison of Hybrid PSO-SA Algorithm and Genetic Algorithm for Classification S. G. Sanjeevi1* A. Naga Nikhila2 Thaseem Khan3 G. Sumathi4 1. Associate Professor, Dept. of Comp. Science & Engg., National Institute of Technology, Warangal, A.P., India 2. Dept. of Comp. Science & Engg., National Institute of Technology, Warangal, A.P., India 3. Dept. of Comp. Science & Engg.,National Institute of Technology, Warangal, A.P., India 4. Dept. of Comp. Science & Engg., National Institute of Technology, Warangal, A.P., India * E-mail of the corresponding author: sgsanjeevi@yahoo.com Abstract In this work, we propose and present a Hybrid particle swarm optimization-Simulated annealing algorithm and compare it with a Genetic algorithm for training respectively neural networks of identical architectures. These neural networks were then tested on a classification task. In particle swarm optimization, behavior of a particle is influenced by the experiential knowledge of the particle as well as socially exchanged information. Particle swarm optimization follows a parallel search strategy. In simulated annealing uphill moves are made in the search space in a stochastic fashion in addition to the downhill moves. Simulated annealing therefore has better scope of escaping local minima and reach a global minimum in the search space. Thus simulated annealing gives a selective randomness to the search. Genetic algorithm performs parallel and randomized search. The goal of training the neural network is to minimize the sum of the squares of the error between the target and observed output values for all the training samples and to deliver good test performance on the test inputs. We compared the performance of the neural networks of identical architectures trained by the Hybrid particle swarm optimization-simulated annealing and Genetic algorithm respectively on a classification task and noted the results obtained. Neural network trained by Hybrid particle swarm optimization-simulated annealing has given better results compared to the neural network trained by the Genetic algorithm in the tests conducted by us. Keywords: Classification, Hybrid particle swarm optimization-Simulated annealing, Simulated Annealing, Genetic algorithm, Neural Network etc. 1. Introduction Classification is an important activity of machine learning. Various algorithms are conventionally used for classification task namely, Decision tree learning using ID3[1], Concept learning using Candidate elimination [2], Neural networks [3], Naïve Bayes classifier [4] are some of the traditional methods used for classification. Ever since back propagation algorithm was invented and popularized by [3], [5] and [6] neural networks were actively used for classification. However, since back-propagation method follows hill climbing approach, it is susceptible to occurrence of local minima. Hence we examine the use of alternative methods for training neural networks. We examine the use of i) Hybrid particle swarm optimization- simulated annealing algorithm ii) Genetic algorithm to train the neural networks. We study and compare the performance of the neural networks trained by these algorithms on a classification task. 2. Architecture of Neural Network Neural network designed for the classification task has the following architecture. It has four input units, three hidden units and three output units in the input layer, hidden layer and output layer respectively. Sigmoid activation functions were used with hidden and output units. Figure 1 shows the architectural diagram of the neural network. Neurons are connected in the feed forward fashion as shown. Neural 37
  • 2. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 Network has 12 weights between input and hidden layer and 9 weights between hidden and output layer. So the total number of weights in the network are 21. Input Hidden Output Figure 1 Architecture of Neural Network used 2.1 .Iris data Neural Network shown in figure 1 is used to perform classification task on IRIS data. The data was taken from the Univ. of California, Irvine (UCI), Machine learning repository. Iris data consists of 150 input- output vector pairs. Each input vector consists of a 4 tuple having four attribute values corresponding to the four input attributes respectively. Based on the input vector, output vector gives class to which it belongs. Each output vector is a 3 tuple and will have a ‘1’ in first, second or third positions and zeros in rest two positions, thereby indicating the class to which the input vector being considered belongs. Hence, we use 1- of-n encoding on the output side for denoting the class value. The data of 150 input-output pairs is divided randomly into two parts to create the training set and test set respectively. Data from training set is used to train the neural network and data from the test set is used for test purposes. Few samples of IRIS data are shown in table 1. Table 1.Sample Of Iris Data Used. S. No. Attr. 1 Attr. 2 Attr. 3 Attr. 4 Class 1 Class 2 Class 3 1 0.224 0.624 0.067 0.043 1 0 0 2 0.749 0.502 0.627 0.541 0 1 0 3 0.557 0.541 0.847 1 0 0 1 4 0.11 0.502 0.051 0.043 1 0 0 5 0.722 0.459 0.663 0.584 0 1 0 3. Genetic Algorithm to Train the Neural Network Error back propagation algorithm is conventionally used to train the neural networks. Error-Back propagation algorithm uses gradient descent search which is based on the concept of hill climbing. Main disadvantage of neural network using back propagation algorithm is that since it uses hill climbing approach it can get stuck at local minima. Hence, here we explore the usage of alternative algorithms instead of conventional backpropagation algorithm to optimize the performance of training a neural network and compare them. The objective function required to be minimized for training the neural network is the error function E. We define the error function E as follows. The error function E is given by 38
  • 3. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 where x is a specific training example and X is the set of all training examples, denotes the target output for the kth output neuron corresponding to training sample x, denotes the observed output for the kth output neuron corresponding to training sample x. Hence, error function computed is a measure of the sum of the squares of the error between target and observed output values for all the output neurons, across all the training samples. Genetic algorithms (GAs) were proposed and enunciated by John Holland at the University of Michigan and popularized by David Goldberg [9, 10]. They have the characteristics of parallel search being undertaken through potential solutions with a random exploration of the search space. They use the evolutionary operators of selection, crossover and mutation on the population of the potential solutions. Unlike backpropagation, GAs are stochastic and hence do not get into the problem of local minimum. Genetic algorithms can be used for optimization. In the problem we have to optimize the error function defined in equation (1) by minimizing it. Hence, the objective function we use is the error function in equation (1) which needs to be minimized. To do this we define the fitness function f as A – E where A is a positive number which is appropriately selected to be sufficiently large so that A – E is positive during the search task undertaken through GA. Task of minimizing the error function is converted to maximizing the value of A – E which is the fitness function f used with the genetic algorithm during the training of the neural network. We are using the genetic algorithm for the purpose of efficiently training the neural network. The idea of efficiently training the neural network is achieved if the neural network predicts with higher accuracy the test outputs corresponding to the test inputs given to the network. GAs use set of binary strings called chromosomes to form a population of binary strings. Each chromosome which is a binary string is a potential solution to the network. Hence, each chromosome needs to encode the set of weights present in a neural network. The neural network shown in figure 1 has 21 weights in it. Hence each member of the population (chromosome) needs to encode these complete set of 21 weights so that we can evaluate the fitness function defined above. Below we describe the encoding of the weights for the neural network shown in figure 1 for each chromosome. 3.1.Encoding of the Weights in the Neural Network The Neural Network in figure 1 has 12 weights between input and hidden layer and 9 weights between hidden and output layer. So the total number of weights in the network are 21. We coded these weights in binary form. Each weight was assumed to vary between +12.75 and -12.75 with a precision of 0.05. This is because weights learned by a typical neural network will be small positive or negative real numbers. Any real number between -12.75 and +12.75 with a precision of 0.05 can be represented by a 9-bit binary string. Thus, -12.75 was represented with a binary string ‘000000000’ and +12.75 was represented with ‘111111111’. For example, number 0.1 is represented by ‘100000010’. Each of the twenty one weights present in the neural network was represented by a separate 9 bit binary string and the set of all the weights present in the neural network is represented by their concatenation having 189 bits binary string. These 189 bits represent one possible assignment of values to the weights in the neural network. In the initial population a set of 50 randomly generated binary strings were created, each binary string having 189 bits. Different binary strings in population represent different potential solutions to the problem of training the neural network. Each binary string represents one possible assignment of weights to the neural network. 3.2. Genetic Algorithm Here we describe the Genetic algorithm used for training the neural network. 1. A set of 50 binary strings, each string having 189 bits were randomly created to form the initial 39
  • 4. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 population. 2. Evaluate the fitness function f for each of the binary strings in the present population P. 3. Elitism: Two of the binary strings with highest fitness values from the population are carried over to the new population being created Pnew. 4. Selection: The selection operation selects the parent strings from the present population for the crossover operation. This is done using the selection probability pi = fi /Σ fi . The roulette wheel concept is used to make the selection of binary strings. The probability pi is determined as the ratio of fitness of a given binary string with the sum of fitness values for all the 50 binary strings in the population P. 5. Crossover: Pairs of chromosomes were selected from population P using selection probability defined above and a single point crossover is performed on the pair to form two offspring. Selection probabilities encourage the selection of high fitness individuals to be selected as a pair. Crossover point was chosen at random and the two strings are interchanged at the point. This process was repeated for each of the 24 pairs of strings selected. Selection of each pair was done using selection probability calculations as explained above. Strings generated after crossover are kept in the new population being created Pnew. 6. Mutation: Mutation involves flipping a randomly selected bit. Mutation was done with a probability of 5% on the strings in Pnew . In the string selected for mutation, a randomly selected bit is flipped. 7. On completing above step the new population Pnew is formed. Make the new population created Pnew as the present population P for the next iteration of GA ie. Set P ←Pnew. Repeat steps 2 to 7 for sufficient number of iterations. Steps 2 to 7 were repeated for 15000 iterations in our work. The string with the highest fitness among the population is returned as the solution and gives the assignment of weights to the neural network after training through the Genetic Algorithm. 4. Hybrid PSO - SA Algorithm In particle swarm optimization [11, 12] a swarm of particles are flown through a multidimensional search space. Position of each particle represents a potential solution. Position of each particle is changed by adding a velocity vector to it. Velocity vector is influenced by the experiential knowledge of the particle as well as socially exchanged information. The experiential knowledge of a particle A describes the distance of the particle A from its own best position since the particle A’s first time step. This best position of a particle is referred to as the personal best of the particle. The global best position in a swarm at a time t is the best position found in the swarm of particles at the time t. The socially exchanged information of a particle A describes the distance of a particle A from the global best position in the swarm at time t. The experiential knowledge and socially exchanged information are also referred to as cognitive and social components respectively. We propose and present here the hybrid PSO-SA algorithm in table 2. Table 2. Hybrid Particle Swarm Optimization-Simulated Annealing Algorithm Create a nx dimensional swarm of ns particles; repeat for each particle i = 1, . . .,ns do // yi denotes the personal best position of the particle i so far // set the personal best position if f (xi) <f (yi) then 40
  • 5. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 yi = xi ; end // ŷ denotes the global best of the swarm so far // set the global best position if f (yi) <f ( ŷ ) then ŷ = yi ; end end for each particle i = 1, . . ., ns do update the velocity vi of particle i using equation (2); vi(t+1) = vi(t) + c1 r1(t)[yi(t) – xi(t) ] + c2 r2(t)[ŷ(t) – xi(t)] (2) // where yi denotes the personal best position of the particle i // and ŷ denotes the global best position of the swarm // and xi denotes the present position vector of particle i. update the position using equation (3); bi(t) = xi (t) //storing present position xi(t+1) = xi(t) + vi (t+1) (3) // applying simulated annealing compute ∆E = (f (xi(t) - f(xi(t+1)) (4) // ∆E = (E value for the previous network before weight change) - (E value for the present network with changed configuration of weights). • if ∆E is positive then // new value of E is smaller and therefore better than previous value. Then // accept the new position and make it the current position else accept the new position xi(t+1) (even though it has a higher E value) with a probability p defined by p= (5) • Revise the temperature T as per schedule defined below. The temperature schedule followed is defined as follows: The starting temperature is taken as T =1050; Current temp = T/log(iterations+1); Current temperature was changed after every 100 iterations using above formula. end 41
  • 6. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 until stopping condition is true ; In equation (2) of table 2, vi ,yi and ŷ are vectors of nx dimensions. vij denotes scalar component of vi in dimension j. vij is calculated as shown in equation 6 where r1j and r2j are random values in the range [0,1] and c1 and c2 are learning factors chosen as c1 = c2 = 2. vij(t+1) = vij(t) + c1 r1j(t)[yij(t) – xij(t) ] + c2 r2j(t)[ŷj(t) – xij) (t)] (6) 4.1. Implementation details of Hybrid PSO-SA algorithm for training a neural network We describe here the implementation details of hybrid PSO-SA algorithm for training the neural network. There are 21 weights in the neural network shown in figure 1. Hybrid PSO-SA algorithm combines particle swarm optimization algorithm with simulated annealing approach in the context of training a neural network. The swarm is initialized with a population of 50 particles. Each particle has 21 weights. Each weight value corresponds to a position in a particular dimension of the particle. Since there are 21 weights for each particle, there are therefore 21 dimensions. Hence, position vector of each particle corresponds to a 21 dimensional weight vector. Position (weight) in each dimension is modified by adding velocity value to it in that dimension. Fitness function corresponding to a position xi , is f (xi) and it denotes the distance of position xi from the final solution. In the hybrid PSO-SA algorithm this fitness function f (xi) is defined as error function E(w). Error function E(w) is a function of the weights in the neural network as defined in equation (1). In the optimization problem of training neural networks fitness function which is error function needs to be minimized. Each particle’s velocity vector is updated by considering the personal best position of the particle, global best position of the entire swarm and the present position vector of the particle as shown in equation (2) of table 2. Velocity of a particle i in dimension j is calculated as shown in equation (6). Hybrid PSO-SA algorithm combines pso algorithm with simulated annealing approach. Each of the 50 particles in the swarm is associated with 21 weights present in the neural network. The error function E(w) which is a function of the weights in the neural network as defined in equation (1) is treated as the fitness function. Error E(w) ( fitness function) needs to be minimized. For each of the 50 particles in the swarm, the solution (position) given by the particle is accepted if the change in error function ∆E as defined in equation 4 is positive, since this indicates error function is reducing in the present iteration compared to the previous iteration value. If ∆Eis not positive then new position is accepted with a probability p given by formula in equation (5). This is implemented by generating a random number between 0 and 1. If the number generated is lesser than p then the new position is accepted, else the previous position of particle is retained without changing. Each particle’s personal best position and the global best position of the swarm are updated after each iteration. Hybrid PSO-SA algorithm was run for 15000 iterations. After 15000 iterations global best position gbest of the swarm is returned as the solution. Hence, stopping criterion for the algorithm was chosen as completion of 15000 iterations. Hybrid PSO-SA algorithm combines parallel search approach of PSO and selective random search and global search properties of simulated annealing [7, 8] and hence combines the advantages of both the approaches. 5. Experiments and Results We have trained the neural network with architecture shown in figure 1 with genetic algorithm described in section 3 on training set taken from the IRIS data. Training set was a subset of samples chosen randomly from the IRIS data. Remaining samples from IRIS data were included in the test set. Performance was observed on the test data set for predicting the class of each sample. Table 3. Results Of Testing For Neural Network Using Genetic Algorithm 42
  • 7. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 Sl. No Samples in Samples in Test Correct Misclassifications Training Set Set classifications 1 100 50 44 6 2 95 55 45 10 3 85 65 53 12 4 75 75 61 14 We have chosen similarly, training sets of varying sizes from the IRIS data and included each time the samples which were not selected for training set into test set. Neural network was trained by each of the training sets and tested the performance of the network on corresponding test sets. Training was done for 15000 iterations with each of the training sets using the genetic algorithm. Results are shown in table 3. We have also trained the neural network with architecture shown in figure 1 with Hybrid PSO-SA algorithm described in section 4, with each of the training sets chosen above in table 3 and tested the performance of the network on corresponding test sets. Training was done using Hybrid PSO-SA algorithm for 15000 iterations with each of the training sets. The results are shown in table 4. Table 4. Results Of Testing For Neural Network Using Hybrid PSO-SA Algorithm Sl.No Samples in Training Samples in Test Correct Misclassifications Set Set Classifications 1 100 50 47 3 2 95 55 51 4 3 85 65 61 4 4 75 75 70 5 Two neural networks with same architecture as shown in figure 1 were used for training and testing with the two algorithms of Hybrid PSO-SA algorithm and Genetic algorithm respectively. Training was performed for same number of 15000 iterations on each neural network. Same training and test sets were used for comparison of neural networks performance with both the algorithms. Results of experiments and testing point out that neural network trained with Hybrid PSO-SA algorithm gives better performance over neural network trained with the Genetic algorithm across the training and test sets used. 6. Conclusion Our objective was to compare the performance of feed-forward neural network trained with Hybrid PSO- SA algorithm with the neural network trained by Genetic algorithm. We have trained the neural networks with identical architectures with the Hybrid PSO-SA algorithm and the Genetic algorithm respectively. The task we have tested using neural networks trained separately using these two algorithms is the IRIS data classification. We found that neural network trained with Hybrid PSO-SA algorithm has given better 43
  • 8. Computer Engineering and Intelligent Systems www.iiste.org ISSN 2222-1719 (Paper) ISSN 2222-2863 (Online) Vol 3, No.2, 2012 classification performance among the two. Neural network trained with Hybrid PSO-SA algorithm combines parallel search approach of PSO and selective random search and global search properties of simulated annealing and hence combines the advantages of both the approaches. GA also uses parallel and randomized search. Hence we compared these two global search approaches. Hybrid PSO-SA algorithm has given better results for training a neural network than the Genetic algorithm. References A Quinlan, J.R.(1986). Induction of decision trees. Machine Learning, 1 (1), 81-106. Mitchell, T.M.(1977). Version spaces: A candidate elimination approach to rule learning. Fifth International Joint Conference on AI (pp.305-310). Cambridge, MA: MIT Press. Rumelhart, D.E., & McClelland, J.L.(1986). Parallel distributed processing: exploration in the microstructure of cognition(Vols.1 & 2). Cambridge, MA : MIT Press. Duda, R.O., & Hart, P.E.(1973). Pattern classification and scene analysis New York: John Wiley & Sons. Werbos, P. (1975). Beyond regression: New tools for prediction and analysis in the behavioral sciences (Ph.D. dissertation). Harvard University. Parker, D.(1985). Learning logic (MIT Technical Report TR-47).MIT Center for Research in Computational Economics and Management Science. Kirkpatrick, S., Gelatt,Jr., C.D., and M.P. Vecchi 1983. Optimization by simulated annealing.Science 220(4598). Russel, S., &Norvig, P. (1995). Artificial intelligence: A modern approach. Englewood Cliffs, NJ:Prentice- Hall. Mitchell, T.M., 1997.Machine Learning, New York: McGraw-Hill. K Goldberg, David E. (1989). Genetic Algorithms in Search Optimization and Machine Learning. Addison Wesley. p. 41.ISBN 0201157675. Kennedy, J., Eberhart, R., 1995. “Particle swarm optimization”, IEEE International Conference on Neural Networks, Perth, WA , (pp.1942 – 1948), Australia. Engelbrecht, A.P., 2007, Computational Intelligence, England: John Wiley & Sons. Sriram G. Sanjeevi has acquired B.E. (Electronics and Communication Engineering) from Osmania University, Hyderabad, India in 1981, M.Tech (Computer Science and Engineering) from M.I.T. Manipal, Mangalore University, India in 1991 and Ph.D (Computer Science and Engineering) from IIT, Bombay, India in 2009. He is currently Associate Professor & Head of Computer Science and Engineering Department, National Institute of Technology, Warangal, India. His research interests are Neural networks, Machine learning and Soft computing. A. N. Nikhila has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in 2011. Her research interests are Neural networks, Machine learning. Thaseem Khan has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in 2011. His research interests are Neural networks, Machine learning. G. Sumathi has acquired B.Tech (Computer Science and Engineering) from NIT Warangal, India in 2011. Her research interests are Neural networks, Machine learning. 44
  • 9. International Journals Call for Paper The IISTE, a U.S. publisher, is currently hosting the academic journals listed below. The peer review process of the following journals usually takes LESS THAN 14 business days and IISTE usually publishes a qualified article within 30 days. Authors should send their full paper to the following email address. More information can be found in the IISTE website : www.iiste.org Business, Economics, Finance and Management PAPER SUBMISSION EMAIL European Journal of Business and Management EJBM@iiste.org Research Journal of Finance and Accounting RJFA@iiste.org Journal of Economics and Sustainable Development JESD@iiste.org Information and Knowledge Management IKM@iiste.org Developing Country Studies DCS@iiste.org Industrial Engineering Letters IEL@iiste.org Physical Sciences, Mathematics and Chemistry PAPER SUBMISSION EMAIL Journal of Natural Sciences Research JNSR@iiste.org Chemistry and Materials Research CMR@iiste.org Mathematical Theory and Modeling MTM@iiste.org Advances in Physics Theories and Applications APTA@iiste.org Chemical and Process Engineering Research CPER@iiste.org Engineering, Technology and Systems PAPER SUBMISSION EMAIL Computer Engineering and Intelligent Systems CEIS@iiste.org Innovative Systems Design and Engineering ISDE@iiste.org Journal of Energy Technologies and Policy JETP@iiste.org Information and Knowledge Management IKM@iiste.org Control Theory and Informatics CTI@iiste.org Journal of Information Engineering and Applications JIEA@iiste.org Industrial Engineering Letters IEL@iiste.org Network and Complex Systems NCS@iiste.org Environment, Civil, Materials Sciences PAPER SUBMISSION EMAIL Journal of Environment and Earth Science JEES@iiste.org Civil and Environmental Research CER@iiste.org Journal of Natural Sciences Research JNSR@iiste.org Civil and Environmental Research CER@iiste.org Life Science, Food and Medical Sciences PAPER SUBMISSION EMAIL Journal of Natural Sciences Research JNSR@iiste.org Journal of Biology, Agriculture and Healthcare JBAH@iiste.org Food Science and Quality Management FSQM@iiste.org Chemistry and Materials Research CMR@iiste.org Education, and other Social Sciences PAPER SUBMISSION EMAIL Journal of Education and Practice JEP@iiste.org Journal of Law, Policy and Globalization JLPG@iiste.org Global knowledge sharing: New Media and Mass Communication NMMC@iiste.org EBSCO, Index Copernicus, Ulrich's Journal of Energy Technologies and Policy JETP@iiste.org Periodicals Directory, JournalTOCS, PKP Historical Research Letter HRL@iiste.org Open Archives Harvester, Bielefeld Academic Search Engine, Elektronische Public Policy and Administration Research PPAR@iiste.org Zeitschriftenbibliothek EZB, Open J-Gate, International Affairs and Global Strategy IAGS@iiste.org OCLC WorldCat, Universe Digtial Library , Research on Humanities and Social Sciences RHSS@iiste.org NewJour, Google Scholar. Developing Country Studies DCS@iiste.org IISTE is member of CrossRef. All journals Arts and Design Studies ADS@iiste.org have high IC Impact Factor Values (ICV).