SlideShare a Scribd company logo
1 of 3
Download to read offline
Methods of Combining Neural Networks and Genetic Algorithms
                                                 Talib S. Hussain
                                               Queen’s University
                                            hussain@qucis.queensu.ca

1. Introduction                                              method, then he can program the network structure
                                                             explicitly. However, if the problem is very complex or
          In the past decade, two areas of research which    has no known solution, the developer may not know
have become very popular are the fields of neural            what structure to give the network. To this end, most
networks (NNs) and genetic algorithms (GAs). Both are        neural network models include a learning rule which can
computational abstractions of biological information         change the network’s structure over the course of
processing systems, and both have captured the               training to arrive at a good final solution. Back-
imaginations of researchers all over the world. In           propagation is the most popular learning rule.
general, NNs are used as learning systems and GAs as
optimisation systems, but as many researchers have           1.2 Genetic Algorithms
discovered, they may be combined in a number of                        A variety of computational models based on
different ways resulting in highly successful adaptive       evolutionary processes have been proposed, and the
systems. In this tutorial, a summary will be given of        most popular models are those known as genetic
these combination methods. This summary is not meant         algorithms.      A genetic algorithm has four main
to be exhaustive, but rather to be indicative of the type    elements: the genetic code, a concise representation for
of research being conducted. For a more detailed             an individual solution; the population, a number of
discussion, see Yao (1993) and Schaffer et al. (1992).       individual solutions; the fitness function, an evaluation
          The tutorial is broken into three sections. In     of the usefulness of an individual; and the propagation
the first section, a brief introduction to the foundations   techniques, a set of methods for generating new
of neural networks and genetic algorithms is given. It is    individuals. The genetic algorithm works as follows.
assumed that the participants have a basic understanding     First, a population of individuals is generated by
of both fields, and this introduction is designed as a       randomly selecting different genes. The fitness of each
short refresher. In the second section, a variety of         individual is then evaluated, and the propagation
approaches to integrating NNs and GAs are presented.         techniques are applied to highly fit individuals to
In the final section, some of the key research issues are    generate a new population - the next generation. The
discussed.                                                   cycle of evaluate and propagate continues until a
                                                             satisfactory solution, hopefully optimal, is found.
1.1 Neural Networks                                                    In a typical genetic algorithm, the genetic code
          To set up the terminology for the rest of the      is a fixed-length bit string and the population is always a
paper, let us review the basics of a neural network. A       fixed size. The three most common propagation
neural network is a computational model consisting of a      techniques are elitism, mutation and crossover. In
number of connected elements, known as neurons. A            elitism, the exact individual survives into the next
neuron is a processing unit that receives input from         generation. In mutation, a new individual is created
outside the network and/or from other neurons, applies a     from an old one by changing a small number of
local transformation to that input, and provides a single    randomly selected bits in its gene. In crossover, a new
output signal which is passed on to other neurons and/or     individual is created from two old ones by randomly
outside the network. Each of the inputs is modified by a     selecting a split point in their genes are creating a new
value associated with the connection. This value is          gene with the left part from one parent and the right part
referred to as the connection strength, or weight, and       from another. In any genetic algorithm, the two key
roughly speaking, represents how much importance the         aspects are the genetic representation and the fitness
neuron attaches to that input source. The local              function. Together, these determine the type of problem
transformation is referred to as the activation function     which is being solved and the possible solutions which
and is usually sigmoidal in nature.                          may be generated.
          A typical neural network is capable of             2. Combining NNs and GAs
representing many functions, as proved by
Komolgorov’s Theorem, but finding the best network           2.1 Supportive and Collaborative
needed to solve a specific problem is a very open-ended              Researchers have combined NNs and GAs in a
problem. If the developer knows the exact solution           number of different ways. Schaffer et al. have noted
that these combinations can be classified into one of two     better convergence. Better still, since GAs are good at
general types - supportive combinations in which the          global search but inefficient at local finely tuned search,
NN and GA are applied sequentially, and collaborative         a hybrid approach combining GAs and gradient descent
combinations in which they are applied simultaneously.        are attractive. (Yao)
          In a supportive approach, the GA and the NN
are applied to two different stages of the problem. The       2.3 Evolution of architectures
most common combination is to use a GA to pre-                           In the second approach, the GA is used to
process the data set that is used to train a NN. For          select general structural parameters and the neural
instance, the GA may be used to reduce the                    learning is used separately to trained the network and
dimensionality of the data space by eliminating               determine its fitness. This includes evolution of both
redundant or unnecessary features. Supportive                 the topology (i.e., connectivity pattern) and activation
combinations are not highly interesting since the GA          functions of each node, although most work has
and NN are used very independently and either can             concentrated on the former and little has been done on
easily be replaced by an alternative technique. Some          the latter.
other possible combinations include: using a NN to                       In architecture evolution, the genetic code can
select the starting population for the GA; using a GA to      be either a direct or indirect encoding of the network’s
analyse the representations of a NN; and using a GA           topology. In a direct encoding, each connection is
and NN to solve the same problem and integrating their        explicitly represented (e.g., a matrix where 1 indicate
responses using a voting scheme. (Schaffer et al.)            the presence of a connection and 0 indicates no
          Alternatively, in a collaborative approach, the     connection).      In an indirect encoding, important
GA and NN are integrated into a single system in which        parameters of the network are represented and the
a population of neural networks is evolved. In other          details of the exact connectivity are left to
words, the goal of the system is to find the optimal          developmental rules (e.g., specify the number of hidden
neural network solution. Such collaborative approaches        nodes and assume full connectivity between layers).
are possible since neural network learning and genetic                   In both cases, the exact neural network is not
algorithms are both form of search. A neural network          specified since the weights are determined by the
learning rule performs a highly constrained search to         initialisation routine and the network’s learning
optimise the network’s structure, while a genetic             algorithm. Thus, the evaluation of a gene is noisy since
algorithm performs a very general population-based            it is dependent upon the evaluation of the trained
search to find an optimally fit gene. Both are examples       network, and the GA finds the best set of architectural
of biased search techniques, and “any algorithm that          parameters rather than the best neural network.
employs a bias to guide its future samples can be
mislead in a search space with the right structure. There     2.4 Evolution of learning rules
is always an Achilles heal.” (Schaffer et al, p. 4) The
                                                                        In the final approach, the GA is used similarly
primary reason researchers have looked at integrating
                                                              to the evolution of architecture, but a parametric
NNs and GAs is the belief that they may compensate for
                                                              representation of the network’s learning rule is also
each other’s search weaknesses.
                                                              encoded in the gene. The genetic coding of topology in
                                                              this case is generally indirect.
2.2 Evolution of Connection Weights
                                                                        Evolving learning rules does not refer simply
          A genetic algorithm can be applied to               to adapting learning algorithm parameters (e.g., learning
optimising a neural network in a variety of ways. Yao         rate, momentum, etc.) but to adapting the learning
has indicated three main approaches.- the evolution of        functions themselves. This is an area of research which
weights, the evolution of topology, and the evolution of      has received little attention. “The biggest problem here
learning rules. In each case, the GA’s genetic code           is how to encode the dynamic behaviour of a learning
varies highly.                                                rule into static genotypes. Trying to develop a universal
          In the first, the GA is used as the learning rule   representation scheme which can specify any kind of
of the NN. The genetic code is a direct encoding of the       dynamic behaviours is clearly impractical let alone the
neural network, with each weight being represented            prohibitive long computation time required to search
explicitly. The population of the GA are all NNs with         such a learning rule space.” (Yao, p. 214)
the same basic topology, but with different weight
values. Mutation and crossover thus affect only the
                                                              3. Issues
weights of the individuals. A key question in such
system is whether to use binary weights or real-valued
ones - the latter increases the search space greatly.                  Collaborative combinations of NNs and GAs
          Using GAs instead of gradient descent               have sparked the interest of a great number of
algorithms to train the weights can result in faster and      researchers because of their obvious analogy to natural
systems.    A wide variety of systems have been               grammatical encoding has recently received some
developed and a number of research issues have been           attention. (Gruau, 1994) Grammar encoding is quite
considered.                                                   powerful since it is compact but can represent a great
                                                              range of networks.
3.1 The Baldwin Effect
          In general, one may wonder whether it really is     4. Conclusions
of any use to have both neural learning and genetic                    Neural networks and genetic algorithms are
search operating in the same system. Perhaps using just       two highly popular areas of research, and integrating
genetic search would work given enough time, or               both techniques can often lead to highly successful
perhaps a very general neural learning technique would        learning systems. The participants of this tutorial are
be sufficiently powerful. This is quite possibly true, but    encouraged to try applying evolutionary neural network
an observation from natural systems known as the              solutions, or even developing new combinations of their
Baldwin Effect provide a clearer answer.                      own.
          The Baldwin Effect states that in an
evolutionary system, successful genes can propagate           References
faster, and in some cases only, if the individuals are
capable of learning. This principle has been clearly          French, R. & Messinger, A. (1994). “Genes, phenes and
demonstrated in an artificial evolutionary system by                   the Baldwin Effect: Learning and evolution in
French & Messinger (1994). Thus, an evolutionary                       a simulated population,” Artificial Life IV,
system with simple individuals which can learn is                      277-282.
generally more successful than one with non-learning          Gruau, F. (1994) “Automatic definition of modular
individuals and probably also better than a single highly              neural networks,” Adaptive Behaviour, 3, 151-
complex learning individual.                                           184.
                                                              Schaffer, D., Whitley, D. & Eshelman, L. (1992)
3.2 Generalisation                                                     “Combinations of Genetic Algorithms and
          In evolving a neural network, attention must be
                                                                       Neural Networks: A survey of the state of the
paid to the trade-off between evolutionary fitness and
                                                                       art,” Proceedings of the International
generalisation ability. In many tasks, the final network
                                                                       Workshop on Combinations of Genetic
is trained on a small set of data and applied to a much
                                                                       Algorithms and Neural Networks. D. Whitley
larger set of data. The goal of the learning is actually to
                                                                       and D. Schaffer (Eds.,) Los Alamitos, CA:
develop a neural network with the best performance on
                                                                       IEEE Computer Society Press, 1-37.
the entire problem and not just the training data.
                                                              Yao, X. (1993) “Evolutionary artificial neural networks”
However, this can easily be overlooked during the
                                                                       International Journal of Neural Systems, 4,
development process.
                                                                       203-222.
          Thus, one must be careful when evolving
neural networks not to select for highly specialised,
poorly generalising networks. This is especially true in
problem areas which are highly dynamic.



3.3 Encoding Methods
          The two main properties of an encoding of a
neural network in a GA are its compactness and
representation capability. A compact encoding is useful
since the GA can then be efficiently applied to problems
requiring large NN solutions. An encoding should be
powerful enough to represent a large class of NNs or
else the GA may not generate very good solutions. For
instance, direct encoding is generally quite powerful in
representation, but not compact, while parameterised
encoding is compact, yet often represents a highly
restrictive set of structures.
          The discussion so far has focused on direct
encoding and parametric encoding of neural network
structure. Other possibilities also exist. In particular,

More Related Content

What's hot

Handwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTHandwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTRishabhTyagi48
 
Success Tips for Campus Recruitment
Success Tips for Campus RecruitmentSuccess Tips for Campus Recruitment
Success Tips for Campus RecruitmentRajendra Sabnis
 
Understanding Black Box Models with Shapley Values
Understanding Black Box Models with Shapley ValuesUnderstanding Black Box Models with Shapley Values
Understanding Black Box Models with Shapley ValuesJonathan Bechtel
 
Ml10 dimensionality reduction-and_advanced_topics
Ml10 dimensionality reduction-and_advanced_topicsMl10 dimensionality reduction-and_advanced_topics
Ml10 dimensionality reduction-and_advanced_topicsankit_ppt
 
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...Universitat Politècnica de Catalunya
 
User-­friendly Metaworkflows in Quantum Chemistry
User-­friendly Metaworkflows in Quantum ChemistryUser-­friendly Metaworkflows in Quantum Chemistry
User-­friendly Metaworkflows in Quantum ChemistrySandra Gesing
 
Data Mining : Concepts and Techniques
Data Mining : Concepts and TechniquesData Mining : Concepts and Techniques
Data Mining : Concepts and TechniquesDeepaR42
 
Introduction to Graph neural networks @ Vienna Deep Learning meetup
Introduction to Graph neural networks @  Vienna Deep Learning meetupIntroduction to Graph neural networks @  Vienna Deep Learning meetup
Introduction to Graph neural networks @ Vienna Deep Learning meetupLiad Magen
 
Density Based Clustering
Density Based ClusteringDensity Based Clustering
Density Based ClusteringSSA KPI
 
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...Analytics India Magazine
 
K nearest neighbor
K nearest neighborK nearest neighbor
K nearest neighborUjjawal
 
An introduction to Deep Learning
An introduction to Deep LearningAn introduction to Deep Learning
An introduction to Deep LearningJulien SIMON
 
Artificial Neural Network report
Artificial Neural Network reportArtificial Neural Network report
Artificial Neural Network reportAnjali Agrawal
 
Text clustering
Text clusteringText clustering
Text clusteringKU Leuven
 
CLUSTER SILHOUETTES.pptx
CLUSTER SILHOUETTES.pptxCLUSTER SILHOUETTES.pptx
CLUSTER SILHOUETTES.pptxagniva pradhan
 

What's hot (20)

Text analysis using python
Text analysis using pythonText analysis using python
Text analysis using python
 
Siamese networks.pptx.pdf
Siamese networks.pptx.pdfSiamese networks.pptx.pdf
Siamese networks.pptx.pdf
 
Handwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPTHandwritten Digit Recognition(Convolutional Neural Network) PPT
Handwritten Digit Recognition(Convolutional Neural Network) PPT
 
Success Tips for Campus Recruitment
Success Tips for Campus RecruitmentSuccess Tips for Campus Recruitment
Success Tips for Campus Recruitment
 
Understanding Black Box Models with Shapley Values
Understanding Black Box Models with Shapley ValuesUnderstanding Black Box Models with Shapley Values
Understanding Black Box Models with Shapley Values
 
Ml10 dimensionality reduction-and_advanced_topics
Ml10 dimensionality reduction-and_advanced_topicsMl10 dimensionality reduction-and_advanced_topics
Ml10 dimensionality reduction-and_advanced_topics
 
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
Interpretability of Convolutional Neural Networks - Eva Mohedano - UPC Barcel...
 
User-­friendly Metaworkflows in Quantum Chemistry
User-­friendly Metaworkflows in Quantum ChemistryUser-­friendly Metaworkflows in Quantum Chemistry
User-­friendly Metaworkflows in Quantum Chemistry
 
Data Mining : Concepts and Techniques
Data Mining : Concepts and TechniquesData Mining : Concepts and Techniques
Data Mining : Concepts and Techniques
 
cv
cvcv
cv
 
Random forest
Random forestRandom forest
Random forest
 
Introduction to Graph neural networks @ Vienna Deep Learning meetup
Introduction to Graph neural networks @  Vienna Deep Learning meetupIntroduction to Graph neural networks @  Vienna Deep Learning meetup
Introduction to Graph neural networks @ Vienna Deep Learning meetup
 
3 Centrality
3 Centrality3 Centrality
3 Centrality
 
Density Based Clustering
Density Based ClusteringDensity Based Clustering
Density Based Clustering
 
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...
Explainable deep learning with applications in Healthcare By Sunil Kumar Vupp...
 
K nearest neighbor
K nearest neighborK nearest neighbor
K nearest neighbor
 
An introduction to Deep Learning
An introduction to Deep LearningAn introduction to Deep Learning
An introduction to Deep Learning
 
Artificial Neural Network report
Artificial Neural Network reportArtificial Neural Network report
Artificial Neural Network report
 
Text clustering
Text clusteringText clustering
Text clustering
 
CLUSTER SILHOUETTES.pptx
CLUSTER SILHOUETTES.pptxCLUSTER SILHOUETTES.pptx
CLUSTER SILHOUETTES.pptx
 

Similar to Methods of Combining Neural Networks and Genetic Algorithms

Genome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniquesGenome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniqueseSAT Journals
 
Whale optimization mirjalili
Whale optimization mirjaliliWhale optimization mirjalili
Whale optimization mirjaliliPrashant Kumar
 
A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...eSAT Journals
 
A clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction ofA clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction ofeSAT Publishing House
 
Multilevel techniques for the clustering problem
Multilevel techniques for the clustering problemMultilevel techniques for the clustering problem
Multilevel techniques for the clustering problemcsandit
 
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...Sarvesh Kumar
 
Analysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String RecognitionAnalysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String RecognitionIDES Editor
 
Cornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 NetsCornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 NetsMark Gerstein
 
Capital market applications of neural networks etc
Capital market applications of neural networks etcCapital market applications of neural networks etc
Capital market applications of neural networks etc23tino
 
Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...BaoTramDuong2
 
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERINGA SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERINGijcsa
 
International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...IJCSEIT Journal
 
Literature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesLiterature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesIOSR Journals
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.docbutest
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.docbutest
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.docbutest
 
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...ijasuc
 

Similar to Methods of Combining Neural Networks and Genetic Algorithms (20)

Genome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniquesGenome structure prediction a review over soft computing techniques
Genome structure prediction a review over soft computing techniques
 
Whale optimization mirjalili
Whale optimization mirjaliliWhale optimization mirjalili
Whale optimization mirjalili
 
A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...A clonal based algorithm for the reconstruction of genetic network using s sy...
A clonal based algorithm for the reconstruction of genetic network using s sy...
 
A clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction ofA clonal based algorithm for the reconstruction of
A clonal based algorithm for the reconstruction of
 
G44083642
G44083642G44083642
G44083642
 
Multilevel techniques for the clustering problem
Multilevel techniques for the clustering problemMultilevel techniques for the clustering problem
Multilevel techniques for the clustering problem
 
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
Model of Differential Equation for Genetic Algorithm with Neural Network (GAN...
 
Analysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String RecognitionAnalysis of Neocognitron of Neural Network Method in the String Recognition
Analysis of Neocognitron of Neural Network Method in the String Recognition
 
Cornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 NetsCornell Pbsb 20090126 Nets
Cornell Pbsb 20090126 Nets
 
Capital market applications of neural networks etc
Capital market applications of neural networks etcCapital market applications of neural networks etc
Capital market applications of neural networks etc
 
Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...Discover How Scientific Data is Used for the Public Good with Natural Languag...
Discover How Scientific Data is Used for the Public Good with Natural Languag...
 
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERINGA SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
A SURVEY ON OPTIMIZATION APPROACHES TO TEXT DOCUMENT CLUSTERING
 
International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...International Journal of Computer Science, Engineering and Information Techno...
International Journal of Computer Science, Engineering and Information Techno...
 
Literature Survey On Clustering Techniques
Literature Survey On Clustering TechniquesLiterature Survey On Clustering Techniques
Literature Survey On Clustering Techniques
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
cec01.doc.doc
cec01.doc.doccec01.doc.doc
cec01.doc.doc
 
Ga 1 conference
Ga 1 conferenceGa 1 conference
Ga 1 conference
 
H43014046
H43014046H43014046
H43014046
 
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
New Generation Routing Protocol over Mobile Ad Hoc Wireless Networks based on...
 

More from ESCOM

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo SomESCOM
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales SomESCOM
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som SlidesESCOM
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som NetESCOM
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networksESCOM
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales KohonenESCOM
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia AdaptativaESCOM
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1ESCOM
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3ESCOM
 
Art2
Art2Art2
Art2ESCOM
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo ArtESCOM
 
Neocognitron
NeocognitronNeocognitron
NeocognitronESCOM
 
Neocognitron
NeocognitronNeocognitron
NeocognitronESCOM
 
Neocognitron
NeocognitronNeocognitron
NeocognitronESCOM
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima CognitronESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORKESCOM
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORKESCOM
 
Counterpropagation
CounterpropagationCounterpropagation
CounterpropagationESCOM
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPESCOM
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1ESCOM
 

More from ESCOM (20)

redes neuronales tipo Som
redes neuronales tipo Somredes neuronales tipo Som
redes neuronales tipo Som
 
redes neuronales Som
redes neuronales Somredes neuronales Som
redes neuronales Som
 
redes neuronales Som Slides
redes neuronales Som Slidesredes neuronales Som Slides
redes neuronales Som Slides
 
red neuronal Som Net
red neuronal Som Netred neuronal Som Net
red neuronal Som Net
 
Self Organinising neural networks
Self Organinising  neural networksSelf Organinising  neural networks
Self Organinising neural networks
 
redes neuronales Kohonen
redes neuronales Kohonenredes neuronales Kohonen
redes neuronales Kohonen
 
Teoria Resonancia Adaptativa
Teoria Resonancia AdaptativaTeoria Resonancia Adaptativa
Teoria Resonancia Adaptativa
 
ejemplo red neuronal Art1
ejemplo red neuronal Art1ejemplo red neuronal Art1
ejemplo red neuronal Art1
 
redes neuronales tipo Art3
redes neuronales tipo Art3redes neuronales tipo Art3
redes neuronales tipo Art3
 
Art2
Art2Art2
Art2
 
Redes neuronales tipo Art
Redes neuronales tipo ArtRedes neuronales tipo Art
Redes neuronales tipo Art
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Neocognitron
NeocognitronNeocognitron
Neocognitron
 
Fukushima Cognitron
Fukushima CognitronFukushima Cognitron
Fukushima Cognitron
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Counterpropagation
CounterpropagationCounterpropagation
Counterpropagation
 
Teoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAPTeoría de Resonancia Adaptativa Art2 ARTMAP
Teoría de Resonancia Adaptativa Art2 ARTMAP
 
Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1Teoría de Resonancia Adaptativa ART1
Teoría de Resonancia Adaptativa ART1
 

Recently uploaded

Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationnomboosow
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfAyushMahapatra5
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfciinovamais
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAssociation for Project Management
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfchloefrazer622
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...fonyou31
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDThiyagu K
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdfQucHHunhnh
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introductionMaksud Ahmed
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajanpragatimahajan3
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfAdmir Softic
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024Janet Corral
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 

Recently uploaded (20)

Interactive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communicationInteractive Powerpoint_How to Master effective communication
Interactive Powerpoint_How to Master effective communication
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Activity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdfActivity 01 - Artificial Culture (1).pdf
Activity 01 - Artificial Culture (1).pdf
 
APM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across SectorsAPM Welcome, APM North West Network Conference, Synergies Across Sectors
APM Welcome, APM North West Network Conference, Synergies Across Sectors
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
Disha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdfDisha NEET Physics Guide for classes 11 and 12.pdf
Disha NEET Physics Guide for classes 11 and 12.pdf
 
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
Ecosystem Interactions Class Discussion Presentation in Blue Green Lined Styl...
 
Measures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SDMeasures of Dispersion and Variability: Range, QD, AD and SD
Measures of Dispersion and Variability: Range, QD, AD and SD
 
1029 - Danh muc Sach Giao Khoa 10 . pdf
1029 -  Danh muc Sach Giao Khoa 10 . pdf1029 -  Danh muc Sach Giao Khoa 10 . pdf
1029 - Danh muc Sach Giao Khoa 10 . pdf
 
microwave assisted reaction. General introduction
microwave assisted reaction. General introductionmicrowave assisted reaction. General introduction
microwave assisted reaction. General introduction
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
social pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajansocial pharmacy d-pharm 1st year by Pragati K. Mahajan
social pharmacy d-pharm 1st year by Pragati K. Mahajan
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
General AI for Medical Educators April 2024
General AI for Medical Educators April 2024General AI for Medical Educators April 2024
General AI for Medical Educators April 2024
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 

Methods of Combining Neural Networks and Genetic Algorithms

  • 1. Methods of Combining Neural Networks and Genetic Algorithms Talib S. Hussain Queen’s University hussain@qucis.queensu.ca 1. Introduction method, then he can program the network structure explicitly. However, if the problem is very complex or In the past decade, two areas of research which has no known solution, the developer may not know have become very popular are the fields of neural what structure to give the network. To this end, most networks (NNs) and genetic algorithms (GAs). Both are neural network models include a learning rule which can computational abstractions of biological information change the network’s structure over the course of processing systems, and both have captured the training to arrive at a good final solution. Back- imaginations of researchers all over the world. In propagation is the most popular learning rule. general, NNs are used as learning systems and GAs as optimisation systems, but as many researchers have 1.2 Genetic Algorithms discovered, they may be combined in a number of A variety of computational models based on different ways resulting in highly successful adaptive evolutionary processes have been proposed, and the systems. In this tutorial, a summary will be given of most popular models are those known as genetic these combination methods. This summary is not meant algorithms. A genetic algorithm has four main to be exhaustive, but rather to be indicative of the type elements: the genetic code, a concise representation for of research being conducted. For a more detailed an individual solution; the population, a number of discussion, see Yao (1993) and Schaffer et al. (1992). individual solutions; the fitness function, an evaluation The tutorial is broken into three sections. In of the usefulness of an individual; and the propagation the first section, a brief introduction to the foundations techniques, a set of methods for generating new of neural networks and genetic algorithms is given. It is individuals. The genetic algorithm works as follows. assumed that the participants have a basic understanding First, a population of individuals is generated by of both fields, and this introduction is designed as a randomly selecting different genes. The fitness of each short refresher. In the second section, a variety of individual is then evaluated, and the propagation approaches to integrating NNs and GAs are presented. techniques are applied to highly fit individuals to In the final section, some of the key research issues are generate a new population - the next generation. The discussed. cycle of evaluate and propagate continues until a satisfactory solution, hopefully optimal, is found. 1.1 Neural Networks In a typical genetic algorithm, the genetic code To set up the terminology for the rest of the is a fixed-length bit string and the population is always a paper, let us review the basics of a neural network. A fixed size. The three most common propagation neural network is a computational model consisting of a techniques are elitism, mutation and crossover. In number of connected elements, known as neurons. A elitism, the exact individual survives into the next neuron is a processing unit that receives input from generation. In mutation, a new individual is created outside the network and/or from other neurons, applies a from an old one by changing a small number of local transformation to that input, and provides a single randomly selected bits in its gene. In crossover, a new output signal which is passed on to other neurons and/or individual is created from two old ones by randomly outside the network. Each of the inputs is modified by a selecting a split point in their genes are creating a new value associated with the connection. This value is gene with the left part from one parent and the right part referred to as the connection strength, or weight, and from another. In any genetic algorithm, the two key roughly speaking, represents how much importance the aspects are the genetic representation and the fitness neuron attaches to that input source. The local function. Together, these determine the type of problem transformation is referred to as the activation function which is being solved and the possible solutions which and is usually sigmoidal in nature. may be generated. A typical neural network is capable of 2. Combining NNs and GAs representing many functions, as proved by Komolgorov’s Theorem, but finding the best network 2.1 Supportive and Collaborative needed to solve a specific problem is a very open-ended Researchers have combined NNs and GAs in a problem. If the developer knows the exact solution number of different ways. Schaffer et al. have noted
  • 2. that these combinations can be classified into one of two better convergence. Better still, since GAs are good at general types - supportive combinations in which the global search but inefficient at local finely tuned search, NN and GA are applied sequentially, and collaborative a hybrid approach combining GAs and gradient descent combinations in which they are applied simultaneously. are attractive. (Yao) In a supportive approach, the GA and the NN are applied to two different stages of the problem. The 2.3 Evolution of architectures most common combination is to use a GA to pre- In the second approach, the GA is used to process the data set that is used to train a NN. For select general structural parameters and the neural instance, the GA may be used to reduce the learning is used separately to trained the network and dimensionality of the data space by eliminating determine its fitness. This includes evolution of both redundant or unnecessary features. Supportive the topology (i.e., connectivity pattern) and activation combinations are not highly interesting since the GA functions of each node, although most work has and NN are used very independently and either can concentrated on the former and little has been done on easily be replaced by an alternative technique. Some the latter. other possible combinations include: using a NN to In architecture evolution, the genetic code can select the starting population for the GA; using a GA to be either a direct or indirect encoding of the network’s analyse the representations of a NN; and using a GA topology. In a direct encoding, each connection is and NN to solve the same problem and integrating their explicitly represented (e.g., a matrix where 1 indicate responses using a voting scheme. (Schaffer et al.) the presence of a connection and 0 indicates no Alternatively, in a collaborative approach, the connection). In an indirect encoding, important GA and NN are integrated into a single system in which parameters of the network are represented and the a population of neural networks is evolved. In other details of the exact connectivity are left to words, the goal of the system is to find the optimal developmental rules (e.g., specify the number of hidden neural network solution. Such collaborative approaches nodes and assume full connectivity between layers). are possible since neural network learning and genetic In both cases, the exact neural network is not algorithms are both form of search. A neural network specified since the weights are determined by the learning rule performs a highly constrained search to initialisation routine and the network’s learning optimise the network’s structure, while a genetic algorithm. Thus, the evaluation of a gene is noisy since algorithm performs a very general population-based it is dependent upon the evaluation of the trained search to find an optimally fit gene. Both are examples network, and the GA finds the best set of architectural of biased search techniques, and “any algorithm that parameters rather than the best neural network. employs a bias to guide its future samples can be mislead in a search space with the right structure. There 2.4 Evolution of learning rules is always an Achilles heal.” (Schaffer et al, p. 4) The In the final approach, the GA is used similarly primary reason researchers have looked at integrating to the evolution of architecture, but a parametric NNs and GAs is the belief that they may compensate for representation of the network’s learning rule is also each other’s search weaknesses. encoded in the gene. The genetic coding of topology in this case is generally indirect. 2.2 Evolution of Connection Weights Evolving learning rules does not refer simply A genetic algorithm can be applied to to adapting learning algorithm parameters (e.g., learning optimising a neural network in a variety of ways. Yao rate, momentum, etc.) but to adapting the learning has indicated three main approaches.- the evolution of functions themselves. This is an area of research which weights, the evolution of topology, and the evolution of has received little attention. “The biggest problem here learning rules. In each case, the GA’s genetic code is how to encode the dynamic behaviour of a learning varies highly. rule into static genotypes. Trying to develop a universal In the first, the GA is used as the learning rule representation scheme which can specify any kind of of the NN. The genetic code is a direct encoding of the dynamic behaviours is clearly impractical let alone the neural network, with each weight being represented prohibitive long computation time required to search explicitly. The population of the GA are all NNs with such a learning rule space.” (Yao, p. 214) the same basic topology, but with different weight values. Mutation and crossover thus affect only the 3. Issues weights of the individuals. A key question in such system is whether to use binary weights or real-valued ones - the latter increases the search space greatly. Collaborative combinations of NNs and GAs Using GAs instead of gradient descent have sparked the interest of a great number of algorithms to train the weights can result in faster and researchers because of their obvious analogy to natural
  • 3. systems. A wide variety of systems have been grammatical encoding has recently received some developed and a number of research issues have been attention. (Gruau, 1994) Grammar encoding is quite considered. powerful since it is compact but can represent a great range of networks. 3.1 The Baldwin Effect In general, one may wonder whether it really is 4. Conclusions of any use to have both neural learning and genetic Neural networks and genetic algorithms are search operating in the same system. Perhaps using just two highly popular areas of research, and integrating genetic search would work given enough time, or both techniques can often lead to highly successful perhaps a very general neural learning technique would learning systems. The participants of this tutorial are be sufficiently powerful. This is quite possibly true, but encouraged to try applying evolutionary neural network an observation from natural systems known as the solutions, or even developing new combinations of their Baldwin Effect provide a clearer answer. own. The Baldwin Effect states that in an evolutionary system, successful genes can propagate References faster, and in some cases only, if the individuals are capable of learning. This principle has been clearly French, R. & Messinger, A. (1994). “Genes, phenes and demonstrated in an artificial evolutionary system by the Baldwin Effect: Learning and evolution in French & Messinger (1994). Thus, an evolutionary a simulated population,” Artificial Life IV, system with simple individuals which can learn is 277-282. generally more successful than one with non-learning Gruau, F. (1994) “Automatic definition of modular individuals and probably also better than a single highly neural networks,” Adaptive Behaviour, 3, 151- complex learning individual. 184. Schaffer, D., Whitley, D. & Eshelman, L. (1992) 3.2 Generalisation “Combinations of Genetic Algorithms and In evolving a neural network, attention must be Neural Networks: A survey of the state of the paid to the trade-off between evolutionary fitness and art,” Proceedings of the International generalisation ability. In many tasks, the final network Workshop on Combinations of Genetic is trained on a small set of data and applied to a much Algorithms and Neural Networks. D. Whitley larger set of data. The goal of the learning is actually to and D. Schaffer (Eds.,) Los Alamitos, CA: develop a neural network with the best performance on IEEE Computer Society Press, 1-37. the entire problem and not just the training data. Yao, X. (1993) “Evolutionary artificial neural networks” However, this can easily be overlooked during the International Journal of Neural Systems, 4, development process. 203-222. Thus, one must be careful when evolving neural networks not to select for highly specialised, poorly generalising networks. This is especially true in problem areas which are highly dynamic. 3.3 Encoding Methods The two main properties of an encoding of a neural network in a GA are its compactness and representation capability. A compact encoding is useful since the GA can then be efficiently applied to problems requiring large NN solutions. An encoding should be powerful enough to represent a large class of NNs or else the GA may not generate very good solutions. For instance, direct encoding is generally quite powerful in representation, but not compact, while parameterised encoding is compact, yet often represents a highly restrictive set of structures. The discussion so far has focused on direct encoding and parametric encoding of neural network structure. Other possibilities also exist. In particular,