SlideShare a Scribd company logo
1 of 12
Download to read offline
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 111
Performance Analysis of Various Activation Functions in
Generalized MLP Architectures of Neural Networks
Bekir Karlik bkarlik@mevlana.edu.tr
Engineering Faculty/Computer Engineering/
Mevlana University, Konya, 42003,
Turkey
A Vehbi Olgac avehbi@live.com
Engineering Faculty/Computer Engineering/
Fatih University, Istanbul, 34500,
Turkey
Abstract
The activation function used to transform the activation level of a unit (neuron) into an output
signal. There are a number of common activation functions in use with artificial neural networks
(ANN). The most common choice of activation functions for multi layered perceptron (MLP) is
used as transfer functions in research and engineering. Among the reasons for this popularity are
its boundedness in the unit interval, the function’s and its derivative’s fast computability, and a
number of amenable mathematical properties in the realm of approximation theory. However,
considering the huge variety of problem domains MLP is applied in, it is intriguing to suspect that
specific problems call for single or a set of specific activation functions. The aim of this study is to
analyze the performance of generalized MLP architectures which has back-propagation algorithm
using various different activation functions for the neurons of hidden and output layers. For
experimental comparisons, Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic Section, and Radial
Bases Function (RBF) were used.
Keywords: Activation Functions, Multi Layered Perceptron, Neural Networks, Performance Analysis
1. INTRODUCTION
One of the most attractive properties of ANNs is the possibility to adapt their behavior to the
changing characteristics of the modeled system. Last decades, many researchers have
investigated a variety of methods to improve ANN performance by optimizing training methods,
learn parameters, or network structure, comparably few works is done towards using activation
functions. Radial basis function (RBF) neural network is one of the most popular neural network
architectures [1]. The standard sigmoid reaches an approximation power comparable to or better
than classes of more established functions investigated in the approximation theory (i.e.,splines
and polynomials) [2]. Jordan presented the logistic function which is a natural representation of
the posterior probability in a binary classification problem [3]. Liu and Yao improved the structure
of Generalized Neural Networks (GNN) with two different activation function types which are
sigmoid and Gaussian basis functions [4]. Sopena et al. presented a number of experiments (with
widely–used benchmark problems) showing that multilayer feed–forward networks with a sine
activation function learn two orders of magnitude faster while generalization capacity increases
(compared to ANNs with logistic activation function) [5]. Dorffner developed the Conic Section
Function Neural Network (CSFNN) which is a unified framework for MLP and RBF networks to
make simultaneous use of advantages of both networks [6]. Bodyanskiy presented a novel
double-wavelet neuron architecture obtained by modification of standard wavelet neuron, and
their learning algorithms are proposed. The proposed architecture allows improving the
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 112
approximation properties of wavelet neuron [7]. All these well-known activations functions are
used into nodes of each layer of MLP to solve different non-linear problems. But, there is no
performance of comparison studies using these activation functions. So, in this study, we have
used five different well-known activation functions such as Bi-polar sigmoid, Uni-polar sigmoid,
Tanh, Conic Section, and Radial Bases Function (RBF) to compare their performances.
2. ACTIVATION FUNCTION TYPES
The most important unit in neural network structure is their net inputs by using a scalar-to-scalar
function called “the activation function or threshold function or transfer function”, output a result
value called the “unit's activation”. An activation function for limiting the amplitude of the output of
a neuron. Enabling in a limited range of functions is usually called squashing functions [8-9]. It
squashes the permissible amplitude range of the output signal to some finite value. Some of the
most commonly used activation functions are to solve non-linear problems. These functions are:
Uni-polar sigmoid, Bi-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF). We
did not care about some activation function such as identity function, step function or binary step
functions as they are not used to solve linear problems.
2.1 Uni-Polar Sigmoid Function
Activation function of Uni-polar sigmoid function is given as follows:
x
e
xg −
+
=
1
1
)( (1)
This function is especially advantageous to use in neural networks trained by back-propagation
algorithms. Because it is easy to distinguish, and this can interestingly minimize the computation
capacity for training. The term sigmoid means ‘S-shaped’, and logistic form of the sigmoid maps
the interval (-∞, ∞) onto (0, 1) as seen in
g(x)
0
1
-6 -4 -2 0 2 4 6
x
FIGURE 1.
g(x)
0
1
-6 -4 -2 0 2 4 6
x
FIGURE 1: Uni-Polar Sigmoid Function
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 113
2.2 Bipolar Sigmoid Function
Activation function of Bi-polar sigmoid function is given by
x
x
e
e
xg −
−
+
−
=
1
1
)( (2)
This function is similar to the sigmoid function. For this type of activation function described in Fig.
2, it goes well for applications that produce output values in the range of [-1, 1].
g(x)
-1
0
1
-6 -4 -2 0 2 4 6
x
FIGURE 2: Bi-Polar Sigmoid Function
2.3 Hyperbolic Tangent Function
This function is easily defined as the ratio between the hyperbolic sine and the cosine functions or
expanded as the ratio of the half‐difference and half‐sum of two exponential functions in the
points x and –x as follows :
xx
xx
ee
ee
x
x
x −
−
+
−
==
)cosh(
)sinh(
)tanh( (3)
Hyperbolic Tangent Function is similar to sigmoid function. Its range outputs between -1 and 1 as
seen in
FIGURE . The following is a graphic of the hyperbolic tangent function for real values of its
argument x:
FIGURE 3: Hyperbolic Tangent Function
2.4 Radial Basis Function
Radial basis function (RBF) is based on Gaussian Curve. It takes a parameter that determines
the center (mean) value of the function used as a desired value (see Fig. 4). A radial basis
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 114
function (RBF) is a real-valued function whose value depends only on the distance from the
origin, so that
)()( xgxg = (4)
or alternatively on the distance from some other point c, called a center, so that
)(),( cxgcxg −= (5)
Sums of radial basis functions are typically used to approximate given functions. This
approximation process can also be interpreted as a simple type of neural network.
RBF are typically used to build up function approximations of the form
)()(
1
i
N
i
i cxgwxy −= ∑=
(6)
where the approximating function y(x) is represented as a sum of N radial basis functions, each
associated with a different center ci, and weighted by an appropriate coefficient wi. The weights wi
can be estimated using the matrix methods of linear least squares, because the approximating
function is linear in the weights. Fig. 4 shows that two unnormalized Gaussian radial basis
functions in one input dimension. The basis function centers are located at c1=0.75 and c2=3.25
[10].
FIGURE 4: Two unnormalized Gaussian radial basis functions in one input dimension
RBF can also be interpreted as a rather simple single-layer type of artificial neural network called
a radial basis function network, with the radial basis functions taking on the role of the activation
functions of the network. It can be shown that any continuous function on a compact interval can
in principle be interpolated with arbitrary accuracy by a sum of this form, if a sufficiently large
number N of radial basis functions is used.
2.5 Conic Section Function
Conic section function (CSF) is based on a section of a cone as the name implies. CSF takes a
parameter that determines the angle value of the function as seen in Fig. 5.
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 115
Figure 5: Conic Section Parabola (ω = 90)
The equation of CSF can be defined as follows:
)(cos)()(
1
1
i
N
i
iiii cawwcaxf −−−= ∑
+
=
(7)
where ia is input coefficient, ci is the center, wi is the weight in Multi Layered Perceptron (MLP),
2w is an opening angle which can be any value in the range of [-π/2, π/2] and determines the
different forms of the decision borders.
The hidden units of neural network need activation functions to introduce non-linearity into the
networks. Non-linearity makes multi-layer networks so effective. The sigmoid functions are the
most widely used functions [10-11]. Activation functions should be chosen to be suited to the
distribution of the target values for the output units. You can think the same for binary outputs
where the tangent hyperbolic and sigmoid functions are effective choices. If the target values are
positive but have no known upper bound, an exponential output activation function can be used.
This work has explained all the variations on the parallel distributed processing idea of neural
networks. The structure of each neural network has very similar parts which perform the
processing.
3. COMPARISON WITH DIFFERENT ACTIVATION FUNCTIONS
In this study, different activation functions depend on different number of iterations for comparing
their performances by using the same data. For all the activation functions, we used the number
of nodes in the hidden layer; firstly 10 nodes, secondly 40 nodes (numbers of iterations are the
same for both of them). After presenting the graphs for different parameters from Figure 6
through Figure 10, interpretations of their results will follow right here.
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 116
FIGURE 6: 100 Iterations - 10 Hidden Neurons - Bi-Polar Sigmoid
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 117
FIGURE 7: 100 Iterations - 10 Hidden Neurons - Uni-Polar Sigmoid
FIGURE 8: 100 Iterations - 10 Hidden Neurons - Tangent Hyperbolic
We used the same number of hidden neurons (10 nodes) and iterations (100 iterations) to
compare the differences between the activation functions used above. It is also requires less
number of iterations.
FIGURE 9: 100 Iterations - 10 Hidden Neurons - Conic Section
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 118
FIGURE 10: 100 Iterations - 10 Hidden Neurons – RBF
According to the Figures from 6 through 15, we found Conic Section function the best activation
function for training. Moreover, it requires less number of iterations than the other activation
functions to solve the non-linear problems. However, with regard to testing, we found that the
accuracy of Tanh activation function was much better than CSF and the other activation
functions. This situation explains that total Mean Square Error (MSE) according to iterations
cannot determine the network accuracy. Hence, we can obtain the real accuracy results upon
testing.
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 119
FIGURE 11: 500 Iterations - 40 Hidden Neurons - Bi-Polar Sigmoid
FIGURE 12: 500 Iterations - 40 Hidden Neurons - Uni-Polar Sigmoid
Figures 11 through 15 show the error graphics for each activation function for different numbers
of nodes in the hidden layer which may vary from 10 up to 40, and the number of iterations from
100 up to 500.
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 120
FIGURE 13: 500 Iterations - 40 Hidden Neurons - Tangent Hyperbolic
FIGURE 14: 500 Iterations - 40 Hidden Neurons - Conic Section
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 121
FIGURE 15: 500 Iterations - 40 Hidden Neurons – RBF
According to the first five graphics generated by the same iteration numbers and the same
number of neurons in the hidden layer, the Tangent Hyperbolic Activation function resulted in the
most successful one for all the test parameters. Conic Section Function and Radial Basis
Function were not able to successfully use the data type and the network parameters. The
second group of result graphs is similar to the first group. But, the accuracy of the algorithms was
observed to increase for the sigmoid and tangent hyperbolic functions. The rest of the functions
used in this study were not successful and not accurate enough in this group. Table 1 shows
those total accuracy and error values for the testing phase.
100 Iterations
10 Hidden Neurons
500 Iterations
40 Hidden Neurons
Error Accuracy (%) Error Accuracy (%)
Bi-Polar S. 0,056 93 0,034 89
Uni-Polar S. 0,026 92 0,006 88
Tanh 0,025 95 0,002 99
Conic 0,001 34 0,045 23
RBF 0,003 30 0,001 19
TABLE 1: Results of the testing phase
4. CONCLUSIONS
In this study, we have used five conventional differentiable and monotonic activation functions for
the evolution of MLP architecture along with Generalized Delta rule learning. These proposed
well-known and effective activation functions are Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic
Section, and Radial Bases Function (RBF). Having compared their performances, simulation
results show that Tanh (hyperbolic tangent) function performs better recognition accuracy than
those of the other functions. In other words, the neural network computed good results when
“Tanh-Tanh” combination of activation functions was used for both neurons (or nodes) of hidden
and output layers.
Bekir Karlik and A. Vehbi Olgac
International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 122
We presented experimental results by proposing five different activation functions for MLP neural
networks architecture along with Generalized Delta Rule Learning to solve the non-linear
problems. These results demonstrate that it is possible to improve the ANN performance through
the use of much effective activation function. According to our experimental study, we can say
that the Tanh activation function can be used in the vast majority of MLP applications as a good
choice to obtain high accuracy. Furthermore, the sole use of validation test cannot approve the
work even if it would be able to predict very low MSE. So, we can only decide about the real
accuracy after a thorough testing of the neural network. In the future work, non–monotic
activation functions will probably be sought for the test results in the learning speed and the
performance of the neural network.
REFERENCES
1. T. Poggio, F. Girosi, “A theory of networks for approximation and learning”. A.I. Memo No.
1140, Artificial Intelligence, Laboratory, Massachusetts Institute of Technology, 1989
2. B DasGupta, G. Schnitger, “The Power of Approximating: A Comparison of Activation
Functions”. In Giles, C. L., Hanson, S. J., and Cowan, J. D., editors, Advances in Neural
Information Processing Systems, 5, pp. 615-622, San Mateo, CA. Morgan Kaufmann
Publishers, 1993
3. M.I. Jordan, “Why the logistic function? A tutorial discussion on probabilities and neural
networks”. Computational Cognitive Science Technical Report 9503, Massachusetts Institute
of Technology, 1995
4. Y. Liu, X. Yao, “Evolutionary Design of Artificial Neural Networks with Different Nodes”. In
Proceedings of the Third IEEE International Conference on Evolutionary Computation, pp.
570-675, 1996
5. J. M. Sopena, E. Romero, R. Alqu´ezar, “Neural networks with periodic and monotonic
activation functions: a comparative study in classification problems”. In Proceedings of the
9th International Conference on Artificial Neural Networks, pp. 323-328, 1999
6. G. Dorffner, “Unified frameworks for MLP and RBFNs: Introducing Conic Section Function
Networks”. Cybernetics and Systems, 25: 511-554, 1994
7. Y. Bodyanskiy, N. Lamonova, O. Vynokurova, “Double-Wavelet Neuron Based on Analytical
Activation Functions”. International Journal Information Theories & Applications, 14: 281-288,
2007
8. G. Cybenko, “Approximation by superposition of a sigmoidal function”. Mathematics of
Control, Signals, and Systems, 2(4):303–314, 1987
9. R. P. Lippmann, “An introduction to computing with neural nets”. IEEE Acoustics, Speech and
Signal Processing, 4(2):4–22, 1987
10. M.D. Buhmann, (2003), Radial Basis Functions: Theory and Implementations, Cambridge
University Press, ISBN 978-0-521-63338-3.
11. B. Widrow, M.A. Lehr, “30 years of adoptive neural netwoks; perceptron, madaline, and back
propagation”. Proc. IEEE, 78: 1415–1442, 1990
12. B. Ciocoiu, “Hybrid Feedforward Neural Networks for Solving Classification Problems”.
Neural Processing Letters, 16(1):81-91, 2002

More Related Content

What's hot

Neural networks1
Neural networks1Neural networks1
Neural networks1Mohan Raj
 
On The Application of Hyperbolic Activation Function in Computing the Acceler...
On The Application of Hyperbolic Activation Function in Computing the Acceler...On The Application of Hyperbolic Activation Function in Computing the Acceler...
On The Application of Hyperbolic Activation Function in Computing the Acceler...iosrjce
 
1.meena tushir finalpaper-1-12
1.meena tushir finalpaper-1-121.meena tushir finalpaper-1-12
1.meena tushir finalpaper-1-12Alexander Decker
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelIJECEIAES
 
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...Yuyun Wabula
 
Neural tool box
Neural tool boxNeural tool box
Neural tool boxMohan Raj
 
Segmentation of Images by using Fuzzy k-means clustering with ACO
Segmentation of Images by using Fuzzy k-means clustering with ACOSegmentation of Images by using Fuzzy k-means clustering with ACO
Segmentation of Images by using Fuzzy k-means clustering with ACOIJTET Journal
 
Real Time Face Detection on GPU Using OPENCL
Real Time Face Detection on GPU Using OPENCLReal Time Face Detection on GPU Using OPENCL
Real Time Face Detection on GPU Using OPENCLcsandit
 
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMLAI2
 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNNaren Chandra Kattla
 
Adaptive modified backpropagation algorithm based on differential errors
Adaptive modified backpropagation algorithm based on differential errorsAdaptive modified backpropagation algorithm based on differential errors
Adaptive modified backpropagation algorithm based on differential errorsIJCSEA Journal
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkNaren Chandra Kattla
 
Optimal neural network models for wind speed prediction
Optimal neural network models for wind speed predictionOptimal neural network models for wind speed prediction
Optimal neural network models for wind speed predictionIAEME Publication
 
Emergence of Invariance and Disentangling in Deep Representations
Emergence of Invariance and Disentangling in Deep RepresentationsEmergence of Invariance and Disentangling in Deep Representations
Emergence of Invariance and Disentangling in Deep RepresentationsSangwoo Mo
 

What's hot (17)

Neural networks1
Neural networks1Neural networks1
Neural networks1
 
On The Application of Hyperbolic Activation Function in Computing the Acceler...
On The Application of Hyperbolic Activation Function in Computing the Acceler...On The Application of Hyperbolic Activation Function in Computing the Acceler...
On The Application of Hyperbolic Activation Function in Computing the Acceler...
 
1.meena tushir finalpaper-1-12
1.meena tushir finalpaper-1-121.meena tushir finalpaper-1-12
1.meena tushir finalpaper-1-12
 
H046014853
H046014853H046014853
H046014853
 
Black-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX modelBlack-box modeling of nonlinear system using evolutionary neural NARX model
Black-box modeling of nonlinear system using evolutionary neural NARX model
 
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...
Hierarchical algorithms of quasi linear ARX Neural Networks for Identificatio...
 
Neural tool box
Neural tool boxNeural tool box
Neural tool box
 
Segmentation of Images by using Fuzzy k-means clustering with ACO
Segmentation of Images by using Fuzzy k-means clustering with ACOSegmentation of Images by using Fuzzy k-means clustering with ACO
Segmentation of Images by using Fuzzy k-means clustering with ACO
 
Real Time Face Detection on GPU Using OPENCL
Real Time Face Detection on GPU Using OPENCLReal Time Face Detection on GPU Using OPENCL
Real Time Face Detection on GPU Using OPENCL
 
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and ArchitecturesMetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
MetaPerturb: Transferable Regularizer for Heterogeneous Tasks and Architectures
 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANN
 
Adaptive modified backpropagation algorithm based on differential errors
Adaptive modified backpropagation algorithm based on differential errorsAdaptive modified backpropagation algorithm based on differential errors
Adaptive modified backpropagation algorithm based on differential errors
 
Electricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural NetworkElectricity Demand Forecasting Using Fuzzy-Neural Network
Electricity Demand Forecasting Using Fuzzy-Neural Network
 
Optimal neural network models for wind speed prediction
Optimal neural network models for wind speed predictionOptimal neural network models for wind speed prediction
Optimal neural network models for wind speed prediction
 
G010334554
G010334554G010334554
G010334554
 
Emergence of Invariance and Disentangling in Deep Representations
Emergence of Invariance and Disentangling in Deep RepresentationsEmergence of Invariance and Disentangling in Deep Representations
Emergence of Invariance and Disentangling in Deep Representations
 
poster-lowe-6
poster-lowe-6poster-lowe-6
poster-lowe-6
 

Similar to Performance Analysis of Activation Functions in Neural Networks

Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...
Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...
Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...IOSRJVSP
 
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Cemal Ardil
 
Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...IOSR Journals
 
Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...IOSR Journals
 
APPLIED MACHINE LEARNING
APPLIED MACHINE LEARNINGAPPLIED MACHINE LEARNING
APPLIED MACHINE LEARNINGRevanth Kumar
 
Efficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of TrussesEfficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of TrussesIRJET Journal
 
Design of airfoil using backpropagation training with mixed approach
Design of airfoil using backpropagation training with mixed approachDesign of airfoil using backpropagation training with mixed approach
Design of airfoil using backpropagation training with mixed approachEditor Jacotech
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithmsaciijournal
 
Efficient design of feedforward network for pattern classification
Efficient design of feedforward network for pattern classificationEfficient design of feedforward network for pattern classification
Efficient design of feedforward network for pattern classificationIOSR Journals
 
Analytical and Systematic Study of Artificial Neural Network
Analytical and Systematic Study of Artificial Neural NetworkAnalytical and Systematic Study of Artificial Neural Network
Analytical and Systematic Study of Artificial Neural NetworkIRJET Journal
 
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...
Classification of Iris Data using Kernel Radial Basis Probabilistic  Neural N...Classification of Iris Data using Kernel Radial Basis Probabilistic  Neural N...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...Scientific Review SR
 
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...Scientific Review
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORijac123
 
Recognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkRecognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkeSAT Journals
 
Recognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkRecognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkeSAT Publishing House
 
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...IJNSA Journal
 
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.Anirbit Mukherjee
 
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...CSCJournals
 

Similar to Performance Analysis of Activation Functions in Neural Networks (20)

Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...
Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...
Performance Analysis of the Sigmoid and Fibonacci Activation Functions in NGA...
 
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
Levenberg marquardt-algorithm-for-karachi-stock-exchange-share-rates-forecast...
 
Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...
 
Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...Digital Implementation of Artificial Neural Network for Function Approximatio...
Digital Implementation of Artificial Neural Network for Function Approximatio...
 
APPLIED MACHINE LEARNING
APPLIED MACHINE LEARNINGAPPLIED MACHINE LEARNING
APPLIED MACHINE LEARNING
 
Efficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of TrussesEfficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of Trusses
 
Design of airfoil using backpropagation training with mixed approach
Design of airfoil using backpropagation training with mixed approachDesign of airfoil using backpropagation training with mixed approach
Design of airfoil using backpropagation training with mixed approach
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
 
Efficient design of feedforward network for pattern classification
Efficient design of feedforward network for pattern classificationEfficient design of feedforward network for pattern classification
Efficient design of feedforward network for pattern classification
 
Analytical and Systematic Study of Artificial Neural Network
Analytical and Systematic Study of Artificial Neural NetworkAnalytical and Systematic Study of Artificial Neural Network
Analytical and Systematic Study of Artificial Neural Network
 
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...
Classification of Iris Data using Kernel Radial Basis Probabilistic  Neural N...Classification of Iris Data using Kernel Radial Basis Probabilistic  Neural N...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...
 
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
 
Ann
Ann Ann
Ann
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
 
Recognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkRecognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural network
 
Recognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural networkRecognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural network
 
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...
GENERALIZED LEGENDRE POLYNOMIALS FOR SUPPORT VECTOR MACHINES (SVMS) CLASSIFIC...
 
N ns 1
N ns 1N ns 1
N ns 1
 
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
My 2hr+ survey talk at the Vector Institute, on our deep learning theorems.
 
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
Comparison Between Levenberg-Marquardt And Scaled Conjugate Gradient Training...
 

More from Waqas Tariq

The Use of Java Swing’s Components to Develop a Widget
The Use of Java Swing’s Components to Develop a WidgetThe Use of Java Swing’s Components to Develop a Widget
The Use of Java Swing’s Components to Develop a WidgetWaqas Tariq
 
3D Human Hand Posture Reconstruction Using a Single 2D Image
3D Human Hand Posture Reconstruction Using a Single 2D Image3D Human Hand Posture Reconstruction Using a Single 2D Image
3D Human Hand Posture Reconstruction Using a Single 2D ImageWaqas Tariq
 
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...Waqas Tariq
 
A Proposed Web Accessibility Framework for the Arab Disabled
A Proposed Web Accessibility Framework for the Arab DisabledA Proposed Web Accessibility Framework for the Arab Disabled
A Proposed Web Accessibility Framework for the Arab DisabledWaqas Tariq
 
Real Time Blinking Detection Based on Gabor Filter
Real Time Blinking Detection Based on Gabor FilterReal Time Blinking Detection Based on Gabor Filter
Real Time Blinking Detection Based on Gabor FilterWaqas Tariq
 
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...Waqas Tariq
 
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...Toward a More Robust Usability concept with Perceived Enjoyment in the contex...
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...Waqas Tariq
 
Collaborative Learning of Organisational Knolwedge
Collaborative Learning of Organisational KnolwedgeCollaborative Learning of Organisational Knolwedge
Collaborative Learning of Organisational KnolwedgeWaqas Tariq
 
A PNML extension for the HCI design
A PNML extension for the HCI designA PNML extension for the HCI design
A PNML extension for the HCI designWaqas Tariq
 
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 Board
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 BoardDevelopment of Sign Signal Translation System Based on Altera’s FPGA DE2 Board
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 BoardWaqas Tariq
 
An overview on Advanced Research Works on Brain-Computer Interface
An overview on Advanced Research Works on Brain-Computer InterfaceAn overview on Advanced Research Works on Brain-Computer Interface
An overview on Advanced Research Works on Brain-Computer InterfaceWaqas Tariq
 
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...Waqas Tariq
 
Principles of Good Screen Design in Websites
Principles of Good Screen Design in WebsitesPrinciples of Good Screen Design in Websites
Principles of Good Screen Design in WebsitesWaqas Tariq
 
Progress of Virtual Teams in Albania
Progress of Virtual Teams in AlbaniaProgress of Virtual Teams in Albania
Progress of Virtual Teams in AlbaniaWaqas Tariq
 
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...Waqas Tariq
 
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...Waqas Tariq
 
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...Waqas Tariq
 
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text Editor
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text EditorDynamic Construction of Telugu Speech Corpus for Voice Enabled Text Editor
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text EditorWaqas Tariq
 
An Improved Approach for Word Ambiguity Removal
An Improved Approach for Word Ambiguity RemovalAn Improved Approach for Word Ambiguity Removal
An Improved Approach for Word Ambiguity RemovalWaqas Tariq
 
Parameters Optimization for Improving ASR Performance in Adverse Real World N...
Parameters Optimization for Improving ASR Performance in Adverse Real World N...Parameters Optimization for Improving ASR Performance in Adverse Real World N...
Parameters Optimization for Improving ASR Performance in Adverse Real World N...Waqas Tariq
 

More from Waqas Tariq (20)

The Use of Java Swing’s Components to Develop a Widget
The Use of Java Swing’s Components to Develop a WidgetThe Use of Java Swing’s Components to Develop a Widget
The Use of Java Swing’s Components to Develop a Widget
 
3D Human Hand Posture Reconstruction Using a Single 2D Image
3D Human Hand Posture Reconstruction Using a Single 2D Image3D Human Hand Posture Reconstruction Using a Single 2D Image
3D Human Hand Posture Reconstruction Using a Single 2D Image
 
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...
Camera as Mouse and Keyboard for Handicap Person with Troubleshooting Ability...
 
A Proposed Web Accessibility Framework for the Arab Disabled
A Proposed Web Accessibility Framework for the Arab DisabledA Proposed Web Accessibility Framework for the Arab Disabled
A Proposed Web Accessibility Framework for the Arab Disabled
 
Real Time Blinking Detection Based on Gabor Filter
Real Time Blinking Detection Based on Gabor FilterReal Time Blinking Detection Based on Gabor Filter
Real Time Blinking Detection Based on Gabor Filter
 
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...
Computer Input with Human Eyes-Only Using Two Purkinje Images Which Works in ...
 
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...Toward a More Robust Usability concept with Perceived Enjoyment in the contex...
Toward a More Robust Usability concept with Perceived Enjoyment in the contex...
 
Collaborative Learning of Organisational Knolwedge
Collaborative Learning of Organisational KnolwedgeCollaborative Learning of Organisational Knolwedge
Collaborative Learning of Organisational Knolwedge
 
A PNML extension for the HCI design
A PNML extension for the HCI designA PNML extension for the HCI design
A PNML extension for the HCI design
 
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 Board
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 BoardDevelopment of Sign Signal Translation System Based on Altera’s FPGA DE2 Board
Development of Sign Signal Translation System Based on Altera’s FPGA DE2 Board
 
An overview on Advanced Research Works on Brain-Computer Interface
An overview on Advanced Research Works on Brain-Computer InterfaceAn overview on Advanced Research Works on Brain-Computer Interface
An overview on Advanced Research Works on Brain-Computer Interface
 
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...
Exploring the Relationship Between Mobile Phone and Senior Citizens: A Malays...
 
Principles of Good Screen Design in Websites
Principles of Good Screen Design in WebsitesPrinciples of Good Screen Design in Websites
Principles of Good Screen Design in Websites
 
Progress of Virtual Teams in Albania
Progress of Virtual Teams in AlbaniaProgress of Virtual Teams in Albania
Progress of Virtual Teams in Albania
 
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...
Cognitive Approach Towards the Maintenance of Web-Sites Through Quality Evalu...
 
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...
USEFul: A Framework to Mainstream Web Site Usability through Automated Evalua...
 
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...
Robot Arm Utilized Having Meal Support System Based on Computer Input by Huma...
 
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text Editor
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text EditorDynamic Construction of Telugu Speech Corpus for Voice Enabled Text Editor
Dynamic Construction of Telugu Speech Corpus for Voice Enabled Text Editor
 
An Improved Approach for Word Ambiguity Removal
An Improved Approach for Word Ambiguity RemovalAn Improved Approach for Word Ambiguity Removal
An Improved Approach for Word Ambiguity Removal
 
Parameters Optimization for Improving ASR Performance in Adverse Real World N...
Parameters Optimization for Improving ASR Performance in Adverse Real World N...Parameters Optimization for Improving ASR Performance in Adverse Real World N...
Parameters Optimization for Improving ASR Performance in Adverse Real World N...
 

Recently uploaded

Capitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitolTechU
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentInMediaRes1
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for BeginnersSabitha Banu
 
Historical philosophical, theoretical, and legal foundations of special and i...
Historical philosophical, theoretical, and legal foundations of special and i...Historical philosophical, theoretical, and legal foundations of special and i...
Historical philosophical, theoretical, and legal foundations of special and i...jaredbarbolino94
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Celine George
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfFraming an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfUjwalaBharambe
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdfssuser54595a
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Educationpboyjonauth
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Celine George
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
MICROBIOLOGY biochemical test detailed.pptx
MICROBIOLOGY biochemical test detailed.pptxMICROBIOLOGY biochemical test detailed.pptx
MICROBIOLOGY biochemical test detailed.pptxabhijeetpadhi001
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupJonathanParaisoCruz
 
Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Jisc
 

Recently uploaded (20)

Capitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptxCapitol Tech U Doctoral Presentation - April 2024.pptx
Capitol Tech U Doctoral Presentation - April 2024.pptx
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 
Alper Gobel In Media Res Media Component
Alper Gobel In Media Res Media ComponentAlper Gobel In Media Res Media Component
Alper Gobel In Media Res Media Component
 
Full Stack Web Development Course for Beginners
Full Stack Web Development Course  for BeginnersFull Stack Web Development Course  for Beginners
Full Stack Web Development Course for Beginners
 
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdfTataKelola dan KamSiber Kecerdasan Buatan v022.pdf
TataKelola dan KamSiber Kecerdasan Buatan v022.pdf
 
Historical philosophical, theoretical, and legal foundations of special and i...
Historical philosophical, theoretical, and legal foundations of special and i...Historical philosophical, theoretical, and legal foundations of special and i...
Historical philosophical, theoretical, and legal foundations of special and i...
 
Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17Computed Fields and api Depends in the Odoo 17
Computed Fields and api Depends in the Odoo 17
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdfFraming an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
Framing an Appropriate Research Question 6b9b26d93da94caf993c038d9efcdedb.pdf
 
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
call girls in Kamla Market (DELHI) 🔝 >༒9953330565🔝 genuine Escort Service 🔝✔️✔️
 
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
18-04-UA_REPORT_MEDIALITERAСY_INDEX-DM_23-1-final-eng.pdf
 
Introduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher EducationIntroduction to ArtificiaI Intelligence in Higher Education
Introduction to ArtificiaI Intelligence in Higher Education
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17Difference Between Search & Browse Methods in Odoo 17
Difference Between Search & Browse Methods in Odoo 17
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
MICROBIOLOGY biochemical test detailed.pptx
MICROBIOLOGY biochemical test detailed.pptxMICROBIOLOGY biochemical test detailed.pptx
MICROBIOLOGY biochemical test detailed.pptx
 
9953330565 Low Rate Call Girls In Rohini Delhi NCR
9953330565 Low Rate Call Girls In Rohini  Delhi NCR9953330565 Low Rate Call Girls In Rohini  Delhi NCR
9953330565 Low Rate Call Girls In Rohini Delhi NCR
 
MARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized GroupMARGINALIZATION (Different learners in Marginalized Group
MARGINALIZATION (Different learners in Marginalized Group
 
Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...Procuring digital preservation CAN be quick and painless with our new dynamic...
Procuring digital preservation CAN be quick and painless with our new dynamic...
 

Performance Analysis of Activation Functions in Neural Networks

  • 1. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 111 Performance Analysis of Various Activation Functions in Generalized MLP Architectures of Neural Networks Bekir Karlik bkarlik@mevlana.edu.tr Engineering Faculty/Computer Engineering/ Mevlana University, Konya, 42003, Turkey A Vehbi Olgac avehbi@live.com Engineering Faculty/Computer Engineering/ Fatih University, Istanbul, 34500, Turkey Abstract The activation function used to transform the activation level of a unit (neuron) into an output signal. There are a number of common activation functions in use with artificial neural networks (ANN). The most common choice of activation functions for multi layered perceptron (MLP) is used as transfer functions in research and engineering. Among the reasons for this popularity are its boundedness in the unit interval, the function’s and its derivative’s fast computability, and a number of amenable mathematical properties in the realm of approximation theory. However, considering the huge variety of problem domains MLP is applied in, it is intriguing to suspect that specific problems call for single or a set of specific activation functions. The aim of this study is to analyze the performance of generalized MLP architectures which has back-propagation algorithm using various different activation functions for the neurons of hidden and output layers. For experimental comparisons, Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF) were used. Keywords: Activation Functions, Multi Layered Perceptron, Neural Networks, Performance Analysis 1. INTRODUCTION One of the most attractive properties of ANNs is the possibility to adapt their behavior to the changing characteristics of the modeled system. Last decades, many researchers have investigated a variety of methods to improve ANN performance by optimizing training methods, learn parameters, or network structure, comparably few works is done towards using activation functions. Radial basis function (RBF) neural network is one of the most popular neural network architectures [1]. The standard sigmoid reaches an approximation power comparable to or better than classes of more established functions investigated in the approximation theory (i.e.,splines and polynomials) [2]. Jordan presented the logistic function which is a natural representation of the posterior probability in a binary classification problem [3]. Liu and Yao improved the structure of Generalized Neural Networks (GNN) with two different activation function types which are sigmoid and Gaussian basis functions [4]. Sopena et al. presented a number of experiments (with widely–used benchmark problems) showing that multilayer feed–forward networks with a sine activation function learn two orders of magnitude faster while generalization capacity increases (compared to ANNs with logistic activation function) [5]. Dorffner developed the Conic Section Function Neural Network (CSFNN) which is a unified framework for MLP and RBF networks to make simultaneous use of advantages of both networks [6]. Bodyanskiy presented a novel double-wavelet neuron architecture obtained by modification of standard wavelet neuron, and their learning algorithms are proposed. The proposed architecture allows improving the
  • 2. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 112 approximation properties of wavelet neuron [7]. All these well-known activations functions are used into nodes of each layer of MLP to solve different non-linear problems. But, there is no performance of comparison studies using these activation functions. So, in this study, we have used five different well-known activation functions such as Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF) to compare their performances. 2. ACTIVATION FUNCTION TYPES The most important unit in neural network structure is their net inputs by using a scalar-to-scalar function called “the activation function or threshold function or transfer function”, output a result value called the “unit's activation”. An activation function for limiting the amplitude of the output of a neuron. Enabling in a limited range of functions is usually called squashing functions [8-9]. It squashes the permissible amplitude range of the output signal to some finite value. Some of the most commonly used activation functions are to solve non-linear problems. These functions are: Uni-polar sigmoid, Bi-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF). We did not care about some activation function such as identity function, step function or binary step functions as they are not used to solve linear problems. 2.1 Uni-Polar Sigmoid Function Activation function of Uni-polar sigmoid function is given as follows: x e xg − + = 1 1 )( (1) This function is especially advantageous to use in neural networks trained by back-propagation algorithms. Because it is easy to distinguish, and this can interestingly minimize the computation capacity for training. The term sigmoid means ‘S-shaped’, and logistic form of the sigmoid maps the interval (-∞, ∞) onto (0, 1) as seen in g(x) 0 1 -6 -4 -2 0 2 4 6 x FIGURE 1. g(x) 0 1 -6 -4 -2 0 2 4 6 x FIGURE 1: Uni-Polar Sigmoid Function
  • 3. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 113 2.2 Bipolar Sigmoid Function Activation function of Bi-polar sigmoid function is given by x x e e xg − − + − = 1 1 )( (2) This function is similar to the sigmoid function. For this type of activation function described in Fig. 2, it goes well for applications that produce output values in the range of [-1, 1]. g(x) -1 0 1 -6 -4 -2 0 2 4 6 x FIGURE 2: Bi-Polar Sigmoid Function 2.3 Hyperbolic Tangent Function This function is easily defined as the ratio between the hyperbolic sine and the cosine functions or expanded as the ratio of the half‐difference and half‐sum of two exponential functions in the points x and –x as follows : xx xx ee ee x x x − − + − == )cosh( )sinh( )tanh( (3) Hyperbolic Tangent Function is similar to sigmoid function. Its range outputs between -1 and 1 as seen in FIGURE . The following is a graphic of the hyperbolic tangent function for real values of its argument x: FIGURE 3: Hyperbolic Tangent Function 2.4 Radial Basis Function Radial basis function (RBF) is based on Gaussian Curve. It takes a parameter that determines the center (mean) value of the function used as a desired value (see Fig. 4). A radial basis
  • 4. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 114 function (RBF) is a real-valued function whose value depends only on the distance from the origin, so that )()( xgxg = (4) or alternatively on the distance from some other point c, called a center, so that )(),( cxgcxg −= (5) Sums of radial basis functions are typically used to approximate given functions. This approximation process can also be interpreted as a simple type of neural network. RBF are typically used to build up function approximations of the form )()( 1 i N i i cxgwxy −= ∑= (6) where the approximating function y(x) is represented as a sum of N radial basis functions, each associated with a different center ci, and weighted by an appropriate coefficient wi. The weights wi can be estimated using the matrix methods of linear least squares, because the approximating function is linear in the weights. Fig. 4 shows that two unnormalized Gaussian radial basis functions in one input dimension. The basis function centers are located at c1=0.75 and c2=3.25 [10]. FIGURE 4: Two unnormalized Gaussian radial basis functions in one input dimension RBF can also be interpreted as a rather simple single-layer type of artificial neural network called a radial basis function network, with the radial basis functions taking on the role of the activation functions of the network. It can be shown that any continuous function on a compact interval can in principle be interpolated with arbitrary accuracy by a sum of this form, if a sufficiently large number N of radial basis functions is used. 2.5 Conic Section Function Conic section function (CSF) is based on a section of a cone as the name implies. CSF takes a parameter that determines the angle value of the function as seen in Fig. 5.
  • 5. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 115 Figure 5: Conic Section Parabola (ω = 90) The equation of CSF can be defined as follows: )(cos)()( 1 1 i N i iiii cawwcaxf −−−= ∑ + = (7) where ia is input coefficient, ci is the center, wi is the weight in Multi Layered Perceptron (MLP), 2w is an opening angle which can be any value in the range of [-π/2, π/2] and determines the different forms of the decision borders. The hidden units of neural network need activation functions to introduce non-linearity into the networks. Non-linearity makes multi-layer networks so effective. The sigmoid functions are the most widely used functions [10-11]. Activation functions should be chosen to be suited to the distribution of the target values for the output units. You can think the same for binary outputs where the tangent hyperbolic and sigmoid functions are effective choices. If the target values are positive but have no known upper bound, an exponential output activation function can be used. This work has explained all the variations on the parallel distributed processing idea of neural networks. The structure of each neural network has very similar parts which perform the processing. 3. COMPARISON WITH DIFFERENT ACTIVATION FUNCTIONS In this study, different activation functions depend on different number of iterations for comparing their performances by using the same data. For all the activation functions, we used the number of nodes in the hidden layer; firstly 10 nodes, secondly 40 nodes (numbers of iterations are the same for both of them). After presenting the graphs for different parameters from Figure 6 through Figure 10, interpretations of their results will follow right here.
  • 6. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 116 FIGURE 6: 100 Iterations - 10 Hidden Neurons - Bi-Polar Sigmoid
  • 7. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 117 FIGURE 7: 100 Iterations - 10 Hidden Neurons - Uni-Polar Sigmoid FIGURE 8: 100 Iterations - 10 Hidden Neurons - Tangent Hyperbolic We used the same number of hidden neurons (10 nodes) and iterations (100 iterations) to compare the differences between the activation functions used above. It is also requires less number of iterations. FIGURE 9: 100 Iterations - 10 Hidden Neurons - Conic Section
  • 8. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 118 FIGURE 10: 100 Iterations - 10 Hidden Neurons – RBF According to the Figures from 6 through 15, we found Conic Section function the best activation function for training. Moreover, it requires less number of iterations than the other activation functions to solve the non-linear problems. However, with regard to testing, we found that the accuracy of Tanh activation function was much better than CSF and the other activation functions. This situation explains that total Mean Square Error (MSE) according to iterations cannot determine the network accuracy. Hence, we can obtain the real accuracy results upon testing.
  • 9. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 119 FIGURE 11: 500 Iterations - 40 Hidden Neurons - Bi-Polar Sigmoid FIGURE 12: 500 Iterations - 40 Hidden Neurons - Uni-Polar Sigmoid Figures 11 through 15 show the error graphics for each activation function for different numbers of nodes in the hidden layer which may vary from 10 up to 40, and the number of iterations from 100 up to 500.
  • 10. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 120 FIGURE 13: 500 Iterations - 40 Hidden Neurons - Tangent Hyperbolic FIGURE 14: 500 Iterations - 40 Hidden Neurons - Conic Section
  • 11. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 121 FIGURE 15: 500 Iterations - 40 Hidden Neurons – RBF According to the first five graphics generated by the same iteration numbers and the same number of neurons in the hidden layer, the Tangent Hyperbolic Activation function resulted in the most successful one for all the test parameters. Conic Section Function and Radial Basis Function were not able to successfully use the data type and the network parameters. The second group of result graphs is similar to the first group. But, the accuracy of the algorithms was observed to increase for the sigmoid and tangent hyperbolic functions. The rest of the functions used in this study were not successful and not accurate enough in this group. Table 1 shows those total accuracy and error values for the testing phase. 100 Iterations 10 Hidden Neurons 500 Iterations 40 Hidden Neurons Error Accuracy (%) Error Accuracy (%) Bi-Polar S. 0,056 93 0,034 89 Uni-Polar S. 0,026 92 0,006 88 Tanh 0,025 95 0,002 99 Conic 0,001 34 0,045 23 RBF 0,003 30 0,001 19 TABLE 1: Results of the testing phase 4. CONCLUSIONS In this study, we have used five conventional differentiable and monotonic activation functions for the evolution of MLP architecture along with Generalized Delta rule learning. These proposed well-known and effective activation functions are Bi-polar sigmoid, Uni-polar sigmoid, Tanh, Conic Section, and Radial Bases Function (RBF). Having compared their performances, simulation results show that Tanh (hyperbolic tangent) function performs better recognition accuracy than those of the other functions. In other words, the neural network computed good results when “Tanh-Tanh” combination of activation functions was used for both neurons (or nodes) of hidden and output layers.
  • 12. Bekir Karlik and A. Vehbi Olgac International Journal of Artificial Intelligence And Expert Systems (IJAE), Volume (1): Issue (4) 122 We presented experimental results by proposing five different activation functions for MLP neural networks architecture along with Generalized Delta Rule Learning to solve the non-linear problems. These results demonstrate that it is possible to improve the ANN performance through the use of much effective activation function. According to our experimental study, we can say that the Tanh activation function can be used in the vast majority of MLP applications as a good choice to obtain high accuracy. Furthermore, the sole use of validation test cannot approve the work even if it would be able to predict very low MSE. So, we can only decide about the real accuracy after a thorough testing of the neural network. In the future work, non–monotic activation functions will probably be sought for the test results in the learning speed and the performance of the neural network. REFERENCES 1. T. Poggio, F. Girosi, “A theory of networks for approximation and learning”. A.I. Memo No. 1140, Artificial Intelligence, Laboratory, Massachusetts Institute of Technology, 1989 2. B DasGupta, G. Schnitger, “The Power of Approximating: A Comparison of Activation Functions”. In Giles, C. L., Hanson, S. J., and Cowan, J. D., editors, Advances in Neural Information Processing Systems, 5, pp. 615-622, San Mateo, CA. Morgan Kaufmann Publishers, 1993 3. M.I. Jordan, “Why the logistic function? A tutorial discussion on probabilities and neural networks”. Computational Cognitive Science Technical Report 9503, Massachusetts Institute of Technology, 1995 4. Y. Liu, X. Yao, “Evolutionary Design of Artificial Neural Networks with Different Nodes”. In Proceedings of the Third IEEE International Conference on Evolutionary Computation, pp. 570-675, 1996 5. J. M. Sopena, E. Romero, R. Alqu´ezar, “Neural networks with periodic and monotonic activation functions: a comparative study in classification problems”. In Proceedings of the 9th International Conference on Artificial Neural Networks, pp. 323-328, 1999 6. G. Dorffner, “Unified frameworks for MLP and RBFNs: Introducing Conic Section Function Networks”. Cybernetics and Systems, 25: 511-554, 1994 7. Y. Bodyanskiy, N. Lamonova, O. Vynokurova, “Double-Wavelet Neuron Based on Analytical Activation Functions”. International Journal Information Theories & Applications, 14: 281-288, 2007 8. G. Cybenko, “Approximation by superposition of a sigmoidal function”. Mathematics of Control, Signals, and Systems, 2(4):303–314, 1987 9. R. P. Lippmann, “An introduction to computing with neural nets”. IEEE Acoustics, Speech and Signal Processing, 4(2):4–22, 1987 10. M.D. Buhmann, (2003), Radial Basis Functions: Theory and Implementations, Cambridge University Press, ISBN 978-0-521-63338-3. 11. B. Widrow, M.A. Lehr, “30 years of adoptive neural netwoks; perceptron, madaline, and back propagation”. Proc. IEEE, 78: 1415–1442, 1990 12. B. Ciocoiu, “Hybrid Feedforward Neural Networks for Solving Classification Problems”. Neural Processing Letters, 16(1):81-91, 2002