SlideShare a Scribd company logo
1 of 7
Download to read offline
IOSR Journal of Computer Engineering (IOSR-JCE)
e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May – Jun. 2015), PP 63-69
www.iosrjournals.org
DOI: 10.9790/0661-17376369 www.iosrjournals.org 63 | Page
A New Classifier Based onRecurrent Neural Network Using
Multiple Binary-Output Networks
Mohammad Fazel Younessy1
and Ehsanolah Kabir2
1
Department of Electrical and Computer Engineering, Islamic Azad University, Science and Research Branch,
Tehran, Iran
2
Department of Electrical Engineering, TarbiatModarres University, Tehran, Iran
Abstract:Conventional classifiers have several outputs equal to the number of classes (N) which imposes so
much complexityto the classifiers,both in training and testing stages. Instead of using one extra-large and
complicated classifier, we created N number of simple binary Recurrent Neural Network with true/false outputs.
Each network is responsible of recognizing its own trained class.We also proposed adecision layer added to
each network, making a final decision from a sequence of outputs. We test our system on features extracted from
Iranshahr database, which is a set of 17,000 black-and-white handwritten images of Iranian city names.
Experimental results authenticate the effectiveness of the proposed method.
Keywords: Recurrent Neural Network,RNN, Classifier, Binary Network.
I. Introduction
Classifiers are essential tools for the applications that need to find a specific type of result among a
given set of results, such as offline/online handwritten/typewritten word recognition[1][2][3][4], speech
recognition[5][6][7][8], and medical diagnosis systems[9][10][11]. Examples of such classifiers are Artificial
Neural Network (ANN), Hidden Markov Models (HMM),support vector machine (SVM), and k-nearest
neighbor (k-NN). Generally, a classifier has a training phase, in which some input samples with known results
are fed into the classifier and the classifier internal weights would be trained by such data. Afterwards, in the
testing phase, the input samples with unknown output results would be applied to the classifier, and the results
would be evaluated.
Conventional classifiers are consisted of high-quantity outputs, which force the classifier to be
internally complicated and huge. Such classifiers all suffer from poor results and slow timing, both in training
stage and testing phase. In our paper, we have proposed a new approach to conquer such problems. Instead of
such complicated networks, we make multiple number of networks, each only consisted of a True/False output.
It means, instead of one classifier with N number of outputs, we use N classifiers with simple outputs.Each
classifier would be trained to have true result only by one class. This leads to small and simple classifiers with
faster response time.
One of the most common classifiers used in practice, is ANN. An ANN is consisted of units called
neurons, which are arranged in layers. Typically, ANN converts an input vector (features) to outputs. The
classifier type that we used in our work is a Recurrent Neural Network (RNN). A simple RNN is a special type
of ANN, with oneor more memory layer(s) added to its feedforward layers.The memory layer is a feedback
path, which makes the recurrence of the system.
A typical ANN has an input vector, which makes one output (vector), whereas, RNN has a sequence of
input vectors, which makes series of outputs. Hence, we added a decision layer to the output of the network to
make the final decision (true or false) out of the sequence of inputs. Some other researchers also tried to make a
classifier by RNN system. However, they all seem to be some complex mathematical problem, which would be
added to the inherent complicated nature of conventional networks. For example, in [12], the authors proposed
an RNN classifier with several recurrent layers added to each other.A valuable system has been presented
in[13][14][15]. This system has addeda decision layer called connectionist temporal classification (CTC) to the
output of the RNN and tried to train the network by a differentiable function. A rapid temporal sampling
combined by an RNN classifier has been used for measurement of areal snow coverage in [16]. The authors in
[17] replaced the simple neurons in RNN with a more complicated neuron set, and called it long-short term
memory (LSTM) system, and used it in manuscript recognition.
The rest of the paper would be organized as follows. In next section we will describe the proposed
classification method in detail.In Section 3, we will show our experimental results on a real database. Finally we
conclude our work in section four, and show a possible future work in that section.
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 64 | Page
II. Proposed ClassifierSystem
In this section we will focus on details of our proposed system. The task of the system is to make a
classifier to classify N outputs. Block diagram of the system is shown inFig. 1. As this figure shows, the system
is composed of a training phase and testing phase.
Fig. 1: Block diagram of proposed system.
In training stage, at first, input data vector would be represented to be fed properly to the network. As
mentioned, in our system, instead of using one huge classifier with N outputs, N number of simple classifiers
would be used. Each network is an RNN classifier with a decision layer added to the output layer. The RNN set
would be expressed as below:
RNNs: RNN1, … , RNNN ( 1 )
Each RNN classifier is responsible for one class. Each classifier will have “True/False” outputs. This
means in test phase, we expect that if samples of class i is applied to the RNNi, then output should be true and if
other samples from another classes are fed into the network, then output should be false. This is an interesting
phenomenon, because each network would be simple, and only is trained for one class.
RNN networks normally need a sequence of inputs and make a sequence of outputs. Our task is to
make a classifier, so the network should have one output (true or false), not a train of outputs. As a result, we
made a decision layer added to input of each RNN. Decision layer would use the sequence of outputs from each
RNN and finally make a final decision which might be true or false.
In the testing phase, similar to the training phase, we send the input vectors to the representation unit,
and then to each RNN classifiers (Recall that each RNN has a decision layer at the output). After this stage, we
have a list of outputs derived from the RNN classifiers. We sort the outputs in the final decision section and
output of the system would be a sorted list of probabilities of each class.
In the following sub-sections we will discuss each part of the main block diagram in detail.
A. Input Representation
Input data to the RNN network is a sequence of vectors. Let’s define Fi as an input vector with the
dimension of U × 1. A sequence of Fi, with length of V would make the input data of the network. Note that,
interestingly,V is dependent on the input vector sequence’s length, and may be different from sample to sample.
Equation below shows the idea:
𝔽 = F1: … : FV
Fi =
f1
⋮
fU
( 2 )
Before sending the input vectors to the networks, we standardized the inputs. We forced the inputs to
have a mean (μ) of 0, and standard deviation (σ) of 1. This procedure would make data optimum for the
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 65 | Page
activation functions inside the RNNs, and main information will not be changed, however, performance of
system would improve significantly[18].
B. Network Architecture
Normally, any arbitrary network can be made with any configuration. We have chosen a heuristic
network topology which is presented in Fig. 2.As this figure shows, the network is consisted of two feedforward
layers, hidden layer and output layer.The network also has a feedback (recurrent) layer, with delay of
Tdelay which has been fed back to the hidden layer (layer 1). As mentioned earlier, output layer of each RNN has
two outputs, one showing the state of being true, and the other one the state of being false.
Fig. 2:Topology of the RNN (Figure from[19]).
For each feedforward layer we use Tanh-Sigmoid transfer function[20], which has a output between -1
and +1. This makes the final outputs of the network to be in the range of −1, +1 , which gives us the
probability of the true/false instead of a hard output of being discrete 0 or 1.
C. Decision Layer
Naturally, an RNN has a train of input vectors, and this leads to a sequence of 2-dimensional (2D)
vectors on the output. As a classifier we need to concatenate the result sequence and make a decision at the end
of output sequence. Hence we made a decision layer at the end of each RNN.As shown inFig. 3, a decision
block is added to the RNN block.
Fig. 3: Single RNN with decision layer added to the outputs.
Suppose V is the length of input sequence. We call the outputs of the RNN as y1iand y2i, which are the
true and false outputs of the RNN respectively. We have:
Yi =
1 , if y1i > y2i
0 , else
Yout =
Yi
V − Tdelay
∙
ωi
Ω
× 100
V
i=Tdelay
Ω = ωi
V
i=Tdelay
( 3 )
Looking at ( 3 ), Yi is the decision made by a single vector at each output of the RNN.Yout is the final
decision of the network taken from train of 2D outputs. Please note that Tdelay is the network recurrence time,
and is a system parameter. Normally outputs before that time is of less value, so we don’t consider them.As
above formula shows, output ofdecision layer, Yout , is in fact the ratio of the number of true outputs to the
sequence length. We added a weight coefficient, ωi, to the calculation to improve the results. These weights,
should go from a small value to bigger values, as the network makes output through the time sequence. This
𝐼𝑛𝑝𝑢𝑡
𝑆𝑒𝑞𝑢𝑒𝑛𝑐𝑒
𝑈×𝑉
𝑅𝑁𝑁𝑖
𝑇𝑟𝑢𝑒
𝐹𝑎𝑙𝑠𝑒
𝐷𝑒𝑐𝑖𝑠𝑖𝑜𝑛
𝐿𝑎𝑦𝑒𝑟
𝑦1𝑖
𝑦2𝑖
𝑌𝑜𝑢𝑡
𝑅𝐸𝐿 𝑜𝑢𝑡
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 66 | Page
phenomenon guarantees that the network’s output at the end of the sequence has more influence on the final
decision. This is due to the fact that, at the start of the network sequence evaluation, the recurrent characteristics
of the network is not so strong, but, as more input vectors (from the input sequence) are fed into the network,
network has more reliable outputs. At the end, to omit the effect of the weights, we have normalized the weights
by summation of ωi, Ω, to remove the effect of the weights.In our systems we used three functions for the
weights: 1. ωi = log(i), 2. ωi = exp(i), and 3. ωi = tanh(i/N)
As a result of all mentioned, Yout shows the percentage of similarity of each input sequence, to the
class that RNN network is trained with. It means, ideally, if a sequence of inputs similar to the specific RNN’s
class is fed, output would be 100% and for the other sequences, output is 0%.
Because of comparison in the first line of ( 3 ), decision made by above formula, is in fact a hard
decision. To preserve the soft information of RNN outputs, we created a reliability function, RELout . This
functions shows how much valid the output of the decision layer is. We formulate RELout as below:
RELout =
1
2
y1i − y2i
V − Tdelay
V
i=Tdelay
∙
ωi
Ω
× 100 ( 4 )
As the above formula shows, RELout is, in fact, the average of distance of true outputs and false
outputs, multiplied by the weights. Please note that the coefficient of
1
2
is due to the fact that y1i and y2i are in
the range of −1, +1 .
D. RNN Training
Some special attention should be given to the training method of the RNNs. Generally, in a database,
there are a few samples for each class. In our case, each RNN is responsible for a specific class. If we give the
few related samples to this RNN, forcing the outputs to true (for backpropagation training), and force the
outputs to be false for the many unrelated samples, the network would be finally trained to false results. For
examples, 3% of the input samples would train the network to true, and 97% of the other samples would train
the network to false outputs. This means network will generate false results at the test time. Hence, we have
added the quantity of true samples virtually, by adding some Gaussian noise to the true samples, making them
about one-third of the total samples.
To initialize the weights of the network, we used a Gaussian distribution with μ, σ = (0, 0.1).
Furthermore, we can retrain the network after one successful training phase. Our simulation showed this technic
may increase total result by 2-3%, and also retraining more than 3 retraining has no sensible effect. The negative
point about the retraining is the time it consumes because of repeating the training phase.
Several algorithms for network training has been proposed in scientific texts, such as Levenberg-
Marquardt[21], Bayesian Regularization[22],[23], BFGS Quasi-Newton[24], Gradient Descent[25] and
Resilient Backpropagation(Rprop)[26]. Among those, the Rprop seems to work better for us.Rprop algorithm
normally uses relatively small amount of memory and is faster than most of the algorithms.
E. RNN Testing
To test the Network, we divided the sample data to 3 sets, 70% for training stage, 15% forvalidation of
training phase, and 15% for final test and report.
II. Experimental Results
A. Test Database
To test our system, we have used a real database called Iranshahr. This database is consisted of more
than 17,000 samples of binary images of handwritten Iranian city names, which has been also used in [27]. The
number of classes (cities) in this database is more than 500, which means each class has about 30 samples. We
have used the method in [28] to extract the feature vectors. This method uses the sliding window technic to
extract image features inside a frame. Then by shifting the window from right to left of the image, makes a
sequence of vectors for each sample image. These features are input vector sequences to our system.
B. Classification Results
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 67 | Page
Table 1 shows the best classification result obtained by the system. The second row in this table shows
the classification rate (the percentage of recognizing correct city classes in the database) and the third row shows
the average reliability of the correct results. As this table shows, the recognition rate and reliability of our
system is relatively high and acceptable.
Table 1: Classification results
Top 1 Top 2 Top 5 Top 10
Classification rate 83.9% 84.7% 87.5% 91.1%
Averagereliability 72.3% 73.1% 78.5% 80.5%
C. Effect of Recurrence Delay
A system parameter that has been taken into consideration is the Tdelay , which is the network
recurrence delay. The effect of this parameter has been shown in Fig. 4. This figure shows that forTdelay = 3,
the rate is at its maximum. Please note that, obviously, if we increase Tdelay , we will miss some recurrence
statesand information of the RNN, and thus decreasing the classification rate.
Fig. 4: Effect of Tdelay on classification rate.
D. Effect of Retraining
Figure below shows the effect of retraining on our system rate. We can see retraining of the network up
to 3 times is a good practice, and more than that has almost no effect on system.
Fig. 5: Effect of retraining times on classification rate.
E. Effect of Weights
We have tested our system by 3 weight types. Table below shows the effect of each weight type on our
system. As the table shows, the logarithm function works the best for us and exponential the worst.
0 1 2 3 4 5
0
20
40
60
80
100
Tdelay
Classificationrate(%)
Top 1
Top 2
Top 5
Top 10
0 1 2 3 4 5
0
20
40
60
80
100
Retraining times
Classificationrate(%)
Top 1
Top 2
Top 5
Top 10
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 68 | Page
Table 2: Effect of weight function on classification rate.
Classification rate
Top 1 Top 2 Top 5 Top 10
Weight
type
ωi = exp(i) 75.3% 76.0% 79.6% 80.5%
ωi = tanh(i/N) 80.2% 81.1% 84.1% 85.5%
ωi = log(i) 83.9% 84.7% 87.5% 91.1%
F. Comparison with conventional RNN Classifier
To compare our method with the conventional classification method, i.e. using one huge classifier with
so many outputs, we trained an RNN network with output layer of 500 neurons (equal to the number of classes
in database), and added a decision layer to make the decision. The classification rate was very poor (less than
50%), and shows that our method is strongly better than the conventional classification method which uses
RNNs.
III. Conclusion and Future Work
In this paper we proposed a good method of reducing the network complexity both in training and
testing stages by making several binary networks instead of one complicated network. We used a Recurrent
Neural Network as a classifier and added a decision layer to make the decision. We conducted our test in a real
database to show the power of the method and a good recognition rate has been derived from the system.
As a future work, researchers could use our method in the field of handwritten/typewritten/speech
recognition systems. Even some extra information, such as length of the inputs,can prune some RNN classifiers,
so the final result would be chosen from a reduced set of the networks.
References
[1] Kumawat, P.; Khatri, A.; Nagaria, B., "Comparative Analysis of Offline Handwriting Recognition Using Invariant Moments with
HMM and Combined SVM-HMM Classifier," Communication Systems and Network Technologies (CSNT), 2013 International
Conference on , vol., no., pp.140,143, 6-8 April 2013.
[2] Maqqor, A.; Halli, A.; Satori, K.; Tairi, H., "Off-line recognition handwriting Arabic words using combination of multiple
classifiers," Information Science and Technology (CIST), 2014 Third IEEE International Colloquium in , vol., no., pp.260,265, 20-
22 Oct. 2014
[3] Wu Wei; Gao Guanglai, "Online Handwriting Mongolia words recognition based on HMM classifier," Control and Decision
Conference, 2009. CCDC '09. Chinese , vol., no., pp.3912,3914, 17-19 June 2009.
[4] Al-Hajj Mohamad, R.; Likforman-Sulem, L.; Mokbel, C., "Combining Slanted-Frame Classifiers for Improved HMM-Based Arabic
Handwriting Recognition," Pattern Analysis and Machine Intelligence, IEEE Transactions on , vol.31, no.7, pp.1165,1177, July
2009
[5] Chung-Hsien Wu; Wei-Bin Liang, "Emotion Recognition of Affective Speech Based on Multiple Classifiers Using Acoustic-
Prosodic Information and Semantic Labels," Affective Computing, IEEE Transactions on , vol.2, no.1, pp.10,21, Jan.-June 2011.
[6] Tsang-Long Pao; Wen-Yuan Liao; Yu-Te Chen; Jun-Heng Yeh; Yun-Maw Cheng; Chien, C.S., "Comparison of Several Classifiers
for Emotion Recognition from Noisy Mandarin Speech," Intelligent Information Hiding and Multimedia Signal Processing, 2007.
IIHMSP 2007. Third International Conference on , vol.1, no., pp.23,26, 26-28 Nov. 2007.
[7] Chandrakala, S.; Sekhar, C.C., "Combination of generative models and SVM based classifier for speech emotion recognition,"
Neural Networks, 2009. IJCNN 2009. International Joint Conference on , vol., no., pp.497,502, 14-19 June 2009.
[8] Kucukbay, S.E.; Sert, M., "Audio-based event detection in office live environments using optimized MFCC-SVM approach,"
Semantic Computing (ICSC), 2015 IEEE International Conference on , vol., no., pp.475,480, 7-9 Feb. 2015
[9] Deng, Z.; Cao, L.; Jiang, Y.; Wang, S., "Minimax Probability TSK Fuzzy System Classifier: A More Transparent and Highly
Interpretable Classification Model," Fuzzy Systems, IEEE Transactions on , vol.PP, no.99, pp.1,1.
[10] Gladence, L.M.; Ravi, T.; Karthi, M., "An enhanced method for detecting congestive heart failure - Automatic Classifier,"
Advanced Communication Control and Computing Technologies (ICACCCT), 2014 International Conference on , vol., no.,
pp.586,590, 8-10 May 2014.
[11] AlObaidi, A.T.S.; Mahmood, N.T., "Modified Full Bayesian Networks Classifiers for Medical Diagnosis," Advanced Computer
Science Applications and Technologies (ACSAT), 2013 International Conference on , vol., no., pp.5,12, 23-24 Dec. 2013
[12] Seong-Whan Lee; Hee-Heon Song, "A new recurrent neural-network architecture for visual pattern recognition," Neural Networks,
IEEE Transactions on , vol.8, no.2, pp.331,340, Mar 1997.
[13] Graves, Alex, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. "Connectionist temporal classification: labelling
unsegmented sequence data with recurrent neural networks." In Proceedings of the 23rd international conference on Machine
learning, pp. 369-376. ACM, 2006.
[14] Graves, Alex, Marcus Liwicki, Santiago Fernández, Roman Bertolami, Horst Bunke, and Jürgen Schmidhuber. "A novel
connectionist system for unconstrained handwriting recognition." Pattern Analysis and Machine Intelligence, IEEE Transactions on
31, no. 5 (2009): 855-868.
[15] A. Graves. Supervised Sequence Labelling with Recurrent Neural Networks. Dissertation, Technische Universität München,
München, July 2008.
[16] Simpson, J.J.; McIntire, T.J., "A recurrent neural network classifier for improved retrievals of areal extent of snow cover,"
Geoscience and Remote Sensing, IEEE Transactions on , vol.39, no.10, pp.2135,2147, Oct 2001.
[17] Frinken, Volkmar, Francisco Zamora-Martinez, Salvador Espana-Boquera, Maria Jose Castro-Bleda, Andreas Fischer, and Horst
Bunke. "Long-short term memory neural networks language modeling for handwriting recognition." In Pattern Recognition (ICPR),
2012 21st International Conference on, pp. 701-704. IEEE, 2012.
[18] LeCun, L. Bottou, G. Orr, and K. Muller. Ecient backprop. In G. Orrand M. K., editors, Neural Networks: Tricks of the trade.
Springer, 1998b.
A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks
DOI: 10.9790/0661-17376369 www.iosrjournals.org 69 | Page
[19] MathWorks™, (2012). Neural Network Toolbox7User’s Guide (R2012b).
[20] Vogl, T.P., J.K. Mangis, A.K. Rigler, W.T. Zink, and D.L. Alkon, "Accelerating the convergence of the backpropagation method,"
Biological Cybernetics, Vol. 59, 1988, pp. 257–263
[21] Kenneth Levenberg (1944). "A Method for the Solution of Certain Non-Linear Problems in Least Squares". Quarterly of Applied
Mathematics 2: 164–168
[22] MacKay, Neural Computation, Vol. 4, No. 3, 1992, pp. 415–447
[23] Foresee and Hagan, Proceedings of the International Joint Conference on Neural Networks, June, 1997
[24] Gill, Murray, & Wright, Practical Optimization, 1981.
[25] Jan A. Snyman (2005). Practical Mathematical Optimization: An Introduction to Basic Optimization Theory and Classical and New
Gradient-Based Algorithms. Springer Publishing.
[26] Riedmiller, M., and H. Braun, “A direct adaptive method for faster backpropagation learning: The RPROP algorithm”, Proceedings
of the IEEE International Conference on Neural Networks, 1993.
[27] Bayesteh, E.; Ahmadifard, A.; Khosravi, Hossein, "A Lexicon Reduction Method Based on Clustering Word Images in Offline
Farsi Handwritten Word Recognition Systems," Machine Vision and Image Processing (MVIP), 2011 7th Iranian , vol., no., pp.1,5,
16-17 Nov. 2011.
[28] M. Dehghan, K. Faez, M. Ahmadi, and M. Shridhar, “Handwritten Farsi (Arabic) Word Recognition: A Holistic Approach Using
Discrete HMM,” Pattern Recognition, vol. 34, no. 5, pp. 1057 1065, 2001.

More Related Content

What's hot

Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
abhishek upadhyay
 
Artificial neural network - Architectures
Artificial neural network - ArchitecturesArtificial neural network - Architectures
Artificial neural network - Architectures
Erin Brunston
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network
Iman Ardekani
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
ESCOM
 
Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?
Victor Miagkikh
 
(Artificial) Neural Network
(Artificial) Neural Network(Artificial) Neural Network
(Artificial) Neural Network
Putri Wikie
 
Secured transmission through multi layer perceptron in wireless communication...
Secured transmission through multi layer perceptron in wireless communication...Secured transmission through multi layer perceptron in wireless communication...
Secured transmission through multi layer perceptron in wireless communication...
ijmnct
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
cscpconf
 

What's hot (20)

Back propagation
Back propagationBack propagation
Back propagation
 
Artifical Neural Network
Artifical Neural NetworkArtifical Neural Network
Artifical Neural Network
 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
 
14 Machine Learning Single Layer Perceptron
14 Machine Learning Single Layer Perceptron14 Machine Learning Single Layer Perceptron
14 Machine Learning Single Layer Perceptron
 
Artificial neural network - Architectures
Artificial neural network - ArchitecturesArtificial neural network - Architectures
Artificial neural network - Architectures
 
Artificial Neural Network
Artificial Neural Network Artificial Neural Network
Artificial Neural Network
 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
 
Artificial Neuron network
Artificial Neuron network Artificial Neuron network
Artificial Neuron network
 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
 
hopfield neural network
hopfield neural networkhopfield neural network
hopfield neural network
 
Fundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural NetworksFundamental, An Introduction to Neural Networks
Fundamental, An Introduction to Neural Networks
 
Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?Learning in Networks: were Pavlov and Hebb right?
Learning in Networks: were Pavlov and Hebb right?
 
(Artificial) Neural Network
(Artificial) Neural Network(Artificial) Neural Network
(Artificial) Neural Network
 
Secured transmission through multi layer perceptron in wireless communication...
Secured transmission through multi layer perceptron in wireless communication...Secured transmission through multi layer perceptron in wireless communication...
Secured transmission through multi layer perceptron in wireless communication...
 
Supervised sequence labelling with recurrent neural networks ch1 6
Supervised sequence labelling with recurrent neural networks ch1 6Supervised sequence labelling with recurrent neural networks ch1 6
Supervised sequence labelling with recurrent neural networks ch1 6
 
Max net
Max netMax net
Max net
 
Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyay
 
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTE...
 
Lecture 9 Perceptron
Lecture 9 PerceptronLecture 9 Perceptron
Lecture 9 Perceptron
 
Adaline madaline
Adaline madalineAdaline madaline
Adaline madaline
 

Similar to A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Output Networks

ANNs have been widely used in various domains for: Pattern recognition Funct...
ANNs have been widely used in various domains for: Pattern recognition  Funct...ANNs have been widely used in various domains for: Pattern recognition  Funct...
ANNs have been widely used in various domains for: Pattern recognition Funct...
vijaym148
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1
ncct
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
IJERD Editor
 

Similar to A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Output Networks (20)

MNN
MNNMNN
MNN
 
ai7.ppt
ai7.pptai7.ppt
ai7.ppt
 
Efficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of TrussesEfficiency of Neural Networks Study in the Design of Trusses
Efficiency of Neural Networks Study in the Design of Trusses
 
ai7.ppt
ai7.pptai7.ppt
ai7.ppt
 
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTORARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
ARTIFICIAL NEURAL NETWORK APPROACH TO MODELING OF POLYPROPYLENE REACTOR
 
DEEPLEARNING recurrent neural networs.pdf
DEEPLEARNING recurrent neural networs.pdfDEEPLEARNING recurrent neural networs.pdf
DEEPLEARNING recurrent neural networs.pdf
 
Web spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithmsWeb spam classification using supervised artificial neural network algorithms
Web spam classification using supervised artificial neural network algorithms
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network AlgorithmsWeb Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
 
A survey research summary on neural networks
A survey research summary on neural networksA survey research summary on neural networks
A survey research summary on neural networks
 
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdfNEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
 
20120140503023
2012014050302320120140503023
20120140503023
 
Multi Layer Network
Multi Layer NetworkMulti Layer Network
Multi Layer Network
 
ANNs have been widely used in various domains for: Pattern recognition Funct...
ANNs have been widely used in various domains for: Pattern recognition  Funct...ANNs have been widely used in various domains for: Pattern recognition  Funct...
ANNs have been widely used in various domains for: Pattern recognition Funct...
 
Neural Networks Ver1
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1
 
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptx
 
Adaptive equalization
Adaptive equalizationAdaptive equalization
Adaptive equalization
 
International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)International Journal of Engineering Research and Development (IJERD)
International Journal of Engineering Research and Development (IJERD)
 
Echo state networks and locomotion patterns
Echo state networks and locomotion patternsEcho state networks and locomotion patterns
Echo state networks and locomotion patterns
 
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
 
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
 

More from iosrjce

More from iosrjce (20)

An Examination of Effectuation Dimension as Financing Practice of Small and M...
An Examination of Effectuation Dimension as Financing Practice of Small and M...An Examination of Effectuation Dimension as Financing Practice of Small and M...
An Examination of Effectuation Dimension as Financing Practice of Small and M...
 
Does Goods and Services Tax (GST) Leads to Indian Economic Development?
Does Goods and Services Tax (GST) Leads to Indian Economic Development?Does Goods and Services Tax (GST) Leads to Indian Economic Development?
Does Goods and Services Tax (GST) Leads to Indian Economic Development?
 
Childhood Factors that influence success in later life
Childhood Factors that influence success in later lifeChildhood Factors that influence success in later life
Childhood Factors that influence success in later life
 
Emotional Intelligence and Work Performance Relationship: A Study on Sales Pe...
Emotional Intelligence and Work Performance Relationship: A Study on Sales Pe...Emotional Intelligence and Work Performance Relationship: A Study on Sales Pe...
Emotional Intelligence and Work Performance Relationship: A Study on Sales Pe...
 
Customer’s Acceptance of Internet Banking in Dubai
Customer’s Acceptance of Internet Banking in DubaiCustomer’s Acceptance of Internet Banking in Dubai
Customer’s Acceptance of Internet Banking in Dubai
 
A Study of Employee Satisfaction relating to Job Security & Working Hours amo...
A Study of Employee Satisfaction relating to Job Security & Working Hours amo...A Study of Employee Satisfaction relating to Job Security & Working Hours amo...
A Study of Employee Satisfaction relating to Job Security & Working Hours amo...
 
Consumer Perspectives on Brand Preference: A Choice Based Model Approach
Consumer Perspectives on Brand Preference: A Choice Based Model ApproachConsumer Perspectives on Brand Preference: A Choice Based Model Approach
Consumer Perspectives on Brand Preference: A Choice Based Model Approach
 
Student`S Approach towards Social Network Sites
Student`S Approach towards Social Network SitesStudent`S Approach towards Social Network Sites
Student`S Approach towards Social Network Sites
 
Broadcast Management in Nigeria: The systems approach as an imperative
Broadcast Management in Nigeria: The systems approach as an imperativeBroadcast Management in Nigeria: The systems approach as an imperative
Broadcast Management in Nigeria: The systems approach as an imperative
 
A Study on Retailer’s Perception on Soya Products with Special Reference to T...
A Study on Retailer’s Perception on Soya Products with Special Reference to T...A Study on Retailer’s Perception on Soya Products with Special Reference to T...
A Study on Retailer’s Perception on Soya Products with Special Reference to T...
 
A Study Factors Influence on Organisation Citizenship Behaviour in Corporate ...
A Study Factors Influence on Organisation Citizenship Behaviour in Corporate ...A Study Factors Influence on Organisation Citizenship Behaviour in Corporate ...
A Study Factors Influence on Organisation Citizenship Behaviour in Corporate ...
 
Consumers’ Behaviour on Sony Xperia: A Case Study on Bangladesh
Consumers’ Behaviour on Sony Xperia: A Case Study on BangladeshConsumers’ Behaviour on Sony Xperia: A Case Study on Bangladesh
Consumers’ Behaviour on Sony Xperia: A Case Study on Bangladesh
 
Design of a Balanced Scorecard on Nonprofit Organizations (Study on Yayasan P...
Design of a Balanced Scorecard on Nonprofit Organizations (Study on Yayasan P...Design of a Balanced Scorecard on Nonprofit Organizations (Study on Yayasan P...
Design of a Balanced Scorecard on Nonprofit Organizations (Study on Yayasan P...
 
Public Sector Reforms and Outsourcing Services in Nigeria: An Empirical Evalu...
Public Sector Reforms and Outsourcing Services in Nigeria: An Empirical Evalu...Public Sector Reforms and Outsourcing Services in Nigeria: An Empirical Evalu...
Public Sector Reforms and Outsourcing Services in Nigeria: An Empirical Evalu...
 
Media Innovations and its Impact on Brand awareness & Consideration
Media Innovations and its Impact on Brand awareness & ConsiderationMedia Innovations and its Impact on Brand awareness & Consideration
Media Innovations and its Impact on Brand awareness & Consideration
 
Customer experience in supermarkets and hypermarkets – A comparative study
Customer experience in supermarkets and hypermarkets – A comparative studyCustomer experience in supermarkets and hypermarkets – A comparative study
Customer experience in supermarkets and hypermarkets – A comparative study
 
Social Media and Small Businesses: A Combinational Strategic Approach under t...
Social Media and Small Businesses: A Combinational Strategic Approach under t...Social Media and Small Businesses: A Combinational Strategic Approach under t...
Social Media and Small Businesses: A Combinational Strategic Approach under t...
 
Secretarial Performance and the Gender Question (A Study of Selected Tertiary...
Secretarial Performance and the Gender Question (A Study of Selected Tertiary...Secretarial Performance and the Gender Question (A Study of Selected Tertiary...
Secretarial Performance and the Gender Question (A Study of Selected Tertiary...
 
Implementation of Quality Management principles at Zimbabwe Open University (...
Implementation of Quality Management principles at Zimbabwe Open University (...Implementation of Quality Management principles at Zimbabwe Open University (...
Implementation of Quality Management principles at Zimbabwe Open University (...
 
Organizational Conflicts Management In Selected Organizaions In Lagos State, ...
Organizational Conflicts Management In Selected Organizaions In Lagos State, ...Organizational Conflicts Management In Selected Organizaions In Lagos State, ...
Organizational Conflicts Management In Selected Organizaions In Lagos State, ...
 

Recently uploaded

Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
9953056974 Low Rate Call Girls In Saket, Delhi NCR
 
Digital Communication Essentials: DPCM, DM, and ADM .pptx
Digital Communication Essentials: DPCM, DM, and ADM .pptxDigital Communication Essentials: DPCM, DM, and ADM .pptx
Digital Communication Essentials: DPCM, DM, and ADM .pptx
pritamlangde
 
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
ssuser89054b
 
DeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakesDeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakes
MayuraD1
 
Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdf
Kamal Acharya
 

Recently uploaded (20)

Electromagnetic relays used for power system .pptx
Electromagnetic relays used for power system .pptxElectromagnetic relays used for power system .pptx
Electromagnetic relays used for power system .pptx
 
Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
Call Girls in South Ex (delhi) call me [🔝9953056974🔝] escort service 24X7
 
Hostel management system project report..pdf
Hostel management system project report..pdfHostel management system project report..pdf
Hostel management system project report..pdf
 
Digital Communication Essentials: DPCM, DM, and ADM .pptx
Digital Communication Essentials: DPCM, DM, and ADM .pptxDigital Communication Essentials: DPCM, DM, and ADM .pptx
Digital Communication Essentials: DPCM, DM, and ADM .pptx
 
School management system project Report.pdf
School management system project Report.pdfSchool management system project Report.pdf
School management system project Report.pdf
 
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX
 
DeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakesDeepFakes presentation : brief idea of DeepFakes
DeepFakes presentation : brief idea of DeepFakes
 
Hospital management system project report.pdf
Hospital management system project report.pdfHospital management system project report.pdf
Hospital management system project report.pdf
 
S1S2 B.Arch MGU - HOA1&2 Module 3 -Temple Architecture of Kerala.pptx
S1S2 B.Arch MGU - HOA1&2 Module 3 -Temple Architecture of Kerala.pptxS1S2 B.Arch MGU - HOA1&2 Module 3 -Temple Architecture of Kerala.pptx
S1S2 B.Arch MGU - HOA1&2 Module 3 -Temple Architecture of Kerala.pptx
 
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...Max. shear stress theory-Maximum Shear Stress Theory ​  Maximum Distortional ...
Max. shear stress theory-Maximum Shear Stress Theory ​ Maximum Distortional ...
 
fitting shop and tools used in fitting shop .ppt
fitting shop and tools used in fitting shop .pptfitting shop and tools used in fitting shop .ppt
fitting shop and tools used in fitting shop .ppt
 
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptxOrlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
Orlando’s Arnold Palmer Hospital Layout Strategy-1.pptx
 
Thermal Engineering Unit - I & II . ppt
Thermal Engineering  Unit - I & II . pptThermal Engineering  Unit - I & II . ppt
Thermal Engineering Unit - I & II . ppt
 
AIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech studentsAIRCANVAS[1].pdf mini project for btech students
AIRCANVAS[1].pdf mini project for btech students
 
Computer Networks Basics of Network Devices
Computer Networks  Basics of Network DevicesComputer Networks  Basics of Network Devices
Computer Networks Basics of Network Devices
 
Thermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - VThermal Engineering-R & A / C - unit - V
Thermal Engineering-R & A / C - unit - V
 
Jaipur ❤CALL GIRL 0000000000❤CALL GIRLS IN Jaipur ESCORT SERVICE❤CALL GIRL IN...
Jaipur ❤CALL GIRL 0000000000❤CALL GIRLS IN Jaipur ESCORT SERVICE❤CALL GIRL IN...Jaipur ❤CALL GIRL 0000000000❤CALL GIRLS IN Jaipur ESCORT SERVICE❤CALL GIRL IN...
Jaipur ❤CALL GIRL 0000000000❤CALL GIRLS IN Jaipur ESCORT SERVICE❤CALL GIRL IN...
 
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
NO1 Top No1 Amil Baba In Azad Kashmir, Kashmir Black Magic Specialist Expert ...
 
Online electricity billing project report..pdf
Online electricity billing project report..pdfOnline electricity billing project report..pdf
Online electricity billing project report..pdf
 
A Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna MunicipalityA Study of Urban Area Plan for Pabna Municipality
A Study of Urban Area Plan for Pabna Municipality
 

A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Output Networks

  • 1. IOSR Journal of Computer Engineering (IOSR-JCE) e-ISSN: 2278-0661,p-ISSN: 2278-8727, Volume 17, Issue 3, Ver. VII (May – Jun. 2015), PP 63-69 www.iosrjournals.org DOI: 10.9790/0661-17376369 www.iosrjournals.org 63 | Page A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Output Networks Mohammad Fazel Younessy1 and Ehsanolah Kabir2 1 Department of Electrical and Computer Engineering, Islamic Azad University, Science and Research Branch, Tehran, Iran 2 Department of Electrical Engineering, TarbiatModarres University, Tehran, Iran Abstract:Conventional classifiers have several outputs equal to the number of classes (N) which imposes so much complexityto the classifiers,both in training and testing stages. Instead of using one extra-large and complicated classifier, we created N number of simple binary Recurrent Neural Network with true/false outputs. Each network is responsible of recognizing its own trained class.We also proposed adecision layer added to each network, making a final decision from a sequence of outputs. We test our system on features extracted from Iranshahr database, which is a set of 17,000 black-and-white handwritten images of Iranian city names. Experimental results authenticate the effectiveness of the proposed method. Keywords: Recurrent Neural Network,RNN, Classifier, Binary Network. I. Introduction Classifiers are essential tools for the applications that need to find a specific type of result among a given set of results, such as offline/online handwritten/typewritten word recognition[1][2][3][4], speech recognition[5][6][7][8], and medical diagnosis systems[9][10][11]. Examples of such classifiers are Artificial Neural Network (ANN), Hidden Markov Models (HMM),support vector machine (SVM), and k-nearest neighbor (k-NN). Generally, a classifier has a training phase, in which some input samples with known results are fed into the classifier and the classifier internal weights would be trained by such data. Afterwards, in the testing phase, the input samples with unknown output results would be applied to the classifier, and the results would be evaluated. Conventional classifiers are consisted of high-quantity outputs, which force the classifier to be internally complicated and huge. Such classifiers all suffer from poor results and slow timing, both in training stage and testing phase. In our paper, we have proposed a new approach to conquer such problems. Instead of such complicated networks, we make multiple number of networks, each only consisted of a True/False output. It means, instead of one classifier with N number of outputs, we use N classifiers with simple outputs.Each classifier would be trained to have true result only by one class. This leads to small and simple classifiers with faster response time. One of the most common classifiers used in practice, is ANN. An ANN is consisted of units called neurons, which are arranged in layers. Typically, ANN converts an input vector (features) to outputs. The classifier type that we used in our work is a Recurrent Neural Network (RNN). A simple RNN is a special type of ANN, with oneor more memory layer(s) added to its feedforward layers.The memory layer is a feedback path, which makes the recurrence of the system. A typical ANN has an input vector, which makes one output (vector), whereas, RNN has a sequence of input vectors, which makes series of outputs. Hence, we added a decision layer to the output of the network to make the final decision (true or false) out of the sequence of inputs. Some other researchers also tried to make a classifier by RNN system. However, they all seem to be some complex mathematical problem, which would be added to the inherent complicated nature of conventional networks. For example, in [12], the authors proposed an RNN classifier with several recurrent layers added to each other.A valuable system has been presented in[13][14][15]. This system has addeda decision layer called connectionist temporal classification (CTC) to the output of the RNN and tried to train the network by a differentiable function. A rapid temporal sampling combined by an RNN classifier has been used for measurement of areal snow coverage in [16]. The authors in [17] replaced the simple neurons in RNN with a more complicated neuron set, and called it long-short term memory (LSTM) system, and used it in manuscript recognition. The rest of the paper would be organized as follows. In next section we will describe the proposed classification method in detail.In Section 3, we will show our experimental results on a real database. Finally we conclude our work in section four, and show a possible future work in that section.
  • 2. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 64 | Page II. Proposed ClassifierSystem In this section we will focus on details of our proposed system. The task of the system is to make a classifier to classify N outputs. Block diagram of the system is shown inFig. 1. As this figure shows, the system is composed of a training phase and testing phase. Fig. 1: Block diagram of proposed system. In training stage, at first, input data vector would be represented to be fed properly to the network. As mentioned, in our system, instead of using one huge classifier with N outputs, N number of simple classifiers would be used. Each network is an RNN classifier with a decision layer added to the output layer. The RNN set would be expressed as below: RNNs: RNN1, … , RNNN ( 1 ) Each RNN classifier is responsible for one class. Each classifier will have “True/False” outputs. This means in test phase, we expect that if samples of class i is applied to the RNNi, then output should be true and if other samples from another classes are fed into the network, then output should be false. This is an interesting phenomenon, because each network would be simple, and only is trained for one class. RNN networks normally need a sequence of inputs and make a sequence of outputs. Our task is to make a classifier, so the network should have one output (true or false), not a train of outputs. As a result, we made a decision layer added to input of each RNN. Decision layer would use the sequence of outputs from each RNN and finally make a final decision which might be true or false. In the testing phase, similar to the training phase, we send the input vectors to the representation unit, and then to each RNN classifiers (Recall that each RNN has a decision layer at the output). After this stage, we have a list of outputs derived from the RNN classifiers. We sort the outputs in the final decision section and output of the system would be a sorted list of probabilities of each class. In the following sub-sections we will discuss each part of the main block diagram in detail. A. Input Representation Input data to the RNN network is a sequence of vectors. Let’s define Fi as an input vector with the dimension of U × 1. A sequence of Fi, with length of V would make the input data of the network. Note that, interestingly,V is dependent on the input vector sequence’s length, and may be different from sample to sample. Equation below shows the idea: 𝔽 = F1: … : FV Fi = f1 ⋮ fU ( 2 ) Before sending the input vectors to the networks, we standardized the inputs. We forced the inputs to have a mean (μ) of 0, and standard deviation (σ) of 1. This procedure would make data optimum for the
  • 3. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 65 | Page activation functions inside the RNNs, and main information will not be changed, however, performance of system would improve significantly[18]. B. Network Architecture Normally, any arbitrary network can be made with any configuration. We have chosen a heuristic network topology which is presented in Fig. 2.As this figure shows, the network is consisted of two feedforward layers, hidden layer and output layer.The network also has a feedback (recurrent) layer, with delay of Tdelay which has been fed back to the hidden layer (layer 1). As mentioned earlier, output layer of each RNN has two outputs, one showing the state of being true, and the other one the state of being false. Fig. 2:Topology of the RNN (Figure from[19]). For each feedforward layer we use Tanh-Sigmoid transfer function[20], which has a output between -1 and +1. This makes the final outputs of the network to be in the range of −1, +1 , which gives us the probability of the true/false instead of a hard output of being discrete 0 or 1. C. Decision Layer Naturally, an RNN has a train of input vectors, and this leads to a sequence of 2-dimensional (2D) vectors on the output. As a classifier we need to concatenate the result sequence and make a decision at the end of output sequence. Hence we made a decision layer at the end of each RNN.As shown inFig. 3, a decision block is added to the RNN block. Fig. 3: Single RNN with decision layer added to the outputs. Suppose V is the length of input sequence. We call the outputs of the RNN as y1iand y2i, which are the true and false outputs of the RNN respectively. We have: Yi = 1 , if y1i > y2i 0 , else Yout = Yi V − Tdelay ∙ ωi Ω × 100 V i=Tdelay Ω = ωi V i=Tdelay ( 3 ) Looking at ( 3 ), Yi is the decision made by a single vector at each output of the RNN.Yout is the final decision of the network taken from train of 2D outputs. Please note that Tdelay is the network recurrence time, and is a system parameter. Normally outputs before that time is of less value, so we don’t consider them.As above formula shows, output ofdecision layer, Yout , is in fact the ratio of the number of true outputs to the sequence length. We added a weight coefficient, ωi, to the calculation to improve the results. These weights, should go from a small value to bigger values, as the network makes output through the time sequence. This 𝐼𝑛𝑝𝑢𝑡 𝑆𝑒𝑞𝑢𝑒𝑛𝑐𝑒 𝑈×𝑉 𝑅𝑁𝑁𝑖 𝑇𝑟𝑢𝑒 𝐹𝑎𝑙𝑠𝑒 𝐷𝑒𝑐𝑖𝑠𝑖𝑜𝑛 𝐿𝑎𝑦𝑒𝑟 𝑦1𝑖 𝑦2𝑖 𝑌𝑜𝑢𝑡 𝑅𝐸𝐿 𝑜𝑢𝑡
  • 4. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 66 | Page phenomenon guarantees that the network’s output at the end of the sequence has more influence on the final decision. This is due to the fact that, at the start of the network sequence evaluation, the recurrent characteristics of the network is not so strong, but, as more input vectors (from the input sequence) are fed into the network, network has more reliable outputs. At the end, to omit the effect of the weights, we have normalized the weights by summation of ωi, Ω, to remove the effect of the weights.In our systems we used three functions for the weights: 1. ωi = log(i), 2. ωi = exp(i), and 3. ωi = tanh(i/N) As a result of all mentioned, Yout shows the percentage of similarity of each input sequence, to the class that RNN network is trained with. It means, ideally, if a sequence of inputs similar to the specific RNN’s class is fed, output would be 100% and for the other sequences, output is 0%. Because of comparison in the first line of ( 3 ), decision made by above formula, is in fact a hard decision. To preserve the soft information of RNN outputs, we created a reliability function, RELout . This functions shows how much valid the output of the decision layer is. We formulate RELout as below: RELout = 1 2 y1i − y2i V − Tdelay V i=Tdelay ∙ ωi Ω × 100 ( 4 ) As the above formula shows, RELout is, in fact, the average of distance of true outputs and false outputs, multiplied by the weights. Please note that the coefficient of 1 2 is due to the fact that y1i and y2i are in the range of −1, +1 . D. RNN Training Some special attention should be given to the training method of the RNNs. Generally, in a database, there are a few samples for each class. In our case, each RNN is responsible for a specific class. If we give the few related samples to this RNN, forcing the outputs to true (for backpropagation training), and force the outputs to be false for the many unrelated samples, the network would be finally trained to false results. For examples, 3% of the input samples would train the network to true, and 97% of the other samples would train the network to false outputs. This means network will generate false results at the test time. Hence, we have added the quantity of true samples virtually, by adding some Gaussian noise to the true samples, making them about one-third of the total samples. To initialize the weights of the network, we used a Gaussian distribution with μ, σ = (0, 0.1). Furthermore, we can retrain the network after one successful training phase. Our simulation showed this technic may increase total result by 2-3%, and also retraining more than 3 retraining has no sensible effect. The negative point about the retraining is the time it consumes because of repeating the training phase. Several algorithms for network training has been proposed in scientific texts, such as Levenberg- Marquardt[21], Bayesian Regularization[22],[23], BFGS Quasi-Newton[24], Gradient Descent[25] and Resilient Backpropagation(Rprop)[26]. Among those, the Rprop seems to work better for us.Rprop algorithm normally uses relatively small amount of memory and is faster than most of the algorithms. E. RNN Testing To test the Network, we divided the sample data to 3 sets, 70% for training stage, 15% forvalidation of training phase, and 15% for final test and report. II. Experimental Results A. Test Database To test our system, we have used a real database called Iranshahr. This database is consisted of more than 17,000 samples of binary images of handwritten Iranian city names, which has been also used in [27]. The number of classes (cities) in this database is more than 500, which means each class has about 30 samples. We have used the method in [28] to extract the feature vectors. This method uses the sliding window technic to extract image features inside a frame. Then by shifting the window from right to left of the image, makes a sequence of vectors for each sample image. These features are input vector sequences to our system. B. Classification Results
  • 5. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 67 | Page Table 1 shows the best classification result obtained by the system. The second row in this table shows the classification rate (the percentage of recognizing correct city classes in the database) and the third row shows the average reliability of the correct results. As this table shows, the recognition rate and reliability of our system is relatively high and acceptable. Table 1: Classification results Top 1 Top 2 Top 5 Top 10 Classification rate 83.9% 84.7% 87.5% 91.1% Averagereliability 72.3% 73.1% 78.5% 80.5% C. Effect of Recurrence Delay A system parameter that has been taken into consideration is the Tdelay , which is the network recurrence delay. The effect of this parameter has been shown in Fig. 4. This figure shows that forTdelay = 3, the rate is at its maximum. Please note that, obviously, if we increase Tdelay , we will miss some recurrence statesand information of the RNN, and thus decreasing the classification rate. Fig. 4: Effect of Tdelay on classification rate. D. Effect of Retraining Figure below shows the effect of retraining on our system rate. We can see retraining of the network up to 3 times is a good practice, and more than that has almost no effect on system. Fig. 5: Effect of retraining times on classification rate. E. Effect of Weights We have tested our system by 3 weight types. Table below shows the effect of each weight type on our system. As the table shows, the logarithm function works the best for us and exponential the worst. 0 1 2 3 4 5 0 20 40 60 80 100 Tdelay Classificationrate(%) Top 1 Top 2 Top 5 Top 10 0 1 2 3 4 5 0 20 40 60 80 100 Retraining times Classificationrate(%) Top 1 Top 2 Top 5 Top 10
  • 6. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 68 | Page Table 2: Effect of weight function on classification rate. Classification rate Top 1 Top 2 Top 5 Top 10 Weight type ωi = exp(i) 75.3% 76.0% 79.6% 80.5% ωi = tanh(i/N) 80.2% 81.1% 84.1% 85.5% ωi = log(i) 83.9% 84.7% 87.5% 91.1% F. Comparison with conventional RNN Classifier To compare our method with the conventional classification method, i.e. using one huge classifier with so many outputs, we trained an RNN network with output layer of 500 neurons (equal to the number of classes in database), and added a decision layer to make the decision. The classification rate was very poor (less than 50%), and shows that our method is strongly better than the conventional classification method which uses RNNs. III. Conclusion and Future Work In this paper we proposed a good method of reducing the network complexity both in training and testing stages by making several binary networks instead of one complicated network. We used a Recurrent Neural Network as a classifier and added a decision layer to make the decision. We conducted our test in a real database to show the power of the method and a good recognition rate has been derived from the system. As a future work, researchers could use our method in the field of handwritten/typewritten/speech recognition systems. Even some extra information, such as length of the inputs,can prune some RNN classifiers, so the final result would be chosen from a reduced set of the networks. References [1] Kumawat, P.; Khatri, A.; Nagaria, B., "Comparative Analysis of Offline Handwriting Recognition Using Invariant Moments with HMM and Combined SVM-HMM Classifier," Communication Systems and Network Technologies (CSNT), 2013 International Conference on , vol., no., pp.140,143, 6-8 April 2013. [2] Maqqor, A.; Halli, A.; Satori, K.; Tairi, H., "Off-line recognition handwriting Arabic words using combination of multiple classifiers," Information Science and Technology (CIST), 2014 Third IEEE International Colloquium in , vol., no., pp.260,265, 20- 22 Oct. 2014 [3] Wu Wei; Gao Guanglai, "Online Handwriting Mongolia words recognition based on HMM classifier," Control and Decision Conference, 2009. CCDC '09. Chinese , vol., no., pp.3912,3914, 17-19 June 2009. [4] Al-Hajj Mohamad, R.; Likforman-Sulem, L.; Mokbel, C., "Combining Slanted-Frame Classifiers for Improved HMM-Based Arabic Handwriting Recognition," Pattern Analysis and Machine Intelligence, IEEE Transactions on , vol.31, no.7, pp.1165,1177, July 2009 [5] Chung-Hsien Wu; Wei-Bin Liang, "Emotion Recognition of Affective Speech Based on Multiple Classifiers Using Acoustic- Prosodic Information and Semantic Labels," Affective Computing, IEEE Transactions on , vol.2, no.1, pp.10,21, Jan.-June 2011. [6] Tsang-Long Pao; Wen-Yuan Liao; Yu-Te Chen; Jun-Heng Yeh; Yun-Maw Cheng; Chien, C.S., "Comparison of Several Classifiers for Emotion Recognition from Noisy Mandarin Speech," Intelligent Information Hiding and Multimedia Signal Processing, 2007. IIHMSP 2007. Third International Conference on , vol.1, no., pp.23,26, 26-28 Nov. 2007. [7] Chandrakala, S.; Sekhar, C.C., "Combination of generative models and SVM based classifier for speech emotion recognition," Neural Networks, 2009. IJCNN 2009. International Joint Conference on , vol., no., pp.497,502, 14-19 June 2009. [8] Kucukbay, S.E.; Sert, M., "Audio-based event detection in office live environments using optimized MFCC-SVM approach," Semantic Computing (ICSC), 2015 IEEE International Conference on , vol., no., pp.475,480, 7-9 Feb. 2015 [9] Deng, Z.; Cao, L.; Jiang, Y.; Wang, S., "Minimax Probability TSK Fuzzy System Classifier: A More Transparent and Highly Interpretable Classification Model," Fuzzy Systems, IEEE Transactions on , vol.PP, no.99, pp.1,1. [10] Gladence, L.M.; Ravi, T.; Karthi, M., "An enhanced method for detecting congestive heart failure - Automatic Classifier," Advanced Communication Control and Computing Technologies (ICACCCT), 2014 International Conference on , vol., no., pp.586,590, 8-10 May 2014. [11] AlObaidi, A.T.S.; Mahmood, N.T., "Modified Full Bayesian Networks Classifiers for Medical Diagnosis," Advanced Computer Science Applications and Technologies (ACSAT), 2013 International Conference on , vol., no., pp.5,12, 23-24 Dec. 2013 [12] Seong-Whan Lee; Hee-Heon Song, "A new recurrent neural-network architecture for visual pattern recognition," Neural Networks, IEEE Transactions on , vol.8, no.2, pp.331,340, Mar 1997. [13] Graves, Alex, Santiago Fernández, Faustino Gomez, and Jürgen Schmidhuber. "Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks." In Proceedings of the 23rd international conference on Machine learning, pp. 369-376. ACM, 2006. [14] Graves, Alex, Marcus Liwicki, Santiago Fernández, Roman Bertolami, Horst Bunke, and Jürgen Schmidhuber. "A novel connectionist system for unconstrained handwriting recognition." Pattern Analysis and Machine Intelligence, IEEE Transactions on 31, no. 5 (2009): 855-868. [15] A. Graves. Supervised Sequence Labelling with Recurrent Neural Networks. Dissertation, Technische Universität München, München, July 2008. [16] Simpson, J.J.; McIntire, T.J., "A recurrent neural network classifier for improved retrievals of areal extent of snow cover," Geoscience and Remote Sensing, IEEE Transactions on , vol.39, no.10, pp.2135,2147, Oct 2001. [17] Frinken, Volkmar, Francisco Zamora-Martinez, Salvador Espana-Boquera, Maria Jose Castro-Bleda, Andreas Fischer, and Horst Bunke. "Long-short term memory neural networks language modeling for handwriting recognition." In Pattern Recognition (ICPR), 2012 21st International Conference on, pp. 701-704. IEEE, 2012. [18] LeCun, L. Bottou, G. Orr, and K. Muller. Ecient backprop. In G. Orrand M. K., editors, Neural Networks: Tricks of the trade. Springer, 1998b.
  • 7. A New Classifier Based on Recurrent Neural Network Using Multiple Binary-Output Networks DOI: 10.9790/0661-17376369 www.iosrjournals.org 69 | Page [19] MathWorks™, (2012). Neural Network Toolbox7User’s Guide (R2012b). [20] Vogl, T.P., J.K. Mangis, A.K. Rigler, W.T. Zink, and D.L. Alkon, "Accelerating the convergence of the backpropagation method," Biological Cybernetics, Vol. 59, 1988, pp. 257–263 [21] Kenneth Levenberg (1944). "A Method for the Solution of Certain Non-Linear Problems in Least Squares". Quarterly of Applied Mathematics 2: 164–168 [22] MacKay, Neural Computation, Vol. 4, No. 3, 1992, pp. 415–447 [23] Foresee and Hagan, Proceedings of the International Joint Conference on Neural Networks, June, 1997 [24] Gill, Murray, & Wright, Practical Optimization, 1981. [25] Jan A. Snyman (2005). Practical Mathematical Optimization: An Introduction to Basic Optimization Theory and Classical and New Gradient-Based Algorithms. Springer Publishing. [26] Riedmiller, M., and H. Braun, “A direct adaptive method for faster backpropagation learning: The RPROP algorithm”, Proceedings of the IEEE International Conference on Neural Networks, 1993. [27] Bayesteh, E.; Ahmadifard, A.; Khosravi, Hossein, "A Lexicon Reduction Method Based on Clustering Word Images in Offline Farsi Handwritten Word Recognition Systems," Machine Vision and Image Processing (MVIP), 2011 7th Iranian , vol., no., pp.1,5, 16-17 Nov. 2011. [28] M. Dehghan, K. Faez, M. Ahmadi, and M. Shridhar, “Handwritten Farsi (Arabic) Word Recognition: A Holistic Approach Using Discrete HMM,” Pattern Recognition, vol. 34, no. 5, pp. 1057 1065, 2001.