2. Choursiya and Gupta: Prediction of Neurological Disorder using Classification Approach
AJCSE/Oct-Dec/Vol 6/Issue 4 56
different areas such as restorative translation and
genomics, the calculation of deep leanings can be
flawlessly coordinated with the human cerebrum
workings which have numerous neural network
layers and make vary from machine learning.
Saloni et al.[3]
in their analysis work, they planned
to discover the examples shaped by various
characterization algorithms. As per them, different
examples help in choosing best one from all. To
the training dataset, the component importance is
connected.Differentfeaturedeterminationalgorithm
is worked in which Fisher sifting is observed before
a decent element positioning network.
Suthaharan[7]
provided a recommendation that
classification with the employ of a quantity of
probabilistic neural network (PNN) variations to
separate between healthy peoples and Parkinson’s
disease (PD) patients. Three PNN types have been
availed as a part of this classification process,
identified with the smoothing factor search:
Incremental search Monte Carlo search and half
breed search. Concrete application has given
conclusion accuracy running in the vicinity of
79% and 81% for new and undiscovered patients.
Arel et al.[8]
employed a deep belief network
(DBN) to anticipate subcortical structures of
PD patients situated on microelectrode records
acquired amid deep brain stimulation.
Wang[4]
depicted diverse types of classification
strategies and their examination for effective
analysis about PD. The reliable analysis about
Parkinson disease abides hard enough to
accomplish with misdiagnosis result to be as high
as 28% of cases. The methodology is shown in
this manuscript reason to productively recognize
normal peoples.
Itisintendedforimagesinwhichnon-doubleesteems
can be covenant with as probabilities (which is not
the situation for characteristic images); its utilization
of best down input amid discernment is constrained
to the affiliated memory in the main two layers; it
does not have a methodical method for managing
perceptual in fluctuations; it expects that division has
just been performed and it does not figure out how to
consecutively take care for on the whole useful parts
of articles when separation is troublesome.[9-12]
PROPOSED METHODOLOGY
In the study of classification, herewith planned to
decide and determine pattern by the classification
algorithm. To the preparation dataset, the
classification algorithm endures connected.
Feature extraction, selection in addition to
classification is the fundamental advances.
Data fraction is additionally utilized as a part of
proposed novel technique and abide allied en route
for the preparation set. On the off chance so that
affecting dataset is not grouped effectively, at that
point, it is increase prepared to information part.
Different advances and calculations have been
connected in the territory of neurological disorder,
including information mining calculations such as
K-NN, PCA, Random Forest, DBN, and Neural
Network.[13-15]
The K-NN working is explained on the basis of
the below steps:
o Step-1: Select the number K of the neighbors
o Step-2: Calculate the Euclidean distance of K
Number of neighbors
o Step-3: Take the K nearest neighbors as per
the calculated Euclidean distance.
o Step-4: Among these k neighbors, count the
number of the data points in each category
o Step-5: Assign the new data points to that
category for which the number of neighbors is
maximum.
o Step-6: The model is ready.
Minimax Algorithm
Minimax algorithm is a recursive or backtracking
algorithm which is used in decision-making and
game theory. It provides an optimal move for the
player assuming that opponent is also playing
optimally.[16,17]
Therefore, in bid to calculate the normalized value
of member of the set regarding observed values;
( )
min( )
max mi
Z
n( )
x x
x x
−
−
=
Where, Z is the normalized value of member of
set regarding observed values x, min and max are
the minimum and maximum values in x given its
range. To transform it into particular range, then;
The general formula is;
v΄= (v-min)/(max-min) * (newmax-newmin) +
newmin
Where, v = old variable, v΄ = transformed variable,
newmin = minimum in reference to the stabilized
dataset, newmax = maximum in reference to the
stabilized dataset.
3. Choursiya and Gupta: Prediction of Neurological Disorder using Classification Approach
AJCSE/Oct-Dec-2021/Vol 6/Issue 4 57
v = [min, max]
v΄ = [newmin, newmax]
Basically, Minmax normalization is a strategy
which transforms linearly from x to y = (x min)/
(max-min), where min and max are the minimum
and maximum values in X, where X is the set
regarding pragmatic values of x. It can be easily
being observed that when x = min, then, y = 0
and when x = max, then, y = 1, this means that
the smallest value in X is epitomized to 0 and the
highestvalueinXisepitomizedto1.Consequently,
the entire ranges of the values of X from min to
max are epitomized to the range 0 to 1.
Minmax normalization aspires en route for
scaling every numerical values of v concerning an
arithmetical aspect A to a precise scope denoted
by [new-min A and new-max A].
The following expression transforms v to the new
value v΄:
v
( min ) min
A
A A
A A A
V min
max min
new max n n
´
ew ew
−
−
− − + +
=
−
Algorithm-I: Min-max normalization algorithm
for prediction of occurrence
Min-max Normalization
BEGIN:
Input: = Each value of data block
Max value of corresponding data block
Min value of corresponding data block
Output: = scaled value
1. Set upper along with lower limit (m,n)//
specific range
2. Identify Max plus Min values (Xmax
, Xmin
)
3. For each (data item) do
( )
*
min
max min
Y n m
X X
X X
m
−
= − +
−
End
For each test data –
– Use Euclidean distance to calculate the
distance between test data and each row of
training. The Euclidean method is the most
used when calculating distance.
– Sorting data set in ascending order based on
the distance value.
– From the sorted data array, choose the topmost
K rows.
– Based on the most appearing class of these
rows, it will assign a class to the test point.
After the above steps, the algorithms are going
to the dimensionally reduction process where the
process of the reducing the variable is done. The
data set is classified and prepared for the analysis
stage.
The customary class has been easily detected
then its goes to the separately normal class
otherwise if not detected then it goes to the
KNN-MinMax classifier. In this process, each
class has been accurately predicted among their
personal identity. After successful prediction,
the result analysis approach follows the detected
intrusions.
EXPERIMENTAL ANALYSIS
The data used in this study gathered from
188 patients with PD (127 men and 91 women)
with ages ranging from 30 to 80 (65.5A± 6.0). The
control group consists of 65 healthy individuals
(20 men and 35 women) with ages varying
between 40 and 80 (61.5 Â ± 8.5).
Attribute Information
Time frequency features, Mel frequency cepstral
coefficients, wavelet transform-based features,
vocal fold features, and TWQT features have been
applied to the neurological disorder (PD) patients
to extract clinically useful information for PD
assessment.[18,19]
Execution of Data Set and Result Analysis
The data were divided into training ~70% and
test data ~30%. The K-NN classifier and MinMax
method were used to store all the training data.
We used five-fold cross-validation method, which
is a reliable technique widely used to assess the
accuracy of the predictive system and for avoiding
the over fitting. The number of neighbors was
used to classify the new example and the distance
function was used to determine the nearest
neighbors. The best accuracy of 79% was obtained
when the number of neighbors taken was k = 1,
using the cosine distance.[20-22]
For this parameter, the comparison between
different existing, modified KNN-MinMax
method is performed in which it is found that the
accuracy rate of existing method is around 89%,
and anticipated scheme is regarding 82.12% that
means our proposed algorithm KNN-MinMax
achieve higher accuracy rate than the earlier
4. Choursiya and Gupta: Prediction of Neurological Disorder using Classification Approach
AJCSE/Oct-Dec/Vol 6/Issue 4 58
proposed methods, depicted in Table 1 which
shows the accuracy calculation for generally used
classification algorithms. It shows that the KNN
algorithm shows better performance as compared
to other classification algorithms.[23]
CONCLUSIONS
In this research study, a neurological disorder
diagnosis is acknowledged by realizing the KNN-
MinMax, which abides solitary regarding the
mainly earliest indicator for neurological disorder.
For this reason, a KNN-MinMax classifier, which
contains an amorphous auto encoder and a softmax
classifier, is proposed. A number of simulations
have been performed auxiliary than two databases
to show the adequacy of KNN-MinMax classifier.
The consequences of the proposed classifier are
contrasted and the aftereffects of the state-of-art
classification technique. The trial comes about
and factual examinations are demonstrated that
the KNN-MinMax classifier is exceptionally
productive classifier for neurological disorder
analysis which is around 0.25% more effective
than the existing proposed methods.
REFERENCES
1. Lokare V, Birari S, Patil O. Application of deep belief
networks for image compression. Int J Comput Sci
Inform Technol 2915;6:4799-803.
2. SriramTV, Rao MV, Narayana GV. Intelligent Parkinson
disease prediction using machine learning algorithms.
Int J Eng Innov Technol 2013;3:1-4.
3. Saloni, Sharma RK, Gupta AK. Voice analysis for
telediagnosis of Parkinson disease using artificial
neural networks and support vector machines. I J Intell
Syst Appl 2015;6:41-7.
4. Wang SC.Artificial neural network. In: Interdisciplinary
Computing in Java Programming. Berlin, Germany:
Springer; 2003. p. 81-100.
5. Shahbakhi M, Far DT, Tahami E. Speech analysis
for analysis for diagnosis of Parkinson’s disease
using genetic algorithm and support vector machine.
J Biomed Sci Eng 2014;7:147-56.
6. Kumar V, Verma L. Binary Classifiers for Health
care Databases: A comparative study of Data Mining
Classification Algorithms in the Diagnosis of Breast
Cancer; 2010.
7. Suthaharan S. Big Data Analytics. Machine Learning
Models and Algorithms for Big Data Classification.
United States: Springer; 2016. p. 31-75.
8. Arel I, Rose DC, Karnowski TP. Deep machine
learning-a new frontier in artificial intelligence research.
In: Computational Intelligence Magazine. New Jersey,
United States: IEEE; 2010. p. 13-8.
9. Caglar MF, Cetisli B, Toprak IB. Automatic
Recognition of Parkinson’s Disease from Sustained
Phonation Tests using ANN and Adaptive Neuro-
Fuzzy Classifier; 2010.
10. MovementDisorderSocietyTaskForceonRatingScales
for Parkinson’s Disease. Unified Parkinson’s Disease
Rating Scale (UPDRS): Status and Recommendations;
2003;18:738 –50
11. Little MA, McSharry PE, Hunter EJ,
Ramig LO. Suitability of dysphonia measurements
for telemonitoring of Parkinson’s disease. IEEE Trans
Biomed Eng 2009;56:1015.
12. Gorunescu M, Gorunescu F, Ene M, El-Darzi E.
A Probabilistic Neural Network-Based Method
for Hepatic Diseases Diagnosis, Buletinul Stiinti,
Universitatea din Pitesti, Seria Matematicasi
Informatica; 2006. p. 59-66.
13. Lee H, Ekanadham C, Ng AY. Sparse deep belief
net model for visual area V2. In: Proceedings of
the 21st
Annual Conference on Neural Information
Processing Systems (NIPS’ 07). United States: MIT
Press; 2007.
14. Kim Y, Lee H, Provost EM. Deep learning for robust
feature generation in audio-visual emotion recognition.
In: Proceedings of the IEEE International Conference
on Acoustics, Speech and Signal Processing (ICASSP’
13), Vancouver, Canada; 2013.
15. Hinton GE. A practical guide to training restricted
Boltzmann machines. In: Neural Networks: Tricks of
the Trade. Berlin, Germany: Springer; 2012. p. 599-619.
16. Kai X, Lei J, Yuqiang C, Wei X. Deep learning:
Yesterday, today, and tomorrow. J Comput Res Dev
2013;50:1799-804.
17. Fischer A, Igel C. Training restricted Boltzmann
machines: An introduction. Pattern Recognit
2014;47:25-39.
18. Yosinski J, Lipson H. Visually debugging restricted
Boltzmann machine training with a 3D example. In:
Representation Learning Work-shop, 29th
International
Conference on Machine Learning; 2012.
19. Tieleman T. Training restricted boltzmann Machines
using approximations to the likelihood gradient. In:
Proceedings of the 25th
International Conference on
Machine Learning. New York, United States: ACM;
2008. p. 1064-71.
20. Patra AK, Ray R, Abdullah AA, Dash SR. Prediction of
Parkinson’s disease using ensemble machine learning
classification from acoustic analysis. Int Conf Biomed
Table 1: Accuracy calculation using classification
algorithms
Algorithm Accuracy
(in%)
F1‑Score Recall Precision
Adoptive
Neural Theory
76.6 0.8484 0.8491 0.8484
k‑NN
Algorithm
81.52 0.8082 0.8152 0.8237
SVM‑kNN 72.22 0.8268 0.8322 0.8375
KNN‑MinMax 92.22 0.8644 0.8644 0.8564
5. Choursiya and Gupta: Prediction of Neurological Disorder using Classification Approach
AJCSE/Oct-Dec-2021/Vol 6/Issue 4 59
Eng 2019;1372:012041.
21. Zong Y, Zheng W, Zhang T, Huang X. Cross-corpus
speech emotion recognition based on domain-adaptive
least squares regression. IEEE Signal Process Lett
2016;23:584-8.
22. Hinton GE. To recognize shapes, first learn to generate
images. In: Computational Neuroscience: Theoretical
Insights into Brain Function. Amsterdam, Netherlands:
Elsevier; 2007.
23. Goschenhofer J, Pfister FM, Yuksel KA, Bisch B,
Fietzek U, Thomas J. Wearable-based Parkinson’s
disease severity monitoring using deep learning. In:
European Conference on Principles of Data Mining and
Knowledge Discovery; 2019. p. 400-15.