SlideShare a Scribd company logo
1 of 5
Download to read offline
Performance Evaluation Of Hepatitis Diagnosis
Using Single and Multi-Classifiers Fusion
Ahmed M. Salah EL-Bohy1
, Atallah I. Hashad2
, and Hussien Saad Taha3
Arab Academy for Science, Technology & Maritime Transport, Cairo, Egypt
1
a.bohy4master@gmail.com, 2
hashad@aast.edu, and3
htaha10@yahoo.com.
Abstract- The goal of our paper is to obtain superior accuracy of
different classifiers or multi-classifiers fusion in diagnosing Hepatitis
using world wide data set from Ljubljana University. We present
an implementation among some of the classification methods which
are defined as the best algorithms in medical field. Then we apply a
fusion between classifiers to get the best multi-classifier fusion
approach. By using confusion matrix to get classification accuracy
which built in 10-fold cross validation technique. The experimental
results show that for all data sets (complete, reduced, and no
missing value) using multi-classifiers fusion achieved better
accuracy than the single ones.
Keywords-Hepatitis; Classification techniques; Fusion ; WEKA.
I. INTRODUCTION
The diagnosis of some diseases like hepatitis is very
difficult task for a doctor, where doctors usually determine
decision by comparing the current test results of patients with
another one who has the same condition. Hepatitis is one of the
most common diseases among Egyptians; as it represents 22%
of hepatitis cases around the world. Which motivates us for
suggesting new methods to improve the outcomes of existing
approaches, as well as to help doctors and specialists to
diagnose hepatitis disease survival. [1]
Hepatitis (Greek) means 'liver' and suffix –itis denotes
'inflammation' of the liver and may be due to infectious Or non-
infectious causes. The five types of hepatitis viruses are
Common infectious causes of liver inflammation and some like
Hepatitis A (HAV), B (HBV) and C (HCV) are more frequently
seen infectious agents. Inflammation may lead to death of the
liver cells (hepatocytes) which severely compromises normal
liver function. Acute HBV Infection (less than 6 months) may
resemble the fever, flu, muscle aches, joint pains and generally
being unwell. Symptoms specifying that states are: dark urine,
loss of appetite, nausea, vomiting, jaundice, pain up the liver.
Chronic hepatitis B is an infection persisting more than 6
months, the clinical features of that state correspond to liver
dysfunction, so signs like these may be noticed: enlarged liver,
splenomegaly, hepatosplenomegaly, jaundice, weakness,
abdominal pain, confusion and abdominal swelling [2]. The
success of treatment depends on an early recognition of the
virus, which achieves more exact and less violent treatment
options and mortality from Hepatitis falls.
Recently, data-mining has become one of the most treasured
tools for operating data in order to produce valuable information
for decision-making [3]. Supervised learning, including
classification is one of the most significant brands in data mining,
with a recognized output variable in the dataset. Classification
methods can achieve high accuracy in classifying mortality cases.
The implementations had been applied with a “WEKA” tool
which stands for the Waikato Environment for Knowledge
Analysis. A lot of papers about applying machine learning
procedures for survivability analysis in the field of Hepatitis
diagnostic. Here are some examples:
Using Support Vector Machines And Wrapper Method for
predicting Hepatitis was introduced achieving maximum
accuracy of (74.55%). [4]. but we note that applying SVM
classifier only get higher accuracy than the mentioned
accuracies even with feature selection.
Improving the accuracy of SVM algorithm using feature
selection [5]. Using SVM with Chi-Square achieved accuracy
of (83.12 %), But we note that applying another classifier
(Logistic, Simple Logistic, SMO, RF, J48) gets higher
accuracy than the mentioned one.
Detection of Hepatitis based on SVM and Data Analysis [6]
using SVM & wrapper the achieved accuracy is (85%) but
we found the results are calculated with 7 attributes out of
25 ones as stated whenever the original number of attributes
is 20 only
The rest of this paper is prearranged like this: In sector II,
Classification algorithms are discussed. In sector III
Evaluation principles are discussed. In sector IV a proposed
model is shown. In sector V reports the experimental results.
Finally, Sector VI introduces the conclusion.
II. Classification Algorithms
A Bayesian belief network is sometimes named a Bayes
net, a belief net, or a causal network, It is a directed, acyclic
graph, indicating conditional dependencies. It can be used to
guess the probability of events. the Bayesian decision rule
assurances minimum error if likelihoods and prior probabilities
are known [7].
Decision Tree (DT) Tree that the root and each interior
node is marked with a question. The arcs represent each possible
answer to the concomitant question. Each leaf node represents a
forecast of a problem solution. A prevalent technique for
classification; Leaf node leads the class to which the
corresponding tuple belongs. It’s Model is a computational
model comprising of the three parts: Decision Tree Algorithm to
create the tree Algorithm that applies the tree to the data,
creation of the tree is the most exciting part. Processing is
mostly a search similar to that in a binary search tree (although
DT may not be binary). Advantages: Easy to understand. Easy to
generate rules[8]
Support Vector Machine (SVM) SVMs are amongst the
best (and many believe are definitely the best)“on-the-shelf ”
supervised learning algorithm. It’s derived from statistics in
1992.SVM is widely used in multiple applications pattern
recognition, classification and regression. The SVMs work on an
underlying principle, which is to insert a hyper-plane between
the classes and orient it in such a way so as to keep it at the
maximum distance from the nearest data points These data
points, which appear closest to the hyper-plane, are known as
Support Vectors. [9]
Sequential Minimal Optimization (SMO) humble, easy to
implement, is generally quicker, and has better scaling properties
for difficult SVM problems than the usual SVM training
algorithm SMO quickly solve the SVM QP problem without
extra matrix storage (the memory used is linear with the training
data set size) and without using numerical QP optimization
steps at all. SMO can be used for online learning. While SMO
has been shown to be operative on sparse data sets and
especially fast for linear SVMs, the algorithm can be extremely
slow on non-sparse data sets and on problems that have many
support vectors. Regression problems are especially prone to
these matters because the inputs are usually non-sparse real
numbers (as opposed to binary inputs) with solutions that have
many support vectors. Because of these restrictions, there have
been limited reports of SMO being successfully used on
regression problems. [10]
Logistic Regression (LR) is a famous well known
classifier, it could be used to extend classification results into a
deeper analysis. It’s not widely used due to its slow response in
data mining especially when compared with SVM in large data
sets (not our case) It models the relationship between a dependent
and one or more independent variables, and consents us to look at
the fit of the model as well as at the significance of the
relationships [11]
Simple Logistic: We use simple logistic regression when we
have one nominal variable with two values (dead/alive,
male/female) and one measurement variable. The nominal
variable is the dependent variable, but the measurement
variable is not it is an independent one.
Simple logistic regression is analogous to linear regression,
except the dependent variable is nominal, not a measurement.
[12]
Random Forest: Random forests change how the
classification or regression trees are constructed. In standard
trees, it uses the best split among all variables to split each node
In a random forest, each node is split using the best among a
subset of predictors arbitrarily chosen at that node. it works by
one of two methods, boosting and bagging.it has the advantages
of: handling thousands of input variables without deleting any
variable, giving estimation of variables importance in the
classification, and It also has an active method for estimating
missing data and keeps accuracy when a large amount of the data
are missing [13].
SGD is an abbreviation for Stochastic Gradient Descent:
the gradient of the loss is estimated each sample at a time and the
model is updated along the way with a decreasing strength
schedule. SGD has been excellently applied to large-scale and
sparse machine learning problems often come across the text
classification and natural language processing [14].
K-star or K* is an instance-based classifier. The class of a
test instance is based on the training instances similar to it, as
determined by some similarity function. The difference between
it and the other instance-based learners in that K* uses an
entropy-based distance function. Instance-based learners
classifying instance by matching it to a database of pre-
classified cases. The fundamental assumption is that similar
instances will have similar classifications [15].
III. EVALUATION PRINCIPLES
Evaluation method is based on the confusion matrix. The
confusion matrix is an imagining implement usually used to
show presentations of classifiers. It is used to display the
relationships between real class attributes and predicted classes.
The grade of efficiency of the classification task is calculated
with the number of exact and unseemly classifications in each
conceivable value of the variables being classified in the
confusion matrix [16]
Table 1. Confusion matrix
For instance, in a 2-class classification problem with two
predefined classes (e.g., Positive, negative) the classified test
cases are divided into four categories:
• True positives (TP) correctly classified as positive instances.
• True negatives (TN) correctly classified negative instances.
• False positives (FP) incorrectly classified negative instances
• False negatives (FN) incorrectly classified positive instances.
To evaluate classifier performance. We define accuracy
term that is defined as the entire number of truly classified
instances divided by the entire number of available instances for
an assumed operational point of a classifier.
Accuracy
TP TN
TP TN FP FN


  
(1)
IV. RROPOSED METHODOLOGY
We propose a reliable method for diagnosing Hepatitis and
better classifying mortality cases that may be caused by
Hepatitis infection based on data mining using WEKA as
follows:
A. Data preprocessing
Pre-processing steps are applied to the data before
classification:
First step is Data Cleansing: eliminating or decreasing
noise and the treatment of missing values. Hepatitis data set has
a lot of missing values, especially in attributes 18th, 15th, and
17th as they are 67, 29, 16 values respectively. After removing
the instances has these missing values the (155) instance data set
converted to (83) instances with minor missing values in other
attributes (Only 9 missing values in the reduced data set). We
Predicted Class
Negative Positive
Outcomes
Negative TN FN
Positive FP TP
also convert the data set into a new (third) one by removing these
(9) missing values to be (80) instance instead of (83) instance.
Second step is Relevance Analysis: Statistical correlation
analysis is used to discard the redundant features from further
analysis. The data set has one irrelevant attribute named ‘sex ’
which has no effect in the classification process;
Final, step is Data Transformation: The dataset is
transformed by normalization, which is one of the greatest public
tools used by inventors of automatic recognition classifications
to get superior results. Data normalization hurries up training
time by initialing the training procedure to reach feature within
the same scale. The aim of normalization is to transform the
attribute values to a small-scale range.
B. Single classification task
Data classification is the process of organizing data into
categories for its most effective and efficient use. A well-
structured data classification system makes it easy to get, handle,
and retrieve necessary data.
Classification consists of conveying a class label to a set of
unclassified cases. Supervised classification: The set of possible
classes is known in advance. Unsupervised classification: Set of
possible classes is unknown. After classification we can try to
assign a name to that class. Unsupervised classification is known
as clustering. The presumed model depends on analyzing the
training dataset. The derivative model characterized in several
procedures, such as simple classification rules, decision trees and
another. Basically data classification is a two-stage process, in
the initial stage; a classifier is built signifying a predefined set of
notions or data classes.
This is the training stage, where a classification technique
builds the classifier by learning from a training dataset and their
related class label columns or attributes. In next stage this model
is used for measurement. In order to deduce the predictive
accuracy of the classifier an independent set of the training
instances is used. We evaluate the most frequently classification
techniques that mentioned in recently published researches in the
medical field to accomplish the highest accuracy classifier’s
result with each data set.
C. Multi-classifiers fusion classification task
Fusion is a combination between 2 or more classifiers to
enhance classifier performance (Accuracy). Our procedure will
be electing the highest 2 classifiers in accuracy calculating
accuracy, then the 3rd, then 4th and so on until accuracy
decreases then stop & take the maximum accuracy obtained as
shown in Figure 1 We name the number of classifiers used in the
fusion process “fusion level” best accuracy with other single
classifiers predicting to improve accuracy. Repeating the same
process till the latest level of fusion, according to the number of
single classifiers to pick the highest accuracy through all
processes.
We propose 2 methods (1st Directly applying a classifier to
the raw data & 2nd preprocess the data before applying
classifier).
Import the Dataset.
 Convert nominal values to binary ones (in case of
preprocessing).
 Normalize each variable of the data set, so that the
values range from 0 to 1 (in case of preprocessing).
 Create a separate training (learn) and testing sets by
indiscriminately drawing out the data for training and
for testing.
 Picking and adapt the learning procedure
 Perform the learning procedure
 Calculate the performance of the model on the test set
Figure 1 .Proposed Hepatitis diagnosis algorithm
V. EXPERIMENTAL RESULTS
According to the mentioned work in sector IV proposed
methodology-A- data preprocessing we now have 3 data sets
with 155, 83, 80 instances respectively & last two data sets has
been preprocessed as mentioned in sector IV proposed
methodology-A- data transformation To calculate the proposed
model, first we calculate accuracy using 9 single classifiers then
we compute accuracy in fusion process up to 9 classifiers
together using the highest classifiers combination in each fusion
level (All results are illustrated graphically in Figure 2).
A. Single classification task
Table 2 shows the comparison between accuracies over
nine single classifiers (Bayes Net , SVM, Logistic, SGD, Simple
Logistic, SMO, K*, J48, and RF). Highest accuracy in each data
set is highlighted in yellow color in Table 2. Using 155 instances
(complete) data set the highest accuracy by SMO classifier is
85.17%. Using 83 instances (reduced) data set the highest
accuracy by K* classifier is 91.67%. Using 80 instances (no
missing value) data set the highest accuracy by Bayes Net
classifier is 88.75%.
SMO classifier achieves better accuracy than SVM
classifier illustrated in previous work for the three data sets
without any preprocessing or feature selection [4-6]
Furthermore, we note that RF classifier achieves the second
rank among single classifiers for the whole three data sets and
it’s highlighted in cyan color in Table 2.
Finally K* classifier bare as a good classifier as it achieves
highest accuracy in one data set and it got the seconed place in
another one equal with RF classifier in 80 instances (no missing
value) data set
Table 2 Single classifier results
# instance
Classifier
155
complete
83
Reduce
d
80
no missing value
BayesNet 83.21 86.81 88.75
SVM 79.38 82.08 83.75
Logistic 82.58 77.92 81.25
SGD 84.54 84.17 81.25
Simple Logistic 83.88 80.42 85.00
SMO 85.17 84.03 85.00
K* 81.96 91.67 87.50
J48 83.79 82.22 86.25
RF 85.13 90.56 87.50
Table 3 shows the comparison between accuracies over multi
classifiers fusion between classifiers up to nine classifiers all
possible combination has been executed as illustrated in
proposed algorithm Figure 1 and the table contains the highest
accuracy at each fusion level.
Highest accuracy in each data set for all fusion levels is
highlighted in yellow color in Table 3. Using 155 instances
(complete) data set the highest fusion accuracy obtained by
using K*, J48, RF, Simple Logistic, and Logistic classifiers is
87.04%. Using 83 instances (reduced) data set the highest
fusion accuracy obtained by using K*, RF, SVM, and
BayesNet classifiers is 92.92%. Using 80 instances (no missing
value) data set the highest fusion accuracy obtained by using
K*, J48, RF, and BayesNet classifiers is 91.25%.
The results in Table 3 indicates
First data preprocessing minimized the fusion level /
number of classifiers by one classifier as we get the best result /
highest accuracy at fusion level 4th level instead of 5th level in
case of 155 instances (complete) data set
Second K*and RF are performed as outstanding
classifiers in fusion for Hepatitis data set in all cases (complete,
Reduced, and no missing value) with / without preprocessing
Finally the importance of BayesNet classifier is appeared
for less or no missing values data as it increases accuracy by a
noticeable amount in both 83 instances (reduced) and 80
instances (no missing value) data sets
Table 3 Fusion classifiers results
# instance
Fusion Level
155
Complete
(%)
83
Reduced
(%)
80
no missing
value (%)
2nd
86.42 91.67 88.75
3rd
86.46 89.31 88.75
4th
86.42 92.92 91.25
5th
87.04 91.67 90.00
6th
87.04 90.42 90.00
7th
86.42 90.42 91.25
8th
85.75 86.67 88.75
9th
85.03 85.42 88.85
Figure 2 Single / Fusion classifier accuracy comparison
VI. CONCLUSION
The experimental results clarified that multi-classifiers
fusion reaches better accuracy than single classifier. For the three
data sets.
Although no unique classifier achieves the highest accuracy
as a single classifier for all data sets but RF classifier playing a
great role as a single classifier for the 3 mentioned data sets
(complete, reduced, and no missing value). Also the K* classifier
is a common master classifier between all fusion combination,
adding RF to it in fusion process gives an outstanding results.
BayesNet classifier consider as a break through dealing with
Hepatitis reduced and no missing value data sets in fusion as it
improves accuracy by 1.25% and 2.5% from best single classifier
respectively and increase it by 3.61% and 2.5% from pervious
fusion level
REFERENCES
[1] World Health Organization at http://www.who.int
[2] Tomasz KANIK. “Hepatitis disease diagnosis using Rough
Set modification of the pre-processing algorithm.”
Information and communication technologies international
conference (2012).
[3] Fayyad, Usama, Gregory Piatetsky-Shapiro, and Padhraic
Smyth."From data mining to knowledge discovery in
databases." AI magazine 17.3 (1996).
[4] A.H.Roslina and A.Noraziah. "Prediction Of Hepatitis
Prognosis Using Support Vector Machines And Wrapper
Method." Seventh International Conference on Fuzzy
Systems and Knowledge Discovery (2010).
[5] Varun Kumar.M , Vijay Sharathi.V and Gayathri Devi.B.R.
" Hepatitis Prediction Model based on Data Mining
Algorithm and Optimal Feature Selection to Improve
Predictive Accuracy." International Journal of Computer
Applications (0975 – 8887) Volume 51– No.19, August
2012.
[6] C. Barath Kumar , M. Varun Kumar , T. Gayathri, and S.
Rajesh Kumar " Analysis and Prediction of Hepatitis Using
Support Vector Machine." International Journal of
Computer Science and Information Technologies, Vol. 5
(2) , 2014, 2235-2237.
[7] Jyotirmay Gadewadikar, Ognjen Kuljaca1, Kwabena
Agyepong, Erol Sarigul, Yufeng Zheng and Ping Zhang,
Exploring Bayesian networks for medical decision support
in breast cancer detection, African Journal of Mathematics
and Computer Science Research Vol. 3(10), pp. 225-231,
October 2010.
[8] Ross Quinlan, (1993) C4.5: Programs for Machine Learning,
Morgan Kaufmann Publishers, San Mateo, CA.
[9] Chen Y., Wang G., and Dong S. “Learning with progressive
transductive support vector machine”, Pattern Recognition
Letters, Vol. 24, Pages: 1845-1855
[10] John C. Platt. “Sequential Minimal Optimization:A Fast
Algorithm for Training Support Vector Machines.”
Technical Report MSR-TR-98-14April.
[11] Mitchell T. “Machine Learning, mcgraw Hill.” (1997).
[12] John H. Mcdonald. “Handbook of Biological Statistics.” At
http://www.strath.ac.uk
[13] Breiman, Leo. "Random forests." Machine learning 45.1
(2001): 5-32.
[14] Y. Tsuruoka, J. Tsujii, and S. Ananiadou. " Stochastic
gradient descent training for l1-regularized log-linear
models with cumulative penalty " AFNLP/ACL 2009.
[15] Mahmood, Deeman Y., and Mohammed A. Hussein.
"Intrusion Detection System Based on K-Star Classifier
and Feature Set Reduction." International Organization
of Scientific Research Journal of Computer Engineering
(IOSR-JCE) Vol 15 (2013): 107-112.
[16] P. Cabena, P. Hadjinian, R. Stadler, J. Verhees and
A.Zanasi, Discovering data mining from concept to
implementation. Uppersaddle River, N.J.: Prentice Hall,
1998.

More Related Content

What's hot

Deep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionDeep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionManmeet Singh
 
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...AIRCC Publishing Corporation
 
Data Preparation and Processing
Data Preparation and ProcessingData Preparation and Processing
Data Preparation and ProcessingMehul Gondaliya
 
Exploratory data analysis data visualization
Exploratory data analysis data visualizationExploratory data analysis data visualization
Exploratory data analysis data visualizationDr. Hamdan Al-Sabri
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysisguest0edcaf
 
Marketing analytics - clustering Types
Marketing analytics - clustering TypesMarketing analytics - clustering Types
Marketing analytics - clustering TypesSuryakumar Thangarasu
 
Lecture 6: Ensemble Methods
Lecture 6: Ensemble Methods Lecture 6: Ensemble Methods
Lecture 6: Ensemble Methods Marina Santini
 
pratik meshram-Unit 5 (contemporary mkt r sch)
pratik meshram-Unit 5 (contemporary mkt r sch)pratik meshram-Unit 5 (contemporary mkt r sch)
pratik meshram-Unit 5 (contemporary mkt r sch)Pratik Meshram
 
Cluster Analysis Assignment 2013-2014(2)
Cluster Analysis Assignment 2013-2014(2)Cluster Analysis Assignment 2013-2014(2)
Cluster Analysis Assignment 2013-2014(2)TIEZHENG YUAN
 
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...IOSR Journals
 
Advanced Statistical Manual for Ayurveda Research
Advanced Statistical Manual for Ayurveda ResearchAdvanced Statistical Manual for Ayurveda Research
Advanced Statistical Manual for Ayurveda ResearchAyurdata
 
Deep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionDeep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionManmeet Singh
 

What's hot (19)

02 Related Concepts
02 Related Concepts02 Related Concepts
02 Related Concepts
 
Deep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionDeep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly Detection
 
Eda sri
Eda sriEda sri
Eda sri
 
Saif_CCECE2007_full_paper_submitted
Saif_CCECE2007_full_paper_submittedSaif_CCECE2007_full_paper_submitted
Saif_CCECE2007_full_paper_submitted
 
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...
Enactment Ranking of Supervised Algorithms Dependence of Data Splitting Algor...
 
Data Preparation and Processing
Data Preparation and ProcessingData Preparation and Processing
Data Preparation and Processing
 
Stat11t chapter2
Stat11t chapter2Stat11t chapter2
Stat11t chapter2
 
Exploratory data analysis data visualization
Exploratory data analysis data visualizationExploratory data analysis data visualization
Exploratory data analysis data visualization
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
 
Cluster Validation
Cluster ValidationCluster Validation
Cluster Validation
 
Marketing analytics - clustering Types
Marketing analytics - clustering TypesMarketing analytics - clustering Types
Marketing analytics - clustering Types
 
Lecture 6: Ensemble Methods
Lecture 6: Ensemble Methods Lecture 6: Ensemble Methods
Lecture 6: Ensemble Methods
 
pratik meshram-Unit 5 (contemporary mkt r sch)
pratik meshram-Unit 5 (contemporary mkt r sch)pratik meshram-Unit 5 (contemporary mkt r sch)
pratik meshram-Unit 5 (contemporary mkt r sch)
 
Cluster Analysis Assignment 2013-2014(2)
Cluster Analysis Assignment 2013-2014(2)Cluster Analysis Assignment 2013-2014(2)
Cluster Analysis Assignment 2013-2014(2)
 
Declarative data analysis
Declarative data analysisDeclarative data analysis
Declarative data analysis
 
Stat11t chapter3
Stat11t chapter3Stat11t chapter3
Stat11t chapter3
 
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
 
Advanced Statistical Manual for Ayurveda Research
Advanced Statistical Manual for Ayurveda ResearchAdvanced Statistical Manual for Ayurveda Research
Advanced Statistical Manual for Ayurveda Research
 
Deep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly DetectionDeep Semi-Supervised Anomaly Detection
Deep Semi-Supervised Anomaly Detection
 

Similar to Performance evaluation of hepatitis diagnosis using single and multi classifiers fusion final

Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...
Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...
Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...ahmedbohy
 
Highly Reliable Hepatitis Diagnosis with multi classifiers
Highly Reliable Hepatitis Diagnosis with multi classifiersHighly Reliable Hepatitis Diagnosis with multi classifiers
Highly Reliable Hepatitis Diagnosis with multi classifiersahmedbohy
 
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...ahmedbohy
 
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...ahmad abdelhafeez
 
A Formal Machine Learning or Multi Objective Decision Making System for Deter...
A Formal Machine Learning or Multi Objective Decision Making System for Deter...A Formal Machine Learning or Multi Objective Decision Making System for Deter...
A Formal Machine Learning or Multi Objective Decision Making System for Deter...Editor IJCATR
 
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...ahmad abdelhafeez
 
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...BRNSSPublicationHubI
 
Classification of Breast Cancer Diseases using Data Mining Techniques
Classification of Breast Cancer Diseases using Data Mining TechniquesClassification of Breast Cancer Diseases using Data Mining Techniques
Classification of Breast Cancer Diseases using Data Mining Techniquesinventionjournals
 
Top 20 Data Science Interview Questions and Answers in 2023.pdf
Top 20 Data Science Interview Questions and Answers in 2023.pdfTop 20 Data Science Interview Questions and Answers in 2023.pdf
Top 20 Data Science Interview Questions and Answers in 2023.pdfAnanthReddy38
 
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISAN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISAIRCC Publishing Corporation
 
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISAN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISijcsit
 
Hellinger distance based drift detection for
Hellinger distance based drift detection forHellinger distance based drift detection for
Hellinger distance based drift detection forDSPG Bangaluru
 
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINERANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINERIJCSEA Journal
 
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...ijcsa
 
IRJET - Survey on Analysis of Breast Cancer Prediction
IRJET - Survey on Analysis of Breast Cancer PredictionIRJET - Survey on Analysis of Breast Cancer Prediction
IRJET - Survey on Analysis of Breast Cancer PredictionIRJET Journal
 
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTION
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTIONMULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTION
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTIONIJDKP
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...ijceronline
 
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...IAEME Publication
 
Comparison between the genetic algorithms optimization and particle swarm opt...
Comparison between the genetic algorithms optimization and particle swarm opt...Comparison between the genetic algorithms optimization and particle swarm opt...
Comparison between the genetic algorithms optimization and particle swarm opt...IAEME Publication
 
Performance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning AlgorithmsPerformance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning AlgorithmsDinusha Dilanka
 

Similar to Performance evaluation of hepatitis diagnosis using single and multi classifiers fusion final (20)

Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...
Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...
Novel Frequency Domain Classification Algorithm Based On Parameter Weight Fac...
 
Highly Reliable Hepatitis Diagnosis with multi classifiers
Highly Reliable Hepatitis Diagnosis with multi classifiersHighly Reliable Hepatitis Diagnosis with multi classifiers
Highly Reliable Hepatitis Diagnosis with multi classifiers
 
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...
Hepatitis diagnosis based on Artificial Intelligence Using Single And multi c...
 
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
 
A Formal Machine Learning or Multi Objective Decision Making System for Deter...
A Formal Machine Learning or Multi Objective Decision Making System for Deter...A Formal Machine Learning or Multi Objective Decision Making System for Deter...
A Formal Machine Learning or Multi Objective Decision Making System for Deter...
 
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
Robust Breast Cancer Diagnosis on Four Different Datasets Using Multi-Classif...
 
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...
Impact of Classification Algorithms on Cardiotocography Dataset for Fetal Sta...
 
Classification of Breast Cancer Diseases using Data Mining Techniques
Classification of Breast Cancer Diseases using Data Mining TechniquesClassification of Breast Cancer Diseases using Data Mining Techniques
Classification of Breast Cancer Diseases using Data Mining Techniques
 
Top 20 Data Science Interview Questions and Answers in 2023.pdf
Top 20 Data Science Interview Questions and Answers in 2023.pdfTop 20 Data Science Interview Questions and Answers in 2023.pdf
Top 20 Data Science Interview Questions and Answers in 2023.pdf
 
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISAN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
 
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISAN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSIS
 
Hellinger distance based drift detection for
Hellinger distance based drift detection forHellinger distance based drift detection for
Hellinger distance based drift detection for
 
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINERANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
ANALYSIS AND COMPARISON STUDY OF DATA MINING ALGORITHMS USING RAPIDMINER
 
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...
PERFORMANCE ANALYSIS OF MULTICLASS SUPPORT VECTOR MACHINE CLASSIFICATION FOR ...
 
IRJET - Survey on Analysis of Breast Cancer Prediction
IRJET - Survey on Analysis of Breast Cancer PredictionIRJET - Survey on Analysis of Breast Cancer Prediction
IRJET - Survey on Analysis of Breast Cancer Prediction
 
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTION
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTIONMULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTION
MULTI MODEL DATA MINING APPROACH FOR HEART FAILURE PREDICTION
 
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...IJCER (www.ijceronline.com) International Journal of computational Engineerin...
IJCER (www.ijceronline.com) International Journal of computational Engineerin...
 
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...
COMPARISON BETWEEN THE GENETIC ALGORITHMS OPTIMIZATION AND PARTICLE SWARM OPT...
 
Comparison between the genetic algorithms optimization and particle swarm opt...
Comparison between the genetic algorithms optimization and particle swarm opt...Comparison between the genetic algorithms optimization and particle swarm opt...
Comparison between the genetic algorithms optimization and particle swarm opt...
 
Performance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning AlgorithmsPerformance Comparision of Machine Learning Algorithms
Performance Comparision of Machine Learning Algorithms
 

Recently uploaded

Thermal Engineering Unit - I & II . ppt
Thermal Engineering  Unit - I & II . pptThermal Engineering  Unit - I & II . ppt
Thermal Engineering Unit - I & II . pptDineshKumar4165
 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapRishantSharmaFr
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756dollysharma2066
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...Call Girls in Nagpur High Profile
 
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...tanu pandey
 
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoorTop Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoordharasingh5698
 
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank  Design by Working Stress - IS Method.pdfIntze Overhead Water Tank  Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank Design by Working Stress - IS Method.pdfSuman Jyoti
 
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...SUHANI PANDEY
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startQuintin Balsdon
 
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Bookingroncy bisnoi
 
Unit 2- Effective stress & Permeability.pdf
Unit 2- Effective stress & Permeability.pdfUnit 2- Effective stress & Permeability.pdf
Unit 2- Effective stress & Permeability.pdfRagavanV2
 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...roncy bisnoi
 
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Bookingroncy bisnoi
 
AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdfankushspencer015
 
Work-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptxWork-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptxJuliansyahHarahap1
 

Recently uploaded (20)

Thermal Engineering Unit - I & II . ppt
Thermal Engineering  Unit - I & II . pptThermal Engineering  Unit - I & II . ppt
Thermal Engineering Unit - I & II . ppt
 
Unleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leapUnleashing the Power of the SORA AI lastest leap
Unleashing the Power of the SORA AI lastest leap
 
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
FULL ENJOY Call Girls In Mahipalpur Delhi Contact Us 8377877756
 
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak HamilCara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
Cara Menggugurkan Sperma Yang Masuk Rahim Biyar Tidak Hamil
 
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...Booking open Available Pune Call Girls Pargaon  6297143586 Call Hot Indian Gi...
Booking open Available Pune Call Girls Pargaon 6297143586 Call Hot Indian Gi...
 
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...Bhosari ( Call Girls ) Pune  6297143586  Hot Model With Sexy Bhabi Ready For ...
Bhosari ( Call Girls ) Pune 6297143586 Hot Model With Sexy Bhabi Ready For ...
 
NFPA 5000 2024 standard .
NFPA 5000 2024 standard                                  .NFPA 5000 2024 standard                                  .
NFPA 5000 2024 standard .
 
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoorTop Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
Top Rated Call Girls In chittoor 📱 {7001035870} VIP Escorts chittoor
 
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank  Design by Working Stress - IS Method.pdfIntze Overhead Water Tank  Design by Working Stress - IS Method.pdf
Intze Overhead Water Tank Design by Working Stress - IS Method.pdf
 
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...
VIP Model Call Girls Kothrud ( Pune ) Call ON 8005736733 Starting From 5K to ...
 
Design For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the startDesign For Accessibility: Getting it right from the start
Design For Accessibility: Getting it right from the start
 
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Walvekar Nagar Call Me 7737669865 Budget Friendly No Advance Booking
 
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
(INDIRA) Call Girl Bhosari Call Now 8617697112 Bhosari Escorts 24x7
 
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar  ≼🔝 Delhi door step de...
Call Now ≽ 9953056974 ≼🔝 Call Girls In New Ashok Nagar ≼🔝 Delhi door step de...
 
Unit 2- Effective stress & Permeability.pdf
Unit 2- Effective stress & Permeability.pdfUnit 2- Effective stress & Permeability.pdf
Unit 2- Effective stress & Permeability.pdf
 
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
Call Girls Pimpri Chinchwad Call Me 7737669865 Budget Friendly No Advance Boo...
 
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance BookingCall Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
Call Girls Wakad Call Me 7737669865 Budget Friendly No Advance Booking
 
AKTU Computer Networks notes --- Unit 3.pdf
AKTU Computer Networks notes ---  Unit 3.pdfAKTU Computer Networks notes ---  Unit 3.pdf
AKTU Computer Networks notes --- Unit 3.pdf
 
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
(INDIRA) Call Girl Aurangabad Call Now 8617697112 Aurangabad Escorts 24x7
 
Work-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptxWork-Permit-Receiver-in-Saudi-Aramco.pptx
Work-Permit-Receiver-in-Saudi-Aramco.pptx
 

Performance evaluation of hepatitis diagnosis using single and multi classifiers fusion final

  • 1. Performance Evaluation Of Hepatitis Diagnosis Using Single and Multi-Classifiers Fusion Ahmed M. Salah EL-Bohy1 , Atallah I. Hashad2 , and Hussien Saad Taha3 Arab Academy for Science, Technology & Maritime Transport, Cairo, Egypt 1 a.bohy4master@gmail.com, 2 hashad@aast.edu, and3 htaha10@yahoo.com. Abstract- The goal of our paper is to obtain superior accuracy of different classifiers or multi-classifiers fusion in diagnosing Hepatitis using world wide data set from Ljubljana University. We present an implementation among some of the classification methods which are defined as the best algorithms in medical field. Then we apply a fusion between classifiers to get the best multi-classifier fusion approach. By using confusion matrix to get classification accuracy which built in 10-fold cross validation technique. The experimental results show that for all data sets (complete, reduced, and no missing value) using multi-classifiers fusion achieved better accuracy than the single ones. Keywords-Hepatitis; Classification techniques; Fusion ; WEKA. I. INTRODUCTION The diagnosis of some diseases like hepatitis is very difficult task for a doctor, where doctors usually determine decision by comparing the current test results of patients with another one who has the same condition. Hepatitis is one of the most common diseases among Egyptians; as it represents 22% of hepatitis cases around the world. Which motivates us for suggesting new methods to improve the outcomes of existing approaches, as well as to help doctors and specialists to diagnose hepatitis disease survival. [1] Hepatitis (Greek) means 'liver' and suffix –itis denotes 'inflammation' of the liver and may be due to infectious Or non- infectious causes. The five types of hepatitis viruses are Common infectious causes of liver inflammation and some like Hepatitis A (HAV), B (HBV) and C (HCV) are more frequently seen infectious agents. Inflammation may lead to death of the liver cells (hepatocytes) which severely compromises normal liver function. Acute HBV Infection (less than 6 months) may resemble the fever, flu, muscle aches, joint pains and generally being unwell. Symptoms specifying that states are: dark urine, loss of appetite, nausea, vomiting, jaundice, pain up the liver. Chronic hepatitis B is an infection persisting more than 6 months, the clinical features of that state correspond to liver dysfunction, so signs like these may be noticed: enlarged liver, splenomegaly, hepatosplenomegaly, jaundice, weakness, abdominal pain, confusion and abdominal swelling [2]. The success of treatment depends on an early recognition of the virus, which achieves more exact and less violent treatment options and mortality from Hepatitis falls. Recently, data-mining has become one of the most treasured tools for operating data in order to produce valuable information for decision-making [3]. Supervised learning, including classification is one of the most significant brands in data mining, with a recognized output variable in the dataset. Classification methods can achieve high accuracy in classifying mortality cases. The implementations had been applied with a “WEKA” tool which stands for the Waikato Environment for Knowledge Analysis. A lot of papers about applying machine learning procedures for survivability analysis in the field of Hepatitis diagnostic. Here are some examples: Using Support Vector Machines And Wrapper Method for predicting Hepatitis was introduced achieving maximum accuracy of (74.55%). [4]. but we note that applying SVM classifier only get higher accuracy than the mentioned accuracies even with feature selection. Improving the accuracy of SVM algorithm using feature selection [5]. Using SVM with Chi-Square achieved accuracy of (83.12 %), But we note that applying another classifier (Logistic, Simple Logistic, SMO, RF, J48) gets higher accuracy than the mentioned one. Detection of Hepatitis based on SVM and Data Analysis [6] using SVM & wrapper the achieved accuracy is (85%) but we found the results are calculated with 7 attributes out of 25 ones as stated whenever the original number of attributes is 20 only The rest of this paper is prearranged like this: In sector II, Classification algorithms are discussed. In sector III Evaluation principles are discussed. In sector IV a proposed model is shown. In sector V reports the experimental results. Finally, Sector VI introduces the conclusion. II. Classification Algorithms A Bayesian belief network is sometimes named a Bayes net, a belief net, or a causal network, It is a directed, acyclic graph, indicating conditional dependencies. It can be used to guess the probability of events. the Bayesian decision rule assurances minimum error if likelihoods and prior probabilities are known [7]. Decision Tree (DT) Tree that the root and each interior node is marked with a question. The arcs represent each possible answer to the concomitant question. Each leaf node represents a forecast of a problem solution. A prevalent technique for classification; Leaf node leads the class to which the corresponding tuple belongs. It’s Model is a computational model comprising of the three parts: Decision Tree Algorithm to create the tree Algorithm that applies the tree to the data, creation of the tree is the most exciting part. Processing is mostly a search similar to that in a binary search tree (although DT may not be binary). Advantages: Easy to understand. Easy to generate rules[8] Support Vector Machine (SVM) SVMs are amongst the best (and many believe are definitely the best)“on-the-shelf ” supervised learning algorithm. It’s derived from statistics in 1992.SVM is widely used in multiple applications pattern
  • 2. recognition, classification and regression. The SVMs work on an underlying principle, which is to insert a hyper-plane between the classes and orient it in such a way so as to keep it at the maximum distance from the nearest data points These data points, which appear closest to the hyper-plane, are known as Support Vectors. [9] Sequential Minimal Optimization (SMO) humble, easy to implement, is generally quicker, and has better scaling properties for difficult SVM problems than the usual SVM training algorithm SMO quickly solve the SVM QP problem without extra matrix storage (the memory used is linear with the training data set size) and without using numerical QP optimization steps at all. SMO can be used for online learning. While SMO has been shown to be operative on sparse data sets and especially fast for linear SVMs, the algorithm can be extremely slow on non-sparse data sets and on problems that have many support vectors. Regression problems are especially prone to these matters because the inputs are usually non-sparse real numbers (as opposed to binary inputs) with solutions that have many support vectors. Because of these restrictions, there have been limited reports of SMO being successfully used on regression problems. [10] Logistic Regression (LR) is a famous well known classifier, it could be used to extend classification results into a deeper analysis. It’s not widely used due to its slow response in data mining especially when compared with SVM in large data sets (not our case) It models the relationship between a dependent and one or more independent variables, and consents us to look at the fit of the model as well as at the significance of the relationships [11] Simple Logistic: We use simple logistic regression when we have one nominal variable with two values (dead/alive, male/female) and one measurement variable. The nominal variable is the dependent variable, but the measurement variable is not it is an independent one. Simple logistic regression is analogous to linear regression, except the dependent variable is nominal, not a measurement. [12] Random Forest: Random forests change how the classification or regression trees are constructed. In standard trees, it uses the best split among all variables to split each node In a random forest, each node is split using the best among a subset of predictors arbitrarily chosen at that node. it works by one of two methods, boosting and bagging.it has the advantages of: handling thousands of input variables without deleting any variable, giving estimation of variables importance in the classification, and It also has an active method for estimating missing data and keeps accuracy when a large amount of the data are missing [13]. SGD is an abbreviation for Stochastic Gradient Descent: the gradient of the loss is estimated each sample at a time and the model is updated along the way with a decreasing strength schedule. SGD has been excellently applied to large-scale and sparse machine learning problems often come across the text classification and natural language processing [14]. K-star or K* is an instance-based classifier. The class of a test instance is based on the training instances similar to it, as determined by some similarity function. The difference between it and the other instance-based learners in that K* uses an entropy-based distance function. Instance-based learners classifying instance by matching it to a database of pre- classified cases. The fundamental assumption is that similar instances will have similar classifications [15]. III. EVALUATION PRINCIPLES Evaluation method is based on the confusion matrix. The confusion matrix is an imagining implement usually used to show presentations of classifiers. It is used to display the relationships between real class attributes and predicted classes. The grade of efficiency of the classification task is calculated with the number of exact and unseemly classifications in each conceivable value of the variables being classified in the confusion matrix [16] Table 1. Confusion matrix For instance, in a 2-class classification problem with two predefined classes (e.g., Positive, negative) the classified test cases are divided into four categories: • True positives (TP) correctly classified as positive instances. • True negatives (TN) correctly classified negative instances. • False positives (FP) incorrectly classified negative instances • False negatives (FN) incorrectly classified positive instances. To evaluate classifier performance. We define accuracy term that is defined as the entire number of truly classified instances divided by the entire number of available instances for an assumed operational point of a classifier. Accuracy TP TN TP TN FP FN      (1) IV. RROPOSED METHODOLOGY We propose a reliable method for diagnosing Hepatitis and better classifying mortality cases that may be caused by Hepatitis infection based on data mining using WEKA as follows: A. Data preprocessing Pre-processing steps are applied to the data before classification: First step is Data Cleansing: eliminating or decreasing noise and the treatment of missing values. Hepatitis data set has a lot of missing values, especially in attributes 18th, 15th, and 17th as they are 67, 29, 16 values respectively. After removing the instances has these missing values the (155) instance data set converted to (83) instances with minor missing values in other attributes (Only 9 missing values in the reduced data set). We Predicted Class Negative Positive Outcomes Negative TN FN Positive FP TP
  • 3. also convert the data set into a new (third) one by removing these (9) missing values to be (80) instance instead of (83) instance. Second step is Relevance Analysis: Statistical correlation analysis is used to discard the redundant features from further analysis. The data set has one irrelevant attribute named ‘sex ’ which has no effect in the classification process; Final, step is Data Transformation: The dataset is transformed by normalization, which is one of the greatest public tools used by inventors of automatic recognition classifications to get superior results. Data normalization hurries up training time by initialing the training procedure to reach feature within the same scale. The aim of normalization is to transform the attribute values to a small-scale range. B. Single classification task Data classification is the process of organizing data into categories for its most effective and efficient use. A well- structured data classification system makes it easy to get, handle, and retrieve necessary data. Classification consists of conveying a class label to a set of unclassified cases. Supervised classification: The set of possible classes is known in advance. Unsupervised classification: Set of possible classes is unknown. After classification we can try to assign a name to that class. Unsupervised classification is known as clustering. The presumed model depends on analyzing the training dataset. The derivative model characterized in several procedures, such as simple classification rules, decision trees and another. Basically data classification is a two-stage process, in the initial stage; a classifier is built signifying a predefined set of notions or data classes. This is the training stage, where a classification technique builds the classifier by learning from a training dataset and their related class label columns or attributes. In next stage this model is used for measurement. In order to deduce the predictive accuracy of the classifier an independent set of the training instances is used. We evaluate the most frequently classification techniques that mentioned in recently published researches in the medical field to accomplish the highest accuracy classifier’s result with each data set. C. Multi-classifiers fusion classification task Fusion is a combination between 2 or more classifiers to enhance classifier performance (Accuracy). Our procedure will be electing the highest 2 classifiers in accuracy calculating accuracy, then the 3rd, then 4th and so on until accuracy decreases then stop & take the maximum accuracy obtained as shown in Figure 1 We name the number of classifiers used in the fusion process “fusion level” best accuracy with other single classifiers predicting to improve accuracy. Repeating the same process till the latest level of fusion, according to the number of single classifiers to pick the highest accuracy through all processes. We propose 2 methods (1st Directly applying a classifier to the raw data & 2nd preprocess the data before applying classifier). Import the Dataset.  Convert nominal values to binary ones (in case of preprocessing).  Normalize each variable of the data set, so that the values range from 0 to 1 (in case of preprocessing).  Create a separate training (learn) and testing sets by indiscriminately drawing out the data for training and for testing.  Picking and adapt the learning procedure  Perform the learning procedure  Calculate the performance of the model on the test set Figure 1 .Proposed Hepatitis diagnosis algorithm
  • 4. V. EXPERIMENTAL RESULTS According to the mentioned work in sector IV proposed methodology-A- data preprocessing we now have 3 data sets with 155, 83, 80 instances respectively & last two data sets has been preprocessed as mentioned in sector IV proposed methodology-A- data transformation To calculate the proposed model, first we calculate accuracy using 9 single classifiers then we compute accuracy in fusion process up to 9 classifiers together using the highest classifiers combination in each fusion level (All results are illustrated graphically in Figure 2). A. Single classification task Table 2 shows the comparison between accuracies over nine single classifiers (Bayes Net , SVM, Logistic, SGD, Simple Logistic, SMO, K*, J48, and RF). Highest accuracy in each data set is highlighted in yellow color in Table 2. Using 155 instances (complete) data set the highest accuracy by SMO classifier is 85.17%. Using 83 instances (reduced) data set the highest accuracy by K* classifier is 91.67%. Using 80 instances (no missing value) data set the highest accuracy by Bayes Net classifier is 88.75%. SMO classifier achieves better accuracy than SVM classifier illustrated in previous work for the three data sets without any preprocessing or feature selection [4-6] Furthermore, we note that RF classifier achieves the second rank among single classifiers for the whole three data sets and it’s highlighted in cyan color in Table 2. Finally K* classifier bare as a good classifier as it achieves highest accuracy in one data set and it got the seconed place in another one equal with RF classifier in 80 instances (no missing value) data set Table 2 Single classifier results # instance Classifier 155 complete 83 Reduce d 80 no missing value BayesNet 83.21 86.81 88.75 SVM 79.38 82.08 83.75 Logistic 82.58 77.92 81.25 SGD 84.54 84.17 81.25 Simple Logistic 83.88 80.42 85.00 SMO 85.17 84.03 85.00 K* 81.96 91.67 87.50 J48 83.79 82.22 86.25 RF 85.13 90.56 87.50 Table 3 shows the comparison between accuracies over multi classifiers fusion between classifiers up to nine classifiers all possible combination has been executed as illustrated in proposed algorithm Figure 1 and the table contains the highest accuracy at each fusion level. Highest accuracy in each data set for all fusion levels is highlighted in yellow color in Table 3. Using 155 instances (complete) data set the highest fusion accuracy obtained by using K*, J48, RF, Simple Logistic, and Logistic classifiers is 87.04%. Using 83 instances (reduced) data set the highest fusion accuracy obtained by using K*, RF, SVM, and BayesNet classifiers is 92.92%. Using 80 instances (no missing value) data set the highest fusion accuracy obtained by using K*, J48, RF, and BayesNet classifiers is 91.25%. The results in Table 3 indicates First data preprocessing minimized the fusion level / number of classifiers by one classifier as we get the best result / highest accuracy at fusion level 4th level instead of 5th level in case of 155 instances (complete) data set Second K*and RF are performed as outstanding classifiers in fusion for Hepatitis data set in all cases (complete, Reduced, and no missing value) with / without preprocessing Finally the importance of BayesNet classifier is appeared for less or no missing values data as it increases accuracy by a noticeable amount in both 83 instances (reduced) and 80 instances (no missing value) data sets Table 3 Fusion classifiers results # instance Fusion Level 155 Complete (%) 83 Reduced (%) 80 no missing value (%) 2nd 86.42 91.67 88.75 3rd 86.46 89.31 88.75 4th 86.42 92.92 91.25 5th 87.04 91.67 90.00 6th 87.04 90.42 90.00 7th 86.42 90.42 91.25 8th 85.75 86.67 88.75 9th 85.03 85.42 88.85 Figure 2 Single / Fusion classifier accuracy comparison
  • 5. VI. CONCLUSION The experimental results clarified that multi-classifiers fusion reaches better accuracy than single classifier. For the three data sets. Although no unique classifier achieves the highest accuracy as a single classifier for all data sets but RF classifier playing a great role as a single classifier for the 3 mentioned data sets (complete, reduced, and no missing value). Also the K* classifier is a common master classifier between all fusion combination, adding RF to it in fusion process gives an outstanding results. BayesNet classifier consider as a break through dealing with Hepatitis reduced and no missing value data sets in fusion as it improves accuracy by 1.25% and 2.5% from best single classifier respectively and increase it by 3.61% and 2.5% from pervious fusion level REFERENCES [1] World Health Organization at http://www.who.int [2] Tomasz KANIK. “Hepatitis disease diagnosis using Rough Set modification of the pre-processing algorithm.” Information and communication technologies international conference (2012). [3] Fayyad, Usama, Gregory Piatetsky-Shapiro, and Padhraic Smyth."From data mining to knowledge discovery in databases." AI magazine 17.3 (1996). [4] A.H.Roslina and A.Noraziah. "Prediction Of Hepatitis Prognosis Using Support Vector Machines And Wrapper Method." Seventh International Conference on Fuzzy Systems and Knowledge Discovery (2010). [5] Varun Kumar.M , Vijay Sharathi.V and Gayathri Devi.B.R. " Hepatitis Prediction Model based on Data Mining Algorithm and Optimal Feature Selection to Improve Predictive Accuracy." International Journal of Computer Applications (0975 – 8887) Volume 51– No.19, August 2012. [6] C. Barath Kumar , M. Varun Kumar , T. Gayathri, and S. Rajesh Kumar " Analysis and Prediction of Hepatitis Using Support Vector Machine." International Journal of Computer Science and Information Technologies, Vol. 5 (2) , 2014, 2235-2237. [7] Jyotirmay Gadewadikar, Ognjen Kuljaca1, Kwabena Agyepong, Erol Sarigul, Yufeng Zheng and Ping Zhang, Exploring Bayesian networks for medical decision support in breast cancer detection, African Journal of Mathematics and Computer Science Research Vol. 3(10), pp. 225-231, October 2010. [8] Ross Quinlan, (1993) C4.5: Programs for Machine Learning, Morgan Kaufmann Publishers, San Mateo, CA. [9] Chen Y., Wang G., and Dong S. “Learning with progressive transductive support vector machine”, Pattern Recognition Letters, Vol. 24, Pages: 1845-1855 [10] John C. Platt. “Sequential Minimal Optimization:A Fast Algorithm for Training Support Vector Machines.” Technical Report MSR-TR-98-14April. [11] Mitchell T. “Machine Learning, mcgraw Hill.” (1997). [12] John H. Mcdonald. “Handbook of Biological Statistics.” At http://www.strath.ac.uk [13] Breiman, Leo. "Random forests." Machine learning 45.1 (2001): 5-32. [14] Y. Tsuruoka, J. Tsujii, and S. Ananiadou. " Stochastic gradient descent training for l1-regularized log-linear models with cumulative penalty " AFNLP/ACL 2009. [15] Mahmood, Deeman Y., and Mohammed A. Hussein. "Intrusion Detection System Based on K-Star Classifier and Feature Set Reduction." International Organization of Scientific Research Journal of Computer Engineering (IOSR-JCE) Vol 15 (2013): 107-112. [16] P. Cabena, P. Hadjinian, R. Stadler, J. Verhees and A.Zanasi, Discovering data mining from concept to implementation. Uppersaddle River, N.J.: Prentice Hall, 1998.