In the current health system, it is very difficult for medical practitioners/ physicians to diagnose the effectiveness of heart contraction. In this research, we proposed a machine learning model to predict heart contraction using an artificial neural network (ANN). We also proposed a novel wrapper-based feature selection utilizing a grey wolf optimization (GWO) to reduce the number of required input attributes. In this work, we compared the results achieved using our method and several conventional machine learning algorithms approaches such as support vector machine, decision tree, Knearest neighbor, naïve bayes, random forest, and logistic regression. Computational results show not only that much fewer features are needed, but also higher prediction accuracy can be achieved around 87%. This work has the potential to be applicable to clinical practice and become a supporting tool for doctors/physicians.
SUPERVISED FEATURE SELECTION FOR DIAGNOSIS OF CORONARY ARTERY DISEASE BASED O...csitconf
Feature Selection (FS) has become the focus of much research on decision support systems
areas for which datasets with tremendous number of variables are analyzed. In this paper we
present a new method for the diagnosis of Coronary Artery Diseases (CAD) founded on Genetic
Algorithm (GA) wrapped Bayes Naïve (BN) based FS.
Basically, CAD dataset contains two classes defined with 13 features. In GA–BN algorithm, GA
generates in each iteration a subset of attributes that will be evaluated using the BN in the
second step of the selection procedure. The final set of attribute contains the most relevant
feature model that increases the accuracy. The algorithm in this case produces 85.50%
classification accuracy in the diagnosis of CAD. Thus, the asset of the Algorithm is then
compared with the use of Support Vector Machine (SVM), Multi-Layer Perceptron (MLP) and
C4.5 decision tree Algorithm. The result of classification accuracy for those algorithms are
respectively 83.5%, 83.16% and 80.85%. Consequently, the GA wrapped BN Algorithm is
correspondingly compared with other FS algorithms. The Obtained results have shown very
promising outcomes for the diagnosis of CAD.
Supervised Feature Selection for Diagnosis of Coronary Artery Disease Based o...cscpconf
Feature Selection (FS) has become the focus of much research on decision support systems areas for which datasets with tremendous number of variables are analyzed. In this paper we
present a new method for the diagnosis of Coronary Artery Diseases (CAD) founded on Genetic Algorithm (GA) wrapped Bayes Naïve (BN) based FS. Basically, CAD dataset contains two classes defined with 13 features. In GA–BN algorithm, GA
generates in each iteration a subset of attributes that will be evaluated using the BN in the second step of the selection procedure. The final set of attribute contains the most relevant feature model that increases the accuracy. The algorithm in this case produces 85.50% classification accuracy in the diagnosis of CAD. Thus, the asset of the Algorithm is then compared with the use of Support Vector Machine (SVM), Multi-Layer erceptron (MLP) and C4.5 decision tree Algorithm. The result of classification accuracy for those algorithms are respectively 83.5%, 83.16% and 80.85%. Consequently, the GA wrapped BN Algorithm is correspondingly compared with other FS algorithms. The Obtained results have shown very promising outcomes for the diagnosis of CAD.
Iganfis Data Mining Approach for Forecasting Cancer Threatsijsrd.com
Healthcare facilities have at their disposal vast amounts of cancer patients’ data. Medical practitioners require more efficient techniques to extract relevant knowledge from this data for accurate decision-making. However the challenge is how to extract and act upon it in a timely manner. If well engineered, the huge data can aid in developing expert systems for decision support that can assist physicians in diagnosing and predicting some debilitating life threatening diseases such as cancer. Expert systems for decision support can reduce the cost, the waiting time, and liberate medical practitioners for more research, as well as reduce errors and mistakes that can be made by humans due to fatigue and tiredness. The process of utilizing health data effectively however, involves many challenges such as the problem of missing feature values, the curse of dimensionality due to a large number of attributes, and the course of actions to determine the features that can lead to more accurate diagnosis. Effective data mining tools can assist in early detection of diseases such as cancer. In This paper, we propose a new approach called IGANFIS. This approach optimally minimizes the number of features using the information gain (IG) algorithm which is usually used in text categorization to select the quality of text. The IG will be used for selecting the quality of cancer features by virtue of reducing them in number. The reduced number quality features dataset will then be applied to the Adaptive Neuro Fuzzy Inference System (ANFIS) to train and test the proposed approach. ANFIS method of training is ideally the hybrid learning algorithm which uses the gradient descent method and Least Square Estimate (LSE) for computing the error measure for each training pair. Each cycle of the ANFIS hybrid learning consists of a forward pass to present the input vector calculating the node outputs layer by layer repeating the process for all data and a backward pass using the steepest descent algorithm to update parameters, a process called back propagation.
Multivariate sample similarity measure for feature selection with a resemblan...IJECEIAES
Feature selection improves the classification performance of machine learning models. It also identifies the important features and eliminates those with little significance. Furthermore, feature selection reduces the dimensionality of training and testing data points. This study proposes a feature selection method that uses a multivariate sample similarity measure. The method selects features with significant contributions using a machine-learning model. The multivariate sample similarity measure is evaluated using the University of California, Irvine heart disease dataset and compared with existing feature selection methods. The multivariate sample similarity measure is evaluated with metrics such as minimum subset selected, accuracy, F1-score, and area under the curve (AUC). The results show that the proposed method is able to diagnose chest pain, thallium scan, and major vessels scanned using X-rays with a high capability to distinguish between healthy and heart disease patients with a 99.6% accuracy.
Supervised machine learning based liver disease prediction approach with LASS...journalBEEI
In this contemporary era, the uses of machine learning techniques are increasing rapidly in the field of medical science for detecting various diseases such as liver disease (LD). Around the globe, a large number of people die because of this deadly disease. By diagnosing the disease in a primary stage, early treatment can be helpful to cure the patient. In this research paper, a method is proposed to diagnose the LD using supervised machine learning classification algorithms, namely logistic regression, decision tree, random forest, AdaBoost, KNN, linear discriminant analysis, gradient boosting and support vector machine (SVM). We also deployed a least absolute shrinkage and selection operator (LASSO) feature selection technique on our taken dataset to suggest the most highly correlated attributes of LD. The predictions with 10 fold cross-validation (CV) made by the algorithms are tested in terms of accuracy, sensitivity, precision and f1-score values to forecast the disease. It is observed that the decision tree algorithm has the best performance score where accuracy, precision, sensitivity and f1-score values are 94.295%, 92%, 99% and 96% respectively with the inclusion of LASSO. Furthermore, a comparison with recent studies is shown to prove the significance of the proposed system.
SUPERVISED FEATURE SELECTION FOR DIAGNOSIS OF CORONARY ARTERY DISEASE BASED O...csitconf
Feature Selection (FS) has become the focus of much research on decision support systems
areas for which datasets with tremendous number of variables are analyzed. In this paper we
present a new method for the diagnosis of Coronary Artery Diseases (CAD) founded on Genetic
Algorithm (GA) wrapped Bayes Naïve (BN) based FS.
Basically, CAD dataset contains two classes defined with 13 features. In GA–BN algorithm, GA
generates in each iteration a subset of attributes that will be evaluated using the BN in the
second step of the selection procedure. The final set of attribute contains the most relevant
feature model that increases the accuracy. The algorithm in this case produces 85.50%
classification accuracy in the diagnosis of CAD. Thus, the asset of the Algorithm is then
compared with the use of Support Vector Machine (SVM), Multi-Layer Perceptron (MLP) and
C4.5 decision tree Algorithm. The result of classification accuracy for those algorithms are
respectively 83.5%, 83.16% and 80.85%. Consequently, the GA wrapped BN Algorithm is
correspondingly compared with other FS algorithms. The Obtained results have shown very
promising outcomes for the diagnosis of CAD.
Supervised Feature Selection for Diagnosis of Coronary Artery Disease Based o...cscpconf
Feature Selection (FS) has become the focus of much research on decision support systems areas for which datasets with tremendous number of variables are analyzed. In this paper we
present a new method for the diagnosis of Coronary Artery Diseases (CAD) founded on Genetic Algorithm (GA) wrapped Bayes Naïve (BN) based FS. Basically, CAD dataset contains two classes defined with 13 features. In GA–BN algorithm, GA
generates in each iteration a subset of attributes that will be evaluated using the BN in the second step of the selection procedure. The final set of attribute contains the most relevant feature model that increases the accuracy. The algorithm in this case produces 85.50% classification accuracy in the diagnosis of CAD. Thus, the asset of the Algorithm is then compared with the use of Support Vector Machine (SVM), Multi-Layer erceptron (MLP) and C4.5 decision tree Algorithm. The result of classification accuracy for those algorithms are respectively 83.5%, 83.16% and 80.85%. Consequently, the GA wrapped BN Algorithm is correspondingly compared with other FS algorithms. The Obtained results have shown very promising outcomes for the diagnosis of CAD.
Iganfis Data Mining Approach for Forecasting Cancer Threatsijsrd.com
Healthcare facilities have at their disposal vast amounts of cancer patients’ data. Medical practitioners require more efficient techniques to extract relevant knowledge from this data for accurate decision-making. However the challenge is how to extract and act upon it in a timely manner. If well engineered, the huge data can aid in developing expert systems for decision support that can assist physicians in diagnosing and predicting some debilitating life threatening diseases such as cancer. Expert systems for decision support can reduce the cost, the waiting time, and liberate medical practitioners for more research, as well as reduce errors and mistakes that can be made by humans due to fatigue and tiredness. The process of utilizing health data effectively however, involves many challenges such as the problem of missing feature values, the curse of dimensionality due to a large number of attributes, and the course of actions to determine the features that can lead to more accurate diagnosis. Effective data mining tools can assist in early detection of diseases such as cancer. In This paper, we propose a new approach called IGANFIS. This approach optimally minimizes the number of features using the information gain (IG) algorithm which is usually used in text categorization to select the quality of text. The IG will be used for selecting the quality of cancer features by virtue of reducing them in number. The reduced number quality features dataset will then be applied to the Adaptive Neuro Fuzzy Inference System (ANFIS) to train and test the proposed approach. ANFIS method of training is ideally the hybrid learning algorithm which uses the gradient descent method and Least Square Estimate (LSE) for computing the error measure for each training pair. Each cycle of the ANFIS hybrid learning consists of a forward pass to present the input vector calculating the node outputs layer by layer repeating the process for all data and a backward pass using the steepest descent algorithm to update parameters, a process called back propagation.
Multivariate sample similarity measure for feature selection with a resemblan...IJECEIAES
Feature selection improves the classification performance of machine learning models. It also identifies the important features and eliminates those with little significance. Furthermore, feature selection reduces the dimensionality of training and testing data points. This study proposes a feature selection method that uses a multivariate sample similarity measure. The method selects features with significant contributions using a machine-learning model. The multivariate sample similarity measure is evaluated using the University of California, Irvine heart disease dataset and compared with existing feature selection methods. The multivariate sample similarity measure is evaluated with metrics such as minimum subset selected, accuracy, F1-score, and area under the curve (AUC). The results show that the proposed method is able to diagnose chest pain, thallium scan, and major vessels scanned using X-rays with a high capability to distinguish between healthy and heart disease patients with a 99.6% accuracy.
Supervised machine learning based liver disease prediction approach with LASS...journalBEEI
In this contemporary era, the uses of machine learning techniques are increasing rapidly in the field of medical science for detecting various diseases such as liver disease (LD). Around the globe, a large number of people die because of this deadly disease. By diagnosing the disease in a primary stage, early treatment can be helpful to cure the patient. In this research paper, a method is proposed to diagnose the LD using supervised machine learning classification algorithms, namely logistic regression, decision tree, random forest, AdaBoost, KNN, linear discriminant analysis, gradient boosting and support vector machine (SVM). We also deployed a least absolute shrinkage and selection operator (LASSO) feature selection technique on our taken dataset to suggest the most highly correlated attributes of LD. The predictions with 10 fold cross-validation (CV) made by the algorithms are tested in terms of accuracy, sensitivity, precision and f1-score values to forecast the disease. It is observed that the decision tree algorithm has the best performance score where accuracy, precision, sensitivity and f1-score values are 94.295%, 92%, 99% and 96% respectively with the inclusion of LASSO. Furthermore, a comparison with recent studies is shown to prove the significance of the proposed system.
Genome feature optimization and coronary artery disease prediction using cuck...CSITiaesprime
Cardiovascular diseases are among the major health ailment issue leading to millions of deaths every year. In recent past, analyzing gene expression data, particularly using machine learning strategies to predict and classify the given unlabeled gene expression record is a generous research issue. Concerning this, a substantial requirement is feature optimization, which is since the overall genes observed in human body are closely 25000 and among them 636 are cardiovascular related genes. Hence, it complexes the process of training the machine learning models using these entire cardiovascular gene features. This manuscript uses bidirectional pooled variance strategy of ANOVA standard to select optimal features. Along the side to surpass the constraint observed in traditional classifiers, which is unstable accuracy at k-fold cross validation, this manuscript proposed a classification strategy that build upon the swarm intelligence technique called cuckoo search. The experimental study indicating that the number of optimal features those selected by proposed model is substantially low that compared to the other contemporary model that selects features using forward feature selection and classifies using support vector machine classifier (FFS&SVM). The experimental study evinced that the proposed model, which selects feature by bidirectional pooled variance estimation and classifies using proposed classification strategy that build on cuckoo search (BPVE&CS) outperformed the selected contemporary model (FFS&SVM).
Propose a Enhanced Framework for Prediction of Heart DiseaseIJERA Editor
Heart disease diagnosis requires more experience and it is a complex task. The Heart MRI, ECG and Stress Test etc are the numbers of medical tests are prescribed by the doctor for examining the heart disease and it is the way of tradition in the prediction of heart disease. Today world, the hidden information of the huge amount of health care data is contained by the health care industry. The effective decisions are made by means of this hidden information. For appropriate results, the advanced data mining techniques with the information which is based on the computer are used. In any empirical sciences, for the inference and categorisation, the new mathematical techniques to be used called Artificial neural networks (ANNs) it also be used to the modelling of the real neural networks. Acting, Wanting, knowing, remembering, perceiving, thinking and inferring are the nature of mental phenomena and these can be understand by using the theory of ANN. The problem of probability and induction can be arised for the inference and classification because these are the powerful instruments of ANN. In this paper, the classification techniques like Naive Bayes Classification algorithm and Artificial Neural Networks are used to classify the attributes in the given data set. The attribute filtering techniques like PCA (Principle Component Analysis) filtering and Information Gain Attribute Subset Evaluation technique for feature selection in the given data set to predict the heart disease symptoms. A new framework is proposed which is based on the above techniques, the framework will take the input dataset and fed into the feature selection techniques block, which selects any one techniques that gives the least number of attributes and then classification task is done using two algorithms, the same attributes that are selected by two classification task is taken for the prediction of heart disease. This framework consumes the time for predicting the symptoms of heart disease which make the user to know the important attributes based on the proposed framework.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
PREDICTION OF MALIGNANCY IN SUSPECTED THYROID TUMOUR PATIENTS BY THREE DIFFER...cscpconf
In the present study, the abilities of three classification methods of data mining namely artificial
neural networks with feed-forward back propagation algorithm, J48 decision tree method and
logistic regression analysis are compared in a medical real dataset. The prediction of
malignancy in suspected thyroid tumour patients is the objective of the study. The accuracy of
the correct predictions (the minimum error rate), the amount of time consuming in the
modelling process and the interpretability and simplicity of the results for clinical experts are
the factors considered to choose the best method
PERFORMANCE OF DATA MINING TECHNIQUES TO PREDICT IN HEALTHCARE CASE STUDY: CH...ijdms
With the promises of predictive analytics in big data, and the use of machine learning algorithms,
predicting future is no longer a difficult task, especially for health sector, that has witnessed a great
evolution following the development of new computer technologies that gave birth to multiple fields of
research. Many efforts are done to cope with medical data explosion on one hand, and to obtain useful
knowledge from it, predict diseases and anticipate the cure on the other hand. This prompted researchers
to apply all the technical innovations like big data analytics, predictive analytics, machine learning and
learning algorithms in order to extract useful knowledge and help in making decisions. In this paper, we
will present an overview on the evolution of big data in healthcare system, and we will apply three learning
algorithms on a set of medical data. The objective of this research work is to predict kidney disease by
using multiple machine learning algorithms that are Support Vector Machine (SVM), Decision Tree (C4.5),
and Bayesian Network (BN), and chose the most efficient one.
Diagnosis of rheumatoid arthritis using an ensemble learning approachcsandit
Rheumatoid arthritis is one of the diseases that it
s cause is unknown yet; exploring the field of
medical data mining can be helpful in early diagnos
is and treatment of the disease. In this
study, a predictive model is suggested that diagnos
es rheumatoid arthritis. The rheumatoid
arthritis dataset was collected from 2,564 patients
referred to rheumatology clinic. For each
patient a record consists of several clinical and d
emographic features is saved. After data
analysis and pre-processing operations, three diffe
rent methods are combined to choose proper
features among all the features. Various data class
ification algorithms were applied on these
features. Among these algorithms Adaboost had the h
ighest precision. In this paper, we
proposed a new classification algorithm entitled CS
-Boost that employs Cuckoo search
algorithm for optimizing the performance of Adaboos
t algorithm. Experimental results show
that the CS-Boost algorithm enhance the accuracy of
Adaboost in predicting of Rheumatoid
Arthritis.
DIAGNOSIS OF RHEUMATOID ARTHRITIS USING AN ENSEMBLE LEARNING APPROACH cscpconf
Rheumatoid arthritis is one of the diseases that its cause is unknown yet; exploring the field of
medical data mining can be helpful in early diagnosis and treatment of the disease. In this
study, a predictive model is suggested that diagnoses rheumatoid arthritis. The rheumatoid
arthritis dataset was collected from 2,564 patients referred to rheumatology clinic. For each
patient a record consists of several clinical and demographic features is saved. After data
analysis and pre-processing operations, three different methods are combined to choose proper
features among all the features. Various data classification algorithms were applied on these
features. Among these algorithms Adaboost had the highest precision. In this paper, we
proposed a new classification algorithm entitled CS-Boost that employs Cuckoo search
algorithm for optimizing the performance of Adaboost algorithm. Experimental results show
that the CS-Boost algorithm enhance the accuracy of Adaboost in predicting of Rheumatoid
Arthritis.
A hybrid approach to medical decision-making: diagnosis of heart disease wit...IJECEIAES
Heart disease is one of the most widely spreading and deadliest diseases across the world. In this study, we have proposed hybrid model for heart disease prediction by employing random forest and support vector machine. With random forest, iterative feature elimination is carried out to select heart disease features that improves predictive outcome of support vector machine for heart disease prediction. Experiment is conducted on the proposed model using test set and the experimental result evidently appears to prove that the performance of the proposed hybrid model is better as compared to an individual random forest and support vector machine. Overall, we have developed more accurate and computationally efficient model for heart disease prediction with accuracy of 98.3%. Moreover, experiment is conducted to analyze the effect of regularization parameter (C) and gamma on the performance of support vector machine. The experimental result evidently reveals that support vector machine is very sensitive to C and gamma.
An Ensemble of Filters and Wrappers for Microarray Data Classification mlaij
The development of microarray technology has suppli
ed a large volume of data to many fields. The gene
microarray analysis and classification have demonst
rated an effective way for the effective diagnosis
of
diseases and cancers. In as much as the data achiev
ing from microarray technology is very noisy and al
so
has thousands of features, feature selection plays
an important role in removing irrelevant and redund
ant
features and also reducing computational complexity
. There are two important approaches for gene
selection in microarray data analysis, the filters
and the wrappers. To select a concise subset of inf
ormative
genes, we introduce a hybrid feature selection whic
h combines two approaches. The fact of the matter i
s
that candidate’s features are first selected from t
he original set via several effective filters. The
candidate
feature set is further refined by more accurate wra
ppers. Thus, we can take advantage of both the filt
ers
and wrappers. Experimental results based on 11 micr
oarray datasets show that our mechanism can be
effected with a smaller feature set. Moreover, thes
e feature subsets can be obtained in a reasonable t
ime
An Ensemble of Filters and Wrappers for Microarray Data Classification mlaij
The development of microarray technology has supplied a large volume of data to many fields. The gene microarray analysis and classification have demonstrated an effective way for the effective diagnosis of diseases and cancers. In as much as the data achieving from microarray technology is very noisy and also has thousands of features, feature selection plays an important role in removing irrelevant and redundant features and also reducing computational complexity. There are two important approaches for gene selection in microarray data analysis, the filters and the wrappers. To select a concise subset of informative genes, we introduce a hybrid feature selection which combines two approaches. The fact of the matter is that candidate’s features are first selected from the original set via several effective filters. The candidate feature set is further refined by more accurate wrappers. Thus, we can take advantage of both the filters and wrappers. Experimental results based on 11 microarray datasets show that our mechanism can be effected with a smaller feature set. Moreover, these feature subsets can be obtained in a reasonable time.
A discriminative-feature-space-for-detecting-and-recognizing-pathologies-of-t...Damian R. Mingle, MBA
Each year it has become more and more difficult for healthcare providers to determine if a patient has a pathology
related to the vertebral column. There is great potential to become more efficient and effective in terms of quality
of care provided to patients through the use of automated systems. However, in many cases automated systems
can allow for misclassification and force providers to have to review more causes than necessary. In this study, we
analyzed methods to increase the True Positives and lower the False Positives while comparing them against stateof-the-art
techniques in the biomedical community. We found that by applying the studied techniques of a data-driven
model, the benefits to healthcare providers are significant and align with the methodologies and techniques utilized
in the current research community.
Classification of Health Care Data Using Machine Learning Techniqueinventionjournals
Diagnosis of any diseases using intelligent system produces high accuracy and treated as classification problem, also classification of health care data using machine learning techniques may be used as intelligent health care system (IHSM). Diagnosis of heart disease is a major issue and commonly found in human being due to changing life style and also need to be identified in advance. This paper explores various machine learning techniques to classify heart data downloaded from UCI repository site. After applying feature selection technique (FST), a decision tree based technique: CART is producing high accuracy with four features followed by other classification techniques.
Machine learning approach for predicting heart and diabetes diseases using da...IAESIJAI
Environmental changes and food habits affect people's health with numerous diseases in today's life. Machine learning is a technique that plays a vital role in predicting diseases from collected data. The health sector has plenty of electronic medical data, which helps this technique to diagnose various diseases quickly and accurately. There has been an improvement in accuracy in medical data analysis as data continues to grow in the medical field. Doctors may have a hard time predicting symptoms accurately. This proposed work utilized Kaggle data to predict and diagnose heart and diabetic diseases. The diseases heart and diabetes are the foremost cause of higher death rates for people. The dataset contains target features for the diagnosis of heart disease. This work finds the target variable for diabetic disease by comparing the patient's blood sugars to normal levels. Blood pressure, body mass index (BMI), and other factors diagnose these diseases and disorders. This work justifies the filter method and principal component analysis for selecting and extracting the feature. The main aim of this work is to highlight the implementation of three ensemble techniques-Adaptive boost, Extreme Gradient boosting, and Gradient boosting-as well as the emphasis placed on the accuracy of the results.
Evolving Efficient Clustering and Classification Patterns in Lymphography Dat...ijsc
Data mining refers to the process of retrieving knowledge by discovering novel and relative patterns from large datasets. Clustering and Classification are two distinct phases in data mining that work to provide an established, proven structure from a voluminous collection of facts. A dominant area of modern-day research in the field of medical investigations includes disease prediction and malady categorization. In this paper, our focus is to analyze clusters of patient records obtained via unsupervised clustering techniques and compare the performance of classification algorithms on the clinical data. Feature selection is a supervised method that attempts to select a subset of the predictor features based on the information gain. The Lymphography dataset comprises of 18 predictor attributes and 148 instances with the class label having four distinct values. This paper highlights the accuracy of eight clustering algorithms in detecting clusters of patient records and predictor attributes and highlights the performance of sixteen classification algorithms on the Lymphography dataset that enables the classifier to accurately perform multi-class categorization of medical data. Our work asserts the fact that the Random Tree algorithm and the Quinlan’s C4.5 algorithm give 100 percent classification accuracy with all the predictor features and also with the feature subset selected by the Fisher Filtering feature selection algorithm.. It is also stated here that the Density Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm offers increased clustering accuracy in less computation time.
EVOLVING EFFICIENT CLUSTERING AND CLASSIFICATION PATTERNS IN LYMPHOGRAPHY DAT...ijsc
Data mining refers to the process of retrieving knowledge by discovering novel and relative patterns from
large datasets. Clustering and Classification are two distinct phases in data mining that work to provide an
established, proven structure from a voluminous collection of facts. A dominant area of modern-day
research in the field of medical investigations includes disease prediction and malady categorization. In
this paper, our focus is to analyze clusters of patient records obtained via unsupervised clustering
techniques and compare the performance of classification algorithms on the clinical data. Feature
selection is a supervised method that attempts to select a subset of the predictor features based on the
information gain. The Lymphography dataset comprises of 18 predictor attributes and 148 instances with
the class label having four distinct values. This paper highlights the accuracy of eight clustering algorithms
in detecting clusters of patient records and predictor attributes and highlights the performance of sixteen
classification algorithms on the Lymphography dataset that enables the classifier to accurately perform
multi-class categorization of medical data. Our work asserts the fact that the Random Tree algorithm and
the Quinlan’s C4.5 algorithm give 100 percent classification accuracy with all the predictor features and
also with the feature subset selected by the Fisher Filtering feature selection algorithm.. It is also stated
here that the Density Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm
offers increased clustering accuracy in less computation time.
SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu...IJECEIAES
In many diseases classification an accurate gene analysis is needed, for which selection of most informative genes is very important and it require a technique of decision in complex context of ambiguity. The traditional methods include for selecting most significant gene includes some of the statistical analysis namely 2-Sample-T-test (2STT), Entropy, Signal to Noise Ratio (SNR). This paper evaluates gene selection and classification on the basis of accurate gene selection using structured complex decision technique (SCDT) and classifies it using fuzzy cluster based nearest neighborclassifier (FC-NNC). The effectiveness of the proposed SCDT and FC-NNC is evaluated for leave one out cross validation metric(LOOCV) along with sensitivity, specificity, precision and F1-score with four different classifiers namely 1) Radial Basis Function (RBF), 2) Multi-layer perception(MLP), 3) Feed Forward(FF) and 4) Support vector machine(SVM) for three different datasets of DLBCL, Leukemia and Prostate tumor. The proposed SCDT &FC-NNC exhibits superior result for being considered more accurate decision mechanism.
CARDIOVASCULAR DISEASE DETECTION USING MACHINE LEARNING AND RISK CLASSIFICATI...indexPub
The global prevalence of heart disease indicates a major public health issue. It causes shortness of breath, weakness, and swollen ankles. Early heart disease diagnosis is difficult with current approaches. Hence, a better heart disease detection tool is needed. Treatment requires more than just diagnosis. Risk classification is critical for accurate diagnosis and treatment. In this analysis, a novel cardiovascular disease (CVD) detection paradigm using machine learning (ML) and risk classification based on a weighted fuzzy system is proposed. The system is developed based on ML algorithms such as artificial neural network (ANN) and Long Short-Term Memory (LSTM) and uses standard feature selection techniques knowns as Principal Component Analysis (PCA). Furthermore, the cross-validation method has been used for learning the best practices of model assessment and for hyperparameter tuning. The accuracy-based performance measuring metrics are used for the assessment of the performances of the classifiers. Finally, the outcomes revealed that the proposed model achieved an accuracy of 94.01% which is higher than another conventional model developed in this domain. Additionally, the proposed system can easily be implemented in healthcare for the identification of heart disease.
An Enhanced Feature Selection Method to Predict the Severity in Brain Tumorijtsrd
The level of severity of brain tumor is captured through MRI and then assessed by the physician for their medical interpretation. The facts behind the MRI images are then analyzed by the physician for further medication and follow up activities. An MRI image composed of large volume of features. It has irrelevant, missing and information which is not certain. In medical data analysis, an MRI image doesn't express facts very clearly to the physician for correct interpretation all the time. It also includes huge amount of redundant information within it. A mathematical model known as rough set theory has been applied to resolve this problem by eliminating the redundancy in medical image data. This paper uses a rough set method to find the severity level of the brain tumor of the given MRI image. Rough set feature selection algorithms are applied over the medical image data to select the prominent features. The classification accuracy of the brain tumor can be improved to a better level by using this rough set approach. The prominent features selected through this approach deliver a set of decision rules for the classification task. A search method based on the particle swarm optimization is proposed in this paper for minimizing the attribute set. This approach is compared with previously existing rough set reduction algorithm for finding the accuracy. The reducts originated from the proposed algorithm is more efficient and can generate decision rules that will better classify the tumor types. The rule based method provided by the rough set method delivers classification accuracy in higher level than other smart methods such as fuzzy rule extraction, neural networks, decision trees and Fuzzy Networks like Fuzzy Min Max Neural Networks. Parthiban J | Dr. B. Sathees Kumar "An Enhanced Feature Selection Method to Predict the Severity in Brain Tumor" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26802.pdfPaper URL: https://www.ijtsrd.com/computer-science/data-miining/26802/an-enhanced-feature-selection-method-to-predict-the-severity-in-brain-tumor/parthiban-j
Development of depth map from stereo images using sum of absolute differences...nooriasukmaningtyas
This article proposes a framework for the depth map reconstruction using stereo images. Fundamentally, this map provides an important information which commonly used in essential applications such as autonomous vehicle navigation, drone’s navigation and 3D surface reconstruction. To develop an accurate depth map, the framework must be robust against the challenging regions of low texture, plain color and repetitive pattern on the input stereo image. The development of this map requires several stages which starts with matching cost calculation, cost aggregation, optimization and refinement stage. Hence, this work develops a framework with sum of absolute difference (SAD) and the combination of two edge preserving filters to increase the robustness against the challenging regions. The SAD convolves using block matching technique to increase the efficiency of matching process on the low texture and plain color regions. Moreover, two edge preserving filters will increase the accuracy on the repetitive pattern region. The results show that the proposed method is accurate and capable to work with the challenging regions. The results are provided by the Middlebury standard dataset. The framework is also efficiently and can be applied on the 3D surface reconstruction. Moreover, this work is greatly competitive with previously available methods.
Model predictive controller for a retrofitted heat exchanger temperature cont...nooriasukmaningtyas
This paper aims to demonstrate the practical aspects of process control theory for undergraduate students at the Department of Chemical Engineering at the University of Bahrain. Both, the ubiquitous proportional integral derivative (PID) as well as model predictive control (MPC) and their auxiliaries were designed and implemented in a real-time framework. The latter was realized through retrofitting an existing plate-and-frame heat exchanger unit that has been operated using an analog PID temperature controller. The upgraded control system consists of a personal computer (PC), low-cost signal conditioning circuit, national instruments USB 6008 data acquisition card, and LabVIEW software. LabVIEW control design and simulation modules were used to design and implement the PID and MPC controllers. The performance of the designed controllers was evaluated while controlling the outlet temperature of the retrofitted plate-and-frame heat exchanger. The distinguished feature of the MPC controller in handling input and output constraints was perceived in real-time. From a pedagogical point of view, realizing the theory of process control through practical implementation was substantial in enhancing the student’s learning and the instructor’s teaching experience.
More Related Content
Similar to Predicting heart failure using a wrapper-based feature selection
Genome feature optimization and coronary artery disease prediction using cuck...CSITiaesprime
Cardiovascular diseases are among the major health ailment issue leading to millions of deaths every year. In recent past, analyzing gene expression data, particularly using machine learning strategies to predict and classify the given unlabeled gene expression record is a generous research issue. Concerning this, a substantial requirement is feature optimization, which is since the overall genes observed in human body are closely 25000 and among them 636 are cardiovascular related genes. Hence, it complexes the process of training the machine learning models using these entire cardiovascular gene features. This manuscript uses bidirectional pooled variance strategy of ANOVA standard to select optimal features. Along the side to surpass the constraint observed in traditional classifiers, which is unstable accuracy at k-fold cross validation, this manuscript proposed a classification strategy that build upon the swarm intelligence technique called cuckoo search. The experimental study indicating that the number of optimal features those selected by proposed model is substantially low that compared to the other contemporary model that selects features using forward feature selection and classifies using support vector machine classifier (FFS&SVM). The experimental study evinced that the proposed model, which selects feature by bidirectional pooled variance estimation and classifies using proposed classification strategy that build on cuckoo search (BPVE&CS) outperformed the selected contemporary model (FFS&SVM).
Propose a Enhanced Framework for Prediction of Heart DiseaseIJERA Editor
Heart disease diagnosis requires more experience and it is a complex task. The Heart MRI, ECG and Stress Test etc are the numbers of medical tests are prescribed by the doctor for examining the heart disease and it is the way of tradition in the prediction of heart disease. Today world, the hidden information of the huge amount of health care data is contained by the health care industry. The effective decisions are made by means of this hidden information. For appropriate results, the advanced data mining techniques with the information which is based on the computer are used. In any empirical sciences, for the inference and categorisation, the new mathematical techniques to be used called Artificial neural networks (ANNs) it also be used to the modelling of the real neural networks. Acting, Wanting, knowing, remembering, perceiving, thinking and inferring are the nature of mental phenomena and these can be understand by using the theory of ANN. The problem of probability and induction can be arised for the inference and classification because these are the powerful instruments of ANN. In this paper, the classification techniques like Naive Bayes Classification algorithm and Artificial Neural Networks are used to classify the attributes in the given data set. The attribute filtering techniques like PCA (Principle Component Analysis) filtering and Information Gain Attribute Subset Evaluation technique for feature selection in the given data set to predict the heart disease symptoms. A new framework is proposed which is based on the above techniques, the framework will take the input dataset and fed into the feature selection techniques block, which selects any one techniques that gives the least number of attributes and then classification task is done using two algorithms, the same attributes that are selected by two classification task is taken for the prediction of heart disease. This framework consumes the time for predicting the symptoms of heart disease which make the user to know the important attributes based on the proposed framework.
PREDICTIVE ANALYTICS IN HEALTHCARE SYSTEM USING DATA MINING TECHNIQUEScscpconf
The health sector has witnessed a great evolution following the development of new computer technologies, and that pushed this area to produce more medical data, which gave birth to multiple fields of research. Many efforts are done to cope with the explosion of medical data on one hand, and to obtain useful knowledge from it on the other hand. This prompted researchers to apply all the technical innovations like big data analytics, predictive analytics, machine learning and learning algorithms in order to extract useful knowledge and help in making decisions. With the promises of predictive analytics in big data, and the use of machine learning
algorithms, predicting future is no longer a difficult task, especially for medicine because predicting diseases and anticipating the cure became possible. In this paper we will present an overview on the evolution of big data in healthcare system, and we will apply a learning algorithm on a set of medical data. The objective is to predict chronic kidney diseases by using Decision Tree (C4.5) algorithm.
PREDICTION OF MALIGNANCY IN SUSPECTED THYROID TUMOUR PATIENTS BY THREE DIFFER...cscpconf
In the present study, the abilities of three classification methods of data mining namely artificial
neural networks with feed-forward back propagation algorithm, J48 decision tree method and
logistic regression analysis are compared in a medical real dataset. The prediction of
malignancy in suspected thyroid tumour patients is the objective of the study. The accuracy of
the correct predictions (the minimum error rate), the amount of time consuming in the
modelling process and the interpretability and simplicity of the results for clinical experts are
the factors considered to choose the best method
PERFORMANCE OF DATA MINING TECHNIQUES TO PREDICT IN HEALTHCARE CASE STUDY: CH...ijdms
With the promises of predictive analytics in big data, and the use of machine learning algorithms,
predicting future is no longer a difficult task, especially for health sector, that has witnessed a great
evolution following the development of new computer technologies that gave birth to multiple fields of
research. Many efforts are done to cope with medical data explosion on one hand, and to obtain useful
knowledge from it, predict diseases and anticipate the cure on the other hand. This prompted researchers
to apply all the technical innovations like big data analytics, predictive analytics, machine learning and
learning algorithms in order to extract useful knowledge and help in making decisions. In this paper, we
will present an overview on the evolution of big data in healthcare system, and we will apply three learning
algorithms on a set of medical data. The objective of this research work is to predict kidney disease by
using multiple machine learning algorithms that are Support Vector Machine (SVM), Decision Tree (C4.5),
and Bayesian Network (BN), and chose the most efficient one.
Diagnosis of rheumatoid arthritis using an ensemble learning approachcsandit
Rheumatoid arthritis is one of the diseases that it
s cause is unknown yet; exploring the field of
medical data mining can be helpful in early diagnos
is and treatment of the disease. In this
study, a predictive model is suggested that diagnos
es rheumatoid arthritis. The rheumatoid
arthritis dataset was collected from 2,564 patients
referred to rheumatology clinic. For each
patient a record consists of several clinical and d
emographic features is saved. After data
analysis and pre-processing operations, three diffe
rent methods are combined to choose proper
features among all the features. Various data class
ification algorithms were applied on these
features. Among these algorithms Adaboost had the h
ighest precision. In this paper, we
proposed a new classification algorithm entitled CS
-Boost that employs Cuckoo search
algorithm for optimizing the performance of Adaboos
t algorithm. Experimental results show
that the CS-Boost algorithm enhance the accuracy of
Adaboost in predicting of Rheumatoid
Arthritis.
DIAGNOSIS OF RHEUMATOID ARTHRITIS USING AN ENSEMBLE LEARNING APPROACH cscpconf
Rheumatoid arthritis is one of the diseases that its cause is unknown yet; exploring the field of
medical data mining can be helpful in early diagnosis and treatment of the disease. In this
study, a predictive model is suggested that diagnoses rheumatoid arthritis. The rheumatoid
arthritis dataset was collected from 2,564 patients referred to rheumatology clinic. For each
patient a record consists of several clinical and demographic features is saved. After data
analysis and pre-processing operations, three different methods are combined to choose proper
features among all the features. Various data classification algorithms were applied on these
features. Among these algorithms Adaboost had the highest precision. In this paper, we
proposed a new classification algorithm entitled CS-Boost that employs Cuckoo search
algorithm for optimizing the performance of Adaboost algorithm. Experimental results show
that the CS-Boost algorithm enhance the accuracy of Adaboost in predicting of Rheumatoid
Arthritis.
A hybrid approach to medical decision-making: diagnosis of heart disease wit...IJECEIAES
Heart disease is one of the most widely spreading and deadliest diseases across the world. In this study, we have proposed hybrid model for heart disease prediction by employing random forest and support vector machine. With random forest, iterative feature elimination is carried out to select heart disease features that improves predictive outcome of support vector machine for heart disease prediction. Experiment is conducted on the proposed model using test set and the experimental result evidently appears to prove that the performance of the proposed hybrid model is better as compared to an individual random forest and support vector machine. Overall, we have developed more accurate and computationally efficient model for heart disease prediction with accuracy of 98.3%. Moreover, experiment is conducted to analyze the effect of regularization parameter (C) and gamma on the performance of support vector machine. The experimental result evidently reveals that support vector machine is very sensitive to C and gamma.
An Ensemble of Filters and Wrappers for Microarray Data Classification mlaij
The development of microarray technology has suppli
ed a large volume of data to many fields. The gene
microarray analysis and classification have demonst
rated an effective way for the effective diagnosis
of
diseases and cancers. In as much as the data achiev
ing from microarray technology is very noisy and al
so
has thousands of features, feature selection plays
an important role in removing irrelevant and redund
ant
features and also reducing computational complexity
. There are two important approaches for gene
selection in microarray data analysis, the filters
and the wrappers. To select a concise subset of inf
ormative
genes, we introduce a hybrid feature selection whic
h combines two approaches. The fact of the matter i
s
that candidate’s features are first selected from t
he original set via several effective filters. The
candidate
feature set is further refined by more accurate wra
ppers. Thus, we can take advantage of both the filt
ers
and wrappers. Experimental results based on 11 micr
oarray datasets show that our mechanism can be
effected with a smaller feature set. Moreover, thes
e feature subsets can be obtained in a reasonable t
ime
An Ensemble of Filters and Wrappers for Microarray Data Classification mlaij
The development of microarray technology has supplied a large volume of data to many fields. The gene microarray analysis and classification have demonstrated an effective way for the effective diagnosis of diseases and cancers. In as much as the data achieving from microarray technology is very noisy and also has thousands of features, feature selection plays an important role in removing irrelevant and redundant features and also reducing computational complexity. There are two important approaches for gene selection in microarray data analysis, the filters and the wrappers. To select a concise subset of informative genes, we introduce a hybrid feature selection which combines two approaches. The fact of the matter is that candidate’s features are first selected from the original set via several effective filters. The candidate feature set is further refined by more accurate wrappers. Thus, we can take advantage of both the filters and wrappers. Experimental results based on 11 microarray datasets show that our mechanism can be effected with a smaller feature set. Moreover, these feature subsets can be obtained in a reasonable time.
A discriminative-feature-space-for-detecting-and-recognizing-pathologies-of-t...Damian R. Mingle, MBA
Each year it has become more and more difficult for healthcare providers to determine if a patient has a pathology
related to the vertebral column. There is great potential to become more efficient and effective in terms of quality
of care provided to patients through the use of automated systems. However, in many cases automated systems
can allow for misclassification and force providers to have to review more causes than necessary. In this study, we
analyzed methods to increase the True Positives and lower the False Positives while comparing them against stateof-the-art
techniques in the biomedical community. We found that by applying the studied techniques of a data-driven
model, the benefits to healthcare providers are significant and align with the methodologies and techniques utilized
in the current research community.
Classification of Health Care Data Using Machine Learning Techniqueinventionjournals
Diagnosis of any diseases using intelligent system produces high accuracy and treated as classification problem, also classification of health care data using machine learning techniques may be used as intelligent health care system (IHSM). Diagnosis of heart disease is a major issue and commonly found in human being due to changing life style and also need to be identified in advance. This paper explores various machine learning techniques to classify heart data downloaded from UCI repository site. After applying feature selection technique (FST), a decision tree based technique: CART is producing high accuracy with four features followed by other classification techniques.
Machine learning approach for predicting heart and diabetes diseases using da...IAESIJAI
Environmental changes and food habits affect people's health with numerous diseases in today's life. Machine learning is a technique that plays a vital role in predicting diseases from collected data. The health sector has plenty of electronic medical data, which helps this technique to diagnose various diseases quickly and accurately. There has been an improvement in accuracy in medical data analysis as data continues to grow in the medical field. Doctors may have a hard time predicting symptoms accurately. This proposed work utilized Kaggle data to predict and diagnose heart and diabetic diseases. The diseases heart and diabetes are the foremost cause of higher death rates for people. The dataset contains target features for the diagnosis of heart disease. This work finds the target variable for diabetic disease by comparing the patient's blood sugars to normal levels. Blood pressure, body mass index (BMI), and other factors diagnose these diseases and disorders. This work justifies the filter method and principal component analysis for selecting and extracting the feature. The main aim of this work is to highlight the implementation of three ensemble techniques-Adaptive boost, Extreme Gradient boosting, and Gradient boosting-as well as the emphasis placed on the accuracy of the results.
Evolving Efficient Clustering and Classification Patterns in Lymphography Dat...ijsc
Data mining refers to the process of retrieving knowledge by discovering novel and relative patterns from large datasets. Clustering and Classification are two distinct phases in data mining that work to provide an established, proven structure from a voluminous collection of facts. A dominant area of modern-day research in the field of medical investigations includes disease prediction and malady categorization. In this paper, our focus is to analyze clusters of patient records obtained via unsupervised clustering techniques and compare the performance of classification algorithms on the clinical data. Feature selection is a supervised method that attempts to select a subset of the predictor features based on the information gain. The Lymphography dataset comprises of 18 predictor attributes and 148 instances with the class label having four distinct values. This paper highlights the accuracy of eight clustering algorithms in detecting clusters of patient records and predictor attributes and highlights the performance of sixteen classification algorithms on the Lymphography dataset that enables the classifier to accurately perform multi-class categorization of medical data. Our work asserts the fact that the Random Tree algorithm and the Quinlan’s C4.5 algorithm give 100 percent classification accuracy with all the predictor features and also with the feature subset selected by the Fisher Filtering feature selection algorithm.. It is also stated here that the Density Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm offers increased clustering accuracy in less computation time.
EVOLVING EFFICIENT CLUSTERING AND CLASSIFICATION PATTERNS IN LYMPHOGRAPHY DAT...ijsc
Data mining refers to the process of retrieving knowledge by discovering novel and relative patterns from
large datasets. Clustering and Classification are two distinct phases in data mining that work to provide an
established, proven structure from a voluminous collection of facts. A dominant area of modern-day
research in the field of medical investigations includes disease prediction and malady categorization. In
this paper, our focus is to analyze clusters of patient records obtained via unsupervised clustering
techniques and compare the performance of classification algorithms on the clinical data. Feature
selection is a supervised method that attempts to select a subset of the predictor features based on the
information gain. The Lymphography dataset comprises of 18 predictor attributes and 148 instances with
the class label having four distinct values. This paper highlights the accuracy of eight clustering algorithms
in detecting clusters of patient records and predictor attributes and highlights the performance of sixteen
classification algorithms on the Lymphography dataset that enables the classifier to accurately perform
multi-class categorization of medical data. Our work asserts the fact that the Random Tree algorithm and
the Quinlan’s C4.5 algorithm give 100 percent classification accuracy with all the predictor features and
also with the feature subset selected by the Fisher Filtering feature selection algorithm.. It is also stated
here that the Density Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm
offers increased clustering accuracy in less computation time.
SCDT: FC-NNC-structured Complex Decision Technique for Gene Analysis Using Fu...IJECEIAES
In many diseases classification an accurate gene analysis is needed, for which selection of most informative genes is very important and it require a technique of decision in complex context of ambiguity. The traditional methods include for selecting most significant gene includes some of the statistical analysis namely 2-Sample-T-test (2STT), Entropy, Signal to Noise Ratio (SNR). This paper evaluates gene selection and classification on the basis of accurate gene selection using structured complex decision technique (SCDT) and classifies it using fuzzy cluster based nearest neighborclassifier (FC-NNC). The effectiveness of the proposed SCDT and FC-NNC is evaluated for leave one out cross validation metric(LOOCV) along with sensitivity, specificity, precision and F1-score with four different classifiers namely 1) Radial Basis Function (RBF), 2) Multi-layer perception(MLP), 3) Feed Forward(FF) and 4) Support vector machine(SVM) for three different datasets of DLBCL, Leukemia and Prostate tumor. The proposed SCDT &FC-NNC exhibits superior result for being considered more accurate decision mechanism.
CARDIOVASCULAR DISEASE DETECTION USING MACHINE LEARNING AND RISK CLASSIFICATI...indexPub
The global prevalence of heart disease indicates a major public health issue. It causes shortness of breath, weakness, and swollen ankles. Early heart disease diagnosis is difficult with current approaches. Hence, a better heart disease detection tool is needed. Treatment requires more than just diagnosis. Risk classification is critical for accurate diagnosis and treatment. In this analysis, a novel cardiovascular disease (CVD) detection paradigm using machine learning (ML) and risk classification based on a weighted fuzzy system is proposed. The system is developed based on ML algorithms such as artificial neural network (ANN) and Long Short-Term Memory (LSTM) and uses standard feature selection techniques knowns as Principal Component Analysis (PCA). Furthermore, the cross-validation method has been used for learning the best practices of model assessment and for hyperparameter tuning. The accuracy-based performance measuring metrics are used for the assessment of the performances of the classifiers. Finally, the outcomes revealed that the proposed model achieved an accuracy of 94.01% which is higher than another conventional model developed in this domain. Additionally, the proposed system can easily be implemented in healthcare for the identification of heart disease.
An Enhanced Feature Selection Method to Predict the Severity in Brain Tumorijtsrd
The level of severity of brain tumor is captured through MRI and then assessed by the physician for their medical interpretation. The facts behind the MRI images are then analyzed by the physician for further medication and follow up activities. An MRI image composed of large volume of features. It has irrelevant, missing and information which is not certain. In medical data analysis, an MRI image doesn't express facts very clearly to the physician for correct interpretation all the time. It also includes huge amount of redundant information within it. A mathematical model known as rough set theory has been applied to resolve this problem by eliminating the redundancy in medical image data. This paper uses a rough set method to find the severity level of the brain tumor of the given MRI image. Rough set feature selection algorithms are applied over the medical image data to select the prominent features. The classification accuracy of the brain tumor can be improved to a better level by using this rough set approach. The prominent features selected through this approach deliver a set of decision rules for the classification task. A search method based on the particle swarm optimization is proposed in this paper for minimizing the attribute set. This approach is compared with previously existing rough set reduction algorithm for finding the accuracy. The reducts originated from the proposed algorithm is more efficient and can generate decision rules that will better classify the tumor types. The rule based method provided by the rough set method delivers classification accuracy in higher level than other smart methods such as fuzzy rule extraction, neural networks, decision trees and Fuzzy Networks like Fuzzy Min Max Neural Networks. Parthiban J | Dr. B. Sathees Kumar "An Enhanced Feature Selection Method to Predict the Severity in Brain Tumor" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-3 | Issue-5 , August 2019, URL: https://www.ijtsrd.com/papers/ijtsrd26802.pdfPaper URL: https://www.ijtsrd.com/computer-science/data-miining/26802/an-enhanced-feature-selection-method-to-predict-the-severity-in-brain-tumor/parthiban-j
Similar to Predicting heart failure using a wrapper-based feature selection (20)
Development of depth map from stereo images using sum of absolute differences...nooriasukmaningtyas
This article proposes a framework for the depth map reconstruction using stereo images. Fundamentally, this map provides an important information which commonly used in essential applications such as autonomous vehicle navigation, drone’s navigation and 3D surface reconstruction. To develop an accurate depth map, the framework must be robust against the challenging regions of low texture, plain color and repetitive pattern on the input stereo image. The development of this map requires several stages which starts with matching cost calculation, cost aggregation, optimization and refinement stage. Hence, this work develops a framework with sum of absolute difference (SAD) and the combination of two edge preserving filters to increase the robustness against the challenging regions. The SAD convolves using block matching technique to increase the efficiency of matching process on the low texture and plain color regions. Moreover, two edge preserving filters will increase the accuracy on the repetitive pattern region. The results show that the proposed method is accurate and capable to work with the challenging regions. The results are provided by the Middlebury standard dataset. The framework is also efficiently and can be applied on the 3D surface reconstruction. Moreover, this work is greatly competitive with previously available methods.
Model predictive controller for a retrofitted heat exchanger temperature cont...nooriasukmaningtyas
This paper aims to demonstrate the practical aspects of process control theory for undergraduate students at the Department of Chemical Engineering at the University of Bahrain. Both, the ubiquitous proportional integral derivative (PID) as well as model predictive control (MPC) and their auxiliaries were designed and implemented in a real-time framework. The latter was realized through retrofitting an existing plate-and-frame heat exchanger unit that has been operated using an analog PID temperature controller. The upgraded control system consists of a personal computer (PC), low-cost signal conditioning circuit, national instruments USB 6008 data acquisition card, and LabVIEW software. LabVIEW control design and simulation modules were used to design and implement the PID and MPC controllers. The performance of the designed controllers was evaluated while controlling the outlet temperature of the retrofitted plate-and-frame heat exchanger. The distinguished feature of the MPC controller in handling input and output constraints was perceived in real-time. From a pedagogical point of view, realizing the theory of process control through practical implementation was substantial in enhancing the student’s learning and the instructor’s teaching experience.
Control of a servo-hydraulic system utilizing an extended wavelet functional ...nooriasukmaningtyas
Servo-hydraulic systems have been extensively employed in various industrial applications. However, these systems are characterized by their highly complex and nonlinear dynamics, which complicates the control design stage of such systems. In this paper, an extended wavelet functional link neural network (EWFLNN) is proposed to control the displacement response of the servo-hydraulic system. To optimize the controller's parameters, a recently developed optimization technique, which is called the modified sine cosine algorithm (M-SCA), is exploited as the training method. The proposed controller has achieved remarkable results in terms of tracking two different displacement signals and handling external disturbances. From a comparative study, the proposed EWFLNN controller has attained the best control precision compared with those of other controllers, namely, a proportional-integralderivative (PID) controller, an artificial neural network (ANN) controller, a wavelet neural network (WNN) controller, and the original wavelet functional link neural network (WFLNN) controller. Moreover, compared to the genetic algorithm (GA) and the original sine cosine algorithm (SCA), the M-SCA has shown better optimization results in finding the optimal values of the controller's parameters.
Decentralised optimal deployment of mobile underwater sensors for covering la...nooriasukmaningtyas
This paper presents the problem of sensing coverage of layers of the ocean in three dimensional underwater environments. We propose distributed control laws to drive mobile underwater sensors to optimally cover a given confined layer of the ocean. By applying this algorithm at first the mobile underwater sensors adjust their depth to the specified depth. Then, they make a triangular grid across a given area. Afterwards, they randomly move to spread across the given grid. These control laws only rely on local information also they are easily implemented and computationally effective as they use some easy consensus rules. The feature of exchanging information just among neighbouring mobile sensors keeps the information exchange minimum in the whole networks and makes this algorithm practicable option for undersea. The efficiency of the presented control laws is confirmed via mathematical proof and numerical simulations.
Evaluation quality of service for internet of things based on fuzzy logic: a ...nooriasukmaningtyas
The development of the internet of thing (IoT) technology has become a major concern in sustainability of quality of service (SQoS) in terms of efficiency, measurement, and evaluation of services, such as our smart home case study. Based on several ambiguous linguistic and standard criteria, this article deals with quality of service (QoS). We used fuzzy logic to select the most appropriate and efficient services. For this reason, we have introduced a new paradigmatic approach to assess QoS. In this regard, to measure SQoS, linguistic terms were collected for identification of ambiguous criteria. This paper collects the results of other work to compare the traditional assessment methods and techniques in IoT. It has been proven that the comparison that traditional valuation methods and techniques could not effectively deal with these metrics. Therefore, fuzzy logic is a worthy method to provide a good measure of QoS with ambiguous linguistic and criteria. The proposed model addresses with constantly being improved, all the main axes of the QoS for a smart home. The results obtained also indicate that the model with its fuzzy performance importance index (FPII) has efficiently evaluate the multiple services of SQoS.
Low power architecture of logic gates using adiabatic techniquesnooriasukmaningtyas
The growing significance of portable systems to limit power consumption in ultra-large-scale-integration chips of very high density, has recently led to rapid and inventive progresses in low-power design. The most effective technique is adiabatic logic circuit design in energy-efficient hardware. This paper presents two adiabatic approaches for the design of low power circuits, modified positive feedback adiabatic logic (modified PFAL) and the other is direct current diode based positive feedback adiabatic logic (DC-DB PFAL). Logic gates are the preliminary components in any digital circuit design. By improving the performance of basic gates, one can improvise the whole system performance. In this paper proposed circuit design of the low power architecture of OR/NOR, AND/NAND, and XOR/XNOR gates are presented using the said approaches and their results are analyzed for powerdissipation, delay, power-delay-product and rise time and compared with the other adiabatic techniques along with the conventional complementary metal oxide semiconductor (CMOS) designs reported in the literature. It has been found that the designs with DC-DB PFAL technique outperform with the percentage improvement of 65% for NOR gate and 7% for NAND gate and 34% for XNOR gate over the modified PFAL techniques at 10 MHz respectively.
A review on techniques and modelling methodologies used for checking electrom...nooriasukmaningtyas
The proper function of the integrated circuit (IC) in an inhibiting electromagnetic environment has always been a serious concern throughout the decades of revolution in the world of electronics, from disjunct devices to today’s integrated circuit technology, where billions of transistors are combined on a single chip. The automotive industry and smart vehicles in particular, are confronting design issues such as being prone to electromagnetic interference (EMI). Electronic control devices calculate incorrect outputs because of EMI and sensors give misleading values which can prove fatal in case of automotives. In this paper, the authors have non exhaustively tried to review research work concerned with the investigation of EMI in ICs and prediction of this EMI using various modelling methodologies and measurement setups.
Smart monitoring system using NodeMCU for maintenance of production machinesnooriasukmaningtyas
Maintenance is an activity that helps to reduce risk, increase productivity, improve quality, and minimize production costs. The necessity for maintenance actions will increase efficiency and enhance the safety and quality of products and processes. On getting these conditions, it is necessary to implement a monitoring system used to observe machines' conditions from time to time, especially the machine parts that often experience problems. This paper presents a low-cost intelligent monitoring system using NodeMCU to continuously monitor machine conditions and provide warnings in the case of machine failure. Not only does it provide alerts, but this monitoring system also generates historical data on machine conditions to the Google Cloud (Google Sheet), includes which machines were down, downtime, issues occurred, repairs made, and technician handling. The results obtained are machine operators do not need to lose a relatively long time to call the technician. Likewise, the technicians assisted in carrying out machine maintenance activities and online reports so that errors that often occur due to human error do not happen again. The system succeeded in reducing the technician-calling time and maintenance workreporting time up to 50%. The availability of online and real-time maintenance historical data will support further maintenance strategy.
Design and simulation of a software defined networkingenabled smart switch, f...nooriasukmaningtyas
Using sustainable energy is the future of our planet earth, this became not only economically efficient but also a necessity for the preservation of life on earth. Because of such necessity, smart grids became a very important issue to be researched. Many literatures discussed this topic and with the development of internet of things (IoT) and smart sensors, smart grids are developed even further. On the other hand, software defined networking is a technology that separates the control plane from the data plan of the network. It centralizes the management and the orchestration of the network tasks by using a network controller. The network controller is the heart of the SDN-enabled network, and it can control other networking devices using software defined networking (SDN) protocols such as OpenFlow. A smart switching mechanism called (SDN-smgrid-sw) for the smart grid will be modeled and controlled using SDN. We modeled the environment that interact with the sensors, for the sun and the wind elements. The Algorithm is modeled and programmed for smart efficient power sharing that is managed centrally and monitored using SDN controller. Also, all if the smart grid elements (power sources) are connected to the IP network using IoT protocols.
Efficient wireless power transmission to remote the sensor in restenosis coro...nooriasukmaningtyas
In this study, the researchers have proposed an alternative technique for designing an asymmetric 4 coil-resonance coupling module based on the series-to-parallel topology at 27 MHz industrial scientific medical (ISM) band to avoid the tissue damage, for the constant monitoring of the in-stent restenosis coronary artery. This design consisted of 2 components, i.e., the external part that included 3 planar coils that were placed outside the body and an internal helical coil (stent) that was implanted into the coronary artery in the human tissue. This technique considered the output power and the transfer efficiency of the overall system, coil geometry like the number of coils per turn, and coil size. The results indicated that this design showed an 82% efficiency in the air if the transmission distance was maintained as 20 mm, which allowed the wireless power supply system to monitor the pressure within the coronary artery when the implanted load resistance was 400 Ω.
Grid reactive voltage regulation and cost optimization for electric vehicle p...nooriasukmaningtyas
Expecting large electric vehicle (EV) usage in the future due to environmental issues, state subsidies, and incentives, the impact of EV charging on the power grid is required to be closely analyzed and studied for power quality, stability, and planning of infrastructure. When a large number of energy storage batteries are connected to the grid as a capacitive load the power factor of the power grid is inevitably reduced, causing power losses and voltage instability. In this work large-scale 18K EV charging model is implemented on IEEE 33 network. Optimization methods are described to search for the location of nodes that are affected most due to EV charging in terms of power losses and voltage instability of the network. Followed by optimized reactive power injection magnitude and time duration of reactive power at the identified nodes. It is shown that power losses are reduced and voltage stability is improved in the grid, which also complements the reduction in EV charging cost. The result will be useful for EV charging stations infrastructure planning, grid stabilization, and reducing EV charging costs.
Topology network effects for double synchronized switch harvesting circuit on...nooriasukmaningtyas
Energy extraction takes place using several different technologies, depending on the type of energy and how it is used. The objective of this paper is to study topology influence for a smart network based on piezoelectric materials using the double synchronized switch harvesting (DSSH). In this work, has been presented network topology for circuit DSSH (DSSH Standard, Independent DSSH, DSSH in parallel, mono DSSH, and DSSH in series). Using simulation-based on a structure with embedded piezoelectric system harvesters, then compare different topology of circuit DSSH for knowledge is how to connect the circuit DSSH together and how to implement accurately this circuit strategy for maximizing the total output power. The network topology DSSH extracted power a technique allows again up to in terms of maximal power output compared with network topology standard extracted at the resonant frequency. The simulation results show that by using the same input parameters the maximum efficiency for topology DSSH in parallel produces 120% more energy than topology DSSH-series. In addition, the energy harvesting by mono-DSSH is more than DSSH-series by 650% and it has exceeded DSSHind by 240%.
Improving the design of super-lift Luo converter using hybrid switching capac...nooriasukmaningtyas
In this article, an improvement to the positive output super-lift Luo converter (POSLC) has been proposed to get high gain at a low duty cycle. Also, reduce the stress on the switch and diodes, reduce the current through the inductors to reduce loss, and increase efficiency. Using a hybrid switch unit composed of four inductors and two capacitors it is replaced by the main inductor in the elementary circuit. It’s charged in parallel with the same input voltage and discharged in series. The output voltage is increased according to the number of components. The gain equation is modeled. The boundary condition between continuous conduction mode (CCM) and discontinuous conduction mode (DCM) has been derived. Passive components are designed to get high output voltage (8 times at D=0.5) and low ripple about (0.004). The circuit is simulated and analyzed using MATLAB/Simulink. Maximum power point tracker (MPPT) controls the converter to provide the most interest from solar energy.
Third harmonic current minimization using third harmonic blocking transformernooriasukmaningtyas
Zero sequence blocking transformers (ZSBTs) are used to suppress third harmonic currents in 3-phase systems. Three-phase systems where singlephase loading is present, there is every chance that the load is not balanced. If there is zero-sequence current due to unequal load current, then the ZSBT will impose high impedance and the supply voltage at the load end will be varied which is not desired. This paper presents Third harmonic blocking transformer (THBT) which suppresses only higher harmonic zero sequences. The constructional features using all windings in single-core and construction using three single-phase transformers explained. The paper discusses the constructional features, full details of circuit usage, design considerations, and simulation results for different supply and load conditions. A comparison of THBT with ZSBT is made with simulation results by considering four different cases
Power quality improvement of distribution systems asymmetry caused by power d...nooriasukmaningtyas
With an increase of non-linear load in today’s electrical power systems, the rate of power quality drops and the voltage source and frequency deteriorate if not properly compensated with an appropriate device. Filters are most common techniques that employed to overcome this problem and improving power quality. In this paper an improved optimization technique of filter applies to the power system is based on a particle swarm optimization with using artificial neural network technique applied to the unified power flow quality conditioner (PSO-ANN UPQC). Design particle swarm optimization and artificial neural network together result in a very high performance of flexible AC transmission lines (FACTs) controller and it implements to the system to compensate all types of power quality disturbances. This technique is very powerful for minimization of total harmonic distortion of source voltages and currents as a limit permitted by IEEE-519. The work creates a power system model in MATLAB/Simulink program to investigate our proposed optimization technique for improving control circuit of filters. The work also has measured all power quality disturbances of the electrical arc furnace of steel factory and suggests this technique of filter to improve the power quality.
Studies enhancement of transient stability by single machine infinite bus sys...nooriasukmaningtyas
Maintaining network synchronization is important to customer service. Low fluctuations cause voltage instability, non-synchronization in the power system or the problems in the electrical system disturbances, harmonics current and voltages inflation and contraction voltage. Proper tunning of the parameters of stabilizer is prime for validation of stabilizer. To overcome instability issues and get reinforcement found a lot of the techniques are developed to overcome instability problems and improve performance of power system. Genetic algorithm was applied to optimize parameters and suppress oscillation. The simulation of the robust composite capacitance system of an infinite single-machine bus was studied using MATLAB was used for optimization purpose. The critical time is an indication of the maximum possible time during which the error can pass in the system to obtain stability through the simulation. The effectiveness improvement has been shown in the system
Renewable energy based dynamic tariff system for domestic load managementnooriasukmaningtyas
To deal with the present power-scenario, this paper proposes a model of an advanced energy management system, which tries to achieve peak clipping, peak to average ratio reduction and cost reduction based on effective utilization of distributed generations. This helps to manage conventional loads based on flexible tariff system. The main contribution of this work is the development of three-part dynamic tariff system on the basis of time of utilizing power, available renewable energy sources (RES) and consumers’ load profile. This incorporates consumers’ choice to suitably select for either consuming power from conventional energy sources and/or renewable energy sources during peak or off-peak hours. To validate the efficiency of the proposed model we have comparatively evaluated the model performance with existing optimization techniques using genetic algorithm and particle swarm optimization. A new optimization technique, hybrid greedy particle swarm optimization has been proposed which is based on the two aforementioned techniques. It is found that the proposed model is superior with the improved tariff scheme when subjected to load management and consumers’ financial benefit. This work leads to maintain a healthy relationship between the utility sectors and the consumers, thereby making the existing grid more reliable, robust, flexible yet cost effective.
Energy harvesting maximization by integration of distributed generation based...nooriasukmaningtyas
The purpose of distributed generation systems (DGS) is to enhance the distribution system (DS) performance to be better known with its benefits in the power sector as installing distributed generation (DG) units into the DS can introduce economic, environmental and technical benefits. Those benefits can be obtained if the DG units' site and size is properly determined. The aim of this paper is studying and reviewing the effect of connecting DG units in the DS on transmission efficiency, reactive power loss and voltage deviation in addition to the economical point of view and considering the interest and inflation rate. Whale optimization algorithm (WOA) is introduced to find the best solution to the distributed generation penetration problem in the DS. The result of WOA is compared with the genetic algorithm (GA), particle swarm optimization (PSO), and grey wolf optimizer (GWO). The proposed solutions methodologies have been tested using MATLAB software on IEEE 33 standard bus system
Intelligent fault diagnosis for power distribution systemcomparative studiesnooriasukmaningtyas
Short circuit is one of the most popular types of permanent fault in power distribution system. Thus, fast and accuracy diagnosis of short circuit failure is very important so that the power system works more effectively. In this paper, a newly enhanced support vector machine (SVM) classifier has been investigated to identify ten short-circuit fault types, including single line-toground faults (XG, YG, ZG), line-to-line faults (XY, XZ, YZ), double lineto-ground faults (XYG, XZG, YZG) and three-line faults (XYZ). The performance of this enhanced SVM model has been improved by using three different versions of particle swarm optimization (PSO), namely: classical PSO (C-PSO), time varying acceleration coefficients PSO (T-PSO) and constriction factor PSO (K-PSO). Further, utilizing pseudo-random binary sequence (PRBS)-based time domain reflectometry (TDR) method allows to obtain a reliable dataset for SVM classifier. The experimental results performed on a two-branch distribution line show the most optimal variant of PSO for short fault diagnosis.
A deep learning approach based on stochastic gradient descent and least absol...nooriasukmaningtyas
More than eighty-five to ninety percentage of the diabetic patients are affected with diabetic retinopathy (DR) which is an eye disorder that leads to blindness. The computational techniques can support to detect the DR by using the retinal images. However, it is hard to measure the DR with the raw retinal image. This paper proposes an effective method for identification of DR from the retinal images. In this research work, initially the Weiner filter is used for preprocessing the raw retinal image. Then the preprocessed image is segmented using fuzzy c-mean technique. Then from the segmented image, the features are extracted using grey level co-occurrence matrix (GLCM). After extracting the fundus image, the feature selection is performed stochastic gradient descent, and least absolute shrinkage and selection operator (LASSO) for accurate identification during the classification process. Then the inception v3-convolutional neural network (IV3-CNN) model is used in the classification process to classify the image as DR image or non-DR image. By applying the proposed method, the classification performance of IV3-CNN model in identifying DR is studied. Using the proposed method, the DR is identified with the accuracy of about 95%, and the processed retinal image is identified as mild DR.
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...Amil Baba Dawood bangali
Contact with Dawood Bhai Just call on +92322-6382012 and we'll help you. We'll solve all your problems within 12 to 24 hours and with 101% guarantee and with astrology systematic. If you want to take any personal or professional advice then also you can call us on +92322-6382012 , ONLINE LOVE PROBLEM & Other all types of Daily Life Problem's.Then CALL or WHATSAPP us on +92322-6382012 and Get all these problems solutions here by Amil Baba DAWOOD BANGALI
#vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore#blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #blackmagicforlove #blackmagicformarriage #aamilbaba #kalajadu #kalailam #taweez #wazifaexpert #jadumantar #vashikaranspecialist #astrologer #palmistry #amliyaat #taweez #manpasandshadi #horoscope #spiritual #lovelife #lovespell #marriagespell#aamilbabainpakistan #amilbabainkarachi #powerfullblackmagicspell #kalajadumantarspecialist #realamilbaba #AmilbabainPakistan #astrologerincanada #astrologerindubai #lovespellsmaster #kalajaduspecialist #lovespellsthatwork #aamilbabainlahore #Amilbabainuk #amilbabainspain #amilbabaindubai #Amilbabainnorway #amilbabainkrachi #amilbabainlahore #amilbabaingujranwalan #amilbabainislamabad
Industrial Training at Shahjalal Fertilizer Company Limited (SFCL)MdTanvirMahtab2
This presentation is about the working procedure of Shahjalal Fertilizer Company Limited (SFCL). A Govt. owned Company of Bangladesh Chemical Industries Corporation under Ministry of Industries.
Immunizing Image Classifiers Against Localized Adversary Attacksgerogepatton
This paper addresses the vulnerability of deep learning models, particularly convolutional neural networks
(CNN)s, to adversarial attacks and presents a proactive training technique designed to counter them. We
introduce a novel volumization algorithm, which transforms 2D images into 3D volumetric representations.
When combined with 3D convolution and deep curriculum learning optimization (CLO), itsignificantly improves
the immunity of models against localized universal attacks by up to 40%. We evaluate our proposed approach
using contemporary CNN architectures and the modified Canadian Institute for Advanced Research (CIFAR-10
and CIFAR-100) and ImageNet Large Scale Visual Recognition Challenge (ILSVRC12) datasets, showcasing
accuracy improvements over previous techniques. The results indicate that the combination of the volumetric
input and curriculum learning holds significant promise for mitigating adversarial attacks without necessitating
adversary training.
Using recycled concrete aggregates (RCA) for pavements is crucial to achieving sustainability. Implementing RCA for new pavement can minimize carbon footprint, conserve natural resources, reduce harmful emissions, and lower life cycle costs. Compared to natural aggregate (NA), RCA pavement has fewer comprehensive studies and sustainability assessments.
Final project report on grocery store management system..pdfKamal Acharya
In today’s fast-changing business environment, it’s extremely important to be able to respond to client needs in the most effective and timely manner. If your customers wish to see your business online and have instant access to your products or services.
Online Grocery Store is an e-commerce website, which retails various grocery products. This project allows viewing various products available enables registered users to purchase desired products instantly using Paytm, UPI payment processor (Instant Pay) and also can place order by using Cash on Delivery (Pay Later) option. This project provides an easy access to Administrators and Managers to view orders placed using Pay Later and Instant Pay options.
In order to develop an e-commerce website, a number of Technologies must be studied and understood. These include multi-tiered architecture, server and client-side scripting techniques, implementation technologies, programming language (such as PHP, HTML, CSS, JavaScript) and MySQL relational databases. This is a project with the objective to develop a basic website where a consumer is provided with a shopping cart website and also to know about the technologies used to develop such a website.
This document will discuss each of the underlying technologies to create and implement an e- commerce website.
Cosmetic shop management system project report.pdfKamal Acharya
Buying new cosmetic products is difficult. It can even be scary for those who have sensitive skin and are prone to skin trouble. The information needed to alleviate this problem is on the back of each product, but it's thought to interpret those ingredient lists unless you have a background in chemistry.
Instead of buying and hoping for the best, we can use data science to help us predict which products may be good fits for us. It includes various function programs to do the above mentioned tasks.
Data file handling has been effectively used in the program.
The automated cosmetic shop management system should deal with the automation of general workflow and administration process of the shop. The main processes of the system focus on customer's request where the system is able to search the most appropriate products and deliver it to the customers. It should help the employees to quickly identify the list of cosmetic product that have reached the minimum quantity and also keep a track of expired date for each cosmetic product. It should help the employees to find the rack number in which the product is placed.It is also Faster and more efficient way.
We have compiled the most important slides from each speaker's presentation. This year’s compilation, available for free, captures the key insights and contributions shared during the DfMAy 2024 conference.
Design and Analysis of Algorithms-DP,Backtracking,Graphs,B&B
Predicting heart failure using a wrapper-based feature selection
1. Indonesian Journal of Electrical Engineering and Computer Science
Vol. 21, No. 3, March 2021, pp. 1530~1539
ISSN: 2502-4752, DOI: 10.11591/ijeecs.v21.i3.pp1530-1539 1530
Journal homepage: http://ijeecs.iaescore.com
Predicting heart failure using a wrapper-based feature selection
Minh Tuan Le1
, Minh Thanh Vo2
, Nhat Tan Pham3
, Son V.T Dao4
1,2
SEE, International University, Vietnam National University Ho Chi Minh City, Ho Chi Minh City, Viet Nam
3,4
SIEM, VNU-International University, Vietnam National University Ho Chi Minh City, Ho Chi Minh City, Viet Nam
Article Info ABSTRACT
Article history:
Received Oct 30, 2020
Revised Dec 7, 2020
Accepted Dec 23, 2020
In the current health system, it is very difficult for medical practitioners/
physicians to diagnose the effectiveness of heart contraction. In this research,
we proposed a machine learning model to predict heart contraction using an
artificial neural network (ANN). We also proposed a novel wrapper-based
feature selection utilizing a grey wolf optimization (GWO) to reduce the
number of required input attributes. In this work, we compared the results
achieved using our method and several conventional machine learning
algorithms approaches such as support vector machine, decision tree, K-
nearest neighbor, naïve bayes, random forest, and logistic regression.
Computational results show not only that much fewer features are needed,
but also higher prediction accuracy can be achieved around 87%. This work
has the potential to be applicable to clinical practice and become a supporting
tool for doctors/physicians.
Keywords:
Feature selection
Grey wolf optimization
Heart failure
Multilayer perceptron
Neural network
This is an open access article under the CC BY-SA license.
Corresponding Author:
Son V.T Dao
School of Industrial and Management
International University
Vietnam National University HCMC
Ho Chi Minh City, Viet Nam
Email: dvtson@hcmiu.edu.vn
1. INTRODUCTION
The term "heart failure" referring to a condition in which the heart's contraction is not as effective as
it should be. The heart is a vital organ in the human body because it pumps blood to every other organ. A
patient who is living vegetative states still needs the heart to survive. Heart failure (HF) is a chronic condition
in which one of the ventricles or atriums on both sides is not able to pump rich oxygen into the body and poor
oxygen into the lungs. There are several common reasons cause heart failure. The majority of (HF) patients
are elderly. Cardiac arrest often gradually and deliberately develops after parts of a heart got weaken and
makes others such as ventricles and atriums do extra workloads to provide enough blood and oxygen to the
body [1-2]. With the ubiquitous application of technology in the medical field, it helps the cost of diagnosis
to be inexpensive. Unfortunately, nowadays the number of patients who have been diagnoses with heart
failure is gradually increasing in suburbs and dramatically increasing in urban areas. Therefore, the earlier for
getting diagnosed, the better off it will be for the patients. Because of the difficulty of diagnosing the process
of a heart failure condition, it might cause a postponement in treatment operation. Therefore, it is crucial to
develop a heart disease prediction system for heart failure to support whoever works in the medical
professional field to diagnose patients with conditions more rapid and accurate. Deep learning and Machine
learning algorithms have been successfully applied to various field [3-4], especially medical field to support
doctor/physician to diagnose various diseases such as heart failure, diabetes. ANN has also been applied by
researchers in the medical field [5-7].
2. Indonesian J Elec Eng & Comp Sci ISSN: 2502-4752
Predicting heart failure using a wrapper-based feature selection (Minh Tuan Le)
1531
In this study, we will use a multilayer perceptron models (MLP) together with preprocessing
methods for predicting heart failure patient. Besides, a metaheuristics-based feature selection algorithm grey
wolf optimization (GWO) also applied to MLP models to enhance the performance and reduce training time.
The result is a benchmark with other common machine learning/deep learning algorithms in the following
such as logistic regression (LR) [8], support vector machine (SVM) [9], k-nearest neighbor classifier (KNN)
[10], naive bayesian classifier (NBC) [11-12], decision tree (DT) [13], and random forest classifier (RFC)
[14] based on the original set of available medical features.
Several data mining methods have been successfully applied to diagnosing heart failure (HF). In
[10], Davide Chicco represents a model to diagnostic the survival rate of patients who have been using
clinical record HF data. In the research [15], a list of the machine learning methods was used for the binary
classification of survival. A random forest classifier outperformed all other methods compared to the other
models, which is very impressive in this age [16]. Guidi et al. (2013) represent a clinical decision support
system (CDSS) for analyzing HF patients and comparing the performance of neural network, support vector
machine, fuzzy-genetic, random forest. In [17], the authors used the detailed clinical data on patients
hospitalized with HF in Ontario, Canada. In the machine learning literature, alternate classification schemes
have been developed such as bootstrap aggregation (bagging), boosting, random forests, and support vector
machines. They also compared the ability of these methods to predict the probability of the presence of heart
failure with preserved ejection fraction (HFPEF).
Feature selection (FS) is a process that commonly selects in machine learning to solve the high
dimensionality problem. In FS, we choose a small number of features but important and usually ignore the
irrelevant and noisy features, in order to make the subsequent analysis easier. According to the redundancy
and relevance. Yu et al., [18] have classified those feature subsets into four different types: noisy and
irrelevant; redundant and weakly relevant, weakly relevant and non-redundant, and strongly relevant. An
irrelevant feature does not require predicting accuracy. Furthermore, many approaches can implement with
filter and wrapper methods such as models, search strategies, feature quality measures, and feature
evaluation. All features play as key factors for determining the hypothesis of the predicting models. Besides
that, the number of features and the size of the hypothesis spaces are directly proportional to each other, and
so on. When the number of features increases, the size of the searching space also increased. One such
outstanding case is that if there are M features with the binary class label in a dataset, it has combination
in the search space.
There are three types of FS methods, which are defined based on the interaction with the learning
model, namely filter, wrapper, and embedded methods. The Filter method selects statistics-based features. It
is independent of the learning algorithm and thus requires less computational time. Statistical measures such
as information gain, chi-square test [19], Fisher score, correlation coefficient, and variance threshold are used
to understand the importance of the features. In contrast, the wrapper method’s performance highly depends
on the classifier. The best subset of features is selected based on the results of the classifier. Wrapper
methods are much more computationally expensive than filter methods since it needs to run simultaneously
with the classifier many times. However, these methods are more accurate than the filter method. Some of the
wrapper examples are recursive feature elimination [20], Sequential feature selection algorithms [21], and
genetic algorithms [22]. Thirdly, the embedded method which utilizes ensemble learning and hybrid learning
methods for feature selection. This method has a collective decision; therefore, its performance is better than
the previous one. One example is the random forest which is less computationally intensive than wrapper
methods. One drawback of the embedded method is that it is specific to a learning model.
Many evolutionary metaheuristics-based feature selection methods are also proposed, many of them
are wrapper type since it has been proven that wrapper provides better performance [23]. Too et al., [24]
proposed a competitive binary grey wolf optimizer (CBGWO), which is based on the grey wolf optimizer
(GWO) proposed by Mirjalili et al. [25], for feature selection problem in EMG signal classification. The
results showed that CBGWO outranked other algorithms in terms of performance for that case study. Many
other wrapper-based feature selection algorithms were also introduced in many previous works to select a
subset of features, including binary grey wolf optimization (BGWO) [26], binary particle swarm optimization
(BPSO) [27], ant colony optimization (ACO) [28], and binary differential evolution (BDE) [29].
2. RESEARCH METHOD
This system includes two steps: data pre-processing which involved outlier detection. Then followed
by a multilayer perceptron (MLP). The outlier detection in this research is using interquartile range (IQR)
method and then applying grey wolf optimizer to optimize the architecture of multilayer perceptron for
classifying the heart failure patients. The detail of this method is described in this section.
3. ISSN: 2502-4752
Indonesian J Elec Eng & Comp Sci, Vol. 21, No. 3, March 2021 : 1530 - 1539
1532
2.1. Data Collection
Heart failure clinical records data set has been used in this research, which is records heart failure
patients from the Faisalabad Institute of Cardiology and the Allied Hospital in Faisalabad (Punjab, Pakistan).
The dataset is available on the UCI repository. The target of this binary classification has two categorial
value: 1-yes (patient is sick) and 0-no (patient is healthy). The attribute for predicting is “DEATH_EVENT”
which contains two categorical values and is considered as a binary classification problem. Table 1 list the
number of instance number of attribute and features of the dataset.
The dataset contains 299 instances and 12 attributes. Each of these attributes is physiological
measurements. The patients in this dataset include 194 men and 105 women and the range of their ages
between 40 and 95 years old. Features, measurements, and range are listed in Table 1.
Table 1. Features, measurement, meaning, and range of the dataset
Feature Explanation Measurement Range
Age Age of the patient years [40, …,95]
Anaemia Decrease of red blood cells or hemoglobin Boolean 0,1
High blood pressure If the patient has hypertension Boolean 0,1
Creatinine phosphokinase
(CPK)
Level of the CPK enzyme in the blood mcg/L [23, ....,7861]
Diabetes If the patient has diabetes Boolean 0, 1
Ejection fraction Percentage of blood leaving the heart at each
contraction
% [14, ...,80]
Sex Platelets in the blood binary 0, 1
Platelets woman or man kiloplatelets/mL [25.01, ...,850.00]
Serum creatinine Level of serum creatinine in the blood mg/dL [0.50, …9.40]
Serum sodium Level of serum sodium in the blood mEq/L [114, ...,148]
Smoking If the patient smokes or not Boolean 0, 1
Time Follow-up period days [4, ...,285]
[target] death event If the patient deceased during the follow-up period Boolean 0, 1
2.2. Data Preprocessing
For preprocessing data, we use the normalization and IQR method of outlier detection. Before
training data, we normalize this dataset since the gradient descent will be effective with normalized (scaled)
values. We may get the values in the different scales if we would not normalize the data. To adjust weights,
our model would take more time to train on this data. However, if we normalize our data by using
normalization techniques, we will have numbers on the same scale which will make our model train much
faster and gradient descent will be effective in this case. IQR Method of outlier detection, which is used for
pre-processing data IQR (short for “interquartile range”) in the middle spread, is also known as the quartile
range of the dataset. This concept is used in statistical analysis to help conclude a set of numbers. IQR is used
for the range of variation because it excludes most outliers of data.
In Figure 1, the minimum, maximum is the minimum and maximum value in the dataset. The
median is also called the second quartile of the data. Q1 is the first quartile of the data, it means that 25% of
the data is lies between minimum and Q1. And Q3 is the third quartile of the data, it says that 75% of the data
lies between maximum and Q3. The equation below is the Inter-Quartile Range or IQR, which is the
difference between Q3 and Q1.
Figure 1. A box-whisker plot
(1)
For detecting the outliers with this technique, we have to define a new range, which is called a
decision range. If any point lies outside this range, it is considered an outlier. This range is (2), (3):
4. Indonesian J Elec Eng & Comp Sci ISSN: 2502-4752
Predicting heart failure using a wrapper-based feature selection (Minh Tuan Le)
1533
(2)
(3)
2.3. Research Methodology
2.3.1. Grey wolf optimizer (GWO)
Swarm intelligence is the way of communication between an individual and a group. The
application of herd intelligence in the fields of industry, science, and commerce has many unique and diverse
applications. Research in herd intelligence can help people manage complex systems. GWO simulates the
way that the wolves look for food and survive by avoiding their enemies (Figure 2). GWO was firstly
introduced by Mirjalili et al., 2014 [25]. Alpha means that the leader gives the decision for a sleeping place,
hunting grey, time to wake up. The second level of gray wolves is beta. The betas are the wolves in herds
under alpha but also commanded another low-level wolf. The lowest rank among the gray wolves is Omega.
They are weak wolves and have to rely on other wolves in the pack. Delta ones are dependent on alphas and
betas, but they are more effective than omega. They are responsible for monitoring territorial boundaries and
warning inside in case of danger, protect and ensure safety for herds, take care of the weak, and illness
wolves in the pack.
Figure 2. Position updating in GWO
To develop the mathematical model, the best solution is considered as alpha. Beta and delta are the
second and the third solution, respectively. The step of GWO is encircling prey is shown (4), (5):
) (4)
(5)
where t shows the current iteration, and are coefficient vectors, is the position vector of a grey
wolf and is the position vector of the prey. The coefficient is indicated in the (6), (7):
) (6)
(7)
where is linearly decreased from 2 to 0, and are random vector in [0, 1].
These equations below define the final position of the wolf :
5. ISSN: 2502-4752
Indonesian J Elec Eng & Comp Sci, Vol. 21, No. 3, March 2021 : 1530 - 1539
1534
(8)
(9)
(10)
(11)
(12)
(13)
(14)
2.3.2. Multilayer perceptron (MLP)
The single-layer perceptron solves only linearly separable problems, but some complex problems
are not linearly separable. Therefore, in order to solve some complex problems, one or more layers are added
in a single layer perceptron, so it is known as a multilayer perceptron (MLP) [30-33]. The MLP network is
also known as a feed-forward neural network having one or more hidden layers as can be seen in Figure 3.
Figure 3. The Architecture of a multilayer perceptron
In Figure 3, the neural network has an input layer with n neurons, one hidden layer with n neurons,
and an output layer
Input layer: call input variable (x1, …, xn), also called the visible layer
Hidden layer: the layer of the node lies between the input and output layer.
Output layer: this layer produces the output variables
The following steps below show the calculation of the MLP output after giving the weights, inputs,
and biases:
The weighted sums of inputs are calculated as follow:
(15)
where shows the th input, represent the number of nodes, is the connection weight from
the th node to the th node and is the threshold of the hidden node.
The calculation of the output of each hidden node:
(16)
6. Indonesian J Elec Eng & Comp Sci ISSN: 2502-4752
Predicting heart failure using a wrapper-based feature selection (Minh Tuan Le)
1535
The final outputs are based on the calculation of the output of hidden nodes:
(17)
(18)
where is the connection weight from to and is the threshold of the output node.
For the definition of the final output, the weights and biases are used. We find the values for weights
and biases to achieve a relationship between the inputs and outputs. In this algorithm, weights and biases
have been adjusted repeatedly for minimizing the actual output vector of the network and output vector.
3. RESULTS AND ANALYSIS
3.1. Performance Evaluation
In this system, the performance of these algorithms is studied based on performance metrics such as
accuracy, precision, recall, F1 score, which are given in these (19-22):
(19)
(20)
(21)
(22)
Where: a true positive (TP): the samples are classified as true (T) while they are (T); a true negative (TN): the
samples are classified as false (F) while they are (F); a false positive (FP): the samples are classified as (T)
while they are (F); A false negative (FN): the samples are classified as (F) while they are (T).
3.2. Analysis Results Using Various Machine Learning Models
In this research, six classifier models LR, KNN, SVM, NB, DT, RFC are used. The dataset is
divided into 80:20, which is 80% of data for training the models and 20% is used for testing the accuracy of
the models. In this research, we apply the removal of the outlier dataset for training.
The bar chart in the Figure 4 indicates the accuracy of machine learning algorithms. As can be seen
from the figure that the random forest classifier is having the highest accuracy with 85% compared to the
other algorithms. LR also achieves good accuracy as compared to SVM.
Figure 4. Classification accuracy with different models
Table 2 indicates the classification results of the six models based on accuracy, recall, precision, F1
score on the heart failure dataset. It can be seen on the table that, among the six methods, the RFC method
performs the best performance, with the highest accuracy 85%, 65.21% recall, 93.75% precision, and 76% F1
score than the other machine learning models.
7. ISSN: 2502-4752
Indonesian J Elec Eng & Comp Sci, Vol. 21, No. 3, March 2021 : 1530 - 1539
1536
Table 2. The prediction results
ID Method Accuracy Recall Precision F1 score
1 LR 0.80 0.4782 1 0.6470
2 KNN 0.75 0.4782 0.7857 0.5945
3 SVM 0.80 0.4782 1 0.6470
4 NBC 0.6833 0.3043 0.7 0.4242
5 DT 0.7667 0.6086 0.7368 0.6666
6 RFC 0.85 0.6521 0.9375 0.7692
Figure 5 shows the confusion matrix of all algorithms. The diagonal elements of the matrix are the
correctly classified number of points for each data layer. From here, the accuracy can be inferred by the sum
of the elements on the diagonal divided by the sum of the elements of the entire matrix. A good model will
give a confusion matrix with the elements on the main diagonal having a big value and the remaining
elements having a small value. It can be seen from the confusion matrix that the main diagonal elements of
the matrix of LR, SVM, and RFC have a higher value.
Figure 5. Confusion matrix of all algorithms
The receiver operating characteristic (ROC) plot is a measurement for evaluating the classifier
performance of each algorithm, as shown in Figure 6.
Figure 6. Receiver operating characteristic (ROC) of all algorithm
8. Indonesian J Elec Eng & Comp Sci ISSN: 2502-4752
Predicting heart failure using a wrapper-based feature selection (Minh Tuan Le)
1537
3.3. Experimental Results of GWO-MLP
In this work, we show the experimental result when grey wolf optimization is applied to multilayer
perceptron (GWO-MLP). The Table 3 shows the samples of the selected feature using grey wolf optimization
and Figure 7 shows Convergence curve of the fitness function.
Table 3. Samples of feature selection
ID Feature Feature selection
1 age Selected
2 anaemia Selected
3 creatinine_phosphokinase x
4 diabetes x
5 ejection_fraction Selected
6 high_blood_pressure Selected
7 platelets x
8 serum_creatinine x
9 serum_sodium x
10 sex Selected
11 smoking x
12 time Selected
Figure 7. Convergence curve of the fitness function
We can see that 6 of the 12 features are selected. After that, this subset of features is trained on the
MLP with 100 epochs, which yields the following result in Figure 8. In this approach, the performance of
GWO-MLP achieved 87% of accuracy, 74% recall, 77% precision, and 76% F1 score. With this approach,
only six out of the original twelve features were selected, the accuracy of the GWO-MLP model was higher
than the other methods, indicating there were unuseful features in the data. Furthermore, the training time of
this data is lower than the other models.
(a) (b)
Figure 8. (a) Training and Validation accuracy result, (b) Training and Validation loss result
4. CONCLUSION
In this paper, we propose a machine learning model to predict heart failure using an ANN. At first, a
wrapper-based feature selection approach using a metaheuristic called GWO to select 6 features out of the
original 12 features. These features are used as inputs for the MLP for the prediction task. Our proposed
results achieve an accuracy of 87%, which shows that our approach outperformed other machine learning
models such as SVM, LR, KNN, NBC, DT, RFC. Furthermore, with fewer features, our machine learning
model is much simpler and requires much less computational effort. Potential future works are listed as
follows: fine-tuning the MLP architecture, i.e the number of hidden layers and hidden nodes, as well as the
activation functions; or optimizing the parameters of the feature selection algorithm for achieving a better
performance.
9. ISSN: 2502-4752
Indonesian J Elec Eng & Comp Sci, Vol. 21, No. 3, March 2021 : 1530 - 1539
1538
REFERENCES
[1] T. Ee, P. Tg, K. Gs, N. Kk, and F. Di, “Heart Failure: Diagnosis, Severity Estimation and Prediction of Adverse
Events Through Machine Learning Techniques.,” Comput Struct Biotechnol J, vol. 15, pp. 26–47, Nov. 2016, doi:
10.1016/j.csbj.2016.11.001.
[2] L. Hussain, I. A. Awan, W. Aziz, S. Saeed, and A. Ali, “Detecting Congestive Heart Failure by Extracting
Multimodal Features and Employing Machine Learning Techniques,” BioMed Research International, vol. 2020,
pp. 1–19, 2020, doi: https://doi.org/10.1155/2020/4281243.
[3] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, Art. no. 7553, May 2015, doi:
10.1038/nature14539.
[4] M. Rungruanganukul and T. Siriborvornratanakul, “Deep Learning Based Gesture Classification for Hand Physical
Therapy Interactive Program,” in Digital Human Modeling and Applications in Health, Safety, Ergonomics and
Risk Management. Posture, Motion and Health, Jul. 2020, pp. 349–358, doi: 10.1007/978-3-030-49904-4_26.
[5] C. S. Dangare and S. S. Apte, “A Data Mining Approach for Prediction of Heart Disease Using Neural Networks”,
International Journal of Computer Engineering and Technology(IJCET), vol. 3, no. 3, pp. 30–40, Oct. 2012.
[6] S. Smiley, “Diagnostic for Heart Disease with Machine Learning,” Medium, Jan. 12, 2020.
https://towardsdatascience.com/diagnostic-for-heart-disease-with-machine-learning-81b064a3c1dd (accessed Sep.
19, 2020).
[7] D. Shen, G. Wu, and H.-I. Suk, “Deep Learning in Medical Image Analysis,” Annual Review of Biomedical
Engineering, vol. 19, no. 1, pp. 221–248, 2017, doi: 10.1146/annurev-bioeng-071516-044442.
[8] R. E. Wright, “Logistic regression,” in Reading and understanding multivariate statistics, Washington, DC, US:
American Psychological Association, 1995, pp. 217–244.
[9] Vladimir N. Vapnik, "Statistical Learning Theory", Canada, A Wiley-Interscience Publication, Sep. 1998
[10] N. S. Altman, “An Introduction to Kernel and Nearest-Neighbor Nonparametric Regression: The American
Statistician" The American Statistician: Vol 46, No 3, pp. 175-185, Aug. 1992, [Online]. Available:
https://doi.org/10.2307/2685209.
[11] K. M. Ting and Z. Zheng, “Improving the Performance of Boosting for Naive Bayesian Classification,” in
Methodologies for Knowledge Discovery and Data Mining, Berlin, Heidelberg, 1999, pp. 296–305, [Online].
Available: https://doi.org/10.1007/3-540-48912-6_41.
[12] C. Kerdvibulvech, “Human Hand Motion Recognition Using an Extended Particle Filter,” in Articulated Motion
and Deformable Objects, Cham, 2014, pp. 71–80, doi: 10.1007/978-3-319-08849-5_8.
[13] J. R. Quinlan, “Induction of decision trees” Mach Learn 1, vol. 1, no. 1, pp. 81–106, Mar. 1986, [Online].
Available: https://doi.org/10.1007/BF00116251.
[14] L. Breiman, “Random Forests,” Machine Learning, vol. 45, no. 1, pp. 5–32, Oct. 2001, doi:
10.1023/A:1010933404324
[15] S. Angraal et al., “Machine Learning Prediction of Mortality and Hospitalization in Heart Failure with Preserved
Ejection Fraction,” JACC: Heart Failure, vol. 8, no. 1, pp. 12–21, Jan. 2020, [Online]. Available:
https://doi.org/10.1016/j.jchf.2019.06.013.
[16] D. Chicco and G. Jurman, “Machine learning can predict survival of patients with heart failure from serum
creatinine and ejection fraction alone,” BMC Medical Informatics and Decision Making, vol. 20, no. 1, p. 16, Feb.
2020, [Online]. Available: https://doi.org/10.1186/s12911-020-1023-5.
[17] P. C. Austin, J. V. Tu, J. E. Ho, D. Levy, and D. S. Lee, “Using methods from the data-mining and machine-
learning literature for disease classification and prediction: a case study examining classification of heart failure
subtypes,” Journal of Clinical Epidemiology, vol. 66, no. 4, pp. 398–407, Apr. 2013, doi:
10.1016/j.jclinepi.2012.11.008.
[18] L. Yu and H. Liu, “Efficient Feature Selection via Analysis of Relevance and Redundancy,” The Journal of
Machine Learning Research, vol. 5, pp. 1205–1224, Dec. 2004.
[19] Y. Yang and J. Pedersen, “A Comparative Study on Feature Selection in Text Categorization,” 1997.
[20] K. Yan and D. Zhang, “Feature selection and analysis on correlated gas sensor data with recursive feature
elimination,” Sensors and Actuators B: Chemical, vol. 212, pp. 353–363, Jun. 2015, [Online]. Available:
https://doi.org/10.1016/j.snb.2015.02.025.
[21] A. Jain and D. Zongker, “Feature selection: evaluation, application, and small sample performance,” IEEE
Transactions on Pattern Analysis and Machine Intelligence, vol. 19, no. 2, pp. 153–158, Feb. 1997, [Online].
Available: https://doi.org/10.1109/34.574797.
[22] C. De Stefano, F. Fontanella, C. Marrocco, and A. Scotto di Freca, “A GA-based feature selection approach with an
application to handwritten character recognition,” Pattern Recognition Letters, vol. 35, pp. 130–141, Jan. 2014,
[Online]. Available: https://doi.org/10.1016/j.patrec.2013.01.026.
[23] E. Zorarpacı and S. A. Özel, “A hybrid approach of differential evolution and artificial bee colony for feature
selection,” Expert Systems with Applications, vol. 62, pp. 91–103, Nov. 2016, [Online]. Available:
https://doi.org/10.1016/j.eswa.2016.06.004.
[24] J. Too, A. R. Abdullah, N. Mohd Saad, N. Mohd Ali, and W. Tee, “A New Competitive Binary Grey Wolf
Optimizer to Solve the Feature Selection Problem in EMG Signals Classification,” Computers, vol. 7, no. 4, Art.
no. 4, Dec. 2018, [Online]. Available: https://doi.org/10.3390/computers7040058.
[25] S. Mirjalili, S. M. Mirjalili, and A. Lewis, “Grey Wolf Optimizer,” Advances in Engineering Software, vol. 69, pp.
46–61, Mar. 2014, [Online]. Available: https://doi.org/10.1016/j.advengsoft.2013.12.007.
10. Indonesian J Elec Eng & Comp Sci ISSN: 2502-4752
Predicting heart failure using a wrapper-based feature selection (Minh Tuan Le)
1539
[26] E. Emary, Hossam M. Zawbaa, “Binary Grey Wolf Optimization Approaches for Feature Selection”
Neurocomputing, vol. 172, pp. 371-381, Jan. 2016, [Online]. Available:
https://doi.org/10.1016/j.neucom.2015.06.083.
[27] L.-Y. Chuang, H.-W. Chang, C.-J. Tu, and C.-H. Yang, “Improved binary PSO for feature selection using gene
expression data” Computational Biology and Chemistry, vol. 32, no. 1, pp. 29–38, Feb. 2008, [Online]. Available:
https://doi.org/10.1016/j.compbiolchem.2007.09.005.
[28] M. H. Aghdam, N. G. Aghaee, M. EhsanBasiri “Text feature selection using ant colony optimization" Expert
systems with applications, vol. 36, pp. 6843-6853, April. 2009, [Online]. Available:
https://doi.org/10.1016/j.eswa.2008.08.022.
[29] X. He, Q. Zhang, N. Sun and Y. Dong, "Feature Selection with Discrete Binary Differential Evolution," 2009
International Conference on Artificial Intelligence and Computational Intelligence, pp. 327-330, 2009, [Online].
Available: https://doi.org/10.1109/AICI.2009.438.
[30] M. Jahangir, H. Afzal, M. Ahmed, K. Khurshid and R. Nawaz, "An expert system for diabetes prediction using auto
tuned multi-layer perceptron," 2017 Intelligent Systems Conference (IntelliSys), pp. 722-728, 2017, [Online].
Available: https://doi.org/10.1109/IntelliSys.2017.8324209.
[31] Jahangir, M., Afzal, H., Ahmed, M. et al, "Auto-MeDiSine: an auto-tunable medical decision support engine using
an automated class outlier detection method and AutoMLP," Neural Computing and Applications, pp. 2621–2633
(2020), [Online]. Available: https://doi.org/10.1007/s00521-019-04137-5.
[32] Chaudhuri, B. B., and U. Bhattacharya. “Efficient Training and Improved Performance of Multilayer Perceptron in
Pattern Classification.” Neurocomputing, 34, no. 1 (September 1, 2000): 11–27. https://doi.org/10.1016/S0925-
2312(00)00305-2.
[33] Orhan, Umut, Mahmut Hekim, and Mahmut Ozer. “EEG Signals Classification Using the K-Means Clustering and
a Multilayer Perceptron Neural Network Model.” Expert Systems with Applications, 38, no. 10 (September 15,
2011): 13475–81. https://doi.org/10.1016/j.eswa.2011.04.149.