In healthcare sector, data are enormous and diverse because it contains a data of different types and getting knowledge from these data is crucial. So to get that knowledge, data mining techniques may be utilized to mine knowledge by building models from healthcare dataset. At present, the classification of heart diseases patients has been a demanding research confront for many researchers. For building a classification model for a these patient, we used four different classification algorithms such as NaiveBayes, MultilayerPerceptron, RandomForest and DecisionTable. The intention behind this work is to classify that whether a patient is tested positive or tested negative for heart diseases, based on some diagnostic measurements integrated into the dataset.
A Heart Disease Prediction Model using Logistic Regressionijtsrd
The early prognosis of cardiovascular diseases can aid in making decisions to lifestyle changes in high risk patients and in turn reduce their complications. Research has attempted to pinpoint the most influential factors of heart disease as well as accurately predict the overall risk using homogenous data mining techniques. Recent research has delved into amalgamating these techniques using approaches such as hybrid data mining algorithms. This paper proposes a rule based model to compare the accuracies of applying rules to the individual results of logistic regression on the Cleveland Heart Disease Database in order to present an accurate model of predicting heart disease. K. Sandhya Rani | M. Sai Manoj | G. Suguna Mani"A Heart Disease Prediction Model using Logistic Regression" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11401.pdf http://www.ijtsrd.com/computer-science/data-miining/11401/a-heart-disease-prediction-model-using-logistic-regression/k-sandhya-rani
A Heart Disease Prediction Model using Logistic Regression By Cleveland DataBaseijtsrd
The early prognosis of cardiovascular diseases can aid in making decisions to lifestyle changes in high risk patients and in turn reduce their complications. Research has attempted to pinpoint the most influential factors of heart disease as well as accurately predict the overall risk using homogenous data mining techniques. Recent research has delved into amalgamating these techniques using approaches such as hybrid data mining algorithms. This paper proposes a rule based model to compare the accuracies of applying rules to the individual results of logistic regression on the Cleveland Heart Disease Database in order to present an accurate model of predicting heart disease. K. Sandhya Rani | M. Sai Chaitanya | G. Sai Kiran"A Heart Disease Prediction Model using Logistic Regression By Cleveland DataBase" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11402.pdf http://www.ijtsrd.com/computer-science/data-miining/11402/a-heart-disease-prediction-model-using-logistic-regression-by-cleveland-database/k-sandhya-rani
Heart Disease Prediction Using Data Mining TechniquesIJRES Journal
There are huge amounts of data in the medical industry which is not processed properly and hence cannot be used effectively in making decisions. We can use data mining techniques to mine these patterns and relationships. This research has developed a prototype Heart Disease Prediction using data mining techniques, namely Neural Network, K-Means Clustering and Frequent Item Set Generation. Using medical profiles such as age, sex, blood pressure and blood sugar it can predict the likelihood patients getting a heart disease. It enables significant knowledge, e.g. patterns, relationships between medical factors related to heart disease to be established. Performance of these techniques is compared through sensitivity, specificity and accuracy. It has been observed that Artificial Neural Networks outperform K Means clustering in all the parameters i.e. Sensitivity, Specificity and Accuracy.
A Heart Disease Prediction Model using Logistic Regressionijtsrd
The early prognosis of cardiovascular diseases can aid in making decisions to lifestyle changes in high risk patients and in turn reduce their complications. Research has attempted to pinpoint the most influential factors of heart disease as well as accurately predict the overall risk using homogenous data mining techniques. Recent research has delved into amalgamating these techniques using approaches such as hybrid data mining algorithms. This paper proposes a rule based model to compare the accuracies of applying rules to the individual results of logistic regression on the Cleveland Heart Disease Database in order to present an accurate model of predicting heart disease. K. Sandhya Rani | M. Sai Manoj | G. Suguna Mani"A Heart Disease Prediction Model using Logistic Regression" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11401.pdf http://www.ijtsrd.com/computer-science/data-miining/11401/a-heart-disease-prediction-model-using-logistic-regression/k-sandhya-rani
A Heart Disease Prediction Model using Logistic Regression By Cleveland DataBaseijtsrd
The early prognosis of cardiovascular diseases can aid in making decisions to lifestyle changes in high risk patients and in turn reduce their complications. Research has attempted to pinpoint the most influential factors of heart disease as well as accurately predict the overall risk using homogenous data mining techniques. Recent research has delved into amalgamating these techniques using approaches such as hybrid data mining algorithms. This paper proposes a rule based model to compare the accuracies of applying rules to the individual results of logistic regression on the Cleveland Heart Disease Database in order to present an accurate model of predicting heart disease. K. Sandhya Rani | M. Sai Chaitanya | G. Sai Kiran"A Heart Disease Prediction Model using Logistic Regression By Cleveland DataBase" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-2 | Issue-3 , April 2018, URL: http://www.ijtsrd.com/papers/ijtsrd11402.pdf http://www.ijtsrd.com/computer-science/data-miining/11402/a-heart-disease-prediction-model-using-logistic-regression-by-cleveland-database/k-sandhya-rani
Heart Disease Prediction Using Data Mining TechniquesIJRES Journal
There are huge amounts of data in the medical industry which is not processed properly and hence cannot be used effectively in making decisions. We can use data mining techniques to mine these patterns and relationships. This research has developed a prototype Heart Disease Prediction using data mining techniques, namely Neural Network, K-Means Clustering and Frequent Item Set Generation. Using medical profiles such as age, sex, blood pressure and blood sugar it can predict the likelihood patients getting a heart disease. It enables significant knowledge, e.g. patterns, relationships between medical factors related to heart disease to be established. Performance of these techniques is compared through sensitivity, specificity and accuracy. It has been observed that Artificial Neural Networks outperform K Means clustering in all the parameters i.e. Sensitivity, Specificity and Accuracy.
Heart Disease Prediction using Machine Learning Algorithmijtsrd
Nowadays, Heart disease has become dangerous to a human being, it effects very badly to human body. If anyone is suffering from heart disease, then it leads to blood clotting. Heart disease prediction is very difficult task to predict in the field of medical science. Affiliation has predicted that 12 million people fail horrendously every year as a result of heart disease. In this paper, we propose a k Nearest Neighbors Algorithm KNN way to deal with improve the exactness of heart determination. We show that k Nearest Neighbors Algorithm KNN have better accuracy than random forest algorithm for viewing heart disease. The k Nearest Neighbors Algorithm give more precise and exact outcome . We have taken 13 attributes in the dataset and a target attribute, by applying machine learning we achieved 84 accuracy in the heart disease detection. Ravi Kumar Singh | Dr. A Rengarajan "Heart Disease Prediction using Machine Learning Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38358.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38358/heart-disease-prediction-using-machine-learning-algorithm/ravi-kumar-singh
Prediction of Heart Disease using Machine Learning Algorithms: A Surveyrahulmonikasharma
According to recent survey by WHO organisation 17.5 million people dead each year. It will increase to 75 million in the year 2030[1].Medical professionals working in the field of heart disease have their own limitation, they can predict chance of heart attack up to 67% accuracy[2], with the current epidemic scenario doctors need a support system for more accurate prediction of heart disease. Machine learning algorithm and deep learning opens new door opportunities for precise predication of heart attack. Paper provideslot information about state of art methods in Machine learning and deep learning. An analytical comparison has been provided to help new researches’ working in this field.
This describes the techniques that are used for prediction of heart diseases using the concept of data mining.It states about IHPDS(Intelligent Heart Disease Prediction System)
One of the major purposes manufacturers incorporate AI or ML in their applications is to ease software computations and to predict precise results. I think compared to any other application, a medical application requires a lot of precise computations and therefore, AI is a perfect solution to enhance performance and productivity. While reading the health-tech news, I came across recent research in this regard, the use of AI in predicting a potential stroke or cardiac arrest. ..
A Survey on Heart Disease Prediction Techniquesijtsrd
Heart disease is the main reason for a huge number of deaths in the world over the last few decades and has evolved as the most life threatening disease. The health care industry is found to be rich in information. So, there is a need to discover hidden patterns and trends in them. For this purpose, data mining techniques can be applied to extract the knowledge from the large sets of data. Many researchers, in recent times have been using several machine learning techniques for predicting the heart related diseases as it can predict the disease effectively. Even though a machine learning technique proves to be effective in assisting the decision makers, still there is a scope for developing an accurate and efficient system to diagnose and predict the heart diseases thereby helping doctors with ease of work. This paper presents a survey of various techniques used for predicting heart disease and reviews their performance. G. Niranjana | Dr I. Elizabeth Shanthi "A Survey on Heart Disease Prediction Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38349.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38349/a-survey-on-heart-disease-prediction-techniques/g-niranjana
Multivariate sample similarity measure for feature selection with a resemblan...IJECEIAES
Feature selection improves the classification performance of machine learning models. It also identifies the important features and eliminates those with little significance. Furthermore, feature selection reduces the dimensionality of training and testing data points. This study proposes a feature selection method that uses a multivariate sample similarity measure. The method selects features with significant contributions using a machine-learning model. The multivariate sample similarity measure is evaluated using the University of California, Irvine heart disease dataset and compared with existing feature selection methods. The multivariate sample similarity measure is evaluated with metrics such as minimum subset selected, accuracy, F1-score, and area under the curve (AUC). The results show that the proposed method is able to diagnose chest pain, thallium scan, and major vessels scanned using X-rays with a high capability to distinguish between healthy and heart disease patients with a 99.6% accuracy.
Heart Disease Prediction using Machine Learning Algorithmijtsrd
Nowadays, Heart disease has become dangerous to a human being, it effects very badly to human body. If anyone is suffering from heart disease, then it leads to blood clotting. Heart disease prediction is very difficult task to predict in the field of medical science. Affiliation has predicted that 12 million people fail horrendously every year as a result of heart disease. In this paper, we propose a k Nearest Neighbors Algorithm KNN way to deal with improve the exactness of heart determination. We show that k Nearest Neighbors Algorithm KNN have better accuracy than random forest algorithm for viewing heart disease. The k Nearest Neighbors Algorithm give more precise and exact outcome . We have taken 13 attributes in the dataset and a target attribute, by applying machine learning we achieved 84 accuracy in the heart disease detection. Ravi Kumar Singh | Dr. A Rengarajan "Heart Disease Prediction using Machine Learning Algorithm" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38358.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38358/heart-disease-prediction-using-machine-learning-algorithm/ravi-kumar-singh
Prediction of Heart Disease using Machine Learning Algorithms: A Surveyrahulmonikasharma
According to recent survey by WHO organisation 17.5 million people dead each year. It will increase to 75 million in the year 2030[1].Medical professionals working in the field of heart disease have their own limitation, they can predict chance of heart attack up to 67% accuracy[2], with the current epidemic scenario doctors need a support system for more accurate prediction of heart disease. Machine learning algorithm and deep learning opens new door opportunities for precise predication of heart attack. Paper provideslot information about state of art methods in Machine learning and deep learning. An analytical comparison has been provided to help new researches’ working in this field.
This describes the techniques that are used for prediction of heart diseases using the concept of data mining.It states about IHPDS(Intelligent Heart Disease Prediction System)
One of the major purposes manufacturers incorporate AI or ML in their applications is to ease software computations and to predict precise results. I think compared to any other application, a medical application requires a lot of precise computations and therefore, AI is a perfect solution to enhance performance and productivity. While reading the health-tech news, I came across recent research in this regard, the use of AI in predicting a potential stroke or cardiac arrest. ..
A Survey on Heart Disease Prediction Techniquesijtsrd
Heart disease is the main reason for a huge number of deaths in the world over the last few decades and has evolved as the most life threatening disease. The health care industry is found to be rich in information. So, there is a need to discover hidden patterns and trends in them. For this purpose, data mining techniques can be applied to extract the knowledge from the large sets of data. Many researchers, in recent times have been using several machine learning techniques for predicting the heart related diseases as it can predict the disease effectively. Even though a machine learning technique proves to be effective in assisting the decision makers, still there is a scope for developing an accurate and efficient system to diagnose and predict the heart diseases thereby helping doctors with ease of work. This paper presents a survey of various techniques used for predicting heart disease and reviews their performance. G. Niranjana | Dr I. Elizabeth Shanthi "A Survey on Heart Disease Prediction Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38349.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38349/a-survey-on-heart-disease-prediction-techniques/g-niranjana
Multivariate sample similarity measure for feature selection with a resemblan...IJECEIAES
Feature selection improves the classification performance of machine learning models. It also identifies the important features and eliminates those with little significance. Furthermore, feature selection reduces the dimensionality of training and testing data points. This study proposes a feature selection method that uses a multivariate sample similarity measure. The method selects features with significant contributions using a machine-learning model. The multivariate sample similarity measure is evaluated using the University of California, Irvine heart disease dataset and compared with existing feature selection methods. The multivariate sample similarity measure is evaluated with metrics such as minimum subset selected, accuracy, F1-score, and area under the curve (AUC). The results show that the proposed method is able to diagnose chest pain, thallium scan, and major vessels scanned using X-rays with a high capability to distinguish between healthy and heart disease patients with a 99.6% accuracy.
Enhanced Detection System for Trust Aware P2P Communication NetworksEditor IJCATR
Botnet is a number of computers that have been set up to forward transmissions to other computers unknowingly to the user
of the system and it is most significant to detect the botnets. However, peer-to-peer (P2P) structured botnets are very difficult to detect
because, it doesn’t have any centralized server. In this paper, we deliver an infrastructure of P2P that will improve the trust of the peers
and its data. In order to identify the botnets we provide a technique called data provenance integrity. It will ensure the correct origin or
source of information and prevents opponents from using host resources. A reputation based trust model is used for selecting the
trusted peer. In this model, each peer has a reputation value which is calculated based on its past activity. Here a hash table is used for
efficient file searching and data stored in it is based on the reputation value.
C omparative S tudy of D iabetic P atient D ata’s U sing C lassification A lg...Editor IJCATR
Data mining refers to extracting knowledge from large amount of data. Real life data mining approaches are
interesting because they often present a different se
t of problems for
diabetic
patient’s
data
.
The
research area to solve
various problems and classification is one of main problem in the field. The research describes algorithmic discussion of J48
,
J48 Graft, Random tree, REP, LAD. Here used to compare the
performance of computing time, correctly classified
instances, kappa statistics, MAE, RMSE, RAE, RRSE and
to find the error rate measurement for different classifiers in
weka .In this paper the
data
classification is diabetic patients data set is develope
d by collecting data from hospital repository
consists of 1865 instances with different attributes. The instances in the dataset are two categories of blood tests, urine t
ests.
Weka tool is used to classify the data is evaluated using 10 fold cross validat
ion and the results are compared. When the
performance of algorithms
,
we found J48 is better algorithm in most of the cases
Comparative Study of Diabetic Patient Data’s Using Classification Algorithm i...Editor IJCATR
Data mining refers to extracting knowledge from large amount of data. Real life data mining approaches are
interesting because they often present a different set of problems for diabetic patient’s data. The research area to solve
various problems and classification is one of main problem in the field. The research describes algorithmic discussion of J48,
J48 Graft, Random tree, REP, LAD. Here used to compare the performance of computing time, correctly classified
instances, kappa statistics, MAE, RMSE, RAE, RRSE and to find the error rate measurement for different classifiers in
weka .In this paper the data classification is diabetic patients data set is developed by collecting data from hospital repository
consists of 1865 instances with different attributes. The instances in the dataset are two categories of blood tests, urine tests.
Weka tool is used to classify the data is evaluated using 10 fold cross validation and the results are compared. When the
performance of algorithms, we found J48 is better algorithm in most of the cases.
The Healthcare industry contains big and complex data that may be required in order to discover fascinating pattern of diseases & makes effective decisions with the help of different machine learning techniques. Advanced data mining techniques are used to discover knowledge in database and for medical research. This paper has analyzed prediction systems for Diabetes, Kidney and Liver disease using more number of input attributes. The data mining classification techniques, namely Support Vector Machine(SVM) and Random Forest (RF) are analyzed on Diabetes, Kidney and Liver disease database. The performance of these techniques is compared, based on precision, recall, accuracy, f_measure as well as time. As a result of study the proposed algorithm is designed using SVM and RF algorithm and the experimental result shows the accuracy of 99.35%, 99.37 and 99.14 on diabetes, kidney and liver disease respectively.
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISijcsit
The Healthcare industry contains big and complex data that may be required in order to discover fascinating pattern of diseases & makes effective decisions with the help of different machine learning techniques. Advanced data mining techniques are used to discover knowledge in database and for medical
research. This paper has analyzed prediction systems for Diabetes, Kidney and Liver disease using more
number of input attributes. The data mining classification techniques, namely Support Vector Machine(SVM) and Random Forest (RF) are analyzed on Diabetes, Kidney and Liver disease database. The performance of these techniques is compared, based on precision, recall, accuracy, f_measure as well
as time. As a result of study the proposed algorithm is designed using SVM and RF algorithm and the experimental result shows the accuracy of 99.35%, 99.37 and 99.14 on diabetes, kidney and liver disease respectively.
EVALUATING THE ACCURACY OF CLASSIFICATION ALGORITHMS FOR DETECTING HEART DISE...mlaij
The healthcare industry generates enormous amounts of complex clinical data that make the prediction of
disease detection a complicated process. In medical informatics, making effective and efficient decisions is
very important. Data Mining (DM) techniques are mainly used to identify and extract hidden patterns and
interesting knowledge to diagnose and predict diseases in medical datasets. Nowadays, heart disease is
considered one of the most important problems in the healthcare field. Therefore, early diagnosis leads to
a reduction in deaths. DM techniques have proven highly effective for predicting and diagnosing heart
diseases. This work utilizes the classification algorithms with a medical dataset of heart disease; namely,
J48, Random Forest, and Naïve Bayes to discover the accuracy of their performance. We also examine the
impact of the feature selection method. A comparative and analysis study was performed to determine the
best technique using Waikato Environment for Knowledge Analysis (Weka) software, version 3.8.6. The
performance of the utilized algorithms was evaluated using standard metrics such as accuracy, sensitivity
and specificity. The importance of using classification techniques for heart disease diagnosis has been
highlighted. We also reduced the number of attributes in the dataset, which showed a significant
improvement in prediction accuracy. The results indicate that the best algorithm for predicting heart
disease was Random Forest with an accuracy of 99.24%.
EVALUATING THE ACCURACY OF CLASSIFICATION ALGORITHMS FOR DETECTING HEART DIS...mlaij
The healthcare industry generates enormous amounts of complex clinical data that make the prediction of
disease detection a complicated process. In medical informatics, making effective and efficient decisions is
very important. Data Mining (DM) techniques are mainly used to identify and extract hidden patterns and
interesting knowledge to diagnose and predict diseases in medical datasets. Nowadays, heart disease is
considered one of the most important problems in the healthcare field. Therefore, early diagnosis leads to
a reduction in deaths. DM techniques have proven highly effective for predicting and diagnosing heart
diseases. This work utilizes the classification algorithms with a medical dataset of heart disease; namely,
J48, Random Forest, and Naïve Bayes to discover the accuracy of their performance. We also examine the
impact of the feature selection method. A comparative and analysis study was performed to determine the
best technique using Waikato Environment for Knowledge Analysis (Weka) software, version 3.8.6. The
performance of the utilized algorithms was evaluated using standard metrics such as accuracy, sensitivity
and specificity. The importance of using classification techniques for heart disease diagnosis has been
highlighted. We also reduced the number of attributes in the dataset, which showed a significant
improvement in prediction accuracy. The results indicate that the best algorithm for predicting heart
disease was Random Forest with an accuracy of 99.24%
Similar to Classification of Heart Diseases Patients using Data Mining Techniques (20)
Our CPM guide includes everything you need to get started in the Critical Path Method - with step-by-step examples, solutions, as well as schedules to help get your next project done faster and easier. The Critical Path Method (CPM) is a simple but powerful technique for analyzing, planning, and scheduling large, complex projects. It is used to determine a project’s critical path—the longest sequence of tasks that must be finished for the entire project to be complete.
CPM, also known as Critical Path Analysis (CPA), identifies dependencies between tasks, and shows which tasks are critical to a project. The Critical Path Method (CPM) is one of the most important concepts in project management, and certainly among the most enduring. But what is the Critical Path Method, exactly? This beginner-friendly guide will help you understand the Critical Path Method and apply it in your projects.
Early iterations of the Critical Path Method can be traced all the way back to the Manhattan Project in the early 1940s. Given the ambition, scale, and importance of this world-altering project, scientists - and the managers behind them - developed a number of techniques to make sure that the project delivered results on time. For a project management technique, the Critical Path Method has quite an illustrious history. One of these techniques was to map out the most important tasks in any project and use that to estimate the project completion date.
The Critical Path Method in project management is a cornerstone of project planning even to this day. How long a project takes often depends on the most important tasks that constitute it.
Effort estimation is a process in which project managers evaluate how much time and money they need for completing a project. This technique is common in software development, where technology professionals define the resources and schedule for developing a new application or releasing an update. These forecasts help create accurate estimates that often require approval before work on a project begins. Effort estimation is a common tool as part of the Agile methodology, which is a framework that divides a project into smaller phases. In this framework, you can estimate the effort for several components of development, including:
Epics: Epics are large projects that several teams manage throughout development. These usually contain several smaller releases and tasks.
Features: A feature is a piece of functionality or design that addresses a user's need. A feature often includes specific acceptance criteria that detail how that part of the product works.
Sprints: A sprint is a short period containing a fraction of work. Often, a few team members complete development tasks in sprints that build toward epics and releases.
Releases: Releases are software packages development teams can deploy. These often contain several epics and features that teams deploy in iterations.
Development teams might estimate the effort for each of these components of the Agile framework or select components depending on the needs of the project.
A software process model is an abstraction of the software development process. The models specify the stages and order of a process. So, think of this as a representation of the order of activities of the process and the sequence in which they are performed. A model will define the following:
1. The tasks to be performed
2. The input and output of each task
3. The pre and post-conditions for each task
4. The flow and sequence of each task
The goal of a software process model is to provide guidance for controlling and coordinating the tasks to achieve the end product and objectives as effectively as possible.
Managing projects and entire programmes is an important part of OSCE activities in the field and by its Institutions and Secretariat. Good programme and project management requires effective planning, proper implementation, monitoring, and evaluation. The Conflict Prevention Centre defines and implements the Organization’s management methodology and tools, and builds the capacity of staff in this area via specialized coaching and training.
Organize your projects with project plans to keep things on track—before you even start. A project plan houses all the necessary details of your project, such as goals, tasks, scope, deadlines, and deliverables. This shows stakeholders a clear roadmap of your project, ensures you have the resources for it, and holds everyone accountable from the start. In this article, we teach you the seven steps to create your own project plan.
What is project? Software Project Vs. Other Types. Activities by
Software Project Mgt. Plans, Methods and Methodologies. Problems with Software Projects.
The HyperText Markup Language or HTML is the standard markup language for documents designed to be displayed in a web browser. It defines the content and structure of web content. It is often assisted by technologies such as Cascading Style Sheets and scripting languages such as JavaScript.
nodemon is a tool that helps develop Node. js based applications by automatically restarting the node application when file changes in the directory are detected. nodemon does not require any additional changes to your code or method of development. nodemon is a replacement wrapper for node.
Node handles these tasks by running asynchronously, which means that reading user input from a terminal isn't as simple as calling a getInput() function.
The Node.js file system module allows you to work with the file system on your computer. To include the File System module, use the require() method: var fs = require('fs'); Common use for the File System module: Read files.
Transaction processing means dividing information processing up into individual, indivisible operations, called transactions, that complete or fail as a whole; a transaction can't remain in an intermediate, incomplete, state (so other processes can't access the transaction's data until either the transaction has
A web server is software and hardware that uses HTTP (Hypertext Transfer Protocol) and other protocols to respond to client requests made over the World Wide Web. The main job of a web server is to display website content through storing, processing and delivering webpages to users. Besides HTTP, web servers also support SMTP (Simple Mail Transfer Protocol) and FTP (File Transfer Protocol), used for email, file transfer and storage.
Web server hardware is connected to the internet and allows data to be exchanged with other connected devices, while web server software controls how a user accesses hosted files. The web server process is an example of the client/server model. All computers that host websites must have web server software.
Web servers are used in web hosting, or the hosting of data for websites and web-based applications -- or web applications.
How do web servers work?
Web server software is accessed through the domain names of websites and ensures the delivery of the site's content to the requesting user. The software side is also comprised of several components, with at least an HTTP server. The HTTP server is able to understand HTTP and URLs. As hardware, a web server is a computer that stores web server software and other files related to a website, such as HTML documents, images and JavaScript files.
When a web browser, like Google Chrome or Firefox, needs a file that's hosted on a web server, the browser will request the file by HTTP. When the request is received by the web server, the HTTP server will accept the request, find the content and send it back to the browser through HTTP.
More specifically, when a browser requests a page from a web server, the process will follow a series of steps. First, a person will specify a URL in a web browser's address bar. The web browser will then obtain the IP address of the domain name -- either translating the URL through DNS (Domain Name System) or by searching in its cache. This will bring the browser to a web server. The browser will then request the specific file from the web server by an HTTP request. The web server will respond, sending the browser the requested page, again, through HTTP. If the requested page does not exist or if something goes wrong, the web server will respond with an error message. The browser will then be able to display the webpage.
Multiple domains also can be hosted on one web server.
Examples of web server uses
Web servers often come as part of a larger package of internet- and intranet-related programs that are used for:
sending and receiving emails;
downloading requests for File Transfer Protocol (FTP) files; and
building and publishing webpages.
Many basic web servers will also support server-side scripting, which is used to employ scripts on a web server that can customize the response to the client. Server-side scripting runs on the server machine and typically has a broad feature set, which includes database access. The server-side scripting
What is as web server?
A web server is a computer that runs websites. It's a computer program that distributes web pages as they are requisitioned. The basic objective of the web server is to store, process and deliver web pages to the users. This intercommunication is done using Hypertext Transfer Protocol (HTTP).
How do web servers work?
Web server software is accessed through the domain names of websites and ensures the delivery of the site's content to the requesting user. The software side is also comprised of several components, with at least an HTTP server. The HTTP server is able to understand HTTP and URLs. As hardware, a web server is a computer that stores web server software and other files related to a website, such as HTML documents, images and JavaScript files.
When a web browser, like Google Chrome or Firefox, needs a file that's hosted on a web server, the browser will request the file by HTTP. When the request is received by the web server, the HTTP server will accept the request, find the content and send it back to the browser through HTTP.
More specifically, when a browser requests a page from a web server, the process will follow a series of steps. First, a person will specify a URL in a web browser's address bar. The web browser will then obtain the IP address of the domain name -- either translating the URL through DNS (Domain Name System) or by searching in its cache. This will bring the browser to a web server. The browser will then request the specific file from the web server by an HTTP request. The web server will respond, sending the browser the requested page, again, through HTTP. If the requested page does not exist or if something goes wrong, the web server will respond with an error message. The browser will then be able to display the webpage.
Multiple domains also can be hosted on one web server.
Examples of web server uses
Web servers often come as part of a larger package of internet- and intranet-related programs that are used for:
sending and receiving emails;
downloading requests for File Transfer Protocol (FTP) files; and
building and publishing webpages.
Many basic web servers will also support server-side scripting, which is used to employ scripts on a web server that can customize the response to the client. Server-side scripting runs on the server machine and typically has a broad feature set, which includes database access. The server-side scripting process will also use Active Server Pages (ASP), Hypertext Preprocessor (PHP) and other scripting languages. This process also allows HTML documents to be created dynamically.
Number System is a method of representing Numbers on the Number Line with the help of a set of Symbols and rules. These symbols range from 0-9 and are termed as digits. Number System is used to perform mathematical computations ranging from great scientific calculations to calculations like counting the number of Toys for a Kid or Number chocolates remaining in the box. Number Systems comprise of multiple types based on the base value for its digits.
What is the Number Line?
A Number line is a representation of Numbers with a fixed interval in between on a straight line. A Number line contains all the types of numbers like natural numbers, rationals, Integers, etc. Numbers on the number line increase while moving Left to Right and decrease while moving from right to left. Ends of a number line are not defined i.e., numbers on a number line range from infinity on the left side of the zero to infinity on the right side of the zero.
Positive Numbers: Numbers that are represented on the right side of the zero are termed as Positive Numbers. The value of these numbers increases on moving towards the right. Positive numbers are used for Addition between numbers. Example: 1, 2, 3, 4, …
Negative Numbers: Numbers that are represented on the left side of the zero are termed as Negative Numbers. The value of these numbers decreases on moving towards the left. Negative numbers are used for Subtraction between numbers. Example: -1, -2, -3, -4, …
Number and Its Types
A number is a value created by the combination of digits with the help of certain rules. These numbers are used to represent arithmetical quantities. A digit is a symbol from a set 10 symbols ranging from 0, 1, 2, 3, 4, 5, 6, 7, 8, and 9. Any combination of digits represents a Number. The size of a Number depends on the count of digits that are used for its creation.
For Example: 123, 124, 0.345, -16, 73, 9, etc.
Types of Numbers
Numbers are of various types depending upon the patterns of digits that are used for their creation. Various symbols and rules are also applied on Numbers which classifies them into a variety of different types:
Number and Its Types
1. Natural Numbers: Natural Numbers are the most basic type of Numbers that range from 1 to infinity. These numbers are also called Positive Numbers or Counting Numbers. Natural Numbers are represented by the symbol N.
Example: 1, 2, 3, 4, 5, 6, 7, and so on.
2. Whole Numbers: Whole Numbers are basically the Natural Numbers, but they also include ‘zero’. Whole numbers are represented by the symbol W.
Example: 0, 1, 2, 3, 4, and so on.
3. Integers: Integers are the collection of Whole Numbers plus the negative values of the Natural Numbers. Integers do not include fraction numbers i.e. they can’t be written in a/b form. The range of Integers is from the Infinity at the Negative end and Infinity at the Positive end, including zero. Integers are represented by the symbol Z.
Example: ...,-4, -3, -2, -1, 0, 1, 2, 3, 4,...
Programming Language
As we know, to communicate with a person, we need a specific language, similarly to communicate with computers, programmers also need a language is called Programming language.
Before learning the programming language, let's understand what is language?
What is Language?
Language is a mode of communication that is used to share ideas, opinions with each other. For example, if we want to teach someone, we need a language that is understandable by both communicators.
What is a Programming Language?
A programming language is a computer language that is used by programmers (developers) to communicate with computers. It is a set of instructions written in any specific language ( C, C++, Java, Python) to perform a specific task.
A programming language is mainly used to develop desktop applications, websites, and mobile applications.
Types of programming language
1. Low-level programming language
Low-level language is machine-dependent (0s and 1s) programming language. The processor runs low- level programs directly without the need of a compiler or interpreter, so the programs written in low-level language can be run very fast.
Low-level language is further divided into two parts -
i. Machine Language
Machine language is a type of low-level programming language. It is also called as machine code or object code. Machine language is easier to read because it is normally displayed in binary or hexadecimal form (base 16) form. It does not require a translator to convert the programs because computers directly understand the machine language programs.
The advantage of machine language is that it helps the programmer to execute the programs faster than the high-level programming language.
ii. Assembly Language
Assembly language (ASM) is also a type of low-level programming language that is designed for specific processors. It represents the set of instructions in a symbolic and human-understandable form. It uses an assembler to convert the assembly language to machine language.
information system, an integrated set of components for collecting, storing, and processing data and for providing information, knowledge, and digital products. Business firms and other organizations rely on information systems to carry out and manage their operations, interact with their customers and suppliers, and compete in the marketplace. Information systems are used to run inter-organizational supply chains and electronic markets. For instance, corporations use information systems to process financial accounts, manage their human resources, and to reach their potential customers with online promotions. Many major companies are built entirely around information systems. These include eBay, a large auction marketplace; Amazon, an expanding electronic mall and provider of cloud computing services; Alibaba, a business-to-business e-marketplace; and Google, a search engine company that derives most of its revenue from keyword advertising on Internet searches. Governments deploy information systems to provide services cost-effectively to citizens. Digital goods—such as electronic books, video products, and software—and online services, such as gaming and social networking, are delivered with information systems. Individuals rely on information systems, generally Internet-based, for conducting much of their personal lives: for socializing, study, shopping, banking, and entertainment.
As major new technologies for recording and processing information were invented over the millennia, new capabilities appeared, and people became empowered. The invention of the printing press by Johannes Gutenberg in the mid-15th century and the invention of a mechanical calculator by Blaise Pascal in the 17th century are but two examples. These inventions led to a profound revolution in the ability to record, process, disseminate, and reach for information and knowledge. This led, in turn, to even deeper changes in individual lives, business organization, and human governance.
The first large-scale mechanical information system was Herman Hollerith’s census tabulator. Invented in time to process the 1890 U.S. census, Hollerith’s machine represented a major step in automation, as well as an inspiration to develop computerized information systems.
One of the first computers used for such information processing was the UNIVAC I, installed at the U.S. Bureau of the Census in 1951 for administrative use and at General Electric in 1954 for commercial use. Beginning in the late 1970s, personal computers brought some of the advantages of information systems to small businesses and to individuals. Early in the same decade, the Internet began its expansion as a global network of networks. In 1991 the World Wide Web, invented by Tim Berners-Lee as a means to access the interlinked information stored in the globally dispersed computers connected by the Internet, began operation and became the principal service delivered on the network. The global penetration of the...
Applications of Computer Science in Pharmacy
Computer is mandatory in this advanced era and pharmacy and related subjects are not exception to it. This review mainly focuses on the various applications, software’s and use of computers in pharmacy. Computer science and technology is deeply utilized in pharmacy field everywhere like in pharmacy colleges, pharmaceutical industries, research centers, hospital pharmacy and many more. Computer significantly reduces the time, expenditure, and manpower required for any kind of work. Development of various softwares makes it trouble-free to handle huge data. In short, computers are playing critical role in pharmacy field, without computers pharmacy research will be long-lasting andexpensive.
Pharmacy field plays a crucial role in patient health care. It is a huge field which is present worldwide. To run pharmacy field professionally and efficiently, it requires huge management and manpower. But nowadays use of computers in pharmacy field reduced the manpower and time. Computers are almost related to every corner of pharmacy field. These are utilized in the drug design technique, retail pharmacy shop, clinical research centers, crude drug identification,drug storage and business management, hospital and clinical pharmacy, in pharmacy colleges for computer-assisted learning.
The Internet is a huge collection of data. It is available with just one click. Various search engines like Google, Yahoo, Rediff, and Bing help in searching online data related to the pharmacy field just one has to enter his or her area of interest in the search engine.
In the Pharmacy field, effective use of computers started in 1980. Since then there is a great demand for computers in the pharmacy field. Computers are having their own advantages like reduction in time, accuracy, and reduction in manpower, speed, multitasking, non-fatigued, high memory, data storage and many more.
Computers in pharmacy are used for the information of drug data, records and files, drug management (creating, modifying, adding and deleting data in patient files to generate reports), business details.
Applications of Computer Science in Pharmacy
Computer is mandatory in this advanced era and pharmacy and related subjects are not exception to it. This review mainly focuses on the various applications, software’s and use of computers in pharmacy. Computer science and technology is deeply utilized in pharmacy field everywhere like in pharmacy colleges, pharmaceutical industries, research centers, hospital pharmacy and many more. Computer significantly reduces the time, expenditure, and manpower required for any kind of work. Development of various softwares makes it trouble-free to handle huge data. In short, computers are playing critical role in pharmacy field, without computers pharmacy research will be long-lasting andexpensive.
Pharmacy field plays a crucial role in patient health care. It is a huge field which is present worldwide. To run pharmacy field professionally and efficiently, it requires huge management and manpower. But nowadays use of computers in pharmacy field reduced the manpower and time. Computers are almost related to every corner of pharmacy field. These are utilized in the drug design technique, retail pharmacy shop, clinical research centers, crude drug identification,drug storage and business management, hospital and clinical pharmacy, in pharmacy colleges for computer assistedlearning.
Internet is huge collection of data. It is available in just one click. Various search engines like Google, Yahoo, Rediff, Bing help in searching online data related to pharmacy field just one have to enter his or her area of interest in search engine.
In Pharmacy field, effective use of computers started from 1980. Since then there is great demand of computers in pharmacy field. Computers are having their own advantages like reduction in time, accuracy, and reduction in man power, speed, multitasking, non-fatiguness, high memory, data storage and many more.
USE OF INTERNET IN PHARMACY
Internet is collection of huge data. And this data is available for us in just a one click. Internet is useful tool in literature survey. Books are also available on the internet. Various research journals can be easily accessed via internet. There are number of web-sites which are related to pharmacy field. Some of these web sites are as follows;
www.phrma.org
Organization representing America's pharmaceutical research companies provides details of drug development, industry news, and health guides.
www.healthcareforums.com
Created to facilitate interaction among healthcare professionals on specific topics which include discussion of cases, research and other relevant issues.
www.astra.com
This is the official web-site of ASTRA pharmaceuticals which produces medications for respiratory tract, cardiovascular and gastrointestinal diseases, and for pain
Show drafts
volume_up
Empowering the Data Analytics Ecosystem: A Laser Focus on Value
The data analytics ecosystem thrives when every component functions at its peak, unlocking the true potential of data. Here's a laser focus on key areas for an empowered ecosystem:
1. Democratize Access, Not Data:
Granular Access Controls: Provide users with self-service tools tailored to their specific needs, preventing data overload and misuse.
Data Catalogs: Implement robust data catalogs for easy discovery and understanding of available data sources.
2. Foster Collaboration with Clear Roles:
Data Mesh Architecture: Break down data silos by creating a distributed data ownership model with clear ownership and responsibilities.
Collaborative Workspaces: Utilize interactive platforms where data scientists, analysts, and domain experts can work seamlessly together.
3. Leverage Advanced Analytics Strategically:
AI-powered Automation: Automate repetitive tasks like data cleaning and feature engineering, freeing up data talent for higher-level analysis.
Right-Tool Selection: Strategically choose the most effective advanced analytics techniques (e.g., AI, ML) based on specific business problems.
4. Prioritize Data Quality with Automation:
Automated Data Validation: Implement automated data quality checks to identify and rectify errors at the source, minimizing downstream issues.
Data Lineage Tracking: Track the flow of data throughout the ecosystem, ensuring transparency and facilitating root cause analysis for errors.
5. Cultivate a Data-Driven Mindset:
Metrics-Driven Performance Management: Align KPIs and performance metrics with data-driven insights to ensure actionable decision making.
Data Storytelling Workshops: Equip stakeholders with the skills to translate complex data findings into compelling narratives that drive action.
Benefits of a Precise Ecosystem:
Sharpened Focus: Precise access and clear roles ensure everyone works with the most relevant data, maximizing efficiency.
Actionable Insights: Strategic analytics and automated quality checks lead to more reliable and actionable data insights.
Continuous Improvement: Data-driven performance management fosters a culture of learning and continuous improvement.
Sustainable Growth: Empowered by data, organizations can make informed decisions to drive sustainable growth and innovation.
By focusing on these precise actions, organizations can create an empowered data analytics ecosystem that delivers real value by driving data-driven decisions and maximizing the return on their data investment.
Opendatabay - Open Data Marketplace.pptxOpendatabay
Opendatabay.com unlocks the power of data for everyone. Open Data Marketplace fosters a collaborative hub for data enthusiasts to explore, share, and contribute to a vast collection of datasets.
First ever open hub for data enthusiasts to collaborate and innovate. A platform to explore, share, and contribute to a vast collection of datasets. Through robust quality control and innovative technologies like blockchain verification, opendatabay ensures the authenticity and reliability of datasets, empowering users to make data-driven decisions with confidence. Leverage cutting-edge AI technologies to enhance the data exploration, analysis, and discovery experience.
From intelligent search and recommendations to automated data productisation and quotation, Opendatabay AI-driven features streamline the data workflow. Finding the data you need shouldn't be a complex. Opendatabay simplifies the data acquisition process with an intuitive interface and robust search tools. Effortlessly explore, discover, and access the data you need, allowing you to focus on extracting valuable insights. Opendatabay breaks new ground with a dedicated, AI-generated, synthetic datasets.
Leverage these privacy-preserving datasets for training and testing AI models without compromising sensitive information. Opendatabay prioritizes transparency by providing detailed metadata, provenance information, and usage guidelines for each dataset, ensuring users have a comprehensive understanding of the data they're working with. By leveraging a powerful combination of distributed ledger technology and rigorous third-party audits Opendatabay ensures the authenticity and reliability of every dataset. Security is at the core of Opendatabay. Marketplace implements stringent security measures, including encryption, access controls, and regular vulnerability assessments, to safeguard your data and protect your privacy.
As Europe's leading economic powerhouse and the fourth-largest hashtag#economy globally, Germany stands at the forefront of innovation and industrial might. Renowned for its precision engineering and high-tech sectors, Germany's economic structure is heavily supported by a robust service industry, accounting for approximately 68% of its GDP. This economic clout and strategic geopolitical stance position Germany as a focal point in the global cyber threat landscape.
In the face of escalating global tensions, particularly those emanating from geopolitical disputes with nations like hashtag#Russia and hashtag#China, hashtag#Germany has witnessed a significant uptick in targeted cyber operations. Our analysis indicates a marked increase in hashtag#cyberattack sophistication aimed at critical infrastructure and key industrial sectors. These attacks range from ransomware campaigns to hashtag#AdvancedPersistentThreats (hashtag#APTs), threatening national security and business integrity.
🔑 Key findings include:
🔍 Increased frequency and complexity of cyber threats.
🔍 Escalation of state-sponsored and criminally motivated cyber operations.
🔍 Active dark web exchanges of malicious tools and tactics.
Our comprehensive report delves into these challenges, using a blend of open-source and proprietary data collection techniques. By monitoring activity on critical networks and analyzing attack patterns, our team provides a detailed overview of the threats facing German entities.
This report aims to equip stakeholders across public and private sectors with the knowledge to enhance their defensive strategies, reduce exposure to cyber risks, and reinforce Germany's resilience against cyber threats.
Classification of Heart Diseases Patients using Data Mining Techniques
1. IJRECE VOL. 6 ISSUE 3 ( JULY - SEPTEMBER 2018) ISSN: 2393-9028 (PRINT) | ISSN: 2348-2281 (ONLINE)
INTERNATIONAL JOURNAL OF RESEARCH IN ELECTRONICS AND COMPUTER ENGINEERING
A UNIT OF I2OR 1495 | P a g e
Classification of Heart Diseases Patients using Data
Mining Techniques
Mukesh Kumar1
, Shankar Shambhu2
, Abha Sharma3
Chitkara University School of Engineering and Technology
Chitkara University, Himachal Pradesh, India
mukesh.kumar@chitkarauniversity.edu.in1
, shankar.shambhu@chitkarauniversity.edu.in2
,
abha.sharma@chitkarauniversity.edu.in3
Abstract— In healthcare sector, data are enormous and
diverse because it contains a data of different types and getting
knowledge from these data is crucial. So to get that
knowledge, data mining techniques may be utilized to mine
knowledge by building models from healthcare dataset. At
present, the classification of heart diseases patients has been a
demanding research confront for many researchers. For
building a classification model for a these patient, we used
four different classification algorithms such as NaiveBayes,
MultilayerPerceptron, RandomForest and DecisionTable. The
intention behind this work is to classify that whether a patient
is tested positive or tested negative for heart diseases, based on
some diagnostic measurements integrated into the dataset.
Keywords— Heart disease, Classification,
MultilayerPerceptron, RandomForest, DecisionTable,
NaiveBayes
I. INTRODUCTION
As we are aware that one of the main reason of death is
heart disease in both men and women worldwide. To prevent
the same we need to take early action to increase survival
chances. In general we are aware of some basic warning signs
of heart attack like discomfort in chest, pain in left arm, short
breath, sweating. Few of major risk factors to heart attack n
clued high BP, high cholesterol, overweight, poor diet, no
physical activity, high stress etc. To prevent heart attack on
should control blood pressure, take healthy diet and control
cholesterol level. Regular exercise is important to maintain
healthy weight. In today’s scenario lifestyle is very hectic so
stress management is very important and of course keeping a
check on alcohol intake.
Researchers all over the world are following different
Machine Learning (ML) techniques to classify and predict the
reason of heart disease. Data mining helps to formulate and
analyse the data to derive data from existing data base [1].
According to American Heart Association (AHA) for the first
time in last 50 years, a statistical report has been released
showing the death rates caused by cardio vascular diseases in
men and women. The age group considered is between 35 to
74 years. These statistics will be helpful to improve and save
life of people worldwide.
II. LITERATURE REVIEW
As per Author [1] data mining is a process through which
one should be able to understand problem definition. Similarly
data collection is very important to identify familiar data.
Once the data is gathered it should be checked for correctness
and structured in a proper way. All these factors are important
to implement data mining in accurate way to produce best
results.
Whereas [2], follows K-Nearest Neighbor algorithm to test
data. KNN is one of the simplest methods and produces
consistent results. Factor considered under this study are age,
minimum and maximum pulse rate and blood pressure. The
comparison of result was based on previous history of patient.
Another study [3] aided towards early diagnosis of the
heart disease and risk factors. The accuracy of results was
around 84% using support vector machine and NaiveBayes.
Factors under consideration for defining accuracy were chest
pain, heart rate and various other attributes. SVM is intended
to produce results at a lower rate and is expected to be used for
future reference.
Comparison with different classifiers has been done to
analyse the best possible method that produces unbiased
estimate. To predict chances of heart disease a total of 10
attributes have been considered. All the classifiers were
implemented using WEKA tool [4].
III. DATASET DESCRIPTION USED FOR PREDICTION
In our research work, we are taking this dataset from
https://github.com/renatopp/arff-
datasets/blob/master/classification/heart.statlog.arff. Table-1
below gives us all the detail regarding our dataset taken into
consideration. The heart diseases patient’s dataset used here
contain 270 different records with 14 attributes. The major
objective of this work is to classify whether a patient is heart
patient or not, based on firm diagnostic measurements
integrated in the dataset.
Data available in real world is incomplete especially in the
area of a medical sector. So to remove unnecessary and noise
in the data, we perform the pre-processing on the data. Pre-
processing of data is a very important stage in this research
work as it affects classification results of the heart diseases
2. IJRECE VOL. 6 ISSUE 3 ( JULY - SEPTEMBER 2018) ISSN: 2393-9028 (PRINT) | ISSN: 2348-2281 (ONLINE)
INTERNATIONAL JOURNAL OF RESEARCH IN ELECTRONICS AND COMPUTER ENGINEERING
A UNIT OF I2OR 1496 | P a g e
patients. Initially, unwanted and noisy data is removed from
the record and secondly, data mining techniques algorithm is
applied to build a classifier model. Classifier Modelling means
selecting diverse techniques and applying them to different
data dataset of the same type. In this research paper, four
different classifications are implemented like NaiveBayes,
MultilayerPerceptron, RandomForest and DecisionTable to
build the best classifier model for our dataset.
Table 1: Dataset description used to implement classification algorithm
S. No. Attribute Description of Attribute Type
1 age Age of the patients Real
2 sex Gender of the patients Binary
3 CP Chest Pain Nominal
4 restbp Resting Blood Pressure Real
5 Cholestoral Serum Cholestoral Real
6 fbs Fasting Blood Sugar Binary
7 restec Resting Electrocardiographic Nominal
8 hr Max Heart Rate Real
9 exercise Exercise induced Angina Binary
10 oldpeak ST depression induced by exercise relative to rest Real
11 slope Slope of the peak exercise ST segment Ordered
12 vessels major vessels (0-3) colored by flourosopy Real
13 thal Thal Nominal
14 Result Present/absent
IV. EXPERIMENTAL SETUP
The dataset is simulated and analyzed in WEKA toolkit.
WEKA is freely available software with already compiled
algorithms of machine learning to perform data mining tasks.
Data mining helps in finding precious information concealed
in enormous amounts of data. For this purpose, we have
WEKA toolkit which has the collection of complied
algorithms of machine learning for data mining purpose. Here,
we have used
K-cross validation test where value of K is 10. We used this
validation test option to minimize process distortion and
recover process efficiency. The four different classifiers
NaiveBayes (NB), MultilayerPerceptron (MP), RandomForest
(RF) and DecisionTable(DT) were simulated on WEKA
toolkit. The results of simulation are showing that the
considered technique gives good results in the literature as
compared to other similar methods, taking into consideration.
Figure 1 show the result of dataset uploaded into the WEKA
tool kit.
3. IJRECE VOL. 6 ISSUE 3 ( JULY - SEPTEMBER 2018) ISSN: 2393-9028 (PRINT) | ISSN: 2348-2281 (ONLINE)
INTERNATIONAL JOURNAL OF RESEARCH IN ELECTRONICS AND COMPUTER ENGINEERING
A UNIT OF I2OR 1497 | P a g e
Figure 1: Environmental Setup of Weka Tool for Implementation
V. RESULT AND DISCUSSION
Here to develop the classification model for patients of
heart we are implemented and analyzed four classification
algorithms NaiveBayes, MultilayerPerceptron, RandomForest
and DecisionTable. In this research paper, for prepare training
dataset to prepare our model and for testing dataset to test the
developed model we used 10-fold cross validation. First, we
check our dataset for baseline accuracy with ZeroR
classification algorithm. The baseline accuracy for our dataset
is 55.55 %, which implies that we need to do a lot of work to
improve the accuracy of our dataset. Table 2 shows the
experimental result of different classification algorithms. We
have conceded some implementation to estimate the accuracy
and effectiveness of multiple classification algorithms for
classifying dataset of heart patients.
Table 2: It shows the performance of different Classifiers used for implementation
Evaluation Criteria for Classification
Algorithm
Classification Algorithms used
NB MP RF DT
Timing to build model (in Sec) 0 Sec 0.43 Sec 0.05 Sec 0.03 Sec
Correctly classified instances 75 69 72 75
Incorrectly classified instances 11 12 14 14
Accuracy (%) 87.20% 85.18 % 83.72 % 84.26 %
Here we show that the NaiveBayes classification algorithm has
performed well with more accuracy as compared to other
algorithm used. In the WEKA tool, a classified instance whose
percentage has accurately classified is called accuracy of the
classifying model. Other evaluation criteria for classification
algorithm implementation are Kappa Statistic (KS), Mean
Absolute Error (MAE), Root Mean Squared Error (RMSE),
Relative Absolute Error (RAE) & Root relative squared Error
(RRSE). In table 3, we illustrate simulation result for a
different algorithm with their evaluation criteria.
Table 3: Training and Simulation error of each classifier used in the implementation
Evaluation Criteria
Classification Algorithms used
NB MP RF DT
Kappa statistic(KS) 0.745 0.745 0.676 0.6868
Mean absolute error(MAE) 0.1705 0.1705 0.2691 0.2656
Root mean squared error (RMSE) 0.3387 0.3387 0.345 0.3643
Relative absolute error (RAE) 33.64 % 33.64 % 53.09 % 52.55 %
Root relative squared error(RRSE) 65.65 % 65.65 % 66.88 % 70.82 %
To choose the algorithms for soaring performance, different
algorithms are implemented and evaluated with respect to
some evaluation criterion on selected dataset. The
classification algorithm which achieves the utmost
performance in provisions of soaring specificity and
sensitivity value is measured by the finest algorithm. From
table 4, it is clear that the NaiveBayes classification algorithm
achieves the maximum value.
The efficiency of the machine learning classifier can be
assessed with numerous measures. The estimate of these
measures basically depends on the contingency table which is
further obtained from the classification algorithm
implemented. Table 4; contain the value of the contingency
table of a particular heart patient dataset.
4. IJRECE VOL. 6 ISSUE 3 ( JULY - SEPTEMBER 2018) ISSN: 2393-9028 (PRINT) | ISSN: 2348-2281 (ONLINE)
INTERNATIONAL JOURNAL OF RESEARCH IN ELECTRONICS AND COMPUTER ENGINEERING
A UNIT OF I2OR 1498 | P a g e
Table 4: Comparison of accuracy measures of each classifier used in an implementation
Classifier TP FP Precision Recall Class
NaiveBayes
(68 percentage split)
0.826 0.075 0.927 0.826 present
0.925 0.174 0.822 0.925 absent
MultilayerPreceptron
( 70 percentage split)
0.826 0.075 0.927 0.826 Present
0.925 0.174 0.822 0.925 Absent
RandomForest
( 68 percentage split)
0.783 0.100 0.900 0.783 Present
0.900 0.217 0.783 0.900 Absent
DecisionTable
( 68 percentage split)
0.787 0.095 0.902 0.787 Present
0.905 0.213 0.792 0.905 Absent
Any classification algorithm performance is extremely
depending on the nature of dataset that is used for training. In
WEKA tool, confusion matrices which are generated after
simulation of classification algorithm are useful to evaluate
different classifiers. In the confusion matrix, columns
represent the predicted classification classes, and rows
represent the actual classes.
Table 5: Confusion Matrix of each classifier used in an implementation
Classifier TP FP Class
NaiveBayes
38 8 Present
3 37 Absent
MultilayerPreceptron
38 8 Present
3 37 Absent
RandomForest
36 10 Present
4 36 Absent
DecisionTable 37 10 Present
4 38 Absent
Based on the above Table 2, we noticeably see that
the maximum accuracy is 87.20% for NaiveBayes and the
minimum accuracy is 83.72% for RandomForest classification
algorithm. These entire algorithm discussed above are tested
using 10-cross validation. In our work, out of 270 different
records, approximately 68% of data are used as training data
and rest of the data is taken as test data.
Table 5. Shows use the confusion matrix of each
classifier used for implementation here. In NaiveBayes
algorithm, out of 86 testing data 75 is tested correctly and 11
are tested incorrectly classified. But in case of RandomForest
out of 86 testing data, 72 are correctly classified and 14 are
incorrectly classified.
VI. CONCLUSTION
In this paper, we are considered the dataset of heart diseases
patients which is further collected at National Institute of
Heart and Digestive and Kidney Diseases. The dataset has 270
instances with fourteen different attributes. We are simulated
NaiveBayes, MultilayerPerceptron, RandomForest and
DecisionTable classification algorithm and found that the
NaiveBayes have the maximum accuracy (87.20%) for
classifying the heart patients whether they are positive or
5. IJRECE VOL. 6 ISSUE 3 ( JULY - SEPTEMBER 2018) ISSN: 2393-9028 (PRINT) | ISSN: 2348-2281 (ONLINE)
INTERNATIONAL JOURNAL OF RESEARCH IN ELECTRONICS AND COMPUTER ENGINEERING
A UNIT OF I2OR 1499 | P a g e
negative. In the future, the results can be utilized to create a
control plan for Heart patients because Heart patients are
normally not identified until a later stage of the disease or the
development of complications.
References
[1] K. Sudhakar, Dr. M. Manimekalai, “Study of Heart Disease
Prediction using Data Mining”, International Journal of
Advanced Research in � Computer Science and Software
Engineering, Volume 4, Issue 1, pp.1157-60, January 2014.
[2] R. Chitra, V. Seenivasagam, “ Review of heart disease
prediction system using data mining and hybrid intelligent
techniques”, ICTACT JOURNAL ON SOFT COMPUTING,
July 2013, volume: 03, issue: 04 pp.605-09.
[3] C. Ordonez, “Association rule discovery with the train and test
approach for heart disease prediction,” IEEE transactions on
Information technology in biomedicine�: a publication of the
IEEE Engineering in Medicine and Biology Society, vol. 10, no.
2,pp.334–43,Apr.2006
[4] H. Chen, S. Y. Huang, PS Hong, CH Cheng, EJ Lin, Heart
Disease Prediction System, Computing in cardiology, pp. 557-
560 2011.
[5] Vikas Chaurasia, and Saurabh Pal, Data Mining Approach to
Detect Heart Dieses, International Journal of Advanced
Computer Science and Information Technology, Vol. 2, No. 4,
pp. 56- 66, 2013.
[6] Abhishek Taneja, Heart Disease Prediction System Using Data
Mining Techniques, Oriental Journal of Computer Science and
technology, Vol. 6, No. 4, 2013.
[7] K. S. Kavitha, K. V. Ramakrishnan, Manoj Kumar Singh,
International Journal of Computer Science, vol. 7, issue 5, pp.
272-283, 2010.
[8] K. Usha Rani, Analysis of Heart Disease Dataset Using Neural
Network approach International Journal of Data Mining &
Knowledge Management Process, Vol. 1, No. 5, September
2011.
[9] V. Sugumaran, V. Muralidharan and K.I. Ramachandran,
Feature selection using Decision Tree and classification through
Proximal Support Vector Machine for fault diagnostics of roller
bearing, Mechanical Systems and Signal Processing, Volume
21, Issue 2, Pages 930-942, February 2007.
[10] Srinivas, K.,” Analysis of coronary heart disease and prediction
of heart attack in coal mining regions using data mining
techniques”, IEEE Transaction on Computer Science and
Education (ICCSE), p(1344 - 1349), 2010.
[11] Yanwei Xing, “Combination Data Mining Methods with New
Medical Data to Predicting Outcome of Coronary Heart
Disease”, IEEE Transactions on Convergence Information
Technology, pp(868 – 872), 21-23 Nov. 2007
[12] S. K. Yadav and Pal S., “Data Mining: A Prediction for
Performance Improvement of Engineering Students using
Classification”, World of Computer Science and Information
Technology (WCSIT), 2(2), 51-56, 2012.