The document describes a dengue detection and prediction system using data mining techniques. Clinical documents are analyzed to extract named entities, symptoms, and other features to generate a feature vector. Various classifiers are trained and evaluated on the vector to identify the best model for predicting dengue. Frequency analysis is also performed to correlate dengue occurrence with symptoms over months. The system achieves appreciable accuracy in detecting and predicting dengue.
Propose a Enhanced Framework for Prediction of Heart DiseaseIJERA Editor
Heart disease diagnosis requires more experience and it is a complex task. The Heart MRI, ECG and Stress Test etc are the numbers of medical tests are prescribed by the doctor for examining the heart disease and it is the way of tradition in the prediction of heart disease. Today world, the hidden information of the huge amount of health care data is contained by the health care industry. The effective decisions are made by means of this hidden information. For appropriate results, the advanced data mining techniques with the information which is based on the computer are used. In any empirical sciences, for the inference and categorisation, the new mathematical techniques to be used called Artificial neural networks (ANNs) it also be used to the modelling of the real neural networks. Acting, Wanting, knowing, remembering, perceiving, thinking and inferring are the nature of mental phenomena and these can be understand by using the theory of ANN. The problem of probability and induction can be arised for the inference and classification because these are the powerful instruments of ANN. In this paper, the classification techniques like Naive Bayes Classification algorithm and Artificial Neural Networks are used to classify the attributes in the given data set. The attribute filtering techniques like PCA (Principle Component Analysis) filtering and Information Gain Attribute Subset Evaluation technique for feature selection in the given data set to predict the heart disease symptoms. A new framework is proposed which is based on the above techniques, the framework will take the input dataset and fed into the feature selection techniques block, which selects any one techniques that gives the least number of attributes and then classification task is done using two algorithms, the same attributes that are selected by two classification task is taken for the prediction of heart disease. This framework consumes the time for predicting the symptoms of heart disease which make the user to know the important attributes based on the proposed framework.
DATA MINING CLASSIFICATION ALGORITHMS FOR KIDNEY DISEASE PREDICTION IJCI JOURNAL
Data mining is a non-trivial process of categorizing valid, novel, potentially useful and ultimately understandable patterns in data. In terms, it accurately state as the extraction of information from a huge database. Data mining is a vital role in several applications such as business organizations, educational institutions, government sectors, health care industry, scientific and engineering. . In the health care
industry, the data mining is predominantly used for disease prediction. Enormous data mining techniques are existing for predicting diseases namely classification, clustering, association rules, summarizations, regression and etc. The main objective of this research work is to predict kidney diseases using classification algorithms such as Naïve Bayes and Support Vector Machine. This research work mainly
focused on finding the best classification algorithm based on the classification accuracy and execution time performance factors. From the experimental results it is observed that the performance of the SVM is better than the Naive Bayes classifier algorithm.
C omparative S tudy of D iabetic P atient D ata’s U sing C lassification A lg...Editor IJCATR
Data mining refers to extracting knowledge from large amount of data. Real life data mining approaches are
interesting because they often present a different se
t of problems for
diabetic
patient’s
data
.
The
research area to solve
various problems and classification is one of main problem in the field. The research describes algorithmic discussion of J48
,
J48 Graft, Random tree, REP, LAD. Here used to compare the
performance of computing time, correctly classified
instances, kappa statistics, MAE, RMSE, RAE, RRSE and
to find the error rate measurement for different classifiers in
weka .In this paper the
data
classification is diabetic patients data set is develope
d by collecting data from hospital repository
consists of 1865 instances with different attributes. The instances in the dataset are two categories of blood tests, urine t
ests.
Weka tool is used to classify the data is evaluated using 10 fold cross validat
ion and the results are compared. When the
performance of algorithms
,
we found J48 is better algorithm in most of the cases
Disease prediction in big data healthcare using extended convolutional neural...IJAAS Team
Diabetes Mellitus is one of the growing fatal diseases all over the world. It leads to complications that include heart disease, stroke, and nerve disease, kidney damage. So, Medical Professionals want a reliable prediction system to diagnose Diabetes. To predict the diabetes at earlier stage, different machine learning techniques are useful for examining the data from different sources and valuable knowledge is synopsized. So, mining the diabetes data in an efficient way is a crucial concern. In this project, a medical dataset has been accomplished to predict the diabetes. The R-Studio and Pypark software was employed as a statistical computing tool for diagnosing diabetes. The PIMA Indian database was acquired from UCI repository will be used for analysis. The dataset was studied and analyzed to build an effective model that predicts and diagnoses the diabetes disease earlier.
Smart Health Prediction Using Data Mining.Data mining is a new powerful technology which is of high interest in computer world. It is a sub field of computer science that uses already existing data in different databases to transform it into new researches and results. It makes use of Artificial Intelligence, machine learning and database management to extract new patterns from large data sets and the knowledge associated with these patterns. The actual task is to extract data by automatic or semi-automatic means. The different parameters included in data mining includes clustering, forecasting, path analysis and predictive analysis.
Propose a Enhanced Framework for Prediction of Heart DiseaseIJERA Editor
Heart disease diagnosis requires more experience and it is a complex task. The Heart MRI, ECG and Stress Test etc are the numbers of medical tests are prescribed by the doctor for examining the heart disease and it is the way of tradition in the prediction of heart disease. Today world, the hidden information of the huge amount of health care data is contained by the health care industry. The effective decisions are made by means of this hidden information. For appropriate results, the advanced data mining techniques with the information which is based on the computer are used. In any empirical sciences, for the inference and categorisation, the new mathematical techniques to be used called Artificial neural networks (ANNs) it also be used to the modelling of the real neural networks. Acting, Wanting, knowing, remembering, perceiving, thinking and inferring are the nature of mental phenomena and these can be understand by using the theory of ANN. The problem of probability and induction can be arised for the inference and classification because these are the powerful instruments of ANN. In this paper, the classification techniques like Naive Bayes Classification algorithm and Artificial Neural Networks are used to classify the attributes in the given data set. The attribute filtering techniques like PCA (Principle Component Analysis) filtering and Information Gain Attribute Subset Evaluation technique for feature selection in the given data set to predict the heart disease symptoms. A new framework is proposed which is based on the above techniques, the framework will take the input dataset and fed into the feature selection techniques block, which selects any one techniques that gives the least number of attributes and then classification task is done using two algorithms, the same attributes that are selected by two classification task is taken for the prediction of heart disease. This framework consumes the time for predicting the symptoms of heart disease which make the user to know the important attributes based on the proposed framework.
DATA MINING CLASSIFICATION ALGORITHMS FOR KIDNEY DISEASE PREDICTION IJCI JOURNAL
Data mining is a non-trivial process of categorizing valid, novel, potentially useful and ultimately understandable patterns in data. In terms, it accurately state as the extraction of information from a huge database. Data mining is a vital role in several applications such as business organizations, educational institutions, government sectors, health care industry, scientific and engineering. . In the health care
industry, the data mining is predominantly used for disease prediction. Enormous data mining techniques are existing for predicting diseases namely classification, clustering, association rules, summarizations, regression and etc. The main objective of this research work is to predict kidney diseases using classification algorithms such as Naïve Bayes and Support Vector Machine. This research work mainly
focused on finding the best classification algorithm based on the classification accuracy and execution time performance factors. From the experimental results it is observed that the performance of the SVM is better than the Naive Bayes classifier algorithm.
C omparative S tudy of D iabetic P atient D ata’s U sing C lassification A lg...Editor IJCATR
Data mining refers to extracting knowledge from large amount of data. Real life data mining approaches are
interesting because they often present a different se
t of problems for
diabetic
patient’s
data
.
The
research area to solve
various problems and classification is one of main problem in the field. The research describes algorithmic discussion of J48
,
J48 Graft, Random tree, REP, LAD. Here used to compare the
performance of computing time, correctly classified
instances, kappa statistics, MAE, RMSE, RAE, RRSE and
to find the error rate measurement for different classifiers in
weka .In this paper the
data
classification is diabetic patients data set is develope
d by collecting data from hospital repository
consists of 1865 instances with different attributes. The instances in the dataset are two categories of blood tests, urine t
ests.
Weka tool is used to classify the data is evaluated using 10 fold cross validat
ion and the results are compared. When the
performance of algorithms
,
we found J48 is better algorithm in most of the cases
Disease prediction in big data healthcare using extended convolutional neural...IJAAS Team
Diabetes Mellitus is one of the growing fatal diseases all over the world. It leads to complications that include heart disease, stroke, and nerve disease, kidney damage. So, Medical Professionals want a reliable prediction system to diagnose Diabetes. To predict the diabetes at earlier stage, different machine learning techniques are useful for examining the data from different sources and valuable knowledge is synopsized. So, mining the diabetes data in an efficient way is a crucial concern. In this project, a medical dataset has been accomplished to predict the diabetes. The R-Studio and Pypark software was employed as a statistical computing tool for diagnosing diabetes. The PIMA Indian database was acquired from UCI repository will be used for analysis. The dataset was studied and analyzed to build an effective model that predicts and diagnoses the diabetes disease earlier.
Smart Health Prediction Using Data Mining.Data mining is a new powerful technology which is of high interest in computer world. It is a sub field of computer science that uses already existing data in different databases to transform it into new researches and results. It makes use of Artificial Intelligence, machine learning and database management to extract new patterns from large data sets and the knowledge associated with these patterns. The actual task is to extract data by automatic or semi-automatic means. The different parameters included in data mining includes clustering, forecasting, path analysis and predictive analysis.
prediction of heart disease using machine learning algorithmsINFOGAIN PUBLICATION
The successful experiment of data mining in highly visible fields like marketing, e-business, and retail has led to its application in other sectors and industries. Healthcare is being discovered among these areas. There is an opulence of data available within the healthcare systems. However, there is a scarcity of useful analysis tool to find hidden relationships in data. This research intends to provide a detailed description of Naïve Bayes and decision tree classifier that are applied in our research particularly in the prediction of Heart Disease. Some experiment has been conducted to compare the execution of predictive data mining technique on the same dataset, and the consequence reveals that Decision Tree outperforms over Bayesian classification.
Heart Disease Prediction Using Data Mining TechniquesIJRES Journal
There are huge amounts of data in the medical industry which is not processed properly and hence cannot be used effectively in making decisions. We can use data mining techniques to mine these patterns and relationships. This research has developed a prototype Heart Disease Prediction using data mining techniques, namely Neural Network, K-Means Clustering and Frequent Item Set Generation. Using medical profiles such as age, sex, blood pressure and blood sugar it can predict the likelihood patients getting a heart disease. It enables significant knowledge, e.g. patterns, relationships between medical factors related to heart disease to be established. Performance of these techniques is compared through sensitivity, specificity and accuracy. It has been observed that Artificial Neural Networks outperform K Means clustering in all the parameters i.e. Sensitivity, Specificity and Accuracy.
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISijcsit
The Healthcare industry contains big and complex data that may be required in order to discover fascinating pattern of diseases & makes effective decisions with the help of different machine learning techniques. Advanced data mining techniques are used to discover knowledge in database and for medical
research. This paper has analyzed prediction systems for Diabetes, Kidney and Liver disease using more
number of input attributes. The data mining classification techniques, namely Support Vector Machine(SVM) and Random Forest (RF) are analyzed on Diabetes, Kidney and Liver disease database. The performance of these techniques is compared, based on precision, recall, accuracy, f_measure as well
as time. As a result of study the proposed algorithm is designed using SVM and RF algorithm and the experimental result shows the accuracy of 99.35%, 99.37 and 99.14 on diabetes, kidney and liver disease respectively.
Psdot 14 using data mining techniques in heartZTech Proje
FINAL YEAR IEEE PROJECTS,
EMBEDDED SYSTEMS PROJECTS,
ENGINEERING PROJECTS,
MCA PROJECTS,
ROBOTICS PROJECTS,
ARM PIC BASED PROJECTS, MICRO CONTROLLER PROJECTS Z Technologies, Chennai
A novel methodology for diagnosing the heart disease using fuzzy databaseeSAT Journals
Abstract A familiar method used for information storing is using a Database. Sometimes the user’s criteria may not be fulfilled due to the presence of vast amount of data in the system of regular database and for the decision making they should be provided with the exact information. In this paper, in the database of the heart disease where there is inaccuracy occurs, accurate information is provided for the users to help them by introducing a medical fuzzy database. The occurrence of uncertainty in the available information for the decision making, the vague values or imprecise in the representation of the data is called Inaccuracy in given data. In this paper, the severity of the patient with the heart disease is diagnosed by introducing a new database system called a fuzzy database management system by utilizing an existing data in the common database systems. Key Words: Fuzzy database, inaccurate data, membership function, linguistic representation, SQL.
A Survey on Heart Disease Prediction Techniquesijtsrd
Heart disease is the main reason for a huge number of deaths in the world over the last few decades and has evolved as the most life threatening disease. The health care industry is found to be rich in information. So, there is a need to discover hidden patterns and trends in them. For this purpose, data mining techniques can be applied to extract the knowledge from the large sets of data. Many researchers, in recent times have been using several machine learning techniques for predicting the heart related diseases as it can predict the disease effectively. Even though a machine learning technique proves to be effective in assisting the decision makers, still there is a scope for developing an accurate and efficient system to diagnose and predict the heart diseases thereby helping doctors with ease of work. This paper presents a survey of various techniques used for predicting heart disease and reviews their performance. G. Niranjana | Dr I. Elizabeth Shanthi "A Survey on Heart Disease Prediction Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38349.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38349/a-survey-on-heart-disease-prediction-techniques/g-niranjana
Prediction of Heart Disease using Machine Learning Algorithms: A Surveyrahulmonikasharma
According to recent survey by WHO organisation 17.5 million people dead each year. It will increase to 75 million in the year 2030[1].Medical professionals working in the field of heart disease have their own limitation, they can predict chance of heart attack up to 67% accuracy[2], with the current epidemic scenario doctors need a support system for more accurate prediction of heart disease. Machine learning algorithm and deep learning opens new door opportunities for precise predication of heart attack. Paper provideslot information about state of art methods in Machine learning and deep learning. An analytical comparison has been provided to help new researches’ working in this field.
HEALTH PREDICTION ANALYSIS USING DATA MININGAshish Salve
Data mining techniques are used for a variety of applications. In healthcare industry, datamining plays an important
role in predicting diseases. For detecting a disease number of tests should be required from the patient. But using data
mining technique the number of tests can be reduced. This reduced test plays an important role in time and performance.
This report analyses data mining techniques which can be used for predicting different types of diseases. This report reviewed
the research papers which mainly concentrate on predicting various disease
OBIA on Coastal Landform Based on Structure Tensor csandit
This paper presents the OBIA method based on structure tensor to identify complex coastal
landforms. That is, develop Hessian matrix by Gabor filtering and calculate multiscale structure
tensor. Extract edge information of image from the trace of structure tensor and conduct
watershed segment of the image. Then, develop texons and create texton histogram. Finally,
obtain the final results by means of maximum likelihood classification with KL divergence as
the similarity measurement. The study findings show that structure tensor could obtain
multiscale and all-direction information with small data redundancy. Moreover, the method
described in the current paper has high classification accuracy
AUTOMATIC GENERATION AND OPTIMIZATION OF TEST DATA USING HARMONY SEARCH ALGOR...csandit
Software testing is the primary phase, which is performed during software development and it is
carried by a sequence of instructions of test inputs followed by expected output. The Harmony
Search (HS) algorithm is based on the improvisation process of music. In comparison to other
algorithms, the HSA has gain popularity and superiority in the field of evolutionary
computation. When musicians compose the harmony through different possible combinations of
the music, at that time the pitches are stored in the harmony memory and the optimization can
be done by adjusting the input pitches and generate the perfect harmony. The test case
generation process is used to identify test cases with resources and also identifies critical
domain requirements. In this paper, the role of Harmony search meta-heuristic search
technique is analyzed in generating random test data and optimized those test data. Test data
are generated and optimized by applying in a case study i.e. a withdrawal task in Bank ATM
through Harmony search. It is observed that this algorithm generates suitable test cases as well
as test data and gives brief details about the Harmony search method. It is used for test data
generation and optimization
Design of IEEE 1149.1 Tap Controller IP Core csandit
An implementation of IEEE 1149.1 TAP controller is presented in this paper. JTAG is an
established technology and industry standard for on-chip boundary scan testing of SoCs. JTAG
TAP controllers are becoming a delivery and control mechanism for Design For Test. The
objective of this work is to design and implement a TAP controller IP core compatible with
IEEE 1149.1-2013 revision of the standard. The test logic architecture also includes the Test
Mode Persistence controller and its associated logic. This work is expected to serve as a ready
to use module that can be directly inserted in to a new digital IC designs with little
modifications.
prediction of heart disease using machine learning algorithmsINFOGAIN PUBLICATION
The successful experiment of data mining in highly visible fields like marketing, e-business, and retail has led to its application in other sectors and industries. Healthcare is being discovered among these areas. There is an opulence of data available within the healthcare systems. However, there is a scarcity of useful analysis tool to find hidden relationships in data. This research intends to provide a detailed description of Naïve Bayes and decision tree classifier that are applied in our research particularly in the prediction of Heart Disease. Some experiment has been conducted to compare the execution of predictive data mining technique on the same dataset, and the consequence reveals that Decision Tree outperforms over Bayesian classification.
Heart Disease Prediction Using Data Mining TechniquesIJRES Journal
There are huge amounts of data in the medical industry which is not processed properly and hence cannot be used effectively in making decisions. We can use data mining techniques to mine these patterns and relationships. This research has developed a prototype Heart Disease Prediction using data mining techniques, namely Neural Network, K-Means Clustering and Frequent Item Set Generation. Using medical profiles such as age, sex, blood pressure and blood sugar it can predict the likelihood patients getting a heart disease. It enables significant knowledge, e.g. patterns, relationships between medical factors related to heart disease to be established. Performance of these techniques is compared through sensitivity, specificity and accuracy. It has been observed that Artificial Neural Networks outperform K Means clustering in all the parameters i.e. Sensitivity, Specificity and Accuracy.
AN ALGORITHM FOR PREDICTIVE DATA MINING APPROACH IN MEDICAL DIAGNOSISijcsit
The Healthcare industry contains big and complex data that may be required in order to discover fascinating pattern of diseases & makes effective decisions with the help of different machine learning techniques. Advanced data mining techniques are used to discover knowledge in database and for medical
research. This paper has analyzed prediction systems for Diabetes, Kidney and Liver disease using more
number of input attributes. The data mining classification techniques, namely Support Vector Machine(SVM) and Random Forest (RF) are analyzed on Diabetes, Kidney and Liver disease database. The performance of these techniques is compared, based on precision, recall, accuracy, f_measure as well
as time. As a result of study the proposed algorithm is designed using SVM and RF algorithm and the experimental result shows the accuracy of 99.35%, 99.37 and 99.14 on diabetes, kidney and liver disease respectively.
Psdot 14 using data mining techniques in heartZTech Proje
FINAL YEAR IEEE PROJECTS,
EMBEDDED SYSTEMS PROJECTS,
ENGINEERING PROJECTS,
MCA PROJECTS,
ROBOTICS PROJECTS,
ARM PIC BASED PROJECTS, MICRO CONTROLLER PROJECTS Z Technologies, Chennai
A novel methodology for diagnosing the heart disease using fuzzy databaseeSAT Journals
Abstract A familiar method used for information storing is using a Database. Sometimes the user’s criteria may not be fulfilled due to the presence of vast amount of data in the system of regular database and for the decision making they should be provided with the exact information. In this paper, in the database of the heart disease where there is inaccuracy occurs, accurate information is provided for the users to help them by introducing a medical fuzzy database. The occurrence of uncertainty in the available information for the decision making, the vague values or imprecise in the representation of the data is called Inaccuracy in given data. In this paper, the severity of the patient with the heart disease is diagnosed by introducing a new database system called a fuzzy database management system by utilizing an existing data in the common database systems. Key Words: Fuzzy database, inaccurate data, membership function, linguistic representation, SQL.
A Survey on Heart Disease Prediction Techniquesijtsrd
Heart disease is the main reason for a huge number of deaths in the world over the last few decades and has evolved as the most life threatening disease. The health care industry is found to be rich in information. So, there is a need to discover hidden patterns and trends in them. For this purpose, data mining techniques can be applied to extract the knowledge from the large sets of data. Many researchers, in recent times have been using several machine learning techniques for predicting the heart related diseases as it can predict the disease effectively. Even though a machine learning technique proves to be effective in assisting the decision makers, still there is a scope for developing an accurate and efficient system to diagnose and predict the heart diseases thereby helping doctors with ease of work. This paper presents a survey of various techniques used for predicting heart disease and reviews their performance. G. Niranjana | Dr I. Elizabeth Shanthi "A Survey on Heart Disease Prediction Techniques" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-5 | Issue-2 , February 2021, URL: https://www.ijtsrd.com/papers/ijtsrd38349.pdf Paper Url: https://www.ijtsrd.com/computer-science/other/38349/a-survey-on-heart-disease-prediction-techniques/g-niranjana
Prediction of Heart Disease using Machine Learning Algorithms: A Surveyrahulmonikasharma
According to recent survey by WHO organisation 17.5 million people dead each year. It will increase to 75 million in the year 2030[1].Medical professionals working in the field of heart disease have their own limitation, they can predict chance of heart attack up to 67% accuracy[2], with the current epidemic scenario doctors need a support system for more accurate prediction of heart disease. Machine learning algorithm and deep learning opens new door opportunities for precise predication of heart attack. Paper provideslot information about state of art methods in Machine learning and deep learning. An analytical comparison has been provided to help new researches’ working in this field.
HEALTH PREDICTION ANALYSIS USING DATA MININGAshish Salve
Data mining techniques are used for a variety of applications. In healthcare industry, datamining plays an important
role in predicting diseases. For detecting a disease number of tests should be required from the patient. But using data
mining technique the number of tests can be reduced. This reduced test plays an important role in time and performance.
This report analyses data mining techniques which can be used for predicting different types of diseases. This report reviewed
the research papers which mainly concentrate on predicting various disease
OBIA on Coastal Landform Based on Structure Tensor csandit
This paper presents the OBIA method based on structure tensor to identify complex coastal
landforms. That is, develop Hessian matrix by Gabor filtering and calculate multiscale structure
tensor. Extract edge information of image from the trace of structure tensor and conduct
watershed segment of the image. Then, develop texons and create texton histogram. Finally,
obtain the final results by means of maximum likelihood classification with KL divergence as
the similarity measurement. The study findings show that structure tensor could obtain
multiscale and all-direction information with small data redundancy. Moreover, the method
described in the current paper has high classification accuracy
AUTOMATIC GENERATION AND OPTIMIZATION OF TEST DATA USING HARMONY SEARCH ALGOR...csandit
Software testing is the primary phase, which is performed during software development and it is
carried by a sequence of instructions of test inputs followed by expected output. The Harmony
Search (HS) algorithm is based on the improvisation process of music. In comparison to other
algorithms, the HSA has gain popularity and superiority in the field of evolutionary
computation. When musicians compose the harmony through different possible combinations of
the music, at that time the pitches are stored in the harmony memory and the optimization can
be done by adjusting the input pitches and generate the perfect harmony. The test case
generation process is used to identify test cases with resources and also identifies critical
domain requirements. In this paper, the role of Harmony search meta-heuristic search
technique is analyzed in generating random test data and optimized those test data. Test data
are generated and optimized by applying in a case study i.e. a withdrawal task in Bank ATM
through Harmony search. It is observed that this algorithm generates suitable test cases as well
as test data and gives brief details about the Harmony search method. It is used for test data
generation and optimization
Design of IEEE 1149.1 Tap Controller IP Core csandit
An implementation of IEEE 1149.1 TAP controller is presented in this paper. JTAG is an
established technology and industry standard for on-chip boundary scan testing of SoCs. JTAG
TAP controllers are becoming a delivery and control mechanism for Design For Test. The
objective of this work is to design and implement a TAP controller IP core compatible with
IEEE 1149.1-2013 revision of the standard. The test logic architecture also includes the Test
Mode Persistence controller and its associated logic. This work is expected to serve as a ready
to use module that can be directly inserted in to a new digital IC designs with little
modifications.
Big Graph : Tools, Techniques, Issues, Challenges and Future Directions csandit
Analyzing interconnection structures among the data through the use of graph algorithms and
graph analytics has been shown to provide tremendous value in many application domains (like
social networks, protein networks, transportation networks, bibliographical networks,
knowledge bases and many more). Nowadays, graphs with billions of nodes and trillions of
edges have become very common. In principle, graph analytics is an important big data
discovery technique. Therefore, with the increasing abundance of large scale graphs, designing
scalable systems for processing and analyzing large scale graphs has become one of the
timeliest problems facing the big data research community. In general, distributed processing of
big graphs is a challenging task due to their size and the inherent irregular structure of graph
computations. In this paper, we present a comprehensive overview of the state-of-the-art to
better understand the challenges of developing very high-scalable graph processing systems. In
addition, we identify a set of the current open research challenges and discuss some promising
directions for future research.
Pentingnya Analisa Support dan Resistance Saat Trading ForexForex Simpro
Belajar teknik trading yang simple dan profile melalui channel YouTube Forex Simpro : https://www.youtube.com/c/ForexSimproTrading
kunjungi website : www.forexsimpro.com
A Survey on Security Risk Management Frameworks in Cloud Computing csandit
Cloud computing technology has experienced exponential growth over the past few years. It
provides many advantages for both individuals and organizations. However, at the same time,
many issues have arisen due to the vast growth of cloud computing. Organizations often have
concerns about the migration and utilization of cloud computing due to the loss of control over
their outsourced resources and cloud computing is vulnerable to risks. Thus, a cloud provider
needs to manage the cloud computing environment risks in order to identify, assess, and
prioritize the risks in order to decrease those risks, improve security, increase confidence in
cloud services, and relieve organizations’ concerns on the issue of using a cloud environment.
Considering that a conventional risk management framework does not fit well with cloud
computing due to the complexity of its environment, research in this area has become
widespread. The aim of this paper is to review the previously proposed risk management
frameworks for cloud computing and to make a comparison between them in order to determine
the strengths and weaknesses of each of them. The review will consider the extent of the
involvement and participation of consumers in cloud computing and other issues.
Effects of The Different Migration Periods on Parallel Multi-Swarm PSO csandit
In recent years, there has been an increasing inter
est in parallel computing. In parallel
computing, multiple computing resources are used si
multaneously in solving a problem. There
are multiple processors that will work concurrently
and the program is divided into different
tasks to be simultaneously solved. Recently, a cons
iderable literature has grown up around the
theme of metaheuristic algorithms. Particle swarm o
ptimization (PSO) algorithm is a popular
metaheuristic algorithm. The parallel comprehensive
learning particle swarm optimization
(PCLPSO) algorithm based on PSO has multiple swarms
based on the master-slave paradigm
and works cooperatively and concurrently. The migra
tion period is an important parameter in
PCLPSO and affects the efficiency of the algorithm.
We used the well-known benchmark
functions in the experiments and analysed the perfo
rmance of PCLPSO using different
migration periods.
AVOIDING DUPLICATED COMPUTATION TO IMPROVE THE PERFORMANCE OF PFSP ON CUDA GPUScsandit
Graphics Processing Units (GPUs) have been emerged as powerful parallel compute platforms for various
application domains. A GPU consists of hundreds or even thousands processor cores and adopts Single
Instruction Multiple Threading (SIMT) architecture. Previously, we have proposed an approach that
optimizes the Tabu Search algorithm for solving the Permutation Flowshop Scheduling Problem (PFSP)
on a GPU by using a math function to generate all different permutations, avoiding the need of placing all
the permutations in the global memory. Based on the research result, this paper proposes another
approach that further improves the performance by avoiding duplicated computation among threads,
which is incurred when any two permutations have the same prefix. Experimental results show that the
GPU implementation of our proposed Tabu Search for PFSP runs up to 1.5 times faster than another GPU
implementation proposed by Czapinski and Barnes
Applications of the Erlang B and C Formulas to Model a Network of Banking Com...csandit
This paper surveys the contributions and applicatio
ns of queueing theory in the field of banking
data networks. We start by highlighting the history
of IT and banks and we continue by
providing information regarding the main prudential
regulations on the banking area as Basel
Accords and green IT regulations, that on one side
generate more computing needs and on the
other side promote conscientious use of the existin
g IT systems.
Continuing with a background of the network technol
ogies used in Economics, the focus will be
on the queueing theory, describing and giving an ov
erview of the most important queueing
models used in economical informatics. While the qu
eueing theory is characterized by its
practical, intuitive and subtle attributes, the que
ueing models are described by a set of 3
factors: an input process, a service process and a
physical configuration of the queue or the
queueing discipline.
The Erlang B and C mathematical definitions of form
ulas for a specific number of s servers, at
the
λ
arrival rate, and the average service time will
be described, used and confirmed by
computer simulations of real queues usually found i
n the banking computing systems.
The goal is to provide sufficient information to co
mputer performance analysts who are
interested in using the queueing theory to model a
network of banking computer systems using
the right simulation model applied in real-life sce
narios, e.g. overcoming the negative impacts
of the European banking regulations while moving to
wards green computing.
Projection Profile Based Number Plate Localization and Recognition csandit
This paper proposes algorithms to localize vehicle
number plates from natural background
images, to segment the characters from the localize
d number plates and to recognize the
segmented characters. The reported system is tested
on a dataset of 560 sample images
captured with different background under various il
luminations. The performance accuracy of
the proposed system has been calculated at each sta
ge, which is 97.1%, 95.4% and 95.72% for
localisation & extraction, character segmentation a
nd character recognition respectively. The
proposed method is also capable of localising and r
ecognising multiple number plates in
images.
Management Architecture for Dynamic Federated Identity Management csandit
We present the concept and design of Dynamic Automa
ted Metadata Exchange (DAME) in
Security Assertion Markup Language (SAML) based use
r authentication and authorization
infrastructures. This approach solves the real-worl
d limitations in scalability of pre-exchanged
metadata in SAML-based federations and inter-federa
tions. The user initiates the metadata
exchange on demand, therefore reducing the size of
the exchanged metadata compared to
traditional metadata aggregation. In order to speci
fy and discuss the necessary changes to
identity federation architectures, we apply the Mun
ich Network Management (MNM) service
model to Federated Identity Management via a truste
d third party (TTP); an overview of all
components and interactions is created. Based on th
is model, the management architecture of
the TTP with its basic management functionalities i
s designed. This management architecture
includes further functionality for automated manage
ment of entities and dynamic federations.
CREATING DATA OUTPUTS FROM MULTI AGENT TRAFFIC MICRO SIMULATION TO ASSIMILATI...csandit
The intensive development of traffic engineering and technologies that are integrated into
vehicles, roads and their surroundings, bring opportunities of real time transport mobility
modeling. Based on such model it is then possible to establish a predictive layer that is capable
of predicting short and long term traffic flow behavior. It is possible to create the real time
model of traffic mobility based on generated data. However, data may have different
geographical, temporal or other constraints, or failures. It is therefore appropriate to develop
tools that artificially create missing data, which can then be assimilated with real data. This
paper presents a mechanism describing strategies of generating artificial data using
microsimulations. It describes traffic microsimulation based on our solution of multiagent
framework over which a system for generating traffic data is built. The system generates data of
a structure corresponding to the data acquired in the real world.
Functional encryption (FE) has more fine-grained control to encrypted data than traditional
encryption schemes. The well-accepted security of FE is indistinguishability-based security
(IND-FE) and simulation-based security (SIMFE), but the security is not sufficient. For
example, if an adversary has the ability to access a vector of ciphertexts and can ask to open
some information of the messages, such as coins used in the encryption or secret key in multikey
setting, whether the privacy of the unopened messages is guaranteed. This is called selective
opening attack (SOA).
In this paper, we propose a stronger security of FE which is secure against SOA (we call SOFE)
and propose a concrete construction of SO-FE scheme in the standard model. Our scheme
is a non-adaptive IND-FE which satisfies selective opening secure in the simulation sense. In
addition, the scheme can encrypt messages of any bit length other than bitwise and it is secure
against SOA-C and SOAK simultaneously while the two attacks were appeared in different
model before. According to the different functionality f, our scheme can specialize as IBE, ABE
and even PE schemes secure against SOA.
Trend removal from raman spectra with local variance estimation and cubic spl...csijjournal
Trend removal is an important problem in most communication systems. Here, we show a proposed
algorithm for trend (background) removal from Raman Spectra by merging local variance estimation and
cubic spline interpolation methods. We found that Raman spectrum does not need a smoothing process to
remove trend from noisy signals. Employing this technique results in more speedy and noiseless systems
than other techniques that use wavelet transformation to suppress noise.
Mining Fuzzy Association Rules from Web Usage Quantitative Data csandit
Web usage mining is the method of extracting interesting patterns from Web usage log file. Web
usage mining is subfield of data mining uses various data mining techniques to produce
association rules. Data mining techniques are used to generate association rules from
transaction data. Most of the time transactions are boolean transactions, whereas Web usage
data consists of quantitative values. To handle these real world quantitative data we used fuzzy
data mining algorithm for extraction of association rules from quantitative Web log file. To
generate fuzzy association rules first we designed membership function. This membership
function is used to transform quantitative values into fuzzy terms. Experiments are carried out
on different support and confidence. Experimental results show the performance of the
algorithm with varied supports and confidence.
VIRTUAL SCENE CONSTRUCTION OF LARGE-SCALE CULTURAL HERITAGE : A FRAMEWORK INI...csandit
Virtual reality technology has been applied to the protection of cultural heritage for about 20 years. However, methods or systems of cultural heritage reported in previous studies are still unable to represent large-scale cultural heritage sites such as the Beijing-Hangzhou Grand Canal, the Struve Geodetic Arc and the boundaries of the Roman Empire. We aimed at
constructing a large-scale cultural-heritage 3-D model with the focus on better management and organization of the scene. Starting from the case study of the Beijing-Hangzhou Grand
Canal, we first explore various remote sensing data suitable for large-scale cultural heritage modeling, and then adopt a 3-D geographic global information system for large-scale 3-D
scene organization and management.
Бизнес-планы по стандартам UNIDO, на получение субсидии, для стартапа и развития бизнеса от консалтинговой компании "СИМЭН". Более подробная информация на сайте www.ciman.ru.
OPTIMIZED PREDICTION IN MEDICAL DIAGNOSIS USING DNA SEQUENCES AND STRUCTURE I...IAEME Publication
The prediction in medical diagnosis is an important role in the field of bio informatics which acts as a vital tool for handling and maintaining human health care in an efficient way. The technologies involves in the process prediction are al complex ones to integrate and implements in an effective way. The optimized prediction in the medical diagnosis reduces the medical treatment complexities along with the time and cost saving benefits. The DNA (Deoxyribonucleic Acid) sequences and structure information will make the disease diagnosis prediction with more accuracy and optimizations for comparisons and conclusions. For the impact of inheritance based diseases the earlier predictions will definitely act as a pivotal process for improving the cure strategies for any human being. The existence case disease diagnosis based prediction will also help and supports the medical treatment in an advanced way both in technical and process wise. This paper proposed Optimized Prediction in Medical Diagnosis Using DNA Sequences and Structure Information with pattern matching and comparison analysis using DNA sequence and structure information for individual patient condition. In future this paper will be extended with artificial intelligence based implementation through genetic algorithm to attain a DNA based Medical Diagnosis Disease Prediction System.
Chronic Kidney Disease prediction is one of the most important issues in healthcare analytics. The most interesting and challenging tasks in day to day life is prediction in medical field. In this paper, we employ some machine learning techniques for predicting the chronic kidney disease using clinical data. We use three machine learning algorithms such as Decision Tree(DT) algorithm, Naive Bayesian (NB) algorithm. The performance of the above models are compared with each other in order to select the best classifier in predicting the chronic kidney disease for given dataset.
Text Extraction Engine to Upgrade Clinical Decision Support SystemIJRTEMJOURNAL
New generation technological improvements lead patients to search their symptoms and
corresponding diagnosis on online resources. In this study, it is aimed to develop a machine learning model to
suit in different availability of users. Most of the current systems allow people to choose related symptom in web
interfaces. In addition to these applications it is aimed to implement a new technique which extracts the text-based
symptoms and its related parameters such as, severity, duration, location, cause, accompanied by any other
indicators. This study is applicable for patient`s everyday language statements besides medical expression of
symptoms for corresponding symptoms which is also supporting initial clinical decision system. Extracted terms
are analyzed for matching diagnosis where an accuracy of 90% has been accomplished.
Text Extraction Engine to Upgrade Clinical Decision Support Systemjournal ijrtem
New generation technological improvements lead patients to search their symptoms and
corresponding diagnosis on online resources. In this study, it is aimed to develop a machine learning model to
suit in different availability of users. Most of the current systems allow people to choose related symptom in web
interfaces. In addition to these applications it is aimed to implement a new technique which extracts the text-based
symptoms and its related parameters such as, severity, duration, location, cause, accompanied by any other
indicators. This study is applicable for patient`s everyday language statements besides medical expression of
symptoms for corresponding symptoms which is also supporting initial clinical decision system. Extracted terms
are analyzed for matching diagnosis where an accuracy of 90% has been accomplished.
The Healthcare industry contains big and complex data that may be required in order to discover fascinating pattern of diseases & makes effective decisions with the help of different machine learning techniques. Advanced data mining techniques are used to discover knowledge in database and for medical research. This paper has analyzed prediction systems for Diabetes, Kidney and Liver disease using more number of input attributes. The data mining classification techniques, namely Support Vector Machine(SVM) and Random Forest (RF) are analyzed on Diabetes, Kidney and Liver disease database. The performance of these techniques is compared, based on precision, recall, accuracy, f_measure as well as time. As a result of study the proposed algorithm is designed using SVM and RF algorithm and the experimental result shows the accuracy of 99.35%, 99.37 and 99.14 on diabetes, kidney and liver disease respectively.
A comprehensive study on disease risk predictions in machine learning IJECEIAES
Over recent years, multiple disease risk prediction models have been developed. These models use various patient characteristics to estimate the probability of outcomes over a certain period of time and hold the potential to improve decision making and individualize care. Discovering hidden patterns and interactions from medical databases with growing evaluation of the disease prediction model has become crucial. It needs many trials in traditional clinical findings that could complicate disease prediction. A Comprehensive study on different strategies used to predict disease is conferred in this paper. Applying these techniques to healthcare data, has improvement of risk prediction models to find out the patients who would get benefit from disease management programs to reduce hospital readmission and healthcare cost, but the results of these endeavors have been shifted.
Enhanced Detection System for Trust Aware P2P Communication NetworksEditor IJCATR
Botnet is a number of computers that have been set up to forward transmissions to other computers unknowingly to the user
of the system and it is most significant to detect the botnets. However, peer-to-peer (P2P) structured botnets are very difficult to detect
because, it doesn’t have any centralized server. In this paper, we deliver an infrastructure of P2P that will improve the trust of the peers
and its data. In order to identify the botnets we provide a technique called data provenance integrity. It will ensure the correct origin or
source of information and prevents opponents from using host resources. A reputation based trust model is used for selecting the
trusted peer. In this model, each peer has a reputation value which is calculated based on its past activity. Here a hash table is used for
efficient file searching and data stored in it is based on the reputation value.
Comparative Study of Diabetic Patient Data’s Using Classification Algorithm i...Editor IJCATR
Data mining refers to extracting knowledge from large amount of data. Real life data mining approaches are
interesting because they often present a different set of problems for diabetic patient’s data. The research area to solve
various problems and classification is one of main problem in the field. The research describes algorithmic discussion of J48,
J48 Graft, Random tree, REP, LAD. Here used to compare the performance of computing time, correctly classified
instances, kappa statistics, MAE, RMSE, RAE, RRSE and to find the error rate measurement for different classifiers in
weka .In this paper the data classification is diabetic patients data set is developed by collecting data from hospital repository
consists of 1865 instances with different attributes. The instances in the dataset are two categories of blood tests, urine tests.
Weka tool is used to classify the data is evaluated using 10 fold cross validation and the results are compared. When the
performance of algorithms, we found J48 is better algorithm in most of the cases.
SUPERVISED FEATURE SELECTION FOR DIAGNOSIS OF CORONARY ARTERY DISEASE BASED O...csitconf
Feature Selection (FS) has become the focus of much research on decision support systems
areas for which datasets with tremendous number of variables are analyzed. In this paper we
present a new method for the diagnosis of Coronary Artery Diseases (CAD) founded on Genetic
Algorithm (GA) wrapped Bayes Naïve (BN) based FS.
Basically, CAD dataset contains two classes defined with 13 features. In GA–BN algorithm, GA
generates in each iteration a subset of attributes that will be evaluated using the BN in the
second step of the selection procedure. The final set of attribute contains the most relevant
feature model that increases the accuracy. The algorithm in this case produces 85.50%
classification accuracy in the diagnosis of CAD. Thus, the asset of the Algorithm is then
compared with the use of Support Vector Machine (SVM), Multi-Layer Perceptron (MLP) and
C4.5 decision tree Algorithm. The result of classification accuracy for those algorithms are
respectively 83.5%, 83.16% and 80.85%. Consequently, the GA wrapped BN Algorithm is
correspondingly compared with other FS algorithms. The Obtained results have shown very
promising outcomes for the diagnosis of CAD.
Similar to DENGUE DETECTION AND PREDICTION SYSTEM USING DATA MINING WITH FREQUENCY ANALYSIS (20)
Welocme to ViralQR, your best QR code generator.ViralQR
Welcome to ViralQR, your best QR code generator available on the market!
At ViralQR, we design static and dynamic QR codes. Our mission is to make business operations easier and customer engagement more powerful through the use of QR technology. Be it a small-scale business or a huge enterprise, our easy-to-use platform provides multiple choices that can be tailored according to your company's branding and marketing strategies.
Our Vision
We are here to make the process of creating QR codes easy and smooth, thus enhancing customer interaction and making business more fluid. We very strongly believe in the ability of QR codes to change the world for businesses in their interaction with customers and are set on making that technology accessible and usable far and wide.
Our Achievements
Ever since its inception, we have successfully served many clients by offering QR codes in their marketing, service delivery, and collection of feedback across various industries. Our platform has been recognized for its ease of use and amazing features, which helped a business to make QR codes.
Our Services
At ViralQR, here is a comprehensive suite of services that caters to your very needs:
Static QR Codes: Create free static QR codes. These QR codes are able to store significant information such as URLs, vCards, plain text, emails and SMS, Wi-Fi credentials, and Bitcoin addresses.
Dynamic QR codes: These also have all the advanced features but are subscription-based. They can directly link to PDF files, images, micro-landing pages, social accounts, review forms, business pages, and applications. In addition, they can be branded with CTAs, frames, patterns, colors, and logos to enhance your branding.
Pricing and Packages
Additionally, there is a 14-day free offer to ViralQR, which is an exceptional opportunity for new users to take a feel of this platform. One can easily subscribe from there and experience the full dynamic of using QR codes. The subscription plans are not only meant for business; they are priced very flexibly so that literally every business could afford to benefit from our service.
Why choose us?
ViralQR will provide services for marketing, advertising, catering, retail, and the like. The QR codes can be posted on fliers, packaging, merchandise, and banners, as well as to substitute for cash and cards in a restaurant or coffee shop. With QR codes integrated into your business, improve customer engagement and streamline operations.
Comprehensive Analytics
Subscribers of ViralQR receive detailed analytics and tracking tools in light of having a view of the core values of QR code performance. Our analytics dashboard shows aggregate views and unique views, as well as detailed information about each impression, including time, device, browser, and estimated location by city and country.
So, thank you for choosing ViralQR; we have an offer of nothing but the best in terms of QR code services to meet business diversity!
Key Trends Shaping the Future of Infrastructure.pdfCheryl Hung
Keynote at DIGIT West Expo, Glasgow on 29 May 2024.
Cheryl Hung, ochery.com
Sr Director, Infrastructure Ecosystem, Arm.
The key trends across hardware, cloud and open-source; exploring how these areas are likely to mature and develop over the short and long-term, and then considering how organisations can position themselves to adapt and thrive.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Essentials of Automations: Optimizing FME Workflows with ParametersSafe Software
Are you looking to streamline your workflows and boost your projects’ efficiency? Do you find yourself searching for ways to add flexibility and control over your FME workflows? If so, you’re in the right place.
Join us for an insightful dive into the world of FME parameters, a critical element in optimizing workflow efficiency. This webinar marks the beginning of our three-part “Essentials of Automation” series. This first webinar is designed to equip you with the knowledge and skills to utilize parameters effectively: enhancing the flexibility, maintainability, and user control of your FME projects.
Here’s what you’ll gain:
- Essentials of FME Parameters: Understand the pivotal role of parameters, including Reader/Writer, Transformer, User, and FME Flow categories. Discover how they are the key to unlocking automation and optimization within your workflows.
- Practical Applications in FME Form: Delve into key user parameter types including choice, connections, and file URLs. Allow users to control how a workflow runs, making your workflows more reusable. Learn to import values and deliver the best user experience for your workflows while enhancing accuracy.
- Optimization Strategies in FME Flow: Explore the creation and strategic deployment of parameters in FME Flow, including the use of deployment and geometry parameters, to maximize workflow efficiency.
- Pro Tips for Success: Gain insights on parameterizing connections and leveraging new features like Conditional Visibility for clarity and simplicity.
We’ll wrap up with a glimpse into future webinars, followed by a Q&A session to address your specific questions surrounding this topic.
Don’t miss this opportunity to elevate your FME expertise and drive your projects to new heights of efficiency.
GDG Cloud Southlake #33: Boule & Rebala: Effective AppSec in SDLC using Deplo...James Anderson
Effective Application Security in Software Delivery lifecycle using Deployment Firewall and DBOM
The modern software delivery process (or the CI/CD process) includes many tools, distributed teams, open-source code, and cloud platforms. Constant focus on speed to release software to market, along with the traditional slow and manual security checks has caused gaps in continuous security as an important piece in the software supply chain. Today organizations feel more susceptible to external and internal cyber threats due to the vast attack surface in their applications supply chain and the lack of end-to-end governance and risk management.
The software team must secure its software delivery process to avoid vulnerability and security breaches. This needs to be achieved with existing tool chains and without extensive rework of the delivery processes. This talk will present strategies and techniques for providing visibility into the true risk of the existing vulnerabilities, preventing the introduction of security issues in the software, resolving vulnerabilities in production environments quickly, and capturing the deployment bill of materials (DBOM).
Speakers:
Bob Boule
Robert Boule is a technology enthusiast with PASSION for technology and making things work along with a knack for helping others understand how things work. He comes with around 20 years of solution engineering experience in application security, software continuous delivery, and SaaS platforms. He is known for his dynamic presentations in CI/CD and application security integrated in software delivery lifecycle.
Gopinath Rebala
Gopinath Rebala is the CTO of OpsMx, where he has overall responsibility for the machine learning and data processing architectures for Secure Software Delivery. Gopi also has a strong connection with our customers, leading design and architecture for strategic implementations. Gopi is a frequent speaker and well-known leader in continuous delivery and integrating security into software delivery.
SAP Sapphire 2024 - ASUG301 building better apps with SAP Fiori.pdfPeter Spielvogel
Building better applications for business users with SAP Fiori.
• What is SAP Fiori and why it matters to you
• How a better user experience drives measurable business benefits
• How to get started with SAP Fiori today
• How SAP Fiori elements accelerates application development
• How SAP Build Code includes SAP Fiori tools and other generative artificial intelligence capabilities
• How SAP Fiori paves the way for using AI in SAP apps
Securing your Kubernetes cluster_ a step-by-step guide to success !KatiaHIMEUR1
Today, after several years of existence, an extremely active community and an ultra-dynamic ecosystem, Kubernetes has established itself as the de facto standard in container orchestration. Thanks to a wide range of managed services, it has never been so easy to set up a ready-to-use Kubernetes cluster.
However, this ease of use means that the subject of security in Kubernetes is often left for later, or even neglected. This exposes companies to significant risks.
In this talk, I'll show you step-by-step how to secure your Kubernetes cluster for greater peace of mind and reliability.
A tale of scale & speed: How the US Navy is enabling software delivery from l...sonjaschweigert1
Rapid and secure feature delivery is a goal across every application team and every branch of the DoD. The Navy’s DevSecOps platform, Party Barge, has achieved:
- Reduction in onboarding time from 5 weeks to 1 day
- Improved developer experience and productivity through actionable findings and reduction of false positives
- Maintenance of superior security standards and inherent policy enforcement with Authorization to Operate (ATO)
Development teams can ship efficiently and ensure applications are cyber ready for Navy Authorizing Officials (AOs). In this webinar, Sigma Defense and Anchore will give attendees a look behind the scenes and demo secure pipeline automation and security artifacts that speed up application ATO and time to production.
We will cover:
- How to remove silos in DevSecOps
- How to build efficient development pipeline roles and component templates
- How to deliver security artifacts that matter for ATO’s (SBOMs, vulnerability reports, and policy evidence)
- How to streamline operations with automated policy checks on container images
Epistemic Interaction - tuning interfaces to provide information for AI supportAlan Dix
Paper presented at SYNERGY workshop at AVI 2024, Genoa, Italy. 3rd June 2024
https://alandix.com/academic/papers/synergy2024-epistemic/
As machine learning integrates deeper into human-computer interactions, the concept of epistemic interaction emerges, aiming to refine these interactions to enhance system adaptability. This approach encourages minor, intentional adjustments in user behaviour to enrich the data available for system learning. This paper introduces epistemic interaction within the context of human-system communication, illustrating how deliberate interaction design can improve system understanding and adaptation. Through concrete examples, we demonstrate the potential of epistemic interaction to significantly advance human-computer interaction by leveraging intuitive human communication strategies to inform system design and functionality, offering a novel pathway for enriching user-system engagements.
2. 54 Computer Science & Information Technology (CS & IT)
they have concentrated primarily around heart attacks [6] [7] [10] [11] [12]. Inspiration drawn
from such work, combined with the increasing rate of dengue cases around the world motivates
us to develop a system to model, predict and analyze dengue instances. The inability to extract
useful information from clinical documents may hamper the health care experts’ efforts from
understanding the relationship between the prevalence of diseases and the associated factors. The
frequency of diseases can also be allied with its time frame. This is especially true in the case of
water-borne and air-borne diseases. Addressing this task will be a major help to doctors, experts
and patients. This relation will enable health care connoisseurs to take preventive measures and
reduce the prevalence of these diseases.
2. PROBLEM STATEMENT
The knowledge available in medical repositories is effectively mined and analyzed using the
proposed system. The input is a set of annotated discharge summaries containing data pertaining
to the disease dengue. Disorder names are extracted from these summaries and looked up in a
summarized UMLS (Unified Medical Language System). The output produced in this step is
supplied to classifiers which then perform detection and prediction. Further, frequency
correlation is performed with the time frame.
3. OVERVIEW OF PROPOSED SYSTEM
Figure 1. Overview of the system
The annotated discharge summaries are supplied to feature extraction algorithms and the
extracted features are in turn used to generate a feature vector. This is supplied as input to a
classification algorithm and a prediction model is developed. The model generated can then be
used to detect and predict the presence of dengue. Finally, a correlation analysis is performed to
determine how the disease is spread over the months.
4. RELATED WORK
N. Aditya Sundar et al [5] use regular factors contributing to heart diseases, including age, sex,
blood sugar and blood pressure, to predict the likelihood of a patient getting a heart disease. Data
mining techniques of Naïve Bayesian classification and WAC (Weighted Associative Classifier)
3. Computer Science & Information Technology (CS & IT) 55
are used to train a model on existing data. Subsequently, patients and nurses can use this model to
supply features and get a prediction on a possible heart attack. Oona Frunza et al [6] present a
machine learning approach that identifies semantic relations between treatments and diseases and
focuses on three semantic relations (prevent, cure and side effect). Later, features were extracted
from unstructured clinical text, and were used to classify the relationship between diseases and
associated treatments. Jyoti Soni et al [7] have developed a predictive data mining algorithm to
predict the presence of heart disease. Fifteen attributes were selected to perform the prediction
and Decision Tree was found to produce the best results. Classification based on clustering
algorithms was found to not perform well. [12] Proposes a Medical Diagnosis System for
predicting the risk of cardiovascular disease. It uses genetic algorithm to determine the weights
for a neural network. This feed forward neural network is subsequently used for classification and
prediction of heart diseases. A data set of 303 instances of heart disease with 14 attributes each is
used for training the system. Devendra Ratnaparkhi, Tushar Mahajan and Vishal Jadhav in their
paper [10] describe a system for prediction of heart disease using Naïve Bayes algorithms. They
further propose a web interface to help healthcare practitioners assess the possibility of a heart
problem in patients. A similar attempt proposes a heart disease prediction system [11] using
Decision Tree and Naïve Bayes and its implementation in .NET platform by I.S.Jenzi et al in
their paper. Some data mining techniques used for modeling and prediction of dengue include
SVM [13], decision tree [14] and neural network [15].
5. SYSTEM DESIGN
The system design is divided into 2 parts.
5.1. Feature Vector Generation
Figure 2. Feature Vector Generation
4. 56 Computer Science & Information Technology (CS & IT)
5.1.1. POS Tagging
The Stanford POS Tagger is used to tag the discharge summaries. An instance of the tagger class
is created. The input data is stored in a folder. The program iterates through the folder
tags all the input files using the tagger instance created. The tagged data is stored in a file.
5.1.2. Key Term Extraction
The key terms such as nouns and adjectives (specified by the tags NN NNP NNS NNPS JJ JJS
etc) are extracted from the tagged data and stored in a file.
Figure 3. Key term extraction pseudo code
5.1.3. Duplicates Removal
The file generated might contain redundant attributes. To avoid this, the duplicates are removed.
The pseudo code for the same is given below
Figure 4. Duplicates removal pseudo code
Computer Science & Information Technology (CS & IT)
The Stanford POS Tagger is used to tag the discharge summaries. An instance of the tagger class
is created. The input data is stored in a folder. The program iterates through the folder
tags all the input files using the tagger instance created. The tagged data is stored in a file.
Figure 2. POS tagging pseudo code
The key terms such as nouns and adjectives (specified by the tags NN NNP NNS NNPS JJ JJS
etc) are extracted from the tagged data and stored in a file.
Figure 3. Key term extraction pseudo code
The file generated might contain redundant attributes. To avoid this, the duplicates are removed.
same is given below
Figure 4. Duplicates removal pseudo code
The Stanford POS Tagger is used to tag the discharge summaries. An instance of the tagger class
is created. The input data is stored in a folder. The program iterates through the folder’s files and
tags all the input files using the tagger instance created. The tagged data is stored in a file.
The key terms such as nouns and adjectives (specified by the tags NN NNP NNS NNPS JJ JJS
The file generated might contain redundant attributes. To avoid this, the duplicates are removed.
5. Computer Science & Information Technology
5.1.4. Dictionary Look Up
UMLS (Unified Medical Language System) serves as a repository of mentions. The UMLS is
used to extract the relevant symptoms from the tagged file. FileSearcher
the FindWordInFile method is used to search for a word in a given file.
Figure 5. Dictionary lookup pseudo code
5.1.5. Temporal Data Extraction
The discharge summaries are fed as input to the temporal data extraction algorithm.
admission months are extracted using regular expressions.
Figure 6. Temporal data extraction pseudo code
5.1.6. Non-Symptomatic Feature Extraction
Non-Symptomatic features such as age, gender, marital status, family history and past medical
history are extracted using regular expressions from the annotated discharge summaries.
Computer Science & Information Technology (CS & IT)
fied Medical Language System) serves as a repository of mentions. The UMLS is
used to extract the relevant symptoms from the tagged file. FileSearcher Class is imported and
the FindWordInFile method is used to search for a word in a given file.
Figure 5. Dictionary lookup pseudo code
5.1.5. Temporal Data Extraction
The discharge summaries are fed as input to the temporal data extraction algorithm.
admission months are extracted using regular expressions.
Figure 6. Temporal data extraction pseudo code
Symptomatic Feature Extraction
Symptomatic features such as age, gender, marital status, family history and past medical
ry are extracted using regular expressions from the annotated discharge summaries.
57
fied Medical Language System) serves as a repository of mentions. The UMLS is
Class is imported and
The discharge summaries are fed as input to the temporal data extraction algorithm. The
Symptomatic features such as age, gender, marital status, family history and past medical
ry are extracted using regular expressions from the annotated discharge summaries.
6. 58 Computer Science & Information Technology (CS & IT)
• The age is computed using the
• The gender can be either M or F (Male or Female)
• The marital status can be Y or N (Yes or No)
• The family history can be Y or N (Yes or No)
• The past medical history can be Y or N (Yes or No)
• The disease can be Y or N (Yes or No)
5.1.7. Feature Vector Generation
The feature vector is the input supplied to the classifier. The features extracted are combined in a
comma separated format and a feature vector is generated. The vector can be represented using
frequency value representation or using binary representation. The frequency value format
implies that, the frequency of occurrence of the feature in the documen
representation on the other hand only considers the presence or absence of the feature in concern.
Dengue has very few prominent symptoms and therefore it is not advisable to use the frequency
value representation to retrieve th
therefore preferred in this case. Non
Figure 7. Feature Vector Generation pseudo code
The feature vector generated is supplied to a
analysis is performed.
Computer Science & Information Technology (CS & IT)
The age is computed using the date of birth of the patient.
The gender can be either M or F (Male or Female)
The marital status can be Y or N (Yes or No)
ory can be Y or N (Yes or No)
The past medical history can be Y or N (Yes or No)
The disease can be Y or N (Yes or No)
5.1.7. Feature Vector Generation
The feature vector is the input supplied to the classifier. The features extracted are combined in a
comma separated format and a feature vector is generated. The vector can be represented using
frequency value representation or using binary representation. The frequency value format
implies that, the frequency of occurrence of the feature in the document is considered. The binary
representation on the other hand only considers the presence or absence of the feature in concern.
Dengue has very few prominent symptoms and therefore it is not advisable to use the frequency
value representation to retrieve them from clinical text. Binary representation of symptoms is
therefore preferred in this case. Non-symptomatic features are represented as nominal attributes.
Figure 7. Feature Vector Generation pseudo code
The feature vector generated is supplied to a set of classifiers. To identify the best classifier an
The feature vector is the input supplied to the classifier. The features extracted are combined in a
comma separated format and a feature vector is generated. The vector can be represented using
frequency value representation or using binary representation. The frequency value format
t is considered. The binary
representation on the other hand only considers the presence or absence of the feature in concern.
Dengue has very few prominent symptoms and therefore it is not advisable to use the frequency
em from clinical text. Binary representation of symptoms is
symptomatic features are represented as nominal attributes.
set of classifiers. To identify the best classifier an
7. Computer Science & Information Technology (CS & IT) 59
5.2. Classification and Analysis
Figure 8. Classification and analysis
5.2.1. Classification
The following is the gist of steps followed during classification process:
1. Prepare Training set
2. Supply training set to the classifiers
3. Build the classification models
4. Save the models that have been built
5. Prepare the test set
6. Evaluate the test set on the saved models
8. 60 Computer Science & Information Technology (CS & IT)
7. Analyze the performance of the classifiers
8. Choose the best classifier
9. Supply unlabeled dataset to the best classifier
10. Obtain the prediction
5.2.2. Frequency Analysis
Frequency analysis aims at correlating the frequency of occurrence of the disease over the months.
Eight most common and highly contributing symptoms for dengue have been chosen. The
occurrences of these symptoms over the months is represented using graphs to give a better
understanding of which symptom contributes the most to the presence of dengue.
6. IMPLEMENTATION
6.1. Data set used
We have used 100 samples of annotated discharge summaries as input to this system. The
personal details of the patients are already preprocessed to ensure patient confidentiality. They
contain details like age, date of birth, date of admission, patient's medical history, medication
administered to the patient during the period of stay in the hospital. And the final diagnosis of the
patient is also mentioned.
6.2. Tagged file
The above dataset is sent to a POS tagger to perform the part of speech tagging. An instance of
the tagger is created and its TagFile method is used to tag the data. This tagged file is sent to a
key term extraction algorithm and the relevant features are extracted. The duplicate terms are
removed from using the duplicates removal algorithm. These terms are stored in a file.
6.3. UMLS Look up
A subset of the UMLS containing terms relevant to the disease are used as basis to perform the
dictionary look up. The file containing the key terms is then compared with the thesaurus and
symptoms that contribute to dengue are stored in another file.
6.4. Feature Extraction and Vector Generation
6.4.1. Symptomatic features
To extract the symptomatic features, the following steps are performed:
1. A file reader object is created
2. The discharge summaries are read line by line
• Each line is split into words
• The words are compared with the file containing filtered output
9. Computer Science & Information Technology
• If there is a match , 1 is written to the feature to the feature vector
3. If there is no match, 0 is written to the vector
6.4.2. Non- Symptomatic feature
The non-symptomatic features are extracted using regular expressions. The features are extracted
and written to the feature vector file.
Snapshot of the generated vector is as shown:
6.5. Classification
The training set is supplied as input to 6 classifiers. Classification analysis was performed on the
classifiers. The steps involved in this analysis are:
• Import the weka and java packages
• Call function useClassifier with the data to be classified as parameter
• Create the classifier object
• Build the classifier model
• Save the model
• Create an Evaluation object
• Cross validate using 10 fold cross validation
• Print the confusion matrix
The results of the analysis are discussed in the Results and Discussions section of the paper.
6.5.1. Prediction on Test Set
The test set contains the samples that aren’t known to the classification model yet. The saved
model is then evaluated on the test set and the accuracy is obtained.
Computer Science & Information Technology (CS & IT)
If there is a match , 1 is written to the feature to the feature vector
, 0 is written to the vector
Symptomatic features
symptomatic features are extracted using regular expressions. The features are extracted
ten to the feature vector file. The feature vector is saved as an arff file.
of the generated vector is as shown:
Figure 9. Feature vector
The training set is supplied as input to 6 classifiers. Classification analysis was performed on the
classifiers. The steps involved in this analysis are:
and java packages
Call function useClassifier with the data to be classified as parameter
Create the classifier object
Build the classifier model
Create an Evaluation object
Cross validate using 10 fold cross validation
matrix
The results of the analysis are discussed in the Results and Discussions section of the paper.
The test set contains the samples that aren’t known to the classification model yet. The saved
model is then evaluated on the test set and the accuracy is obtained.
61
symptomatic features are extracted using regular expressions. The features are extracted
The training set is supplied as input to 6 classifiers. Classification analysis was performed on the
The results of the analysis are discussed in the Results and Discussions section of the paper.
The test set contains the samples that aren’t known to the classification model yet. The saved
10. 62 Computer Science & Information Technology (CS & IT)
6.5.2. Prediction on Unlabeled Dataset
Unlabeled dataset is fed to the saved model. The disease label is a "?" in this case. The model
then predicts the labels for these samples.
6.5.3. Graphical User Interface
A GUI was developed to simplify access to the dengue detection system. Separate panels, one for
researchers and another for common users were developed. Researchers can upload a folder
consisting of discharge summaries which will be used as the training set. Common users can
indicate which symptoms they are experiencing and get a prediction from the system.
Figure 10. Patient GUI
Figure 11. Researcher GUI
11. Computer Science & Information Technology (CS & IT) 63
6.6. Frequency Analysis
To perform frequency analysis, we have used bar charts. The bar charts are generated using
JFreeCharts. The correlations of the spread of the symptoms and in turn the disease over the
months are reported briefly to give a clear picture to the researchers. This feature is only available
to the researchers.
Figure 12. Fever vs month
7. RESULTS AND DISCUSSIONS
The feature vector is supplied to various supervised learning algorithms and classifier models are
generated. LibSVM is integrated software for support vector classification, regression and
distribution estimation. It supports multi-class classification. Logistic regression classifier uses a
sigmoid function to perform the classification. Multilayer perceptron is a classifier based on
Artificial Neural Networks. Each layer is completely connected to the next layer in the network.
Naïve Bayes methods are a set of supervised learning methods based on applying Bayes theorem
with the naïve assumption of independence between every pair of features. The Sequential
Minimal Optimizer uses John Plat’s sequential minimal optimization algorithm for training a
support vector classifier. It also normalizes all attributes by default. The Simple Logistic Classifier
is used for building linear logistic regression models. These classifiers are subject to two types of
classifications – 10-fold cross-validation and percentage split (2/3rd
training and 1/3rd
test).
Accuracies obtained from the 2 methods are compared. In addition, accuracy of the various
classifiers are analyzed based on five performance metrics (Accuracy, Kappa statistics, Mean
absolute error, Root mean squared error, Relative absolute error) [16] and the best model is
chosen.
• Accuracy: The number of samples that are correctly classified from the given 100
input samples.
• Kappa Statistic: The Kappa Statistic can be defined as measuring degree of agreement
between two sets of categorized data. Kappa result varies between 0 to 1intervals.
Higher the value of Kappa means stronger the agreement/ bonding. If Kappa = 1, then
there is perfect agreement. If Kappa = 0, then there is no agreement. If values of Kappa
statics are varying in the range of 0.40 to 0.59 considered as moderate, 0.60 to 0.79
considered as substantial, and above 0.80 considered as outstanding.
12. 64 Computer Science & Information Technology (CS & IT)
• Mean Absolute Error:
divided by number of predictions. It is measure set of predicted value to actual value
i.e. how close a predicted model to actual model. The lower the value of MAE the
better the classification.
• Root Mean Squared Error :
squares error divided by number of predictions. It is measure the differences
values predicted by a model and the values actually observed. Small value of RMSE
means better accuracy of model. Lower the value of RMSE, better the prediction and
accuracy.
• Relative Absolute Error:
measurement to the accepted measurement. A lower percentage indicated better
prediction and accuracy.
Figure 13. Classifier analysis using 10
Based on the above analysis, SMO is identified to be the most optimal classifier
7.1 Analysis and correlation
The predicted results are visualized in graphical form subsequent to prediction. Counts of
occurrences of various symptoms over the months are depicted using bar charts, and these values
are compared with the graphs
maximum manifestation of all symptoms was found to be September. This was also the month
with maximum cases of dengue, according to the prediction. This inference was also corroborated
by the graph generated from the initial training dataset, and we gather from these graphs that
August, September and October are the months most vulnerable to dengue.
Computer Science & Information Technology (CS & IT)
Mean Absolute Error: Mean absolute error can be defined as sum of absolute errors
divided by number of predictions. It is measure set of predicted value to actual value
a predicted model to actual model. The lower the value of MAE the
cation.
Root Mean Squared Error : Root mean square error is defined as square root of sum of
squares error divided by number of predictions. It is measure the differences
values predicted by a model and the values actually observed. Small value of RMSE
means better accuracy of model. Lower the value of RMSE, better the prediction and
Relative Absolute Error: Relative error is the ratio of the absolute erro
measurement to the accepted measurement. A lower percentage indicated better
prediction and accuracy.
Figure 13. Classifier analysis using 10-fold cross validation
Based on the above analysis, SMO is identified to be the most optimal classifier.
Analysis and correlation
The predicted results are visualized in graphical form subsequent to prediction. Counts of
occurrences of various symptoms over the months are depicted using bar charts, and these values
are compared with the graphs generated for the original training dataset. The month with
maximum manifestation of all symptoms was found to be September. This was also the month
with maximum cases of dengue, according to the prediction. This inference was also corroborated
h generated from the initial training dataset, and we gather from these graphs that
August, September and October are the months most vulnerable to dengue.
ned as sum of absolute errors
divided by number of predictions. It is measure set of predicted value to actual value
a predicted model to actual model. The lower the value of MAE the
ned as square root of sum of
squares error divided by number of predictions. It is measure the differences between
values predicted by a model and the values actually observed. Small value of RMSE
means better accuracy of model. Lower the value of RMSE, better the prediction and
Relative error is the ratio of the absolute error of the
measurement to the accepted measurement. A lower percentage indicated better
The predicted results are visualized in graphical form subsequent to prediction. Counts of
occurrences of various symptoms over the months are depicted using bar charts, and these values
generated for the original training dataset. The month with
maximum manifestation of all symptoms was found to be September. This was also the month
with maximum cases of dengue, according to the prediction. This inference was also corroborated
h generated from the initial training dataset, and we gather from these graphs that
13. Computer Science & Information Technology (CS & IT) 65
Figure 14. Overview of all symptoms spread over the months
Figure 15. Occurrences of all symptoms over the months
8. CONCLUSION
To conclude, we have discussed, in this report, the detailed design and related algorithms for a
system to identify disorder mentions from clinical text and correlate its frequency with the time
frame. The annotated discharge summaries are tagged and feature extraction algorithms are used
to obtain the features relevant to the disease, Dengue. This is followed by the generation of a
feature vector (Binary representation). This vector is then used to train and build various
classification models and SMO is found to produce the best results. The model generated further
aids in the prediction of the disease. Bar graphs are then used to succinctly represent this
correlation. Additionally the correlation of training samples with time frame was compared with
the correlation obtained from predicted results and the disease occurrence was found to
concentrate in the months of August, September and October in both the cases.
9. LIMITATIONS
Our system uses only 15 features. Extracting more features might increase the accuracy of the
model. The feature vector is depicted using the binary representation. Using the frequency value
representation might improve overall classification.
10. FUTURE WORK
As a part of our future work, we intend to write an implementation to produce bag of words and
extract more features to produce an extensive analysis. Further, we also intend to implement
14. 66 Computer Science & Information Technology (CS & IT)
tagging of the discharge summaries using BIOS tagging [5]. Whenever hospitals receive new
samples showing a tendency for dengue, those samples must be integrated with the existing
training set. This was, the training and predictive capacity of the model will grow, possible giving
better results in the future. To provide and up-to date analysis, we could extend the project to be
used as a desktop app or browser plugin which will automatically synchronize with new data
received from the hospitals' end.
REFERENCES
[1] Sameer Pradhan, Noemie Elhadad, Wendy Chapman, Suresh Manandhar & Guergana Savova, (July
2014) “Analysis of Clinical Text”, SemEval-2014 Task 7.
[2] Melinda Katona & Richard Farkas, (June 2014) “SZTE-NLP: Clinical Text Analysis with Named
Entity Recognition”, SemEval-2014.
[3] Koldo Gojenola, Maite Oronoz, Alicia Perez & Arantza Casillas, (December 2014) “IxaMed:
Applying Freeling and a Perceptron Sequential Tagger at the Shared Task on Analyzing Clinical
Texts”, SemEval-2014.
[4] Parth Pathak, Pinal Patel, Vishal Panchal, Narayan Choudhary, Amrish Patel & Gautam Joshi, (July
2014) “ezDI: A Hybrid CRF and SVM based Model for Detecting and Encoding Disorder Mentions
in Clinical Notes”, SemEval-2014.
[5] Oana Frunza, Diana Inkpen & Thomas Tran, (June 2011) “A Machine LearningApproach for
Identifying Disease-Treatment Relations in Short Texts”, IEEE transactions on knowledge and data
engineering, vol. 23, Issue no. 6.
[6] Deepali Chandna, (2014) “Diagnosis of Heart Disease Using Data Mining Algorithm”, International
Journal of Computer Science and Information Technologies, Vol. 5 (2), pp1678-1680.
[7] Jyoti Soni, Ujma Ansari, Dipesh Sharma & Sunita Soni , (March 2011) “Predictive Data Mining for
Medical Diagnosis: An Overview of Heart Disease Prediction”, International Journal of Computer
Applications, vol 17.
[8] Smitha T & Dr.V Sundaram, (2012) “Knowledge Discovery from Real Time Database using Data
Mining Technique”, International Journal of Scientific and Research Publications, Volume 2, Issue 4.
[9] M.A.Nishara Banu & B Gomathy, (Nov-Dec 2013) “Disease Predicting System Using Data Mining
Techniques”, International Journal of Technical Research and Applications, Volume 1, Issue 5, PP.
41-45
[10] Devendra Ratnaparkhi, Tushar Mahajan & Vishal Jadhav, (November 2015) “Heart Disease
Prediction System Using Data Mining Technique”, International Research Journal of Engineering and
Technology, Volume: 02 Issue: 08.
[11] I.S.Jenzi, P.Priyanka & Dr.P.Alli, (March 2013) “A Reliable Classifier Model Using Data Mining
Approach for Heart Disease Prediction”, International Journal of Advanced Research in Computer
Science and Software Engineering, Volume 3, Issue 3.
[12] N. G. Bhuvaneswari Amma, (February 2012) “Cardiovascular Disease Prediction System using
Genetic algorithm and Neural Network”, International Conference on Computing, Communication
and Applications, IEEE, pp1-5.
15. Computer Science & Information Technology (CS & IT) 67
[13] A.Shameem Fathima & D.Manimeglai, (March 2012) “Predictive Analysis for the Arbovirus-Dengue
using SVM Classification”, International Journal of Engineering and Technology, Volume 2 No. 3
[14] Daranee Thitipryoonwongse, Prapat Suriyaphol & Nuanwan Soonthornphisaj, (2012) “A Data
Mining Framework for Building Dengue Infection Disease Model”, 26thAnnual Conference of the
Japanese Society for Artificial Intelligence
[15] N.Subitha & Dr.A.Padmapriya, (August 2013) “Diagnosis for Dengue Fever Using Spatial Data
Mining”, International Journal of Computer Trends and Technology, Volume 4 Issue 8
[16] Yugal kumar & G. Sahoo, (July 2012) “Analysis of Parametric & Non Parametric Classifiers for
Classification Technique using WEKA”, I.J. Information Technology and Computer Science, Volume
7, pp43-49.
AUTHORS
Nandini V is currently pursuing her final year, Computer Science and Engineering in
SSN College of Engineering. She has published a paper on Machine Vision in the
ARPN Journal of Engineering and Applied Sciences. Her research interests include
Artificial Intelligence, Robotics, Machine Learning, Machine Vision and Data Mining.
Sriranjitha R is currently pursuing her final year, Computer Science and Engineering
in SSN College of Engineering. She is a member of CSI (Computer Society of India).
Her research interests include Machine Learning, Artificial Intelligence, Data Mining
and Data Structures.
Yazhini T P is currently pursuing her final year, Computer Science and Engineering in
SSN College of Engineering. Her research interests include Computer Networks, Data
Mining, Artificial Intelligence and Web Technology.