— The automation of fault detection in material
science is getting popular because of less cost and time. Steel
plates fault detection is an important material science problem.
Data mining techniques deal with data analysis of large data.
Decision trees are very popular classifiers because of their simple
structures and accuracy. A classifier ensemble is a set of
classifiers whose individual decisions are combined in to classify
new examples. Classifiers ensembles generally perform better
than single classifier. In this paper, we show the application of
decision tree ensembles for steel plates faults prediction. The
results suggest that Random Subspace and AdaBoost.M1 are the
best ensemble methods for steel plates faults prediction with
prediction accuracy more than 80%. We also demonstrate that if
insignificant features are removed from the datasets, the
performance of the decision tree ensembles improve for steel
plates faults prediction. The results suggest the future
development of steel plate faults analysis tools by using decision
tree ensembles.
Comprehensive Survey of Data Classification & Prediction Techniquesijsrd.com
In this paper, we present an literature survey of the modern data classification and prediction algorithms. All these algorithms are very important in real world applications like- heart disease prediction, cancer prediction etc. Classification of data is a very popular and computationally expensive task. The fundamentals of data classification are also discussed in brief.
Comparative study of various supervisedclassification methodsforanalysing def...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Column store decision tree classification of unseen attribute setijma
A decision tree can be used for clustering of frequently used attributes to improve tuple reconstruction time
in column-stores databases. Due to ad-hoc nature of queries, strongly correlative attributes are grouped
together using a decision tree to share a common minimum support probability distribution. At the same
time in order to predict the cluster for unseen attribute set, the decision tree may work as a classifier. In
this paper we propose classification and clustering of unseen attribute set using decision tree to improve
tuple reconstruction time.
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...csandit
Feature selection is a problem closely related to dimensionality reduction. A commonly used
approach in feature selection is ranking the individual features according to some criteria and
then search for an optimal feature subset based on an evaluation criterion to test the optimality.
The objective of this work is to predict more accurately the presence of Learning Disability
(LD) in school-aged children with reduced number of symptoms. For this purpose, a novel
hybrid feature selection approach is proposed by integrating a popular Rough Set based feature
ranking process with a modified backward feature elimination algorithm. The approach follows
a ranking of the symptoms of LD according to their importance in the data domain. Each
symptoms significance or priority values reflect its relative importance to predict LD among the
various cases. Then by eliminating least significant features one by one and evaluating the
feature subset at each stage of the process, an optimal feature subset is generated. The
experimental results shows the success of the proposed method in removing redundant
attributes efficiently from the LD dataset without sacrificing the classification performance.
Ant System and Weighted Voting Method for Multiple Classifier Systems IJECEIAES
Combining multiple classifiers is considered as a general solution for classification tasks. However, there are two problems in combining multiple classifiers: constructing a diverse classifier ensemble; and, constructing an appropriate combiner. In this study, an improved multiple classifier combination scheme is propose. A diverse classifier ensemble is constructed by training them with different feature set partitions. The ant system-based algorithm is used to form the optimal feature set partitions. Weighted voting is used to combine the classifiers’ outputs by considering the strength of the classifiers prior to voting. Experiments were carried out using k-NN ensembles on benchmark datasets from the University of California, Irvine, to evaluate the credibility of the proposed method. Experimental results showed that the proposed method has successfully constructed better k-NN ensembles. Further more the proposed method can be used to develop other multiple classifier systems.
SURVEY ON CLASSIFICATION ALGORITHMS USING BIG DATASETEditor IJMTER
Data mining environment produces a large amount of data that need to be analyzed.
Using traditional databases and architectures, it has become difficult to process, manage and analyze
patterns. To gain knowledge about the Big Data a proper architecture should be understood.
Classification is an important data mining technique with broad applications to classify the various
kinds of data used in nearly every field of our life. Classification is used to classify the item
according to the features of the item with respect to the predefined set of classes. This paper put a
light on various classification algorithms including j48, C4.5, Naive Bayes using large dataset.
Comprehensive Survey of Data Classification & Prediction Techniquesijsrd.com
In this paper, we present an literature survey of the modern data classification and prediction algorithms. All these algorithms are very important in real world applications like- heart disease prediction, cancer prediction etc. Classification of data is a very popular and computationally expensive task. The fundamentals of data classification are also discussed in brief.
Comparative study of various supervisedclassification methodsforanalysing def...eSAT Publishing House
IJRET : International Journal of Research in Engineering and Technology is an international peer reviewed, online journal published by eSAT Publishing House for the enhancement of research in various disciplines of Engineering and Technology. The aim and scope of the journal is to provide an academic medium and an important reference for the advancement and dissemination of research results that support high-level learning, teaching and research in the fields of Engineering and Technology. We bring together Scientists, Academician, Field Engineers, Scholars and Students of related fields of Engineering and Technology.
Column store decision tree classification of unseen attribute setijma
A decision tree can be used for clustering of frequently used attributes to improve tuple reconstruction time
in column-stores databases. Due to ad-hoc nature of queries, strongly correlative attributes are grouped
together using a decision tree to share a common minimum support probability distribution. At the same
time in order to predict the cluster for unseen attribute set, the decision tree may work as a classifier. In
this paper we propose classification and clustering of unseen attribute set using decision tree to improve
tuple reconstruction time.
Feature Selection : A Novel Approach for the Prediction of Learning Disabilit...csandit
Feature selection is a problem closely related to dimensionality reduction. A commonly used
approach in feature selection is ranking the individual features according to some criteria and
then search for an optimal feature subset based on an evaluation criterion to test the optimality.
The objective of this work is to predict more accurately the presence of Learning Disability
(LD) in school-aged children with reduced number of symptoms. For this purpose, a novel
hybrid feature selection approach is proposed by integrating a popular Rough Set based feature
ranking process with a modified backward feature elimination algorithm. The approach follows
a ranking of the symptoms of LD according to their importance in the data domain. Each
symptoms significance or priority values reflect its relative importance to predict LD among the
various cases. Then by eliminating least significant features one by one and evaluating the
feature subset at each stage of the process, an optimal feature subset is generated. The
experimental results shows the success of the proposed method in removing redundant
attributes efficiently from the LD dataset without sacrificing the classification performance.
Ant System and Weighted Voting Method for Multiple Classifier Systems IJECEIAES
Combining multiple classifiers is considered as a general solution for classification tasks. However, there are two problems in combining multiple classifiers: constructing a diverse classifier ensemble; and, constructing an appropriate combiner. In this study, an improved multiple classifier combination scheme is propose. A diverse classifier ensemble is constructed by training them with different feature set partitions. The ant system-based algorithm is used to form the optimal feature set partitions. Weighted voting is used to combine the classifiers’ outputs by considering the strength of the classifiers prior to voting. Experiments were carried out using k-NN ensembles on benchmark datasets from the University of California, Irvine, to evaluate the credibility of the proposed method. Experimental results showed that the proposed method has successfully constructed better k-NN ensembles. Further more the proposed method can be used to develop other multiple classifier systems.
SURVEY ON CLASSIFICATION ALGORITHMS USING BIG DATASETEditor IJMTER
Data mining environment produces a large amount of data that need to be analyzed.
Using traditional databases and architectures, it has become difficult to process, manage and analyze
patterns. To gain knowledge about the Big Data a proper architecture should be understood.
Classification is an important data mining technique with broad applications to classify the various
kinds of data used in nearly every field of our life. Classification is used to classify the item
according to the features of the item with respect to the predefined set of classes. This paper put a
light on various classification algorithms including j48, C4.5, Naive Bayes using large dataset.
Multimodal Biometrics Recognition by Dimensionality Diminution MethodIJERA Editor
Multimodal biometric system utilizes two or more character modalities, e.g., face, ear, and fingerprint,
Signature, plamprint to improve the recognition accuracy of conventional unimodal methods. We propose a new
dimensionality reduction method called Dimension Diminish Projection (DDP) in this paper. DDP can not only
preserve local information by capturing the intra-modal geometry, but also extract between-class relevant
structures for classification effectively. Experimental results show that our proposed method performs better
than other algorithms including PCA, LDA and MFA.
K-Medoids Clustering Using Partitioning Around Medoids for Performing Face Re...ijscmcj
Face recognition is one of the most unobtrusive biometric techniques that can be used for access control as well as surveillance purposes. Various methods for implementing face recognition have been proposed with varying degrees of performance in different scenarios. The most common issue with effective facial biometric systems is high susceptibility of variations in the face owing to different factors like changes in pose, varying illumination, different expression, presence of outliers, noise etc. This paper explores a novel technique for face recognition by performing classification of the face images using unsupervised learning approach through K-Medoids clustering. Partitioning Around Medoids algorithm (PAM) has been used for performing K-Medoids clustering of the data. The results are suggestive of increased robustness to noise and outliers in comparison to other clustering methods. Therefore the technique can also be used to increase the overall robustness of a face recognition system and thereby increase its invariance and make it a reliably usable biometric modality.
Engineering Research Publication
Best International Journals, High Impact Journals,
International Journal of Engineering & Technical Research
ISSN : 2321-0869 (O) 2454-4698 (P)
www.erpublication.org
We conducted comparative analysis of different supervised dimension reduction techniques by integrating a set of different data splitting algorithms and demonstrate the relative efficacy of learning algorithms dependence of sample complexity. The issue of sample complexity discussed in the dependence of data splitting algorithms. In line with the expectations, every supervised learning classifier demonstrated different capability for different data splitting algorithms and no way to calculate overall ranking of techniques was directly available. We specifically focused the classifier ranking dependence of data splitting algorithms and devised a model built on weighted average rank Weighted Mean Rank Risk Adjusted Model (WMRRAM) for consent ranking of learning classifier algorithms.
Comparison Study of Decision Tree Ensembles for RegressionSeonho Park
Nowadays, decision tree ensemble methods are widely used for solving classification and regression problem due to their rigorousness and robustness. To compare with classification, the performance in regression problem so far has not been yet addressed in detail. In this presentation, we review the state-of-art decision tree ensemble methodology in scikit-learn and xgboost for regression. Also, empirical study results are illustrated to compare their performance and computational efficiency.
For years, the Machine Learning community has focused on developing efficient
algorithms that can produce very accurate classifiers. However, it is often much easier
to find several good classifiers based on dataset combination, instead of single classifier
applied on deferent datasets. The advantages of using classifier dataset combinations
instead of a single one are twofold: it helps lowering the computational complexity by
using simpler models, and it can improve the classification accuracy and performance.
Most Data mining applications are based on pattern matching algorithms, thus improving
the performance of the classification has a positive impact on the quality of the overall
data mining task. Since combination strategies proved very useful in improving the
performance, these techniques have become very important in applications such as
Cancer detection, Speech Technology and Natural Language Processing .The aim of this
paper is basically to propose proprietary metric, Normalized Geometric Index (NGI)
based on the latent properties of datasets for improving the accuracy of data mining tasks.
Clustering technology has been applied in numerous applications. It can enhance the performance
of information retrieval systems, it can also group Internet users to help improve the click-through rate of
on-line advertising, etc. Over the past few decades, a great many data clustering algorithms have been
developed, including K-Means, DBSCAN, Bi-Clustering and Spectral clustering, etc. In recent years, two
new data clustering algorithms have been proposed, which are affinity propagation (AP, 2007) and density
peak based clustering (DP, 2014). In this work, we empirically compare the performance of these two latest
data clustering algorithms with state-of-the-art, using 6 external and 2 internal clustering validation metrics.
Our experimental results on 16 public datasets show that, the two latest clustering algorithms, AP and DP,
do not always outperform DBSCAN. Therefore, to find the best clustering algorithm for a specific dataset, all
of AP, DP and DBSCAN should be considered. Moreover, we find that the comparison of different clustering
algorithms is closely related to the clustering evaluation metrics adopted. For instance, when using the
Silhouette clustering validation metric, the overall performance of K-Means is as good as AP and DP. This
work has important reference values for researchers and engineers who need to select appropriate clustering
algorithms for their specific applications.
Exploratory data analysis using xgboost package in RSatoshi Kato
Explain HOW-TO procedure exploratory data analysis using xgboost (EDAXGB), such as feature importance, sensitivity analysis, feature contribution and feature interaction. It is just based on using built-in predict() function in R package.
All of the sample codes are available at: https://github.com/katokohaku/EDAxgboost
EXTRACTING USEFUL RULES THROUGH IMPROVED DECISION TREE INDUCTION USING INFORM...ijistjournal
Classification is widely used technique in the data mining domain, where scalability and efficiency are the immediate problems in classification algorithms for large databases. We suggest improvements to the existing C4.5 decision tree algorithm. In this paper attribute oriented induction (AOI) and relevance analysis are incorporated with concept hierarchy’s knowledge and HeightBalancePriority algorithm for construction of decision tree along with Multi level mining. The assignment of priorities to attributes is done by evaluating information entropy, at different levels of abstraction for building decision tree using HeightBalancePriority algorithm. Modified DMQL queries are used to understand and explore the shortcomings of the decision trees generated by C4.5 classifier for education dataset and the results are compared with the proposed approach.
EXTRACTING USEFUL RULES THROUGH IMPROVED DECISION TREE INDUCTION USING INFORM...ijistjournal
Classification is widely used technique in the data mining domain, where scalability and efficiency are the immediate problems in classification algorithms for large databases. We suggest improvements to the existing C4.5 decision tree algorithm. In this paper attribute oriented induction (AOI) and relevance analysis are incorporated with concept hierarchy’s knowledge and HeightBalancePriority algorithm for construction of decision tree along with Multi level mining. The assignment of priorities to attributes is done by evaluating information entropy, at different levels of abstraction for building decision tree using HeightBalancePriority algorithm. Modified DMQL queries are used to understand and explore the shortcomings of the decision trees generated by C4.5 classifier for education dataset and the results are compared with the proposed approach.
Using particle swarm optimization to solve test functions problemsriyaniaes
In this paper the benchmarking functions are used to evaluate and check the particle swarm optimization (PSO) algorithm. However, the functions utilized have two dimension but they selected with different difficulty and with different models. In order to prove capability of PSO, it is compared with genetic algorithm (GA). Hence, the two algorithms are compared in terms of objective functions and the standard deviation. Different runs have been taken to get convincing results and the parameters are chosen properly where the Matlab software is used. Where the suggested algorithm can solve different engineering problems with different dimension and outperform the others in term of accuracy and speed of convergence.
The pertinent single-attribute-based classifier for small datasets classific...IJECEIAES
Classifying a dataset using machine learning algorithms can be a big challenge when the target is a small dataset. The OneR classifier can be used for such cases due to its simplicity and efficiency. In this paper, we revealed the power of a single attribute by introducing the pertinent single-attributebased-heterogeneity-ratio classifier (SAB-HR) that used a pertinent attribute to classify small datasets. The SAB-HR’s used feature selection method, which used the Heterogeneity-Ratio (H-Ratio) measure to identify the most homogeneous attribute among the other attributes in the set. Our empirical results on 12 benchmark datasets from a UCI machine learning repository showed that the SAB-HR classifier significantly outperformed the classical OneR classifier for small datasets. In addition, using the H-Ratio as a feature selection criterion for selecting the single attribute was more effectual than other traditional criteria, such as Information Gain (IG) and Gain Ratio (GR).
The importance of employee satisfaction and work motivation is growing all the time in the companies. Many researchers have been made to find out the effect the job satisfaction and motivation have in the productivity of the company. This paper is about the employee satisfaction in a Pharmaceutical company in India. This paper wanted to find out in practice what the level of employee satisfaction in a company is. After the target organization had been found, the research question will composed: what is the level of employee satisfaction in. The main subjects will be leadership and motivation, and the affect they have on employee satisfaction. To find out the results for the research, questionnaires will be delivering to the employees, in the company. The purpose of this kind of research is to find out which factors could be improved in the target company and how to make employees enjoy their work every day.
Keywords- employee satisfaction, work motivation, leadership, Expectation, Wages.
Multimodal Biometrics Recognition by Dimensionality Diminution MethodIJERA Editor
Multimodal biometric system utilizes two or more character modalities, e.g., face, ear, and fingerprint,
Signature, plamprint to improve the recognition accuracy of conventional unimodal methods. We propose a new
dimensionality reduction method called Dimension Diminish Projection (DDP) in this paper. DDP can not only
preserve local information by capturing the intra-modal geometry, but also extract between-class relevant
structures for classification effectively. Experimental results show that our proposed method performs better
than other algorithms including PCA, LDA and MFA.
K-Medoids Clustering Using Partitioning Around Medoids for Performing Face Re...ijscmcj
Face recognition is one of the most unobtrusive biometric techniques that can be used for access control as well as surveillance purposes. Various methods for implementing face recognition have been proposed with varying degrees of performance in different scenarios. The most common issue with effective facial biometric systems is high susceptibility of variations in the face owing to different factors like changes in pose, varying illumination, different expression, presence of outliers, noise etc. This paper explores a novel technique for face recognition by performing classification of the face images using unsupervised learning approach through K-Medoids clustering. Partitioning Around Medoids algorithm (PAM) has been used for performing K-Medoids clustering of the data. The results are suggestive of increased robustness to noise and outliers in comparison to other clustering methods. Therefore the technique can also be used to increase the overall robustness of a face recognition system and thereby increase its invariance and make it a reliably usable biometric modality.
Engineering Research Publication
Best International Journals, High Impact Journals,
International Journal of Engineering & Technical Research
ISSN : 2321-0869 (O) 2454-4698 (P)
www.erpublication.org
We conducted comparative analysis of different supervised dimension reduction techniques by integrating a set of different data splitting algorithms and demonstrate the relative efficacy of learning algorithms dependence of sample complexity. The issue of sample complexity discussed in the dependence of data splitting algorithms. In line with the expectations, every supervised learning classifier demonstrated different capability for different data splitting algorithms and no way to calculate overall ranking of techniques was directly available. We specifically focused the classifier ranking dependence of data splitting algorithms and devised a model built on weighted average rank Weighted Mean Rank Risk Adjusted Model (WMRRAM) for consent ranking of learning classifier algorithms.
Comparison Study of Decision Tree Ensembles for RegressionSeonho Park
Nowadays, decision tree ensemble methods are widely used for solving classification and regression problem due to their rigorousness and robustness. To compare with classification, the performance in regression problem so far has not been yet addressed in detail. In this presentation, we review the state-of-art decision tree ensemble methodology in scikit-learn and xgboost for regression. Also, empirical study results are illustrated to compare their performance and computational efficiency.
For years, the Machine Learning community has focused on developing efficient
algorithms that can produce very accurate classifiers. However, it is often much easier
to find several good classifiers based on dataset combination, instead of single classifier
applied on deferent datasets. The advantages of using classifier dataset combinations
instead of a single one are twofold: it helps lowering the computational complexity by
using simpler models, and it can improve the classification accuracy and performance.
Most Data mining applications are based on pattern matching algorithms, thus improving
the performance of the classification has a positive impact on the quality of the overall
data mining task. Since combination strategies proved very useful in improving the
performance, these techniques have become very important in applications such as
Cancer detection, Speech Technology and Natural Language Processing .The aim of this
paper is basically to propose proprietary metric, Normalized Geometric Index (NGI)
based on the latent properties of datasets for improving the accuracy of data mining tasks.
Clustering technology has been applied in numerous applications. It can enhance the performance
of information retrieval systems, it can also group Internet users to help improve the click-through rate of
on-line advertising, etc. Over the past few decades, a great many data clustering algorithms have been
developed, including K-Means, DBSCAN, Bi-Clustering and Spectral clustering, etc. In recent years, two
new data clustering algorithms have been proposed, which are affinity propagation (AP, 2007) and density
peak based clustering (DP, 2014). In this work, we empirically compare the performance of these two latest
data clustering algorithms with state-of-the-art, using 6 external and 2 internal clustering validation metrics.
Our experimental results on 16 public datasets show that, the two latest clustering algorithms, AP and DP,
do not always outperform DBSCAN. Therefore, to find the best clustering algorithm for a specific dataset, all
of AP, DP and DBSCAN should be considered. Moreover, we find that the comparison of different clustering
algorithms is closely related to the clustering evaluation metrics adopted. For instance, when using the
Silhouette clustering validation metric, the overall performance of K-Means is as good as AP and DP. This
work has important reference values for researchers and engineers who need to select appropriate clustering
algorithms for their specific applications.
Exploratory data analysis using xgboost package in RSatoshi Kato
Explain HOW-TO procedure exploratory data analysis using xgboost (EDAXGB), such as feature importance, sensitivity analysis, feature contribution and feature interaction. It is just based on using built-in predict() function in R package.
All of the sample codes are available at: https://github.com/katokohaku/EDAxgboost
EXTRACTING USEFUL RULES THROUGH IMPROVED DECISION TREE INDUCTION USING INFORM...ijistjournal
Classification is widely used technique in the data mining domain, where scalability and efficiency are the immediate problems in classification algorithms for large databases. We suggest improvements to the existing C4.5 decision tree algorithm. In this paper attribute oriented induction (AOI) and relevance analysis are incorporated with concept hierarchy’s knowledge and HeightBalancePriority algorithm for construction of decision tree along with Multi level mining. The assignment of priorities to attributes is done by evaluating information entropy, at different levels of abstraction for building decision tree using HeightBalancePriority algorithm. Modified DMQL queries are used to understand and explore the shortcomings of the decision trees generated by C4.5 classifier for education dataset and the results are compared with the proposed approach.
EXTRACTING USEFUL RULES THROUGH IMPROVED DECISION TREE INDUCTION USING INFORM...ijistjournal
Classification is widely used technique in the data mining domain, where scalability and efficiency are the immediate problems in classification algorithms for large databases. We suggest improvements to the existing C4.5 decision tree algorithm. In this paper attribute oriented induction (AOI) and relevance analysis are incorporated with concept hierarchy’s knowledge and HeightBalancePriority algorithm for construction of decision tree along with Multi level mining. The assignment of priorities to attributes is done by evaluating information entropy, at different levels of abstraction for building decision tree using HeightBalancePriority algorithm. Modified DMQL queries are used to understand and explore the shortcomings of the decision trees generated by C4.5 classifier for education dataset and the results are compared with the proposed approach.
Using particle swarm optimization to solve test functions problemsriyaniaes
In this paper the benchmarking functions are used to evaluate and check the particle swarm optimization (PSO) algorithm. However, the functions utilized have two dimension but they selected with different difficulty and with different models. In order to prove capability of PSO, it is compared with genetic algorithm (GA). Hence, the two algorithms are compared in terms of objective functions and the standard deviation. Different runs have been taken to get convincing results and the parameters are chosen properly where the Matlab software is used. Where the suggested algorithm can solve different engineering problems with different dimension and outperform the others in term of accuracy and speed of convergence.
The pertinent single-attribute-based classifier for small datasets classific...IJECEIAES
Classifying a dataset using machine learning algorithms can be a big challenge when the target is a small dataset. The OneR classifier can be used for such cases due to its simplicity and efficiency. In this paper, we revealed the power of a single attribute by introducing the pertinent single-attributebased-heterogeneity-ratio classifier (SAB-HR) that used a pertinent attribute to classify small datasets. The SAB-HR’s used feature selection method, which used the Heterogeneity-Ratio (H-Ratio) measure to identify the most homogeneous attribute among the other attributes in the set. Our empirical results on 12 benchmark datasets from a UCI machine learning repository showed that the SAB-HR classifier significantly outperformed the classical OneR classifier for small datasets. In addition, using the H-Ratio as a feature selection criterion for selecting the single attribute was more effectual than other traditional criteria, such as Information Gain (IG) and Gain Ratio (GR).
The importance of employee satisfaction and work motivation is growing all the time in the companies. Many researchers have been made to find out the effect the job satisfaction and motivation have in the productivity of the company. This paper is about the employee satisfaction in a Pharmaceutical company in India. This paper wanted to find out in practice what the level of employee satisfaction in a company is. After the target organization had been found, the research question will composed: what is the level of employee satisfaction in. The main subjects will be leadership and motivation, and the affect they have on employee satisfaction. To find out the results for the research, questionnaires will be delivering to the employees, in the company. The purpose of this kind of research is to find out which factors could be improved in the target company and how to make employees enjoy their work every day.
Keywords- employee satisfaction, work motivation, leadership, Expectation, Wages.
The purpose of paper is to recommend strategies to
increase customer loyalty through complaint management and as
a tool to manage risk. The paper encompasses the theoretical
concepts which emerge from the extensive review of literature on
complaints and risk. It was found that complaints and risk have a
significant relation and through complaint management, risk can
be reduced. The study has proposed COMPSAT Grid (reinforced
with literature review) demonstrating the state of Banks based on
no. of complaints and loyalty of customers. COMPSAT Grid can
become a base to design the strategies to increase customer’s
loyalty. The study is limited to the customer’s perceived risk. The
paper stresses on the importance of complaints in managing the
risk. Through COMPSAT grid the service providers may
modulate existing strategies to increase customer loyalty. The
concepts will establish complaint management as a basis of
marketing strategy modulation. The model is a theoretical
approach which is based on the concepts
We present a case of murder with a blunt object.
On the body of the deceased were identified specific traces in
the form of suffusions. Four people were suspected for the
murder. During the investigative four pairs of sneakers were
obtained. A full forensic examination was performed which
excluded three of the four pairs of suspected shoes. The forth
pair was a match. This kind of research is extremely important
in forensic practice and theory, as they allow forensic experts
to identify the objects that left specific traces on the body of a
victim or deceased. They are also extremely important during
investigation of criminal offenses, especially when there is more
than one suspect.
The present study was conducted at Lucknow
District in Uttar Pradesh. The purpose of this study is to
document how being perform in extra-curricular activities can
influence development in academics, social skills, and high school
completion. In this paper we study the possible influence of
extracurricular activities on student’s performance of eighth-and
ninth graders. 120 students of age group between 13 to 16 years
comprised the sample of the study. Self-made questionnaire for
school students were administered. Data was analyzed in term of
percentage and t-test analysis. The statistical analysis revealed
that all the 6 types of extracurricular activities, viz. Yoga, Horse
riding, Sport activities, Dance, Music and Indoor and outdoor
activities together showed significant role in some extracurricular
activities and Student’s performance of Government and Private
School. Students who participate in extracurricular activities
generally benefit from the many opportunities afforded them.
Benefits of participating in extracurricular activities included
having better grades, having higher standardized test scores and
higher educational attainment, attending school more regularly,
and having higher a higher self-concept. Those who participate in
out-of-school activities often have higher grade point averages, a
decrease in absenteeism, and an increased connectedness to the
school. Finally, we discuss the possible influence of
extracurricular activities on student’s performance and whether
such participation is advisable
Abledate Online Mature Online dating, undoubtedly has made its symbol in the online world. Abledate Online relationship can discover you that special somebody in life. Abledate can switch the typical traditional connections that you have had in the previous. Abledate online dating website is a type of internet dating that is also enjoyable if you know how to make it really unique. Abledate online relationship is presently the hot craze and many persons are turning to our internet relationship sites to find buddies, romance, love and relationships. Abledate Online relationship might be the best solution to many that are dating. Abledate might be less complicated for them to locate somebody special, but Abledate might just be a way to date without a time constraint in USA. Hurry up start dating after signup at Abledate.com
Caterer | Asian Restaurants - Chef Patricks KitchenLance Libot
Chef Patrick’s Kitchen is a local Chinese restaurant that focuses in delighting the palates of each and every guest with various fresh and delicious dishes all at affordable prices.
World Wide Web is large sized repository of
interlinked hypertext documents accessed via the Internet. Web
may contain text, images, video, and other multimedia data. The
user navigates through this using hyperlink. Search Engine gives
millions of results and applies Web mining techniques to order the
results. The sorted order of search results is obtained by applying
some special algorithms called—Page ranking algorithms. The
algorithm measures the importance of the pages by analyzing the
number of inlinked and outlinked pages. Our proposed system is
built on an idea that to rank the relevant pages higher in the
retrieved document set, an analysis of both page‘s text substance
and links information is required. The proposed approach is
based on the assumption that the effective weight of a term in a
page is computed by adding the weight of a term in the current
page and additional weight of the term in the linked pages. In
this chapter, we first study the nature of web pages, the various
link analysis ranking algorithms and their limitations and then
show the comparative analysis of the ranking scores obtained
through these approaches with our new suggested ranking
approach.
Waste is actually the biggest feed stock available for
processing, to produce useful and usable products. The
increasing amount of waste is a characteristic of the modern
human, though this discloses a more luxury life it presents an
environmental hazard that cannot be ignored. Following this
understanding we decided to impact on method of utilizing waste
and converting it into useful and usable products as well as
reducing the nuisance of waste. The methodology followed comes
in steps; first a random trial to detect the use of unsorted waste
and evaluating the equipment design, secondly improvement of
blending of different components of waste, thirdly improvement
of facilities for uniformity of heat and pressure, finally arriving
at suitable formula regarding the ratio of the different waste
components to give uniformity and better hold of the product.
International Journal of Engineering Research and Applications (IJERA) is an open access online peer reviewed international journal that publishes research and review articles in the fields of Computer Science, Neural Networks, Electrical Engineering, Software Engineering, Information Technology, Mechanical Engineering, Chemical Engineering, Plastic Engineering, Food Technology, Textile Engineering, Nano Technology & science, Power Electronics, Electronics & Communication Engineering, Computational mathematics, Image processing, Civil Engineering, Structural Engineering, Environmental Engineering, VLSI Testing & Low Power VLSI Design etc.
Novel Ensemble Tree for Fast Prediction on Data StreamsIJERA Editor
Data Streams are sequential set of data records. When data appears at highest speed and constantly, so predicting
the class accordingly to the time is very essential. Currently Ensemble modeling techniques are growing
speedily in Classification of Data Stream. Ensemble learning will be accepted since its benefit to manage huge
amount of data stream, means it will manage the data in a large size and also it will be able to manage concept
drifting. Prior learning, mostly focused on accuracy of ensemble model, prediction efficiency has not considered
much since existing ensemble model predicts in linear time, which is enough for small applications and
accessible models workings on integrating some of the classifier. Although real time application has huge
amount of data stream so we required base classifier to recognize dissimilar model and make a high grade
ensemble model. To fix these challenges we developed Ensemble tree which is height balanced tree indexing
structure of base classifier for quick prediction on data streams by ensemble modeling techniques. Ensemble
Tree manages ensembles as geodatabases and it utilizes R tree similar to structure to achieve sub linear time
complexity
Proposing an Appropriate Pattern for Car Detection by Using Intelligent Algor...Editor IJCATR
Nowadays, the automotive industry has attracted the attention of consumers, and product quality is considered as an
essential element in current competitive markets. Security and comfort are the main criteria and parameters of selecting a car.
Therefore, standard dataset of CAR involving six features and characteristics and 1728 instances have been used. In this paper, it
has been tried to select a car with the best characteristics by using intelligent algorithms (Random Forest, J48, SVM,
NaiveBayse) and combining these algorithms with aggregated classifiers such as Bagging and AdaBoostMI. In this study, speed
and accuracy of intelligent algorithms in identifying the best car have been taken into account.
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHESVikash Kumar
IMAGE CLASSIFICATION USING KNN, RANDOM FOREST AND SVM ALGORITHM ON GLAUCOMA DATASETS AND EXPLAIN THE ACCURACY, SENSITIVITY, AND SPECIFICITY OF EACH AND EVERY ALGORITHMS
Bank - Loan Purchase Modeling
This case is about a bank which has a growing customer base. Majority of these customers are liability customers (depositors) with varying size of deposits. The number of customers who are also borrowers (asset customers) is quite small, and the bank is interested in expanding this base rapidly to bring in more loan business and in the process, earn more through the interest on loans. In particular, the management wants to explore ways of converting its liability customers to personal loan customers (while retaining them as depositors). A campaign that the bank ran last year for liability customers showed a healthy conversion rate of over 9% success. This has encouraged the retail marketing department to devise campaigns with better target marketing to increase the success ratio with a minimal budget. The department wants to build a model that will help them identify the potential customers who have a higher probability of purchasing the loan. This will increase the success ratio while at the same time reduce the cost of the campaign. The dataset has data on 5000 customers. The data include customer demographic information (age, income, etc.), the customer's relationship with the bank (mortgage, securities account, etc.), and the customer response to the last personal loan campaign (Personal Loan). Among these 5000 customers, only 480 (= 9.6%) accepted the personal loan that was offered to them in the earlier campaign.
Our job is to build the best model which can classify the right customers who have a higher probability of purchasing the loan. We are expected to do the following:
EDA of the data available. Showcase the results using appropriate graphs
Apply appropriate clustering on the data and interpret the output .
Build appropriate models on both the test and train data (CART & Random Forest). Interpret all the model outputs and do the necessary modifications wherever eligible (such as pruning).
Check the performance of all the models that you have built (test and train). Use all the model performance measures you have learned so far. Share your remarks on which model performs the best.
Regularized Weighted Ensemble of Deep Classifiers ijcsa
Ensemble of classifiers increases the performance of the classification since the decision of many experts
are fused together to generate the resultant decision for prediction making. Deep learning is a classification algorithm where along with the basic learning technique, fine tuning learning is done for improved precision of learning. Deep classifier ensemble learning is having a good scope of research.Feature subset selection is another for creating individual classifiers to be fused for ensemble learning. All these ensemble techniques faces ill posed problem of overfitting. Regularized weighted ensemble of deep support vector machine performs the prediction analysis on the three UCI repository problems IRIS,Ionosphere and Seed data set, thereby increasing the generalization of the boundary plot between the
classes of the data set. The singular value decomposition reduced norm 2 regularization with the two level
deep classifier ensemble gives the best result in our experiments.
The International Journal of Engineering and Science (The IJES)theijes
The International Journal of Engineering & Science is aimed at providing a platform for researchers, engineers, scientists, or educators to publish their original research results, to exchange new ideas, to disseminate information in innovative designs, engineering experiences and technological skills. It is also the Journal's objective to promote engineering and technology education. All papers submitted to the Journal will be blind peer-reviewed. Only original articles will be published.
sis of health condition is very challenging task for every human being because life is directly related to health
condition. Data mining based classification is one of the important applications for classification of data. In this
research work, we have used various classification techniques for classification of thyroid data. CART gives highest
accuracy 99.47% as best model. Feature selection plays very important role to computationally efficient and increase
the performance of model. This research work focus on Info Gain and Gain Ratio feature selection technique to
reduce the irrelevant features from original data set and computationally increase the performance of model. We have
applied both the feature selection techniques on best model i. e. CART. Our proposed CART-Info Gain and CARTGain
Ratio gives 99.47% and 99.20% accuracy with 25 and 3 feature respectively.
Recent joint surgery studies reveal increased
revisions and resurfacing of the metal on metal hip joints. Metal
on metal hip implants were developed more than thirty years ago
and their application has been refined because of availability of
advanced manufacturing techniques and partly by advancements
in material science and engineering. Development of composite
materials may provide greater durability to metal-on-metal hip
implants .This review article is a study of the latest literature of
metal-on-metal hip implants and its various modeling techniques.
Numbers of methods are used for convergence and numerical
solution to investigate the performance of metal-on-metal hip
implant for accurate stable solution. This paper presents analysis
done by various researchers on metal-on-metal hip implants for
wear, lubrication, fatigue, bio-tribo-corrosion, design, toxicity
and resurfacing. After in vivo and in vitro studies, it is found that
all these methods have limitations. There is a need of more
insight for lubrication analysis, geometry of bearings, materials
and input parameters. The information provided in this work is
intended as an aid in the assessment of metal-on-metal hip joints.
Background Hospital contributes significantly tangible and intangible resources on a concurred plan by the scheduling of surgery on the OT list. Postponement decreases efficiency by declining throughput leads to wastage of resources hence burden to the nation. Patients and their family face economic and emotional implication due to the postponement. Postponement rate being a quality indicator controls check mechanism could be developed from the results. Postponement of elective scheduled operations results in inefficient use of the operating room (OR) time on the day of surgery. Inconvenience to patients and families are also caused by postponements. Moreover, the day of surgery (DOS) postponement creates logistic and financial burden associated with extended hospital stay and repetitions of pre-operative preparations to an extent of repetition of investigations in some cases causing escalated costs, wastage of time and reduced income. Methodology A cross-sectional study was done in the operation theaters of a tertiary care hospital in which total ten operation theaters of General Surgery Data of scheduled, performed and postponed surgeries was collected from all the operation theater with effect from March 1st to September 30th, 2018. A questionnaire was developed to find out the reasons for the postponement for all hospital’s stakeholders (surgeons, Anesthetist, Nursing Officer) and they were further evaluated time series analysis of scheduling of Operation Theater for moving average technique. Results Total 958 surgeries were scheduled and 772 surgeries performed were and 186 surgeries were postponed with a postponement rate of 19.42% in the cardiac surgery department during the study period. Month-wise postponement Rate exponential smoothing of time series data shows the dynamic of operating suits. To test throughput Postponement rate was plotted the postponed surgeries and on regression analysis is in a perfect linear relationship.
Introduction: Postponement of elective scheduled operations results in inefficient use of operating room (OR) time on the day of surgery. Inconvenience to patients and families also caused by postponements. Moreover, day of surgery (DOS) postponement creates logistic and financial burden associated with extended hospital stay and repetitions of pre-operative preparations to an extend of repetition of investigations in some cases causing escalated costs, wastage of time and reduced income. Methodology: A cross sectional study was done in the operation theaters of a tertiary care hospital in which total ten operation theaters of General Surgery Data of scheduled, performed and postponed surgeries was collected from all the operation theater with effect from march 1st to September 30th 2018. A questionnaire was developed to find out the reasons for the postponement for all hospital’s stakeholders (Surgeons, Anesthetist, Nursing officer) and they were further evaluated Time series analysis of scheduling of Operation Theater for Moving average Technique. Results: total 2,466 surgeries were scheduled and 1,980 surgeries were performed and 486 surgeries were postponed in the general surgery department during the study period. Month wise postponement forecast was in accordance with the performed surgeries and on regression analysis postponed surgeries were in perfect linear relationship with the postponement Rate.
In the present paper the experimental study of
Nanotechnology involves high cost for Lab set-up and the
experimentation processes were also slow. Attempt has also
been made to discuss the contributions towards the societal
change in the present convergence of Nano-systems and
information technologies. one cannot rely on experimental
nanotechnology alone. As such, the Computer- simulations and
modeling are one of the foundations of computational
nanotechnology. The computer modeling and simulations
were also referred as computational experimentations. The
accuracy of such Computational nano-technology based
experiment generally depends on the accuracy of the following
things: Intermolecular interaction, Numerical models and
Simulation schemes used. The essence of nanotechnology is
therefore size and control because of the diversity of
applications the plural term nanotechnology is preferred by
some nevertheless they all share the common feature of control
at the nanometer scale the latter focusing on the observation
and study of phenomena at the nanometer scale. In this paper,
a brief study of Computer-Simulation techniques as well as
some Experimental result
Solar cell absorber Kesterite- type Cu2ZnSnS4 (CZTS) thin films have been prepared by Chemical Bath Deposition (CBD). UV–vis absorption spectra measurement indicated that the band gap of as-synthesized CZTS was about1.68 eV, which was near the optimum value for photovoltaic solar conversion in a single-band-gap device. The polycrystalline CZTS thin films with kieserite crystal structure have been obtained by XRD. The average of crystalline size of CZTS is 27 nm
Multilevel inverters play a crucial part in the
areas of high and medium voltage applications. Among the three
main multilevel inverters used, the capacitor clamped multilevel
inverter(CCMLI) has advantage with respect to voltage
redundancies. This work proposes a switching pattern to improve
the performance of chosen H-bridge type CCMLI over
conventional CCMLI. The PWM technique used in this work is
Phase Opposition Disposition PWM(PODPWM). The
performance of proposed H-bridge type CCMLI is verified
through MATLAB-Simulink based simulation. It has been
observed that the THD is low in chosen CCMLI compared to
conventional CCMLI.
- In this paper, we introduce a practical mechanism of
compressing a binary phase code modulation (BPCM) signal
according to Barker code with 13 chips in presence of additive
white Gaussian noise (AWGN) by using a digital matched filter
(DMF) corresponding to time domain convolution algorithm of
input and reference signals using Cyclone II EP2C70F896C6
FPGA from ALTERA placed on education and development
board DE2-70 with the following parameters: frequency of
BPCM signal fIF=2 MHz, sampling frequency
f MHz SAM 50
,pulse period
T 200s
, pulse width
S 13sc
, chip width
CH 1sc
, compressing factor
KCOM 13
, SNRinp=1/1, 1/2, 1/3, 1/4, 1/5 and processing
gain factor SNRout/SNRinp=11.14 dB.
The results of filter operation are evaluated using a digital
oscilloscope GDS-1052U to display the input and output signals
for different SNRinp.
Flooding is one of the most devastating natural
disasters in Nigeria. The impact of flooding on human activities
cannot be overemphasized. It can threaten human lives, their
property, environment and the economy. Different techniques
exist to manage and analyze the impact of flooding. Some of these
techniques have not been effective in management of flood
disaster. Remote sensing technique presents itself as an effective
and efficient means of managing flood disaster. In this study,
SPOT-10 image was used to perform land cover/ land use
classification of the study area. Advanced Space borne Thermal
Emission and Reflection Radiometer (ASTER) image of 2010 was
used to generate the Digital Elevation Model (DEM). The image
focal statistics were generated using the Spatial Analyst/
Neighborhood/Focal Statistics Tool in ArcMap. The contour map
was produced using the Spatial Analyst/ Surface/ Contour Tools.
The DEM generated from the focal statistics was reclassified into
different risk levels based on variation of elevation values. The
depression in the DEM was filled and used to create the flow
direction map. The flow accumulation map was produced using
the flow direction data as input image. The stream network and
watershed were equally generated and the stream vectorized. The
reclassified DEM, stream network and vectorized land cover
classes were integrated and used to analyze the impact of flood on
the classes. The result shows that 27.86% of the area studied will
be affected at very high risk flood level, 35.63% at high risk,
17.90% at moderate risk, 10.72% at low risk, and 7.89% at no
risk flood level. Built up area class will be mostly affected at very
high risk flood level while farmland will be affected at high risk
flood level. Oshoro, Imhekpeme, and Weppa communities will be
affected at very high risk flood inundation while Ivighe, Uneme,
Igoide and Iviari communities will be at risk at high risk flood
inundation level. It is recommended among others that buildings
that fall within the “Very High Risk” area should be identified
and occupants possibly relocated to other areas such as the “No
Risk” area.
Without water, humans cannot live. Since time began,
we have lived by the water and vast tracts of waterless land have
been abandoned as it is too difficult to inhabit. At any given
moment, the earth’s atmosphere contains 4,000 cubic miles of
water, which is just 0.000012% of the 344 million cubic miles of
water on earth. Nature maintains this ratio via evaporation and
condensation, irrespective of the activities of man.
There is a certain need for an alternative to solve the water
scarcity. Obtaining water from the atmosphere is nothing new -
since the beginning of time, nature’s continuous hydrologic cycle
of evaporation and condensation in the form of rain or snow has
been the sole source and means of regenerating wholesome water
for all forms of life on earth.
An effective method to generate water is by the separation of
moisture present in air by condensation. In this study, the water
present in air is condensed on the surface of a container and then
collected in an external jacket provided on the container.
Insulations are provided to optimize the inner temperature of the
container.
The method is although uncommon but has certain advantages
which make it a success. The process is economical and does not
require a lot of utilities. It also helps in further reducing the
carbon footprint.
In every moment of functioning the Li-Ion
battery must provide the power required by the user, to have a
long operating life and to and to provide high reliability in
operation. The methods for analysis and testing batteries are
ensuring that all these conditions imposed to the batteries are
met by being tested depending on their intended use.
The success rate of real estate project is
decreasing as there is large scale of project and participation of
entities. It is necessary to study the risk factors involved in the
project. This paper focused on types of risks involved in the
project, risk factors, risk management tools & techniques.
Identification of risk of the project in terms of the total cost of the
project has been divided under Technical, Financial, Sociopolitical
and Statutory cost centers. Large real estate projects
have to tackle the following issues: land acquisition, skilledlabour
shortage, non-availability of skilled project managers, and
mechanization of the construction process to cater to the growing
demands. Non- availability of supporting infrastructure, political
issues like instability of the government leading to regulatory
issues, social issues, marketing forms an important part in these
projects as this is a onetime investment and the purchase cycle is
long , long development period makes the same project be at
different points in the real estate value cycle.
- In the present scenario carbon emission and sand
mining are major concern due to its hazardous effect to
environment and making serious imbalance to the ecosystem.
Various studies have been conducted to reduce severe effect on
environment, using byproducts like copper slag as partial
replacement of fine aggregate. Different researchers have also
revealed numerous uses of copper slag as a replacing agent in
determining the strength of concrete. A comprehensive review of
studies has been presented in this paper for scope of replacement
of fine aggregate from copper slag in concrete
- Security is a concept similar to being cautious
or alert against any danger. Network security is the condition of
being protected against any danger or loss. Thus safety plays a
important role in bank transactions where disclosure of any data
results in big loss. We can define networking as the combination
of two or more computers for the purpose of resource sharing.
Resources here include files, database, emails etc. It is the
protection of these resources from unauthorized users that
brought the development of network security. It is a measure
incorporated to protect data during their transmission and also
to ensure the transmitted is protected and authentic.
Security of online bank transactions here has been
improved by increasing the number of bits while establishing the
SSL connection as well as in RSA asymmetric key encryption
along with SHA1 used for digital signature to authenticate the
user
Background: Septoplasty is a common surgical
procedure performed by otolaryngologists for the correction of
deviated nasal septum. This surgery may be associated with
numerous complications. To minimize these complications,
otolaryngologists frequently pack both nasal cavities with
different types of nasal packing. Despite all its advantages,
nasal packing is also associated with some disadvantages. To
avoid these issues, many surgeons use suturing techniques to
obviate the need for packing after surgery.
Objective: To determine the efficacy and safety of trans-septal
suture technique in preventing complications and decreasing
morbidity after septoplasty in comparison with nasal packing.
Patients and methods: Prospective comparative study. This
study was conducted in the department of Otolaryngology -
Head and Neck Surgery, Rizgary Teaching Hospital - Erbil,
from the 6th of May 2014 to the 30th of November 2014.
A total of 60 patients aged 18-45 years, undergoing septoplasty,
were included in the study. Before surgery, patients were
randomly divided into two equal groups. Group (A) with transseptal
suture technique was compared with group (B) in which
nasal packing with Merocel was done. Postoperative morbidity
in terms of pain, bleeding, postnasal drip, sleep disturbance,
dysphagia, headache and epiphora along with postoperative
complications including septal hematoma, septal perforation,
crustation and synechiae formation were assessed over a follow
up period of four weeks.
Results: Out of 60 patients, 37 patients were males (61.7%)
and 23 patients were females (38.3%). Patients with nasal
packing had significantly more postoperative pain (P<0.05)><0.05). There was no significant difference between
the two groups with respect to nasal bleeding, septal
hematoma, septal perforation, crustation and synechiae
formation.
Conclusion: Septoplasty can be safely performed using transseptal
suturing technique without nasal packing.
The basic reason behind the need to
monitor water quality is to verify whether the examined
water quality is suitable for intended usage or not. This
study is conducted on Al -Shamiya al- sharqi drain in
Diwaniya city in Iraq to make valid assessment for the
level of parameters measured and to realize their effects
on irrigation. In order to assess the drainage water
quality for irrigation purposes with a high accuracy, the
Irrigation Water Quality Index (IWQI) will be examined
and upgraded (integrated with GIS) to make a
classification for drainage water. For this purpose, ten
samples of drainage water were taken from different ten
location of the stuay area. The collected samples were
analyzed chemically for different elements which affect
water quality for irrigation.These elements are :
Calcium(Ca+2), Sodium(Na+
), Magnesium(Mg+2),
Chloride( ), Potassium(K+
), Bicarbonate(HCO3),
Nitrate(NO3), Sulfate( , Phosphate( , Electrical
Conductivity(EC), Total Dissolved Solids (TDS), Total
Suspended Solids (TSS) and pH-values (PH). Sodium
Adsorption Ratio (SAR) and Sodium Content (Na%)
have been also calculated. Results suggest that, the use of
GIS and Water Quality Index (WQI) methods could
provide an extremely interesting as well as efficient tool
to water resource management. The results analysis of
(IWQI) maps confirms that: 52% of the drainage water
in study area falls within the "Low restriction" (LR) and
47%of study area has water with (Moderate
restriction)(MR),While 1% of drainage water in the
study area classified as (Sever restriction) (SR). So, the
drainage water should be used with the soil having high
permeability with some constraints imposed on types of
plant for specified tolerance of salts
The cable-hoisting method and rail cable-lifting
method are widely used in the construction of suspension bridge.
This paper takes a suspension bridge in Hunan as an example,
and expounds the two construction methods, and analyzes their
respective merits and disadvantages.
Baylis-Hillman reaction has been achieved on
different organic motifs but with completion times of three to
six days. Micellar medium of CTAB in water along with the
organic base DABCO has been used to effect the BaylisHillman
reaction on a steroidal nucleus of Withaferin-A for the
first time with different aromatic aldehydes within a day to
synthesize a library of BH adducts (W1a –W14a) and (W1bW14b)
as a mixture of two isomers and W15 as a single
compound. The isomers were separated on column and the
major components were chosen for bio-evaluation. Cytotoxic
activity of the synthesized compounds was screened against a
panel of four cancer cell lines Lung A-549, Breast MCF-7,
Colon HCT-116 and Leukemia THP-1 along with 5-florouracil
and Mitomycin-C as references. All the compounds exhibited
promising activity against screened cell lines and were found to
possess enhaunced activity than parent compound. BH adducts
with aromatic systems having methoxy and nitro groups were
found to be more active.
This paper presents the details on the
experimental investigation carried out to get the desired fresh
properties of the SCC. Tests were performed on various mixtures
to obtain the required SCC. In the present research work we
have replaced 15% of cement with class F fly ash. By varying the
quantity of water and sand the mortar mix was prepared. Later
varying percentage of coarse aggregate was added to the mortar
to obtain the desired SCC.
The batteries used in electric and hybrid vehicles
consists of several cells with voltages between 3.6V battery and
4.2 V in series or parallel combinations of configurations for
obtaining the necessary available voltages in the operation of a
hybrid electric vehicle. How malfunction of a single cell affects
the behavior of the entire battery pack, BMS main function is to
protect individual cells against over-discharge, overload or
overheating. This is done by correct balancing of the cells. In
addition BMS estimates the battery charge status
This project aims at using (PD-MCPWM) Phase
disposition multi carrier pulse width modulation technique to
reduce leakage current in a transformerless cascaded multilevel
inverter for PV systems. Advantages of transformerless PV
inverter topology is as follows, simple structure, low weight and
provides higher efficiency , but however this topology provides a
path for the leakage current to flow through the parasitic
capacitance formed between the PV module and the ground.
Modulation technique reduces leakage current with an added
advantage without adding any extra components.
More from International Journal of Technical Research & Application (20)
DevOps and Testing slides at DASA ConnectKari Kakkonen
My and Rik Marselis slides at 30.5.2024 DASA Connect conference. We discuss about what is testing, then what is agile testing and finally what is Testing in DevOps. Finally we had lovely workshop with the participants trying to find out different ways to think about quality and testing in different parts of the DevOps infinity loop.
PHP Frameworks: I want to break free (IPC Berlin 2024)Ralf Eggert
In this presentation, we examine the challenges and limitations of relying too heavily on PHP frameworks in web development. We discuss the history of PHP and its frameworks to understand how this dependence has evolved. The focus will be on providing concrete tips and strategies to reduce reliance on these frameworks, based on real-world examples and practical considerations. The goal is to equip developers with the skills and knowledge to create more flexible and future-proof web applications. We'll explore the importance of maintaining autonomy in a rapidly changing tech landscape and how to make informed decisions in PHP development.
This talk is aimed at encouraging a more independent approach to using PHP frameworks, moving towards a more flexible and future-proof approach to PHP development.
LF Energy Webinar: Electrical Grid Modelling and Simulation Through PowSyBl -...DanBrown980551
Do you want to learn how to model and simulate an electrical network from scratch in under an hour?
Then welcome to this PowSyBl workshop, hosted by Rte, the French Transmission System Operator (TSO)!
During the webinar, you will discover the PowSyBl ecosystem as well as handle and study an electrical network through an interactive Python notebook.
PowSyBl is an open source project hosted by LF Energy, which offers a comprehensive set of features for electrical grid modelling and simulation. Among other advanced features, PowSyBl provides:
- A fully editable and extendable library for grid component modelling;
- Visualization tools to display your network;
- Grid simulation tools, such as power flows, security analyses (with or without remedial actions) and sensitivity analyses;
The framework is mostly written in Java, with a Python binding so that Python developers can access PowSyBl functionalities as well.
What you will learn during the webinar:
- For beginners: discover PowSyBl's functionalities through a quick general presentation and the notebook, without needing any expert coding skills;
- For advanced developers: master the skills to efficiently apply PowSyBl functionalities to your real-world scenarios.
JMeter webinar - integration with InfluxDB and GrafanaRTTS
Watch this recorded webinar about real-time monitoring of application performance. See how to integrate Apache JMeter, the open-source leader in performance testing, with InfluxDB, the open-source time-series database, and Grafana, the open-source analytics and visualization application.
In this webinar, we will review the benefits of leveraging InfluxDB and Grafana when executing load tests and demonstrate how these tools are used to visualize performance metrics.
Length: 30 minutes
Session Overview
-------------------------------------------
During this webinar, we will cover the following topics while demonstrating the integrations of JMeter, InfluxDB and Grafana:
- What out-of-the-box solutions are available for real-time monitoring JMeter tests?
- What are the benefits of integrating InfluxDB and Grafana into the load testing stack?
- Which features are provided by Grafana?
- Demonstration of InfluxDB and Grafana using a practice web application
To view the webinar recording, go to:
https://www.rttsweb.com/jmeter-integration-webinar
Builder.ai Founder Sachin Dev Duggal's Strategic Approach to Create an Innova...Ramesh Iyer
In today's fast-changing business world, Companies that adapt and embrace new ideas often need help to keep up with the competition. However, fostering a culture of innovation takes much work. It takes vision, leadership and willingness to take risks in the right proportion. Sachin Dev Duggal, co-founder of Builder.ai, has perfected the art of this balance, creating a company culture where creativity and growth are nurtured at each stage.
Dev Dives: Train smarter, not harder – active learning and UiPath LLMs for do...UiPathCommunity
💥 Speed, accuracy, and scaling – discover the superpowers of GenAI in action with UiPath Document Understanding and Communications Mining™:
See how to accelerate model training and optimize model performance with active learning
Learn about the latest enhancements to out-of-the-box document processing – with little to no training required
Get an exclusive demo of the new family of UiPath LLMs – GenAI models specialized for processing different types of documents and messages
This is a hands-on session specifically designed for automation developers and AI enthusiasts seeking to enhance their knowledge in leveraging the latest intelligent document processing capabilities offered by UiPath.
Speakers:
👨🏫 Andras Palfi, Senior Product Manager, UiPath
👩🏫 Lenka Dulovicova, Product Program Manager, UiPath
Software Delivery At the Speed of AI: Inflectra Invests In AI-Powered QualityInflectra
In this insightful webinar, Inflectra explores how artificial intelligence (AI) is transforming software development and testing. Discover how AI-powered tools are revolutionizing every stage of the software development lifecycle (SDLC), from design and prototyping to testing, deployment, and monitoring.
Learn about:
• The Future of Testing: How AI is shifting testing towards verification, analysis, and higher-level skills, while reducing repetitive tasks.
• Test Automation: How AI-powered test case generation, optimization, and self-healing tests are making testing more efficient and effective.
• Visual Testing: Explore the emerging capabilities of AI in visual testing and how it's set to revolutionize UI verification.
• Inflectra's AI Solutions: See demonstrations of Inflectra's cutting-edge AI tools like the ChatGPT plugin and Azure Open AI platform, designed to streamline your testing process.
Whether you're a developer, tester, or QA professional, this webinar will give you valuable insights into how AI is shaping the future of software delivery.
Transcript: Selling digital books in 2024: Insights from industry leaders - T...BookNet Canada
The publishing industry has been selling digital audiobooks and ebooks for over a decade and has found its groove. What’s changed? What has stayed the same? Where do we go from here? Join a group of leading sales peers from across the industry for a conversation about the lessons learned since the popularization of digital books, best practices, digital book supply chain management, and more.
Link to video recording: https://bnctechforum.ca/sessions/selling-digital-books-in-2024-insights-from-industry-leaders/
Presented by BookNet Canada on May 28, 2024, with support from the Department of Canadian Heritage.
Slack (or Teams) Automation for Bonterra Impact Management (fka Social Soluti...Jeffrey Haguewood
Sidekick Solutions uses Bonterra Impact Management (fka Social Solutions Apricot) and automation solutions to integrate data for business workflows.
We believe integration and automation are essential to user experience and the promise of efficient work through technology. Automation is the critical ingredient to realizing that full vision. We develop integration products and services for Bonterra Case Management software to support the deployment of automations for a variety of use cases.
This video focuses on the notifications, alerts, and approval requests using Slack for Bonterra Impact Management. The solutions covered in this webinar can also be deployed for Microsoft Teams.
Interested in deploying notification automations for Bonterra Impact Management? Contact us at sales@sidekicksolutionsllc.com to discuss next steps.
Knowledge engineering: from people to machines and back
A STUDY OF DECISION TREE ENSEMBLES AND FEATURE SELECTION FOR STEEL PLATES FAULTS DETECTION
1. International Journal of Technical Research and Applications e-ISSN: 2320-8163,
www.ijtra.com Volume 2, Issue 4 (July-Aug 2014), PP 127-131
127 | P a g e
A STUDY OF DECISION TREE ENSEMBLES
AND FEATURE SELECTION FOR STEEL
PLATES FAULTS DETECTION
Sami M. Halawani
Faculty of Computing and Information Technology,
King Abdulaziz University, Jeddah, Saudi Arabia.
Abstract— The automation of fault detection in material
science is getting popular because of less cost and time. Steel
plates fault detection is an important material science problem.
Data mining techniques deal with data analysis of large data.
Decision trees are very popular classifiers because of their simple
structures and accuracy. A classifier ensemble is a set of
classifiers whose individual decisions are combined in to classify
new examples. Classifiers ensembles generally perform better
than single classifier. In this paper, we show the application of
decision tree ensembles for steel plates faults prediction. The
results suggest that Random Subspace and AdaBoost.M1 are the
best ensemble methods for steel plates faults prediction with
prediction accuracy more than 80%. We also demonstrate that if
insignificant features are removed from the datasets, the
performance of the decision tree ensembles improve for steel
plates faults prediction. The results suggest the future
development of steel plate faults analysis tools by using decision
tree ensembles.
Index Terms— Material informatics, Steel plates faults, Data
mining, Decision trees, Ensembles.
I. INTRODUCTION
Materials informatics [1] is a field of study that applies the
principles of data mining techniques to material science. It is
very important to predict the material behavior correctly. As
we cannot do the very large number of experiments without
high cost and time, the prediction of material properties by
computer methods is gaining ground. A fault is defined as an
abnormal behavior. Defects or Faults detection is an important
problem in material science [2]. The timely detection of faults
may save lot of time and money. The steel industry is one of
the areas which have fault detection problem. The task is to
detect the type of defects, steel plates have. Some of the defects
are Pastry, Z-Scratch, K-Scatch, Stains, Dirtiness, and Bumps
etc. One of the tradition methods is to have manual inspection
of each steel plate to find out the defects. This method is time
consuming and need lot of efforts. The automation of fault
detection technique is emerging as a powerful technique for
fault detection [3]. This process relies on data mining
techniques [4] to predict the fault detection. These techniques
use past data (the data consist of features and the output that is
to be predicted by using features) to construct models (which
are also called classifiers) and these models are used to predict
the faults.
Classifiers ensembles are popular data mining techniques
[5]. Classifier ensembles are combination of base models; the
decision of each base model is combined to get the final
decision. A decision tree [6] is very popular classifier which
has been successfully used for various applications. Decision
tree ensembles have been very accurate classifiers and have
shown excellent performance in different applications [7].
It has also been shown that not all features that are used for
prediction are useful [8]. Removing some of the insignificant
features may improve the performance of the classifiers.
Hence, it is important to analyze the data to remove
insignificant features.
Various data mining techniques [9, 10] have been used to
predict the steel plate faults, however, there is no detailed study
to show the performance of different kinds of decision tree
ensembles for predicting steel plate faults. The advantage of
the decision tree ensembles is that they give accurate
predictions [11]. Hence, we expect that decision tree ensembles
will perform well for steel plate fault predictions. As there has
been no study to see the effect of removing insignificant
features on prediction accuracy for steel plate faults, it will be
useful to do this study for better prediction results. There are
two objectives of the paper;
1- There are many decision tree ensemble methods. In this
paper we will find out which method is the best for steel plates
prediction problem.
2- To investigate the effect of removing insignificant
features on the prediction accuracy for decision tree ensembles
for steel plates faults.
II. RELATED WORKS
In this section, we will discuss various data mining
techniques that were used in this paper.
Decision trees
Decision trees are very popular tools for classification as
they produce rules that can be easily analyzed by human
2. International Journal of Technical Research and Applications e-ISSN: 2320-8163,
www.ijtra.com Volume 2, Issue 4 (July-Aug 2014), PP 127-131
128 | P a g e
beings. A decision tree can be used to classify an example by
starting at the root of the tree and moving through it until a leaf
node, which provides the rules for classification of the
example. There are various methods to produce decision trees,
however C4.5 [6] and CART trees [12] are the most popular
decision tree methods.
Classifier Ensembles
Ensembles [5] are a combination of multiple base models;
the final classification depends on the combined outputs of
individual models. Classifier ensembles have shown to produce
better results than single models, provided the classifiers are
accurate and diverse.
Several methods have been proposed to build decision tree
ensembles. In these methods, randomization is introduced to
build diverse decision trees. Bagging [13] and Boosting [14]
introduce randomization by manipulating the training data
supplied to each classifier. Ho [15] proposes Random
Subspaces that selects random subsets of input features for
training an ensemble. Breiman [16] combines Random
Subspaces technique with Bagging to create Random Forests.
To build a tree, it uses a bootstrap replica of the training
sample, then during the tree growing phase, at each node the
optimal split is selected from a random subset of size K of
candidate features, then during the tree growing phase, at each
node the optimal split is selected from a random subset of size
K of candidate features. We will discuss these techniques in
detail.
Random Subspaces
Ho [15] presents Random Subspaces (RS) ensembles. In
this method, diverse datasets are created by selecting random
subsets of a feature space. Each decision tree in an ensemble
learns on one dataset from the pool of different datasets.
Results of these trees are combined to get the final result. This
simple method is quite competitive to other ensemble methods.
Experiments suggest that RS is good when there is certain
redundancy in features. For datasets where there is no
redundancy, redundancy needs to be introduced artificially by
concatenating new features that are linear combinations of
original features to the original features and treating this as the
data.
Bagging
Bagging [13] generates different bootstrap training datasets
from the original training dataset and uses each of them to train
one of the classifiers in the ensemble. For example, to create a
training set of N data points, it selects one point from the
training dataset, N times without replacement. Each point has
equal probability of selection. In one training dataset, some of
the points get selected more than once, whereas some of them
are not selected at all. Different training datasets are created by
this process. When different classifiers of the ensemble are
trained on different training datasets, diverse classifiers are
created. Bagging does more to reduce the variance part of the
error of the base classifier than the bias part of the error.
Adaboost.M1
Boosting [14] generates a sequence of classifiers with different
weight distribution over the training set. In each iteration, the
learning algorithm is invoked to minimize the weighted error,
and it returns a hypothesis. The weighted error of this
hypothesis is computed and applied to update the weight on the
training examples. The final classifier is constructed by a
weighted vote of the individual classifiers. Each classifier is
weighted according to its accuracy on the weighted training set
that it has trained on. The key idea, behind Boosting is to
concentrate on data points that are hard to classify by
increasing their weights so that the probability of their
selection in the next round is increased. In subsequent iteration,
therefore, Boosting tries to solve more difficult learning
problems. Boosting reduces both bias and variance parts of the
error. As it concentrates on hard to classify data points, this
leads to the decrease in the bias. At the same time classifiers
are trained on different training data sets so it helps in reducing
the variance. Boosting has difficulty in learning when the
dataset is noisy.
Random Forests
Random Forests [16] are very popular decision tree
ensembles. It combines bagging with random subspace. For
each decision tree, a dataset is created by bagging procedure.
During the tree growing phase, at each node, k attributes are
selected randomly and the node is split by the best attribute
from these k attributes. Breiman [16] shows that Random
Forests are quite competitive to Adaboost. However, Random
Forests can handle mislabeled data points better than Adaboost
can. Due to its robustness of the Random Forests, they are
widely used.
Feature Selection
Not all features are useful for prediction [17, 18]. It is
important to find out insignificant features that may cause error
in prediction. There are many methods for feature selection,
however, Relief [19] is a very popular feature selection
method. Relief is based on feature estimation. Relief assigns a
value of relevance to each feature. All features with higher
values than the user given threshold value are selected.
III. RESULTS AND DISCUSSION
All the experiments were carried out by using WEKA
software [20]. We did the experiments with Random Subspace,
Bagging, AdaBoost.M1 and Random Forests modules. For the
Random subspace, Bagging and AdaBoost.M1 modules, we
carried out experiments with J48 tree (the implementation of
C4.5 tree). The size of the ensembles was set to 50. All the
other default parameters were used in the experiments. We also
carried out experiment with single J48 tree. Relief module was
used for selecting most important features. 10-cross fold
strategy was used in the experiments. The experiments were
done on the dataset taken from UCI repository [21]. The
3. International Journal of Technical Research and Applications e-ISSN: 2320-8163,
www.ijtra.com Volume 2, Issue 4 (July-Aug 2014), PP 127-131
129 | P a g e
dataset has 27 independent features and 7 classes; Pastry, Z-
Scratch, K-Scatch, Stains, Dirtiness, and Bumps and Other-
Deffects. The information about the features is provided in
Table 1. We compared the prediction error of all the methods.
Low error suggests better performance.
Table 1. – The information about the features of steel plates
Feature
Number
Feature Name Feature
Number
Feature Name
1 X_Minimum 15 Edges_Index
2 X_Maximum 16 Empty_Index
3 Y_Minimum 17 Square_Index
4 Y_Maximum 18 Outside_X_Inde
x
5 Pixels_Areas 19 Edges_X_Index
6 X_Perimeter 20 Edges_Y_Index
7 Y_Perimeter 21 Outside_Global_In
dex
8 Sum_of_Luminosity 22 LogOfAreas
9 Minimum_of_Lumin
osity
23 Log_X_Index
10 Maximum_of_Lumin
osity
24 Log_Y_Index
11 Length_of_Conveyer 25 Orientation_Index
12 TypeOfSteel_A300 26 Luminosity_Index
13 TypeOfSteel_A400 27 SigmoidOfAreas
14 Steel_Plate_Thicknes
s
Table 2- Classification error of various methods with all the
features used for the training and prediction.
Classification Method Classification Error in %
Random Subspace 19.62
Bagging 19.93
AdaBoost.M1 19.83
Random Forests 20.76
Single Tree 23.95
Table 3- Classification error of various methods with 20
most important features, calculated from Relief method, used
for the training and prediction.
Classification Method Classification Error in %
Random Subspace 20.60
Bagging 19.19
AdaBoost.M1 18.08
Random Forests 20.04
Single Tree 23.13
Table 4- Classification error of various methods with
15 most important features, calculated from Relief method,
used for the training and prediction.
Classification Method Classification Error in %
Random Subspace 21.32
Bagging 21.70
AdaBoost.M1 21.38
Random Forests 22.35
Single Tree 24.67
19
19.5
20
20.5
21
21.5
22
0 10 20 30
PredictionErrorin%
Number of Features
Fig. 1. Prediction error vs number of features graph for
Bagging method.
17
18
19
20
21
22
0 10 20 30
PredictionErrorin%
Number of Features
4. International Journal of Technical Research and Applications e-ISSN: 2320-8163,
www.ijtra.com Volume 2, Issue 4 (July-Aug 2014), PP 127-131
130 | P a g e
Fig. 2. Prediction error vs number of features graph for
AdaBoot.M1 method.
Discussion
The results with all features, 20 most important features and 15
most important features are shown in Table 2, Table 3 and
Table 4 respectively. Results suggest that with all features,
Random Subspace performed best. Adaboost.M1 came second.
Results for all the classifier methods except Random Subspace
improved when the best 20 features were selected. For
example, AdaBoost.M1 had 18.08 % classification error with
15 features, whereas with all features the error was 19.83 %. It
suggests that some of the features were insignificant so
removing those features improved the performance. As
discussed for Random Forests, the ensemble method is useful
when we have large number of features, hence removing
features had adverse effect on the performance. With 10 most
important features, the performance of all the classification
methods degraded, it suggests that we removed some of the
significant features from the datasets that had adverse effect on
the performance of the classification methods. Results for
Bagging and AdaBoost.M1 are presented in Fig. 1 and Fig. 2
respectively. These figures show that by removing the
insignificant features the error first decreases and then
increases. This suggests that the dataset has more than 15
important features.
The best performance is achieved with AdaBoost.M1
(18.08 % error) with 20 features. It shows that a good
combination of ensemble method and feature selection method
can be useful for steel plates defects. Results suggest that all
the ensembles method performed better than single trees. This
demonstrates the efficacy of the decision tree ensemble
methods for steel plates defects problem.
IV. CONCLUSION AND FUTURE WORK
Data mining techniques are useful for predicting material
properties. In this paper, we show that decision tree ensembles
can be used to predict the steel plate faults. We carried out
experiments with different decision tree ensemble methods.
We found that AdaBoost.M1 performed best (18.08 % error)
when we removed 12 insignificant features. In this paper, we
showed that decision tree ensembles particularly Random
subspace and AdaBoost.M1 are very useful for steel plates
faults prediction. We also observed that removing 7
insignificant features improves the performance of decision
tree ensembles. In this paper, we applied decision tree
ensembles methods with Relief feature selection method for
predicting steel plates faults, in future we will apply ensembles
of neural networks [22] and support vector machines [23] for
predicting for predicting steel plates faults. In future, we will
use other feature section methods [17, 18] to study their
performance for steel plate faults prediction.
REFERENCES
[1] Rajan, K., Materialstoday, 2005, 8, 10.
DOI: 10.1016/S1369-7021(05)71123-8
[2] Kelly, A., Knowles, K. M., Crystallography and Crystal
Defects, John Wiley & Sons, 2012.
DOI: 10.1002/9781119961468
[3] Perzyk, M; Kochanski, A., Kozlowski, J.; Soroczynski, A.;
Biernacki R, Information Sciences, 2014, 259, 380-392. DOI
DOI: 10.1016/j.ins.2013.10.019
[4] Han, J; Kamber, M;, Pei, J; Data Mining: Concepts and
Techniques, Morgan Kaufmann, 2011.
DOI: 10.1145/565117.565130
[5] Kuncheva, L. I., Combining pattern classifiers: methods and
algorithms. Wiley-IEEE Press, New York, 2004.
DOI: .wiley.com/10.1002/0471660264.index
[6] Quinlan, J. R., C4.5: Programs for machine learning. CA:
Morgan Kaufmann, San Mateo, 1993.
DOI: 10.1007/BF00993309
[7] Ahmad, A. Halawani S., and Albidewi, T., Expert systems with
Applications, Elsevier, 2012, 39, 7, 6396-6401.
DOI: 10.1016/j.eswa.2011.12.029
[8] Ahmad, A.; Dey, L., Pattern Recognition Letters, 2005, 26, 1,
43–56.
DOI: 10.1016/j.patrec.2004.08.015
[9] Buscema, M.; Terzi, S.; Tastle, W.; Fuzzy Information
Processing Society (NAFIPS), 2010 Annual Meeting of the
North American New Meta-Classifier,in NAFIPS 2010, Toronto
(CANADA), 2010.
DOI: 10.1109/NAFIPS.2010.5548298
[10] Fakhr, M.; Elsayad, A., Journal of Computer Science, 2012, 8,
4, 506-514.
DOI: 10.3844/jcssp.2012.506.514
[11] Dietterich, T. G., In Proceedings of conference multiple
classifier systems, Lecture notes in computer science, 2000,
1857, pp. 1–15.
DOI: 10.1007/3-540-45014-9_1
[12] Breiman, L., J. Friedman, R. Olshen and C. Stone, Classification
and Regression Trees. 1st Edn., Chapman and Hall, London,
1984.
ISBN: 0412048418
[13] Breiman, L., Mach. Learn., 1996, 24, 2 123–140.
DOI: 10.1023/A:1018054314350
[14] Freund, Y. and R.E. Schapire,. J. Comput. Syst. Sci., 1997, 55,
119-139.
DOI: 10.1006/jcss.1997.1504
[15] Ho, T. K.; IEEE Transactions on Pattern Analysis and Machine
Intelligence, 1998, 20, 8, 832–844.
DOI: 10.1109/34.709601
[16] Breiman, L., Machine Learning, 2001, 45, 1, 5–32.
DOI: 10.1023/A:1010933404324
[17] Saeys, Y.; I. Inza, I.; Larranaga, P, Bioinformatics, 2007, 23,
19, 2507–2517.
DOI: 10.1093/bioinformatics/btm344
[18] Jain, A.; Zongker, D, IEEE Trans. Pattern Anal. Mach. Intell.,
1997, 19, 2, 153–158.
DOI: 10.1109/34.574797
5. International Journal of Technical Research and Applications e-ISSN: 2320-8163,
www.ijtra.com Volume 2, Issue 4 (July-Aug 2014), PP 127-131
131 | P a g e
[19] Kira, K.; Rendell, L.;. AAAI-92 Proceedings, 1992.
ISBN:0-262-51063-4
[20] Witten, I. H.; Frank, E., Data mining: practical machine learning
tools with java implementations. Morgan Kaufmann, San
Francisco, 2000.
ISBN: 0120884070
[21] Frank, A.; Asuncion, A.; UCI machine learning repository
Irvine. University of California, 2010.
http://archive.ics.uci.edu/ml/
[22] Bishop, C.M.. Pattern Recognition and Machine Learning. 1st
Edn., Springer-Verlag, New York, ISBN-10: 9780387310732,
2006.
ISBN: 0387310738
[23] Vapnik, V., The Nature of Statistical Learning Theory.
Springer-Verlag, New York, 1995.