SlideShare a Scribd company logo
Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com 
ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 
www.ijera.com 91|P a g e 
Comparision of methods for combination of multiple classifiers 
that predict behavior patterns 
Anuja V. Deshpande*, Prof. Dharmeshkumar Mistry** 
*(Department of Computer Science, D.J Sanghvi College of Engineering, Mumbai-56) 
** (Department of Computer Science, D.J Sanghvi College of Engineering, Mumbai-56) 
ABSTRACT 
Predictive analysis include techniques fromdata mining that analyze current and historical data and make 
predictions about the future. Predictive analytics is used in actuarial science, financial services, retail, travel, 
healthcare, insurance, pharmaceuticals, marketing, telecommunications and other fields.Predicting patterns can 
be considered as a classification problem and combining the different classifiers gives better results. We will 
study and compare three methods used to combine multiple classifiers. Bayesian networks perform 
classification based on conditional probability. It is ineffective and easy to interpret as it assumes that the 
predictors are independent. Tree augmented naïve Bayes (TAN) constructs a maximum weighted spanning tree 
that maximizes the likelihood of the training data, to perform classification.This tree structure eliminates the 
independent attribute assumption of naïve Bayesian networks. Behavior-knowledge space method works in two 
phases and can provide very good performances if large and representative data sets are available. 
Keywords– Bayesian networks, Behavior-knowledge space, TAN. 
I. INTRODUCTION 
There is an increasing demand for predicting 
behavior in all industries. Ecommerce, Credit risk, 
Insurance fraud are just some of the applications 
which benefit from behavior prediction. Predicting 
behavior patterns can be seen as a classification 
problem. It has been observed that a combination of 
different classifiers produces better results as 
compared to that of a single classifier. Different 
classifiers usually have different methodologies and 
features, which generally complement each other. 
Hence, cooperation between these classifiers can be 
optimized to reduce errors and increase the 
performance of classification. The paper is divided as 
follows. We first understand the problem of 
combining different classifiers, in section 2, we 
discuss the various methods for combining classifiers 
and in section 3, we make a comparative study of 
these methods. Finally, in section 4 conclusions are 
given.If ek denotes classifier k (where k=1,….K) and 
k is the total number of classifiers[4]. Let C1,...CM be 
mutually exclusive and exhaustive set of patterns and 
M represents the total number of pattern classes[4]. 
A= {1, …., M} be a set which consists of all class 
index numbers.[4] x is the unknown input pattern and 
ek(x)=jk means classifier k assigns input x to class 
푗푘 ∈ 퐴 ∪ { 푀 + 1 }. If 
푗푘 ∪ 퐴1, it means classifier k accepts x, otherwise it 
rejects x[4]. The combination problem can then be 
stated as,”When k classifiers give their individual 
decisions to the unknown input, what is the method 
which can combine them efficiently and produce the 
final decision?”[4]. It can be formulated as: 
(1) 
Where E is the panel of multiple classifiers which 
gives z one definitive class j[4]. 
Various methods have been proposed to combine 
classifiers. In this paper we will study and compare 
three types of classifier combining methods i.e. 
Bayesian networks, tree augmented naïve Bayes 
(TAN) and Behavior-Knowledge Space method. 
II. Methods for Combining Multiple 
Classifiers 
2.1 Naïve Bayes Network 
Fig 1. Naïve Bayes Network 
A Bayesian network is also called as Bayes network, 
belief network, Bayesian model or probabilistic 
directed acyclic graphical model. It is a probabilistic 
graphical model (a type of statistical model) that 
represents a set of random variables and their 
RESEARCH ARTICLE OPEN ACCESS
Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com 
ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 
www.ijera.com 92|P a g e 
conditional dependencies via a directed acyclic graph (DAG).For example For example, a Bayesian network could represent the probabilistic relationships between diseases customer behavior and their predilection to buy a product. Given certain attributes, the network can be used to compute whether a customer is likely to buy a given product. Conventionally, Bayesian networks are DAGs whose nodes represent random variables. These variables may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies and nodes that are not connected represent variables that are conditionally independent of each other. The probability function associated with each node takes as input a particular set of values for the node's parent variables and gives the probability of the variable represented by the node. For example, the probability function could be represented by a table of 2m entries, one entry for each of the 2m possible combinations of its parents being true or false, if the parents are m Boolean variables. Advantages of this approach are that it is easy to understand and fast to train. The probability of classifier selection of class j where 1≤j≤M as its classified class when true class was class i where 1≤i≤M is defined as: 
(2) Where ek(x) is a class label selected by classifier k as the true class for an input x[3]. The belief function for class i can be expressed by the sum of conditional probabilities as follows: 
(3) 
where ɳ is a normalization coefficient that satisfies [3].The belief function BEL(i) is the product of the contributions from all classifiers for class i, and represents the total validity for class i.[3] Taking the class label whose BEL value is the largest makes the final decision[3]. The combining rule is shown below: 
(4) 
This approach assumes that all features are independent of each other. No structure learning procedure is required and hence this structure is easy to construct and works efficiently. To allow more complex networks, the Tree Augmented Naïve Bayes (TAN) network is proposed. 2.2 Tree Augmented Naïve Bayes Networks (TAN) 
Fig 2. A TAN Bayes net The figure depicts a TAN Bayes net for combining multiple classifiers. Here, X1, X2, X3, X4 represent the different rules. The TAN model, while retaining the basic structure of Naïve Bayes, also permits each attribute to have at most one other parent, allowing the model to capture dependencies between attributes[1]. Which arcs to include in the 'augmented' network is decided by the algorithm by making a complete graph between all the non-class attributes, where the weight of each edge is given as the conditional mutual information between those two attributes. A maximum weight spanning tree is constructed over this graph, and the edges that appear in the spanning tree are added to the network[1]. Given the independence assumptions in the tree T , the posterior probability is: 
(5) where xj(k) stands for the parent of variable xk in the tree T , and x0 for the null. We now need to keep a counter for the number of training instances, a counter for each class label, and a counter for each attribute value, parent value and class label triplet[2]. TAN maintains the computational simplicity if Naïve Bayes while increasing the performance. 2.3 Behavior-knowledge Space Method. This method has two stages (1) the knowledge modeling stage, responsible for extracting knowledge from behavior of classifiers and constructing a K- dimensional behavior-knowledge space; and (2) the operation stage that is carried out for each test sample and which combines decisions generated from individual classifiers, enters a specific unit of the constructed space, and makes a final decision by a rule which utilizes the knowledge inside the unit [4]. 
A behavior-knowledge space (BKS) is a K- dimensional space where each dimension
Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com 
ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 
www.ijera.com 93|P a g e 
corresponds to the decision of one classifier and has M+1 possible decision values from the set {1,2, …., M+1}[4]. If the decision of the classifier belongs to the set {1, …., M}, the classifier accepts the input sample, else the classifier rejects the input sample [4]. Each unit of the BKS contains three kinds of data (1) the total number of incoming samples, (2) the best representative class and (3) the number of samples belonging to each class [4]. The first i.e. the knowledge modeling stage, uses the learning set of samples with the expected class labels and the recognized class labels to construct the BKS [4]. The values of T and R are computed as follows: 푇푒 1 …..푒 퐾 = 푛푒 1 ……푒(퐾) (푚)푀푚 =1 푅e(1)…..e(K)={j|ne(1)……e(K)(j)= max1≤m≤Mne(1)……e(K) (m)} (6) Where, ne(1)…….e(K) (m)= the number of incoming samples belonging to class m in BKS Te(1)…….e(K)=total number of samples in BKS Re(1)…..e(K)= best representative class for the BKS[4]. In the operation stage, the final decision is made by the following rule: E(x)=Re(1)…..e(K) , when Te(1)…….e(K)>0and 푛푒 1 …..푒(푘) (푅푒 1 ……푒(퐾)) 푇푒 1 ……푒(푘) ≥휆 , M+1 , otherwise (7) Where λis a threshold which controls the reliable degree of the decision [4]. Table 1.Two dimensional behavior knowledge space 
each cell in the table means the intersection of the decision values from the individual classifiers and becomes a basic unit of computation in BKS approach[3]. 
III. A Comparative Study 
3.1. Independent clause assumption 
A naïve Bayesian network needs the clauses to be independent. There may be strong dependencies among clauses that are not realized. Tree augmented naïve Bayes (TAN) allows us to capture the dependencies between different attributes and thus, improves performance of the former. But, even TAN has limitations. It allows each attribute to have at most one parent, thus restricting dependency. Behavior-knowledge space approach does not assume independencies among classifiers at all and hence, performs better in most cases than Bayesian networks and Tree augmented Bayesian networks. 3.2. Working In naïve Bayesian networks, the outcome of each clause is represented as a random variable, the value of which depends on the examples classification. The tree augmented naïve Bayes algorithm works by making a complete graph between all the non-class attributes, where the weight of each edge is given as the conditional mutual information between those two attributes [1]. A maximum weight spanning tree is constructed over this graph, and the edges that appear in the spanning tree are added to the network [1]. In contrast to these approaches which rely on making trees, the Behavior-knowledge space approach works in two stages. First, by constructing a knowledge space and second, by making the final decision in the operation stage. 3.3. Performance Naïve Bayesian networks assume classifier independency and are thus easy to interpret but inefficient. They are also too cautious about classifying something as positive [1]. TAN excel in handling imprecise rules and provide an advantage in situations with imprecise rules and a preponderance of negative examples, such as these link discovery domains [1]. The Behavior-knowledge space method can provide very good performances if large and representative data sets are available [5]. Otherwise over fitting is likely to occur, and the generalization error quickly increases [5]. 
IV. CONCLUSION 
The assumptions, working and performance of three different approaches has been discussed. The Naïve Bayesian network, easiest to understand is very ineffective. The tree augmented naïve Bayes improves the performance of naïve Bayes to some extent, but still does not completely eliminate classifier independency assumption. The Behavior- knowledge space method overcomes this limitation and also has adaptive learning ability. It can also automatically find out the best threshold for a given required performance. Future works can include attempts to combine the different methods suggested here to achieve better results.
Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com 
ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 
www.ijera.com 94|P a g e 
REFERENCES 
Journal Papers: [1] Jesse Davis, Vitor Santos Costa, Irene M. Ong, David Page and Ines Dutra, Using Bayesian Classifiers to Combine Rules 
[2] Josep Roure Alcobé, Incremental Learning of Tree Augmented Naive Bayes Classifiers [3] Eunju Kim, Wooju Kim, Yillbyung Lee, Combination of multiple classifiers for the customer’s purchase behavior prediction, Decision Support Systems 34 (2002) 167– 175 Proceedings Papers: [4] Huang, Y.S., Suen, C.Y., The behavior- knowledge space method for combination of multiple classifiers,Computer Vision and Pattern Recognition, 1993. Proceedings CVPR '93., 1993 IEEE Computer Society Conference, 1993, 347-352. Chapters in Books: [5] Terry Windeatt, Fabio Roli, The Behavior Knowledge Space Fusion Method: Analysis of Generalization Error and Strategies for Performance Improvement, Multiple Classifier Systems, (4th International Workshop, MCS 2003 Guildford, UK, June 11–13, 2003 Proceedings, Springer Berlin Heidelberg).

More Related Content

What's hot

An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
idescitation
 
Clusters techniques
Clusters techniquesClusters techniques
Clusters techniques
rajshreemuthiah
 
Feature selection in multimodal
Feature selection in multimodalFeature selection in multimodal
Feature selection in multimodal
ijcsa
 
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
K-MEDOIDS CLUSTERING  USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...K-MEDOIDS CLUSTERING  USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
ijscmc
 
Unsupervised learning clustering
Unsupervised learning clusteringUnsupervised learning clustering
Unsupervised learning clustering
Dr Nisha Arora
 
Data clustering
Data clustering Data clustering
Data clustering
GARIMA SHAKYA
 
My8clst
My8clstMy8clst
My8clst
ketan533
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
SSA KPI
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
IJERA Editor
 
Clustering techniques final
Clustering techniques finalClustering techniques final
Clustering techniques final
Benard Maina
 
Efficient classification of big data using vfdt (very fast decision tree)
Efficient classification of big data using vfdt (very fast decision tree)Efficient classification of big data using vfdt (very fast decision tree)
Efficient classification of big data using vfdt (very fast decision tree)
eSAT Journals
 
An approximate possibilistic
An approximate possibilisticAn approximate possibilistic
An approximate possibilistic
csandit
 
3.1 clustering
3.1 clustering3.1 clustering
3.1 clustering
Krish_ver2
 
Capter10 cluster basic
Capter10 cluster basicCapter10 cluster basic
Capter10 cluster basic
Houw Liong The
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
guest0edcaf
 
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
cscpconf
 

What's hot (20)

An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
An Automatic Medical Image Segmentation using Teaching Learning Based Optimiz...
 
Clusters techniques
Clusters techniquesClusters techniques
Clusters techniques
 
Feature selection in multimodal
Feature selection in multimodalFeature selection in multimodal
Feature selection in multimodal
 
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
K-MEDOIDS CLUSTERING  USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...K-MEDOIDS CLUSTERING  USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
K-MEDOIDS CLUSTERING USING PARTITIONING AROUND MEDOIDS FOR PERFORMING FACE R...
 
Unsupervised learning clustering
Unsupervised learning clusteringUnsupervised learning clustering
Unsupervised learning clustering
 
Data clustering
Data clustering Data clustering
Data clustering
 
Cluster analysis
Cluster analysisCluster analysis
Cluster analysis
 
My8clst
My8clstMy8clst
My8clst
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
 
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...A Combined Approach for Feature Subset Selection and Size Reduction for High ...
A Combined Approach for Feature Subset Selection and Size Reduction for High ...
 
Cluster analysis
Cluster analysisCluster analysis
Cluster analysis
 
Clustering techniques final
Clustering techniques finalClustering techniques final
Clustering techniques final
 
Dbm630 lecture09
Dbm630 lecture09Dbm630 lecture09
Dbm630 lecture09
 
Efficient classification of big data using vfdt (very fast decision tree)
Efficient classification of big data using vfdt (very fast decision tree)Efficient classification of big data using vfdt (very fast decision tree)
Efficient classification of big data using vfdt (very fast decision tree)
 
An approximate possibilistic
An approximate possibilisticAn approximate possibilistic
An approximate possibilistic
 
3.1 clustering
3.1 clustering3.1 clustering
3.1 clustering
 
Cluster analysis
Cluster analysisCluster analysis
Cluster analysis
 
Capter10 cluster basic
Capter10 cluster basicCapter10 cluster basic
Capter10 cluster basic
 
Cluster Analysis
Cluster AnalysisCluster Analysis
Cluster Analysis
 
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
EFFECTIVENESS PREDICTION OF MEMORY BASED CLASSIFIERS FOR THE CLASSIFICATION O...
 

Viewers also liked

Porque no vacunarse 14 4 10
Porque no vacunarse 14 4 10Porque no vacunarse 14 4 10
Porque no vacunarse 14 4 10Jose Luis
 
Heber beloni cajado_silva
Heber beloni cajado_silvaHeber beloni cajado_silva
Heber beloni cajado_silva
alemaun_vr
 
Portfólio - Estúdio AMBIENTES VIRTUAIS 3D
Portfólio - Estúdio AMBIENTES VIRTUAIS 3DPortfólio - Estúdio AMBIENTES VIRTUAIS 3D
Portfólio - Estúdio AMBIENTES VIRTUAIS 3D
Estúdio AMBIENTES VIRTUAIS 3D
 
Challenging
ChallengingChallenging
ChallengingGIA VER
 
Islamabad Office Market Snapshot 2016
Islamabad Office Market Snapshot 2016Islamabad Office Market Snapshot 2016
Islamabad Office Market Snapshot 2016Kamil Noorani
 
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
IJERA Editor
 
La formación de docentes Philippe Perrenaud
La formación de docentes Philippe PerrenaudLa formación de docentes Philippe Perrenaud
La formación de docentes Philippe Perrenaud
María Eugenia Gallo Arbeláez
 
Business Savvy From the Start: Excel in an Entry Level Job
Business Savvy From the Start: Excel in an Entry Level JobBusiness Savvy From the Start: Excel in an Entry Level Job
Business Savvy From the Start: Excel in an Entry Level Job
Rasmussen College
 
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-2104103 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410guestee6378
 
Tumor Detection Based On Symmetry Information
Tumor Detection Based On Symmetry InformationTumor Detection Based On Symmetry Information
Tumor Detection Based On Symmetry Information
IJERA Editor
 

Viewers also liked (20)

Ponencia colombia
Ponencia colombiaPonencia colombia
Ponencia colombia
 
Porque no vacunarse 14 4 10
Porque no vacunarse 14 4 10Porque no vacunarse 14 4 10
Porque no vacunarse 14 4 10
 
Pasos Simples para Salvar al Mundo
Pasos Simples para Salvar al MundoPasos Simples para Salvar al Mundo
Pasos Simples para Salvar al Mundo
 
Heber beloni cajado_silva
Heber beloni cajado_silvaHeber beloni cajado_silva
Heber beloni cajado_silva
 
Legal Note - Atenas
Legal Note - AtenasLegal Note - Atenas
Legal Note - Atenas
 
Inteligencia colectiva
Inteligencia colectivaInteligencia colectiva
Inteligencia colectiva
 
Consejos para disfrutar de una casa sana electricidad
Consejos para disfrutar de una casa sana electricidadConsejos para disfrutar de una casa sana electricidad
Consejos para disfrutar de una casa sana electricidad
 
Portfólio - Estúdio AMBIENTES VIRTUAIS 3D
Portfólio - Estúdio AMBIENTES VIRTUAIS 3DPortfólio - Estúdio AMBIENTES VIRTUAIS 3D
Portfólio - Estúdio AMBIENTES VIRTUAIS 3D
 
Challenging
ChallengingChallenging
Challenging
 
Islamabad Office Market Snapshot 2016
Islamabad Office Market Snapshot 2016Islamabad Office Market Snapshot 2016
Islamabad Office Market Snapshot 2016
 
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
Direct Torque Control of Induction Motor Drive Fed from a Photovoltaic Multil...
 
Limitación auditiva
Limitación auditivaLimitación auditiva
Limitación auditiva
 
Comprensión y creación de sentido
Comprensión y creación de sentidoComprensión y creación de sentido
Comprensión y creación de sentido
 
La formación de docentes sigloXXI Philippe Perrenaud
La formación de docentes sigloXXI Philippe PerrenaudLa formación de docentes sigloXXI Philippe Perrenaud
La formación de docentes sigloXXI Philippe Perrenaud
 
La formación de docentes Philippe Perrenaud
La formación de docentes Philippe PerrenaudLa formación de docentes Philippe Perrenaud
La formación de docentes Philippe Perrenaud
 
Business Savvy From the Start: Excel in an Entry Level Job
Business Savvy From the Start: Excel in an Entry Level JobBusiness Savvy From the Start: Excel in an Entry Level Job
Business Savvy From the Start: Excel in an Entry Level Job
 
2
22
2
 
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-2104103 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410
3 action care_-plan_de_marketing_y_beneficios_para_el_franquiciado_-210410
 
2
22
2
 
Tumor Detection Based On Symmetry Information
Tumor Detection Based On Symmetry InformationTumor Detection Based On Symmetry Information
Tumor Detection Based On Symmetry Information
 

Similar to Comparision of methods for combination of multiple classifiers that predict behavior patterns

Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problem
nooriasukmaningtyas
 
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
IJRES Journal
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
IRJET Journal
 
Optimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering AlgorithmOptimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering Algorithm
IJERA Editor
 
Analysis of machine learning algorithms for character recognition: a case stu...
Analysis of machine learning algorithms for character recognition: a case stu...Analysis of machine learning algorithms for character recognition: a case stu...
Analysis of machine learning algorithms for character recognition: a case stu...
nooriasukmaningtyas
 
Application of combined support vector machines in process fault diagnosis
Application of combined support vector machines in process fault diagnosisApplication of combined support vector machines in process fault diagnosis
Application of combined support vector machines in process fault diagnosis
Dr.Pooja Jain
 
K means report
K means reportK means report
K means report
Gaurav Handa
 
20070702 Text Categorization
20070702 Text Categorization20070702 Text Categorization
20070702 Text Categorization
midi
 
Machine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.pptMachine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.ppt
Anshika865276
 
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITYPROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
IAEME Publication
 
Mapping Subsets of Scholarly Information
Mapping Subsets of Scholarly InformationMapping Subsets of Scholarly Information
Mapping Subsets of Scholarly Information
Paul Houle
 
An Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution AlgorithmAn Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution Algorithm
IOSR Journals
 
Neural nw k means
Neural nw k meansNeural nw k means
Neural nw k means
Eng. Dr. Dennis N. Mwighusa
 
Classification Techniques: A Review
Classification Techniques: A ReviewClassification Techniques: A Review
Classification Techniques: A Review
IOSRjournaljce
 
K-means Clustering Method for the Analysis of Log Data
K-means Clustering Method for the Analysis of Log DataK-means Clustering Method for the Analysis of Log Data
K-means Clustering Method for the Analysis of Log Data
idescitation
 
nnml.ppt
nnml.pptnnml.ppt
nnml.ppt
yang947066
 
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
IJECEIAES
 
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
IOSR Journals
 
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques  Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
ijsc
 

Similar to Comparision of methods for combination of multiple classifiers that predict behavior patterns (20)

Max stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problemMax stable set problem to found the initial centroids in clustering problem
Max stable set problem to found the initial centroids in clustering problem
 
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
 
1376846406 14447221
1376846406  144472211376846406  14447221
1376846406 14447221
 
Optimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering AlgorithmOptimising Data Using K-Means Clustering Algorithm
Optimising Data Using K-Means Clustering Algorithm
 
Analysis of machine learning algorithms for character recognition: a case stu...
Analysis of machine learning algorithms for character recognition: a case stu...Analysis of machine learning algorithms for character recognition: a case stu...
Analysis of machine learning algorithms for character recognition: a case stu...
 
Application of combined support vector machines in process fault diagnosis
Application of combined support vector machines in process fault diagnosisApplication of combined support vector machines in process fault diagnosis
Application of combined support vector machines in process fault diagnosis
 
K means report
K means reportK means report
K means report
 
20070702 Text Categorization
20070702 Text Categorization20070702 Text Categorization
20070702 Text Categorization
 
Machine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.pptMachine Learning and Artificial Neural Networks.ppt
Machine Learning and Artificial Neural Networks.ppt
 
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITYPROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
PROFIT AGENT CLASSIFICATION USING FEATURE SELECTION EIGENVECTOR CENTRALITY
 
Mapping Subsets of Scholarly Information
Mapping Subsets of Scholarly InformationMapping Subsets of Scholarly Information
Mapping Subsets of Scholarly Information
 
An Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution AlgorithmAn Adaptive Masker for the Differential Evolution Algorithm
An Adaptive Masker for the Differential Evolution Algorithm
 
Neural nw k means
Neural nw k meansNeural nw k means
Neural nw k means
 
Classification Techniques: A Review
Classification Techniques: A ReviewClassification Techniques: A Review
Classification Techniques: A Review
 
K-means Clustering Method for the Analysis of Log Data
K-means Clustering Method for the Analysis of Log DataK-means Clustering Method for the Analysis of Log Data
K-means Clustering Method for the Analysis of Log Data
 
nnml.ppt
nnml.pptnnml.ppt
nnml.ppt
 
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
Comparison of Iris dataset classification with Gaussian naïve Bayes and decis...
 
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
Efficient Disease Classifier Using Data Mining Techniques: Refinement of Rand...
 
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques  Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
Methodological Study Of Opinion Mining And Sentiment Analysis Techniques
 

Recently uploaded

Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
AJAYKUMARPUND1
 
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
Amil Baba Dawood bangali
 
DESIGN A COTTON SEED SEPARATION MACHINE.docx
DESIGN A COTTON SEED SEPARATION MACHINE.docxDESIGN A COTTON SEED SEPARATION MACHINE.docx
DESIGN A COTTON SEED SEPARATION MACHINE.docx
FluxPrime1
 
weather web application report.pdf
weather web application report.pdfweather web application report.pdf
weather web application report.pdf
Pratik Pawar
 
Courier management system project report.pdf
Courier management system project report.pdfCourier management system project report.pdf
Courier management system project report.pdf
Kamal Acharya
 
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSE
TECHNICAL TRAINING MANUAL   GENERAL FAMILIARIZATION COURSETECHNICAL TRAINING MANUAL   GENERAL FAMILIARIZATION COURSE
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSE
DuvanRamosGarzon1
 
Cosmetic shop management system project report.pdf
Cosmetic shop management system project report.pdfCosmetic shop management system project report.pdf
Cosmetic shop management system project report.pdf
Kamal Acharya
 
HYDROPOWER - Hydroelectric power generation
HYDROPOWER - Hydroelectric power generationHYDROPOWER - Hydroelectric power generation
HYDROPOWER - Hydroelectric power generation
Robbie Edward Sayers
 
Planning Of Procurement o different goods and services
Planning Of Procurement o different goods and servicesPlanning Of Procurement o different goods and services
Planning Of Procurement o different goods and services
JoytuBarua2
 
Vaccine management system project report documentation..pdf
Vaccine management system project report documentation..pdfVaccine management system project report documentation..pdf
Vaccine management system project report documentation..pdf
Kamal Acharya
 
MCQ Soil mechanics questions (Soil shear strength).pdf
MCQ Soil mechanics questions (Soil shear strength).pdfMCQ Soil mechanics questions (Soil shear strength).pdf
MCQ Soil mechanics questions (Soil shear strength).pdf
Osamah Alsalih
 
Architectural Portfolio Sean Lockwood
Architectural Portfolio Sean LockwoodArchitectural Portfolio Sean Lockwood
Architectural Portfolio Sean Lockwood
seandesed
 
ASME IX(9) 2007 Full Version .pdf
ASME IX(9)  2007 Full Version       .pdfASME IX(9)  2007 Full Version       .pdf
ASME IX(9) 2007 Full Version .pdf
AhmedHussein950959
 
Immunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary AttacksImmunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary Attacks
gerogepatton
 
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Dr.Costas Sachpazis
 
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdfHybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
fxintegritypublishin
 
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
bakpo1
 
WATER CRISIS and its solutions-pptx 1234
WATER CRISIS and its solutions-pptx 1234WATER CRISIS and its solutions-pptx 1234
WATER CRISIS and its solutions-pptx 1234
AafreenAbuthahir2
 
Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.
PrashantGoswami42
 
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfCOLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
Kamal Acharya
 

Recently uploaded (20)

Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
Pile Foundation by Venkatesh Taduvai (Sub Geotechnical Engineering II)-conver...
 
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
NO1 Uk best vashikaran specialist in delhi vashikaran baba near me online vas...
 
DESIGN A COTTON SEED SEPARATION MACHINE.docx
DESIGN A COTTON SEED SEPARATION MACHINE.docxDESIGN A COTTON SEED SEPARATION MACHINE.docx
DESIGN A COTTON SEED SEPARATION MACHINE.docx
 
weather web application report.pdf
weather web application report.pdfweather web application report.pdf
weather web application report.pdf
 
Courier management system project report.pdf
Courier management system project report.pdfCourier management system project report.pdf
Courier management system project report.pdf
 
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSE
TECHNICAL TRAINING MANUAL   GENERAL FAMILIARIZATION COURSETECHNICAL TRAINING MANUAL   GENERAL FAMILIARIZATION COURSE
TECHNICAL TRAINING MANUAL GENERAL FAMILIARIZATION COURSE
 
Cosmetic shop management system project report.pdf
Cosmetic shop management system project report.pdfCosmetic shop management system project report.pdf
Cosmetic shop management system project report.pdf
 
HYDROPOWER - Hydroelectric power generation
HYDROPOWER - Hydroelectric power generationHYDROPOWER - Hydroelectric power generation
HYDROPOWER - Hydroelectric power generation
 
Planning Of Procurement o different goods and services
Planning Of Procurement o different goods and servicesPlanning Of Procurement o different goods and services
Planning Of Procurement o different goods and services
 
Vaccine management system project report documentation..pdf
Vaccine management system project report documentation..pdfVaccine management system project report documentation..pdf
Vaccine management system project report documentation..pdf
 
MCQ Soil mechanics questions (Soil shear strength).pdf
MCQ Soil mechanics questions (Soil shear strength).pdfMCQ Soil mechanics questions (Soil shear strength).pdf
MCQ Soil mechanics questions (Soil shear strength).pdf
 
Architectural Portfolio Sean Lockwood
Architectural Portfolio Sean LockwoodArchitectural Portfolio Sean Lockwood
Architectural Portfolio Sean Lockwood
 
ASME IX(9) 2007 Full Version .pdf
ASME IX(9)  2007 Full Version       .pdfASME IX(9)  2007 Full Version       .pdf
ASME IX(9) 2007 Full Version .pdf
 
Immunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary AttacksImmunizing Image Classifiers Against Localized Adversary Attacks
Immunizing Image Classifiers Against Localized Adversary Attacks
 
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
Sachpazis:Terzaghi Bearing Capacity Estimation in simple terms with Calculati...
 
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdfHybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
Hybrid optimization of pumped hydro system and solar- Engr. Abdul-Azeez.pdf
 
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
一比一原版(SFU毕业证)西蒙菲莎大学毕业证成绩单如何办理
 
WATER CRISIS and its solutions-pptx 1234
WATER CRISIS and its solutions-pptx 1234WATER CRISIS and its solutions-pptx 1234
WATER CRISIS and its solutions-pptx 1234
 
Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.Quality defects in TMT Bars, Possible causes and Potential Solutions.
Quality defects in TMT Bars, Possible causes and Potential Solutions.
 
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdfCOLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
COLLEGE BUS MANAGEMENT SYSTEM PROJECT REPORT.pdf
 

Comparision of methods for combination of multiple classifiers that predict behavior patterns

  • 1. Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 www.ijera.com 91|P a g e Comparision of methods for combination of multiple classifiers that predict behavior patterns Anuja V. Deshpande*, Prof. Dharmeshkumar Mistry** *(Department of Computer Science, D.J Sanghvi College of Engineering, Mumbai-56) ** (Department of Computer Science, D.J Sanghvi College of Engineering, Mumbai-56) ABSTRACT Predictive analysis include techniques fromdata mining that analyze current and historical data and make predictions about the future. Predictive analytics is used in actuarial science, financial services, retail, travel, healthcare, insurance, pharmaceuticals, marketing, telecommunications and other fields.Predicting patterns can be considered as a classification problem and combining the different classifiers gives better results. We will study and compare three methods used to combine multiple classifiers. Bayesian networks perform classification based on conditional probability. It is ineffective and easy to interpret as it assumes that the predictors are independent. Tree augmented naïve Bayes (TAN) constructs a maximum weighted spanning tree that maximizes the likelihood of the training data, to perform classification.This tree structure eliminates the independent attribute assumption of naïve Bayesian networks. Behavior-knowledge space method works in two phases and can provide very good performances if large and representative data sets are available. Keywords– Bayesian networks, Behavior-knowledge space, TAN. I. INTRODUCTION There is an increasing demand for predicting behavior in all industries. Ecommerce, Credit risk, Insurance fraud are just some of the applications which benefit from behavior prediction. Predicting behavior patterns can be seen as a classification problem. It has been observed that a combination of different classifiers produces better results as compared to that of a single classifier. Different classifiers usually have different methodologies and features, which generally complement each other. Hence, cooperation between these classifiers can be optimized to reduce errors and increase the performance of classification. The paper is divided as follows. We first understand the problem of combining different classifiers, in section 2, we discuss the various methods for combining classifiers and in section 3, we make a comparative study of these methods. Finally, in section 4 conclusions are given.If ek denotes classifier k (where k=1,….K) and k is the total number of classifiers[4]. Let C1,...CM be mutually exclusive and exhaustive set of patterns and M represents the total number of pattern classes[4]. A= {1, …., M} be a set which consists of all class index numbers.[4] x is the unknown input pattern and ek(x)=jk means classifier k assigns input x to class 푗푘 ∈ 퐴 ∪ { 푀 + 1 }. If 푗푘 ∪ 퐴1, it means classifier k accepts x, otherwise it rejects x[4]. The combination problem can then be stated as,”When k classifiers give their individual decisions to the unknown input, what is the method which can combine them efficiently and produce the final decision?”[4]. It can be formulated as: (1) Where E is the panel of multiple classifiers which gives z one definitive class j[4]. Various methods have been proposed to combine classifiers. In this paper we will study and compare three types of classifier combining methods i.e. Bayesian networks, tree augmented naïve Bayes (TAN) and Behavior-Knowledge Space method. II. Methods for Combining Multiple Classifiers 2.1 Naïve Bayes Network Fig 1. Naïve Bayes Network A Bayesian network is also called as Bayes network, belief network, Bayesian model or probabilistic directed acyclic graphical model. It is a probabilistic graphical model (a type of statistical model) that represents a set of random variables and their RESEARCH ARTICLE OPEN ACCESS
  • 2. Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 www.ijera.com 92|P a g e conditional dependencies via a directed acyclic graph (DAG).For example For example, a Bayesian network could represent the probabilistic relationships between diseases customer behavior and their predilection to buy a product. Given certain attributes, the network can be used to compute whether a customer is likely to buy a given product. Conventionally, Bayesian networks are DAGs whose nodes represent random variables. These variables may be observable quantities, latent variables, unknown parameters or hypotheses. Edges represent conditional dependencies and nodes that are not connected represent variables that are conditionally independent of each other. The probability function associated with each node takes as input a particular set of values for the node's parent variables and gives the probability of the variable represented by the node. For example, the probability function could be represented by a table of 2m entries, one entry for each of the 2m possible combinations of its parents being true or false, if the parents are m Boolean variables. Advantages of this approach are that it is easy to understand and fast to train. The probability of classifier selection of class j where 1≤j≤M as its classified class when true class was class i where 1≤i≤M is defined as: (2) Where ek(x) is a class label selected by classifier k as the true class for an input x[3]. The belief function for class i can be expressed by the sum of conditional probabilities as follows: (3) where ɳ is a normalization coefficient that satisfies [3].The belief function BEL(i) is the product of the contributions from all classifiers for class i, and represents the total validity for class i.[3] Taking the class label whose BEL value is the largest makes the final decision[3]. The combining rule is shown below: (4) This approach assumes that all features are independent of each other. No structure learning procedure is required and hence this structure is easy to construct and works efficiently. To allow more complex networks, the Tree Augmented Naïve Bayes (TAN) network is proposed. 2.2 Tree Augmented Naïve Bayes Networks (TAN) Fig 2. A TAN Bayes net The figure depicts a TAN Bayes net for combining multiple classifiers. Here, X1, X2, X3, X4 represent the different rules. The TAN model, while retaining the basic structure of Naïve Bayes, also permits each attribute to have at most one other parent, allowing the model to capture dependencies between attributes[1]. Which arcs to include in the 'augmented' network is decided by the algorithm by making a complete graph between all the non-class attributes, where the weight of each edge is given as the conditional mutual information between those two attributes. A maximum weight spanning tree is constructed over this graph, and the edges that appear in the spanning tree are added to the network[1]. Given the independence assumptions in the tree T , the posterior probability is: (5) where xj(k) stands for the parent of variable xk in the tree T , and x0 for the null. We now need to keep a counter for the number of training instances, a counter for each class label, and a counter for each attribute value, parent value and class label triplet[2]. TAN maintains the computational simplicity if Naïve Bayes while increasing the performance. 2.3 Behavior-knowledge Space Method. This method has two stages (1) the knowledge modeling stage, responsible for extracting knowledge from behavior of classifiers and constructing a K- dimensional behavior-knowledge space; and (2) the operation stage that is carried out for each test sample and which combines decisions generated from individual classifiers, enters a specific unit of the constructed space, and makes a final decision by a rule which utilizes the knowledge inside the unit [4]. A behavior-knowledge space (BKS) is a K- dimensional space where each dimension
  • 3. Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 www.ijera.com 93|P a g e corresponds to the decision of one classifier and has M+1 possible decision values from the set {1,2, …., M+1}[4]. If the decision of the classifier belongs to the set {1, …., M}, the classifier accepts the input sample, else the classifier rejects the input sample [4]. Each unit of the BKS contains three kinds of data (1) the total number of incoming samples, (2) the best representative class and (3) the number of samples belonging to each class [4]. The first i.e. the knowledge modeling stage, uses the learning set of samples with the expected class labels and the recognized class labels to construct the BKS [4]. The values of T and R are computed as follows: 푇푒 1 …..푒 퐾 = 푛푒 1 ……푒(퐾) (푚)푀푚 =1 푅e(1)…..e(K)={j|ne(1)……e(K)(j)= max1≤m≤Mne(1)……e(K) (m)} (6) Where, ne(1)…….e(K) (m)= the number of incoming samples belonging to class m in BKS Te(1)…….e(K)=total number of samples in BKS Re(1)…..e(K)= best representative class for the BKS[4]. In the operation stage, the final decision is made by the following rule: E(x)=Re(1)…..e(K) , when Te(1)…….e(K)>0and 푛푒 1 …..푒(푘) (푅푒 1 ……푒(퐾)) 푇푒 1 ……푒(푘) ≥휆 , M+1 , otherwise (7) Where λis a threshold which controls the reliable degree of the decision [4]. Table 1.Two dimensional behavior knowledge space each cell in the table means the intersection of the decision values from the individual classifiers and becomes a basic unit of computation in BKS approach[3]. III. A Comparative Study 3.1. Independent clause assumption A naïve Bayesian network needs the clauses to be independent. There may be strong dependencies among clauses that are not realized. Tree augmented naïve Bayes (TAN) allows us to capture the dependencies between different attributes and thus, improves performance of the former. But, even TAN has limitations. It allows each attribute to have at most one parent, thus restricting dependency. Behavior-knowledge space approach does not assume independencies among classifiers at all and hence, performs better in most cases than Bayesian networks and Tree augmented Bayesian networks. 3.2. Working In naïve Bayesian networks, the outcome of each clause is represented as a random variable, the value of which depends on the examples classification. The tree augmented naïve Bayes algorithm works by making a complete graph between all the non-class attributes, where the weight of each edge is given as the conditional mutual information between those two attributes [1]. A maximum weight spanning tree is constructed over this graph, and the edges that appear in the spanning tree are added to the network [1]. In contrast to these approaches which rely on making trees, the Behavior-knowledge space approach works in two stages. First, by constructing a knowledge space and second, by making the final decision in the operation stage. 3.3. Performance Naïve Bayesian networks assume classifier independency and are thus easy to interpret but inefficient. They are also too cautious about classifying something as positive [1]. TAN excel in handling imprecise rules and provide an advantage in situations with imprecise rules and a preponderance of negative examples, such as these link discovery domains [1]. The Behavior-knowledge space method can provide very good performances if large and representative data sets are available [5]. Otherwise over fitting is likely to occur, and the generalization error quickly increases [5]. IV. CONCLUSION The assumptions, working and performance of three different approaches has been discussed. The Naïve Bayesian network, easiest to understand is very ineffective. The tree augmented naïve Bayes improves the performance of naïve Bayes to some extent, but still does not completely eliminate classifier independency assumption. The Behavior- knowledge space method overcomes this limitation and also has adaptive learning ability. It can also automatically find out the best threshold for a given required performance. Future works can include attempts to combine the different methods suggested here to achieve better results.
  • 4. Anuja V. Deshpande Int. Journal of Engineering Research and Applications www.ijera.com ISSN : 2248-9622, Vol. 4, Issue 9( Version 3), September 2014, pp.91-94 www.ijera.com 94|P a g e REFERENCES Journal Papers: [1] Jesse Davis, Vitor Santos Costa, Irene M. Ong, David Page and Ines Dutra, Using Bayesian Classifiers to Combine Rules [2] Josep Roure Alcobé, Incremental Learning of Tree Augmented Naive Bayes Classifiers [3] Eunju Kim, Wooju Kim, Yillbyung Lee, Combination of multiple classifiers for the customer’s purchase behavior prediction, Decision Support Systems 34 (2002) 167– 175 Proceedings Papers: [4] Huang, Y.S., Suen, C.Y., The behavior- knowledge space method for combination of multiple classifiers,Computer Vision and Pattern Recognition, 1993. Proceedings CVPR '93., 1993 IEEE Computer Society Conference, 1993, 347-352. Chapters in Books: [5] Terry Windeatt, Fabio Roli, The Behavior Knowledge Space Fusion Method: Analysis of Generalization Error and Strategies for Performance Improvement, Multiple Classifier Systems, (4th International Workshop, MCS 2003 Guildford, UK, June 11–13, 2003 Proceedings, Springer Berlin Heidelberg).