SlideShare a Scribd company logo
1 of 4
Download to read offline
IJSRD - International Journal for Scientific Research & Development| Vol. 1, Issue 3, 2013 | ISSN (online): 2321-0613
All rights reserved by www.ijsrd.com 775
Incremental Discretization for Naïve Bayes Learning using FIFFD
Mr. Kunal Khimani 1
Mr. Kamal Sutaria 2
Ms. Kruti Khalpada 3
1
Gujarat Technological University PG School, Ahmedabad, Gujarat
2
Asst. Prof., C.E. Department, VVP Engineering College, Rajkot Gujarat
3
Institute of Technology, Nirma University, Ahmedabad, Gujarat
Abstract—Incremental Flexible Frequency Discretization
(IFFD) is a recently proposed discretization approach for
Naïve Bayes (NB). IFFD performs satisfactory by setting
the minimal interval frequency for discretized intervals as a
fixed number. In this paper, we first argue that this setting
cannot guarantee that the selecting MinBinSize is on always
optimal for all the different datasets. So the performance of
Naïve Bayes is not good in terms of classification error. We
thus proposed a sequential search method for NB: named
Flexible IFFD. Experiments were conducted on 4 datasets
from UCI machine learning repository and performance was
compared between NB trained on the data discretized by
FIFFD, IFFD, and PKID.
Keywords: Discretization, incremental, Naïve Bayes.
I. INTRODUCTION
Naive-Bayes classifiers are widely employed for
classification tasks because of their efficiency and efficacy.
Naïve Bayesian classifiers are simple, robust, and also
support incremental training. Its efficiency is witnessed
widespread deployment in classification task. Naïve
Bayesian classifiers have long been a core technique in
information retrieval. Naive-Bayesian learning needs to
estimate probabilities for each attribute-class pair. The naïve
Bayesian classifier provides a very simple and yet
surprisingly accurate technique for machine learning. The
naïve Bayesian classifier provides a very simple and yet
surprisingly accurate technique for machine learning. When
classifying an instance, naïve Bayesian classifiers assume
the attribute conditionally independent of each other given
the class; then apply the Bayes’ theorem to estimate the
probability of each class given the instance. The class with
the highest probability is chosen as the class of instance.
An attribute can be either qualitative or
quantitative. Discretization produces a qualitative attribute
from a quantitative attribute. Naive-Bayes classifiers can be
trained on the resulting qualitative attributes instead of the
original quantitative attributes then it increase the efficiency
of classifier. Two terminologies bias and variance are
widely used in NB discretization, which are interval
frequency (the number of training instances in one interval)
and interval number (the number of discretized intervals
produced by a specific discretization algorithm).So we have
to very careful about this two problem arising during
discretization. Yang proposed the proportion k-interval
discretization technique (PKID). PKID works based on the
fact that there is a Tradeoff between interval number,
interval frequency and the bias, variance component in the
classification error decomposition. Also, “large interval
frequency incurs low variance but high bias whereas large
interval number produces low bias and high variance".
However, PKID does not work well with small data sets,
which have at most 1200 instances. Then Ying and Webb
proposed another technique called Fixed Frequency
Discretization (FFD). FFD discretizes the training instances
into a set of intervals which contain approximately the same
number of m instances, where m is a parameter specified by
user. Note that, in FFD the interval frequency is fixed for
each interval without considering the number of training
instances. The larger the training data size is, the larger the
number of intervals is produced. However, the interval
frequency will not change.
One another thing that the above both Fixed
Frequency Discretization (FFD) and proportional K Interval
Discretization (PKID) are not support incremental approach.
Ideally, discretization should also be incremental in order to
be coupled with NB. When receiving a new training
instance, incremental discretization is expected to be able to
adjust intervals’ boundaries and statistics, using only the
current intervals and this new instance instead of re-
accessing previous training data. Unfortunately, the
majority of existing discretization methods are not oriented
to incremental learning. To update discretized intervals
with new instances, they need to add those new
instances into previous training data, and then re-
discretize on basis of the updated complete training data
set. This is detrimental to NB’s efficiency by inevitably
slowing down its learning process. Incremental Flexible
Frequency Discretization (IFFD) is the first incremental
discretization technique proposed for NB. IFFD sets the
interval frequency ranging from MinBinSize to maxBinSize
instead of single value m. The number MinBinSize and
maxBinSize stand for the minimal and maximal interval
frequency.
Some preliminary research has been already done
to enhance incremental discretization for NB. A
representative method, named PiD, proposed by Gama and
Pinto is based on two layer histograms and is efficient in
term of time and space complexity. The argument can be,
that setting the MinBinSize as a fixed number does not
guarantee that the classification performance of NB is
optimum. There exists a most suitable MinBinSize for each
dataset. Finally, propose a new incremental discretization
method: FIFFD using a sequential search.
Computerized Numerical Control (CNC) cutting
has various distinct advantages over the other cutting
technologies, such as no thermal distortion, high machining
versatility, high an effective technology for processing
various engineering materials. The mechanism and rate of
material removal during CNC machining depends both on
the type of tool and on a range flexibility and small cutting
Incremental Discretization for Naïve Bayes Learning using FIFFD
(IJSRD/Vol. 1/Issue 3/2013/0093)
All rights reserved by www.ijsrd.com 776
forces, and has been proven to be of cutting parameters.
CNC can machining the hard and brittle materials like
Steels, Non-ferrous alloys Ti alloys, Metal Matrix
Composite, Ceramic Matrix Composite, Concrete , Stone –
Granite , Wood , Reinforced plastics, Metal Polymer
Laminates .
II. DISCRETIZATION FOR NAÏVE BAYES
CLASSIFICATION
A. Naïve Bayes Classifier (NB)
Assume that an instance I is a vector of attribute
values 1, 2,..,X X Xn  , each value being an observation of
an attribute   1,X i ni  . Each instance can have a class
label  , ,,1 2C C C Ci n , being a value of the class variable
C. If an instance has a known class label, it is a training
instance. If an instance has no known class label, it is a
testing instance. The dataset of training instances is called
the training dataset. The dataset of testing instances is called
the testing dataset.
To classify an instance  1 2, ,..., nI x x x , NB estimates
the probability of each class label given I,  |iP C c I
using Formula (0.1, 0.2, 0.3,0.4). Formula (1.2) follows
(1.1) because P(I) is invariant across different class
labels and can be canceled. Formula (1.4) follows (1.3)
because of NB’s attributes independent assumption. It then
assigns the class with the highest probability to I. NB is
called naïve because it assumes that attributes are
conditionally independent of each other given the class
label. Although its assumption is sometimes violated, NB
is able to offer surprisingly good classification accuracy
in addition to its very high learning efficiency, which makes
NB popular with numerous real-world classification
applications.
   
 
|i iP C c P I C c
P I
 
 (0.1)
   |i iP C c P I C c   (0.2)
   1, 2,, |i n iP C c P X X X C c     (0.3)
   1
|
n
i j j i
j
P C c P X x C c

    (0.4)
In naïve-Bayes classifier, the class type must be qualitative
while the attribute type can be either qualitative or
quantitative. When an attribute jX is quantitative, it often
has a large or even infinite number of values. As a result, the
conditional probability that jX takes a particular value jx
given the class label ic covers very few instances if there is
any at all. Hence it is not reliable to estimate
 |j j iP X x C c  according to the observed instances. One
common practice to solve the problem of quantitative data
for NB is discretization.
Fig. 1: Block diagram for Discretization with NB
B. Discretization
Discretization is a popular approach to transforming
quantitative attributes into qualitative ones for NB. It groups
sorted values of a quantitative attribute into a sequence of
intervals, treats each interval as a qualitative value, and
maps every quantitative value into a qualitative value
according to which interval it belongs to. In the paper, the
boundaries among intervals are sometimes referred to as cut
points. The number of instances in an interval is referred to
as interval frequency. The total number of intervals
produced by discretization is referred to as interval number.
Incremental discretization aims at efficiently updating
discretization intervals and associated statistics upon
receiving each new training instance. Ideally, it does not
require to access historical training instances to carry out
the update. Instead it only needs the current intervals (with
associated statistics) and the new instance.
1) Incremental Flexible Frequency Discretization
In this section, we propose a novel incremental
discretization method, FIFFD. It is motivated by the pros
and cons of Incremental Flexible frequency discretization
(IFFD) in the context of naive-Bayes learning and
incremental learning.
a) Incremental Flexible Frequency Discretization (IFFD)
IFFD sets its interval frequency to be a range [minBinsize,
maxBinsize) instead of a single value m. The two
arguments, minBinsize and maxBinsize, are respectively the
minimum and maximum frequency that IFFD allows
intervals to assume. Whenever a new value arrives, IFFD
first inserts it into the interval that the value falls into. IFFD
then checks whether the updated interval’s frequency
reaches maxBinsize. If not, it accepts the change and update
statistics accordingly. If yes, IFFD splits the overflowed
interval into two intervals under the condition that any of the
resulting intervals has its frequency no less than minBinsize.
Otherwise, even if the interval overflows because of the
insertion, IFFD does not split it, in order to prevent high
classification variance. In the current implementation of
IFFD, minBinsize is set as 30, and maxBinsize is set as
twice of minBinsize. Assume minBinsize = 3 and hence
maxBinsize = 6. When the new attribute value “5.2” comes,
IFFD inserts it into the second interval {4.5, 5.1, 5.9}. That
interval is hence changed into {4.5, 5.1, 5.2, 5.9} whose
frequency (equal to 4) is still within [3, 6). So what we need
do is only to modify NB’s conditional probably related to
the second interval. Assume another two new attribute
values “5.4, 5.5” have come and are again inserted into the
second interval. This time, the interval {4.5, 5.1, 5.2, 5.4,
5.5, and 5.9} has a frequency as 6, reaching maxBinSize.
New
Instanc
e
Incorpo
rate
instance
Update
Discreti
zed
Data
Classifier
Discretized
Data
Incremental Discretization for Naïve Bayes Learning using FIFFD
(IJSRD/Vol. 1/Issue 3/2013/0093)
All rights reserved by www.ijsrd.com 777
Hence IFFD will split it into {4.5, 5.1, 5.2} and {5.4, 5.5,
5.9} whose frequencies are both within [3, 6). Then we only
need to recalculate NB’s conditional probabilities related to
those two intervals. By this means, IFFD makes the
update process local, affecting a minimum number of
intervals and associated statistics. As a result, incremental
discretization can be carried out very efficiently.
2) Flexible IFFD
The proposed new method FIFFD is based on the following
drawback of IFFD: there exists a most suitable MinBinSize
for the discretization intervals for each numeric attribute as
the values of numeric attributes has some distribution.
Though such a cumulative distribution does not necessarily
be Gaussian distribution, if we could approximate the
distribution using the optimal minimal discretization interval
frequency (MinBinSize), it will in turn benefit the
classification performance. It is hard to show that such an
optimal interval frequency exists theoretically because our
knowledge is very few about the data distribution, especially
for unseen data. FIFFD works as follows: instead of setting
the MinBinSize as 30 for all the data sets, we set a search
space for the most suitable MinBinSize ranging from 1 up to
the range specified by user. FIFFD works in rounds by
testing each MinBinSize values, in each round, we do a
sequential search on these range of values and set the
current value as MinBinSize and discretize the data using
IFFD based on the current MinBinSize value, we record the
classification error for each round, if the classification error
reduces once a MinBinSize is set, we will update the
MinBinSize, this search process is terminated until all
values ranging specified by user have been searched or the
classification error no longer reduces. The pseudo-code of
FIFFD is listed in Algorithm. In FIFFD, we also set the
maxBinSize as twice of MinBinSize. cut Points is the set of
cut points of discretization intervals. Counter is the
conditional probability table of the classifier. IFFD will
update the cut Points and counter according to the new
attribute value V. class Label is the class label of V. Note
the FIFFD is a sequential search based supervised approach;
the search efficiency for optimal MinBinSize is still efficient
in the context of incremental learning. Therefore, the
efficiency of FIFFD is comparable to that of IFFD.
3) Algorithm: Flexible IFFD
FIFFD (cut Points, counter, V, class Label, range) Generate
the discretized data with most suitable minBinsize value.
INPUT: V: input data,
Range: it specify the search space range.
Counter: counter is the conditional probability table.
Cut Points: cut Points is the set of cut points of
discretization intervals.
Class Label: class Label is the class label of V.
OUTPUT: discretized intervals with its most suitable
binning value.
METHOD:
Do a sequential search up to specified range and set the
current value as minBinsize,
While TRUE do
Test whether V is greater than the last cut point
If V is larger than the last cut point then
Insert V into the last interval;
Update the corresponding interval frequency;
Record changed interval;
Else
Check for other intervals;
Find the cut point and insert values in to the interval;
Update particular interval;
If frequency exceeded maximum size of interval
Get new cut points;
Insert new cut points in to cut points;
Calculate counter for each cut point;
Note down current MinBinSize and NB classification error;
Get new value for MinBinSize;
End while
Return ideal bin size;
III. RESULT ANALYSIS
A. Dataset Descriptions
In this section, we will justify our claim on the existence of
optimal minBinsize and evaluate our new discretization
method FIFFD for NB with other alternatives, including
PKID and IFFD.
We did our experiments on 4 datasets from UCI machine
learning repository. Datasets information is summarized in
Table 1. Size denotes the number of instances in a dataset,
Qa. Is the number of numeric attributes, Cat. is the number
of categorical attributes, and C means the number of
different class values. We listed the empirical result for the
existence of optimal minBinsize for each dataset in Figure 1.
Sr. No. Dataset Attributes Records Class
1 Glass 10 428 7
2 Emotion 78 1186 2
3 Sick 30 3772 2
4 Pima 9 10000 2
5 Adult 15 32560 2
6 Census 14 48998 17
Table. 1: Dataset information
Dataset FIFFD IFFD_NB PKID_NB
Glass 95.79% 82.24% 84.11%
Emotion 91.39% 89.12% 89.20%
Sick 97.00% 96.95% 96.87%
Pima 96.64% 92.47% 88.02%
Adult 84.25% 82.14% 81.82%
Census 46.97% 46.22% 46.76%
Table. 2: Naïve Bayes Accuracy comparison
Table 2 indicates that the classification performance of NB
with FIFFD is much better than that of NB with IFFD and
PKID. NB with FIFFD outperforms NB with PKID and NB
with IFFD on all 6 datasets we have tested. The reason is
that NB with FIFFD used a sequential search approach and
tried to improve the classification performance of NB as
much as possible.
Incremental Discretization for Naïve Bayes Learning using FIFFD
(IJSRD/Vol. 1/Issue 3/2013/0093)
All rights reserved by www.ijsrd.com 778
B. Analysis
Figure 2 shows the performance of accuracy study which
has been carried out on different size of datasets. The
accuracy of the proposed system has been tested for both
IFFD and PKID method. The experiment shows that the
accuracy is improved in each case for the proposed system.
It is provided that our method is best.
Fig. 2: Accuracy performance
Figure 3 show the classification error rate of Naïve Bayes
trained on most suitable Binning and MinBinSize is ranging
from 1 to 45 in below figure. It is easily concluded that a
most suitable BinSize is exists for each dataset. Fig (a)
shows the error rate of FIFFD is minimal when the
MinBinSize is 1 for Glass, Emotion and Census datasets. If
we increase the value of MinBinSize then the performance
tends to be worse. Fig (b) shows the error rate of FIFFD is
minimal when MinBinSize is 30 for Sick, 25 for German, 37
for Magik Gamma, 38 for Ecoli datasets. If we decrease the
value, performance tends to be worse.
Fig. 3: Classification Error Rate of NB
IV. CONCLUSION
We experimentally found out that a most suitable BinSize
exists for each and every datasets. The previous incremental
discretization methods for Naïve Bayes learning were
having the problem of fixed interval size that is not ideal for
all data sets. The proposed system based on sequential
search that is the incremental discretization with FIFFD can
find the ideal interval size which can make the Naïve Bayes
classifier more efficient by reducing the classification error
rate. So NB with FIFFD is much better than that of Naïve
Bayes with PKID and IFFD.
FUTURE EXTENSION
There still exists some scope for the improvement in the
proposed system. One can prove it theoretically that why
such kind of most suitable interval size or binning exists.
The second one is if try to know something more about the
data distribution and use such a domain knowledge to direct
the process of discretization.
REFERENCES
[1] Pat Langley, Wayne IBA and Kevin Thompson. “An
analysis of Bayesian Classifiers”, Tenth national
conference on AI. [Page no. 223-228].1992.
[2] Harry Z, Charles L”A Fundamental issue in Naïve
Bayes” computer science, university of new bunswick.
[Page no. 1-5]
[3] Geoffrey Webb “A Comparative Study of Discretization
Methods for naïve bayes classification “in proceeding
of PKAW. [Page no. 159-173].2002.
[4] Y.Yang “Proportional k-Interval Discretization for
Naive-Bayes Classifiers” ECML. [Page no. 564-575]
2001.
[5] Ying Yang “Weighted Proportional k-Interval
Discretization for Naive-Bayes Classifiers” In
proceedings PAKDD. [Page no. 501-512] 2003.
[6] Carlos Pinto “Partition incremental discretization” in
proceedings IEEE.
[Page no. 168-174].2005.
[7] Y.Yang “Discretization for Naïve bayes learning:
managing Bias and variance” Machin learning 74(1).
[Page no. 39-74].2009.
[8] LU, YANG and GEOFFERY I. WEB. “Incremental
Discretization for Naïve-Bayes Classifier.” ADMA.
[Page no. 223-238].2006
[9] Y.Yang “Discretization for Naïve bayes learning”
Ph.D.Thesis, School of Computer Science and Software
Engineering Monash Uni, Australia.2
Accuracy(%)
Dataset
Accuracy
OB_NB
IFFD_NB
PKID_NB

More Related Content

What's hot

11.compression technique using dct fractal compression
11.compression technique using dct fractal compression11.compression technique using dct fractal compression
11.compression technique using dct fractal compressionAlexander Decker
 
Compression technique using dct fractal compression
Compression technique using dct fractal compressionCompression technique using dct fractal compression
Compression technique using dct fractal compressionAlexander Decker
 
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...ijesajournal
 
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDING
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDINGFUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDING
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDINGIJCSEA Journal
 
Intrusion Detection System for Classification of Attacks with Cross Validation
Intrusion Detection System for Classification of Attacks with Cross ValidationIntrusion Detection System for Classification of Attacks with Cross Validation
Intrusion Detection System for Classification of Attacks with Cross Validationinventionjournals
 
3ways to improve semantic segmentation
3ways to improve semantic segmentation3ways to improve semantic segmentation
3ways to improve semantic segmentationFrozen Paradise
 
31 8949 punithavathi ramesh sub.id.69 (edit)new
31 8949 punithavathi ramesh   sub.id.69 (edit)new31 8949 punithavathi ramesh   sub.id.69 (edit)new
31 8949 punithavathi ramesh sub.id.69 (edit)newIAESIJEECS
 
New Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionNew Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionIJERA Editor
 
IRJET - Handwritten Bangla Digit Recognition using Capsule Network
IRJET -  	  Handwritten Bangla Digit Recognition using Capsule NetworkIRJET -  	  Handwritten Bangla Digit Recognition using Capsule Network
IRJET - Handwritten Bangla Digit Recognition using Capsule NetworkIRJET Journal
 
Knowledge distillation deeplab
Knowledge distillation deeplabKnowledge distillation deeplab
Knowledge distillation deeplabFrozen Paradise
 
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...IRJET Journal
 
Representation and recognition of handwirten digits using deformable templates
Representation and recognition of handwirten digits using deformable templatesRepresentation and recognition of handwirten digits using deformable templates
Representation and recognition of handwirten digits using deformable templatesAhmed Abd-Elwasaa
 
A combination of decision tree learning and clustering
A combination of decision tree learning and clusteringA combination of decision tree learning and clustering
A combination of decision tree learning and clusteringChinnapat Kaewchinporn
 

What's hot (18)

11.compression technique using dct fractal compression
11.compression technique using dct fractal compression11.compression technique using dct fractal compression
11.compression technique using dct fractal compression
 
Compression technique using dct fractal compression
Compression technique using dct fractal compressionCompression technique using dct fractal compression
Compression technique using dct fractal compression
 
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...
PERFORMANCE EVALUATION OF FUZZY LOGIC AND BACK PROPAGATION NEURAL NETWORK FOR...
 
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDING
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDINGFUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDING
FUZZY SET THEORETIC APPROACH TO IMAGE THRESHOLDING
 
Intrusion Detection System for Classification of Attacks with Cross Validation
Intrusion Detection System for Classification of Attacks with Cross ValidationIntrusion Detection System for Classification of Attacks with Cross Validation
Intrusion Detection System for Classification of Attacks with Cross Validation
 
3ways to improve semantic segmentation
3ways to improve semantic segmentation3ways to improve semantic segmentation
3ways to improve semantic segmentation
 
Ijetcas14 527
Ijetcas14 527Ijetcas14 527
Ijetcas14 527
 
31 8949 punithavathi ramesh sub.id.69 (edit)new
31 8949 punithavathi ramesh   sub.id.69 (edit)new31 8949 punithavathi ramesh   sub.id.69 (edit)new
31 8949 punithavathi ramesh sub.id.69 (edit)new
 
Et25897899
Et25897899Et25897899
Et25897899
 
New Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral RecognitionNew Approach of Preprocessing For Numeral Recognition
New Approach of Preprocessing For Numeral Recognition
 
IRJET - Handwritten Bangla Digit Recognition using Capsule Network
IRJET -  	  Handwritten Bangla Digit Recognition using Capsule NetworkIRJET -  	  Handwritten Bangla Digit Recognition using Capsule Network
IRJET - Handwritten Bangla Digit Recognition using Capsule Network
 
Knowledge distillation deeplab
Knowledge distillation deeplabKnowledge distillation deeplab
Knowledge distillation deeplab
 
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...
IRJET - Study on the Effects of Increase in the Depth of the Feature Extracto...
 
Representation and recognition of handwirten digits using deformable templates
Representation and recognition of handwirten digits using deformable templatesRepresentation and recognition of handwirten digits using deformable templates
Representation and recognition of handwirten digits using deformable templates
 
O017429398
O017429398O017429398
O017429398
 
Dr24751755
Dr24751755Dr24751755
Dr24751755
 
Be24365370
Be24365370Be24365370
Be24365370
 
A combination of decision tree learning and clustering
A combination of decision tree learning and clusteringA combination of decision tree learning and clustering
A combination of decision tree learning and clustering
 

Similar to Incremental Discretization for Naive Bayes Learning using FIFFD

Hypothesis on Different Data Mining Algorithms
Hypothesis on Different Data Mining AlgorithmsHypothesis on Different Data Mining Algorithms
Hypothesis on Different Data Mining AlgorithmsIJERA Editor
 
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...IRJET Journal
 
Analytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion miningAnalytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion miningcsandit
 
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...cscpconf
 
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGcsandit
 
IRJET- Machine Learning and Deep Learning Methods for Cybersecurity
IRJET- Machine Learning and Deep Learning Methods for CybersecurityIRJET- Machine Learning and Deep Learning Methods for Cybersecurity
IRJET- Machine Learning and Deep Learning Methods for CybersecurityIRJET Journal
 
Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)Fatimakhan325
 
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHES
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHESIMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHES
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHESVikash Kumar
 
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...IRJET Journal
 
Feature Subset Selection for High Dimensional Data using Clustering Techniques
Feature Subset Selection for High Dimensional Data using Clustering TechniquesFeature Subset Selection for High Dimensional Data using Clustering Techniques
Feature Subset Selection for High Dimensional Data using Clustering TechniquesIRJET Journal
 
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...csandit
 
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...cscpconf
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...IRJET Journal
 
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...csandit
 
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...IRJET Journal
 
Quality Prediction in Fingerprint Compression
Quality Prediction in Fingerprint CompressionQuality Prediction in Fingerprint Compression
Quality Prediction in Fingerprint CompressionIJTET Journal
 
AI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
AI Class Topic 6: Easy Way to Learn Deep Learning AI TechnologiesAI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
AI Class Topic 6: Easy Way to Learn Deep Learning AI TechnologiesValue Amplify Consulting
 
IRJET - Face Recognition in Digital Documents with Live Image
IRJET - Face Recognition in Digital Documents with Live ImageIRJET - Face Recognition in Digital Documents with Live Image
IRJET - Face Recognition in Digital Documents with Live ImageIRJET Journal
 

Similar to Incremental Discretization for Naive Bayes Learning using FIFFD (20)

Hx3115011506
Hx3115011506Hx3115011506
Hx3115011506
 
Hypothesis on Different Data Mining Algorithms
Hypothesis on Different Data Mining AlgorithmsHypothesis on Different Data Mining Algorithms
Hypothesis on Different Data Mining Algorithms
 
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...
Enhancing Classification Accuracy of K-Nearest Neighbors Algorithm using Gain...
 
Analytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion miningAnalytical study of feature extraction techniques in opinion mining
Analytical study of feature extraction techniques in opinion mining
 
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
Radial Basis Function Neural Network (RBFNN), Induction Motor, Vector control...
 
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MININGANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
ANALYTICAL STUDY OF FEATURE EXTRACTION TECHNIQUES IN OPINION MINING
 
IRJET- Machine Learning and Deep Learning Methods for Cybersecurity
IRJET- Machine Learning and Deep Learning Methods for CybersecurityIRJET- Machine Learning and Deep Learning Methods for Cybersecurity
IRJET- Machine Learning and Deep Learning Methods for Cybersecurity
 
Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)Types of Machine Learnig Algorithms(CART, ID3)
Types of Machine Learnig Algorithms(CART, ID3)
 
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHES
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHESIMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHES
IMAGE CLASSIFICATION USING DIFFERENT CLASSICAL APPROACHES
 
PNN and inversion-B
PNN and inversion-BPNN and inversion-B
PNN and inversion-B
 
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...
IRJET - Factors Affecting Deployment of Deep Learning based Face Recognition ...
 
Feature Subset Selection for High Dimensional Data using Clustering Techniques
Feature Subset Selection for High Dimensional Data using Clustering TechniquesFeature Subset Selection for High Dimensional Data using Clustering Techniques
Feature Subset Selection for High Dimensional Data using Clustering Techniques
 
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
 
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
 
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
 
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
IMPROVING SUPERVISED CLASSIFICATION OF DAILY ACTIVITIES LIVING USING NEW COST...
 
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...
IRJET- Evaluation of Classification Algorithms with Solutions to Class Imbala...
 
Quality Prediction in Fingerprint Compression
Quality Prediction in Fingerprint CompressionQuality Prediction in Fingerprint Compression
Quality Prediction in Fingerprint Compression
 
AI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
AI Class Topic 6: Easy Way to Learn Deep Learning AI TechnologiesAI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
AI Class Topic 6: Easy Way to Learn Deep Learning AI Technologies
 
IRJET - Face Recognition in Digital Documents with Live Image
IRJET - Face Recognition in Digital Documents with Live ImageIRJET - Face Recognition in Digital Documents with Live Image
IRJET - Face Recognition in Digital Documents with Live Image
 

More from ijsrd.com

IoT Enabled Smart Grid
IoT Enabled Smart GridIoT Enabled Smart Grid
IoT Enabled Smart Gridijsrd.com
 
A Survey Report on : Security & Challenges in Internet of Things
A Survey Report on : Security & Challenges in Internet of ThingsA Survey Report on : Security & Challenges in Internet of Things
A Survey Report on : Security & Challenges in Internet of Thingsijsrd.com
 
IoT for Everyday Life
IoT for Everyday LifeIoT for Everyday Life
IoT for Everyday Lifeijsrd.com
 
Study on Issues in Managing and Protecting Data of IOT
Study on Issues in Managing and Protecting Data of IOTStudy on Issues in Managing and Protecting Data of IOT
Study on Issues in Managing and Protecting Data of IOTijsrd.com
 
Interactive Technologies for Improving Quality of Education to Build Collabor...
Interactive Technologies for Improving Quality of Education to Build Collabor...Interactive Technologies for Improving Quality of Education to Build Collabor...
Interactive Technologies for Improving Quality of Education to Build Collabor...ijsrd.com
 
Internet of Things - Paradigm Shift of Future Internet Application for Specia...
Internet of Things - Paradigm Shift of Future Internet Application for Specia...Internet of Things - Paradigm Shift of Future Internet Application for Specia...
Internet of Things - Paradigm Shift of Future Internet Application for Specia...ijsrd.com
 
A Study of the Adverse Effects of IoT on Student's Life
A Study of the Adverse Effects of IoT on Student's LifeA Study of the Adverse Effects of IoT on Student's Life
A Study of the Adverse Effects of IoT on Student's Lifeijsrd.com
 
Pedagogy for Effective use of ICT in English Language Learning
Pedagogy for Effective use of ICT in English Language LearningPedagogy for Effective use of ICT in English Language Learning
Pedagogy for Effective use of ICT in English Language Learningijsrd.com
 
Virtual Eye - Smart Traffic Navigation System
Virtual Eye - Smart Traffic Navigation SystemVirtual Eye - Smart Traffic Navigation System
Virtual Eye - Smart Traffic Navigation Systemijsrd.com
 
Ontological Model of Educational Programs in Computer Science (Bachelor and M...
Ontological Model of Educational Programs in Computer Science (Bachelor and M...Ontological Model of Educational Programs in Computer Science (Bachelor and M...
Ontological Model of Educational Programs in Computer Science (Bachelor and M...ijsrd.com
 
Understanding IoT Management for Smart Refrigerator
Understanding IoT Management for Smart RefrigeratorUnderstanding IoT Management for Smart Refrigerator
Understanding IoT Management for Smart Refrigeratorijsrd.com
 
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...ijsrd.com
 
A Review: Microwave Energy for materials processing
A Review: Microwave Energy for materials processingA Review: Microwave Energy for materials processing
A Review: Microwave Energy for materials processingijsrd.com
 
Web Usage Mining: A Survey on User's Navigation Pattern from Web Logs
Web Usage Mining: A Survey on User's Navigation Pattern from Web LogsWeb Usage Mining: A Survey on User's Navigation Pattern from Web Logs
Web Usage Mining: A Survey on User's Navigation Pattern from Web Logsijsrd.com
 
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEM
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEMAPPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEM
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEMijsrd.com
 
Making model of dual axis solar tracking with Maximum Power Point Tracking
Making model of dual axis solar tracking with Maximum Power Point TrackingMaking model of dual axis solar tracking with Maximum Power Point Tracking
Making model of dual axis solar tracking with Maximum Power Point Trackingijsrd.com
 
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...ijsrd.com
 
Study and Review on Various Current Comparators
Study and Review on Various Current ComparatorsStudy and Review on Various Current Comparators
Study and Review on Various Current Comparatorsijsrd.com
 
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...ijsrd.com
 
Defending Reactive Jammers in WSN using a Trigger Identification Service.
Defending Reactive Jammers in WSN using a Trigger Identification Service.Defending Reactive Jammers in WSN using a Trigger Identification Service.
Defending Reactive Jammers in WSN using a Trigger Identification Service.ijsrd.com
 

More from ijsrd.com (20)

IoT Enabled Smart Grid
IoT Enabled Smart GridIoT Enabled Smart Grid
IoT Enabled Smart Grid
 
A Survey Report on : Security & Challenges in Internet of Things
A Survey Report on : Security & Challenges in Internet of ThingsA Survey Report on : Security & Challenges in Internet of Things
A Survey Report on : Security & Challenges in Internet of Things
 
IoT for Everyday Life
IoT for Everyday LifeIoT for Everyday Life
IoT for Everyday Life
 
Study on Issues in Managing and Protecting Data of IOT
Study on Issues in Managing and Protecting Data of IOTStudy on Issues in Managing and Protecting Data of IOT
Study on Issues in Managing and Protecting Data of IOT
 
Interactive Technologies for Improving Quality of Education to Build Collabor...
Interactive Technologies for Improving Quality of Education to Build Collabor...Interactive Technologies for Improving Quality of Education to Build Collabor...
Interactive Technologies for Improving Quality of Education to Build Collabor...
 
Internet of Things - Paradigm Shift of Future Internet Application for Specia...
Internet of Things - Paradigm Shift of Future Internet Application for Specia...Internet of Things - Paradigm Shift of Future Internet Application for Specia...
Internet of Things - Paradigm Shift of Future Internet Application for Specia...
 
A Study of the Adverse Effects of IoT on Student's Life
A Study of the Adverse Effects of IoT on Student's LifeA Study of the Adverse Effects of IoT on Student's Life
A Study of the Adverse Effects of IoT on Student's Life
 
Pedagogy for Effective use of ICT in English Language Learning
Pedagogy for Effective use of ICT in English Language LearningPedagogy for Effective use of ICT in English Language Learning
Pedagogy for Effective use of ICT in English Language Learning
 
Virtual Eye - Smart Traffic Navigation System
Virtual Eye - Smart Traffic Navigation SystemVirtual Eye - Smart Traffic Navigation System
Virtual Eye - Smart Traffic Navigation System
 
Ontological Model of Educational Programs in Computer Science (Bachelor and M...
Ontological Model of Educational Programs in Computer Science (Bachelor and M...Ontological Model of Educational Programs in Computer Science (Bachelor and M...
Ontological Model of Educational Programs in Computer Science (Bachelor and M...
 
Understanding IoT Management for Smart Refrigerator
Understanding IoT Management for Smart RefrigeratorUnderstanding IoT Management for Smart Refrigerator
Understanding IoT Management for Smart Refrigerator
 
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...
DESIGN AND ANALYSIS OF DOUBLE WISHBONE SUSPENSION SYSTEM USING FINITE ELEMENT...
 
A Review: Microwave Energy for materials processing
A Review: Microwave Energy for materials processingA Review: Microwave Energy for materials processing
A Review: Microwave Energy for materials processing
 
Web Usage Mining: A Survey on User's Navigation Pattern from Web Logs
Web Usage Mining: A Survey on User's Navigation Pattern from Web LogsWeb Usage Mining: A Survey on User's Navigation Pattern from Web Logs
Web Usage Mining: A Survey on User's Navigation Pattern from Web Logs
 
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEM
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEMAPPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEM
APPLICATION OF STATCOM to IMPROVED DYNAMIC PERFORMANCE OF POWER SYSTEM
 
Making model of dual axis solar tracking with Maximum Power Point Tracking
Making model of dual axis solar tracking with Maximum Power Point TrackingMaking model of dual axis solar tracking with Maximum Power Point Tracking
Making model of dual axis solar tracking with Maximum Power Point Tracking
 
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...
A REVIEW PAPER ON PERFORMANCE AND EMISSION TEST OF 4 STROKE DIESEL ENGINE USI...
 
Study and Review on Various Current Comparators
Study and Review on Various Current ComparatorsStudy and Review on Various Current Comparators
Study and Review on Various Current Comparators
 
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...
Reducing Silicon Real Estate and Switching Activity Using Low Power Test Patt...
 
Defending Reactive Jammers in WSN using a Trigger Identification Service.
Defending Reactive Jammers in WSN using a Trigger Identification Service.Defending Reactive Jammers in WSN using a Trigger Identification Service.
Defending Reactive Jammers in WSN using a Trigger Identification Service.
 

Recently uploaded

Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleAlluxio, Inc.
 
Risk Management in Engineering Construction Project
Risk Management in Engineering Construction ProjectRisk Management in Engineering Construction Project
Risk Management in Engineering Construction ProjectErbil Polytechnic University
 
Ch10-Global Supply Chain - Cadena de Suministro.pdf
Ch10-Global Supply Chain - Cadena de Suministro.pdfCh10-Global Supply Chain - Cadena de Suministro.pdf
Ch10-Global Supply Chain - Cadena de Suministro.pdfChristianCDAM
 
Indian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptIndian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptMadan Karki
 
DM Pillar Training Manual.ppt will be useful in deploying TPM in project
DM Pillar Training Manual.ppt will be useful in deploying TPM in projectDM Pillar Training Manual.ppt will be useful in deploying TPM in project
DM Pillar Training Manual.ppt will be useful in deploying TPM in projectssuserb6619e
 
Mine Environment II Lab_MI10448MI__________.pptx
Mine Environment II Lab_MI10448MI__________.pptxMine Environment II Lab_MI10448MI__________.pptx
Mine Environment II Lab_MI10448MI__________.pptxRomil Mishra
 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionMebane Rash
 
Industrial Safety Unit-IV workplace health and safety.ppt
Industrial Safety Unit-IV workplace health and safety.pptIndustrial Safety Unit-IV workplace health and safety.ppt
Industrial Safety Unit-IV workplace health and safety.pptNarmatha D
 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm Systemirfanmechengr
 
Engineering Drawing section of solid
Engineering Drawing     section of solidEngineering Drawing     section of solid
Engineering Drawing section of solidnamansinghjarodiya
 
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...Erbil Polytechnic University
 
Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...121011101441
 
National Level Hackathon Participation Certificate.pdf
National Level Hackathon Participation Certificate.pdfNational Level Hackathon Participation Certificate.pdf
National Level Hackathon Participation Certificate.pdfRajuKanojiya4
 
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTION
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTIONTHE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTION
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTIONjhunlian
 
Earthing details of Electrical Substation
Earthing details of Electrical SubstationEarthing details of Electrical Substation
Earthing details of Electrical Substationstephanwindworld
 
Autonomous emergency braking system (aeb) ppt.ppt
Autonomous emergency braking system (aeb) ppt.pptAutonomous emergency braking system (aeb) ppt.ppt
Autonomous emergency braking system (aeb) ppt.pptbibisarnayak0
 
Configuration of IoT devices - Systems managament
Configuration of IoT devices - Systems managamentConfiguration of IoT devices - Systems managament
Configuration of IoT devices - Systems managamentBharaniDharan195623
 
11. Properties of Liquid Fuels in Energy Engineering.pdf
11. Properties of Liquid Fuels in Energy Engineering.pdf11. Properties of Liquid Fuels in Energy Engineering.pdf
11. Properties of Liquid Fuels in Energy Engineering.pdfHafizMudaserAhmad
 
Transport layer issues and challenges - Guide
Transport layer issues and challenges - GuideTransport layer issues and challenges - Guide
Transport layer issues and challenges - GuideGOPINATHS437943
 

Recently uploaded (20)

Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at Scale
 
Risk Management in Engineering Construction Project
Risk Management in Engineering Construction ProjectRisk Management in Engineering Construction Project
Risk Management in Engineering Construction Project
 
Ch10-Global Supply Chain - Cadena de Suministro.pdf
Ch10-Global Supply Chain - Cadena de Suministro.pdfCh10-Global Supply Chain - Cadena de Suministro.pdf
Ch10-Global Supply Chain - Cadena de Suministro.pdf
 
Indian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptIndian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.ppt
 
DM Pillar Training Manual.ppt will be useful in deploying TPM in project
DM Pillar Training Manual.ppt will be useful in deploying TPM in projectDM Pillar Training Manual.ppt will be useful in deploying TPM in project
DM Pillar Training Manual.ppt will be useful in deploying TPM in project
 
Mine Environment II Lab_MI10448MI__________.pptx
Mine Environment II Lab_MI10448MI__________.pptxMine Environment II Lab_MI10448MI__________.pptx
Mine Environment II Lab_MI10448MI__________.pptx
 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of Action
 
Industrial Safety Unit-IV workplace health and safety.ppt
Industrial Safety Unit-IV workplace health and safety.pptIndustrial Safety Unit-IV workplace health and safety.ppt
Industrial Safety Unit-IV workplace health and safety.ppt
 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm System
 
Engineering Drawing section of solid
Engineering Drawing     section of solidEngineering Drawing     section of solid
Engineering Drawing section of solid
 
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...
Comparative study of High-rise Building Using ETABS,SAP200 and SAFE., SAFE an...
 
Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...
 
National Level Hackathon Participation Certificate.pdf
National Level Hackathon Participation Certificate.pdfNational Level Hackathon Participation Certificate.pdf
National Level Hackathon Participation Certificate.pdf
 
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTION
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTIONTHE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTION
THE SENDAI FRAMEWORK FOR DISASTER RISK REDUCTION
 
Earthing details of Electrical Substation
Earthing details of Electrical SubstationEarthing details of Electrical Substation
Earthing details of Electrical Substation
 
Autonomous emergency braking system (aeb) ppt.ppt
Autonomous emergency braking system (aeb) ppt.pptAutonomous emergency braking system (aeb) ppt.ppt
Autonomous emergency braking system (aeb) ppt.ppt
 
Configuration of IoT devices - Systems managament
Configuration of IoT devices - Systems managamentConfiguration of IoT devices - Systems managament
Configuration of IoT devices - Systems managament
 
11. Properties of Liquid Fuels in Energy Engineering.pdf
11. Properties of Liquid Fuels in Energy Engineering.pdf11. Properties of Liquid Fuels in Energy Engineering.pdf
11. Properties of Liquid Fuels in Energy Engineering.pdf
 
Transport layer issues and challenges - Guide
Transport layer issues and challenges - GuideTransport layer issues and challenges - Guide
Transport layer issues and challenges - Guide
 
Designing pile caps according to ACI 318-19.pptx
Designing pile caps according to ACI 318-19.pptxDesigning pile caps according to ACI 318-19.pptx
Designing pile caps according to ACI 318-19.pptx
 

Incremental Discretization for Naive Bayes Learning using FIFFD

  • 1. IJSRD - International Journal for Scientific Research & Development| Vol. 1, Issue 3, 2013 | ISSN (online): 2321-0613 All rights reserved by www.ijsrd.com 775 Incremental Discretization for Naïve Bayes Learning using FIFFD Mr. Kunal Khimani 1 Mr. Kamal Sutaria 2 Ms. Kruti Khalpada 3 1 Gujarat Technological University PG School, Ahmedabad, Gujarat 2 Asst. Prof., C.E. Department, VVP Engineering College, Rajkot Gujarat 3 Institute of Technology, Nirma University, Ahmedabad, Gujarat Abstract—Incremental Flexible Frequency Discretization (IFFD) is a recently proposed discretization approach for Naïve Bayes (NB). IFFD performs satisfactory by setting the minimal interval frequency for discretized intervals as a fixed number. In this paper, we first argue that this setting cannot guarantee that the selecting MinBinSize is on always optimal for all the different datasets. So the performance of Naïve Bayes is not good in terms of classification error. We thus proposed a sequential search method for NB: named Flexible IFFD. Experiments were conducted on 4 datasets from UCI machine learning repository and performance was compared between NB trained on the data discretized by FIFFD, IFFD, and PKID. Keywords: Discretization, incremental, Naïve Bayes. I. INTRODUCTION Naive-Bayes classifiers are widely employed for classification tasks because of their efficiency and efficacy. Naïve Bayesian classifiers are simple, robust, and also support incremental training. Its efficiency is witnessed widespread deployment in classification task. Naïve Bayesian classifiers have long been a core technique in information retrieval. Naive-Bayesian learning needs to estimate probabilities for each attribute-class pair. The naïve Bayesian classifier provides a very simple and yet surprisingly accurate technique for machine learning. The naïve Bayesian classifier provides a very simple and yet surprisingly accurate technique for machine learning. When classifying an instance, naïve Bayesian classifiers assume the attribute conditionally independent of each other given the class; then apply the Bayes’ theorem to estimate the probability of each class given the instance. The class with the highest probability is chosen as the class of instance. An attribute can be either qualitative or quantitative. Discretization produces a qualitative attribute from a quantitative attribute. Naive-Bayes classifiers can be trained on the resulting qualitative attributes instead of the original quantitative attributes then it increase the efficiency of classifier. Two terminologies bias and variance are widely used in NB discretization, which are interval frequency (the number of training instances in one interval) and interval number (the number of discretized intervals produced by a specific discretization algorithm).So we have to very careful about this two problem arising during discretization. Yang proposed the proportion k-interval discretization technique (PKID). PKID works based on the fact that there is a Tradeoff between interval number, interval frequency and the bias, variance component in the classification error decomposition. Also, “large interval frequency incurs low variance but high bias whereas large interval number produces low bias and high variance". However, PKID does not work well with small data sets, which have at most 1200 instances. Then Ying and Webb proposed another technique called Fixed Frequency Discretization (FFD). FFD discretizes the training instances into a set of intervals which contain approximately the same number of m instances, where m is a parameter specified by user. Note that, in FFD the interval frequency is fixed for each interval without considering the number of training instances. The larger the training data size is, the larger the number of intervals is produced. However, the interval frequency will not change. One another thing that the above both Fixed Frequency Discretization (FFD) and proportional K Interval Discretization (PKID) are not support incremental approach. Ideally, discretization should also be incremental in order to be coupled with NB. When receiving a new training instance, incremental discretization is expected to be able to adjust intervals’ boundaries and statistics, using only the current intervals and this new instance instead of re- accessing previous training data. Unfortunately, the majority of existing discretization methods are not oriented to incremental learning. To update discretized intervals with new instances, they need to add those new instances into previous training data, and then re- discretize on basis of the updated complete training data set. This is detrimental to NB’s efficiency by inevitably slowing down its learning process. Incremental Flexible Frequency Discretization (IFFD) is the first incremental discretization technique proposed for NB. IFFD sets the interval frequency ranging from MinBinSize to maxBinSize instead of single value m. The number MinBinSize and maxBinSize stand for the minimal and maximal interval frequency. Some preliminary research has been already done to enhance incremental discretization for NB. A representative method, named PiD, proposed by Gama and Pinto is based on two layer histograms and is efficient in term of time and space complexity. The argument can be, that setting the MinBinSize as a fixed number does not guarantee that the classification performance of NB is optimum. There exists a most suitable MinBinSize for each dataset. Finally, propose a new incremental discretization method: FIFFD using a sequential search. Computerized Numerical Control (CNC) cutting has various distinct advantages over the other cutting technologies, such as no thermal distortion, high machining versatility, high an effective technology for processing various engineering materials. The mechanism and rate of material removal during CNC machining depends both on the type of tool and on a range flexibility and small cutting
  • 2. Incremental Discretization for Naïve Bayes Learning using FIFFD (IJSRD/Vol. 1/Issue 3/2013/0093) All rights reserved by www.ijsrd.com 776 forces, and has been proven to be of cutting parameters. CNC can machining the hard and brittle materials like Steels, Non-ferrous alloys Ti alloys, Metal Matrix Composite, Ceramic Matrix Composite, Concrete , Stone – Granite , Wood , Reinforced plastics, Metal Polymer Laminates . II. DISCRETIZATION FOR NAÏVE BAYES CLASSIFICATION A. Naïve Bayes Classifier (NB) Assume that an instance I is a vector of attribute values 1, 2,..,X X Xn  , each value being an observation of an attribute   1,X i ni  . Each instance can have a class label  , ,,1 2C C C Ci n , being a value of the class variable C. If an instance has a known class label, it is a training instance. If an instance has no known class label, it is a testing instance. The dataset of training instances is called the training dataset. The dataset of testing instances is called the testing dataset. To classify an instance  1 2, ,..., nI x x x , NB estimates the probability of each class label given I,  |iP C c I using Formula (0.1, 0.2, 0.3,0.4). Formula (1.2) follows (1.1) because P(I) is invariant across different class labels and can be canceled. Formula (1.4) follows (1.3) because of NB’s attributes independent assumption. It then assigns the class with the highest probability to I. NB is called naïve because it assumes that attributes are conditionally independent of each other given the class label. Although its assumption is sometimes violated, NB is able to offer surprisingly good classification accuracy in addition to its very high learning efficiency, which makes NB popular with numerous real-world classification applications.       |i iP C c P I C c P I    (0.1)    |i iP C c P I C c   (0.2)    1, 2,, |i n iP C c P X X X C c     (0.3)    1 | n i j j i j P C c P X x C c      (0.4) In naïve-Bayes classifier, the class type must be qualitative while the attribute type can be either qualitative or quantitative. When an attribute jX is quantitative, it often has a large or even infinite number of values. As a result, the conditional probability that jX takes a particular value jx given the class label ic covers very few instances if there is any at all. Hence it is not reliable to estimate  |j j iP X x C c  according to the observed instances. One common practice to solve the problem of quantitative data for NB is discretization. Fig. 1: Block diagram for Discretization with NB B. Discretization Discretization is a popular approach to transforming quantitative attributes into qualitative ones for NB. It groups sorted values of a quantitative attribute into a sequence of intervals, treats each interval as a qualitative value, and maps every quantitative value into a qualitative value according to which interval it belongs to. In the paper, the boundaries among intervals are sometimes referred to as cut points. The number of instances in an interval is referred to as interval frequency. The total number of intervals produced by discretization is referred to as interval number. Incremental discretization aims at efficiently updating discretization intervals and associated statistics upon receiving each new training instance. Ideally, it does not require to access historical training instances to carry out the update. Instead it only needs the current intervals (with associated statistics) and the new instance. 1) Incremental Flexible Frequency Discretization In this section, we propose a novel incremental discretization method, FIFFD. It is motivated by the pros and cons of Incremental Flexible frequency discretization (IFFD) in the context of naive-Bayes learning and incremental learning. a) Incremental Flexible Frequency Discretization (IFFD) IFFD sets its interval frequency to be a range [minBinsize, maxBinsize) instead of a single value m. The two arguments, minBinsize and maxBinsize, are respectively the minimum and maximum frequency that IFFD allows intervals to assume. Whenever a new value arrives, IFFD first inserts it into the interval that the value falls into. IFFD then checks whether the updated interval’s frequency reaches maxBinsize. If not, it accepts the change and update statistics accordingly. If yes, IFFD splits the overflowed interval into two intervals under the condition that any of the resulting intervals has its frequency no less than minBinsize. Otherwise, even if the interval overflows because of the insertion, IFFD does not split it, in order to prevent high classification variance. In the current implementation of IFFD, minBinsize is set as 30, and maxBinsize is set as twice of minBinsize. Assume minBinsize = 3 and hence maxBinsize = 6. When the new attribute value “5.2” comes, IFFD inserts it into the second interval {4.5, 5.1, 5.9}. That interval is hence changed into {4.5, 5.1, 5.2, 5.9} whose frequency (equal to 4) is still within [3, 6). So what we need do is only to modify NB’s conditional probably related to the second interval. Assume another two new attribute values “5.4, 5.5” have come and are again inserted into the second interval. This time, the interval {4.5, 5.1, 5.2, 5.4, 5.5, and 5.9} has a frequency as 6, reaching maxBinSize. New Instanc e Incorpo rate instance Update Discreti zed Data Classifier Discretized Data
  • 3. Incremental Discretization for Naïve Bayes Learning using FIFFD (IJSRD/Vol. 1/Issue 3/2013/0093) All rights reserved by www.ijsrd.com 777 Hence IFFD will split it into {4.5, 5.1, 5.2} and {5.4, 5.5, 5.9} whose frequencies are both within [3, 6). Then we only need to recalculate NB’s conditional probabilities related to those two intervals. By this means, IFFD makes the update process local, affecting a minimum number of intervals and associated statistics. As a result, incremental discretization can be carried out very efficiently. 2) Flexible IFFD The proposed new method FIFFD is based on the following drawback of IFFD: there exists a most suitable MinBinSize for the discretization intervals for each numeric attribute as the values of numeric attributes has some distribution. Though such a cumulative distribution does not necessarily be Gaussian distribution, if we could approximate the distribution using the optimal minimal discretization interval frequency (MinBinSize), it will in turn benefit the classification performance. It is hard to show that such an optimal interval frequency exists theoretically because our knowledge is very few about the data distribution, especially for unseen data. FIFFD works as follows: instead of setting the MinBinSize as 30 for all the data sets, we set a search space for the most suitable MinBinSize ranging from 1 up to the range specified by user. FIFFD works in rounds by testing each MinBinSize values, in each round, we do a sequential search on these range of values and set the current value as MinBinSize and discretize the data using IFFD based on the current MinBinSize value, we record the classification error for each round, if the classification error reduces once a MinBinSize is set, we will update the MinBinSize, this search process is terminated until all values ranging specified by user have been searched or the classification error no longer reduces. The pseudo-code of FIFFD is listed in Algorithm. In FIFFD, we also set the maxBinSize as twice of MinBinSize. cut Points is the set of cut points of discretization intervals. Counter is the conditional probability table of the classifier. IFFD will update the cut Points and counter according to the new attribute value V. class Label is the class label of V. Note the FIFFD is a sequential search based supervised approach; the search efficiency for optimal MinBinSize is still efficient in the context of incremental learning. Therefore, the efficiency of FIFFD is comparable to that of IFFD. 3) Algorithm: Flexible IFFD FIFFD (cut Points, counter, V, class Label, range) Generate the discretized data with most suitable minBinsize value. INPUT: V: input data, Range: it specify the search space range. Counter: counter is the conditional probability table. Cut Points: cut Points is the set of cut points of discretization intervals. Class Label: class Label is the class label of V. OUTPUT: discretized intervals with its most suitable binning value. METHOD: Do a sequential search up to specified range and set the current value as minBinsize, While TRUE do Test whether V is greater than the last cut point If V is larger than the last cut point then Insert V into the last interval; Update the corresponding interval frequency; Record changed interval; Else Check for other intervals; Find the cut point and insert values in to the interval; Update particular interval; If frequency exceeded maximum size of interval Get new cut points; Insert new cut points in to cut points; Calculate counter for each cut point; Note down current MinBinSize and NB classification error; Get new value for MinBinSize; End while Return ideal bin size; III. RESULT ANALYSIS A. Dataset Descriptions In this section, we will justify our claim on the existence of optimal minBinsize and evaluate our new discretization method FIFFD for NB with other alternatives, including PKID and IFFD. We did our experiments on 4 datasets from UCI machine learning repository. Datasets information is summarized in Table 1. Size denotes the number of instances in a dataset, Qa. Is the number of numeric attributes, Cat. is the number of categorical attributes, and C means the number of different class values. We listed the empirical result for the existence of optimal minBinsize for each dataset in Figure 1. Sr. No. Dataset Attributes Records Class 1 Glass 10 428 7 2 Emotion 78 1186 2 3 Sick 30 3772 2 4 Pima 9 10000 2 5 Adult 15 32560 2 6 Census 14 48998 17 Table. 1: Dataset information Dataset FIFFD IFFD_NB PKID_NB Glass 95.79% 82.24% 84.11% Emotion 91.39% 89.12% 89.20% Sick 97.00% 96.95% 96.87% Pima 96.64% 92.47% 88.02% Adult 84.25% 82.14% 81.82% Census 46.97% 46.22% 46.76% Table. 2: Naïve Bayes Accuracy comparison Table 2 indicates that the classification performance of NB with FIFFD is much better than that of NB with IFFD and PKID. NB with FIFFD outperforms NB with PKID and NB with IFFD on all 6 datasets we have tested. The reason is that NB with FIFFD used a sequential search approach and tried to improve the classification performance of NB as much as possible.
  • 4. Incremental Discretization for Naïve Bayes Learning using FIFFD (IJSRD/Vol. 1/Issue 3/2013/0093) All rights reserved by www.ijsrd.com 778 B. Analysis Figure 2 shows the performance of accuracy study which has been carried out on different size of datasets. The accuracy of the proposed system has been tested for both IFFD and PKID method. The experiment shows that the accuracy is improved in each case for the proposed system. It is provided that our method is best. Fig. 2: Accuracy performance Figure 3 show the classification error rate of Naïve Bayes trained on most suitable Binning and MinBinSize is ranging from 1 to 45 in below figure. It is easily concluded that a most suitable BinSize is exists for each dataset. Fig (a) shows the error rate of FIFFD is minimal when the MinBinSize is 1 for Glass, Emotion and Census datasets. If we increase the value of MinBinSize then the performance tends to be worse. Fig (b) shows the error rate of FIFFD is minimal when MinBinSize is 30 for Sick, 25 for German, 37 for Magik Gamma, 38 for Ecoli datasets. If we decrease the value, performance tends to be worse. Fig. 3: Classification Error Rate of NB IV. CONCLUSION We experimentally found out that a most suitable BinSize exists for each and every datasets. The previous incremental discretization methods for Naïve Bayes learning were having the problem of fixed interval size that is not ideal for all data sets. The proposed system based on sequential search that is the incremental discretization with FIFFD can find the ideal interval size which can make the Naïve Bayes classifier more efficient by reducing the classification error rate. So NB with FIFFD is much better than that of Naïve Bayes with PKID and IFFD. FUTURE EXTENSION There still exists some scope for the improvement in the proposed system. One can prove it theoretically that why such kind of most suitable interval size or binning exists. The second one is if try to know something more about the data distribution and use such a domain knowledge to direct the process of discretization. REFERENCES [1] Pat Langley, Wayne IBA and Kevin Thompson. “An analysis of Bayesian Classifiers”, Tenth national conference on AI. [Page no. 223-228].1992. [2] Harry Z, Charles L”A Fundamental issue in Naïve Bayes” computer science, university of new bunswick. [Page no. 1-5] [3] Geoffrey Webb “A Comparative Study of Discretization Methods for naïve bayes classification “in proceeding of PKAW. [Page no. 159-173].2002. [4] Y.Yang “Proportional k-Interval Discretization for Naive-Bayes Classifiers” ECML. [Page no. 564-575] 2001. [5] Ying Yang “Weighted Proportional k-Interval Discretization for Naive-Bayes Classifiers” In proceedings PAKDD. [Page no. 501-512] 2003. [6] Carlos Pinto “Partition incremental discretization” in proceedings IEEE. [Page no. 168-174].2005. [7] Y.Yang “Discretization for Naïve bayes learning: managing Bias and variance” Machin learning 74(1). [Page no. 39-74].2009. [8] LU, YANG and GEOFFERY I. WEB. “Incremental Discretization for Naïve-Bayes Classifier.” ADMA. [Page no. 223-238].2006 [9] Y.Yang “Discretization for Naïve bayes learning” Ph.D.Thesis, School of Computer Science and Software Engineering Monash Uni, Australia.2 Accuracy(%) Dataset Accuracy OB_NB IFFD_NB PKID_NB