SlideShare a Scribd company logo
1 of 12
Download to read offline
1
Varying parameter in classification
based on imprecise probabilities
SMPS
Bristol, September 2006
J. Abellán, S. Moral, M. Gómez, A. Masegosa
Department of Computer Science and AI
University of Granada
2
Index
1. Classifiying with Decision Trees.
2. Decision Trees with Imprecise
Probabilities.
2.1. Imprecise Dirichlet Model.
2.2. Classification Method.
3. Experimentation.
4. Conclusions and Future Work.
3
1. Classifying with Decision Trees
In a classification problem, there is a data set D with values of
a set of attribute variables X (petal width, petal length, sepal
width, sepal length) and a class variable C (Iris Flower
Type:Setosa, Versicolor, Virginica).
A decision tree is a tree model with an attribute variable in
each node and a value of the class variable in the leafs.
Petal-W
small large
Setal-W
small large
Petal-L
small large
Versicolor Setosa ViginicaVersicolor
4
1. Classifying with Decision Trees
In a Decision Tree, the path from the root node to another node
defines a configuration. A configuration σ is a m-tuple with a
set of fixed values for several attribute variables (Petal-Width =
small, Sepal-Width = small).
ID3 of Quinlan is based in the Shannon entropy. It is measured
the reduction in the uncertainty that provokes the introduction
of one variable in the model:
The entropy measures the uncertainty about C.
)|()()|( XCHCHXCInfoGain −=
∑∑ ⋅−==⋅=−=
j
jj
j c
cc
c
jj
N
n
N
n
cCPcCPCH )(log)|(log)|()|(:Enropy 22 σ
σ
σ
σ
σσσ
5
1. Classifying with Decision Trees
Example: Iris Subtype Classifcation.
IG(Class|Petal-W) =0.26
IG(Class|Petal-L) =0.15
IG(Class|Sepal-L) =0.12
IG(Class|Sepal-W)=0.11
Petal-W
small large
Setal-W
small large
Petal-L
small large
Versicolor Setosa ViginicaVersicolor
IG(Class|Petal-W=large, Petal-L) =0.13
IG(Class|Petal-W=large, Setal-W) =0.08
IG(Class|Petal-W=large, Setal-L) =-0.05
IG(Class|Petal-W=small, Setal-W) =0.11
IG(Class|Petal-W=small, Petal-L) =0.04
IG(Class|Petal-W=small, Setal-L) =-0.02
IG(Class|Petal-W=small, Setal-W=small, Petal-L) = - 0.08
IG(Class|Petal-W=small, Setal-W=small, Petal-W) = - 0.12
?
6
2. Decision Trees with Imprecise Probabilities
Recently, Abellán,Moral [7] have introduced a new
algorithm to build decision trees based on Imprecise
Dirichlet Model of Walley [22].
The authors use of the maximum entropy on credal
sets as a good measure of total uncertainty [7,16].
In this way, the structure of the decision tree is exactly
the same, the difference is in the criterium to select
variables in the ramification.
Using this new entropy criterium, the performance is
better than ID3.
7
2.1. Imprecise Dirichlet Model (IDM) [22]
In this model, the probabilities are estimated as an interval.
Given a data set D and a configuration σ. We shall consider a
credal set for the class variable C respect to σ defined by the
set of probabilities distributions, p = (p1,p2,…,pk), such that:
These intervals are obtained in basis of IDM and they depends
of the real value s > 0 that is a hyperparemeter of the model.
For the IDM, Walley[22] suggests a value for s between 1 and 2.
“Classification performance effects of varying the paremeter s”.
σ
CΡ
[ ])|(),|(,)|( σσσ σ
σ
σ
σ
jj
cc
jj cPcP
sN
sn
sN
n
cCPp jj
=
⎥
⎥
⎦
⎤
⎢
⎢
⎣
⎡
+
+
+
∈==
8
2.2 The Classification Method
This method [7] evaluates a single variable and, also, a couple
of variable for the ramification in each step.
There will be considered the following functions:
If Max{Inf1(X)} > Max{Inf2(Y,Z)}, it is selected X.
If Max{Inf2(Y,Z)} > Max{Inf1(X)}, it is selected Y if
Inf(Y)>Inf(Z) and it is selected Z, otherwise.
When the inclusion of any variable does not reduce the
uncertainty of the class variable, a leaf is introduced.The most
frequence class in the subset of data compatible with its
configuration is associated to this leaf.
( )}{
),(1 i
i
xX
C
x
i
TU
N
n
XInf =
Ρ⋅= ∑ Iσ
σ
σ
σ { }
( )ji
ji
yYxX
C
yx
ji
TU
N
n
YXInf
==
Ρ= ∑
,
,
,
),,(2
Iσ
σ
σ
σ
9
3. Experimentation
The evaluation was carried out over 8 data sets from
UCI repository:
Discretized with Fayyad and Irani’s procedure [14].
It was considered the following set of value for s:
{0.5, 0.1, 1.5, 2, 2.5, 3, 3.5, 4, 5, 6, 8, k/2}
10
3. Experimentation
Results:
Average obtained taking the best results
possible for every data set is 77.1 %.
s ≥ 3.5
s ≤ 2.0
11
4. Conclusions and Future Work
First Explorative Study of the results of our
classification method varying parameters.
It is posible to improve the result changing the value
of s.
It is proposed s=1.5. It is between the values
proposed by Walley [22].
More studies and experiments are necessary to
ascertain the ideal relationship between the value of
s and some characteristics of the data base.
12
END!
Thanks and Questions?

More Related Content

Similar to Varying parameter in classification based on imprecise probabilities

Interpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional dataInterpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional datatuxette
 
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...Loc Nguyen
 
An Experimental Study of Diabetes Disease Prediction System Using Classificat...
An Experimental Study of Diabetes Disease Prediction System Using Classificat...An Experimental Study of Diabetes Disease Prediction System Using Classificat...
An Experimental Study of Diabetes Disease Prediction System Using Classificat...IOSRjournaljce
 
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...IJRES Journal
 
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural NetworksQuantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural NetworksValentin De Bortoli
 
Subgroup identification for precision medicine. a comparative review of 13 me...
Subgroup identification for precision medicine. a comparative review of 13 me...Subgroup identification for precision medicine. a comparative review of 13 me...
Subgroup identification for precision medicine. a comparative review of 13 me...SuciAidaDahhar
 
LE03.doc
LE03.docLE03.doc
LE03.docbutest
 
Elastic path2path (International Conference on Image Processing'18)
Elastic path2path (International Conference on Image Processing'18)Elastic path2path (International Conference on Image Processing'18)
Elastic path2path (International Conference on Image Processing'18)TamalBatabyal
 
Hyers ulam rassias stability of exponential primitive mapping
Hyers  ulam rassias stability of exponential primitive mappingHyers  ulam rassias stability of exponential primitive mapping
Hyers ulam rassias stability of exponential primitive mappingAlexander Decker
 
An enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataAn enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataeSAT Journals
 
An enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataAn enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataeSAT Publishing House
 
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...Ioannis Partalas
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modulesChristian Robert
 
Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...Loc Nguyen
 
block-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfblock-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfJunghyun Lee
 

Similar to Varying parameter in classification based on imprecise probabilities (20)

Interpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional dataInterpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional data
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
 
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
Learning dyadic data and predicting unaccomplished co-occurrent values by mix...
 
An Experimental Study of Diabetes Disease Prediction System Using Classificat...
An Experimental Study of Diabetes Disease Prediction System Using Classificat...An Experimental Study of Diabetes Disease Prediction System Using Classificat...
An Experimental Study of Diabetes Disease Prediction System Using Classificat...
 
Eu33884888
Eu33884888Eu33884888
Eu33884888
 
Eu33884888
Eu33884888Eu33884888
Eu33884888
 
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
A Mathematical Programming Approach for Selection of Variables in Cluster Ana...
 
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural NetworksQuantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
 
Subgroup identification for precision medicine. a comparative review of 13 me...
Subgroup identification for precision medicine. a comparative review of 13 me...Subgroup identification for precision medicine. a comparative review of 13 me...
Subgroup identification for precision medicine. a comparative review of 13 me...
 
LE03.doc
LE03.docLE03.doc
LE03.doc
 
Elastic path2path (International Conference on Image Processing'18)
Elastic path2path (International Conference on Image Processing'18)Elastic path2path (International Conference on Image Processing'18)
Elastic path2path (International Conference on Image Processing'18)
 
Hyers ulam rassias stability of exponential primitive mapping
Hyers  ulam rassias stability of exponential primitive mappingHyers  ulam rassias stability of exponential primitive mapping
Hyers ulam rassias stability of exponential primitive mapping
 
An enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataAn enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical data
 
An enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical dataAn enhanced fuzzy rough set based clustering algorithm for categorical data
An enhanced fuzzy rough set based clustering algorithm for categorical data
 
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...
Aggressive Sampling for Multi-class to Binary Reduction with Applications to ...
 
better together? statistical learning in models made of modules
better together? statistical learning in models made of modulesbetter together? statistical learning in models made of modules
better together? statistical learning in models made of modules
 
Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...Extreme bound analysis based on correlation coefficient for optimal regressio...
Extreme bound analysis based on correlation coefficient for optimal regressio...
 
block-mdp-masters-defense.pdf
block-mdp-masters-defense.pdfblock-mdp-masters-defense.pdf
block-mdp-masters-defense.pdf
 
Lesson 26
Lesson 26Lesson 26
Lesson 26
 
AI Lesson 26
AI Lesson 26AI Lesson 26
AI Lesson 26
 

More from NTNU

An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...NTNU
 
Bagging Decision Trees on Data Sets with Classification Noise
Bagging Decision Trees on Data Sets with Classification NoiseBagging Decision Trees on Data Sets with Classification Noise
Bagging Decision Trees on Data Sets with Classification NoiseNTNU
 
lassification with decision trees from a nonparametric predictive inference p...
lassification with decision trees from a nonparametric predictive inference p...lassification with decision trees from a nonparametric predictive inference p...
lassification with decision trees from a nonparametric predictive inference p...NTNU
 
Locally Averaged Bayesian Dirichlet Metrics
Locally Averaged Bayesian Dirichlet MetricsLocally Averaged Bayesian Dirichlet Metrics
Locally Averaged Bayesian Dirichlet MetricsNTNU
 
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...NTNU
 
An interactive approach for cleaning noisy observations in Bayesian networks ...
An interactive approach for cleaning noisy observations in Bayesian networks ...An interactive approach for cleaning noisy observations in Bayesian networks ...
An interactive approach for cleaning noisy observations in Bayesian networks ...NTNU
 
Learning classifiers from discretized expression quantitative trait loci
Learning classifiers from discretized expression quantitative trait lociLearning classifiers from discretized expression quantitative trait loci
Learning classifiers from discretized expression quantitative trait lociNTNU
 
Split Criterions for Variable Selection Using Decision Trees
Split Criterions for Variable Selection Using Decision TreesSplit Criterions for Variable Selection Using Decision Trees
Split Criterions for Variable Selection Using Decision TreesNTNU
 
A Semi-naive Bayes Classifier with Grouping of Cases
A Semi-naive Bayes Classifier with Grouping of CasesA Semi-naive Bayes Classifier with Grouping of Cases
A Semi-naive Bayes Classifier with Grouping of CasesNTNU
 
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...NTNU
 
Interactive Learning of Bayesian Networks
Interactive Learning of Bayesian NetworksInteractive Learning of Bayesian Networks
Interactive Learning of Bayesian NetworksNTNU
 
A Bayesian approach to estimate probabilities in classification trees
A Bayesian approach to estimate probabilities in classification treesA Bayesian approach to estimate probabilities in classification trees
A Bayesian approach to estimate probabilities in classification treesNTNU
 
A Bayesian Random Split to Build Ensembles of Classification Trees
A Bayesian Random Split to Build Ensembles of Classification TreesA Bayesian Random Split to Build Ensembles of Classification Trees
A Bayesian Random Split to Build Ensembles of Classification TreesNTNU
 
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...NTNU
 
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...NTNU
 
Evaluating query-independent object features for relevancy prediction
Evaluating query-independent object features for relevancy predictionEvaluating query-independent object features for relevancy prediction
Evaluating query-independent object features for relevancy predictionNTNU
 
Effects of Highly Agreed Documents in Relevancy Prediction
Effects of Highly Agreed Documents in Relevancy PredictionEffects of Highly Agreed Documents in Relevancy Prediction
Effects of Highly Agreed Documents in Relevancy PredictionNTNU
 
Conference poster 6
Conference poster 6Conference poster 6
Conference poster 6NTNU
 

More from NTNU (18)

An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
An Importance Sampling Approach to Integrate Expert Knowledge When Learning B...
 
Bagging Decision Trees on Data Sets with Classification Noise
Bagging Decision Trees on Data Sets with Classification NoiseBagging Decision Trees on Data Sets with Classification Noise
Bagging Decision Trees on Data Sets with Classification Noise
 
lassification with decision trees from a nonparametric predictive inference p...
lassification with decision trees from a nonparametric predictive inference p...lassification with decision trees from a nonparametric predictive inference p...
lassification with decision trees from a nonparametric predictive inference p...
 
Locally Averaged Bayesian Dirichlet Metrics
Locally Averaged Bayesian Dirichlet MetricsLocally Averaged Bayesian Dirichlet Metrics
Locally Averaged Bayesian Dirichlet Metrics
 
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
Application of a Selective Gaussian Naïve Bayes Model for Diffuse-Large B-Cel...
 
An interactive approach for cleaning noisy observations in Bayesian networks ...
An interactive approach for cleaning noisy observations in Bayesian networks ...An interactive approach for cleaning noisy observations in Bayesian networks ...
An interactive approach for cleaning noisy observations in Bayesian networks ...
 
Learning classifiers from discretized expression quantitative trait loci
Learning classifiers from discretized expression quantitative trait lociLearning classifiers from discretized expression quantitative trait loci
Learning classifiers from discretized expression quantitative trait loci
 
Split Criterions for Variable Selection Using Decision Trees
Split Criterions for Variable Selection Using Decision TreesSplit Criterions for Variable Selection Using Decision Trees
Split Criterions for Variable Selection Using Decision Trees
 
A Semi-naive Bayes Classifier with Grouping of Cases
A Semi-naive Bayes Classifier with Grouping of CasesA Semi-naive Bayes Classifier with Grouping of Cases
A Semi-naive Bayes Classifier with Grouping of Cases
 
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
Combining Decision Trees Based on Imprecise Probabilities and Uncertainty Mea...
 
Interactive Learning of Bayesian Networks
Interactive Learning of Bayesian NetworksInteractive Learning of Bayesian Networks
Interactive Learning of Bayesian Networks
 
A Bayesian approach to estimate probabilities in classification trees
A Bayesian approach to estimate probabilities in classification treesA Bayesian approach to estimate probabilities in classification trees
A Bayesian approach to estimate probabilities in classification trees
 
A Bayesian Random Split to Build Ensembles of Classification Trees
A Bayesian Random Split to Build Ensembles of Classification TreesA Bayesian Random Split to Build Ensembles of Classification Trees
A Bayesian Random Split to Build Ensembles of Classification Trees
 
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
An Experimental Study about Simple Decision Trees for Bagging Ensemble on Dat...
 
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
Selective Gaussian Naïve Bayes Model for Diffuse Large-B-Cell Lymphoma Classi...
 
Evaluating query-independent object features for relevancy prediction
Evaluating query-independent object features for relevancy predictionEvaluating query-independent object features for relevancy prediction
Evaluating query-independent object features for relevancy prediction
 
Effects of Highly Agreed Documents in Relevancy Prediction
Effects of Highly Agreed Documents in Relevancy PredictionEffects of Highly Agreed Documents in Relevancy Prediction
Effects of Highly Agreed Documents in Relevancy Prediction
 
Conference poster 6
Conference poster 6Conference poster 6
Conference poster 6
 

Recently uploaded

Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPirithiRaju
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)riyaescorts54
 
Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)DHURKADEVIBASKAR
 
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...lizamodels9
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxpriyankatabhane
 
Twin's paradox experiment is a meassurement of the extra dimensions.pptx
Twin's paradox experiment is a meassurement of the extra dimensions.pptxTwin's paradox experiment is a meassurement of the extra dimensions.pptx
Twin's paradox experiment is a meassurement of the extra dimensions.pptxEran Akiva Sinbar
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxkessiyaTpeter
 
Pests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPirithiRaju
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Patrick Diehl
 
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxSTOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxMurugaveni B
 
Forest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantForest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantadityabhardwaj282
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRlizamodels9
 
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxMicrophone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxpriyankatabhane
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxSwapnil Therkar
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxBerniceCayabyab1
 
TOPIC 8 Temperature and Heat.pdf physics
TOPIC 8 Temperature and Heat.pdf physicsTOPIC 8 Temperature and Heat.pdf physics
TOPIC 8 Temperature and Heat.pdf physicsssuserddc89b
 
Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentationtahreemzahra82
 

Recently uploaded (20)

Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdfPests of soyabean_Binomics_IdentificationDr.UPR.pdf
Pests of soyabean_Binomics_IdentificationDr.UPR.pdf
 
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
(9818099198) Call Girls In Noida Sector 14 (NOIDA ESCORTS)
 
Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)Recombinant DNA technology( Transgenic plant and animal)
Recombinant DNA technology( Transgenic plant and animal)
 
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...
Best Call Girls In Sector 29 Gurgaon❤️8860477959 EscorTs Service In 24/7 Delh...
 
Speech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptxSpeech, hearing, noise, intelligibility.pptx
Speech, hearing, noise, intelligibility.pptx
 
Twin's paradox experiment is a meassurement of the extra dimensions.pptx
Twin's paradox experiment is a meassurement of the extra dimensions.pptxTwin's paradox experiment is a meassurement of the extra dimensions.pptx
Twin's paradox experiment is a meassurement of the extra dimensions.pptx
 
Engler and Prantl system of classification in plant taxonomy
Engler and Prantl system of classification in plant taxonomyEngler and Prantl system of classification in plant taxonomy
Engler and Prantl system of classification in plant taxonomy
 
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptxSOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
SOLUBLE PATTERN RECOGNITION RECEPTORS.pptx
 
Pests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdfPests of castor_Binomics_Identification_Dr.UPR.pdf
Pests of castor_Binomics_Identification_Dr.UPR.pdf
 
Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?Is RISC-V ready for HPC workload? Maybe?
Is RISC-V ready for HPC workload? Maybe?
 
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptxSTOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
STOPPED FLOW METHOD & APPLICATION MURUGAVENI B.pptx
 
Forest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are importantForest laws, Indian forest laws, why they are important
Forest laws, Indian forest laws, why they are important
 
Hot Sexy call girls in Moti Nagar,🔝 9953056974 🔝 escort Service
Hot Sexy call girls in  Moti Nagar,🔝 9953056974 🔝 escort ServiceHot Sexy call girls in  Moti Nagar,🔝 9953056974 🔝 escort Service
Hot Sexy call girls in Moti Nagar,🔝 9953056974 🔝 escort Service
 
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Munirka Delhi 💯Call Us 🔝8264348440🔝
 
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCRCall Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
Call Girls In Nihal Vihar Delhi ❤️8860477959 Looking Escorts In 24/7 Delhi NCR
 
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptxMicrophone- characteristics,carbon microphone, dynamic microphone.pptx
Microphone- characteristics,carbon microphone, dynamic microphone.pptx
 
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptxAnalytical Profile of Coleus Forskohlii | Forskolin .pptx
Analytical Profile of Coleus Forskohlii | Forskolin .pptx
 
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptxGenBio2 - Lesson 1 - Introduction to Genetics.pptx
GenBio2 - Lesson 1 - Introduction to Genetics.pptx
 
TOPIC 8 Temperature and Heat.pdf physics
TOPIC 8 Temperature and Heat.pdf physicsTOPIC 8 Temperature and Heat.pdf physics
TOPIC 8 Temperature and Heat.pdf physics
 
Harmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms PresentationHarmful and Useful Microorganisms Presentation
Harmful and Useful Microorganisms Presentation
 

Varying parameter in classification based on imprecise probabilities

  • 1. 1 Varying parameter in classification based on imprecise probabilities SMPS Bristol, September 2006 J. Abellán, S. Moral, M. Gómez, A. Masegosa Department of Computer Science and AI University of Granada
  • 2. 2 Index 1. Classifiying with Decision Trees. 2. Decision Trees with Imprecise Probabilities. 2.1. Imprecise Dirichlet Model. 2.2. Classification Method. 3. Experimentation. 4. Conclusions and Future Work.
  • 3. 3 1. Classifying with Decision Trees In a classification problem, there is a data set D with values of a set of attribute variables X (petal width, petal length, sepal width, sepal length) and a class variable C (Iris Flower Type:Setosa, Versicolor, Virginica). A decision tree is a tree model with an attribute variable in each node and a value of the class variable in the leafs. Petal-W small large Setal-W small large Petal-L small large Versicolor Setosa ViginicaVersicolor
  • 4. 4 1. Classifying with Decision Trees In a Decision Tree, the path from the root node to another node defines a configuration. A configuration σ is a m-tuple with a set of fixed values for several attribute variables (Petal-Width = small, Sepal-Width = small). ID3 of Quinlan is based in the Shannon entropy. It is measured the reduction in the uncertainty that provokes the introduction of one variable in the model: The entropy measures the uncertainty about C. )|()()|( XCHCHXCInfoGain −= ∑∑ ⋅−==⋅=−= j jj j c cc c jj N n N n cCPcCPCH )(log)|(log)|()|(:Enropy 22 σ σ σ σ σσσ
  • 5. 5 1. Classifying with Decision Trees Example: Iris Subtype Classifcation. IG(Class|Petal-W) =0.26 IG(Class|Petal-L) =0.15 IG(Class|Sepal-L) =0.12 IG(Class|Sepal-W)=0.11 Petal-W small large Setal-W small large Petal-L small large Versicolor Setosa ViginicaVersicolor IG(Class|Petal-W=large, Petal-L) =0.13 IG(Class|Petal-W=large, Setal-W) =0.08 IG(Class|Petal-W=large, Setal-L) =-0.05 IG(Class|Petal-W=small, Setal-W) =0.11 IG(Class|Petal-W=small, Petal-L) =0.04 IG(Class|Petal-W=small, Setal-L) =-0.02 IG(Class|Petal-W=small, Setal-W=small, Petal-L) = - 0.08 IG(Class|Petal-W=small, Setal-W=small, Petal-W) = - 0.12 ?
  • 6. 6 2. Decision Trees with Imprecise Probabilities Recently, Abellán,Moral [7] have introduced a new algorithm to build decision trees based on Imprecise Dirichlet Model of Walley [22]. The authors use of the maximum entropy on credal sets as a good measure of total uncertainty [7,16]. In this way, the structure of the decision tree is exactly the same, the difference is in the criterium to select variables in the ramification. Using this new entropy criterium, the performance is better than ID3.
  • 7. 7 2.1. Imprecise Dirichlet Model (IDM) [22] In this model, the probabilities are estimated as an interval. Given a data set D and a configuration σ. We shall consider a credal set for the class variable C respect to σ defined by the set of probabilities distributions, p = (p1,p2,…,pk), such that: These intervals are obtained in basis of IDM and they depends of the real value s > 0 that is a hyperparemeter of the model. For the IDM, Walley[22] suggests a value for s between 1 and 2. “Classification performance effects of varying the paremeter s”. σ CΡ [ ])|(),|(,)|( σσσ σ σ σ σ jj cc jj cPcP sN sn sN n cCPp jj = ⎥ ⎥ ⎦ ⎤ ⎢ ⎢ ⎣ ⎡ + + + ∈==
  • 8. 8 2.2 The Classification Method This method [7] evaluates a single variable and, also, a couple of variable for the ramification in each step. There will be considered the following functions: If Max{Inf1(X)} > Max{Inf2(Y,Z)}, it is selected X. If Max{Inf2(Y,Z)} > Max{Inf1(X)}, it is selected Y if Inf(Y)>Inf(Z) and it is selected Z, otherwise. When the inclusion of any variable does not reduce the uncertainty of the class variable, a leaf is introduced.The most frequence class in the subset of data compatible with its configuration is associated to this leaf. ( )}{ ),(1 i i xX C x i TU N n XInf = Ρ⋅= ∑ Iσ σ σ σ { } ( )ji ji yYxX C yx ji TU N n YXInf == Ρ= ∑ , , , ),,(2 Iσ σ σ σ
  • 9. 9 3. Experimentation The evaluation was carried out over 8 data sets from UCI repository: Discretized with Fayyad and Irani’s procedure [14]. It was considered the following set of value for s: {0.5, 0.1, 1.5, 2, 2.5, 3, 3.5, 4, 5, 6, 8, k/2}
  • 10. 10 3. Experimentation Results: Average obtained taking the best results possible for every data set is 77.1 %. s ≥ 3.5 s ≤ 2.0
  • 11. 11 4. Conclusions and Future Work First Explorative Study of the results of our classification method varying parameters. It is posible to improve the result changing the value of s. It is proposed s=1.5. It is between the values proposed by Walley [22]. More studies and experiments are necessary to ascertain the ideal relationship between the value of s and some characteristics of the data base.