The document discusses methods for communicating classification uncertainty to end-users. It presents the misclassification method and reclassification method for estimating classification errors. The misclassification method bases error rates on true class sizes and is robust to changes in class proportions, while the reclassification method can be biased by such changes. Additional methods are proposed to estimate variance and handle multiclass problems and shifting feature distributions. The goal is to help non-experts intuitively understand uncertainty in automated classification.
15. How to communicate
the uncertainty?
15
Here the Octopus appeared.
(Ā½Ī¦ )-(Ļāā )
How precise is this?
May 2018
16. Communication Problems
16
Why should we communicate the uncertainty?
Make informed decisions when
choosing and tuning classifiers
Estimate noise and biases in
classification results
17. Communication Problems
17
Why should we communicate the uncertainty?
Make informed decisions when
choosing and tuning classifiers
Estimate noise and biases in
classification results
20. Issues with Classifier Evaluation
20
Itās very tedious
Iām not confident
in my decisions
21. Issues with Classifier Evaluation
21
Itās very tedious
Iām not confident
in my decisions
The terminology
confuses me
22. Issues with Classifier Evaluation
22
Itās very tedious
Iām not confident
in my decisions
The terminology
confuses me
I often confuse
FP and FN
23. Issues with Classifier Evaluation
23
Itās very tedious
Iām not confident
in my decisions
The terminology
confuses me
I donāt understand
the impact on
end-results
I often confuse
FP and FN
39. Communication Problems
39
Why should we communicate the uncertainty?
Make informed decisions when
choosing and tuning classifiers
Estimate noise and biases in
classification results
40. Communication Problems
40
Why should we communicate the uncertainty?
Make informed decisions when
choosing and tuning classifiers
Estimate noise and biases in
classification results
41. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
42. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
Class A
Class A increases a lot
43. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
Class A
Class B
Class A increases a lot
Minority Class B increases too
44. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
Class A
Class B
Class A increases a lot
Minority Class B increases too
Class A items misclassified as
Class B increase too
45. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
Class A
Class B
Class A increases a lot
Minority Class B increases too
Class A items misclassified as
Class B increase too
Does Class B increase only
because of errors from Class A?
46. Issues with Estimating Classification Errors
Time
Count of Items per Class over time
NumberofItems
Class A
Class B
Class A increases a lot
Does Class B increase only
because of errors from Class A?
Minority Class B increases too
Class A items misclassified as
Class B increase too
Within the items classified as Class B,
how many truly belong to Class A?
51. Reclassification Method
Number of items
truly belonging to Class X
and classified as Class Y
Error rates based on
output class size
(e.g., Precision)
52. Reclassification Method
Number of items
truly belonging to Class X
and classified as Class Y
Error rates based on
output class size
(e.g., Precision)
Total mumber of items
classified as Class Y
(output class size)
61. Results with UCI datasets
Class size estimates for 100 random splits in training, test and target sets
(NaĆÆve bayes classifier with 10-fold cross validation)
70. Variance Estimation
Methods exist to estimate the variance
of the Reclassification & Misclassification methods
They are applicable for test sets randomly sampled
within the target sets
Target Set
Test Set
71. Variance Estimation
Methods exist to estimate the variance
of the Reclassification & Misclassification methods
They are applicable for test sets randomly sampled
within the target sets
They are not applicable for disjoint test and target sets
Target Set
Test Set
Target Set
Test Set
73. Sample-to-Sample Method
The Sample-to-Sample method addresses disjoint test and target set
randomly sampled within the same population
Target Set Test Set
Population
with actual Class X
79. Sample-to-Sample Method
Distribution of target setās error rates
estimated from test setās error rates
Normal distribution
explainable with the
Central Limit Theorem
83. Sample-to-Sample Method
Distribution of target setās error rates
estimated from test setās error rates
We use the class size estimates
from the Misclassification method
Variance w.r.t. target set
88. Application of Sample-to-Sample Method
Letās start with binary problems,
we can expressing their solutions in a simpler form
89. Application of Sample-to-Sample Method
Letās start with binary problems,
we can expressing their solutions in a simpler form
These are ratios of random variablesā¦
(Cauchy distribution)
90. Application of Sample-to-Sample Method
Letās start with binary problems,
we can expressing their solutions in a simpler form
These are ratios of random variablesā¦
(Cauchy distribution)
...but they are correlated
(Fiellerās theorem)
91. Application of Sample-to-Sample Method
Fiellerās Theorem estimates confidence intervalsā limits
for ratios of correlated random variables
92. Application of Sample-to-Sample Method
Evaluation of Sample-to-Sample applied with Fiellerās theorem
using estimated to derive
93. Application of Sample-to-Sample Method
Evaluation of Sample-to-Sample applied with Fiellerās theorem
using estimated to derive
We achieve
accurate confidence intervals
for class size estimates
94. Application of Sample-to-Sample Method
Evaluation of Sample-to-Sample applied with Fiellerās theorem
using estimated to derive
ā¦but intervals can be
very large for small class sizes
We achieve
accurate confidence intervals
for class size estimates
95. Application of Sample-to-Sample Method
Evaluation of Sample-to-Sample applied with Fiellerās theorem
using estimated to derive
We achieve
accurate confidence intervals
for class size estimates
ā¦but intervals can be
very large for small class sizes
ā¦and inaccurate for very small
class sizes or error rates
96. Application of Sample-to-Sample Method
Multiclass problems are difficult to express as a ratio
compatible with Fiellerās theorem
ā¦but bootstrapping and simulations can address multiclass problems
100. Future Work
100
Variance estimation for multiclass problems
Guidelines for balancing the sizes of test and target sets
(smaller training sets but larger test sets may improve error estimation)
101. Future Work
101
Variance estimation for multiclass problems
Predict variance magnitude without knowledge of the target sets
(Maximum Determinant method)
Guidelines for balancing the sizes of test and target sets
(smaller training sets but larger test sets may improve error estimation)
102. Future Work
102
Variance estimation for multiclass problems
Handle shifts of error rates and feature distributions
(domain adaptation, e.g., with Bayesian classifiers)
Predict variance magnitude without knowledge of the target sets
(Maximum Determinant method)
Guidelines for balancing the sizes of test and target sets
(smaller training sets but larger test sets may improve error estimation)
103. Future Work
103
Variance estimation for multiclass problems
Fully-specified guidelines for choosing between
Reclassification or Misclassification methods. Or none.
(depending on number of classes, class sizes in test and target sets,
error rate magnitude, shifts of features distribution)
Handle shifts of error rates and feature distributions
(domain adaptation, e.g., with Bayesian classifiers)
Predict variance magnitude without knowledge of the target sets
(Maximum Determinant method)
Guidelines for balancing the sizes of test and target sets
(smaller training sets but larger test sets may improve error estimation)
105. Future Work
105
Visualizations of variance estimates
e.g., for potential target sets
Uncertainty propagation in pipelines of classifiers
e.g., with different test sets
106. Future Work
106
Visualizations of variance estimates
e.g., for potential target sets
Uncertainty propagation in pipelines of classifiers
e.g., with different test sets
Identify individual misclassifications
110. October 2017 IEEE Conference on Data Science and Advanced Analytics (DSAA)
Varying Feature Distributions
111. Varying Feature Distributions
If the feature distributions vary between test and target set,
classifiers may behave differently
112. Varying Feature Distributions
The error rates may systematically differ between test and target sets,
and the Misclassification and Reclassification Methods
can greatly worsen the classification biases
If the feature distributions vary between test and target set,
classifiers may behave differently
115. Varying Feature Distributions
Regressions can be fit to infer error rates from feature values,
but this approach is more complex with the Misclassification Method
Future work is required to handle varying feature distributions
116. Varying Feature Distributions
Regressions can be fit to infer error rates from feature values,
but this approach is more complex with the Misclassification Method
Future work is required to handle varying feature distributions
ā¦but the Misclassification Method can be used to
refine priors in Bayesian classifiers
(i.e., the unconditional class probabilities)
124. Maximum Determinant Method
When starting an application, several classifiers may be available
with no knowledge of the potential target sets
125. Maximum Determinant Method
To choose a classifier, the Maximum Determinant Method aims at
predicting which classifier yields the smallest variance
when applying the Misclassification Method
When starting an application, several classifiers may be available
with no knowledge of the potential target sets
128. Maximum Determinant Method
Hypothesis: The higher the determinant of the error rate matrix,
the lower the results variance.
Inspired by
Cramerās rule
132. Maximum Determinant Method
What are the parameters of relationship between the determinant
and the variance of misclassification results?
(number of classes, class sizes in test and target sets, error rate magnitude)
Initial results are promising but theory must be established
133. Maximum Determinant Method
Initial results are promising but theory must be established
Binary problems for which the method is irrelevant?
What are the parameters of relationship between the determinant
and the variance of misclassification results?
(number of classes, class sizes in test and target sets, error rate magnitude)
134. Maximum Determinant Method
Initial results are promising but theory must be established
Problems for which Misclassification or Ratio-to-TP error rates
provide better predictors?
Binary problems for which the method is irrelevant?
What are the parameters of relationship between the determinant
and the variance of misclassification results?
(number of classes, class sizes in test and target sets, error rate magnitude)
151. Interactions of Uncertainty Factors
151
Poor images
yield more errors
Ground-Truth
Quality
Classification
Errors
Image Quality
152. Interactions of Uncertainty Factors
152
Typhoons yield poor images? (bias)
What confidence intervals? (noise)
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
153. Interactions of Uncertainty Factors
153
Missing videos?
Sampling
Coverage
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
154. Interactions of Uncertainty Factors
154
Some species often move
in & out the field of view
Sampling
Coverage
Duplicated
Individuals
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
155. Interactions of Uncertainty Factors
155
Fields of view
target specific habitats
Sampling
Coverage
Duplicated
Individuals
Field of View
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
156. Interactions of Uncertainty Factors
156
Fields of view
target specific habitats
and shift overtime
Sampling
Coverage
Duplicated
Individuals
Field of View
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
157. Interactions of Uncertainty Factors
157
Sampling
Coverage
Duplicated
Individuals
Field of View
Ground-Truth
Quality
Classification
Errors
Biases & Noise
in Specific Output
Image Quality
160. Lessons Learned
160
Uncertainty factors arise from the system
and its deployement conditions
Investigations should include
domain experts and technical experts
ā¦and non-experts!
161. Lessons Learned
161
Uncertainty factors arise from the system
and its deployement conditions
Investigations should include
domain experts and technical experts
ā¦and non-experts!
People need to feel comfortable
to engage in criticism
178. 178
Issues Tackled
Some metrics
conceal uncertainty
The metrics omit which species
are confused with another
Using one single type of curve
can hide differences
179. 179
Issues Tackled
Some metrics
conceal uncertainty
Using one single type of curve
can hide differences
ā¦and omit species proportions
The metrics omit which species
are confused with another
Editor's Notes
disclosing CV errors to domain experts
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Put visualization in background
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
The central limit theorem (CLT) is a statistical theory that states that given a sufficiently large sample size from a population with a finite level of variance,
- the mean of all samples from the same population will be approximately equal to the mean of the population
- all of the samples will follow an approximate normal distribution
The central limit theorem (CLT) is a statistical theory that states that given a sufficiently large sample size from a population with a finite level of variance,
- the mean of all samples from the same population will be approximately equal to the mean of the population
- all of the samples will follow an approximate normal distribution
Edit figure!
Put visualization in background
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Edit figure!
Put visualization in background
Put visualization in background
Put visualization in background
disclosing CV errors to domain experts
Put visualization in background
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
disclosing CV errors to domain experts
Put visualization in background
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.
Explain project for images
Then why computer vision
and then text bubble.