SlideShare a Scribd company logo
1 of 47
Testing, Testing, 1…2…3:
The Basics of Diagnostic
Testing
Brian N. Eisen
The Eisen Law Firm, Co., L.P.A.
1300 E. 9th St., Suite 1801
Cleveland, OH 44114
(216) 687-0900
brian@eisenlawfirm.com
Diagnostic Tests are
Important
But are They Worth a Damn?
Our Population
The Perfect Diagnostic Test
Diagnostic Tests are NOT Perfect
Three Key Terms
1. SENSITIVITY
2. SPECIFICITY
3. PREDICTIVE VALUE
SENSITIVITY
Sensitivity is a measure of how good a test
is at correctly identifying those individuals
with the condition or disease.
It is defined as the fraction of the diseased
who test positive.
Sensitivity answers the question, “If we
tested only sick people, what percentage of
them would the test correctly identify”?
Sensitivity: Look ONLY at those
with the Condition
If our test is 80%
Sensitive:
If our test is 100%
Sensitive:
All 20 subjects WITH THE
CONDITION will test
positive!
16 of 20 subjects WITH
THE CONDITION will test
positive.
Real Life Example #1
High Sensitivity
Most Home Pregnancy Tests
Claim to be >95% Sensitive
Real Life Example #2
Low Sensitivity
H1N1 Rapid Test Only 40% Sensitive
So, for Every 10 People
with the Flu…
The Test Only Gets 4 Correct!
# Positive Tests (4) + # Negative Tests (6) = 1
All with flu (10) All with flu (10)
SENSITIVITY + FALSE NEGATIVE RATE = 1
Sensitivity and False Negatives are
Complementary
Here’s Another Way to Look at it:
Those who
test + are
“True
Positives”
Those who
test - are
“False
Negatives”
Sensitivity = TP/(TP +FN) X 100
Sensitivity and False Negatives are
Complementary
• Sensitivity + False Negative Rate = 1
• A test with 40% Sensitivity has a 60% False
Negative Rate
• A test with 75% Sensitivity has _____ %
False Negatives
25
Q: How Can you Develop a Test
that is 100% Sensitive?
A: Make it Always Positive!
Sensitivity: So What?
• Sensitivity + False Negative Rate =1
• The greater the Sensitivity, the Fewer False
Negatives
• So, a Very Sensitive Test will have Very Few
False Negatives
• LESSON: IF YOU GET A NEGATIVE, IT’S
PROBABLY A TRUE NEGATIVE!
• RULE: SnOUT (SENSITIVE, NEGATIVE,
OUT)
SPECIFICITY
Specificity is a measure of how good a test is
at correctly identifying those individuals who
do not have the disease.
It is defined as the fraction of the non-
diseased who test negative.
Specificity answers the question, “If we
tested only those who are well (do not have
the condition or disease), what percentage
would the test correctly identify”?
Specificity: Look ONLY at those
WITHOUT the Condition
If our test is 100%
Specific:
All 80 subjects WITHOUT
THE CONDITION will test
negative!
If our test is 75%
Specific:
60 of 80 subjects WITHOUT
THE CONDITION will test
negative.
Here’s Another Way to Look at it:
Those who
test + are
“False
Positives”
Those who
test - are
“True
Negatives”
Specificity = TN/(TN +FP) X 100
Specificity and False Positives are
Complementary
• Specificity + False Positive Rate = 1
• A test with 40% Specificity has a 60% False
Positive Rate
• A test with 75% Specificity has _____ % False
Positives
25
Specificity: So What?
• Specificity + False Positives =1
• The greater the Specificity, the Fewer False
Positives
• So, a Very Specific Test will have Very Few
False Positives
• LESSON: IF YOU GET A POSITIVE, IT’S
PROBABLY A TRUE POSITIVE!
• RULE: SpIN (SPECIFIC, POSITIVE, IN)
Real Life Example #1
High Specificity
WHAT DOES THIS MEAN?
27% Sensitive
• If you test a hospitalized inpatient with a
UTI, only 27% chance it will be positive.
• 73% of hospitalized patients with UTI will
have a false negative test.
• Not very good to rule out UTI.
94% Specific
• If you test a hospitalized inpatient without a
UTI, there is a 94% chance the test will be
negative.
• Only a few (6%) of patients without UTI
will have a falsely positive test.
• Good test to rule in UTI.
SpIN
Real Life Example #2:
Low Specificity
WHAT DOES THIS MEAN?
100% Sensitive
• If someone has mediastinitis, his CT will be
positive always.
• No one will have a falsely negative CT (no
one with mediastinitis will have a negative
CT).
• If CT is negative, you can rule out
mediastinitis.
SnOUT
33% Specific
• Only 1/3 people without mediastinitis will
have a negative CT.
• 2/3 people will have a positive CT, even
though they don’t have mediastinitis.
• Not a good test to rule in mediastinitis.
In Practice
• When a doctor tells you she ruled out a
condition with a negative result on a
particular test, find out its sensitivity
(SnOUT). If it’s not very high, go after her.
• When a doctor tells you she ruled in a
condition with a positive result on a
particular test, find out its specificity
(SpIN). If it’s not very high, go after her.
Questions Answered
Sensitivity answers the question, “If
we tested only sick people, what
percentage of them would the test
correctly identify”?
Specificity answers the question, “If we
tested only those who are well (do not
have the condition or disease), what
percentage would the test correctly
identify”?
Population  Individual
• In the clinical setting, a more important
question is:
• If the test results are positive (or
negative) in a given patient, what is
the probability that this patient has
(or does not have) the disease?
• In other words:
What proportion of patients who test
positive (or negative) actually have (or do
not have) the disease in question?
PREDICTIVE VALUE
Positive predictive value is the
proportion of all people with positive
tests who have the disease.
Negative predictive value is the
proportion of all people with negative
tests who do not have the disease.
PPV: Chance a Positive Result is
Correct
In this example, 20 out of 100 positive
results are correct. PPV = 20%
NPV: Chance Negative Result is
Correct
In this example, 80 out of 100 negative
results are correct. NPV = 80%
CRITICAL THING TO NOTE ABOUT
PREDICTIVE VALUES:
Unlike Sensitivity and Specificity,
PREDICTIVE VALUES DEPEND
UPON THE PREVALENCE OF A
DISEASE IN THE POPULATION.
THE HIGHER THE PREVALENCE,
THE HIGHER THE POSITIVE PREDICTIVE VALUE
When almost everyone has it, the chance
of a positive test being right is high!
THE LOWER THE PREVALENCE,
THE HIGHER THE NEGATIVE PREDICTIVE VALUE
When almost no one has it, the chance of
a negative test being right is high!
In Practice
• A doctor tells you he ordered an invasive
treatment because the patient tested positive
in a test known to have a “high PPV.”
• Q: Is this a sound decision?
• A: It depends.
– Is the patient from the population where the
PPV was determined?
If the patient comes from a population with a much
lower prevalence, PPV would be much lower, and it
might be risky to attempt the treatment!
If the Published PPV comes
from this population:
You don’t want to apply it to a
member of this population:
CALCULATING THE RATES
A test is used in 50 people with disease
and 50 people without. These are the
results:
Disease
+ -
Test
+ 48 3 51
- 2 47 49
50 50 100
Sensitivity = 48/50 = 96%
Specificity = 47/50 = 94%
Positive predictive value = 48/51 = 94%
Negative predictive value = 47/49 = 96%
Disease
+ -
Test
+ 48 3 51
- 2 47 49
50 50 100
Now, let’s take this test out into a
population where 2% of people have the
disease, not 50% as in the previous
example. Assume there are 10,000 people,
and the same sensitivity and specificity as
before, namely 96% and 94%, respectively
Disease
+ -
Test
+ 192 588 780
- 8 9,212 9,220
200 9,800 10,000
• What is the positive predictive value
now?
192/780 = 24.6%
• When the prevalence of disease is 50%,
94% of positive tests indicate disease. But
when prevalence is only 2% -- less than
one in four -- positive test results indicate
a person with disease, and 2% actually
would represents a very common disease.
Brian’s Cheat Sheet

More Related Content

Similar to OAJ presentation final draft

8 screening.pptxscreening.pptxscreening.
8 screening.pptxscreening.pptxscreening.8 screening.pptxscreening.pptxscreening.
8 screening.pptxscreening.pptxscreening.Wasihun Aragie
 
Evidence-based diagnosis
Evidence-based diagnosisEvidence-based diagnosis
Evidence-based diagnosisHesham Gaber
 
Disease screening and screening test validity
Disease screening and screening test validityDisease screening and screening test validity
Disease screening and screening test validityTampiwaChebani
 
Screening and diagnostic testing
Screening and diagnostic  testingScreening and diagnostic  testing
Screening and diagnostic testingamitakashyap1
 
2016 veterinary diagnostics
2016 veterinary diagnostics2016 veterinary diagnostics
2016 veterinary diagnosticsPerez Eric
 
Probability.pdf.pdf and Statistics for R
Probability.pdf.pdf and Statistics for RProbability.pdf.pdf and Statistics for R
Probability.pdf.pdf and Statistics for RSakhileKhoza2
 
CAT 1 -MPH 5101 - FOUNDATIONS OF EPIDEMIOLOGY (1).pptx
CAT 1 -MPH 5101 -  FOUNDATIONS OF EPIDEMIOLOGY  (1).pptxCAT 1 -MPH 5101 -  FOUNDATIONS OF EPIDEMIOLOGY  (1).pptx
CAT 1 -MPH 5101 - FOUNDATIONS OF EPIDEMIOLOGY (1).pptxShafici Almis
 
screening-140217071714-phpapp02.pdf
screening-140217071714-phpapp02.pdfscreening-140217071714-phpapp02.pdf
screening-140217071714-phpapp02.pdfSYEDZIYADFURQAN
 
Validity and Reliability
Validity and Reliability Validity and Reliability
Validity and Reliability Tauseef Jawaid
 
Evidence based diagnosis
Evidence based diagnosisEvidence based diagnosis
Evidence based diagnosisHesham Al-Inany
 
Evaluating diagnostic tests.pptx
Evaluating diagnostic tests.pptxEvaluating diagnostic tests.pptx
Evaluating diagnostic tests.pptxangelabraver1
 
7 Lecture 8 Outbrek invesitgation.pptx
7 Lecture 8 Outbrek invesitgation.pptx7 Lecture 8 Outbrek invesitgation.pptx
7 Lecture 8 Outbrek invesitgation.pptxmergawekwaya
 
Screening test (basic concepts)
Screening test (basic concepts)Screening test (basic concepts)
Screening test (basic concepts)Tarek Tawfik Amin
 
Sdc smacc wrestling with risk
Sdc smacc wrestling with riskSdc smacc wrestling with risk
Sdc smacc wrestling with riskSimon Carley
 
Dr Amit Diagnostic Tests.pptx
Dr Amit Diagnostic Tests.pptxDr Amit Diagnostic Tests.pptx
Dr Amit Diagnostic Tests.pptxAmitBhondve3
 

Similar to OAJ presentation final draft (20)

Evidence Based Diagnosis
Evidence Based DiagnosisEvidence Based Diagnosis
Evidence Based Diagnosis
 
8 screening.pptxscreening.pptxscreening.
8 screening.pptxscreening.pptxscreening.8 screening.pptxscreening.pptxscreening.
8 screening.pptxscreening.pptxscreening.
 
Validity and Screening Test
Validity and Screening TestValidity and Screening Test
Validity and Screening Test
 
Evidence-based diagnosis
Evidence-based diagnosisEvidence-based diagnosis
Evidence-based diagnosis
 
Disease screening and screening test validity
Disease screening and screening test validityDisease screening and screening test validity
Disease screening and screening test validity
 
Screening and diagnostic testing
Screening and diagnostic  testingScreening and diagnostic  testing
Screening and diagnostic testing
 
2016 veterinary diagnostics
2016 veterinary diagnostics2016 veterinary diagnostics
2016 veterinary diagnostics
 
Probability.pdf.pdf and Statistics for R
Probability.pdf.pdf and Statistics for RProbability.pdf.pdf and Statistics for R
Probability.pdf.pdf and Statistics for R
 
CAT 1 -MPH 5101 - FOUNDATIONS OF EPIDEMIOLOGY (1).pptx
CAT 1 -MPH 5101 -  FOUNDATIONS OF EPIDEMIOLOGY  (1).pptxCAT 1 -MPH 5101 -  FOUNDATIONS OF EPIDEMIOLOGY  (1).pptx
CAT 1 -MPH 5101 - FOUNDATIONS OF EPIDEMIOLOGY (1).pptx
 
screening-140217071714-phpapp02.pdf
screening-140217071714-phpapp02.pdfscreening-140217071714-phpapp02.pdf
screening-140217071714-phpapp02.pdf
 
Validity and Reliability
Validity and Reliability Validity and Reliability
Validity and Reliability
 
Evidence based diagnosis
Evidence based diagnosisEvidence based diagnosis
Evidence based diagnosis
 
Evaluating diagnostic tests.pptx
Evaluating diagnostic tests.pptxEvaluating diagnostic tests.pptx
Evaluating diagnostic tests.pptx
 
7 Lecture 8 Outbrek invesitgation.pptx
7 Lecture 8 Outbrek invesitgation.pptx7 Lecture 8 Outbrek invesitgation.pptx
7 Lecture 8 Outbrek invesitgation.pptx
 
LHR.pptx
LHR.pptxLHR.pptx
LHR.pptx
 
Diagnostic test and agreement.pptx
Diagnostic test and agreement.pptxDiagnostic test and agreement.pptx
Diagnostic test and agreement.pptx
 
Screening test (basic concepts)
Screening test (basic concepts)Screening test (basic concepts)
Screening test (basic concepts)
 
Sdc smacc wrestling with risk
Sdc smacc wrestling with riskSdc smacc wrestling with risk
Sdc smacc wrestling with risk
 
Dr Amit Diagnostic Tests.pptx
Dr Amit Diagnostic Tests.pptxDr Amit Diagnostic Tests.pptx
Dr Amit Diagnostic Tests.pptx
 
Clinical epidemiology
Clinical epidemiologyClinical epidemiology
Clinical epidemiology
 

OAJ presentation final draft

  • 1. Testing, Testing, 1…2…3: The Basics of Diagnostic Testing Brian N. Eisen The Eisen Law Firm, Co., L.P.A. 1300 E. 9th St., Suite 1801 Cleveland, OH 44114 (216) 687-0900 brian@eisenlawfirm.com
  • 2. Diagnostic Tests are Important But are They Worth a Damn?
  • 5. Diagnostic Tests are NOT Perfect
  • 6. Three Key Terms 1. SENSITIVITY 2. SPECIFICITY 3. PREDICTIVE VALUE
  • 7. SENSITIVITY Sensitivity is a measure of how good a test is at correctly identifying those individuals with the condition or disease. It is defined as the fraction of the diseased who test positive. Sensitivity answers the question, “If we tested only sick people, what percentage of them would the test correctly identify”?
  • 8. Sensitivity: Look ONLY at those with the Condition
  • 9. If our test is 80% Sensitive: If our test is 100% Sensitive: All 20 subjects WITH THE CONDITION will test positive! 16 of 20 subjects WITH THE CONDITION will test positive.
  • 10. Real Life Example #1 High Sensitivity Most Home Pregnancy Tests Claim to be >95% Sensitive
  • 11. Real Life Example #2 Low Sensitivity H1N1 Rapid Test Only 40% Sensitive
  • 12. So, for Every 10 People with the Flu…
  • 13. The Test Only Gets 4 Correct!
  • 14. # Positive Tests (4) + # Negative Tests (6) = 1 All with flu (10) All with flu (10) SENSITIVITY + FALSE NEGATIVE RATE = 1 Sensitivity and False Negatives are Complementary
  • 15. Here’s Another Way to Look at it: Those who test + are “True Positives” Those who test - are “False Negatives” Sensitivity = TP/(TP +FN) X 100
  • 16. Sensitivity and False Negatives are Complementary • Sensitivity + False Negative Rate = 1 • A test with 40% Sensitivity has a 60% False Negative Rate • A test with 75% Sensitivity has _____ % False Negatives 25
  • 17. Q: How Can you Develop a Test that is 100% Sensitive? A: Make it Always Positive!
  • 18. Sensitivity: So What? • Sensitivity + False Negative Rate =1 • The greater the Sensitivity, the Fewer False Negatives • So, a Very Sensitive Test will have Very Few False Negatives • LESSON: IF YOU GET A NEGATIVE, IT’S PROBABLY A TRUE NEGATIVE! • RULE: SnOUT (SENSITIVE, NEGATIVE, OUT)
  • 19. SPECIFICITY Specificity is a measure of how good a test is at correctly identifying those individuals who do not have the disease. It is defined as the fraction of the non- diseased who test negative. Specificity answers the question, “If we tested only those who are well (do not have the condition or disease), what percentage would the test correctly identify”?
  • 20. Specificity: Look ONLY at those WITHOUT the Condition
  • 21. If our test is 100% Specific: All 80 subjects WITHOUT THE CONDITION will test negative! If our test is 75% Specific: 60 of 80 subjects WITHOUT THE CONDITION will test negative.
  • 22. Here’s Another Way to Look at it: Those who test + are “False Positives” Those who test - are “True Negatives” Specificity = TN/(TN +FP) X 100
  • 23. Specificity and False Positives are Complementary • Specificity + False Positive Rate = 1 • A test with 40% Specificity has a 60% False Positive Rate • A test with 75% Specificity has _____ % False Positives 25
  • 24. Specificity: So What? • Specificity + False Positives =1 • The greater the Specificity, the Fewer False Positives • So, a Very Specific Test will have Very Few False Positives • LESSON: IF YOU GET A POSITIVE, IT’S PROBABLY A TRUE POSITIVE! • RULE: SpIN (SPECIFIC, POSITIVE, IN)
  • 25. Real Life Example #1 High Specificity WHAT DOES THIS MEAN?
  • 26. 27% Sensitive • If you test a hospitalized inpatient with a UTI, only 27% chance it will be positive. • 73% of hospitalized patients with UTI will have a false negative test. • Not very good to rule out UTI.
  • 27. 94% Specific • If you test a hospitalized inpatient without a UTI, there is a 94% chance the test will be negative. • Only a few (6%) of patients without UTI will have a falsely positive test. • Good test to rule in UTI. SpIN
  • 28. Real Life Example #2: Low Specificity WHAT DOES THIS MEAN?
  • 29. 100% Sensitive • If someone has mediastinitis, his CT will be positive always. • No one will have a falsely negative CT (no one with mediastinitis will have a negative CT). • If CT is negative, you can rule out mediastinitis. SnOUT
  • 30. 33% Specific • Only 1/3 people without mediastinitis will have a negative CT. • 2/3 people will have a positive CT, even though they don’t have mediastinitis. • Not a good test to rule in mediastinitis.
  • 31. In Practice • When a doctor tells you she ruled out a condition with a negative result on a particular test, find out its sensitivity (SnOUT). If it’s not very high, go after her. • When a doctor tells you she ruled in a condition with a positive result on a particular test, find out its specificity (SpIN). If it’s not very high, go after her.
  • 32. Questions Answered Sensitivity answers the question, “If we tested only sick people, what percentage of them would the test correctly identify”? Specificity answers the question, “If we tested only those who are well (do not have the condition or disease), what percentage would the test correctly identify”?
  • 33. Population  Individual • In the clinical setting, a more important question is: • If the test results are positive (or negative) in a given patient, what is the probability that this patient has (or does not have) the disease? • In other words: What proportion of patients who test positive (or negative) actually have (or do not have) the disease in question?
  • 34. PREDICTIVE VALUE Positive predictive value is the proportion of all people with positive tests who have the disease. Negative predictive value is the proportion of all people with negative tests who do not have the disease.
  • 35. PPV: Chance a Positive Result is Correct In this example, 20 out of 100 positive results are correct. PPV = 20%
  • 36. NPV: Chance Negative Result is Correct In this example, 80 out of 100 negative results are correct. NPV = 80%
  • 37. CRITICAL THING TO NOTE ABOUT PREDICTIVE VALUES: Unlike Sensitivity and Specificity, PREDICTIVE VALUES DEPEND UPON THE PREVALENCE OF A DISEASE IN THE POPULATION.
  • 38. THE HIGHER THE PREVALENCE, THE HIGHER THE POSITIVE PREDICTIVE VALUE When almost everyone has it, the chance of a positive test being right is high!
  • 39. THE LOWER THE PREVALENCE, THE HIGHER THE NEGATIVE PREDICTIVE VALUE When almost no one has it, the chance of a negative test being right is high!
  • 40. In Practice • A doctor tells you he ordered an invasive treatment because the patient tested positive in a test known to have a “high PPV.” • Q: Is this a sound decision? • A: It depends. – Is the patient from the population where the PPV was determined? If the patient comes from a population with a much lower prevalence, PPV would be much lower, and it might be risky to attempt the treatment!
  • 41. If the Published PPV comes from this population:
  • 42. You don’t want to apply it to a member of this population:
  • 43. CALCULATING THE RATES A test is used in 50 people with disease and 50 people without. These are the results: Disease + - Test + 48 3 51 - 2 47 49 50 50 100
  • 44. Sensitivity = 48/50 = 96% Specificity = 47/50 = 94% Positive predictive value = 48/51 = 94% Negative predictive value = 47/49 = 96% Disease + - Test + 48 3 51 - 2 47 49 50 50 100
  • 45. Now, let’s take this test out into a population where 2% of people have the disease, not 50% as in the previous example. Assume there are 10,000 people, and the same sensitivity and specificity as before, namely 96% and 94%, respectively Disease + - Test + 192 588 780 - 8 9,212 9,220 200 9,800 10,000
  • 46. • What is the positive predictive value now? 192/780 = 24.6% • When the prevalence of disease is 50%, 94% of positive tests indicate disease. But when prevalence is only 2% -- less than one in four -- positive test results indicate a person with disease, and 2% actually would represents a very common disease.