SlideShare a Scribd company logo
1 of 27
Item Analysis
Purpose of Item Analysis
– Evaluates the quality of each item

– Rationale: the quality of items determines the
  quality of test (i.e., reliability & validity)

– May suggest ways of improving the
 measurement of a test

– Can help with understanding why certain
  tests predict some criteria but not others
Item Analysis
 When  analyzing the test items, we have several
 questions about the performance of each item. Some
 of these questions include:
     Are the items congruent with the test objectives?
     Are the items valid? Do they measure what they're
     supposed to measure?
     Are the items reliable? Do they measure consistently?
     How long does it take an examinee to complete each
     item?
     What items are most difficult to answer correctly?
     What items are easy?
     Are there any poor performing items that need to be
     discarded?
Types of Item Analyses for CTT

Three major types:
   1. Assess quality of the distractors

   2. Assess difficulty of the items
   3. Assess how well an item
   differentiates between high and low
   performers
DISTRACTOR ANALYSIS

 A. Multiple-Choke
 B. Multiply-Choice
 C. Multiple-Choice
 D. Multi-Choice
Distractor Analysis
First question of item analysis: How many
people choose each response?
If there is only one best response, then all
     other response options are distractors.
Example from in-class assignment (N = 35):

Which method has the best internal consistency?   #
a) projective test                                1
b) peer ratings                                   1
c) forced choice                                  21
d) differences n.s.                               12
Distractor Analysis (cont’d)
A perfect test item would have 2 characteristics:
      1. Everyone who knows the item gets it right
      2. People who do not know the item will have
responses equally distributed across the wrong answers.

It is not desirable to have one of the distractors chosen
more often than the correct answer.

 This result indicates a potential problem with the
question. This distractor may be too similar to the correct
answer and/or there may be something in either the stem
or the alternatives that is misleading.
Distractor Analysis (cont’d)
 Calculate the # of people expected to choose each of the
 distractors. If random same expected number for each
 wrong response (Figure 10-1).




# of Persons     N answering incorrectly       14
Exp. To Choose                             =        = 4.7
Distractor       Number of distractors         3
Distractor Analysis (cont’d)
When the number of persons choosing a distractor
significantly exceeds the number expected, there are 2
possibilities:

1. It is possible that the choice reflects partial knowledge

2. The item is a poorly worded trick question

 unpopular distractors may lower item and test difficulty
because it is easily eliminated

 extremely popular is likely to lower the reliability and
validity of the test
Item Difficulty Analysis
 Description   and How to Compute
        ex: a) (6 X 3) + 4 = ?
            b) 9 [1n(-3.68) X (1 – 1n(+3.68))] = ?

 Itis often difficult to explain or define difficulty in
  terms of some intrinsic characteristic of the item

 The  only common thread of difficult items is that
  individuals did not know the answer
Item Difficulty




 Percentage of test takers who respond correctly


What if p = .00
What if p = 1.00?
Item Difficulty
– An item with a p value of .0 or 1.0 does not
  contribute to measuring individual differences and
  thus is certain to be useless
– When comparing 2 test scores, we are interested in
  who had the higher score or the differences in
  scores
– p value of .5 have most variation so seek items in
  this range and remove those with extreme values
– can also be examined to determine proportion
  answering in a particular way for items that don’t
  have a “correct” answer
Item Difficulty (cont.)

What is the best p-value?
– most optimal p-value = .50
– maximum discrimination between good
  and poor performers


Should we only choose items of .50?
When shouldn’t we?
Should we only choose items of .50?

Not necessarily ...

 When  wanting to screen the very top group of
  applicants (i.e., admission to university or medical
  school).

Cutoffs may be much higher

 Other institutions want a minimum level (i.e., minimum
  reading level)

Cutoffs may be much lower
Item Difficulty (cont.)
Interpreting the p-value...

     example:
       100 people take a test
       15 got question 1 right


What is the p-value?
Is this an easy or hard item?
Item Difficulty (cont.)
Interpreting the p-value...

     example:
       100 people take a test
       70 got question 1 right


What is the p-value?
Is this an easy or hard item?
Item Difficulty (cont’d)
General Rules of Item Difficulty…

p low (< .20)             difficult test item
p moderate (.20 - .80)    moderately diff.
p high (> .80)            easy item
ITEM DISCRIMINATION
... The extent to which an item
differentiates people on the
behavior that the test is designed
to assess.


the computed difference between
the percentage of high achievers
and the percentage of
low achievers who got the item
right.
Item Discrimination (cont.)

 compares   the performance of upper
 group (with high test scores) and lower
 group (low test scores) on each item--%
 of test takers in each group who were
 correct
Item Discrimination (cont’d):
      Discrimination Index (D)
   Divide sample into TOP half and
    BOTTOM half (or TOP and BOTTOM
    third)
   Compute Discrimination Index (D)
Item Discrimination
D   =U-L

U = # in the upper group correct response
           Total # in upper group

L = # in the lower group correct response
            Total # in lower group

The higher the value of D, the more adequately
 the item discriminates (The highest value is 1.0)
Item Discrimination
 seekitems with high positive numbers (those
 who do well on the test tend to get the item
 correct)

 negative  numbers (lower scorers on test more
 likely to get item correct) and low positive
 numbers (about the same proportion of low and
 high scorers get the item correct) don’t
 discriminate well and are discarded
Item Discrimination (cont’d):
        Item-Total Correlation
Correlation between each item (a correct response
usually receives a score of 1 and an incorrect a score
of zero) and the total test score.
To which degree do item and test measures the same
thing?

Positive -item discriminates between high and low
scores

Near 0 - item does not discriminate between high & low

Negative - scores on item and scores on test disagree
Item Discrimination (cont’d):
     Item-Total Correlation
Item-total correlations are directly
    related to reliability.
         Why?
Because the more each item correlates
with the test as a whole, the higher all
items correlate with each other
( = higher alpha, internal consistency)
Quantitative Item Analysis
 Inter-item correlation matrix displays the
  correlation of each item with every other
  item
 provides important information for
  increasing the test’s internal consistency
 each item should be highly correlated
  with every other item measuring the same
  construct and not correlated with items
  measuring a different construct
Quantitative Item Analysis
 itemsthat are not highly correlated with
 other items measuring the same
 construct can and should be dropped to
 increase internal consistency
Item Discrimination (cont’d):
        Interitem Correlation
Possible causes for low inter-item correlation:
a. Item badly written (revise)
b. Item measures other attribute than rest of
   the test (discard)
c. Item correlated with some items, but not
   with others: test measures 2 distinct
   attributes (subtests or subscales)

More Related Content

What's hot

What's hot (20)

Item Analysis
Item AnalysisItem Analysis
Item Analysis
 
Principles of Test Construction 1
Principles of Test Construction 1Principles of Test Construction 1
Principles of Test Construction 1
 
Characteristics of a good test
Characteristics  of a good testCharacteristics  of a good test
Characteristics of a good test
 
Item analysis
Item analysisItem analysis
Item analysis
 
Item analysis with spss software
Item analysis with spss softwareItem analysis with spss software
Item analysis with spss software
 
Characteristics of a good test
Characteristics of a good testCharacteristics of a good test
Characteristics of a good test
 
Item analysis
Item analysisItem analysis
Item analysis
 
Item Analysis
Item AnalysisItem Analysis
Item Analysis
 
Qualities of a Good Test
Qualities of a Good TestQualities of a Good Test
Qualities of a Good Test
 
Types of Scores & Types of Standard Scores
Types of Scores & Types of Standard ScoresTypes of Scores & Types of Standard Scores
Types of Scores & Types of Standard Scores
 
Quantitative Item Analysis
Quantitative Item Analysis Quantitative Item Analysis
Quantitative Item Analysis
 
Item analysis ppt
Item analysis pptItem analysis ppt
Item analysis ppt
 
Item Analysis
Item AnalysisItem Analysis
Item Analysis
 
New item analysis
New item analysisNew item analysis
New item analysis
 
Item analysis
Item analysisItem analysis
Item analysis
 
Test construction essay tests
Test construction  essay testsTest construction  essay tests
Test construction essay tests
 
Assessment: Grading & Student Evaluation
Assessment: Grading & Student EvaluationAssessment: Grading & Student Evaluation
Assessment: Grading & Student Evaluation
 
Test item formats: definition, types, pros and cons
Test item formats: definition, types, pros and consTest item formats: definition, types, pros and cons
Test item formats: definition, types, pros and cons
 
Test Administration, Test administration, Test-taking Strategies
Test Administration, Test administration, Test-taking StrategiesTest Administration, Test administration, Test-taking Strategies
Test Administration, Test administration, Test-taking Strategies
 
Writing objective test items
Writing objective test itemsWriting objective test items
Writing objective test items
 

Viewers also liked

Item Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexItem Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexMr. Ronald Quileste, PhD
 
Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validationKEnkenken Tan
 
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...Andrea Varga
 
Item discrimination
Item discriminationItem discrimination
Item discriminationBasil Ahamed
 
K to 12 Grade 3 LAPG FILIPINO Reviewer
K to 12 Grade 3 LAPG FILIPINO Reviewer K to 12 Grade 3 LAPG FILIPINO Reviewer
K to 12 Grade 3 LAPG FILIPINO Reviewer LiGhT ArOhL
 
Types of test questions
Types of test questionsTypes of test questions
Types of test questionsMa Tamonte
 
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)2. paghahanda ng mga kagamitang tanaw dinig (julius m.)
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)Julius Morite
 
Reviewer filipino
Reviewer filipinoReviewer filipino
Reviewer filipinoayamvicn
 
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)Kedamien Riley
 
Kinds of tests and testing
Kinds of tests and testingKinds of tests and testing
Kinds of tests and testingMaury Martinez
 
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer LiGhT ArOhL
 
Types of Tests
Types of TestsTypes of Tests
Types of TestsIvy Autor
 
Distracter Analysis - Index of Effectiveness
Distracter Analysis - Index of EffectivenessDistracter Analysis - Index of Effectiveness
Distracter Analysis - Index of EffectivenessMr. Ronald Quileste, PhD
 
Kagamitang panturo presentation
Kagamitang panturo presentationKagamitang panturo presentation
Kagamitang panturo presentationbhe pestijo
 
KINDS OF TEST
KINDS OF TESTKINDS OF TEST
KINDS OF TESTyamih 88
 
K to 12 Grade 3 FILIPINO NAT (National Achievement Test) Reviewer
K to 12 Grade 3 FILIPINO  NAT (National Achievement Test) ReviewerK to 12 Grade 3 FILIPINO  NAT (National Achievement Test) Reviewer
K to 12 Grade 3 FILIPINO NAT (National Achievement Test) ReviewerLiGhT ArOhL
 
Nat reviewer in filipino vi
Nat reviewer in filipino viNat reviewer in filipino vi
Nat reviewer in filipino viMaricel Conales
 

Viewers also liked (20)

Item Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty IndexItem Analysis - Discrimination and Difficulty Index
Item Analysis - Discrimination and Difficulty Index
 
Item analysis and validation
Item analysis and validationItem analysis and validation
Item analysis and validation
 
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...
Semantic similarity of distractors in multiple- choice tests: extrinsic evalu...
 
Item discrimination
Item discriminationItem discrimination
Item discrimination
 
Instructional Strategies
Instructional StrategiesInstructional Strategies
Instructional Strategies
 
K to 12 Grade 3 LAPG FILIPINO Reviewer
K to 12 Grade 3 LAPG FILIPINO Reviewer K to 12 Grade 3 LAPG FILIPINO Reviewer
K to 12 Grade 3 LAPG FILIPINO Reviewer
 
Types of test questions
Types of test questionsTypes of test questions
Types of test questions
 
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)2. paghahanda ng mga kagamitang tanaw dinig (julius m.)
2. paghahanda ng mga kagamitang tanaw dinig (julius m.)
 
Paghahanda ng mga kagamitang tanaw dinig
Paghahanda ng mga kagamitang tanaw dinigPaghahanda ng mga kagamitang tanaw dinig
Paghahanda ng mga kagamitang tanaw dinig
 
Reviewer filipino
Reviewer filipinoReviewer filipino
Reviewer filipino
 
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)
Paghahanda at ebalwasyon ng kagamitang panturo (Sining at Agham sa Pagtuturo)
 
Fil 2112
Fil 2112Fil 2112
Fil 2112
 
Kinds of tests and testing
Kinds of tests and testingKinds of tests and testing
Kinds of tests and testing
 
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer
K to 12 Grade 3 LAPG FILIPINO GRAMMAR Reviewer
 
Types of Tests
Types of TestsTypes of Tests
Types of Tests
 
Distracter Analysis - Index of Effectiveness
Distracter Analysis - Index of EffectivenessDistracter Analysis - Index of Effectiveness
Distracter Analysis - Index of Effectiveness
 
Kagamitang panturo presentation
Kagamitang panturo presentationKagamitang panturo presentation
Kagamitang panturo presentation
 
KINDS OF TEST
KINDS OF TESTKINDS OF TEST
KINDS OF TEST
 
K to 12 Grade 3 FILIPINO NAT (National Achievement Test) Reviewer
K to 12 Grade 3 FILIPINO  NAT (National Achievement Test) ReviewerK to 12 Grade 3 FILIPINO  NAT (National Achievement Test) Reviewer
K to 12 Grade 3 FILIPINO NAT (National Achievement Test) Reviewer
 
Nat reviewer in filipino vi
Nat reviewer in filipino viNat reviewer in filipino vi
Nat reviewer in filipino vi
 

Similar to Item analysis

Item Development and Analysis WorksheetStudent Name.docx
Item Development and Analysis WorksheetStudent Name.docxItem Development and Analysis WorksheetStudent Name.docx
Item Development and Analysis WorksheetStudent Name.docxsleeperfindley
 
Topic 8b Item Analysis
Topic 8b Item AnalysisTopic 8b Item Analysis
Topic 8b Item AnalysisYee Bee Choo
 
Practical Language Testing by Fulcher (2010)
Practical Language Testing by Fulcher (2010)Practical Language Testing by Fulcher (2010)
Practical Language Testing by Fulcher (2010)Mahsa Farahanynia
 
Item analysis by Shabbir Sohal
Item analysis by Shabbir SohalItem analysis by Shabbir Sohal
Item analysis by Shabbir SohalShabbir Sohal
 
CHAPTER 6 Assessment of Learning 1
CHAPTER 6 Assessment of Learning 1CHAPTER 6 Assessment of Learning 1
CHAPTER 6 Assessment of Learning 1FriasKentOmer
 
ITEM ANALYSIS.pptx
ITEM ANALYSIS.pptxITEM ANALYSIS.pptx
ITEM ANALYSIS.pptxRizaGarganza
 
TEST ITEM ANALYSIS PRESENTATION 2022.ppt
TEST ITEM ANALYSIS PRESENTATION 2022.pptTEST ITEM ANALYSIS PRESENTATION 2022.ppt
TEST ITEM ANALYSIS PRESENTATION 2022.pptJennilynDescargar
 
Analyzingandusingtestitemdata 101012035435-phpapp02
Analyzingandusingtestitemdata 101012035435-phpapp02Analyzingandusingtestitemdata 101012035435-phpapp02
Analyzingandusingtestitemdata 101012035435-phpapp02cezz gonzaga
 
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...GalataaAGoobanaa
 
Administering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentAdministering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentNema Grace Medillo
 
Fdu item analysis (1).ppt revised by dd
Fdu item analysis (1).ppt revised by ddFdu item analysis (1).ppt revised by dd
Fdu item analysis (1).ppt revised by dddettmore
 
educatiinar.pptx
educatiinar.pptxeducatiinar.pptx
educatiinar.pptxNithuNithu7
 
item analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysisitem analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysisswatisheth8
 

Similar to Item analysis (20)

Item analysis2
Item analysis2Item analysis2
Item analysis2
 
Item Development and Analysis WorksheetStudent Name.docx
Item Development and Analysis WorksheetStudent Name.docxItem Development and Analysis WorksheetStudent Name.docx
Item Development and Analysis WorksheetStudent Name.docx
 
Topic 8b Item Analysis
Topic 8b Item AnalysisTopic 8b Item Analysis
Topic 8b Item Analysis
 
Item analysis report
Item analysis report Item analysis report
Item analysis report
 
item analysics
item analysicsitem analysics
item analysics
 
Item
ItemItem
Item
 
Practical Language Testing by Fulcher (2010)
Practical Language Testing by Fulcher (2010)Practical Language Testing by Fulcher (2010)
Practical Language Testing by Fulcher (2010)
 
Item analysis by Shabbir Sohal
Item analysis by Shabbir SohalItem analysis by Shabbir Sohal
Item analysis by Shabbir Sohal
 
CHAPTER 6 Assessment of Learning 1
CHAPTER 6 Assessment of Learning 1CHAPTER 6 Assessment of Learning 1
CHAPTER 6 Assessment of Learning 1
 
ITEM ANALYSIS.pptx
ITEM ANALYSIS.pptxITEM ANALYSIS.pptx
ITEM ANALYSIS.pptx
 
TEST ITEM ANALYSIS PRESENTATION 2022.ppt
TEST ITEM ANALYSIS PRESENTATION 2022.pptTEST ITEM ANALYSIS PRESENTATION 2022.ppt
TEST ITEM ANALYSIS PRESENTATION 2022.ppt
 
Analyzingandusingtestitemdata 101012035435-phpapp02
Analyzingandusingtestitemdata 101012035435-phpapp02Analyzingandusingtestitemdata 101012035435-phpapp02
Analyzingandusingtestitemdata 101012035435-phpapp02
 
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...
ITEM ANALYSIS 2023.pptx uses for exam development especially national examina...
 
Administering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessmentAdministering, analyzing, and improving the test or assessment
Administering, analyzing, and improving the test or assessment
 
Fdu item analysis (1).ppt revised by dd
Fdu item analysis (1).ppt revised by ddFdu item analysis (1).ppt revised by dd
Fdu item analysis (1).ppt revised by dd
 
Item analysis
Item analysisItem analysis
Item analysis
 
educatiinar.pptx
educatiinar.pptxeducatiinar.pptx
educatiinar.pptx
 
Unit. 6.doc
Unit. 6.docUnit. 6.doc
Unit. 6.doc
 
item analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysisitem analysis.pptx education pnc item analysis
item analysis.pptx education pnc item analysis
 
tools of research
tools of researchtools of research
tools of research
 

Item analysis

  • 2. Purpose of Item Analysis – Evaluates the quality of each item – Rationale: the quality of items determines the quality of test (i.e., reliability & validity) – May suggest ways of improving the measurement of a test – Can help with understanding why certain tests predict some criteria but not others
  • 3. Item Analysis  When analyzing the test items, we have several questions about the performance of each item. Some of these questions include: Are the items congruent with the test objectives? Are the items valid? Do they measure what they're supposed to measure? Are the items reliable? Do they measure consistently? How long does it take an examinee to complete each item? What items are most difficult to answer correctly? What items are easy? Are there any poor performing items that need to be discarded?
  • 4. Types of Item Analyses for CTT Three major types: 1. Assess quality of the distractors 2. Assess difficulty of the items 3. Assess how well an item differentiates between high and low performers
  • 5. DISTRACTOR ANALYSIS A. Multiple-Choke B. Multiply-Choice C. Multiple-Choice D. Multi-Choice
  • 6. Distractor Analysis First question of item analysis: How many people choose each response? If there is only one best response, then all other response options are distractors. Example from in-class assignment (N = 35): Which method has the best internal consistency? # a) projective test 1 b) peer ratings 1 c) forced choice 21 d) differences n.s. 12
  • 7. Distractor Analysis (cont’d) A perfect test item would have 2 characteristics: 1. Everyone who knows the item gets it right 2. People who do not know the item will have responses equally distributed across the wrong answers. It is not desirable to have one of the distractors chosen more often than the correct answer.  This result indicates a potential problem with the question. This distractor may be too similar to the correct answer and/or there may be something in either the stem or the alternatives that is misleading.
  • 8. Distractor Analysis (cont’d) Calculate the # of people expected to choose each of the distractors. If random same expected number for each wrong response (Figure 10-1). # of Persons N answering incorrectly 14 Exp. To Choose = = 4.7 Distractor Number of distractors 3
  • 9. Distractor Analysis (cont’d) When the number of persons choosing a distractor significantly exceeds the number expected, there are 2 possibilities: 1. It is possible that the choice reflects partial knowledge 2. The item is a poorly worded trick question  unpopular distractors may lower item and test difficulty because it is easily eliminated  extremely popular is likely to lower the reliability and validity of the test
  • 10. Item Difficulty Analysis  Description and How to Compute ex: a) (6 X 3) + 4 = ? b) 9 [1n(-3.68) X (1 – 1n(+3.68))] = ?  Itis often difficult to explain or define difficulty in terms of some intrinsic characteristic of the item  The only common thread of difficult items is that individuals did not know the answer
  • 11. Item Difficulty Percentage of test takers who respond correctly What if p = .00 What if p = 1.00?
  • 12. Item Difficulty – An item with a p value of .0 or 1.0 does not contribute to measuring individual differences and thus is certain to be useless – When comparing 2 test scores, we are interested in who had the higher score or the differences in scores – p value of .5 have most variation so seek items in this range and remove those with extreme values – can also be examined to determine proportion answering in a particular way for items that don’t have a “correct” answer
  • 13. Item Difficulty (cont.) What is the best p-value? – most optimal p-value = .50 – maximum discrimination between good and poor performers Should we only choose items of .50? When shouldn’t we?
  • 14. Should we only choose items of .50? Not necessarily ...  When wanting to screen the very top group of applicants (i.e., admission to university or medical school). Cutoffs may be much higher  Other institutions want a minimum level (i.e., minimum reading level) Cutoffs may be much lower
  • 15. Item Difficulty (cont.) Interpreting the p-value... example: 100 people take a test 15 got question 1 right What is the p-value? Is this an easy or hard item?
  • 16. Item Difficulty (cont.) Interpreting the p-value... example: 100 people take a test 70 got question 1 right What is the p-value? Is this an easy or hard item?
  • 17. Item Difficulty (cont’d) General Rules of Item Difficulty… p low (< .20) difficult test item p moderate (.20 - .80) moderately diff. p high (> .80) easy item
  • 18. ITEM DISCRIMINATION ... The extent to which an item differentiates people on the behavior that the test is designed to assess. the computed difference between the percentage of high achievers and the percentage of low achievers who got the item right.
  • 19. Item Discrimination (cont.)  compares the performance of upper group (with high test scores) and lower group (low test scores) on each item--% of test takers in each group who were correct
  • 20. Item Discrimination (cont’d): Discrimination Index (D)  Divide sample into TOP half and BOTTOM half (or TOP and BOTTOM third)  Compute Discrimination Index (D)
  • 21. Item Discrimination D =U-L U = # in the upper group correct response Total # in upper group L = # in the lower group correct response Total # in lower group The higher the value of D, the more adequately the item discriminates (The highest value is 1.0)
  • 22. Item Discrimination  seekitems with high positive numbers (those who do well on the test tend to get the item correct)  negative numbers (lower scorers on test more likely to get item correct) and low positive numbers (about the same proportion of low and high scorers get the item correct) don’t discriminate well and are discarded
  • 23. Item Discrimination (cont’d): Item-Total Correlation Correlation between each item (a correct response usually receives a score of 1 and an incorrect a score of zero) and the total test score. To which degree do item and test measures the same thing? Positive -item discriminates between high and low scores Near 0 - item does not discriminate between high & low Negative - scores on item and scores on test disagree
  • 24. Item Discrimination (cont’d): Item-Total Correlation Item-total correlations are directly related to reliability. Why? Because the more each item correlates with the test as a whole, the higher all items correlate with each other ( = higher alpha, internal consistency)
  • 25. Quantitative Item Analysis  Inter-item correlation matrix displays the correlation of each item with every other item  provides important information for increasing the test’s internal consistency  each item should be highly correlated with every other item measuring the same construct and not correlated with items measuring a different construct
  • 26. Quantitative Item Analysis  itemsthat are not highly correlated with other items measuring the same construct can and should be dropped to increase internal consistency
  • 27. Item Discrimination (cont’d): Interitem Correlation Possible causes for low inter-item correlation: a. Item badly written (revise) b. Item measures other attribute than rest of the test (discard) c. Item correlated with some items, but not with others: test measures 2 distinct attributes (subtests or subscales)