SlideShare a Scribd company logo
1 of 26
Item Response Theory

Advance Psychometric Theory
          CPS723P
      Dr. Carlo Magno
Importance of Test Theories
• Estimate examinee ability and
  how the contribution of error
  might be minimized
• Disattenuation of variables
• Reporting true scores or ability
  scores and associated
  confidence
Psychometric History
• Lord (1952, 1953) and other
  psychometricians were interested in
  psychometric models with which to assess
  examinees independently of the particular
  choice of items or assessment tasks that
  were used in the assessment.
• Measurement practices would be enhanced
  if item and test statistics would be made
  sample independent.
• Birnbaum (1957, 1958)
• George Rasch (1960)
• Wright (1968)
Limitations of the CTT
• Item difficulty and item discrimination
  are group dependent.
• The p and r values are dependent on
  the examinee sample from which they
  are taken.
• Scores are entirely test dependent.
• No basis to predict the performance of
  examinees on an item.
Assumptions in IRT
• Unidimensionality
  – Examinee performance is a single
    ability
• Response → Dichotomous
  – The relationship of examinee
    performance on each item and the
    ability measured by the test is
    described as monotonically
    increasing.
• Monotonicity of item performance
  and ability is typified in an item
  characteristic curve (ICC).
• Examinees with more ability have
  higher probabilities for giving
  correct answers to items than
  lower ability students
  (Hambleton, 1989).
• Mathematical model
                                   linking the observable
                                   dichotomously scored
                                   data (item performance)
                   b
          a                        to the unobservable data
                                   (ability)
c
                                 • Pi(θ) gives the probability
                                   of a correct response to
                                   item i as a function if
                                   ability (θ)
                                 • b is the probability of a
    b=item difficulty              correct answer (1+c)/2
    a=item discrimination
    c=psuedoguessing parameter
• Two-parameter
          model: c=0
        • One-parameter
a         model: c=0, a=1
    b
• Three items
  showing
  different item
  difficulties (b)
• Different levels
  of item
  discrimination
Polychotomous IRT Models
• Having more than 2 points in the
  responses (ex. 4 point scale)
• Partial credit model
• Graded response model
• Nominal model
• Rating scale model
Graded Response model for a 5-
point scale
• In IRT measurement framework,
  ability estimates of an examinee
  obtained from a test that vary difficulty
  will be the same.
• Because of the unchanging ability,
  measurement errors are smaller
• True score is determined each test.
• Item parameters are independent on
  the particular examinee sample used.
• Measurement error is estimated at
  each ability level.
Test Characteristic Curve (TCC)
                • TCC: Sum of ICC that
                  make up a test or
                  assessment and can be
                  used to predict scores of
                  examinees at given ability
                  levels.
                       TCC(Ѳ)=∑Pi(Ѳ)
                • Links the true score to the
                  underlying ability
                  measures by the test.
                • TCC shift to the right of
                  the ability scale=difficult
                  items
Item Information Function
          • I(Ѳ), Contribution of
            particular items to the
            assessment of ability.
          • Items with higher
            discriminating power
            contribute more to
            measurement precision
            than items with lower
            discriminating power.
          • Items tend to make their
            best contribution to
            measurement precision
            around their b value.
Item Information Function
1
                                                                       2
                                                                                                            2
                         1           2            3
0.8
                                                                    1.5


0.6                                           4
                                                                                             1

                                                                      1
0.4



0.2                                                                  0.5
                                                                                                                3

                                                                                                    4
  0                                                                    0

      –3   –2     –1         0           1    2       3                    –3   –2      –1         0                1   2   3
                       Ability (θ)                                                            Ability (θ)

            Four item characteristic curves                                     Item information for four test items


                        Figure 6: Item characteristics curves and corresponding item information functions
their corresponding IFF




Test Information Function
•   The sum of item information functions in a test.
•   Higher values of the a parameter increase the
    amount of information an item provides.
•   The lower the c parameter, the more information an
    item provides.
•
•   The more information provided by an assessment at
    a particular level, the smaller the errors associated
    with ability estimation.
2




1.5




  1




 0.5




      0

                                    0                                3

                             Ability (θ)



          Figure 7: Test information function for a four–item test
Item Parameter Invariance

          • Item/test characteristic
            functions and item/test
            information functions are
            integral features of IRT.
Benefits of Item
Response Models
• Item statistics that are independent of the
  groups from which they were estimated.
• Scores describing examinee proficiency or
  ability that are not dependent on test
  difficulty.
• Test models that provide a basis for
  matching items or assessment tasks to
  ability levels.
• Models that do not require strict parallel
  tests or assessments for assessing
  reliability.
Application of IRT on
Test Development
• Item Analysis
  – Determining sample invariant item
    parameters.
  – Utilizing goodness-of-fit criteria to
    detect items that do not fit the
    specified response model (χ2,
    analysis of residuals).
Application of IRT on
Test Development
• Item Selection
  – Assess the contribution of each
    item the test information function
    independent of other items.
– Using item information functions:
  • Describe the shape of the desired test
    information function vs. desired range
    abilities.
  • Select items with information functions
    that will fill up the hard to fill areas
    under the target information function
  • Calculate the test information function
    for the selected assessment material.
  • Continue selecting materials until the
    test information function approximates
    the target information function to a
    satisfactory degree.
• Item banking
  – Test developers can build an
    assessment to fit any desired test
    information function with items
    having sufficient properties.
  – Comparisons of items can be made
    across dissimilar samples.

More Related Content

What's hot

Test Reliability and Validity
Test Reliability and ValidityTest Reliability and Validity
Test Reliability and ValidityBrian Ebie
 
The wilcoxon matched pairs signed-ranks test
The wilcoxon matched pairs signed-ranks testThe wilcoxon matched pairs signed-ranks test
The wilcoxon matched pairs signed-ranks testRegent University
 
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋としてYoshitake Takebayashi
 
MANN WHITNEY U TEST.pptx
MANN WHITNEY U TEST.pptxMANN WHITNEY U TEST.pptx
MANN WHITNEY U TEST.pptxRichelleDordas1
 
Characteristics of a good test
Characteristics of a good test Characteristics of a good test
Characteristics of a good test Arash Yazdani
 
Introduction to Computerized Adaptive Testing (CAT)
Introduction to Computerized Adaptive Testing (CAT)Introduction to Computerized Adaptive Testing (CAT)
Introduction to Computerized Adaptive Testing (CAT)Nathan Thompson
 
7. logistics regression using spss
7. logistics regression using spss7. logistics regression using spss
7. logistics regression using spssDr Nisha Arora
 
Tetrachoric Correlation Coefficient
Tetrachoric Correlation CoefficientTetrachoric Correlation Coefficient
Tetrachoric Correlation CoefficientSharlaine Ruth
 
Validity and Reliability - Research Mangement
Validity and Reliability - Research MangementValidity and Reliability - Research Mangement
Validity and Reliability - Research MangementVinu Arpitha
 
Introduction to statistics 1
Introduction to statistics 1Introduction to statistics 1
Introduction to statistics 1Anwar Afridi
 
Reliability And Validity
Reliability And ValidityReliability And Validity
Reliability And ValidityJames Penny
 
Recode Age Variable in SPSS
Recode  Age  Variable in SPSSRecode  Age  Variable in SPSS
Recode Age Variable in SPSSRuth M Tappin
 
Kruskal Wallis test, Friedman test, Spearman Correlation
Kruskal Wallis test, Friedman test, Spearman CorrelationKruskal Wallis test, Friedman test, Spearman Correlation
Kruskal Wallis test, Friedman test, Spearman CorrelationRizwan S A
 
The kolmogorov smirnov test
The kolmogorov smirnov testThe kolmogorov smirnov test
The kolmogorov smirnov testSubhradeep Mitra
 
如何选择数据分析方法
如何选择数据分析方法如何选择数据分析方法
如何选择数据分析方法小熊 张
 

What's hot (20)

Test Reliability and Validity
Test Reliability and ValidityTest Reliability and Validity
Test Reliability and Validity
 
The wilcoxon matched pairs signed-ranks test
The wilcoxon matched pairs signed-ranks testThe wilcoxon matched pairs signed-ranks test
The wilcoxon matched pairs signed-ranks test
 
01 validity and its type
01 validity and its type01 validity and its type
01 validity and its type
 
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として
単一事例研究法と統計的推測:ベイズ流アプローチを架け橋として
 
MANN WHITNEY U TEST.pptx
MANN WHITNEY U TEST.pptxMANN WHITNEY U TEST.pptx
MANN WHITNEY U TEST.pptx
 
Characteristics of a good test
Characteristics of a good test Characteristics of a good test
Characteristics of a good test
 
Introduction to Computerized Adaptive Testing (CAT)
Introduction to Computerized Adaptive Testing (CAT)Introduction to Computerized Adaptive Testing (CAT)
Introduction to Computerized Adaptive Testing (CAT)
 
7. logistics regression using spss
7. logistics regression using spss7. logistics regression using spss
7. logistics regression using spss
 
Tetrachoric Correlation Coefficient
Tetrachoric Correlation CoefficientTetrachoric Correlation Coefficient
Tetrachoric Correlation Coefficient
 
Validity and Reliability - Research Mangement
Validity and Reliability - Research MangementValidity and Reliability - Research Mangement
Validity and Reliability - Research Mangement
 
Introduction to statistics 1
Introduction to statistics 1Introduction to statistics 1
Introduction to statistics 1
 
Reliability And Validity
Reliability And ValidityReliability And Validity
Reliability And Validity
 
elementary statistic
elementary statisticelementary statistic
elementary statistic
 
Recode Age Variable in SPSS
Recode  Age  Variable in SPSSRecode  Age  Variable in SPSS
Recode Age Variable in SPSS
 
Nonparametric Statistics
Nonparametric StatisticsNonparametric Statistics
Nonparametric Statistics
 
Kruskal Wallis test, Friedman test, Spearman Correlation
Kruskal Wallis test, Friedman test, Spearman CorrelationKruskal Wallis test, Friedman test, Spearman Correlation
Kruskal Wallis test, Friedman test, Spearman Correlation
 
Validity
ValidityValidity
Validity
 
Two Means Independent Samples
Two Means Independent Samples  Two Means Independent Samples
Two Means Independent Samples
 
The kolmogorov smirnov test
The kolmogorov smirnov testThe kolmogorov smirnov test
The kolmogorov smirnov test
 
如何选择数据分析方法
如何选择数据分析方法如何选择数据分析方法
如何选择数据分析方法
 

Similar to Irt 1 pl, 2pl, 3pl.pdf

11 adaptive testing-irt
11 adaptive testing-irt11 adaptive testing-irt
11 adaptive testing-irt宥均 林
 
Introduction to Item Response Theory
Introduction to Item Response TheoryIntroduction to Item Response Theory
Introduction to Item Response TheoryOpenThink Labs
 
HUDE 225Take Home Directions You are a psychologist working a.docx
HUDE 225Take Home Directions You are a psychologist working a.docxHUDE 225Take Home Directions You are a psychologist working a.docx
HUDE 225Take Home Directions You are a psychologist working a.docxwellesleyterresa
 
DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...Hakka Labs
 
Algorithm evaluation using item response theory
Algorithm evaluation using item response theoryAlgorithm evaluation using item response theory
Algorithm evaluation using item response theoryCSIRO
 
introduction to Statistical Theory.pptx
 introduction to Statistical Theory.pptx introduction to Statistical Theory.pptx
introduction to Statistical Theory.pptxDr.Shweta
 
06 distance learning standards-qti
06 distance learning standards-qti06 distance learning standards-qti
06 distance learning standards-qti宥均 林
 
Feature selection
Feature selectionFeature selection
Feature selectionDong Guo
 
LAK13 linkedup tutorial_evaluation_framework
LAK13 linkedup tutorial_evaluation_frameworkLAK13 linkedup tutorial_evaluation_framework
LAK13 linkedup tutorial_evaluation_frameworkHendrik Drachsler
 
Feature enginnering and selection
Feature enginnering and selectionFeature enginnering and selection
Feature enginnering and selectionDavis David
 
KnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProjectKnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProjectMarciano Moreno
 
Latent factor models for Collaborative Filtering
Latent factor models for Collaborative FilteringLatent factor models for Collaborative Filtering
Latent factor models for Collaborative Filteringsscdotopen
 
Handling Missing Attributes using Matrix Factorization 
Handling Missing Attributes using Matrix Factorization Handling Missing Attributes using Matrix Factorization 
Handling Missing Attributes using Matrix Factorization CS, NcState
 
Machine Learning for objective QoE assessment: Science, Myths and a look to t...
Machine Learning for objective QoE assessment: Science, Myths and a look to t...Machine Learning for objective QoE assessment: Science, Myths and a look to t...
Machine Learning for objective QoE assessment: Science, Myths and a look to t...Förderverein Technische Fakultät
 
Basic Engineering Design (Part 6): Test and Evaluate
Basic Engineering Design (Part 6): Test and EvaluateBasic Engineering Design (Part 6): Test and Evaluate
Basic Engineering Design (Part 6): Test and EvaluateDenise Wilson
 
Deependra pal, amul ppt , factor analysis
Deependra pal, amul ppt , factor analysis Deependra pal, amul ppt , factor analysis
Deependra pal, amul ppt , factor analysis Deependra Pal
 

Similar to Irt 1 pl, 2pl, 3pl.pdf (20)

11 adaptive testing-irt
11 adaptive testing-irt11 adaptive testing-irt
11 adaptive testing-irt
 
Introduction to Item Response Theory
Introduction to Item Response TheoryIntroduction to Item Response Theory
Introduction to Item Response Theory
 
HUDE 225Take Home Directions You are a psychologist working a.docx
HUDE 225Take Home Directions You are a psychologist working a.docxHUDE 225Take Home Directions You are a psychologist working a.docx
HUDE 225Take Home Directions You are a psychologist working a.docx
 
DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...DataEngConf SF16 - Three lessons learned from building a production machine l...
DataEngConf SF16 - Three lessons learned from building a production machine l...
 
Algorithm evaluation using item response theory
Algorithm evaluation using item response theoryAlgorithm evaluation using item response theory
Algorithm evaluation using item response theory
 
introduction to Statistical Theory.pptx
 introduction to Statistical Theory.pptx introduction to Statistical Theory.pptx
introduction to Statistical Theory.pptx
 
NCME_040916
NCME_040916NCME_040916
NCME_040916
 
Deep Feature Synthesis
Deep Feature SynthesisDeep Feature Synthesis
Deep Feature Synthesis
 
Presentations
PresentationsPresentations
Presentations
 
06 distance learning standards-qti
06 distance learning standards-qti06 distance learning standards-qti
06 distance learning standards-qti
 
Feature selection
Feature selectionFeature selection
Feature selection
 
LAK13 linkedup tutorial_evaluation_framework
LAK13 linkedup tutorial_evaluation_frameworkLAK13 linkedup tutorial_evaluation_framework
LAK13 linkedup tutorial_evaluation_framework
 
Week 11 collecting_data
Week 11 collecting_dataWeek 11 collecting_data
Week 11 collecting_data
 
Feature enginnering and selection
Feature enginnering and selectionFeature enginnering and selection
Feature enginnering and selection
 
KnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProjectKnowledgeFromDataAtScaleProject
KnowledgeFromDataAtScaleProject
 
Latent factor models for Collaborative Filtering
Latent factor models for Collaborative FilteringLatent factor models for Collaborative Filtering
Latent factor models for Collaborative Filtering
 
Handling Missing Attributes using Matrix Factorization 
Handling Missing Attributes using Matrix Factorization Handling Missing Attributes using Matrix Factorization 
Handling Missing Attributes using Matrix Factorization 
 
Machine Learning for objective QoE assessment: Science, Myths and a look to t...
Machine Learning for objective QoE assessment: Science, Myths and a look to t...Machine Learning for objective QoE assessment: Science, Myths and a look to t...
Machine Learning for objective QoE assessment: Science, Myths and a look to t...
 
Basic Engineering Design (Part 6): Test and Evaluate
Basic Engineering Design (Part 6): Test and EvaluateBasic Engineering Design (Part 6): Test and Evaluate
Basic Engineering Design (Part 6): Test and Evaluate
 
Deependra pal, amul ppt , factor analysis
Deependra pal, amul ppt , factor analysis Deependra pal, amul ppt , factor analysis
Deependra pal, amul ppt , factor analysis
 

More from Carlo Magno

Assessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxAssessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxCarlo Magno
 
Social and Emotional Learning
Social and Emotional LearningSocial and Emotional Learning
Social and Emotional LearningCarlo Magno
 
Educational assessment in the 4 ir
Educational assessment in the 4 irEducational assessment in the 4 ir
Educational assessment in the 4 irCarlo Magno
 
The process of research mentoring
The process of research mentoringThe process of research mentoring
The process of research mentoringCarlo Magno
 
Quality management services sustainability training
Quality management services sustainability trainingQuality management services sustainability training
Quality management services sustainability trainingCarlo Magno
 
Managing technology integration in schools
Managing technology integration in schoolsManaging technology integration in schools
Managing technology integration in schoolsCarlo Magno
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teachingCarlo Magno
 
Empowering educators on technology integration
Empowering educators on technology integrationEmpowering educators on technology integration
Empowering educators on technology integrationCarlo Magno
 
Designing an online lesson
Designing an online lessonDesigning an online lesson
Designing an online lessonCarlo Magno
 
Curriculum integration
Curriculum integrationCurriculum integration
Curriculum integrationCarlo Magno
 
Accountability in Developing Student Learning
Accountability in Developing Student LearningAccountability in Developing Student Learning
Accountability in Developing Student LearningCarlo Magno
 
The Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementThe Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementCarlo Magno
 
Guiding your child on their career decision making
Guiding your child on their career decision makingGuiding your child on their career decision making
Guiding your child on their career decision makingCarlo Magno
 
Assessing Science Inquiry Skills
Assessing Science Inquiry SkillsAssessing Science Inquiry Skills
Assessing Science Inquiry SkillsCarlo Magno
 
Assessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumAssessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumCarlo Magno
 
Quantitative analysis in language research
Quantitative analysis in language researchQuantitative analysis in language research
Quantitative analysis in language researchCarlo Magno
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teachingCarlo Magno
 
Hallmarks of textbook
Hallmarks of textbookHallmarks of textbook
Hallmarks of textbookCarlo Magno
 
managing the learner centered-classroom
managing the learner centered-classroommanaging the learner centered-classroom
managing the learner centered-classroomCarlo Magno
 
Assessing learning objectives
Assessing learning objectivesAssessing learning objectives
Assessing learning objectivesCarlo Magno
 

More from Carlo Magno (20)

Assessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptxAssessment Using the SOLO Framework.pptx
Assessment Using the SOLO Framework.pptx
 
Social and Emotional Learning
Social and Emotional LearningSocial and Emotional Learning
Social and Emotional Learning
 
Educational assessment in the 4 ir
Educational assessment in the 4 irEducational assessment in the 4 ir
Educational assessment in the 4 ir
 
The process of research mentoring
The process of research mentoringThe process of research mentoring
The process of research mentoring
 
Quality management services sustainability training
Quality management services sustainability trainingQuality management services sustainability training
Quality management services sustainability training
 
Managing technology integration in schools
Managing technology integration in schoolsManaging technology integration in schools
Managing technology integration in schools
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teaching
 
Empowering educators on technology integration
Empowering educators on technology integrationEmpowering educators on technology integration
Empowering educators on technology integration
 
Designing an online lesson
Designing an online lessonDesigning an online lesson
Designing an online lesson
 
Curriculum integration
Curriculum integrationCurriculum integration
Curriculum integration
 
Accountability in Developing Student Learning
Accountability in Developing Student LearningAccountability in Developing Student Learning
Accountability in Developing Student Learning
 
The Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School ImprovementThe Instructional leader: TOwards School Improvement
The Instructional leader: TOwards School Improvement
 
Guiding your child on their career decision making
Guiding your child on their career decision makingGuiding your child on their career decision making
Guiding your child on their career decision making
 
Assessing Science Inquiry Skills
Assessing Science Inquiry SkillsAssessing Science Inquiry Skills
Assessing Science Inquiry Skills
 
Assessment in the Social Studies Curriculum
Assessment in the Social Studies CurriculumAssessment in the Social Studies Curriculum
Assessment in the Social Studies Curriculum
 
Quantitative analysis in language research
Quantitative analysis in language researchQuantitative analysis in language research
Quantitative analysis in language research
 
Integrating technology in teaching
Integrating technology in teachingIntegrating technology in teaching
Integrating technology in teaching
 
Hallmarks of textbook
Hallmarks of textbookHallmarks of textbook
Hallmarks of textbook
 
managing the learner centered-classroom
managing the learner centered-classroommanaging the learner centered-classroom
managing the learner centered-classroom
 
Assessing learning objectives
Assessing learning objectivesAssessing learning objectives
Assessing learning objectives
 

Irt 1 pl, 2pl, 3pl.pdf

  • 1. Item Response Theory Advance Psychometric Theory CPS723P Dr. Carlo Magno
  • 2. Importance of Test Theories • Estimate examinee ability and how the contribution of error might be minimized • Disattenuation of variables • Reporting true scores or ability scores and associated confidence
  • 3. Psychometric History • Lord (1952, 1953) and other psychometricians were interested in psychometric models with which to assess examinees independently of the particular choice of items or assessment tasks that were used in the assessment. • Measurement practices would be enhanced if item and test statistics would be made sample independent. • Birnbaum (1957, 1958) • George Rasch (1960) • Wright (1968)
  • 4. Limitations of the CTT • Item difficulty and item discrimination are group dependent. • The p and r values are dependent on the examinee sample from which they are taken. • Scores are entirely test dependent. • No basis to predict the performance of examinees on an item.
  • 5. Assumptions in IRT • Unidimensionality – Examinee performance is a single ability • Response → Dichotomous – The relationship of examinee performance on each item and the ability measured by the test is described as monotonically increasing.
  • 6. • Monotonicity of item performance and ability is typified in an item characteristic curve (ICC). • Examinees with more ability have higher probabilities for giving correct answers to items than lower ability students (Hambleton, 1989).
  • 7. • Mathematical model linking the observable dichotomously scored data (item performance) b a to the unobservable data (ability) c • Pi(θ) gives the probability of a correct response to item i as a function if ability (θ) • b is the probability of a b=item difficulty correct answer (1+c)/2 a=item discrimination c=psuedoguessing parameter
  • 8. • Two-parameter model: c=0 • One-parameter a model: c=0, a=1 b
  • 9. • Three items showing different item difficulties (b)
  • 10. • Different levels of item discrimination
  • 11.
  • 12. Polychotomous IRT Models • Having more than 2 points in the responses (ex. 4 point scale) • Partial credit model • Graded response model • Nominal model • Rating scale model
  • 13. Graded Response model for a 5- point scale
  • 14. • In IRT measurement framework, ability estimates of an examinee obtained from a test that vary difficulty will be the same. • Because of the unchanging ability, measurement errors are smaller • True score is determined each test. • Item parameters are independent on the particular examinee sample used. • Measurement error is estimated at each ability level.
  • 15. Test Characteristic Curve (TCC) • TCC: Sum of ICC that make up a test or assessment and can be used to predict scores of examinees at given ability levels. TCC(Ѳ)=∑Pi(Ѳ) • Links the true score to the underlying ability measures by the test. • TCC shift to the right of the ability scale=difficult items
  • 16. Item Information Function • I(Ѳ), Contribution of particular items to the assessment of ability. • Items with higher discriminating power contribute more to measurement precision than items with lower discriminating power. • Items tend to make their best contribution to measurement precision around their b value.
  • 18. 1 2 2 1 2 3 0.8 1.5 0.6 4 1 1 0.4 0.2 0.5 3 4 0 0 –3 –2 –1 0 1 2 3 –3 –2 –1 0 1 2 3 Ability (θ) Ability (θ) Four item characteristic curves Item information for four test items Figure 6: Item characteristics curves and corresponding item information functions
  • 19. their corresponding IFF Test Information Function • The sum of item information functions in a test. • Higher values of the a parameter increase the amount of information an item provides. • The lower the c parameter, the more information an item provides. • • The more information provided by an assessment at a particular level, the smaller the errors associated with ability estimation.
  • 20. 2 1.5 1 0.5 0 0 3 Ability (θ) Figure 7: Test information function for a four–item test
  • 21. Item Parameter Invariance • Item/test characteristic functions and item/test information functions are integral features of IRT.
  • 22. Benefits of Item Response Models • Item statistics that are independent of the groups from which they were estimated. • Scores describing examinee proficiency or ability that are not dependent on test difficulty. • Test models that provide a basis for matching items or assessment tasks to ability levels. • Models that do not require strict parallel tests or assessments for assessing reliability.
  • 23. Application of IRT on Test Development • Item Analysis – Determining sample invariant item parameters. – Utilizing goodness-of-fit criteria to detect items that do not fit the specified response model (χ2, analysis of residuals).
  • 24. Application of IRT on Test Development • Item Selection – Assess the contribution of each item the test information function independent of other items.
  • 25. – Using item information functions: • Describe the shape of the desired test information function vs. desired range abilities. • Select items with information functions that will fill up the hard to fill areas under the target information function • Calculate the test information function for the selected assessment material. • Continue selecting materials until the test information function approximates the target information function to a satisfactory degree.
  • 26. • Item banking – Test developers can build an assessment to fit any desired test information function with items having sufficient properties. – Comparisons of items can be made across dissimilar samples.