SlideShare a Scribd company logo
1 of 37
Validity and Reliability
Good Assessments 
 Regardless of its other characteristics, 
the most important characteristics in 
determining the usefulness of 
assessment information are its validity 
and reliability.
Validity 
Validity has been defined as referring to 
the appropriateness, correctness, 
meaningfulness, and usefulness of the 
specific inferences researchers make 
based on the data they collect.
Validity 
 Validity is an evaluation of the 
adequacy and appropriateness of the 
interpretations and uses of assessment 
results. 
 Validity is always determined by a 
judgment made by the test user.
Nature of Validity 
 Appropriateness of the interpretation and use made 
of the results of an assessment procedure for a 
given group of individuals 
 A matter of degree 
 Specific to some particular use or interpretation 
 A unitary concept 
 Overall evaluate judgment
Sources of Validity 
 Test content 
 Response process 
 Internal structure
Evidence of Validity 
There are 3 types of evidence a researcher might 
collect: 
Content-related evidence of validity 
Content and format of the instrument 
Criterion-related evidence of validity 
Relationship between scores obtained using the instrument and 
scores obtained 
Construct-related evidence of validity 
Psychological construct being measured by the instrument
Content Validity 
 Compare the assessment tasks to the 
specifications describing the task domain under 
consideration 
 The extent to which an assessment procedure 
adequately represents the content of the 
assessment domain being sampled
Content Validity 
Example: 
The content validity of the instrument has been 
validated by four HR managers to identify the 
relevancy of the human resource management 
practices in automobile industry. 
Their CV were attached together in Appendices 2 for 
reference.
Example 
After the pilot studies, the PES instrument was subjected to 
content validation by a panel of senior teachers and 
educationists (Credentials enclosed in Appendix G). 
All members of the panel worked independently. Each was 
presented with a set of the instrument and informed of the 
purpose of the instrument. 
They were then requested to study the items and decide on the 
suitability of the items. They were also asked if any other items 
should be included to fulfil the purpose of the instrument, and 
to comment on any part of the scale's items that they felt needed 
amending or clarification.
Example 
Appendix G: Credentials of Panel for Instrument Validation 
Name, Qualification and Experience 
1. DR. FOO SAY FOOI 
He is a lecturer in Educational Administration, Faculty of Educational Studies, 
Universiti Putra Malaysia. 
2. ASSOC. PROF. DR. TURIMAN SUANDI 
A lecturer in the Faculty of Educational Studies, Universiti Putra Malaysia. Among 
others, he lectures on Research Design. Presently Deputy Dean for Development 
and Students. 
3. ASSOC. PROF. DR. ZAIDATOL AKMALIAH L. PIHIE 
A lecturer in the field of Educational Administration at the Faculty of Educational 
Studies, Universiti Putra Malaysia. She had written articles and books related to 
educational management and administration. She is the Deputy Dean in the 
FPPUPM.
Construct Validity 
 Establish the meaning of assessment results by controlling 
the development of the assessment, evaluate the cognitive 
procession used by students to perform tasks, evaluate the 
relationships of the scores with other relevant measures, 
and experimentally determine what factors influence 
performance. 
 The extent to which empirical evidence confirms that an 
inferred construct exists and that a given assessment 
procedure is measuring the inferred construct accurately
Criterion Validity 
 Compare assessment results with another 
performance obtained at a later date or with 
another measure of performance obtained 
concurrently (for estimating present status). 
 The degree to which performance on an 
assessment procedure accurately predicts a 
student’s performance on an external criterion
Factors Influence Validity 
 Unclear directions 
 Reading vocabulary and sentence structure 
too difficult 
 Ambiguity 
 Inadequate time limits 
 Overemphasis of easy-to-assess aspects of 
domain at the expense of important but 
difficult-to-assess aspects
Factors Influence Validity 
 Test items inappropriate for the outcomes 
being measured 
 Poorly constructed test items 
 Test too short 
 Improper arrangement of items 
 Identifiable pattern of answers 
 First impression 
 Personal theories
Reliability 
 Refers to the consistency of measurement, 
that is, how consistent test scores or other 
assessment results are from one 
measurement to another.
Nature of Reliability 
 Refers to the results obtained with an instrument 
and not to the instrument itself 
 An estimate of reliability always refers to a 
particular type of consistency (time, task, students, 
rater) 
 Reliability is a necessary but not sufficient 
condition for validity 
 Reliability is assessed primarily with statistical 
indices
Terminology 
 Correlation Coefficient: a statistic that indicates 
the degree of relationship between any two sets of 
scores obtained from the same group of individuals
Methods of estimating Reliability 
 Test-retest 
 Equivalent-forms 
 Internal Consistency 
 Split-half 
 Kuder-Richardson 
 Alpha Coefficient 
 Interrater 
What method?
Test-Retest 
 Give the same test twice to the same group 
with some time interval between tests, from 
several minutes to several days
Equivalent Forms 
 Give two forms of the test to the same group 
in close succession
Test-Retest with equivalent forms 
 Give two forms of the test to the same group 
with an increased time interval between 
forms
Internal-Consistency Methods 
There are several internal-consistency methods that require 
only one administration of an instrument. 
Split-half Procedure: involves scoring two halves of a test 
separately for each subject and calculating the correlation 
coefficient between the two scores. 
Kuder-Richardson Approaches: (KR20 and KR21) requires 3 
pieces of information: 
Number of items on the test 
The mean 
The standard deviation 
Considered the most frequent method for determining internal 
consistency 
Alpha Coefficient: a general form of the KR20 used to 
calculate the reliability of items that are not scored right vs. 
wrong.
Internal Consistency 
If Likert scales are used to represent the response 
choices, analysis for internal consistency can be 
accomplished using Cronbach’s Alpha. 
Job Satisfaction Questionnaie 
No. Component Cronbach Alpha 
1 Intrinsic Factors .75 
2 Extrinsic Factors .83
Interrater 
 Give a set of student responses requiring 
judgmental scoring to two or more raters 
and have them independently score the 
responses
Factors Influencing 
Reliability Measures 
 Number of assessment tasks 
 Spread of scores 
 Objectivity
Validity-Reliability 
 Reliability is a necessary but 
insufficient condition for validity. 
 Reliability (consistency) of 
measurement is needed to obtain valid 
results, but we can have reliability 
without validity.
Interactions of 
Validity and Reliability 
Lack of validity 
and reliability 
Good reliability 
But poor validity 
Good validity 
and reliability
Other Factors Influencing 
Validity and Reliability
Objectivity 
 Refers to the degree to which equally 
competent scorers obtain the same 
results. 
 The test items (objective type) 
 The results (not influenced by scorer’s 
judgment or opinion)
Practicality 
 Economical from the viewpoint of both 
time and money. 
 Easily administered and scored. 
 Produce results that can be accurately 
interpreted and applied by available 
school personnel
Usability 
 Ease of administration 
 Time required for administration 
 Ease of interpretation and application 
 Availability of equivalent or 
comparable forms 
 Cost of testing
The Standard Error of 
Measurement 
 The index used in educational assessment to 
describe the consistency of a particular 
person’s performance(s) 
 Reflection of the consistency of an 
individual’s scores 
 The higher the reliability of a test, the 
smaller the SE of measurable will be
Teachers’ Ethical Responsibilities 
Regarding Assessment 
 Make fair and impartial decisions 
 Construct and administer fair and clear 
assessments 
 Motivate pupils to do their best 
 Teach pupils the varied types of assessments 
 Provide opportunities for pupils to practice 
test approaches 
 Make reasonable accommodations for 
students with disabilities
Summary 
 Concept of Validity 
 Factors influencing validity 
 Concept of Reliability 
 Factors influencing reliability
Questions 
How important is the concept of ‘validity’ in 
classroom assessment? 
To what extent do reliability concerns affect 
the construction of a good test?
References 
 Linn, R. L., & Miller, M. D. (2005). 
Measurement and assessment in testing. 
International Edition. 9th Ed. New Jersey: 
Pearson Merrill Prentice Hall 
 Airasian, P.W. (2005). Classroom 
assessment: Concepts and applications. 5th 
Ed. Boston: McGrawHill.

More Related Content

What's hot

Validity, its types, measurement & factors.
Validity, its types, measurement & factors.Validity, its types, measurement & factors.
Validity, its types, measurement & factors.
Maheen Iftikhar
 
Characteristics of a good test
Characteristics of a good testCharacteristics of a good test
Characteristics of a good test
cyrilcoscos
 
reliability presentation.pptx
reliability presentation.pptxreliability presentation.pptx
reliability presentation.pptx
Ramsha Makhdum
 
Assessment and evaluation for learning
Assessment and evaluation for learningAssessment and evaluation for learning
Assessment and evaluation for learning
arihantcollege9
 

What's hot (20)

Validity, its types, measurement & factors.
Validity, its types, measurement & factors.Validity, its types, measurement & factors.
Validity, its types, measurement & factors.
 
Good test , Reliability and Validity of a good test
Good test , Reliability and Validity of a good testGood test , Reliability and Validity of a good test
Good test , Reliability and Validity of a good test
 
Characteristics of a good test
Characteristics of a good testCharacteristics of a good test
Characteristics of a good test
 
Validity and reliability in assessment.
Validity and reliability in assessment. Validity and reliability in assessment.
Validity and reliability in assessment.
 
reliability presentation.pptx
reliability presentation.pptxreliability presentation.pptx
reliability presentation.pptx
 
Test administration (Educational Assessment)
Test administration (Educational Assessment)Test administration (Educational Assessment)
Test administration (Educational Assessment)
 
Norm referenced and Criterion Referenced Test
Norm referenced and Criterion Referenced TestNorm referenced and Criterion Referenced Test
Norm referenced and Criterion Referenced Test
 
Validity in Assessment
Validity in AssessmentValidity in Assessment
Validity in Assessment
 
Validity and objectivity of tests
Validity and objectivity of testsValidity and objectivity of tests
Validity and objectivity of tests
 
01 validity and its type
01 validity and its type01 validity and its type
01 validity and its type
 
Characteristics of Assessment
Characteristics of Assessment Characteristics of Assessment
Characteristics of Assessment
 
Reliability types
Reliability typesReliability types
Reliability types
 
Assessment and evaluation for learning
Assessment and evaluation for learningAssessment and evaluation for learning
Assessment and evaluation for learning
 
Qualities of a Good Test
Qualities of a Good TestQualities of a Good Test
Qualities of a Good Test
 
Types of test items
Types of test itemsTypes of test items
Types of test items
 
Meaning of Test, Testing and Evaluation
Meaning of Test, Testing and EvaluationMeaning of Test, Testing and Evaluation
Meaning of Test, Testing and Evaluation
 
Individual vs group test
Individual vs group testIndividual vs group test
Individual vs group test
 
Distinction among measurement, assessment and evaluation
Distinction among measurement, assessment and evaluationDistinction among measurement, assessment and evaluation
Distinction among measurement, assessment and evaluation
 
Reliability
ReliabilityReliability
Reliability
 
Administration/Conducting the Test
Administration/Conducting the TestAdministration/Conducting the Test
Administration/Conducting the Test
 

Viewers also liked

Validity, reliability & practicality
Validity, reliability & practicalityValidity, reliability & practicality
Validity, reliability & practicality
Samcruz5
 
Presentation Validity & Reliability
Presentation Validity & ReliabilityPresentation Validity & Reliability
Presentation Validity & Reliability
songoten77
 
Reliability
ReliabilityReliability
Reliability
Roi Xcel
 
Reliability & validity
Reliability & validityReliability & validity
Reliability & validity
Amin Altayeb
 
Safety-Critical Embedded Systems Course
Safety-Critical Embedded Systems CourseSafety-Critical Embedded Systems Course
Safety-Critical Embedded Systems Course
paupo
 
Lecture 3 measurement, reliability and validity (
Lecture   3 measurement, reliability and validity (Lecture   3 measurement, reliability and validity (
Lecture 3 measurement, reliability and validity (
La Islaa
 
Development of health measurement scales – part 2
Development of health measurement scales – part 2Development of health measurement scales – part 2
Development of health measurement scales – part 2
Rizwan S A
 
A haunted house analysis (Virginia Woolf)
A haunted house analysis (Virginia Woolf)A haunted house analysis (Virginia Woolf)
A haunted house analysis (Virginia Woolf)
Fera Komalasari
 
Reliability for Total Quality Management (TQM)
Reliability for Total Quality Management (TQM)Reliability for Total Quality Management (TQM)
Reliability for Total Quality Management (TQM)
Lavinia Layson
 

Viewers also liked (20)

Validity, reliability & practicality
Validity, reliability & practicalityValidity, reliability & practicality
Validity, reliability & practicality
 
Presentation Validity & Reliability
Presentation Validity & ReliabilityPresentation Validity & Reliability
Presentation Validity & Reliability
 
Validity and reliability of questionnaires
Validity and reliability of questionnairesValidity and reliability of questionnaires
Validity and reliability of questionnaires
 
validity its types and importance
validity its types and importancevalidity its types and importance
validity its types and importance
 
B.tech admission in idia
B.tech admission in idiaB.tech admission in idia
B.tech admission in idia
 
Reliability
ReliabilityReliability
Reliability
 
Reliability and validity
Reliability and validityReliability and validity
Reliability and validity
 
8. validity and reliability of research instruments
8. validity and reliability of research instruments8. validity and reliability of research instruments
8. validity and reliability of research instruments
 
Reliability & validity
Reliability & validityReliability & validity
Reliability & validity
 
Reliability And Validity
Reliability And ValidityReliability And Validity
Reliability And Validity
 
Safety-Critical Embedded Systems Course
Safety-Critical Embedded Systems CourseSafety-Critical Embedded Systems Course
Safety-Critical Embedded Systems Course
 
Reliability bachman 1990 chapter 6
Reliability bachman 1990 chapter 6Reliability bachman 1990 chapter 6
Reliability bachman 1990 chapter 6
 
Lecture 3 measurement, reliability and validity (
Lecture   3 measurement, reliability and validity (Lecture   3 measurement, reliability and validity (
Lecture 3 measurement, reliability and validity (
 
Development of health measurement scales – part 2
Development of health measurement scales – part 2Development of health measurement scales – part 2
Development of health measurement scales – part 2
 
Reliability and validity1
Reliability and validity1Reliability and validity1
Reliability and validity1
 
good test Characteristics
good test Characteristics  good test Characteristics
good test Characteristics
 
1 Reliability and Validity in Physical Therapy Tests
1  Reliability and Validity in Physical Therapy Tests1  Reliability and Validity in Physical Therapy Tests
1 Reliability and Validity in Physical Therapy Tests
 
Reliability in Language Testing
Reliability in Language Testing Reliability in Language Testing
Reliability in Language Testing
 
A haunted house analysis (Virginia Woolf)
A haunted house analysis (Virginia Woolf)A haunted house analysis (Virginia Woolf)
A haunted house analysis (Virginia Woolf)
 
Reliability for Total Quality Management (TQM)
Reliability for Total Quality Management (TQM)Reliability for Total Quality Management (TQM)
Reliability for Total Quality Management (TQM)
 

Similar to Week 8 & 9 - Validity and Reliability

Test characteristics
Test characteristicsTest characteristics
Test characteristics
Samcruz5
 
Evaluation and measurement nursing education
Evaluation and measurement nursing educationEvaluation and measurement nursing education
Evaluation and measurement nursing education
parvathysree
 
research-instruments (1).pptx
research-instruments (1).pptxresearch-instruments (1).pptx
research-instruments (1).pptx
JCronus
 
Assessing learning in Instructional Design
Assessing learning in Instructional DesignAssessing learning in Instructional Design
Assessing learning in Instructional Design
leesha roberts
 
constructionoftests-211015110341 (1).pptx
constructionoftests-211015110341 (1).pptxconstructionoftests-211015110341 (1).pptx
constructionoftests-211015110341 (1).pptx
GajeSingh9
 

Similar to Week 8 & 9 - Validity and Reliability (20)

Validity and reliability
Validity and reliabilityValidity and reliability
Validity and reliability
 
Principles of assessment
Principles of assessmentPrinciples of assessment
Principles of assessment
 
LESSON 6 JBF 361.pptx
LESSON 6 JBF 361.pptxLESSON 6 JBF 361.pptx
LESSON 6 JBF 361.pptx
 
Validity of a Research Tool
Validity of a Research ToolValidity of a Research Tool
Validity of a Research Tool
 
Understanding reliability and validity
Understanding reliability and validityUnderstanding reliability and validity
Understanding reliability and validity
 
Understanding Reliability and Validity.pptx
Understanding Reliability and Validity.pptxUnderstanding Reliability and Validity.pptx
Understanding Reliability and Validity.pptx
 
Understanding Reliability and Validity
Understanding Reliability and ValidityUnderstanding Reliability and Validity
Understanding Reliability and Validity
 
EM&E.pptx
EM&E.pptxEM&E.pptx
EM&E.pptx
 
Test characteristics
Test characteristicsTest characteristics
Test characteristics
 
Evaluation and measurement nursing education
Evaluation and measurement nursing educationEvaluation and measurement nursing education
Evaluation and measurement nursing education
 
research-instruments (1).pptx
research-instruments (1).pptxresearch-instruments (1).pptx
research-instruments (1).pptx
 
Assessing learning in Instructional Design
Assessing learning in Instructional DesignAssessing learning in Instructional Design
Assessing learning in Instructional Design
 
Measurement & Evaluation pptx
Measurement & Evaluation pptxMeasurement & Evaluation pptx
Measurement & Evaluation pptx
 
Construction of Tests
Construction of TestsConstruction of Tests
Construction of Tests
 
constructionoftests-211015110341 (1).pptx
constructionoftests-211015110341 (1).pptxconstructionoftests-211015110341 (1).pptx
constructionoftests-211015110341 (1).pptx
 
Concept and nature of measurment and evaluation (1)
Concept and nature of measurment and evaluation (1)Concept and nature of measurment and evaluation (1)
Concept and nature of measurment and evaluation (1)
 
Validity
ValidityValidity
Validity
 
Learning_activity1_Medina.Llerena_Karen.Tamara.
Learning_activity1_Medina.Llerena_Karen.Tamara.Learning_activity1_Medina.Llerena_Karen.Tamara.
Learning_activity1_Medina.Llerena_Karen.Tamara.
 
Validity Evidence
Validity EvidenceValidity Evidence
Validity Evidence
 
Validity, reliabiltiy and alignment to determine the effectiveness of assessment
Validity, reliabiltiy and alignment to determine the effectiveness of assessmentValidity, reliabiltiy and alignment to determine the effectiveness of assessment
Validity, reliabiltiy and alignment to determine the effectiveness of assessment
 

More from Syamsul Nor Azlan Mohamad

More from Syamsul Nor Azlan Mohamad (20)

Seni Seramik Slide PPT
Seni Seramik Slide PPTSeni Seramik Slide PPT
Seni Seramik Slide PPT
 
Modul Pedagogi Responsif Budaya Kraf Tradisional Seramik Pendidikan Seni Visual
Modul Pedagogi Responsif Budaya Kraf Tradisional Seramik Pendidikan Seni VisualModul Pedagogi Responsif Budaya Kraf Tradisional Seramik Pendidikan Seni Visual
Modul Pedagogi Responsif Budaya Kraf Tradisional Seramik Pendidikan Seni Visual
 
Tanglung Slide PPT
Tanglung Slide PPTTanglung Slide PPT
Tanglung Slide PPT
 
Modul 3 Seni Sulaman
Modul 3 Seni SulamanModul 3 Seni Sulaman
Modul 3 Seni Sulaman
 
Modul 2 Seni Tekat
Modul 2 Seni TekatModul 2 Seni Tekat
Modul 2 Seni Tekat
 
Modul 1 Tanglung
Modul 1 TanglungModul 1 Tanglung
Modul 1 Tanglung
 
Seni Tekat Slide PPT
Seni Tekat Slide PPTSeni Tekat Slide PPT
Seni Tekat Slide PPT
 
Seni Sulaman 1 Slide PPT
Seni Sulaman 1 Slide PPTSeni Sulaman 1 Slide PPT
Seni Sulaman 1 Slide PPT
 
Seni Sulaman 2 Slide PPT
Seni Sulaman 2 Slide PPTSeni Sulaman 2 Slide PPT
Seni Sulaman 2 Slide PPT
 
Seni Sulaman 4 Slide PPT
Seni Sulaman 4 Slide PPTSeni Sulaman 4 Slide PPT
Seni Sulaman 4 Slide PPT
 
Seni Sulaman 3 Slide PPT
Seni Sulaman 3 Slide PPTSeni Sulaman 3 Slide PPT
Seni Sulaman 3 Slide PPT
 
RPH Responsif Budaya
RPH Responsif BudayaRPH Responsif Budaya
RPH Responsif Budaya
 
Modul 3 Seni Sulaman
Modul 3 Seni SulamanModul 3 Seni Sulaman
Modul 3 Seni Sulaman
 
Modul 2 Seni Tekat
Modul 2 Seni TekatModul 2 Seni Tekat
Modul 2 Seni Tekat
 
Modul 1 Permainan Tanglung
Modul 1 Permainan TanglungModul 1 Permainan Tanglung
Modul 1 Permainan Tanglung
 
A Guideline on How to Write and Publish a Review Paper by prof. dr. saidur ra...
A Guideline on How to Write and Publish a Review Paper by prof. dr. saidur ra...A Guideline on How to Write and Publish a Review Paper by prof. dr. saidur ra...
A Guideline on How to Write and Publish a Review Paper by prof. dr. saidur ra...
 
Teaching Design and Technology
Teaching Design and Technology Teaching Design and Technology
Teaching Design and Technology
 
Specific Approaches To Curriculum Development
Specific Approaches To Curriculum DevelopmentSpecific Approaches To Curriculum Development
Specific Approaches To Curriculum Development
 
National philosophy of education
National philosophy of educationNational philosophy of education
National philosophy of education
 
How to evaluate an art program
How to evaluate an art programHow to evaluate an art program
How to evaluate an art program
 

Recently uploaded

Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
negromaestrong
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
Chris Hunter
 
Gardella_PRCampaignConclusion Pitch Letter
Gardella_PRCampaignConclusion Pitch LetterGardella_PRCampaignConclusion Pitch Letter
Gardella_PRCampaignConclusion Pitch Letter
MateoGardella
 

Recently uploaded (20)

Unit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptxUnit-IV- Pharma. Marketing Channels.pptx
Unit-IV- Pharma. Marketing Channels.pptx
 
Key note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdfKey note speaker Neum_Admir Softic_ENG.pdf
Key note speaker Neum_Admir Softic_ENG.pdf
 
Z Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot GraphZ Score,T Score, Percential Rank and Box Plot Graph
Z Score,T Score, Percential Rank and Box Plot Graph
 
Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17Advanced Views - Calendar View in Odoo 17
Advanced Views - Calendar View in Odoo 17
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
ICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptxICT Role in 21st Century Education & its Challenges.pptx
ICT Role in 21st Century Education & its Challenges.pptx
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Unit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptxUnit-IV; Professional Sales Representative (PSR).pptx
Unit-IV; Professional Sales Representative (PSR).pptx
 
psychiatric nursing HISTORY COLLECTION .docx
psychiatric  nursing HISTORY  COLLECTION  .docxpsychiatric  nursing HISTORY  COLLECTION  .docx
psychiatric nursing HISTORY COLLECTION .docx
 
Class 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdfClass 11th Physics NEET formula sheet pdf
Class 11th Physics NEET formula sheet pdf
 
Application orientated numerical on hev.ppt
Application orientated numerical on hev.pptApplication orientated numerical on hev.ppt
Application orientated numerical on hev.ppt
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024Mehran University Newsletter Vol-X, Issue-I, 2024
Mehran University Newsletter Vol-X, Issue-I, 2024
 
fourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writingfourth grading exam for kindergarten in writing
fourth grading exam for kindergarten in writing
 
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptxBasic Civil Engineering first year Notes- Chapter 4 Building.pptx
Basic Civil Engineering first year Notes- Chapter 4 Building.pptx
 
Seal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptxSeal of Good Local Governance (SGLG) 2024Final.pptx
Seal of Good Local Governance (SGLG) 2024Final.pptx
 
Unit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptxUnit-V; Pricing (Pharma Marketing Management).pptx
Unit-V; Pricing (Pharma Marketing Management).pptx
 
Making and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdfMaking and Justifying Mathematical Decisions.pdf
Making and Justifying Mathematical Decisions.pdf
 
Gardella_PRCampaignConclusion Pitch Letter
Gardella_PRCampaignConclusion Pitch LetterGardella_PRCampaignConclusion Pitch Letter
Gardella_PRCampaignConclusion Pitch Letter
 
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17  How to Extend Models Using Mixin ClassesMixin Classes in Odoo 17  How to Extend Models Using Mixin Classes
Mixin Classes in Odoo 17 How to Extend Models Using Mixin Classes
 

Week 8 & 9 - Validity and Reliability

  • 2. Good Assessments  Regardless of its other characteristics, the most important characteristics in determining the usefulness of assessment information are its validity and reliability.
  • 3. Validity Validity has been defined as referring to the appropriateness, correctness, meaningfulness, and usefulness of the specific inferences researchers make based on the data they collect.
  • 4. Validity  Validity is an evaluation of the adequacy and appropriateness of the interpretations and uses of assessment results.  Validity is always determined by a judgment made by the test user.
  • 5. Nature of Validity  Appropriateness of the interpretation and use made of the results of an assessment procedure for a given group of individuals  A matter of degree  Specific to some particular use or interpretation  A unitary concept  Overall evaluate judgment
  • 6. Sources of Validity  Test content  Response process  Internal structure
  • 7. Evidence of Validity There are 3 types of evidence a researcher might collect: Content-related evidence of validity Content and format of the instrument Criterion-related evidence of validity Relationship between scores obtained using the instrument and scores obtained Construct-related evidence of validity Psychological construct being measured by the instrument
  • 8. Content Validity  Compare the assessment tasks to the specifications describing the task domain under consideration  The extent to which an assessment procedure adequately represents the content of the assessment domain being sampled
  • 9. Content Validity Example: The content validity of the instrument has been validated by four HR managers to identify the relevancy of the human resource management practices in automobile industry. Their CV were attached together in Appendices 2 for reference.
  • 10. Example After the pilot studies, the PES instrument was subjected to content validation by a panel of senior teachers and educationists (Credentials enclosed in Appendix G). All members of the panel worked independently. Each was presented with a set of the instrument and informed of the purpose of the instrument. They were then requested to study the items and decide on the suitability of the items. They were also asked if any other items should be included to fulfil the purpose of the instrument, and to comment on any part of the scale's items that they felt needed amending or clarification.
  • 11. Example Appendix G: Credentials of Panel for Instrument Validation Name, Qualification and Experience 1. DR. FOO SAY FOOI He is a lecturer in Educational Administration, Faculty of Educational Studies, Universiti Putra Malaysia. 2. ASSOC. PROF. DR. TURIMAN SUANDI A lecturer in the Faculty of Educational Studies, Universiti Putra Malaysia. Among others, he lectures on Research Design. Presently Deputy Dean for Development and Students. 3. ASSOC. PROF. DR. ZAIDATOL AKMALIAH L. PIHIE A lecturer in the field of Educational Administration at the Faculty of Educational Studies, Universiti Putra Malaysia. She had written articles and books related to educational management and administration. She is the Deputy Dean in the FPPUPM.
  • 12. Construct Validity  Establish the meaning of assessment results by controlling the development of the assessment, evaluate the cognitive procession used by students to perform tasks, evaluate the relationships of the scores with other relevant measures, and experimentally determine what factors influence performance.  The extent to which empirical evidence confirms that an inferred construct exists and that a given assessment procedure is measuring the inferred construct accurately
  • 13. Criterion Validity  Compare assessment results with another performance obtained at a later date or with another measure of performance obtained concurrently (for estimating present status).  The degree to which performance on an assessment procedure accurately predicts a student’s performance on an external criterion
  • 14. Factors Influence Validity  Unclear directions  Reading vocabulary and sentence structure too difficult  Ambiguity  Inadequate time limits  Overemphasis of easy-to-assess aspects of domain at the expense of important but difficult-to-assess aspects
  • 15. Factors Influence Validity  Test items inappropriate for the outcomes being measured  Poorly constructed test items  Test too short  Improper arrangement of items  Identifiable pattern of answers  First impression  Personal theories
  • 16. Reliability  Refers to the consistency of measurement, that is, how consistent test scores or other assessment results are from one measurement to another.
  • 17. Nature of Reliability  Refers to the results obtained with an instrument and not to the instrument itself  An estimate of reliability always refers to a particular type of consistency (time, task, students, rater)  Reliability is a necessary but not sufficient condition for validity  Reliability is assessed primarily with statistical indices
  • 18. Terminology  Correlation Coefficient: a statistic that indicates the degree of relationship between any two sets of scores obtained from the same group of individuals
  • 19. Methods of estimating Reliability  Test-retest  Equivalent-forms  Internal Consistency  Split-half  Kuder-Richardson  Alpha Coefficient  Interrater What method?
  • 20. Test-Retest  Give the same test twice to the same group with some time interval between tests, from several minutes to several days
  • 21. Equivalent Forms  Give two forms of the test to the same group in close succession
  • 22. Test-Retest with equivalent forms  Give two forms of the test to the same group with an increased time interval between forms
  • 23. Internal-Consistency Methods There are several internal-consistency methods that require only one administration of an instrument. Split-half Procedure: involves scoring two halves of a test separately for each subject and calculating the correlation coefficient between the two scores. Kuder-Richardson Approaches: (KR20 and KR21) requires 3 pieces of information: Number of items on the test The mean The standard deviation Considered the most frequent method for determining internal consistency Alpha Coefficient: a general form of the KR20 used to calculate the reliability of items that are not scored right vs. wrong.
  • 24. Internal Consistency If Likert scales are used to represent the response choices, analysis for internal consistency can be accomplished using Cronbach’s Alpha. Job Satisfaction Questionnaie No. Component Cronbach Alpha 1 Intrinsic Factors .75 2 Extrinsic Factors .83
  • 25. Interrater  Give a set of student responses requiring judgmental scoring to two or more raters and have them independently score the responses
  • 26. Factors Influencing Reliability Measures  Number of assessment tasks  Spread of scores  Objectivity
  • 27. Validity-Reliability  Reliability is a necessary but insufficient condition for validity.  Reliability (consistency) of measurement is needed to obtain valid results, but we can have reliability without validity.
  • 28. Interactions of Validity and Reliability Lack of validity and reliability Good reliability But poor validity Good validity and reliability
  • 29. Other Factors Influencing Validity and Reliability
  • 30. Objectivity  Refers to the degree to which equally competent scorers obtain the same results.  The test items (objective type)  The results (not influenced by scorer’s judgment or opinion)
  • 31. Practicality  Economical from the viewpoint of both time and money.  Easily administered and scored.  Produce results that can be accurately interpreted and applied by available school personnel
  • 32. Usability  Ease of administration  Time required for administration  Ease of interpretation and application  Availability of equivalent or comparable forms  Cost of testing
  • 33. The Standard Error of Measurement  The index used in educational assessment to describe the consistency of a particular person’s performance(s)  Reflection of the consistency of an individual’s scores  The higher the reliability of a test, the smaller the SE of measurable will be
  • 34. Teachers’ Ethical Responsibilities Regarding Assessment  Make fair and impartial decisions  Construct and administer fair and clear assessments  Motivate pupils to do their best  Teach pupils the varied types of assessments  Provide opportunities for pupils to practice test approaches  Make reasonable accommodations for students with disabilities
  • 35. Summary  Concept of Validity  Factors influencing validity  Concept of Reliability  Factors influencing reliability
  • 36. Questions How important is the concept of ‘validity’ in classroom assessment? To what extent do reliability concerns affect the construction of a good test?
  • 37. References  Linn, R. L., & Miller, M. D. (2005). Measurement and assessment in testing. International Edition. 9th Ed. New Jersey: Pearson Merrill Prentice Hall  Airasian, P.W. (2005). Classroom assessment: Concepts and applications. 5th Ed. Boston: McGrawHill.