1. www.ou.edu.vn
Click to edit Master subtitle style
HCMC OPEN UNIVERSITY GRADUATE SCHOOL
MINISTRY OF EDUCATION AND TRAINING
HO CHI MINH CITY OPEN UNIVERSITY
TEST CONSIDERATION: REALIABILITY &
VALIDITY
Presenters : Group 5 Lý Tuấn Phú
Đặng Kiều Anh
Nguyễn Duy Cường
Nguyễn Thị Kim Loan
Mai Xuân Ái
Trần Thị Kim Ngân
July, Hochiminh City, 2013
2. I. Reliability
- Introduction
- Factors that affect language test scores
- Classical true score measurement theory
- Generalizability theory
- Standard error of measurement: interpreting individual test scores
within classical true score and generalizability theory
- Item response theory
- Reliability of criterion-referenced test scores
- Factors that affect reliability estimates
- Systematic measurement error
II. Validation/ Validity
- Introduction
- Reliability and validity revisited
- Validity as a unitary concept
- The evidential basis of validity
- Test bias
- The consequential or ethical basis of validity
- Postmortem: face validity
2
3. Relationship between reliability & validity:
complementary aspects
(1) (Reliability) to minimize the effects of
measurement error, and
(2) (Validity) to maximize the effects of the language
abilities we want to measure.
Investigation of reliability: <= we must identify
sources of error and estimate the degree of their
effects on test scores <= distinguishing the effects
of the language abilities we want to measure from
the effects of other factors: Complex problem!
5. Different factors will affect different individuals
differently.
Designing and developing language tests: to
minimize their effects on test performance :
◦ test method,
◦ random factors
◦ personal attributes: Sources of test bias (test invalidity)
Sources of measurement error
• ‘Mean’ ( 𝑥): the average of the scores of a
given group of test takers.
• ‘Variance’ (𝑠2
): how much individual scores
vary from the group mean.
6. Classical true score (CTS) measurement theory
consists of a set of assumptions about the
relationships between actual, or observed test scores
and the factors that affect these scores.
Concept 1: True score and error score
1. An observed score on a test comprises 2 factors: a
true score (an individual’s level of ability) & an
error score (factors other than the ability being
tested).
2. The relationship between true and error scores:
error scores are unsystematic, or random, and are
uncorrelated with true scores.
7. Concept 2: Parallel tests
Two tests are parallel if, for every group of
persons taking both tests, (1) the true score on one
test is equal to the true score on the other, and (2)
the error variances for the two tests are equal.
8. 3. Reliability of observed scores:
a. Reliability as the correlation between parallel tests:
If the observed scores on two parallel tests are
highly correlated, this indicates that effects of the error
scores are minimal, and that they can be considered
reliable indicators of the ability being measured.
b. Reliability and measurement error as proportions of
observed score variance
If an individual's observed score on a test is
composed of a true score and an error score, the greater
the proportion of true score, the less the proportion of
error score, and thus the more reliable the observed
score.
9. 3 approaches to estimating reliability: (p.173 – p.
185)
1. Internal consistency estimates: are concerned
primarily with sources of error from within the test
and scoring procedures.
2. Stability estimates indicate how consistent test
scores are over time.
3. Equivalence estimates provide an indication of the
extent to which scores on alternate forms of a test
are equivalent.
10. Problems with the classical true score model:
+ The CTS model treats error variance as
homogeneous in origin.
+ The CTS model considers all error to be random,
and consequently fails to distinguish systematic error
from random error.
https://www.youtube.com/watch?v=CSI-1Zk6oeM
https://www.youtube.com/watch?v=k84ksLUWKKc
11. Constitutes a theory and set of procedures
for specifying and estimating the relative
effects of different factors on observed test
scores
=> Provide a means for relating the uses or
interpretations of test scores to the way test
users specify and interpret dif. factors, or
sources of errors.
12. A given measure or score is treated as a
sample from a hypothetical universe of
possible measures.
Interpreting a test score = generalizing from
a single measure to a universe of measures.
(on the basis of an individual’s performance
on a test => generalize to her performance in
other contexts).
The more reliable the sample of performance,
or test score is, the more generalizable it is.
14. The application of G-theory to test
development and use:
generalizability study (‘G-study’)
decision study (‘D-study’)
specify the dif. sources of variance,
estimate the relative importance of these dif.
sources simultaneously,
employ these estimates in the interpretation
and use of test scores.
15. Universe of generalization is a domain of
uses or abilities to which we want test scores
to generalize.
Universe of measures are types of test scores
we would be willing to accept as indicators of
the ability to be measured for the purpose
intended.
16. Are whom we are going to make decisions or
inferences.
The degree of generalizability determines the
way we define the population.
Ex: using test results for making decisions
about 1 group => this group is population of
persons.
Using a test with more than one group
(entrance or placement tests) => generalizing
beyond a particular group.
17. If we could obtain measures for an individual
under all the different conditions specified in
the universe of possible measures, his
average score on these measures might be
considered the best indicator of his ability.
=> is defined as the mean of a person’s scores
on all measures from the universe of possible
measures.
18. The standard error of measurement is the
indicator of how much we would expect an
individual’s test scores to vary, given a
particular level of reliability.
When investigating the amount of
measurement error in individual test scores,
we are looking at differences b/w test takers’
obtained scores and their true scores.
19. The error score is the difference between an
obtained score and the true score.
The more reliable the test is, the closer the
obtained scores will cluster around the true score
mean => smaller standard deviation of errors.
The less reliable the test, the greater the
standard deviation.
Because of the importance in the interpretation
of test scores, the standard deviation of the error
scores has a name: the standard error of
measurement (SEM).
20. SEM provides a means for applying estimates
of reliability to the interpretation and use of
individuals’ observed test scores
its primary advantage: makes test users
aware of how much variability in observed
scores to expect as a result of measurement
error.
21.
22. Norm-referenced (NR) test scores:
maximize inter-individual score
differences or score variance
Criterion-referenced (CR) test scores:
- provide information about an
individual’s relative ‘mastery’ of an
ability domain
- develop to be representative of the
criterion ability
- occur in educational programs and
language classrooms
- commonly use achievement tests
28. The examination of validity : examining the validity
of a given use of test scores is a complex process that
must involve the examination of both the evidence that
supports that interpretation or use and the ethical values
that provide the basis or justification for that
interpretation or use ( Messick 1975,1980, 1989).
In test validation we are not examining the validity of
the test content or of even the test scores themselves,
but rather the validity of the way we interpret or use
the information gathered through the testing
procedure.
Validity is not simply a function of the content and
procedure of the test itself, it must consider how test
takers perform also.
29. Reliability is a requirement for validity
The investigation of reliability and validity can be viewed as
complementary aspects of identifying, estimating, & interpreting
different sources of variance in test scores.
The investigation of reliability is concerned with answering the
question : How much variance in test scores is due to
measurement error?, How much variance is due to factors other
than measurement error?
Validity is concerned with identifying the factors that produce
the reliable variance in test scores,
The question addressed : what specific abilities account for the
reliable variance in test score ?
II. 2. Reliability and validity revisited
The relationship b/w reliability & validity
30. Definition of validation.
The relationship between reliability and validity, viewing the
estimation of reliability as an essential requisite of validation.
The framework proposed by Messick (1989) for considering
validity as a unitary through multifaceted concept.
The evidential basis for validity
Construct validity ( includes content relevance, criterion
relatedness)
Test bias ( including culture, test content, personality characteristics
of test takers, sex, age).
The ethical, or consequential basis of test use.
II. Validation/ Validity
31. Another way to distinguish reliability from validity : to
consider the theoretical frameworks upon which
they depends.
In estimating reliability we are concerned primarily with
examining variance in test scores themselves.
In validity, we must consider other sources of variance, and
utilize the theory of abilities that we hypothesize will affect test
performance.
The process if validation must look beyond reliability and
examine the relationship b/w test performance and factors
outside the test itself.
The relationship b/w reliability & validity (cont)
32. However, the distinguishing of reliability & validity is
still not clear , due to :
Different test methods from each other
Abilities from test methods
The relationship b/w reliability & validity (cont)
33. The classic statement of the relationship b/w reliability
& validity by Campbell and Fiske (1959) :
Agreement b/w similar
Measures of the same trait
( for example, correlation b/w
scores on parallel tests)
Agreement b/w different
measures of the same trait
(for example, correlation
b/w scores on a multiple
Choice test of grammar
& ratings of grammar on
An oral interview)
The relationship b/w reliability & validity (cont)
34. In many cases, the distinctiveness of the test methods is
not so clear.
->We must carefully consider not only similarities in the
test content, but also similarities in the test methods, in
order to determine whether correlations b/w tests should
be interpreted as estimators of reliability or as evidence
supporting validity.
Language testing has a very special and complex problem
when its comes to traits and methods-> It’s difficult for
language test to distinguish traits and methods.
35. Source of
Justification
Evidential
Basis
Consequential
Basis
Function of outcome
of testing
Test
interpretation Test use
Construct
validity
Construct validity+
Value implications
Construct validity+
Relevance/
Utility
Construct validity+
Relevance/
Utility+
Social
consequences
II. 2. Validity as a unitary concept
36.
37. content relevance requires ‘the specification
of the behavioral domain in question and
the attendant specification of the task or
test domain’ (Messick 1980: 1017).
content coverage, or the extent to which the
tasks required in the test adequately
represent the behavioral domain in question.
38. The examination of the
content relevance and
content coverage is a
necessary part of the validity
process
39. may be level of ability as defined by group
membership, individuals’ performance on
another test of the ability in question, or their
relative success in performing some task that
involves this ability.
40. Information on concurrent criterion relatedness is
undoubtedly the most commonly used in
language testing.
There are 2 forms:
(1) examining differences in test performance
among groups of individuals at different levels of
language ability.
(2) examining correlations among various
measures of a given ability.
41. need to collect data a relationship between
scores on the test and job or course
performance
can largely ignore the question of what abilities
are being measured
42. Construct validity is indeed the unifying concept
integrates criterion and content considerations
into a common framework for testing rational
hypotheses about theoretically relevant
relationships.
(Messick 1980: 1015)
Construct validation requires both logical
analysis and empirical investigation.
43. the test developer involved in the process of construct
validation is likely to collect several types of empirical
evidence. These may include any or all of the following:
(1) the examination of patterns of correlations among
item scores and test scores, and between
characteristics of and tests and scores on items and
tests;
(2) analyses and modeling of the processes underlying
test performance;
(3) studies of group differences;
(4) studies of changes over time
(5) investigation of the effects of experimental
treatment (Messick 1989)
44. Correlational evidence is derived €ram a
family of statistical procedures that examine
the relationships among variables, or
measures.
A correlation is a functional relationship
between two measures.
Correlational approaches to construct
validation may utilize both
exploratory and confirmatory modes.
45. It is impossible to make clear, unambiguous
inferences regarding the influence of various
factors on test scores on the basis of a single
correlation between two tests.
46. A commonly used procedure for interpreting a
large number of correlations is factor analysis
47. Characteristic: each measure is considered to
be a combination of trait and method, and
tests are included in the design so as to
combine multiple traits with multiple
methods.
Advantage: permits the investigator to
examine patterns of both convergence and
discrimination among correlations.
Convergence is essentially what
48. Analysis of data: many ways
(1) the direct inspection of convergent and
discriminant correlations
(2) the analysis of variance
(3) confirmatory factor analysis
49. individuals are assigned .at random to two or more
groups, each of which is given a different treatment. At
the end of the treatment, observations are made to
investigate differences among the different groups.
There are two distinguishing characteristics of a true
experimental design. The first is that of randomization,
which means that (1) a sample of subjects is randomly
selected from a population, and (2) the individuals in
this random sample are then randomly assigned to two
or more groups for comparison.
The second characteristic is that of experimental
intervention, or treatment. This means that the
different groups of subjects are exposed to distinct
treatments, or sets of circumstances, as part of the
experiment.
50. the process of construct validation is a complex
and continuous undertaking, involving both ( 1)
theoretical, logical analysis leading to empirically
testable hypotheses, and (2) a variety of
appropriate approaches to empirical observation
and analysis
The result of this process of construct
validation will be a statement regarding the extent
to which the test under consideration provides a
valid basis for making inferences about the given
ability with respect to the types of individuals and
contexts that have provided the setting for the
validation research.
51. 51
What is test bias?
Systematic differences in test performance ,
resulted by the differences in individual
characteristics
Examples: Gender Difference in Mathematical
Ability
A reliable mathematics test to a representative
groups of males and females.
On average, males have higher scores than
females
=> Tendency to interpret that: “males have
greater mathematical ability than female”
52. 52
However, test score should not be interpreted to reflect
purely mathematical ability.
The differences b/w test scores due to test score bias,
NOT due to differences in true mathematical abilitt
=> Differences in group performance do not indicate
test bias.
=> The systematic differences which are not logically
related to the ability in the questions/ tests => test is
biased
54. 54
+ Cultural background
Cultural differences (Britre (1968, 1973: Britre and Brown. 1971)
The problem of cultural content ((Plaister (1967) and Condon
(1975))
In item response theory, some items in multiple - choice
vocabulary are in favor of one linguistic and cultural subgroups
(Chen and Henning (1985))
Aptitude tests: possibly biased toward culturally different
groups (Zeidner (1986))
+ Background knowledge
Prior knowledge affects test performance (Chacevycn et al.
(1982))
In ESP testing, students' performance: affected as much by their
prior knowledge as by their language proficiency
+ Cognitive characteristics
Cognitive factors influence language acquisition (Brown 1987)
55. 55
Cognitive styles/ learning styles:
+ field- dependent/ independent
a field-independent learning style is defined
by a tendency to separate details from the
surrounding context ( cited from
http://www.teachingenglish.org.uk/knowledge-
database/field-independent-learners)
a field-dependent learning style, which is
defined by a relative inability to distinguish
detail from other information around it
56. 56
Example
Field-independent learners tend to rely less on the
teacher or other learners for support.
=> Psychological differences
Ambiguity tolerance/ intolerance : cognitive
flexibility
Tolerance of ambiguity: one's acceptance of
confusing situation and a lack of clear r line
demarcation (Ely (1989)),
One facet of personality characteristics : related to
risk taking . Those who can tolerate ambiguity are
more likely to take risks in language learning, an
essential of making progress on the language
acquisition
(As cited in Grace ,1997)
57. 57
Test: serve the need of an educational system or of
society
The use of language tests reflect in microcosm the
role of test in general as instrument of social policy
The role of tests can be described via kinds of tests
+ placement
+ diagnosis
+ selection (based in the proficiency/ achievement )
+ evaluation
+ making decisions
The issues involved in the ethics of tests:
+ numerous
+ vary across societies, cultures, testing contexts
58. 58
=> focus on the rights of individual test takers
:
+ secrecy
+ access to information
+ privacy
+ confidentiality
+ consent
+ the balance b/w individual rights and the
values of the society
59. 59
As test developers and test users, people need to
consider:
+ the rights & interests of test takers
+ the responsibilities of institutions for
making decisions based on tests
+ public interest
These considerations are political, dynamic, and
vary across societies
These considerations have implications for the
practice of teachers' profession, kinds of tests to
be developed an the ways in which test usefulness
is justified .
60. 60
We must move out of the comfortable combines of
applies linguistic and psychometric theory into the
arena of public policy.
Hulin et at. (1983)
"it is important to realize that testing and social policy
a=cannot be totally separated and that questions about
the use of tests can not be addressed without
considering existing social forces, whatever they are (p.
285)
4 areas of considerations in the ethical use and
interpretation of test results (Messick (1980, 1988b)
+ construct validity/ the evidence supports the
interpretation of test scores
+ value systems that inform test use
+ practical usefulness of the test
+ the consequences to the educational system or
society of using test results for a particular purpose
61. 61
In short , complete evidence should be
provided:
+ to prove that tests are used as
valid indicators of the abilities which
are appropriate to the intended use
+ to determine the use of test
62. 62
Test validity: the appeal or appearance of a
test
Measure what it is supposed to measure.
Test appearance has a considerable effect on
the acceptability of tests to both test takers
and test users.
Test talkers will take the test seriously
enough to try the best or not. Accept/ not
accept the test. Test is useful or not.
=> test takers' reaction influent the validity
and reliability of tests.