This document discusses assessment literacy for language teachers. It explains that assessment literacy refers to the knowledge and skills teachers need regarding assessment, evaluation, and testing. It is important for setting global standards. The document outlines some of the issues with over-reliance on numerical test scores and argues that teachers need to understand different assessment methods and choose those appropriate for their classroom goals and situations. It also discusses criterion-referenced assessment and the increasing focus on performance-based evaluation in language teaching.
2. The Role of Assessment in the
Language Classroom
ELT Career & Professional
Development
May 27, 2012
Kahoko Matsumoto
Tokai University
3. 3
Contents
1. Testing from a Japanese Perspective
3. Why Criterion-referenced Assessment
Now?
4. Increasing Importance of Performance
Assessment
2. What Language Teachers Should Know
(Assessment Literacy)
5. What Language Teachers Should Know
(Revisited)
4. 1. Testing from a Japanese
Perspective
<The changes in the view on competence>
1.Divisible Competence (- 1950’s)
Audio-lingual/Grammar-translation period. Competence was
measured separately (discrete-point test).
2.Unitary Competence (1960’s – 1980’s)
Early days of communicative approach. The existence of
“general (universal) competence” was assumed.
(objective, data-based integrative tests→standardized test)
3. Multidimensional Competence (1990’s -)
More attention to performance (output) and its assessment.
There are various degrees of relationship among different
abilities. (skills-integrated tests, project/process-based
assessment, alternative assessment, etc.)
6. <General theories of competence>
Canale & Swain (1980)
Grammatical, Discoursal, Sociolinguistic, Strategic
Backman (1990)
Organizational(grammatical/textual) and
Pragmatic(illocutional/sociolinguistic)
*How much of these functional, real-life communicative
abilities can be measured by tests?
* How much of listening, reading, writing and speaking
abilities can be measured by grammar and vocabulary
test items?
* What do notorious “translation” items measure?
7. 7
<The danger of quantified test
results>
7
20 30 50 80 9060
10 points 10 points 10 points
Do these 10 points represent the same ability difference?
➾No! Also a couple of points are always within the standard
error of measurement
8. 8
<Validity and reliability>
Listening Ability
Items in Listening
Test A
Items in Listening
Test B
Both tests can be highly reliable,
but they measure quite different
constructs (factors) of listening.
→In classroom achievement/
diagnostic tests, validity (whether
the test measures what has been
taught) is more important!
9. 9
2. What Language Teachers Should
Know (Assessment Literacy)
1) Taylor (2009) claims that the term assessment
literacy encompasses what language teachers
need to know about assessment matters. This
consists of the level of skill, knowledge and
understanding of assessment principles necessary
to maintain good, effective practice of teaching.
Also teachers have to make effort for this to be
shared by other test stakeholder groups including
government officials, policy planners, the media
and the general public.
10. 10
2) Inbar-Lourie (2008) says that Language
Assessment Literacy involves additional
competencies such as language-specific
techniques in addition to general assessment
literacy skills. Her idea of the Language
Assessment Literacy (LAL) competencies
for educational purposes can be broken up into
several subsections; ‘why, what and how.’
11. 11
‘Why’ refers to the rationale behind the testing. It looks to
analyze assessment culture and the concepts of validity and
ethics that go along with it. The author wishes to stress the
vital role of language assessment in making crucial decisions
in other areas of the learning spectrum.
‘What’ deals with the current theories regarding assessment
and the validity and reliability of the same. In relation to
language assessment, this relates to contemporary debates
on issues such as the norms of English as an International
Language.
‘How’ can look at test construction or development or the
role of assessment in a language curriculum.
12. 12
<Why: purpose of a test>
1) to screen or stream students
→norm-referenced proficiency tests
(reliability>validity)
2) to make diagnosis of students’ improvements
and provide informed support
→ criterion-referenced diagnostic/achievement
tests (validity>reliability)
13. 13
<What: goals of learning/assessment>
●Different kinds of validity:
- Content validity
- Concurrent validity
- Criterion validity
- Construct validity
- Predictive validity
●Nature of assessment
- summative vs. formative
- internal vs. external
- self-reflective assessment vs. rater assessment
*avoiding subjective/objective dichotomy
14. 14
<How: development of a solid rubric>
1) Content/ability to cover→validity
2) Assessment criteria
3) Nature of task
4) Students’ ability range→difficulty level
5) Weighting among test items (global vs. local
abilities)
15. 15
<How: choosing appropriate types
of test
1) test of receptive skills
- multiple-choice test
- cloze test
- essay test, etc.
2) performance test
- recitation/translation
- short response to a stimulus/question
- summary creation
- expression of an opinion
- free writing/speaking in response to a prompt
* individual vs. interactive tasks
3) integrated skills test
- controlled
- project/process-based assessment
16. 3. Why Criterion-based
Assessment Now?
1) Established descriptive (explanatory) criteria help
increased accountability and transparency with added
validity
ex. Can-do Statements (CDSs)
= functional, qualitative statements of what learners
can really do in various communicative situations
2) Increasing use of Common European Framework of
Reference for Languages: Learning, Teaching and
Assessment=CEFR)
* verifiable by both quantitative and qualitative
measures
21. 21
<Statistical vs. qualitative>
- Validity is sometimes
questionable.
- Test scores don’t always
represent real-life
abilities.
- The consistency among
objectives, teaching and
assessment is secured.
- Sometimes hard to verify
the results.
Statistical Approach
(IRT, etc.)
Qualitative
Approach
* We can be benefitted from both approaches, if we know
which to use for different types of assessment.
24. 4. Increased Importance of
Performance Assessment
1) Increased importance placed on English as a
lingua franca and also on worldwide internet
communication by writing (Warschauer, 2000;
Warschauer & Ware, 2006)
2) Rapid globalization and increased use of
online verbal communication tools (video-
conferencing, free learning websites,
webinars, etc.)
27. 27
5. What Language Teachers Should
Know (Revisited)
1) Taylor (2009) claims that the term assessment
literacy encompasses what language teachers
need to know about assessment matters. This
consists of the level of skill, knowledge and
understanding of assessment principles necessary
to maintain good, effective practice of teaching.
Also teachers have to make effort for this to be
shared by other test stakeholder groups including
government officials, policy planners, the media
and the general public.
28. 28
2) Inbar-Lourie (2008) says that Language
Assessment Literacy involves additional
competencies such as language-specific
techniques in addition to general assessment
literacy skills. Her idea of the Language
Assessment Literacy (LAL) competencies
for educational purposes can be broken up into
several subsections; ‘why, what and how.’
29. 29
‘Why’ refers to the rationale behind the testing. It looks to
analyze assessment culture and the concepts of validity and
ethics that go along with it. The author wishes to stress the
vital role of language assessment in making crucial decisions
in other areas of the learning spectrum.
‘What’ deals with the current theories regarding assessment
and the validity and reliability of the same. In relation to
language assessment, this relates to contemporary debates
on issues such as the norms of English as an International
Language.
‘How’ can look at test construction or development or the
role of assessment in a language curriculum.
30. Reference
Backman, L.F. (1990). Fundamental Considerations in Language Testing.
Oxford: Oxford University Press.
Canale, M and M. Swain. (1980). Theoretical bases of communicative approaches to
second language teaching and testing. Applied Linguistics 1, 1-47.
Cummins, J. (1979) Cognitive/academic language proficiency, linguistic
interdependence, the optimum age question and some other matters. Working
Papers on Bilingualism, No. 19, 121-129.
Demauro, G.(1992). Examination of the relationships among TSE, TWE, and TOEFL
scores. Language Testing, 9(2), 149-161.
Inbar-Lourie. O. (2008). Constructing a language assessment knowledge base:
A focus on language assessment courses. Language Testing, 25(3), 385-402.
Matsumoto, K. (2011). Studies on the Correlations of Listening Ability and
Productive Abilities. In JACET Testing SIG (eds.) Studies on L2 Listening:
Teaching and Assessing. 69-84.
Shojima, K. (2007) Japanese Journal for Research on Testing, 3, 161-178. (in
Japanese with English abstract). Scholastic achievement structure of National
Center Test 2006 by self-organizing map.
Taylor, L. (2009). Developing assessment literacy. Annual Review of Applied
Linguistics. 29, 21-36.
Warschauer, M. (2000). The changing global economy and the future of the
English teaching, TESOL Quarterly, 32, 511-535.
Warschauer, M. , & Ware, P. (2006). Automated writing evaluation: Defining the
classroom research agenda, Language Teaching Research, 10(2), 219-233. 30
31. Thank you for listening!
For questions and comments:
mkahoko@tsc.u-tokai.ac.jp
31