READING TEST SPECIFICATIONS 
ASSIGNMENT-01: 4.8 
GROUP MEMBERS: 
BILAL H. YASEEN GS39550 
NURUL NAJWA BINTI ABU BAKAR GS36081 
HANI SHAKIR GS39584 
LECTURER: 
PROF. DR. CHAN SWEE HENG 
Faculty of Modern Languages and Communication 
Universiti Putra Malaysia 
November 2014
READING TEST SPECIFICATIONS: 4.8 
• Introduction: 
How can we help students to learn in an efficient, effective, 
attractive and accessible way? There is no simple straight ward 
answer to this question; depending on the specific situation, 
solution X will work best for person Y. It’s generally acknowledged 
that we can improve testing consider by making the conditions to 
optimal specifications and then use this acknowledge to design new 
test events. 
We are going to introduce the concepts of test design, and examine 
what a test designer must know in order to create high-quality test 
designing, and to discuss the nature of this specifications, how it 
can be modeled in terms of rules, and how the rules are derived. 
We also will use three examples (Designs / Contexts) to introduce 
the different modeling concepts and use the term test designer, to 
describe those who learning design task to perform.
READING TEST SPECIFICATIONS: 4.8 
THE PURPOSE OF THE TEST 
This is a reading comprehension assessment. 
• To measure the testees' performance based on 
the curriculum during teaching and learning 
sessions in the whole school semester. 
• This test is under local control - used to place 
learners into classes for the next school semester 
• To discover how much the students have 
achieved in their L2 learning.
READING TEST SPECIFICATIONS: 4.8 
THE PURPOSE OF THE TEST Cont. 
• This test is low-stakes or summative. 
• However, teachers use the information to 
make decisions about which classes are more 
suitable for the learners to be in for the next 
semester of school, according to their levels of 
performance.
READING TEST SPECIFICATIONS: 4.8 
The Test Takers 
• Grade 4 ESL students of Baghdad Secondary School, Iraq. 
• The assessment is designed to measure the academic 
achievement of all test takers with diverse background of 
Grade 4. 
• Items in the assessment are written in plain language, 
without unnecessarily complex syntactical structure 
because they are adolescent ESL learners with the kind of 
language needs. 
• This assessment is carefully designed to provide 
accommodations for this type of learner’s level; teenage 
ESL students. 
• The passage and items are accessible.
READING TEST SPECIFICATIONS: 4.8 
Context and Modality 
• The Modality: Reading skill (Comprehension). 
• The context: In this test design the tester chooses THREE efficient contexts 
which is: 
• Multiple choice item 2) True and False item. 3) Matching item. 
The Test design: 
Reading comprehension test for measuring the competence and the 
performance of the testees. (Achievement test). 
• ACHIEVEMENT TEST 
Achievement test attempts to measure what an individual has learned-his 
or her present level of performance; in this design the testees will be 
examined for his/her comprehension in the reading skill.
READING TEST SPECIFICATIONS: 4.8 
In planning a test, the Designer can proceed the 
following : 
1. Identify the puspose of the test. 
2. Identify the best specification. 
3. Select the contents of the test. 
4. Consider the form of the test. 
5. Write the test item. 
6. Consider the layout of the test. 
7. Consider the timing of the test. 
8. Plan the scoring of the test.
READING TEST SPECIFICATIONS: 4.8 
1. IDENTIFY THE PURPOSE OF THE TEST 
1.To diagnose a student’s strength, weakness 
and difficulties. 
2.To measure achievement. 
3.To measure aptitude and potential. 
4.To identify readiness for a program.
READING TEST SPECIFICATIONS: 4.8 
2. IDENTIFY THE TEST SPECIFICATION 
The test specification includes : 
1. Which program objectives and student learning 
outcomes will be addressed. 
2. Which content areas will be addressed. 
3. The relative weightings, balance and coverage of items. 
4. The total number of items in the test. 
5. The number of questions required to address a 
particular element of a program or learning outcome 
6. The exact item in the test.
3. SELECT THE CONTENTS OF THE TEST 
Gronlund and Linn (1990), suggest that an item analysis will need to 
consider : 
1. The suitability of the format of each item for the (learning) 
objective (appropriateness) 
2. The ability of each item to enable students to demonstrate their 
performance of the (learning) objective (relevance) 
3. Clarity of the task for each item 
4. The straight forwardness of the task 
5. The independence of each item (i.e. where the influence of other 
items of the tests minimal and where successful completion of one 
another) 
6. The adequacy of coverage of each (learning) objective by the items 
of the test.
READING TEST SPECIFICATIONS: 4.8 
4. CONSIDER THE FORM OF THE TEST 
The designer (tester) will need to consider 
whether the test will be undertaken individually, 
or in a group, and what form it will take. 
Comprehension test, for example, can be 
conducted if the designer feels that reading will 
obstruct the true purpose of the test.
READING TEST SPECIFICATIONS: 4.8 
5. WRITE THE TEST ITEMS 
In this test we tackled three contexts: 
1. Multiple Choice. 
2. True-false Items. 
3. Matching.
A. CONSTRUCTING MULTIPLE CHOICE ITEMS 
1. Design each item to measure a specific objective. 
• (Direct MCQ) Example: 
Q/ Which of these statements correctly summarizes 
how the author of this passage feels about robots? 
a. Robots are old. 
b. Robots are confusing. 
c. Robots are helpful. 
d. Robots are dangerous.
READING TEST SPECIFICATIONS: 4.8 
• (MCQ Indirect question) Example: 
Q/ What is the main purpose of paragraph 2? 
A. to show how easy it is to make a robot. 
B. to tell what a robot is . 
C. to describe the things a robot can do. 
D. to explain the difference between a robot and a machine.
READING TEST SPECIFICATIONS: 4.8 
2. State both stem and options as simply and directly 
as possible. 
We are sometimes tempted to make multiple choice items too 
wordy. A good rule is to get directly to the point. 
( MCQ cloze) Example: 
Q/ Long ago, people ………… robots. 
a. played 
b. saw 
c. imagined 
d. broke
READING TEST SPECIFICATIONS: 4.8 
3. Make certain that intended answer is clearly the only 
correct one. 
Q/ According to the passage, when was the first real robot made? 
a. 1961 
b. 1900 
c. 2003 
d. 2000 years ago
READING TEST SPECIFICATIONS: 4.8 
4. USE ITEM INDICES TO ACCEPT, DISCARD, 
OR REVISE ITEMS 
 Item facility (level of difficulty): Is the extent to which 
an item is easy or difficult for the proposed group of 
test-takers. IF shows how easy or difficult the 
particular item that’s proved in the test. 
 Item discrimination (discriminating power): Item 
discrimination is the extent to which an item 
differentiates between high- and low-ability test-takers. 
It measures how well the test items are 
arranged to identify the differences in the students 
competence.
4. USE ITEM INDICES TO ACCEPT, DISCARD, 
OR REVISE ITEMS. Cont. 
Distractor efficiency: Distractor efficiency is 
one more important measure of multiple 
choice item’s value in a test. The efficiency of 
distractors is the extent to which: - 
a. The distractors “lure” a sufficient number of 
testees especially lower-ability ones 
b. Those responses are somewhat evenly 
distributed across all distractors.
READING TEST SPECIFICATIONS: 4.8 
B. TRUE-FALSE ITEMS 
The true-false item typically present a declarative statement 
that the students must mark as either true or false. Instructors 
generally use true-false items to measure the recall of factual 
knowledge such as name, events, dates, definitions, etc. But 
this format has the potential to measure higher levels of 
cognitive ability, such as comprehension of significant ideas 
and their application in solving problems. 
Examples: 
 A robot is a special kind of machine. (T or F) 
 A robot is a human being. (T or F)
READING TEST SPECIFICATIONS: 4.8 
c. MATCHING 
A matching exercise typically consists of a list 
of questions or problems to be answered along 
with a list of responses. The examinee is 
required to make an association between each 
question and a response.
READING TEST SPECIFICATIONS: 4.8 
EXAMPLE 
Match the following items in list (A) to their responses 
in list (B): 
A B 
1. Something special is NOT A. 1961 
2. Robots may be used to B. Machine 
3. When was the first real robot made? C. Normal 
4. Who is the first one imagined robots? D. Make things 
5. Most robots just look like E. Homer 
F. 2000 
G. Tired
6. CONSIDER THE LAYOUT 
• The Clarity of the Instruction: 
 What to do (Objective) 
 How long to take (Time) 
 How many items to attempt (Quantity) 
 What kind of response is required 
 How and where to enter the response (Answer placing) 
• The Location and Sequence of Items 
 The progression from the easy to the more difficult items of the test. 
• The Visual Layout of the Page 
 Minimize the unnecessary visual material or words.
READING TEST SPECIFICATIONS: 4.8 
7. CONSIDER THE TIMING OF THE TEST 
• The timing refers to two areas: 
1. When the test will take place 
2. The time allowances to be given to the test and 
its component items 
8. PLAN THE SCORING OF THE TEST
SCORING SYSTEM DESIGN AND 
DESCRIPTIONS 
Multiple- 
Choice Item 
True and 
False Item 
Matching Item 
• 4 answer choices (only one 
correct answer). 
• generic scoring rubric, 1 point 
for correct answer or 0 point 
for incorrect (dichotomous). 
• Only 2 options (1 correct and 1 incorrect 
response). 
• Scored dichotomously (choose whether the 
statement is ‘True’ or ‘False’). 
• If they choose the correctly - 1 point 
• If they don’t - 0 point 
• 5 questions and 7 options can be chosen to 
match with the questions. 
• only 1 correct response for every item 
• 1 point – if they match correctly 
• no point - if they match it incorrectly
READING TEST SPECIFICATIONS: 4.8 
Multiple-Choice Item 
SCORE CATEGORY AND DESCRIPTIONS 
CORRECT ANSWER (A, B, C, D OR E) 
These responses represent one correct choice (1 point) 
INCORRECT ANSWER (A, B, C, D OR E) 
These responses represent one incorrect choice (0 point) 
True and False Item 
SCORE CATEGORY AND DESCRIPTIONS 
CORRECT ANSWER (TRUE OR FALSE OPTION) 
These responses represent one correct choice (1 point) 
INCORRECT ANSWER (TRUE OR FALSE OPTION) 
These responses represent one incorrect choice (0 point)
READING TEST SPECIFICATIONS: 4.8 
Matching Item 
SCORE CATEGORY AND DESCRIPTIONS 
CORRECT MATCH 
These responses represent one correct match (1 point) 
INCORRECT MATCH 
These responses represent one incorrect match (0 point)
References 
Alderson, J. C. (1988). Testing English for specific purposes – how specific can we get? In Hughes, A. (ed.), 
Testing English for University Study. London: Modern English Publications and the British Council, 16–18. 
Allman, C. (2005). Building Assessment Initiatives for Schools: An Accommodations Guide for Parents and 
Students with Visual Impairments. American Foundation for the Blind. Available online: 
http://www.afb.org/Section.asp?SectionID=58&TopicID=264&DocumentID=2762. 
Assessment Reform Group (1999). Assessment for Learning: 10 principles. Retrieved from 
http://www.qca.org.uk/qca_4336.aspx , 12 May 2009. 
Bachman, L. F. (2004). Statistical Analyses for Language Assessment. Cambridge: Cambridge University Press. 
Bernstein, J. (1999). Phone Pass TM Testing: Structure and construct. Menlo Park, CA: Ordinate. 
Carroll, J.B. (1961). Fundamental considerations in testing for English language proficiency of foreign students. 
In Allen, H.B. (Ed.), Teaching English as a Second Language. McGraw Hill, New York, pp. 364-372. 
Center for Applied Linguistics. (2007). Foreign Language Assessment Directory. Retrieved 3 July 2009, from 
http://www.cal.org/CALWebDB/FLAD 
Cohen, A. S. and Wollack, J. A. (2006). Test administration, security, scoring and reporting. In Brennan, R. L. 
(ed.), Educational Measurement. 4th edition. New York: American Council on Education/Praeger, 355–386. 
Fulcher, G. (2010). Communicative language testing. In de Bot, K. (ed.), The Mouton Handbook of Teaching 
English as a Foreign Language. The Hague: Mouton.

Reading test specifications assignment-01-ppt

  • 1.
    READING TEST SPECIFICATIONS ASSIGNMENT-01: 4.8 GROUP MEMBERS: BILAL H. YASEEN GS39550 NURUL NAJWA BINTI ABU BAKAR GS36081 HANI SHAKIR GS39584 LECTURER: PROF. DR. CHAN SWEE HENG Faculty of Modern Languages and Communication Universiti Putra Malaysia November 2014
  • 2.
    READING TEST SPECIFICATIONS:4.8 • Introduction: How can we help students to learn in an efficient, effective, attractive and accessible way? There is no simple straight ward answer to this question; depending on the specific situation, solution X will work best for person Y. It’s generally acknowledged that we can improve testing consider by making the conditions to optimal specifications and then use this acknowledge to design new test events. We are going to introduce the concepts of test design, and examine what a test designer must know in order to create high-quality test designing, and to discuss the nature of this specifications, how it can be modeled in terms of rules, and how the rules are derived. We also will use three examples (Designs / Contexts) to introduce the different modeling concepts and use the term test designer, to describe those who learning design task to perform.
  • 3.
    READING TEST SPECIFICATIONS:4.8 THE PURPOSE OF THE TEST This is a reading comprehension assessment. • To measure the testees' performance based on the curriculum during teaching and learning sessions in the whole school semester. • This test is under local control - used to place learners into classes for the next school semester • To discover how much the students have achieved in their L2 learning.
  • 4.
    READING TEST SPECIFICATIONS:4.8 THE PURPOSE OF THE TEST Cont. • This test is low-stakes or summative. • However, teachers use the information to make decisions about which classes are more suitable for the learners to be in for the next semester of school, according to their levels of performance.
  • 5.
    READING TEST SPECIFICATIONS:4.8 The Test Takers • Grade 4 ESL students of Baghdad Secondary School, Iraq. • The assessment is designed to measure the academic achievement of all test takers with diverse background of Grade 4. • Items in the assessment are written in plain language, without unnecessarily complex syntactical structure because they are adolescent ESL learners with the kind of language needs. • This assessment is carefully designed to provide accommodations for this type of learner’s level; teenage ESL students. • The passage and items are accessible.
  • 6.
    READING TEST SPECIFICATIONS:4.8 Context and Modality • The Modality: Reading skill (Comprehension). • The context: In this test design the tester chooses THREE efficient contexts which is: • Multiple choice item 2) True and False item. 3) Matching item. The Test design: Reading comprehension test for measuring the competence and the performance of the testees. (Achievement test). • ACHIEVEMENT TEST Achievement test attempts to measure what an individual has learned-his or her present level of performance; in this design the testees will be examined for his/her comprehension in the reading skill.
  • 7.
    READING TEST SPECIFICATIONS:4.8 In planning a test, the Designer can proceed the following : 1. Identify the puspose of the test. 2. Identify the best specification. 3. Select the contents of the test. 4. Consider the form of the test. 5. Write the test item. 6. Consider the layout of the test. 7. Consider the timing of the test. 8. Plan the scoring of the test.
  • 8.
    READING TEST SPECIFICATIONS:4.8 1. IDENTIFY THE PURPOSE OF THE TEST 1.To diagnose a student’s strength, weakness and difficulties. 2.To measure achievement. 3.To measure aptitude and potential. 4.To identify readiness for a program.
  • 9.
    READING TEST SPECIFICATIONS:4.8 2. IDENTIFY THE TEST SPECIFICATION The test specification includes : 1. Which program objectives and student learning outcomes will be addressed. 2. Which content areas will be addressed. 3. The relative weightings, balance and coverage of items. 4. The total number of items in the test. 5. The number of questions required to address a particular element of a program or learning outcome 6. The exact item in the test.
  • 10.
    3. SELECT THECONTENTS OF THE TEST Gronlund and Linn (1990), suggest that an item analysis will need to consider : 1. The suitability of the format of each item for the (learning) objective (appropriateness) 2. The ability of each item to enable students to demonstrate their performance of the (learning) objective (relevance) 3. Clarity of the task for each item 4. The straight forwardness of the task 5. The independence of each item (i.e. where the influence of other items of the tests minimal and where successful completion of one another) 6. The adequacy of coverage of each (learning) objective by the items of the test.
  • 11.
    READING TEST SPECIFICATIONS:4.8 4. CONSIDER THE FORM OF THE TEST The designer (tester) will need to consider whether the test will be undertaken individually, or in a group, and what form it will take. Comprehension test, for example, can be conducted if the designer feels that reading will obstruct the true purpose of the test.
  • 12.
    READING TEST SPECIFICATIONS:4.8 5. WRITE THE TEST ITEMS In this test we tackled three contexts: 1. Multiple Choice. 2. True-false Items. 3. Matching.
  • 13.
    A. CONSTRUCTING MULTIPLECHOICE ITEMS 1. Design each item to measure a specific objective. • (Direct MCQ) Example: Q/ Which of these statements correctly summarizes how the author of this passage feels about robots? a. Robots are old. b. Robots are confusing. c. Robots are helpful. d. Robots are dangerous.
  • 14.
    READING TEST SPECIFICATIONS:4.8 • (MCQ Indirect question) Example: Q/ What is the main purpose of paragraph 2? A. to show how easy it is to make a robot. B. to tell what a robot is . C. to describe the things a robot can do. D. to explain the difference between a robot and a machine.
  • 15.
    READING TEST SPECIFICATIONS:4.8 2. State both stem and options as simply and directly as possible. We are sometimes tempted to make multiple choice items too wordy. A good rule is to get directly to the point. ( MCQ cloze) Example: Q/ Long ago, people ………… robots. a. played b. saw c. imagined d. broke
  • 16.
    READING TEST SPECIFICATIONS:4.8 3. Make certain that intended answer is clearly the only correct one. Q/ According to the passage, when was the first real robot made? a. 1961 b. 1900 c. 2003 d. 2000 years ago
  • 17.
    READING TEST SPECIFICATIONS:4.8 4. USE ITEM INDICES TO ACCEPT, DISCARD, OR REVISE ITEMS  Item facility (level of difficulty): Is the extent to which an item is easy or difficult for the proposed group of test-takers. IF shows how easy or difficult the particular item that’s proved in the test.  Item discrimination (discriminating power): Item discrimination is the extent to which an item differentiates between high- and low-ability test-takers. It measures how well the test items are arranged to identify the differences in the students competence.
  • 18.
    4. USE ITEMINDICES TO ACCEPT, DISCARD, OR REVISE ITEMS. Cont. Distractor efficiency: Distractor efficiency is one more important measure of multiple choice item’s value in a test. The efficiency of distractors is the extent to which: - a. The distractors “lure” a sufficient number of testees especially lower-ability ones b. Those responses are somewhat evenly distributed across all distractors.
  • 19.
    READING TEST SPECIFICATIONS:4.8 B. TRUE-FALSE ITEMS The true-false item typically present a declarative statement that the students must mark as either true or false. Instructors generally use true-false items to measure the recall of factual knowledge such as name, events, dates, definitions, etc. But this format has the potential to measure higher levels of cognitive ability, such as comprehension of significant ideas and their application in solving problems. Examples:  A robot is a special kind of machine. (T or F)  A robot is a human being. (T or F)
  • 20.
    READING TEST SPECIFICATIONS:4.8 c. MATCHING A matching exercise typically consists of a list of questions or problems to be answered along with a list of responses. The examinee is required to make an association between each question and a response.
  • 21.
    READING TEST SPECIFICATIONS:4.8 EXAMPLE Match the following items in list (A) to their responses in list (B): A B 1. Something special is NOT A. 1961 2. Robots may be used to B. Machine 3. When was the first real robot made? C. Normal 4. Who is the first one imagined robots? D. Make things 5. Most robots just look like E. Homer F. 2000 G. Tired
  • 22.
    6. CONSIDER THELAYOUT • The Clarity of the Instruction:  What to do (Objective)  How long to take (Time)  How many items to attempt (Quantity)  What kind of response is required  How and where to enter the response (Answer placing) • The Location and Sequence of Items  The progression from the easy to the more difficult items of the test. • The Visual Layout of the Page  Minimize the unnecessary visual material or words.
  • 23.
    READING TEST SPECIFICATIONS:4.8 7. CONSIDER THE TIMING OF THE TEST • The timing refers to two areas: 1. When the test will take place 2. The time allowances to be given to the test and its component items 8. PLAN THE SCORING OF THE TEST
  • 24.
    SCORING SYSTEM DESIGNAND DESCRIPTIONS Multiple- Choice Item True and False Item Matching Item • 4 answer choices (only one correct answer). • generic scoring rubric, 1 point for correct answer or 0 point for incorrect (dichotomous). • Only 2 options (1 correct and 1 incorrect response). • Scored dichotomously (choose whether the statement is ‘True’ or ‘False’). • If they choose the correctly - 1 point • If they don’t - 0 point • 5 questions and 7 options can be chosen to match with the questions. • only 1 correct response for every item • 1 point – if they match correctly • no point - if they match it incorrectly
  • 25.
    READING TEST SPECIFICATIONS:4.8 Multiple-Choice Item SCORE CATEGORY AND DESCRIPTIONS CORRECT ANSWER (A, B, C, D OR E) These responses represent one correct choice (1 point) INCORRECT ANSWER (A, B, C, D OR E) These responses represent one incorrect choice (0 point) True and False Item SCORE CATEGORY AND DESCRIPTIONS CORRECT ANSWER (TRUE OR FALSE OPTION) These responses represent one correct choice (1 point) INCORRECT ANSWER (TRUE OR FALSE OPTION) These responses represent one incorrect choice (0 point)
  • 26.
    READING TEST SPECIFICATIONS:4.8 Matching Item SCORE CATEGORY AND DESCRIPTIONS CORRECT MATCH These responses represent one correct match (1 point) INCORRECT MATCH These responses represent one incorrect match (0 point)
  • 27.
    References Alderson, J.C. (1988). Testing English for specific purposes – how specific can we get? In Hughes, A. (ed.), Testing English for University Study. London: Modern English Publications and the British Council, 16–18. Allman, C. (2005). Building Assessment Initiatives for Schools: An Accommodations Guide for Parents and Students with Visual Impairments. American Foundation for the Blind. Available online: http://www.afb.org/Section.asp?SectionID=58&TopicID=264&DocumentID=2762. Assessment Reform Group (1999). Assessment for Learning: 10 principles. Retrieved from http://www.qca.org.uk/qca_4336.aspx , 12 May 2009. Bachman, L. F. (2004). Statistical Analyses for Language Assessment. Cambridge: Cambridge University Press. Bernstein, J. (1999). Phone Pass TM Testing: Structure and construct. Menlo Park, CA: Ordinate. Carroll, J.B. (1961). Fundamental considerations in testing for English language proficiency of foreign students. In Allen, H.B. (Ed.), Teaching English as a Second Language. McGraw Hill, New York, pp. 364-372. Center for Applied Linguistics. (2007). Foreign Language Assessment Directory. Retrieved 3 July 2009, from http://www.cal.org/CALWebDB/FLAD Cohen, A. S. and Wollack, J. A. (2006). Test administration, security, scoring and reporting. In Brennan, R. L. (ed.), Educational Measurement. 4th edition. New York: American Council on Education/Praeger, 355–386. Fulcher, G. (2010). Communicative language testing. In de Bot, K. (ed.), The Mouton Handbook of Teaching English as a Foreign Language. The Hague: Mouton.