Assessment principles


Published on

  • Be the first to comment

  • Be the first to like this

No Downloads
Total views
On SlideShare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide

Assessment principles

  1. 1. Assessment Principles Carlo Magno, PhDCounseling and Educational Psychology Department De La Salle University, Manila
  2. 2. ASSESSMENT COMPETENCIES FOR TEACHERS • Constructed by the AFT, NCME, NEA: • Teachers should be skilled in: 1.choosing assessment methods appropriate for instructional decisions. 2.Administering, scoring, and interpreting the results of both externally produced and teacher produced assessment methods. 3.Using assessment results when making decisions about individual students, planning teaching, and developing curriculum and school improvement.American Federation of Teachers, National Council on Measurement andEvaluation, and National Education Association in the United States of America.
  3. 3. ASSESSMENT COMPETENCIES FOR TEACHERS4. Developing valid pupil grading procedures that use pupil assessment.5. Communicating assessment results to students, parents, other lay audiences, and other educators.6. Recognizing unethical, illegal, and otherwise inappropriate assessment methods and uses of assessment information.
  4. 4. SHIFTS IN ASSESSMENT• Testing Alternative assessment• Paper and pencil Performance assessment• Multiple choice Supply• Single correct answer Many correct answer• Summative Formative• Outcome only Process and Outcome• Skill focused Task-based• Isolated facts Application of knowledge• Decontextualized task Contextualized task
  5. 5. Assessment Literacy• (1) Assessment comes with a clear purpose• (2) focusing on achievement targets• (3) selecting proper assessment methods• (4) sampling student achievement
  6. 6. ALTERNATIVE FORMS OF ASSESSMENT• Performance based assessment• Authentic assessment• Portfolio assessment
  7. 7. OBJECTIVES• 1. Distinguish performance- based assessment with the traditional paper and pencil tests.• 2. Construct tasks that are performance based.• Design a rubric to assess a performance based task
  8. 8. TERMS• Authentic assessment• Direct assessment• Alternative assessment• Performance testing• Performance assessment• Changes are taking place in assessment
  9. 9. METHOD• Assessment should measure what is really important in the curriculum.• Assessment should look more like instructional activities than like tests.• Educational assessment should approximate the learning tasks of interest, so that, when students practice for the assessment, some useful learning takes place.
  10. 10. WHAT IS PERFORMANCE ASSESSMENT?• Testing that requires a student to create an answer or a product that demonstrates his/her knowledge or skills (Rudner & Boston, 1991).
  11. 11. FEATURES OF PERFORMANCE ASSESSMENT• Intended to assess what it is that students know and can do with the emphasis on doing.• Have a high degree of realism about them.• Involve: (a) activities for which there is no correct answer, (b) assessing groups rather than individuals, (c) testing that would continue over an extended period of time, (d) self-evaluation of performances.• Likely use open-ended tasks aimed at assessing higher level cognitive skills.
  12. 12. PUSH ON PERFORMANCE ASSESSMENT• Bring testing methods more in line with instruction.• Assessment should approximate closely what it is students should know and be able to do.
  13. 13. EMPHASIS OF PERFORMANCE ASSESSMENT• Should assess higher level cognitive skills rather than narrow and lower level discreet skills.• Direct measures of skills of interest.
  14. 14. CHARACTERISTICS OF PERFORMANCE- BASED ASSESSMENT• Students perform, create, construct, produce, or do something.• Deep understanding and/or reasoning skills are needed and assessed.• Involves sustained work, often days and weeks.• Calls on students to explain, justify, and defend.• Performance is directly observable.• Involves engaging in ideas of importance and substance.• Relies on trained assessor’s judgments for scoring• Multiple criteria and standards are prespecified and public• There is no single correct answer.• If authentic, the performance is grounded in real world contexts and constraints.
  15. 15. VARIATION OF AUTHENTICITYRelatively authentic Somewhat authentic AuthenticIndicate which Design a garden Create a gardenparts of a gardendesign areaccurateWrite a paper on Write a proposal to Write a proposal tozoning change fictitious present to city zoning laws council to change zoning lawsExplain what Show how to Play a basketballwould you teach to perform basketball game.students learning skills in practicebasketball
  16. 16. • Answer worksheet 2
  17. 17. CONSTRUCTING PERFORMANCE BASED TASKS1. Identify the performance task in which students will be engaged2. Develop descriptions of the task and the context in which the performance is to be conducted.3. Write the specific question, prompt, or problem that the student will receive.• Structure: Individual or group?• Content: Specific or integrated?• Complexity: Restricted or extended?
  18. 18. COMPLEXITY OF TASK• Restricted-type task – Narrowly defined and require brief responses – Task is structured and specific – Ex: • Construct a bar graph from data provided • Demonstrate a shorter conversation in French about what is on a menu • Read an article from the newspaper and answer questions • Flip a coin ten times. Predict what the next ten flips of the coin will be, and explain why. • Listen to the evening news on television and explain if you believe the stories are biased. • Construct a circle, square, and triangle from provided materials that have the same circumference.
  19. 19. • Extended-type task – Complex, elaborate, and time-consuming. – Often include collaborative work with small group of students. – Requires the use of a variety of information – Examples: • Design a playhouse and estimate cost of materials and labor • Plan a trip to another country: Include the budget and itinerary, and justify why you want to visit certain places • Conduct a historical reenactment (e. g. impeachment trial of ERAP) • Diagnose and repair a car problem • Design an advertising campaign for a new or existing product
  20. 20. IDENTIFYING PERFORMANCE TASK DESCRIPTION• Prepare a task description• Listing of specifications to ensure that essential if criteria are met• Includes the ff.: – Content and skill targets to be assessed – Description of student activities • Group or individual • Help allowed – Resources needed – Teacher role – Administrative process – Scoring procedures
  21. 21. PERFORMANCE-BASED TASK QUESTION PROMPT• Task prompts and questions will be based on the task descriptions.• Clearly identifies the outcomes, outlines what the students are encourage dot do, explains criteria for judgment.
  22. 22. PROMPT:
  23. 23. PERFORMANCE CRITERIA• What you look for in student responses to evaluate their progress toward meeting the learning target.• Dimensions of traits in performance that are used to illustrate understanding, reasoning, and proficiency.• Start with identifying the most important dimensions of the performance• What distinguishes an adequate to an inadequate demonstration of the target?
  24. 24. EXAMPLE OF CRITERIA• Learning target: – Students will be able to write a persuasive paper to encourage the reader to accept a specific course of action or point of view.• Criteria: – Appropriateness of language for the audience – Plausibility and relevance of supporting arguments. – Level of detail presented – Evidence of creative, innovative thinking – Clarity of expression – Organization of ideas
  25. 25. • Watch video of Cody Green
  26. 26. RATING SCALES• Indicate the degree to which a particular dimension is present.• Three kinds: Numerical, qualitative, combined qualitative/quantitative
  27. 27. • Numerical Scale – Numbers of a continuum to indicate different level of proficiency in terms of frequency or qualityExample:No Understanding 12345 Complete understandingNo organization 12345 Clear organizationEmergent reader 12345 Fluent reader
  28. 28. • Qualitative scale – Uses verbal descriptions to indicate student performance. – Provides a way to check the whether each dimension was evidenced. • Type A: Indicate different gradations of the dimension • Type B: Checklist
  29. 29. • Example of Type A: – Minimal, partial, complete – Never, seldom, occasionally, frequently, always – Consistent, sporadically, rarely – None, some, complete – Novice, intermediate, advance, superior – Inadequate, needs improvement, good excellent – Excellent, proficient, needs improvement – Absent, developing, adequate, fully developed – Limited, partial, thorough – Emerging, developing, achieving – Not there yet, shows growth, proficient – Excellent, good, fair, poor
  30. 30. • Example of Type A: Checklist
  31. 31. • Holistic scale – The category of the scale contains several criteria, yielding a single score that gives an overall impression or ratingExample level 4: Sophisticated understanding of text indicated with constructed meaning level 3: Solid understanding of text indicated with some constructed meaning level 2: Partial understanding of text indicated with tenuous constructed meaning level 1: superficial understanding of text with little or no constructed meaning
  33. 33. • Analytic Scale – One in which each criterion receives a separate score. ExampleCriteria Outstanding Competent Marginal 5 4 3 2 1Creative ideasLogicalorganizationRelevance of detailVariety in wordsand sentencesVivid images
  34. 34. RUBRICS• When scoring criteria are combined with a rating scale, a complete scoring guideline is produced or rubric.• A scoring guide that uses criteria to differentiate between levels of student proficiency.
  36. 36. GUIDELINES IN CREATING A RUBRIC1. Be sure the criteria focus on important aspects of the performance2. Match the type of rating with the purpose of the assessment3. The descriptions of the criteria should be directly observable4. The criteria should be written so that students, parents, and others understand them.5. The characteristics and traits used in the scale should be clearly and specifically defined.6. Take appropriate steps to minimize scoring frame
  37. 37. PORTFOLIO ASSESSMENT: EXPLORATION • Have you ever done a portfolio? • Tell me about this experience. Did you enjoy it? • What elements did you include in your portfolio? • Are the materials placed in the portfolio required?
  38. 38. What are Portfolios?• Purposeful, systematic process of collecting and evaluating student products to document progress toward the attainment of learning targets or show evidence that a learning target has been achieved.• Includes student participation in the selection and student self-reflection.• “A collection of artifacts accompanied by a reflective narrative that not only helps the learner to understand and extend learning, but invites the reader of the portfolio to gain insight about learning and the learner (Porter & Cleland, 1995)
  39. 39. Characteristics of Portfolio assessment• Clearly defined purpose and learning targets• Systematic and organized collection of student products• Preestablished guidelines for what will be included• Student selection of some works that will be included• Student self-reflection and self-evaluation• Progress documented with specific products and/or evaluations• Portfolio conferences between students and teachers
  40. 40. A portfolio is:• Purposeful• Systematic and well-organized• Prestablished guidelines are set-up• Students are engaged in the selection of some materials• Clear and well-specified scoring criteria
  41. 41. Purpose of Portfolio• Showcase portfolio: Selection of best works. Student chooses work, profile are accomplishments and individual profile emerges.• Documentation portfolio: Like a scrapbook of information and examples. Inlcudes observations, tests, checklists, and rating scales.• Evaluation portfolio: More standardized. Assess student learning with self-
  42. 42. Advantages of portfolio• Students are actively involved in self-evaluation and self- reflection• Involves collaborative assessment• Ongoing process where students demonstrate performance, evaluate , revise , and produce quality work.• Focus on self-improvement rather than comparison with others• Students become more engaged in learning because both instruction and assessment shift from teacher controlled to mix of internal and external control.• Products help teachers diagnose learning difficulties• clarify reasons for evaluation• Flexible
  43. 43. Disadvatntages• Scoring difficulties may lead to low reliability• Teacher training needed• Time-consuming to develop criteria, score and meet students• Students may not make good selections of which of which material to include• Sampling of student products may lead to weak generalization
  44. 44. Steps in Planning andImplementing Portfolio Assessment1. Determine the the purpose2. Identify physical structure3. Determine sources of content4. Determine sources of content5. Determine student reflective guidelines and scoring criteria6. Review with students7. Portfolio content supplied by teacher and/or student8. Student self-evaluation of contents9. Teacher evaluation of content and student self-evaluation10. Student-teacher conference11. Portfolios returned to students for school
  45. 45. Purpose• Based on specific learning targets• Ideal for assessing product, skill, and reasoning targetsUses:• Showcase portfolio-to illustrate what students are capable of doing• Evaluation of portfolio-standardization of what to include• For parents-what will make sense to parents“Provide specific attention to purpose and corresponding implications when implementing a portfolio.”
  46. 46. Physical structure• What will it look like?• How large will the portfolios be?• Where are they stored so that students can easily access them?• Will it be in folders or scrap books?• How will the works be arranged in the portfolio?• What materials are needed to separate the works in the portfolio?
  47. 47. Sources of content• Work samples• Student and teacher evaluationsGuidelines:• Select categories that will allow you to meet the pupose of the portfolio.• Show improvement in the portfolio• Provide feedback on the students on the procedures they are putting together• Provide indicator system
  48. 48. Self-reflective guidelines and scoring• Establish guidelines for student self- reflection and the scoring criteria• Scoring guidelines are explained to the students before they begin instruction
  49. 49. Implementing portfolio assessment• Review with students: Explain to students what is involved in doing a portfolio.• Begin with learning targets• Show examples• Give opportunities to ask questions• Provide just enough structure so that they can get started without telling them exactly what to do.• Selection of content will depend on the age and previos experience of students• Students and teachers decide together what to include with nonrestrictive guidelines
  50. 50. Some organization• Include table of contents• Brief description of activities• Date produced• Date submitted• Date evaluated
  51. 51. Student self-evaluations• Reflective and self-evaluation activities need to be taught.• Some guide questions for students: – Can you tell me what you did? – What did you like best abut this sample of your writing? – What will you do next?• Self-reflective questions: – What did you learn from writing this piece? – What would you have done differently if you had more time? – What are your greatest strengths and weaknesses in this sample? – What would you do differently if you did this over?
  52. 52. Peer evaluations• Analysis and constructive, supportive criticism of strategies, styles, and other concrete aspects of the product.• Can include comments or a review by parentsTeacher evaluations:• Checklist of content• Portfolio structure evaluation: selection of samples, thoroughness, appearance, self- reflection, and organization.
  53. 53. Student-teacher conferences• Conference is conducted with students before returning the portfolio• Scheduled throughout the school year; some have it monthly• Clarify purposes and procedure with students, answer questions and establish trust• Give guidelines to prepare for each conference• Allow the students to do most of the talking• Have students compare your reflections with theirs• Weaknesses and areas for improvement need to be communicated –show them what is possible for progress
  54. 54. Student-teacher conferences• At the end of the conference there is an action plan for the future• Limit the conference to no more than 10 minutes• Students are encouraged to take notes• Focus on one or two major areas of each conference-helps to have a thoughtful discussion
  55. 55. Advance Organizer1 The Test Blueprint Outline of the Test Development Process Table of Specifications2 Designing Selected-Response Items Binary-choice items Instructions in Writing Binary Type of Items Multiple-choice items Guidelines in Writing Multiple-choice Items Matching items Guidelines in Writing Multiple-choice Items3 Designing Constructed-Response Items Short-answer items Guidelines in Writing Short Answer Items Essay items4 Designing Interpretive Exercise Guidelines in Writing Intepretive Exercise 60 Examples of Interpretive Exercise
  56. 56. Objectives• 1. Explain the theories and concepts that rationalize the practice of assessment.• 2. Make a table of specifications of the test items.• 3. Design pen-and-paper tests that are aligned to the learning intents.• 4. Justify the advantages and disadvantages of any pen-and-paper test.• 5. Evaluate the test items according to the guidelines presented. 61
  57. 57. Outline of Test Development Process• 1. Specify the ultimate goals of the education process• 2. Derive from these the goals of the portion of the system under study• 3. Specify these goals in terms of expected student behavior. If relevant, specify the acceptance level of successful learning.• 4. Determine the relative emphasis or importance of various objectives, their content, and their behaviors.• 5. Select or develop situations that will elicit the desired behavior in the appropriate context or environment, assuming the student has learned it.• 6. Assemble a sample of such situations that together represent accurately the emphasis on content and behavior previously determined. 62
  58. 58. Outline of Test Development Process• 7. Provide for the recording of responses in a form that will facilitate scoring but will not so distort the nature of the behavior elicited that it is no longer a true sample or index of the behavior desired.• 8. Establish scoring criteria and guides to provide objective and unbiased judgment.• 9. Try out the instrument in preliminary form.• 10. Revise the sample of situations on the basis of tryout information.• 11. Analyze reliability, validity, and score distribution in accordance with the projected use of scores.• 12. Develop test norms and a manual, and reproduce and distribute the test. 63
  59. 59. Test Length• The test must be of sufficient length to yield reliable scores• The longer the test, the more the reliable the results• The test should be valid if it is reliable• For the grade school, one must consider the stamina and attention span of the pupils• The test should be long enough to be adequately reliable and short enough to be administered 64
  60. 60. Test Instruction• It is the function of the test instructions to furnish the learning experiences needed in order to enable each examinee to understand clearly what he is being asked to do?• Instructions may be oral, a combination of written and oral instruction is probably desirable, except with very young children.• Clear concise and specific. 65
  61. 61. Test layout• The arrangement of the test items influences the speed and accuracy of the examinee• Utilize the space available while retaining readability.• Items of the same type should be grouped together• Arrange test items from easiest to most difficult as a means of reducing test anxiety.• The test should be ordered first by type then by content• Each item should be completed in the column and page in which it is started.• If the reference material is needed, it should occur on the same page as the item• If you are using numbers to identify items it is better to use letters for the options 66
  62. 62. Scoring the test• Use separate answer sheets• Punched key• Overlay key• Strip keyPlight of the student• The teacher should discuss with the class the content areas and levels of the cognitive domain to be examined• The discussion should utilize a vocabulary and a level of complexity appropriate to the development level of the student• Types of test• Examples of test type 67
  63. 63. Table of Specifications One Grid TOS Content Outline No. of items1. Table of specifications 102. Test and Item characteristics 203. Test layout 54. Test instructions 55. Reproducing the test 56. Test length 57. Scoring the test 5TOTAL 55 68
  64. 64. Table of SpecificationsTwo Grid TOSWeight Content Knowledge Comprehension Application No. of items(Time Outline 30% 40% 30% by contentFrame) area 35% 1. Table of specifications 1 4 4 9 30% 2. Test and Item characteristics 2 3 3 8 10% 3. Test layout 1 1 0 2 5% 4. Test instructions 0 1 0 1 5% 5. Reproducing the test 1 0 0 1 5% 6. Test length 1 0 1 2 10% 7. Scoring the test 2 1 0 3 8 10 8 26 The number of items in a cell is computed using the formula: Given time items = X percentage of cognitive skill X total number of items 69 Total time
  65. 65. Classification of test Items• Selected Response – Binary Choices – Multiple Choice – Matching Type• Constructed Response “Supply Test” – Short Form answers - identification – Completion – fill in the blanks, cloz test – Essay• Performance Type – Paper and pencil type – Identification type – Simulation 70
  66. 66. Item Writing Commandments• Thou shall not produce opaque directions to students regarding how to respond to your instructions (opaque directions)• Thou shall not employ ambiguous statements in your assessment item (ambiguous statements)• Thou shall not unintentionally provide students with clues regarding appropriate response (unintended clues)• Thou shall not employ complex syntax in your assessment item (complex syntax)• Thou shall not use vocabulary that is more advanced than required (Difficult vocabulary) 71
  67. 67. SHORT ANSWER ITEMS• 1. Word the item so that the answer is both brief and definite.• 2. Do not take statements directly from books to use as a basis for short answer items.• 3. A direct question is generally more acceptable than an incomplete statement.• 4. The answer to be expressed in numerical units indicate the type of answer wanted.• 5. Blanks for answers should be equal in length.• 6. Do not use to many blanks. 72
  68. 68. Writing supply items1. Require short, definite, clear-cut, and explicit answersFAULTY: Earnest Hemingway wrote______IMPROVED: The Old Man and the Sea was written by _______. Who wrote The Old man and the Sea?2. Avoid multimutilated statementsFAULTY: _____ pointed out in ____ the freedom of thought in America was seriously hampered by ___, ____, & __.IMPROVED: That freedom of thought in America was seriously hampered by social pressures toward conformity was pointed out in 1830 by ______. 73
  69. 69. Writing supply items3. If several equal answers equal credit should be given to each one.4. Specify and announce in advance whether scoring will take spelling into account.5. In testing for comprehension of terms and knowledge of definition, it is often better to supply the term and require a definition than to provide a definition and require the term.FAULTY: What is the general measurement term describing the consistency with which items in a test measure the same thing?IMPROVED: Define “internal consistency reliability.” 74
  70. 70. Writing supply items6. It is generally recommended that in completion items the blanks come at the end of the statement.FAULTY: A (an) ________ is the index obtained by dividing a mental age score by chronological age and multiplying by 100.IMPROVED: The index obtained by dividing a mental age score by chronological age and multiplying by 100 is called a (an) ________7. Minimize the use of textbook expressions and stereotyped language.FAULTY: The power to declare war is vested in ______IMPROVED: Which national legislative body has the authority to declare war? 75
  71. 71. Writing supply items8. Specify the terms in which the response is to be given.FAULTY: Where does the Security Council of the United Nations hold its meeting?IMPROVED: In what city of the United States does the Security Council of the United Nations hold its meeting?FAULTY: If a circle has 4-inch diameter, its area is_____IMPROVED: A circle has 4-inch diameter. Its area in square inches correct to two decimal places, is _____9. In general, direct questions are preferable to incomplete declarative sentences.FAULTY: Gold was discovered in California in the year ___IMPROVED: In what year was gold discovered in California? 76
  72. 72. Writing supply items10. Avoid extraneous clues to the correct answerFAULTY: A fraction whose denominator is greater than its numerator is a _____IMPROVED: Fractions whose denominator are greater than their numerators are called _____ 77
  73. 73. ALTERNATIVE RESPONSE ITEM• 1. Avoid broad general statements if they are to be judged true or false.• 2. Avoid trivial statements.• 3. Avoid the use of negative statements.• 4. Avoid long complex sentences.• 5. Avoid including two ideas in one statement unless cause and effect relationship are being measured.• 6. If the opinion is being used, attribute it to some source unless the ability to identify opinion is being specifically measured.• 7. True statements and false statements should be equal in length.• 8. The number of true and false statements should be approximately equal. 78
  74. 74. Writing TRUE-FALSE Items1. Avoid the use of “specific determiners”FAULTY: No picture-no sound in a television set may indicate a bad 5U4G.IMPROVED: A bad 5U4G tube in a television set will result in no picture sound.2. Base true-false items upon statements that are absolutely true or false, without qualifications or exceptions.FAULTY: World War II was fought in Europe and the Far East.IMPROVED: The primary combat locations in terms of military personnel during World War II were Europe and the Far East. 79
  75. 75. Writing TRUE-FALSE Items3. Avoid negative stated items when possible and eliminate all double negatives.FAULTY: It is not frequently observed that copper turns green as a result of oxidation.IMPROVED: Copper will turn green upon oxidizing.4. Use quantitative and precise rather than qualitative language where possible.FAULTY: Many people voted for Gloria Arroyo in the 2003 Presidential election.IMPROVED: Gloria Arroyo received more than 60 percent of the popular votes cast in the Presidential election of 2003. 80
  76. 76. Writing TRUE-FALSE Items5. Avoid stereotypic and textbook statements.FAULTY: From time to time efforts have been made to explode the notion that there may be a cause-and-effect relationship between arboreal life and primate anatomy.IMPROVED: There is a known relationship between primate anatomy and arboreal life.6. Avoid making the true items consistently longer than the false items.7. Avoid the use of unfamiliar or esoteric language.FAULTY: According to some peripatetic politicos, the raison d’etre for capital punishment is retribution.IMPROVED: According to some politicians, justification for the existence of capital punishment can be traced to the Biblical statement, “An eye for an eye.” 81
  77. 77. Writing TRUE-FALSE Items8. Avoid complex sentences with many dependent clauses.FAULTY: Jane Austen, an American novelist born in 1790, was a prolific writer and is best known for her novel Pride and Prejudice, which was published in 1820.IMPROVED: Jane Austen is best known for her novel Pride and prejudice.9. It is suggested that the crucial elements of an item be placed at the end of the statement.FAULTY: Oxygen reduction occurs more readily because carbon monoxide combines with hemoglobin faster than oxygen does.IMPROVED: Carbon monoxide poisoning occurs because carbon monoxide dissolves delicate lung tissue. 82
  78. 78. Writing Matching Type Test1. Matching Exercises should be complete on a single page.2. Use response categories that are related but mutually exclusive.3. Keep the number of stimuli relatively small (10-15), and let the number of possible responses exceed the number of stimuli by two or three.4. The direction should clearly specify the basis for matching stimuli and responses.5. Keep the statements in the response column short and list them in some logical order 83
  79. 79. FAULTY: Match List A with List B. You will be given one point for each correct match. List A List B a. cotton gin a. Eli Whitney b. reaper b. Alexander Graham Bell c. wheel c. David Brinkley d. TU54G tube d. Louisa May Alcott e. steamboat e. None of these• Directions failed to specify the basis for matching• List are enumerated identically• Responses not listed logically• Lacks homogeneity• Equal number of elements• Use of “None of the above” 84
  80. 80. IMPROVED: Famous inventions are listed in the left-hand column and inventors in the right-hand column below. Place the letter corresponding to the inventor in the space next to the invention for which he is famous. Each match is worth 1 point, and “None of these” may be the correct answer. Inventors may be used more than once. Inventions Inventors__ 1. steamboat a. Alexander Graham-Bell__ 2. cotton skin b. Robert Fulton__ 3. sewing machine c. Elias Howe__ 4. reaper d. Cyrus McCormick e. Eli Whitney f. None of these 85
  81. 81. Writing Multiple Choice1. It is recommended that the stem be a direct question.2. The stem should pose a clear, define, explicit, and singular problem.FAULTY: Salvador Dali is a. a famous Indian. b. important in international law. c. known for his surrealistic art. d. the author of many avant-garde plays.IMPROVED: With which one of the fine arts is Salvador Dali associated? a. surrealistic painting b. avant-garde theatre c. polytonal symphonic music d. impressionistic poetry 86
  82. 82. Writing Multiple Choice3. Include in the stem any words that might otherwise be repeated in each response.FAULTY: Milk can be pasteurized at home by a. heating it to a temperature of 130o b. Heating it to a temperature of 145o c. Heating it to a temperature of 160o d. Heating it to a temperature of 175oIMPROVED: The minimum temperature that can be used to pasteurize milk at home is: a. 130o b. 145o c. 160o d. 175o 87
  83. 83. Writing Multiple Choice4. Items should be stated simply and understandably, excluding all nonfunctional words from stem and alternatives.FAULTY: Although the experimental research, particularly that by Hansmocker must be considered equivocal and assumptions viewed as too restrictive, most testing experts would recommend as the easiest method of significantly improving paper-and-pencil achievement test reliability to a. increase the size of the group being tested. b. increase the differential weighting of items. c. increase the objective of scoring. d. increase the number of items. e. increase the amount of testing time.IMPROVED: Assume a 10-item, 10-minute paper-and-pencil multiple choice achievement test has a reliability of .40. The easiest way of increasing the reliability to .80 would be to increased a. group size b. scoring objectivity c. differential item scoring weights d. the number of items e. testing time 88
  84. 84. Writing Multiple Choice5. Avoid interrelated items6. Avoid negatively stated itemsFAULTY: None of the following cities is a state capital except a. Bangor b. Los Angeles c. Denver d. New HavenIMPROVED: Which of the following cities is a state capital? a. Bangor b. Los Angeles c. Denver d. New Haven 89
  85. 85. Writing Multiple Choice7. Avoid making the correct alternative systematically different from other options8. If possible the alternatives should be presented in some logical, numerical, or systematic order.9. Response alternatives should be mutually exclusive.FAULTY: Who wrote Harry Potter and the Goblet of Fire? a. J. K. Rowling b. Manny Paquiao c. Lea Salonga d. Mark TwainIMPROVED: Who wrote Penrod? a. J. K. Rowling b. J. R. R. Tolkien c. V. Hugo d. L. Carrol 90
  86. 86. Writing Multiple Choice10. Make all responses plausible and attractive to the less knowledgeable and skillful student.FAULTY: Which of the following statements makes clear the meaning of the word “electron”? a. An electronic tool b. Neutral particles c. Negative particles d. A voting machine e. The nuclei of atomsIMPROVED: Which of the following phrases is a description of an “electron”? a. Neutral particle b. Negative particle c. Neutralized proton d. Radiated particle e. Atom nucleus 91
  87. 87. Writing Multiple Choice11. The response alternative “None of the above” should be used with caution, if at all.FAULTY: What is the area of a right triangle whose sides adjacent to the right angle are 4 inches long respectively? a. 7 b. 12 c. 25 d. None of the aboveIMPROVED: What is the area of a right triangle whose sides adjacent to the right angle are 4 inches and 3 inches respectively? a. 6 sq. inches b. 7 sq. inches c. 12 sq. inches d. 25 sq. inches e. None of the above 92
  88. 88. Writing Multiple Choice12. Make options grammatically parallel to each other and consistent with the stem.FAULTY: As compared with the American factory worker in the early part of the 19th century, the American factory worker at the close of the century a. was working long hours b. received greater social security benefits c. was to receive lower money wages d. was less likely to belong to a labor union. e. became less likely to have personal contact with employersIMPROVED: As compared with the American factory worker in the early part of the century, the American factory worker at the close of the century a. worked longer hours. b. had more social security. c. received lower money wages. d. was less likely to belong to a labor union e. had less personal contact with his employer 93
  89. 89. Writing Multiple Choice13. Avoid such irrelevant cues as “common elements” and “pat verbal associations.”FAULTY: The “standard error of estimate’ refer to a. the objectivity of scoring. b. the percentage of reduced error variance. c. an absolute amount of possible error. d. the amount of error in estimating criterion scores.IMPROVED: The “standard error of estimate” is most directly related to which of the following test characteristic? a. Objectivity b. Reliability c. Validity d. Usability e. Specificity 94
  90. 90. Writing Multiple Choice14. In testing for understanding of a term or concept, it is generally preferable to present the term in the stem and alternative definitions in the options.FAULTY: What name is given to the group of complex organic compounds that occur in small quantities in natural foods that are essential to normal nutrition? a. Calorie b. Minerals c. Nutrients d. VitaminsIMPROVED: Which of the following statements is the best description of a vitamin?15. Use objective items – items’ whose correct answers are agreed by experts 95
  91. 91. Factual Knowledge• The Monroe Doctrine was announced about 10 years after the a. Revolutionary War b. War of 1812 c. Civil War d. Spanish-American WarConceptual Knowledge2. Which of the following statements of the relationship between market price and normal price is true? a. Over a short period of time, market price varies directly with changes in normal price. b. Over a long period of time, market price tends to equal normal price. c. Market price is usually lower than normal price. d. Over a long period of time, market price determines normal price. 96
  92. 92. Translation from symbolic form to another form, or vice versa 3. Which of the graphs below best represent the supply situation where a monopolist maintains a uniform price regardless of the amounts which people buy? A B C D S S S S SPrice Price Price Price S S Quantity Quantity Quantity Quantity 97
  93. 93. ApplicationIn the following items (4-8) you are to judge the effects of a particular policy on the distribution of income. In each case assume that there are no other changes in policy that would counteract the effect of the policy described in the item. Mark the item:A. If the policy described would tend to reduce the existing degree of inequality in the distribution of income,B. If the policy described would tend to increase the existing degree of inequality in the distribution of income, orC. If the policy described would have no effect, or an indeterminate effect, on the distribution of income.__ 4. Increasingly progressive income taxes.__ 5. Confiscation of rent on unimproved__ 6. Introduction of a national sales tax__ 7. Increasing the personal exemptions from income taxes__ 8. Distributing a subsidy to sharecroppers on southern farms 98
  94. 94. Analysis9. An assumption basic to Lindsay’s preference for voluntary associations rather than government order… is a belief a. that government is not organized to make the best use of experts b. that freedom of speech, freedom of meeting, freedom of association, and possible only under a system of voluntary associations. c. in the value of experiment and initiative as a means of attaining an ever improving society d. in the benefits of competition 99
  95. 95. Judgments in terms of external criteriaFor items 14-16, assume that in doing research for a paper about the English language you find a statement by Otto Jespersen that contradicts one point of view in a language you have always accepted. Indicate which of the statements would be significant in determining the value of Jespersen’s statement. For the purpose of these items, you may assume that these statements are accurate. Mark each item using the following key.A. Significant positively – that is, might lead you to trust his statement and to revise your own opinion.B. Significant negatively – that is, night lead you to distrust his statementC. Has no significance__ 14. Mr. Jesperson was professor of English at Copenhagen University__ 15. The statement in question was taken from the very first article that Jespersen published__ 16. Mr. Jespersen’s books are frequently referred to in other works that you consult. 100
  96. 96. Essay Questions• 1. Ask questions or set tasks that will require the examinee to demonstrate a command of essential knowledge.• 2. Ask questions that are determinate, in the sense that experts could agree that one answer is better than another.• 3. Define the examinee’s task as completely and specifically as possible without interfering with measurements of the achievement intended.• 4. In general, give preference to more specific questions that can be more answered briefly.• 5. Avoid giving the examinee a choice among optional questions unless special circumstances make such option necessary.• 6. Test the questions by writing an ideal answer 101
  97. 97. Types of Essays:• General – extensiveness of responses• Restrictive Response – reliable scoringLearning outcomes measured by Essay:• Explain cause-effect relationship• Describe applications of principles• Present relevant arguments• Formulate tangible hypothesis• Formulate valid conclusions• State necessary assumptions• Describe the limitations of data• Explain methods and procedures• Produce, organize, and express ideas• Integrate learnings in different areas• Create original forms• Evaluate the worth of ideas 102
  98. 98. Understanding:A. Comparison of two phenomena on a single designated basis: Compare the writers of the English Renaissance to those of the nineteenth century with respect to their ability tot describe natureB. Comparison of two phenomena in general Compare the French and Russian RevolutionsC. Explanation of the use or exact meaning of a phrase or statement The book of John begins “In the beginning was the word…” From what philosophical system does this statement derive?D. Summary of a text or some portion of it State the central theme of the Communist ManifestoE. Statement of an artist’s purpose in the selection or organization of material Why did Hemingway describe in detail the episode in which Gordon, lying wounded, engage the oncoming enemy? What was Beethoven’s purpose in deviating from the orthodox form of a symphony in Symphony No. 6? 103
  99. 99. Application:A. Causes or effects Why may too frequent reliance on penicillin for the treatment of minor ailments eventually result in its diminished effectiveness against major invasion of body tissues by infectious bacteria?B. Analysis Why was Hamlet torn by conflicting desires?C. Statement of relationship It is said that intelligence correlates with school achievement at about .65. Explain this relationshipD. Illustrations or examples of principles Name three examples of uses of the lever in typical American homesE. Application of rules or principles Would you weigh more or less on the moon? On the sun? Explain.F. Reorganization of facts Some writers have said that the American Revolution was not merely a political revolution against England but also a social revolution, within the colonies, of the poor against the wealthy. Using the same evidence what other conclusion is possible? 104
  100. 100. Judgment:A. Decision for or against Should members of the Communist Party be allowed to teach in American colleges? Why or why not?B. Discussion Discuss the likelihood that four-year private liberal arts colleges will gradually be replaced by junior colleges and state universities.C. Criticism of the adequacy, correctness, or relevance of a statement The discovery of penicillin has often been called an accident. Comment on the adequacy of this explanation.D. Formulation of new questions What should one find out in order to explain why some students of high intelligence fail in school? 105
  101. 101. Designing Interpretive Exercise• Guidelines in Writing Interpretive Exercise• 1. Select an introductory that is in harmony with the objectives of the course. – Amount of emphasis of various interpretive skills is a factor. – Do not overload test takers with interpretive items in a particular area. – Selection of introductory should be guided by general emphasis to be given to the measurement of complex achievement.• 2. Select introductory material that is appropriate to the curricular experience and reading ability of the examinees. 106
  102. 102. Guidelines in Writing Interpretive Exercise• 3. Select introductory material that is new to pupils.• 4. Select introductory material that is brief but meaningful.• 5. Revise introductory material for clarity, conciseness, and greater interpretive value.• 6. Construct test items that require analysis and interpretation of introductory material.• 7. Make the number of items roughly proportional to the length of the introductory material.• 8. Observe all suggestions for constructing objective test items. 107
  103. 103. • Ability to Recognize the Relevance of Information 108
  104. 104. • Ability to Recognize Warranted and Unwarranted Generalizations 109
  105. 105. • Ability to Recognize Inferences 110
  106. 106. • Ability to Interpret Experimental Findings 111
  107. 107. • Ability to Apply Principles 112
  108. 108. • Ability to Recognize Assumptions 113
  109. 109. Reading comprehension• Bem (1975) has argued that androgynous people are “better off” than their sex-typed counterparts 35. What is the independent variable in the because they are not constrained by rigid sex- study? role concepts and are freer to respond to a wider variety of situations. Seeking to test this hypothesis, Bem exposed masculine, feminine, a.Situations calling for independence and and androgynous men and women to situations that called for independence (a masculine nurturance attribute) or nurturance (a feminine attribute). The b.Situation to make the sex type react test for masculine independence assessed the subject’s willingness to resist social pressure by c.Situations to make the androgynous be refusing to agree with peers who gave bogus flexible judgments when rating cartoons for funniness (for example, several peers might say that a very d.Situations like sex type, androgynous and funny cartoon was hilarious). Nurturance or sex role concepts feminine expressiveness, was measured by observing the behavior of the subject when left alone for ten minutes with a 5-month old baby. 36. What are the levels of the IV? The result confirmed Bem’s hypothesis. Both the masculine sex-typed and the androgynous subjects were more independent (less a.masculine attribute and feminine attribute conforming) on the ‘independence” test than feminine sex-typed individuals. Furthermore, both b.rating cartoons and taking care of a baby the feminine and the androgynous subjects were c.independence and nurturance more “nurturant” than the masculine sex-typed individuals when interacting with the baby. Thus, d.flexibility and rigidity the androgynous subjects were quite flexible, they performed as masculine subjects did on the “feminine” task. 114
  110. 110. Interpreting DiagramsInstruction. Study the following illustrations and answer the following questions. 101. Which group received the treatment?Figure 1 A b. group B b.c. none of the above 102. Why did group B remain stable across the experiment? a. there is an EV b. there is no treatment c. there is the occurence of ceiling effect 103. What is the problem at the start of the experiment? Pretest Posttest a.the groups are nonequivalent b.the groups are competing with each other c. the treatment took place immediately 115
  111. 111. Analysis of Test ResultsReliability, Validity, and Item Analysis
  112. 112. Learning Content• Levels of Measurement• Correlation Coefficient• Reliability• Validity• Item Analysis
  113. 113. Objectives• 1. Determine the use of the different ways of establishing an assessment tools’ validity and reliability.• 2. Familiarize on the different methods of establishing an assessment tools’ validity and reliability.• 3. Assess how good an assessment tool is by determining the index of validity, reliability, item discrimination, and item difficulty.
  114. 114. Levels of Measurement• Nominal• Ordinal• Interval• Ratio
  115. 115. Correlation Coefficient• Relationship of two variables (X & Y)• Direction• Positive Negative X Y
  116. 116. Degree of Relationship• 0.80 – 1.00 Very High relationship• 0.6 – 0.79 High Relationship• 0.40 – 0.59 Substantial/Marked relationship• 0.20 – 0.39 Low relationship• 0.00 – 0.19 Negligible relationship
  117. 117. Testing for Significance• Nominal: Phi Coefficient• Ordinal: Spearman rho• Interval & Ratio: Pearson r• Interval with nominal: Point biserial• Decision rule:• If p value < α=.05: significant relationship• If p value > α=.05: no significant relationship
  118. 118. Variance• R2• Square the correlation coefficient• Interpretation: percentage of time that the variability in X accounts for the variability in Y.
  119. 119. Reliability• Consistency of scores Obtained by the same person when retested with the identical test or with an equivalent form of the test
  120. 120. Test-Retest Reliability• Repeating the identical test on a second occasion• Temporal stability• When variables are stable ex: motor coordination, finger dexterity, aptitude, capacity to learn• Correlate the scores from the first test and second test.· The higher the correlation the more reliable
  121. 121. Alternate Form/Parallel Form• Same person is tested with one form on the first occasion and with another equivalent form on the second• Equivalence;• Temporal stability and consistency of response• Used for personality and mental ability tests• Correlate scores on the first form and scores on the second form
  122. 122. Split half• Two scores are obtained for each person by dividing the test into equivalent halves• Internal consistency;• Homogeneity of items• Used for personality and mental ability tests• The test should have many items• Correlate scores of the odd and even numbered items• Convert the obtained correlation coefficient into a coefficient estimate using Spearman Brown•
  123. 123. Kuder Richardson (KR #20/KR #21)• When computing for binary (e.g., true/false) items• Consistency of responses to all items• Used if there is a correct answer (right or wrong)• Use KR #20 or KR #21 formula
  124. 124. Coefficient Alpha• The reliability that would result if all values for each item were standardized (z transformed)• Consistency of responses to all items• Homogeneity of items• Used for personality tests with multiple scored-items• Use the cronbach’s alpha formula
  125. 125. Inter-item reliability• Consistency of responses to all items• Homogeneity of items• Used for personality tests with multiple scored-items• Each item is correlated with every item in the test
  126. 126. Scorer Reliability• Having a sample of test papers independently scored by two examiners• To decrease examiner or scorer variance• Clinical instruments employed in intensive individual tests ex. projective tests• The two scores from the two raters obtained are correlated with each other
  127. 127. Validity• Degree to which the test actually measures what it purports to measure
  128. 128. Content Validity• Systematic examination of the test content to determine whether it covers a representative sample of the behavior domain to be measured.• More appropriate for achievement tests & teacher made tests• Items are based on instructional objectives, course syllabi & textbooks• Consultation with experts• Making test-specifications
  129. 129. Criterion-Prediction Validity• Prediction from the test to any criterion situation over time interval• Hiring job applicants, selecting students for admission to college, assigning military personnel to occupational training programs• Test scores are correlated with other criterion measures ex: mechanical aptitude and job performance as a machinist
  130. 130. Concurrent validity• Tests are administered to a group on whom criterion data are already available• Diagnosing for existing status ex. entrance exam scores of students for college with their average grade for their senior year.• Correlate the test score with the other existing measure
  131. 131. Construct Validity • The extent to which the test may be said to measure a theoretical construct or trait. • Used for personality tests. Measures that are multidimensional ∀ • Correlate a new test with a similar earlier test as measured approximately the same general behavior ∀ • Factor analysis ∀ • Comparison of the upper and lower group ∀ • Point-biserial correlation (pass and fail with total test score) ∀ • Correlate subtest with the entire test
  132. 132. Convergent Validity• The test should correlate significantly from variables it is related to• Commonly for personality measures• Multitrait-multidimensional matrix
  133. 133. Divergent Validity• The test should not correlate significantly from variables from which it should differ• Commonly for personality measures• Multitrait-multidimensional matrix
  134. 134. Item Analysis• Item Difficulty – The percentage of respondents who answered an item correctly• Item Discrimination – Degree to which an item differentiates correctly among test takers in the behavior that the test is designed to measure.
  135. 135. Difficulty Index• Difficulty Index Remark• .76 or higher Easy Item• .25 to .75 Average Item• .24 or lower Difficult Item
  136. 136. Index Discrimination• .40 and above - Very good item• .30 - .39 - Good item• .20 - .29 - Reasonably Good item• .10 - .19 - Marginal item• Below .10 - Poor item