Faculty assessment presentation april 2014


Published on

  • Be the first to comment

  • Be the first to like this

No Downloads
Total views
On SlideShare
From Embeds
Number of Embeds
Embeds 0
No embeds

No notes for slide

Faculty assessment presentation april 2014

  1. 1. Designing Assessment tasks : Rita Kizito march 2014 Rita Kizito march 2014 Workshop for Earth Sciences Department April 2014
  2. 2. In this workshop we intend to… • Link assessment to module learning outcomes • Learn how to set assessment activities at the appropriate NQF level • Briefly interrogate assessment (purpose, types , features) • Refine/develop assessment tasks hand-out #1 – UWC Assessment policy
  3. 3. Integrate outcomes and assessment Knowledge reproduction/ creation Recall Comprehension Application Analysis Synthesis Evaluation Attitudes Inquiry focused Ethically, environmentally and socially aware Motivated Skills Skilled communicators Autonomous and collaborative Problem solving
  4. 4. Are your outcomes well designed ? Specific Provide details of aspect of expectation Meaningful Written in understandable language Appropriate Suit learner’s abilities & experiences Realistic Achievable in given time constraints Testable Some measure of progress/achievement (Butcher, et al., p. 41, 2006)
  5. 5. Bloom’s taxonomy for generating outcomes and assessment tasks
  6. 6. Bigg’s (2003) SOLO taxonomy: Single point Unanticipate d extension Logical related answer Multiple unrelated points Lower level outcomes Higher level outcomes
  7. 7. Bigg’s (2003) SOLO taxonomy: hand-out #2 – Learning taxonomy SOLO Taxonomy Structure of Observable Learning Outcome Prestructural Unistructural Multistructural Relational Extended abstract Quantitative Qualitative
  8. 8. Level Descriptors - SA Qualifications Framework To ensure coherence in learning achievement and to facilitate assessment at the appropriate levels. NQF Level Education Level Qualification 10 Doctorate 9 Masters 8 PG Diploma/Cert Honours 7 Bachelor (ord.) degree 6 Diploma 5 Certificate 4 Matric 3 2 1 HIGHER EDUCATION FURTHER EDUCATION GENERAL EDUCATION
  9. 9. Level Descriptors KNOWLEDGE COMPREHENSION APPLICATION ANALYSIS SYNTHESIS EVALUATIO N list name identify show define recognize recall state summarize explain put into your own words interpret describe compare paraphrase differentiate demonstrate visualize find more information about Restate solve illustrate calculate use interpret relate manipulate apply classify modify put into practice analyze organize deduce choose contrast compare distinguish design hypothesize support schematize write report discuss plan devise compare create construct Evaluate Choose Estimate Judge Defend Criticize Justify YEAR 1 YEARS 1, 2 YEARS 2, 3 YEAR 3 YEARS 3, 4 YEAR 4
  10. 10. Level Descriptors - SA Qualifications Framework What should the students know about the learning area/subject? What types of problems should the student be able to solve? …in what contexts? …with what methods/procedures? How should students obtain and process information? How should students communicate? How independent should the students be? Which values should students uphold? hand-out #3 – level descriptors
  11. 11. “We first have to be clear about what we want students to learn, and then teach and assess accordingly in an aligned system of instruction” (Biggs, 1996). Curriculum alignment
  12. 12. Teacher’s Intentions Student’ s Activity Exam’s Assess ment e.g. Explain Relate Prove Apply e.g. Memorize describe e.g. Memorize describe Unaligned course
  13. 13. Unaligned course Teacher’s Intentions Student’s Activity Exam’s Assessment e.g. Explain Relate Prove apply e.g. Explain Relate Prove apply e.g. Explain Relate Prove apply
  14. 14. Why assess? -two key purposes • Making judgements about student learning for certification/ grading • Helping to prompt and promote further learning/ for monitoring learning progress Summative assessment Formative assessment
  15. 15. What does assessment do ? • Defines what students will concentrate on when learning • Affects how they learn • Specifies what counts as learning • Provides information about shortfalls between performance and specification • Simulates conversation about, and reflection, on improvement Ndagire Kizito 04-04-2014 Peter Knight, 2001
  16. 16. Types of assessment • Laboratory Work •Class Presentations • Essays / Research • Online Tests • Problem Based Learning Case studies Projects • Reflective Journal • Multiple choice questions (MCQs)• Portfolio • Group work
  17. 17. Assessment features Reliability refers to the degree to which an assessment tool produces stable and consistent results. •Test-retest reliability correlating tests given twice over a period of time to a group of individuals. •Parallel forms reliability correlating different versions of an assessment tool (probing the same construct, skill, knowledge base, etc.) to the same group of individuals. •Inter-rater reliability is a measure of the degree to which different judges or raters agree in their assessment decisions. •Internal consistency reliability is an evaluation of the degree to which different test items that probe the same construct produce similar results. similar results
  18. 18. Assessment features Validity refers to how well a test measures what it is purported to measure. •Face Validity ascertains that the measure appears to be assessing the intended construct under study •Construct Validity is used to ensure that the measure is actually measuring what it is intended to measure (i.e. the construct), and not other variables. Ways to improve validity •Make sure your goals and outcomes are clearly defined and operationalized. •Match your assessment measures to your learning outcomes. Use outside reviewers. •If possible, compare your measure with other measures, or data that may be available.
  19. 19. Rubrics/ Assessment criteria • A rubric is a scoring /set of expectations used to judge student performance. It shows students how well they have performed on an assessment task. • Uses assessment criteria and levels of performance to break down a task in parts explaining what is required . • It can be used for a large number of tasks (essays, research projects, oral presentations, portfolios, etc.) and is especially useful for assessing complex , subjective subjects.
  20. 20. Rubrics/ Assessment criteria Examples of rubrics . hand-outs # 4 & 5 – Rubrics
  21. 21. Task 1 1. Identify one outcome from your module outline and develop an assessment task ( you can develop more than one task). 2. Create a rubric for one of the tasks in which you develop criteria (2 or more) and levels of expected achievement for that task answer. 3. Allow a colleague to review the tasks and make comments.
  22. 22. Task 2 Compare the two past question papers and answer the following questions 1.Are hey appropriate for the grade levels they have been prepared for? 2.Which one would you give to your students? Why? 3.How would you modify each one to fit your own context? hand-outs # 6 & 7 – Past papers
  23. 23. Hints for writing exam papers 1. Don’t do it on your own! Get one or two colleagues to do your questions. 2. Have your intended learning outcomes in front of you as your draft your questions. 3. Keep your sentences short. 4. Work out what you’re really testing. 5. Don’t measure the same things again and again. 6. Include data or information in questions to reduce the emphasis on memory.
  24. 24. Hints for writing exam papers 7. Make the question layout easy to follow. 8. Write out an answer to your own question. 9. Decide what the assessment criteria will be. 10.Work out a tight marking scheme. 11.Use the question itself to show how marks are to be allocated. 12.Try your questions out. 13.Proof-read your exam questions carefully hand-out # 8 A moderation checklist for exam papers hand-out # 9 Moderation checklist RGU
  25. 25. Biggs, J. (1996). Enhancing teaching through constructive alignment. Higher education, 32(3), 347-364. Butcher, K. R. (2006). Learning from text with diagrams: Promoting mental model development and inference generation. Journal of Educational Psychology, 98(1), 182. Knight, P. (2001). A Briefing on Key Concepts: Formative and summative, criterion and norm-referenced assessment. Learning and Teaching Support Network. SAQA (2000), The South African Qualifications Authority Level Descriptors for The South African National Qualifications Framework, ttp://www.saqa.org.za/docs/misc/level_descriptors.pdf References
  26. 26. Thank - you