Assessing Online WritingCAA Conference, 201020-21 July, 2010bobby.elliott@sqa.org.uk
http://www.scribd.com/bobbyelliott
QuestionDo you feel confident about assessingstudents’ online writing?
Text messageText keycode to 07624806527SmartphoneGo to poll4.com and enter keywordTwitterTweet keyword with comment to @pollWebhttp://www.polleverywhere.com/multiple_choice_polls/MTA0ODQwOTAwOA/webHow confident are you when assessing students’ online writing?Very: 24697Somewhat: 25157Not at all: 25192
Summary of presentationResearch questionMethodologyFindingsRecommendations
Some background stuffDefinition of assessmentCharacteristics of assessmentNew affordances
AssessmentLatin root: assidere: “to sit beside”“Observing learning”
Definition of assessment“Assessment is an on-going process aimed at understanding and improving student learning. It involves making our expectations explicit and public; setting appropriate criteria and high standards for learning quality; systematically gathering, analysing, and interpreting evidence to determine how well performance matches those expectations and standards; and using the resulting information to document, explain, and improve performance.”
Characteristics of assessmentValidityReliabilityFidelity
Fidelity“Fidelity is the extent to which elements that contribute to a course grade are correctly identified as academic achievement”.“Many academics cannot help but be impressed by the prodigious time and persistence that’s some students apparently invest in producing responses to an assessment task. However, effort is clearly an input variable and therefore does not fall within the definition of academic achievement”.“Assessment in which early understandings are assessed, recorded and counted, misrepresents the level of achievement reached at the end of the course.”
New AffordancesLarge-scale projectsDeveloped co-operatively and collaborativelyImproved authenticityReal life valueLarge audienceRicher contentMore reflectiveMore transparentPeer reviewed/assessed
Transforming Assessment ProjectAn Australian Learning and Teaching Council Fellowship specifically looking at the use of e-assessment within online learning environments, particularly those using one or more Web 2.0 or virtual world technologies.http://www.transformingassessment.com/geoffrey.crisp@adelaide.edu.au
Challenges“Most of its advocates [of writing using Web 2.0 tools] offer no guidance on how to conduct assessment that comes to grips with its unique features, its difference from previous forms of student writing, and staff marking or its academic administration.”“The few extant examples appear to encourage and assess superficial learning, and to gloss over the assessment opportunities and implications of Web 2.0’s distinguishing features.”
Research QuestionWhat is the current practice in the construction of rubrics to assess online discussions?
Methodology20 papers128 criteria/statements33 different criteria10 categories of criterion
Small PrintSmall sampleNot statistically robustNot generalisableNo claims about validity or reliability“However, a clear pattern emerged from the rubrics to give confidence that a more systematic approach may not have produced a significantly different outcome. The pattern of rubrics in the literature was established early in the review and was re-inforced as the sample was increased.”
Findings: Expressing CriteriaExpression varied considerably from rubric to rubric.Some were expressed as criteria, some were written as statements, and some were simply examples of contribution at a particular performance point.Little consistency in terminologyFor example, the quality of academic discourse was expressed in terms of “depth of insight”, “intellectual challenge”, “quality of knowledge”, and “much thought”.Subjective language was commonWords such as “good”, “substantive” and “regular” (without further definition) were often used.Language used varied in formalityWhile the majority of rubrics used formal English, some used a conversational style of language including such phrases as “Your messages didn’t leave me jazzed”.
Findings:Frequency of criteria
Findings:Criteria categorised
Findings:Most common criteriaParticipationAcademic discourseEtiquetteLearning objectivesCritical thinking
Findings: CommentaryOf the 10 categories, five broadly relate to academic standards and five relate to such things as the number of messages posted, how often students visited the forum or the attitude of the student (such as her “being positive”).Eight rubrics made direct or indirect reference to the associated learning objectives. 12 rubrics made no reference to the learning outcomes whatsoever.Eighteen rubrics rewarded participation, which was the single most common category.Four rubrics did not include any criteria relating to academic discourse whatsoever, preferring to focus solely on participation or etiquette or other non-academic criteria.Only one rubric made explicit reference to “learning objectives”.
Case studiesRovai: Online and traditional assessment: what is the difference? (2000)14 criteria for the assessment of online discussions, aligned against three grades (A-C)10 relate to the number of messages posted or the style of the message (“caring”, “compassionate”, “rude”) or other non-achievement variables.4 criteria relate to academic standards (including one that relates to the standard of spelling and grammar).Hazari: Strategy for Assessment of Online Course Discussions, (2004)20 criteria (for awarding points ranging from 1 to 5).10 relate to non-achievements such as “Demonstrates leadership in discussions” and “Posted regularly throughout the week”.Some criteria were crude: “Time between posting indicated student had read and considered substantial number of student postings before responding”
ConclusionsWe’re not sure what we’re assessingWe’re assessing the wrong thingsWe don’t know how to assess online writingWe’re not sure what online writing isWe’re unclear about how it’s differentWe don’t use rubricsWe write rubrics differentlyThe rubrics we use are pretty poorWe’re being unfair to students
Characteristics of a good rubric
RecommendationsAssess learner contributions to online discussions.Use a rubric to carry out the assessment.Design rubrics to have high fidelity.Incorporate the unique affordances of online writing into rubrics.
ChallengesEmbracing online writingAcknowledging the new affordancesReconsidering how we assess studentsCreating new assessment activitiesDesigning new marking rubricsBetter qualityImproved fidelityEmbedding the new affordances
bobby.elliott@sqa.org.uk

CAA conference 2010: Assessing online writing

  • 1.
    Assessing Online WritingCAAConference, 201020-21 July, 2010bobby.elliott@sqa.org.uk
  • 2.
  • 4.
    QuestionDo you feelconfident about assessingstudents’ online writing?
  • 5.
    Text messageText keycodeto 07624806527SmartphoneGo to poll4.com and enter keywordTwitterTweet keyword with comment to @pollWebhttp://www.polleverywhere.com/multiple_choice_polls/MTA0ODQwOTAwOA/webHow confident are you when assessing students’ online writing?Very: 24697Somewhat: 25157Not at all: 25192
  • 6.
    Summary of presentationResearchquestionMethodologyFindingsRecommendations
  • 7.
    Some background stuffDefinitionof assessmentCharacteristics of assessmentNew affordances
  • 8.
    AssessmentLatin root: assidere:“to sit beside”“Observing learning”
  • 9.
    Definition of assessment“Assessmentis an on-going process aimed at understanding and improving student learning. It involves making our expectations explicit and public; setting appropriate criteria and high standards for learning quality; systematically gathering, analysing, and interpreting evidence to determine how well performance matches those expectations and standards; and using the resulting information to document, explain, and improve performance.”
  • 10.
  • 11.
    Fidelity“Fidelity is theextent to which elements that contribute to a course grade are correctly identified as academic achievement”.“Many academics cannot help but be impressed by the prodigious time and persistence that’s some students apparently invest in producing responses to an assessment task. However, effort is clearly an input variable and therefore does not fall within the definition of academic achievement”.“Assessment in which early understandings are assessed, recorded and counted, misrepresents the level of achievement reached at the end of the course.”
  • 12.
    New AffordancesLarge-scale projectsDevelopedco-operatively and collaborativelyImproved authenticityReal life valueLarge audienceRicher contentMore reflectiveMore transparentPeer reviewed/assessed
  • 13.
    Transforming Assessment ProjectAnAustralian Learning and Teaching Council Fellowship specifically looking at the use of e-assessment within online learning environments, particularly those using one or more Web 2.0 or virtual world technologies.http://www.transformingassessment.com/geoffrey.crisp@adelaide.edu.au
  • 14.
    Challenges“Most of itsadvocates [of writing using Web 2.0 tools] offer no guidance on how to conduct assessment that comes to grips with its unique features, its difference from previous forms of student writing, and staff marking or its academic administration.”“The few extant examples appear to encourage and assess superficial learning, and to gloss over the assessment opportunities and implications of Web 2.0’s distinguishing features.”
  • 15.
    Research QuestionWhat isthe current practice in the construction of rubrics to assess online discussions?
  • 16.
    Methodology20 papers128 criteria/statements33different criteria10 categories of criterion
  • 17.
    Small PrintSmall sampleNotstatistically robustNot generalisableNo claims about validity or reliability“However, a clear pattern emerged from the rubrics to give confidence that a more systematic approach may not have produced a significantly different outcome. The pattern of rubrics in the literature was established early in the review and was re-inforced as the sample was increased.”
  • 18.
    Findings: Expressing CriteriaExpressionvaried considerably from rubric to rubric.Some were expressed as criteria, some were written as statements, and some were simply examples of contribution at a particular performance point.Little consistency in terminologyFor example, the quality of academic discourse was expressed in terms of “depth of insight”, “intellectual challenge”, “quality of knowledge”, and “much thought”.Subjective language was commonWords such as “good”, “substantive” and “regular” (without further definition) were often used.Language used varied in formalityWhile the majority of rubrics used formal English, some used a conversational style of language including such phrases as “Your messages didn’t leave me jazzed”.
  • 19.
  • 20.
  • 21.
    Findings:Most common criteriaParticipationAcademicdiscourseEtiquetteLearning objectivesCritical thinking
  • 22.
    Findings: CommentaryOf the10 categories, five broadly relate to academic standards and five relate to such things as the number of messages posted, how often students visited the forum or the attitude of the student (such as her “being positive”).Eight rubrics made direct or indirect reference to the associated learning objectives. 12 rubrics made no reference to the learning outcomes whatsoever.Eighteen rubrics rewarded participation, which was the single most common category.Four rubrics did not include any criteria relating to academic discourse whatsoever, preferring to focus solely on participation or etiquette or other non-academic criteria.Only one rubric made explicit reference to “learning objectives”.
  • 23.
    Case studiesRovai: Onlineand traditional assessment: what is the difference? (2000)14 criteria for the assessment of online discussions, aligned against three grades (A-C)10 relate to the number of messages posted or the style of the message (“caring”, “compassionate”, “rude”) or other non-achievement variables.4 criteria relate to academic standards (including one that relates to the standard of spelling and grammar).Hazari: Strategy for Assessment of Online Course Discussions, (2004)20 criteria (for awarding points ranging from 1 to 5).10 relate to non-achievements such as “Demonstrates leadership in discussions” and “Posted regularly throughout the week”.Some criteria were crude: “Time between posting indicated student had read and considered substantial number of student postings before responding”
  • 24.
    ConclusionsWe’re not surewhat we’re assessingWe’re assessing the wrong thingsWe don’t know how to assess online writingWe’re not sure what online writing isWe’re unclear about how it’s differentWe don’t use rubricsWe write rubrics differentlyThe rubrics we use are pretty poorWe’re being unfair to students
  • 25.
  • 26.
    RecommendationsAssess learner contributionsto online discussions.Use a rubric to carry out the assessment.Design rubrics to have high fidelity.Incorporate the unique affordances of online writing into rubrics.
  • 27.
    ChallengesEmbracing online writingAcknowledgingthe new affordancesReconsidering how we assess studentsCreating new assessment activitiesDesigning new marking rubricsBetter qualityImproved fidelityEmbedding the new affordances
  • 28.

Editor's Notes

  • #2 Bobby Elliott from SQA.Happy to receive e-mails after the event.Small piece of research done as part of a larger research project.Interesting results.30 slides in 30 minutes...
  • #3 Paper is in the book.Also available online.Paper focuses on online discussion boards but my presentation is more general and relates to online writing.
  • #4 3 rules of speaking…Never do a lecture for the first time.Never tell jokes.Never use technology – especially online tech.Serious message in this slide… teachers are becoming more accountable.
  • #5 Quick survey before we begin.
  • #6 Text, phone, tweet or web.This is an example of a free web 2 service.
  • #7 High on criticism. Low on solution. My speciality.
  • #8 Before we begin, some background stuff…
  • #9 The original Latin derivative emphasises the formative nature of assessment.The second definition is more modern and is a nice, simple explanation of the word.
  • #10 Best definition of assessment that I have came across.By Angelo, 1995.Notice the bits in bold…“making expectations explicit”, “having criteria”, “systematically analysing”…Support for the use of a rubric.
  • #11 You will be familiar with two of these characteristics (validity and reliability).But fidelity is new.The idea was introduced by a guy called Sadler (2009) in a paper about assessment.
  • #12 Definition in bold.It’s very similar to validity but focuses on grading.It’s a moot point if we need it but I think it adds something to our vocabulary.Mention effort – we are guilty of this. He call this a “non-achievement”.Give audience time to read the slide.Criticism of continuous assessment. Maybe particularly harmful to boys.
  • #13 Several recent papers have referred to the new “affordances” of online writing.This slide shows what’s special about writing online compared to traditional writing.The major initiative is an Australian project entitled “Transforming Assessment”. See next slide.The new affordances are the new capabilities that online writing provides.Go thru each one.
  • #15 In summary: we’re not using the new capabilities that Web 2 provides.We’re assessing it in a business-as-usual way.Give example of online discussions assessed through “your best two posts” – essay-type contributions.We’re assessing online writing like we assess offline writing.
  • #16 My research was focused on online discussion boards.
  • #17 I looked at 20 rubrics.Consisting of 128 separate criteria (around 6-7 per rubric).Many criteria were basically the same thing expressed in different ways…So I grouped them into 33 different criteriaThen I coded these criteria using a coding system with 10 types of criteria.
  • #18 My sample was small.And the selection of the rubrics are not random or systematic.So I’m not saying much about validity or reliability of my findings.BUT the more I looked at, the more a pattern emerged.In fact, the pattern emerged very early in the review.
  • #19 So here are my findings.
  • #20 This is the table that shows the original 128 criteria grouped into different criteria (similar criteria are grouped).It reduced the 128 criteria to 33.For example, of the 128 criteria in the 20 rubrics, 12 related to quantity of messages – probably 12 of the 20 rubrics had one criterion relating to quantity.Notice some interesting categories: frequency of posting (9 criteria), attitude (7 criteria), showing respect (5 criteria).
  • #21 Here are the 10 codes that I used.So, for example, 18 rubrics (out of 20) made mention of participation (e.g. the number of posts) and this accounted for 36 of the 128 criteria (around one quarter).Also, 8 rubrics had criteria relating to etiquette (manners) – and this accounted for 10 (of 128) criteria.
  • #22 This slide shows the top five criteria.Notice that two relate to “non-achievements” to use Sadler’s terminology.The most common criterion related to participation: how often students posted or how many replies they made or how many times they logged in.Criteria relating to etiquette were the 3rd most common! This means being polite or supportive or “nice”!We would never reward “being nice” in an essay – so why online? Is “being nice” a learning objective?
  • #23 Let audience read my findings.18/20 rubrics rewarded participation.Few (1) were directly related to learning objectives.
  • #24 Rovai’s paper is probably the best known and most widely use.But Sadler would give him a poor grade for fidelity… 4/14… 30%Another well known author in this area is Hazari. He does better achieving 50%.But these papers are sort of seminal works on writing rubrics… but see the last bullet!It’s not “time between posting” that measures quality.
  • #25 The present position is not satisfactory.
  • #26 I’ve tried to illustrate best practice with this list of characteristics.But I haven’t actually created a rubric!Some characteristics are obvious e.g. 3.Nos 2 and 4 and 6 and 8 are more interesting.
  • #27 These recommendations are from my paper.
  • #28 The current position is not good.We have to try to improve how we assess online writing.We need to be more consistent.We need to embrace the new affordances of digital writing.We need to reconsider the assessment tasks we set students.And how to mark these – using rubrics.