Your SlideShare is downloading. ×
C:value added
Upcoming SlideShare
Loading in...5

Thanks for flagging this SlideShare!

Oops! An error has occurred.

Saving this for later? Get the SlideShare app to save on your phone or tablet. Read anywhere, anytime – even offline.
Text the download link to your phone
Standard text messaging rates apply

C:value added


Published on

Value-added assessment

Value-added assessment

Published in: Education, Technology
1 Like
  • Be the first to comment

No Downloads
Total Views
On Slideshare
From Embeds
Number of Embeds
Embeds 0
No embeds

Report content
Flagged as inappropriate Flag as inappropriate
Flag as inappropriate

Select your reason for flagging this presentation as inappropriate.

No notes for slide


  • 1. Value-Added
    Carol Thompson
    July 2010
  • 2. Where we are:
    Adequate Yearly Progress (AYP)
    NCLB measures school performance based on AYP, which calculates the percentage of various student populations that annually met or exceed the state’s academic standards, toward a universal fixed point.
  • 3. No Child Left Behind
    Set measureable goals
    Reduce achievement gap among subpopulations
    Federal funding provided
    All students should reach proficiency standards in reading and math within 12 years
  • 4. With NCLB
    States can choose own testing.
    Standardized tests measure achievement levels, but not change.
    Systematic evaluation is necessary to determine which educational approaches are working and which are not.
  • 5. Assumptions
    IF . . .
    good schools are those that have high test scores, . . .
    then . . .
  • 6. Assumptions
    Bad schools are those that have low test scores.
  • 7. Assumptions
    Students with higher scores are learning more.
    HOWEVER . . .
    Schools that admit students with low scores and raise them to average are better schools than those that admit high scoring students and graduate them at the same level.
  • 8. comparison
    Under NCLB we make comparisons across cohorts.
    Example: This year’s reading scores in 5th grade are higher than last year’s reading scores. Does that show improvement or were this year’s students ahead of last year’s students?
    Value-added measures the progress of each child where they are as they progress through school.
    Value-added measures
    not just
  • 9. Purpose of Value-Added Data
    The value-added approach focuses on changes in test scores over time, rather than on a single test score at a given moment.
    Inequity of standardized-results:
    Some schools doing a good job of teaching severely disadvantaged students could be sanctioned
    Some poor schools ‘shepherding’ top students could be imitated,
    Some excellent schools using effective strategies to help high-risk students may NOT be imitated due to below average test scores.
    (Crane, 2002)
  • 10. Value added
    Value-added Assessment
    Analysis of student achievement that uses longitudinal student achievement data to obtain estimates of the impact schools or teachers have on student learning as measured by test scores. It measures individual student progress from a relative starting point.
  • 11. Value-added
    Levels the playing field,.
    Calculates a projected test score for a student in a given grade and subject, and
    Bases the projected score on the student’s prior academic achievement.
  • 12. Value–Added
    Is a way of analyzing test data that measures both teaching and learning,
    Allows us to see how teachers add to a student’s knowledge over and above what their families and communities do, and
    Separates student effects from school effects.
  • 13. Value-added
    Achievement test data is available for each student each year,
    Highly correlated scales of curricular objectives are used,
    The scales have sufficient stretch to measure progress of previously low and high scoring students, and
    Scales have appropriate repeatabiities,
    All data that meet these conditions can be used in a value-added assessment system, regardless of test source. (Crane, 2002)
  • 14. School effectivenes
    The “differences in schooling effectiveness is the dominant factor affecting the speed that students move up the ‘ramp’ of curriculum.” (Sanders, 2000).
  • 15. Teacher quality
    “An accountability system based on the academic progress of student populations is one that will hold people accountable for things over which they have control, rather than for things over which they do not. Teachers do have primary control of the rate of academic progress of their students. (Sanders, 2000)
  • 16. Achievement vs. growth
    Achievement is best predicted by family income.
    Growth is best predicted by the quality of instruction. Teachers have control over the quality of their instruction.
  • 17. Classroom patterns
    Disproportionate attention is given in classrooms to ‘bubble kids’—those who score just below the state standard for NCLB; more difficult kids get help as a result.
    Research from Tennessee revealed three classroom patterns, shed, reverse shed, and tepee. The names reflect the shape of the slope.
    Even though these patterns can be found in any classroom, they occur disproportionately in some circumstances.
    Data is based on three year averages.
  • 18. Shed pattern
    Typically found in low-income communities.
    Elementary teachers focus on the low achievers in order to get high gains while previous high-achievers get low gains.
  • 19. Shed pattern
    If this pattern continues through elementary, there will be few high achieving children by middle school.
  • 20. Reverse shed pattern
    Opposite focus found in high income communities.
    Teachers concentrate on their highest performers.
  • 21. Reverse shed pattern
    Low achievers get low gains while previous high achievers continue to get high gains.
  • 22. Tepee pattern
    Prevalent in most classrooms in most communities.
    Teachers focus on the average student; also called teaching to the middle.
  • 23. Tepee pattern
    This results in previously average achievers getting high gains and both previously low and high achievers getting low gains.
  • 24. Teaching to the Middle
    Biggest impediment to higher achievement are the years where individual students aren’t making realistic growth. A pattern can exist whereby lower scoring students are given opportunities to make progress but earlier high achieving students are held to the same pace and place as their lower achieving peers. As this pattern repeats over the years it becomes a self-fulfilling prophecy that these students lose ground (Sanders, 2002).
  • 25. Teaching to the Middle
    Value-added provides valuable data for teachers to use in focusing their instructional practices with planned intentionality.
    As educators we need to remember that high achieving students come from housing projects, remote rural areas, and million dollar homes.
  • 26. Teacher variability
    Once again, data suggests that teacher quality may be the single most important in-school factor determining how much students learn.
    (Crane, 2002, Sanders & Rivers, 1996; Jordan, Mendro, & Weerasinghe, 1997; Haycock, 1998)
  • 27. Teacher Quality
    The Sanders and Rivers study found that good teachers raised students math scores at least 2-3 years into the future (regardless of class size)
    The Sanders study (1991-1995) found the top 1/5 of teachers raised students’ achievement test scores 39 percentile points more than teachers of bottom 1/5. (These effects were consistent whether heterogeneous or homogeneous grouping or excelling vs. struggling students.)
  • 28. Teacher Quality
    Layered model—all student achievement data for 5 years is used simultaneously over all subjects tested, linking each year’s data to the current and previous teachers.: teachers given credit for teaching beyond standards assessed as well as not held accountable for one year’s poor performance due to student illness, drug involvement, etc.
    Accommodates ‘real world’ situations; different modes of instruction such as self-contained, team teaching, fractured student records, and data from non-vertically scaled tests.
  • 29. Teacher Variability
    Struggling students aren’t randomly distributed into classrooms. They are found disproportionately in classrooms where they receive poor instruction.
  • 30. Teacher variability
    The best way to improve test scores is to improve teacher quality. This can be done by applying lessons learned from value-added analysis to teacher education and professional development.
    Crane, 2002
  • 31. Teacher variability
    Personnel decisions can also be made given value-added data.
    Incentives could be offered.
    Help offered to teachers who raised test scores the least.
    Teachers promoted to instructional leadership positions who raised scores the most.
    Over time, lowest performing teachers would leave the profession.
  • 32. Teacher effectiveness
    “Of all the factors we study—class size, ethnicity, location, poverty—these pale to triviality in the face of teacher effectiveness.” Sanders address to Metropolitan School Board in Nashville, 2010.
  • 33. Sustained growth over time
    By having sustained growth over time, we will begin to close the gap and all students will still be making gains.
  • 34. Sustained Academic Growth
    Data is not the enemy but a valuable tool in an educator’s toolbox they can use to fine tune instruction to provide the best opportunity for every child. (Sanders, 2000)
  • 35. Advantages of value-added
    Tests don’t have to be nearly so closely aligned as they do when judgments are made from a single year of test results each year.
    Reporting of simple test averages is singularly inaccurate. Disaggregation of socioeconomic strata is an improvement on raw score reporting.
    Use of regression models are sometimes labeled as ‘value-added’. (Sanders, 2000)
  • 36. Additional information
    A value-added analysis is statistical and there is always a margin of error: they are better able to make distinctions at either end with more accuracy than the middle.
    Distinguishing between true learning and teaching to the test: can limit effects by changing the tests frequently.
  • 37. Local example of value-added
    Dallas measures individual student progress from a relative starting point. It measures only that knowledge that the school/teacher is responsible for imparting, calculating the value-added by each school. (Toch, 2005)
  • 38. Benefits of Value-Added
    Provide a more accurate picture of school, districts, and states success/failure.
    Generate objective measures of teacher performance that could be used to raise teacher quality.
    Provide useful data for evaluating school reform programs (Crane, 2002).
    Use of up to five years data for each student with different subjects tested and item analysis (40+ each subtest) could result in over 1,000 items collectively contributing to information array for that student; not a snapshot in time (Sanders, 2000).
  • 39. Provides data
    To improve data-driven decision making
    To build professional learning communities
    To differentiate instruction
    To measure through growth, not just achievement
  • 40. Resources
    Crane, J. (2002). The promise of value-added testing. Policy Report from
    Doran H. & Fleischman, S. (2005). Challenges of value-addedassessment. Educational Leadership. 63:3
    Hershberg, T. (2004). Value-added assessment. The Center for Greater
    Holland, R. (2010). How to build a better teacher. Hoover Institution Stanford University.
    Sanders, W. (2000). Value-added assessment from student achievement data: opportunities and hurdles. Journal of Personnel Evaluation in Education 14:4.
    Schneider, C. (2002). Can value added assessment raise the level of
    student accomplishment? Peer Review.
    Toch, T. (2005). Measure for measure. Washington Monthly.
    Zurawsky, C. (2004). Teachers matter: Evidence from value-added
    assessments. Research Points. 2:2
    (2007). Roundtable discussion on value-added analysis of student achievement: a summary of findings. Working Group on Teacher Quality