SlideShare a Scribd company logo
From a 
Superintendent’s 
Perspective: 
Using data wisely 
Andy Hegedus, Ed.D. 
September 2014
Trying to gauge my audience 
and adjust my speed . . . 
• How many of you think your literacy with 
assessments is “Good” or better? 
• How many of you have a fine tuned 
assessment program? 
• How many of you think your practical 
knowledge about using data for systemic 
improvement is “Good” or better?
The Bottom Line First: 
Here’s what you must have 
• An adequate depth of knowledge to ask really 
good questions and the tenacity to do so 
• A person willing to “Speak the Truth, With 
Love, To Power” 
• A big lever – 
“What gets measured (and attended to), 
gets done” 
The measurement and reinforcement system 
is your responsibility
My Purpose 
• Increase your understanding about 
various urgent assessment related topics 
– Ask better questions 
– Useful for making all types of decisions with 
data
Three main topics 
• Assessment basics + 
• Improving your assessment program 
• Data culture
Go forth thoughtfully 
with care 
• What we’ve known to be true is now being 
shown to be true 
– Using data thoughtfully improves student 
achievement and growth rates 
– 12% mathematics, 13% reading 
• There are dangers present however 
– Unintended Consequences 
Slotnik, W. J. , Smith, M. D., It’s more than money, February 2013, retrieved from 
http://www.ctacusa.com/PDFs/MoreThanMoney-report.pdf
Remember the old adage? 
“What gets measured (and attended to), 
gets done”
• NCLB 
An infamous example 
– Cast light on inequities 
– Improved performance of “Bubble Kids” 
– Narrowed taught curriculum 
The same dynamic happens 
inside your schools
A patient’s health 
doesn’t change 
because we know 
their blood pressure 
It’s our response that 
makes all the 
difference 
It’s what we do that counts
Be considerate of the continuum 
of stakes involved 
Support 
Compensate 
Terminate 
Increasing levels of required rigor 
Increasing risk
Assessment basics 
(in a teacher evaluation frame)
Three primary conditions 
1. Alignment between the content assessed and 
the content to be taught 
2. Selection of an appropriate assessment 
• Used for the purpose for which it was designed 
(proficiency vs. growth) 
• Can accurately measure the knowledge of all 
students 
• Adequate sensitivity to growth 
3. Adjust for context/control for factors outside a 
teacher’s direct control (value-added)
Two approaches we like 
1. Assessment results used 
wisely as part of a 
dialogue to help teachers 
set and meet challenging 
goals 
2. Use of tests as a “yellow 
light” to identify teachers 
who may be in need of 
additional support or are 
ready for more
What question is being answered in support of 
using data in evaluating teachers? 
Is the progress produced 
by this teacher 
dramatically different 
than teaching peers who 
deliver instruction to 
comparable students in 
comparable situations?
Marcus’ growth 
College readiness standard 
Marcus Normal Growth Needed Growth
There are four key steps required to 
The Test 
answer this question 
Top-Down Model 
The Growth Metric 
The Evaluation 
The Rating
Let’s begin at the beginning 
The Test 
The Growth Metric 
The Evaluation 
The Rating
What is measured should be 
aligned to what is to be taught 
3rd Grade 
ELA 
Teacher? 
3rd Grade 
ELA 
Standards 
3rd Grade 
Social 
Studies 
Teacher? 
1. Answer questions to demonstrate 
understanding of text…. 
2. Determine the main idea of a 
Elem. Art 
Teacher? 
text…. 
3. Determine the meaning of general 
academic and domain specific 
words… 
Would you use a general 
reading assessment in the 
evaluation of a…. 
~30% of teachers teach in tested subjects and grades 
The Other 69 Percent: Fairly Rewarding the Performance of Teachers of Nontested Subjects and Grades, 
http://www.cecr.ed.gov/guides/other69Percent.pdf
What is measured should be 
aligned to what is to be taught 
• Assessments should align with the 
teacher’s instructional responsibility 
– Specific advanced content 
• HS teachers teaching discipline specific content 
– Especially 11th and 12th grade 
• MS teachers teaching HS content to advanced students 
– Non-tested subjects 
• School-wide results are more likely “professional 
responsibility” rather than reflecting competence 
– HS teachers providing remedial services
The purpose and design of the 
instrument is significant 
• Many assessments are 
not designed to 
measure growth 
• Others do not measure 
growth equally well for 
all students
Let’s ensure we have similar 
meaning 
Adult 
Reading 
5th Grade 
x 
Beginning Literacy 
x 
Time 1 Time 2 
Status 
Two assumptions: 
1. Measurement accuracy, 
and 
2. Vertical interval scale
Accurately measuring 
growth 
depends on 
accurately measuring 
achievement
What does it take to accurately 
measure achievement? 
Questions 
surrounding the 
student’s 
achievement level 
The more 
questions the 
merrier
Teachers encounter a distribution 
of student performance 
Adult 
Reading 
5th 
Grade 
x 
x 
x 
x 
x x x 
x 
x 
x 
x 
x 
x 
x 
x 
Beginning Literacy 
Grade Level 
Performance
Adaptive testing works differently 
Item bank can 
span full 
range of 
achievement
How about accurately 
measuring height? 
What if the yardstick 
stopped in the middle of 
his back?
Items available need to match student 
ability 
California STAR NWEA MAP
How about accurately 
measuring height? 
What if we could only 
mark within a pre-defined 
six inch range?
5th Grade 
Level Items 
These differences impact 
measurement error 
.12 
.10 
.08 
.06 
.04 
.02 
.00 
Information 
190 200 210 220 230 240 260 
Scale Score 
Fully 
Adaptive Test 
Significantly 
Different Error 
250 
Constrained 
Adaptive or 
Paper/Pencil 
Test
To determine growth, 
achievement 
measurements 
must be related through 
a scale
Let’s measure height again 
If I was measured as: 
5’ 9” 
And a year later I was: 
1.82m 
Did I grow? 
Yes. ~ 2.5” 
How do you know?
Traditional assessment uses items 
reflecting the grade level standards 
Adult 
Reading 
6th Grade 
5th Grade 
4th Grade 
Grade Level Standards 
Beginning Literacy 
Traditional 
Assessment Item Bank
Traditional assessment uses items 
reflecting the grade level standards 
Adult 
Reading 
6th Grade 
5th Grade 
4th Grade 
Grade Level Standards 
Grade Level Standards 
Beginning Literacy 
Overlap allows 
linking and scale 
construction 
Grade Level Standards
Error can change your life! 
• Think of a high stakes test – 
State Summative 
– Designed mainly to identify if a student is proficient or 
not 
• Do they do that well? 
• 93% correct on Proficiency determination 
• Does it go off design well? 
• 75% correct on Performance Levels determination 
*Testing: Not an Exact Science, Education Policy Brief, Delaware Education Research & Development Center, May 2004, 
http://dspace.udel.edu:8080/dspace/handle/19716/244
Using tests in high stakes ways 
creates new dynamic 
• Tests specifically designed to inform classroom 
instruction and school improvement in 
formative ways 
No incentive in the system for 
inaccurate data
8.00 
6.00 
4.00 
2.00 
0.00 
-2.00 
-4.00 
-6.00 
New phenomenon when used as part of 
a compensation program 
Mean value-added growth by school 
1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69 71 
Students taking 10+ minutes longer spring than fall All other students
Other consequence 
When teachers are evaluated 
on growth using a once per 
year assessment, one teacher 
who cheats disadvantages the 
next teacher
Lessons? 
• What were some things you learned? 
• What practices do you want to reinforce? 
• What do you need to do differently? 
• Think – Pair 
– 2 min to make some notes 
– 3 min to share with a neighbor
Testing is complete . . . 
What is useful to answer our question? 
The Test 
The Growth Metric 
The Evaluation 
The Rating
The problem with spring-spring 
testing 
Teacher 1 Summer Teacher 2 
4/14 5/14 6/14 7/14 8/14 9/14 10/14 11/14 12/14 1/15 2/15 3/15 4/15
A better approach 
• When possible use a spring – fall – spring 
approach 
• Measure summer loss and incentivize schools 
and teachers to minimize it 
• Measure teacher performance fall to spring, 
giving as much instructional time as possible 
between assessments 
• Monitor testing conditions to minimize 
gaming of fall or spring results
100 
90 
80 
70 
60 
50 
40 
30 
20 
10 
0 
Grade 2 Grade 3 Grade 4 Grade 5 Grade 6 Grade 7 Grade 8 
Reading 
Math 
The metric matters - 
Let’s go underneath “Proficiency” 
Difficulty of New York Proficient Cut Score 
National Percentile 
College 
Readiness 
New York Linking Study: A Study of the Alignment of the NWEA RIT Scale with the New York State (NYS) Testing Program, November 2013
The metric matters - 
Let’s go underneath “Proficiency” 
Dahlin, M. and Durant, S., The State of Proficiency, Kingsbury Center at NWEA, July 2011
What actually happened? 
Estimated Proficiency Rates For Six NY Districts 
With Proficiency Cut Scores Changed 
87 
4th Grade Mathematics 
55 
100 
80 
60 
40 
20 
0 
-20 
-40 -31 
2012 2013 Reported Change
What actually happened? 
Estimated Proficiency Rates 
For Six NY Districts 
4th Grade Mathematics 
With Proficiency Cut Scores 
87 
Changed 
55 
100 
80 
60 
40 
20 
0 
-20 
-40 -31 
2012 2013 Reported Change 
46 
55 
10 
60 
50 
40 
30 
20 
10 
0 
Estimated Proficiency Rates 
For Six NY Districts 
4th Grade Mathematics 
With 2013 Proficiency Cut 
Scores Applied 
2012 2013 Actual Change
Number of Students 
What gets measured and attended to 
Mathematics 
Fall RIT 
No Change 
Down 
Up 
really does matter 
Proficiency College Readiness 
One district’s change in 5th grade mathematics performance 
relative to the KY proficiency cut scores
Number of Students 
Changing from Proficiency to Growth 
means all kids matter 
Mathematics 
Student’s score in fall 
Below projected 
growth 
Met or above 
projected growth 
Number of 5th grade students meeting projected 
mathematics growth in the same district
Lessons? 
• What were some things you learned? 
• What practices do you want to reinforce? 
• What do you need to do differently? 
• Think – Pair – Share 
– 2 min to make some notes 
– 3 min to share with a neighbor
How can we make it fair? 
The Test 
The Growth Metric 
The Evaluation 
The Rating
Without context what is 
“Good”? 
Scale Norms Study 
Adult 
Literacy 
Beginning 
Reading 
National 
Percentile 
College Readiness 
ACT 
Benchmarks 
State Test 
Performance Levels 
“Meets” 
Proficiency 
Common 
Core 
Proficient 
Performance Levels
Normative data for growth is a 
bit different 
Fall 
Score 
Basic 
Factors 
Grade: 5th 
Subject: 
Reading 
7 
points 
FRL vs. non-FRL? 
IEP vs. non-IEP? 
ESL vs. non-ESL? 
Starting 
Achievement 
Instructional 
Weeks 
Typical growth 
Outside of a teacher’s direct control
A Visual Representation of 
Value Added 
Spring 5th Grade 
Test 
Student A 
Spring Score 209 
Score 207 
(Average Spring Score for Similar 
Students) 
Value Added 
(+2 Score) 
Student A 
Fall Score 200 
Fall 5th Grade 
Test
Consider . . . 
• What if I skip this step? 
– Comparison is likely against normative data 
so the comparison is to “typical kids in 
typical settings” 
• How fair is it to disregard context? 
– Good teacher – bad school 
– Good teacher – challenging kids
Value-added is science 
• Value added models can control for a variety 
of classroom, school level, and other 
conditions 
– Proven statistical methods 
– All attempt to minimize error 
– Variables outside controls are assumed as random
A variety of errors means more 
stability only at the extremes 
• Control for measurement 
error 
– All models attempt to address 
this issue 
• Population size 
• Multiple data points 
– Error is compounded with 
combining two test events 
– Many teachers’ value-added 
scores will fall within the range 
of statistical error
12.00 
11.00 
10.00 
9.00 
8.00 
7.00 
6.00 
5.00 
4.00 
3.00 
2.00 
1.00 
0.00 
-1.00 
-2.00 
-3.00 
-4.00 
-5.00 
-6.00 
-7.00 
-8.00 
-9.00 
-10.00 
-11.00 
-12.00 
Average Growth Index Score and Range 
Range of teacher value-added 
estimates 
Mathematics Growth Index Distribution by Teacher - Validity Filtered 
Q5 
Q4 
Q3 
Q2 
Q1 
Each line in this display represents a single teacher. The graphic 
shows the average growth index score for each teacher (green 
line), plus or minus the standard error of the growth index estimate 
(blue line). We removed students who had tests of questionable 
validity and teachers with fewer than 20 students.
With one teacher, 
error means a lot
Assumption of randomness can 
have risk implications 
• Value-added models assume that variation is 
caused by randomness if not controlled for 
explicitly 
– Young teachers are assigned disproportionate 
numbers of students with poor discipline records 
– Parent requests for the “best” teachers are 
honored 
• Sound educational reasons for placement are 
likely to be defensible
Instability at the tails of the 
distribution 
“The findings indicate that these modeling 
choices can significantly influence outcomes for 
individual teachers, particularly those in the tails 
of the performance distribution who are most 
likely to be targeted by high-stakes policies.” 
Ballou, D., Mokher, C. and Cavalluzzo, L. (2012) Using Value-Added Assessment for Personnel 
Decisions: How Omitted Variables and Model Specification Influence Teachers’ Outcomes. 
LA Times Teacher #1 
LA Times Teacher #2
How tests are used to evaluate 
teachers 
The Test 
The Growth Metric 
The Evaluation 
The Rating
Translation into ratings can be 
difficult to inform with data 
• How would you 
translate a rank order 
to a rating? 
• Data can be provided 
• Value judgment 
ultimately the basis 
for setting cut scores 
for points or rating
Decisions are value based, 
not empirical 
• What is far below a 
district’s expectation is 
subjective 
• What about 
• Obligation to help 
teachers improve? 
• Quality of replacement 
teachers?
Even multiple measures need to 
be used well 
• System for combining elements and 
producing a rating is also a value based 
decision 
– Multiple measures and principal judgment 
must be included 
– Evaluate the extremes to make sure it 
makes sense
Leadership Courage Is A Key 
5 
4 
3 
2 
1 
0 
Ratings can be driven by the assessment 
Teacher 1 Teacher 2 Teacher 3 
Observation Assessment 
Real 
or 
Noise?
Big Message 
If evaluators do not differentiate 
their ratings, 
then all differentiation comes from 
the test
Lessons? 
• What were some things you learned? 
• What practices do you want to reinforce? 
• What do you need to do differently? 
• Think – Pair – Share 
– 2 min to make some notes 
– 3 min to share with a neighbor 
– 2 min for two report outs on anything so far
Improving your 
assessment program
• Read the sheet 
and highlight 
anything 
interesting to 
you 
Let’s Define 
Types of Assessments
The pursuit of compliance is exhausting 
because it is always a moving target. 
Governors move on, the party in power 
gets replaced, a new president is elected, 
and all want to put their own stamp on 
education. 
It is saner and less exhausting to define 
your own course and align compliance 
requirements to that.
Seven standards that define the 
purpose driven assessment system
The purposes of all assessments are 
defined and the assessments are valid 
1 and useful for their purposes
Teachers are educated in the proper 
administration and application of the 
assessments used in their classrooms 2
Redundant, mis-aligned, or unused 
3assessments are eliminated
Assessment results are aligned to 
the needs of their audiences 4
Assessment results are delivered in a 
timely and useful manner 5
The metrics and incentives used 
encourage a focus on all learners 6
The assessment program contributes 
to a climate of transparency and 
objectivity with a long-term focus 7
1. Typical assessment 
purposes 
• Identify student learning needs 
• Identify groupings of students for instruction 
• Guide instruction 
• Course placement 
• Determine eligibility for programs 
• Award credits and/or assign grades 
• Evaluate proficiency 
• Monitor student progress 
• Predict proficiency 
• Project achievement of a goal 
• Formative and summative evaluation of programs 
• Formative evaluation to support school and teacher improvement 
• Report student achievement, growth, and progress to the 
community and stakeholders 
• Summative evaluation of schools and teachers
1. Assessment Purpose Survey 
To increase value… 
Identify gaps between: 
1. How critical is this data 
to your work? 
2. How do you actually 
use this data? 
Take 10 min to fill this 
out and 5 min to pair and 
discuss areas of biggest 
gap
Compare 
assessments and 
their purposes to 
find unnecessary 
overlaps 
Take 10 min to 
fill this out and 5 
min to pair and 
discuss areas of 
redundancy 
3. Eliminate waste
Lessons? 
• What were some things you learned? 
• What practices do you want to reinforce? 
• What do you need to do differently? 
• Think – Pair – Share 
– 2 min to make some notes 
– 3 min to share with a neighbor 
– 2 min for two report outs on this section
Data Culture
Education Organizations 
Mature 
Poor to 
Fair 
Fair to 
Good 
Good to 
Great 
Great to 
Excellent 
Achieving the 
basics of literacy 
and numeracy 
Getting the 
foundations 
in place 
Shaping the 
professional 
Improving 
through 
peers and 
innovation 
Data use does too 
Barber, M., Chijioke, C., & Mourshed, M. (2011). How the world’s most improved school systems keep getting better. McKinsey & Company.
Data Use Continuum 
Poor to 
Fair 
Fair to 
Good 
Good to 
Great 
Great to 
Excellent 
One on One 
Within 
Teams 
Within 
the Walls 
Across the 
Walls 
Requires a shift in the culture
What would you do with this?
Reflection Time 
• Where are your pockets of most maturity? 
• Least maturity? 
• What is causing the differences? 
Think - 2 min
Research on data use in 
school improvement 
• Education problems are “Wicked” 
– Problem boundaries are ill-defined 
– No definitive solutions 
– Highly resistant to change 
– Problem and solutions depend on 
perspective 
– Changes are consequential 
Data can only take you so far
Research on data use in 
school improvement 
• Use data as a platform for deeper 
conversations 
• Define your problem well 
– Problem title and description 
– Magnitude 
– Location 
– Duration
Research on data use in 
school improvement 
• Part of a continuous improvement 
process 
–Data conversations 
• Collaborative 
• Embedded in culture 
• Structured process 
• Love, Nancy – Using data to improve learning for all 
• Lipton & Wellman – Got data? Now what? 
• NWEA Data Coaching
Final Question
What are your biggest 
take-aways? 
• Think – 2 min 
• Pair – 3 min 
• Share – 2 min - Two people
Last thing 
• Materials in your Conference App 
Presentation available on Slideshare.net
More information 
Contacting me: 
E-mail: andy.hegedus@nwea.org 
Contacting us: 
Exhibit Booth or Liz Kaplan

More Related Content

What's hot

Ed Reform Lecture - University of Arkansas
Ed Reform Lecture - University of ArkansasEd Reform Lecture - University of Arkansas
Ed Reform Lecture - University of ArkansasJohn Cronin
 
Teacher evaluation presentation3 mass
Teacher evaluation presentation3  massTeacher evaluation presentation3  mass
Teacher evaluation presentation3 mass
John Cronin
 
Teacher evaluation and goal setting connecticut
Teacher evaluation and goal setting   connecticutTeacher evaluation and goal setting   connecticut
Teacher evaluation and goal setting connecticut
John Cronin
 
Using Assessment data
Using Assessment dataUsing Assessment data
Using Assessment data
Dublin City University
 
Data and assessment powerpoint presentation 2015
Data and assessment powerpoint presentation 2015Data and assessment powerpoint presentation 2015
Data and assessment powerpoint presentation 2015
Erica Zigelman
 
Assessment literacy
Assessment literacyAssessment literacy
Assessment literacy
Vandy PacettiDonelson
 
Data Summer
Data SummerData Summer
Data Summer
mark.richardson
 
Improving student learning through programme assessment
Improving student learning through programme assessmentImproving student learning through programme assessment
Improving student learning through programme assessment
Tansy Jessop
 
McMulen Data Vision Action 11.14.13
McMulen Data Vision Action 11.14.13McMulen Data Vision Action 11.14.13
McMulen Data Vision Action 11.14.13
Glenn E. Malone, EdD
 
Teacher evaluations using technology
Teacher evaluations using technologyTeacher evaluations using technology
Teacher evaluations using technology
Richard Voltz
 
Using Data for Continuos School Improvement
Using Data for Continuos School ImprovementUsing Data for Continuos School Improvement
Using Data for Continuos School Improvement
lindamtz88
 
Presentations morning session 22 January 2018 HEFCE open event “Using data to...
Presentations morning session 22 January 2018 HEFCE open event “Using data to...Presentations morning session 22 January 2018 HEFCE open event “Using data to...
Presentations morning session 22 January 2018 HEFCE open event “Using data to...
Bart Rienties
 
Seven purposes presentation
Seven purposes presentationSeven purposes presentation
Seven purposes presentation
John Cronin
 
Equity Audit
Equity AuditEquity Audit
Equity AuditMatt Hise
 
Study on students dropouts in advanced certificate in pre school education pr...
Study on students dropouts in advanced certificate in pre school education pr...Study on students dropouts in advanced certificate in pre school education pr...
Study on students dropouts in advanced certificate in pre school education pr...
The Open University of Sri Lanka
 
Teacher Evaluation Rubrics
Teacher Evaluation Rubrics Teacher Evaluation Rubrics
Teacher Evaluation Rubrics
Haroon Baig
 
Coaching as part of teacher evaluation process v2.0
Coaching as part of teacher evaluation process v2.0Coaching as part of teacher evaluation process v2.0
Coaching as part of teacher evaluation process v2.0Richard Voltz
 
Inaugural lecture
Inaugural lectureInaugural lecture
Inaugural lecture
Tansy Jessop
 

What's hot (20)

Ed Reform Lecture - University of Arkansas
Ed Reform Lecture - University of ArkansasEd Reform Lecture - University of Arkansas
Ed Reform Lecture - University of Arkansas
 
Teacher evaluation presentation3 mass
Teacher evaluation presentation3  massTeacher evaluation presentation3  mass
Teacher evaluation presentation3 mass
 
Teacher evaluation and goal setting connecticut
Teacher evaluation and goal setting   connecticutTeacher evaluation and goal setting   connecticut
Teacher evaluation and goal setting connecticut
 
Using Assessment data
Using Assessment dataUsing Assessment data
Using Assessment data
 
Data and assessment powerpoint presentation 2015
Data and assessment powerpoint presentation 2015Data and assessment powerpoint presentation 2015
Data and assessment powerpoint presentation 2015
 
Oerc june 2014 final ppt combined
Oerc june 2014 final ppt combinedOerc june 2014 final ppt combined
Oerc june 2014 final ppt combined
 
Assessment literacy
Assessment literacyAssessment literacy
Assessment literacy
 
Data Summer
Data SummerData Summer
Data Summer
 
Improving student learning through programme assessment
Improving student learning through programme assessmentImproving student learning through programme assessment
Improving student learning through programme assessment
 
McMulen Data Vision Action 11.14.13
McMulen Data Vision Action 11.14.13McMulen Data Vision Action 11.14.13
McMulen Data Vision Action 11.14.13
 
Teacher evaluations using technology
Teacher evaluations using technologyTeacher evaluations using technology
Teacher evaluations using technology
 
Using Data for Continuos School Improvement
Using Data for Continuos School ImprovementUsing Data for Continuos School Improvement
Using Data for Continuos School Improvement
 
Presentations morning session 22 January 2018 HEFCE open event “Using data to...
Presentations morning session 22 January 2018 HEFCE open event “Using data to...Presentations morning session 22 January 2018 HEFCE open event “Using data to...
Presentations morning session 22 January 2018 HEFCE open event “Using data to...
 
Seven purposes presentation
Seven purposes presentationSeven purposes presentation
Seven purposes presentation
 
Non-cognitive Skills_CES_FF_161203_Final
Non-cognitive Skills_CES_FF_161203_FinalNon-cognitive Skills_CES_FF_161203_Final
Non-cognitive Skills_CES_FF_161203_Final
 
Equity Audit
Equity AuditEquity Audit
Equity Audit
 
Study on students dropouts in advanced certificate in pre school education pr...
Study on students dropouts in advanced certificate in pre school education pr...Study on students dropouts in advanced certificate in pre school education pr...
Study on students dropouts in advanced certificate in pre school education pr...
 
Teacher Evaluation Rubrics
Teacher Evaluation Rubrics Teacher Evaluation Rubrics
Teacher Evaluation Rubrics
 
Coaching as part of teacher evaluation process v2.0
Coaching as part of teacher evaluation process v2.0Coaching as part of teacher evaluation process v2.0
Coaching as part of teacher evaluation process v2.0
 
Inaugural lecture
Inaugural lectureInaugural lecture
Inaugural lecture
 

Similar to NYSCOSS Conference Superintendents Training on Assessment 9 14

Continuous improvement presentation 2014
Continuous improvement presentation 2014Continuous improvement presentation 2014
Continuous improvement presentation 2014
mrppittman
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growthJohn Cronin
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
John Cronin
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growthJohn Cronin
 
Aeiou of k 3 literary checkpoints-3 3
Aeiou of k 3 literary checkpoints-3 3Aeiou of k 3 literary checkpoints-3 3
Aeiou of k 3 literary checkpoints-3 3
Keith Eades
 
Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Martin Brown
 
MSDE Presentation on Student Learning Objectives: MSEA 2013 Convention
MSDE Presentation on Student Learning Objectives: MSEA 2013 ConventionMSDE Presentation on Student Learning Objectives: MSEA 2013 Convention
MSDE Presentation on Student Learning Objectives: MSEA 2013 Conventionmarylandeducators
 
Acis assessment presentation for posting
Acis assessment presentation for postingAcis assessment presentation for posting
Acis assessment presentation for postingJonathan Martin
 
Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Martin Brown
 
Non-Cognitive Testing
Non-Cognitive TestingNon-Cognitive Testing
Non-Cognitive Testing
jwilliams77
 
Lesssons from reform_around_the_world_0
Lesssons from reform_around_the_world_0Lesssons from reform_around_the_world_0
Lesssons from reform_around_the_world_0leadingcurriculum
 
Wsu Greg Lobdell September 2008 Data And Decision Making
Wsu Greg Lobdell September 2008 Data And Decision MakingWsu Greg Lobdell September 2008 Data And Decision Making
Wsu Greg Lobdell September 2008 Data And Decision MakingWSU Cougars
 
Map ppt for parents 5.2013
Map ppt for parents 5.2013Map ppt for parents 5.2013
Map ppt for parents 5.2013gurubesar
 
RtIMTSS SPE 501-Spr
  RtIMTSS                                       SPE 501-Spr  RtIMTSS                                       SPE 501-Spr
RtIMTSS SPE 501-Spr
VannaJoy20
 
Teaching, learning and assessment
Teaching, learning and assessmentTeaching, learning and assessment
Teaching, learning and assessment
Jon Nicholls
 
Portfolio-Assessment.pdf
Portfolio-Assessment.pdfPortfolio-Assessment.pdf
Portfolio-Assessment.pdf
LynneSamson1
 
Assessment of Students and Environments
Assessment of Students and EnvironmentsAssessment of Students and Environments
Assessment of Students and Environments
Kirsten Haugen
 
Why a programme view? Why TESTA?
Why a programme view? Why TESTA?Why a programme view? Why TESTA?
Why a programme view? Why TESTA?
Tansy Jessop
 
Developing Student Confidence
Developing Student ConfidenceDeveloping Student Confidence
Developing Student Confidence
mmcdowell13
 

Similar to NYSCOSS Conference Superintendents Training on Assessment 9 14 (20)

Continuous improvement presentation 2014
Continuous improvement presentation 2014Continuous improvement presentation 2014
Continuous improvement presentation 2014
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Connecticut mesuring and modeling growth
Connecticut   mesuring and modeling growthConnecticut   mesuring and modeling growth
Connecticut mesuring and modeling growth
 
Aeiou of k 3 literary checkpoints-3 3
Aeiou of k 3 literary checkpoints-3 3Aeiou of k 3 literary checkpoints-3 3
Aeiou of k 3 literary checkpoints-3 3
 
Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Sse workshop 2 spring 2014
Sse workshop 2 spring 2014
 
MSDE Presentation on Student Learning Objectives: MSEA 2013 Convention
MSDE Presentation on Student Learning Objectives: MSEA 2013 ConventionMSDE Presentation on Student Learning Objectives: MSEA 2013 Convention
MSDE Presentation on Student Learning Objectives: MSEA 2013 Convention
 
Acis assessment presentation for posting
Acis assessment presentation for postingAcis assessment presentation for posting
Acis assessment presentation for posting
 
Sse workshop 2 spring 2014
Sse workshop 2 spring 2014Sse workshop 2 spring 2014
Sse workshop 2 spring 2014
 
Group1 Assessment
Group1 AssessmentGroup1 Assessment
Group1 Assessment
 
Non-Cognitive Testing
Non-Cognitive TestingNon-Cognitive Testing
Non-Cognitive Testing
 
Lesssons from reform_around_the_world_0
Lesssons from reform_around_the_world_0Lesssons from reform_around_the_world_0
Lesssons from reform_around_the_world_0
 
Wsu Greg Lobdell September 2008 Data And Decision Making
Wsu Greg Lobdell September 2008 Data And Decision MakingWsu Greg Lobdell September 2008 Data And Decision Making
Wsu Greg Lobdell September 2008 Data And Decision Making
 
Map ppt for parents 5.2013
Map ppt for parents 5.2013Map ppt for parents 5.2013
Map ppt for parents 5.2013
 
RtIMTSS SPE 501-Spr
  RtIMTSS                                       SPE 501-Spr  RtIMTSS                                       SPE 501-Spr
RtIMTSS SPE 501-Spr
 
Teaching, learning and assessment
Teaching, learning and assessmentTeaching, learning and assessment
Teaching, learning and assessment
 
Portfolio-Assessment.pdf
Portfolio-Assessment.pdfPortfolio-Assessment.pdf
Portfolio-Assessment.pdf
 
Assessment of Students and Environments
Assessment of Students and EnvironmentsAssessment of Students and Environments
Assessment of Students and Environments
 
Why a programme view? Why TESTA?
Why a programme view? Why TESTA?Why a programme view? Why TESTA?
Why a programme view? Why TESTA?
 
Developing Student Confidence
Developing Student ConfidenceDeveloping Student Confidence
Developing Student Confidence
 

More from NWEA

Nd evaluations using growth data 4 13
Nd evaluations using growth data 4 13Nd evaluations using growth data 4 13
Nd evaluations using growth data 4 13
NWEA
 
Dylan Wiliam seminar for district leaders accelerate learning with formative...
Dylan Wiliam seminar for district leaders  accelerate learning with formative...Dylan Wiliam seminar for district leaders  accelerate learning with formative...
Dylan Wiliam seminar for district leaders accelerate learning with formative...
NWEA
 
SC Assessment Summit March 2013
SC Assessment Summit March 2013SC Assessment Summit March 2013
SC Assessment Summit March 2013
NWEA
 
Assessment Program Alignment: Making Essential Connections Between Assessment...
Assessment Program Alignment: Making Essential Connections Between Assessment...Assessment Program Alignment: Making Essential Connections Between Assessment...
Assessment Program Alignment: Making Essential Connections Between Assessment...
NWEA
 
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
NWEA
 
KLT TLC Leader Materials Set Excerpt
KLT TLC Leader Materials Set ExcerptKLT TLC Leader Materials Set Excerpt
KLT TLC Leader Materials Set Excerpt
NWEA
 
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
NWEA
 
Predicting Proficiency… How MAP Predicts State Test Performance
Predicting Proficiency… How MAP Predicts State Test PerformancePredicting Proficiency… How MAP Predicts State Test Performance
Predicting Proficiency… How MAP Predicts State Test Performance
NWEA
 
Connecting the Dots: CCSS, DI, NWEA, Help!
Connecting the Dots: CCSS, DI, NWEA, Help!Connecting the Dots: CCSS, DI, NWEA, Help!
Connecting the Dots: CCSS, DI, NWEA, Help!
NWEA
 
What's New at NWEA: Keeping Learning on Track
What's New at NWEA: Keeping Learning on TrackWhat's New at NWEA: Keeping Learning on Track
What's New at NWEA: Keeping Learning on Track
NWEA
 
What’s New at NWEA: Power of Teaching
What’s New at NWEA: Power of TeachingWhat’s New at NWEA: Power of Teaching
What’s New at NWEA: Power of Teaching
NWEA
 
What’s New at NWEA: Skills Pointer
What’s New at NWEA: Skills PointerWhat’s New at NWEA: Skills Pointer
What’s New at NWEA: Skills Pointer
NWEA
 
Finding Meaning in NWEA Data
Finding Meaning in NWEA DataFinding Meaning in NWEA Data
Finding Meaning in NWEA Data
NWEA
 
An Alternative Method to Rate Teacher Performance
An Alternative Method to Rate Teacher PerformanceAn Alternative Method to Rate Teacher Performance
An Alternative Method to Rate Teacher Performance
NWEA
 
Data Driven Learning and the iPad
Data Driven Learning and the iPadData Driven Learning and the iPad
Data Driven Learning and the iPad
NWEA
 
21st Century Teaching and Learning
21st Century Teaching and Learning21st Century Teaching and Learning
21st Century Teaching and Learning
NWEA
 
Grading and Reporting Student Learning
Grading and Reporting Student LearningGrading and Reporting Student Learning
Grading and Reporting Student Learning
NWEA
 
You Want Us To Do What???
You Want Us To Do What???You Want Us To Do What???
You Want Us To Do What???
NWEA
 
MAK Mitchell Keynote Address
MAK Mitchell Keynote AddressMAK Mitchell Keynote Address
MAK Mitchell Keynote Address
NWEA
 
Using DesCartes Instructional Ladders to Plan for Differentiated Instruction
Using DesCartes Instructional Ladders to Plan for Differentiated InstructionUsing DesCartes Instructional Ladders to Plan for Differentiated Instruction
Using DesCartes Instructional Ladders to Plan for Differentiated Instruction
NWEA
 

More from NWEA (20)

Nd evaluations using growth data 4 13
Nd evaluations using growth data 4 13Nd evaluations using growth data 4 13
Nd evaluations using growth data 4 13
 
Dylan Wiliam seminar for district leaders accelerate learning with formative...
Dylan Wiliam seminar for district leaders  accelerate learning with formative...Dylan Wiliam seminar for district leaders  accelerate learning with formative...
Dylan Wiliam seminar for district leaders accelerate learning with formative...
 
SC Assessment Summit March 2013
SC Assessment Summit March 2013SC Assessment Summit March 2013
SC Assessment Summit March 2013
 
Assessment Program Alignment: Making Essential Connections Between Assessment...
Assessment Program Alignment: Making Essential Connections Between Assessment...Assessment Program Alignment: Making Essential Connections Between Assessment...
Assessment Program Alignment: Making Essential Connections Between Assessment...
 
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
Predicting Student Performance on the MSP-HSPE: Understanding, Conducting, an...
 
KLT TLC Leader Materials Set Excerpt
KLT TLC Leader Materials Set ExcerptKLT TLC Leader Materials Set Excerpt
KLT TLC Leader Materials Set Excerpt
 
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
What's New at NWEA: Children’s Progress Academic Assessment (CPAA)
 
Predicting Proficiency… How MAP Predicts State Test Performance
Predicting Proficiency… How MAP Predicts State Test PerformancePredicting Proficiency… How MAP Predicts State Test Performance
Predicting Proficiency… How MAP Predicts State Test Performance
 
Connecting the Dots: CCSS, DI, NWEA, Help!
Connecting the Dots: CCSS, DI, NWEA, Help!Connecting the Dots: CCSS, DI, NWEA, Help!
Connecting the Dots: CCSS, DI, NWEA, Help!
 
What's New at NWEA: Keeping Learning on Track
What's New at NWEA: Keeping Learning on TrackWhat's New at NWEA: Keeping Learning on Track
What's New at NWEA: Keeping Learning on Track
 
What’s New at NWEA: Power of Teaching
What’s New at NWEA: Power of TeachingWhat’s New at NWEA: Power of Teaching
What’s New at NWEA: Power of Teaching
 
What’s New at NWEA: Skills Pointer
What’s New at NWEA: Skills PointerWhat’s New at NWEA: Skills Pointer
What’s New at NWEA: Skills Pointer
 
Finding Meaning in NWEA Data
Finding Meaning in NWEA DataFinding Meaning in NWEA Data
Finding Meaning in NWEA Data
 
An Alternative Method to Rate Teacher Performance
An Alternative Method to Rate Teacher PerformanceAn Alternative Method to Rate Teacher Performance
An Alternative Method to Rate Teacher Performance
 
Data Driven Learning and the iPad
Data Driven Learning and the iPadData Driven Learning and the iPad
Data Driven Learning and the iPad
 
21st Century Teaching and Learning
21st Century Teaching and Learning21st Century Teaching and Learning
21st Century Teaching and Learning
 
Grading and Reporting Student Learning
Grading and Reporting Student LearningGrading and Reporting Student Learning
Grading and Reporting Student Learning
 
You Want Us To Do What???
You Want Us To Do What???You Want Us To Do What???
You Want Us To Do What???
 
MAK Mitchell Keynote Address
MAK Mitchell Keynote AddressMAK Mitchell Keynote Address
MAK Mitchell Keynote Address
 
Using DesCartes Instructional Ladders to Plan for Differentiated Instruction
Using DesCartes Instructional Ladders to Plan for Differentiated InstructionUsing DesCartes Instructional Ladders to Plan for Differentiated Instruction
Using DesCartes Instructional Ladders to Plan for Differentiated Instruction
 

Recently uploaded

Home assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdfHome assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdf
Tamralipta Mahavidyalaya
 
A Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in EducationA Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in Education
Peter Windle
 
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
siemaillard
 
TESDA TM1 REVIEWER FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
TESDA TM1 REVIEWER  FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...TESDA TM1 REVIEWER  FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
TESDA TM1 REVIEWER FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
EugeneSaldivar
 
The Challenger.pdf DNHS Official Publication
The Challenger.pdf DNHS Official PublicationThe Challenger.pdf DNHS Official Publication
The Challenger.pdf DNHS Official Publication
Delapenabediema
 
Instructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptxInstructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptx
Jheel Barad
 
Operation Blue Star - Saka Neela Tara
Operation Blue Star   -  Saka Neela TaraOperation Blue Star   -  Saka Neela Tara
Operation Blue Star - Saka Neela Tara
Balvir Singh
 
Synthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptxSynthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptx
Pavel ( NSTU)
 
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdfAdversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
Po-Chuan Chen
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
Vivekanand Anglo Vedic Academy
 
Chapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptxChapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptx
Mohd Adib Abd Muin, Senior Lecturer at Universiti Utara Malaysia
 
Model Attribute Check Company Auto Property
Model Attribute  Check Company Auto PropertyModel Attribute  Check Company Auto Property
Model Attribute Check Company Auto Property
Celine George
 
2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
Sandy Millin
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
Vikramjit Singh
 
"Protectable subject matters, Protection in biotechnology, Protection of othe...
"Protectable subject matters, Protection in biotechnology, Protection of othe..."Protectable subject matters, Protection in biotechnology, Protection of othe...
"Protectable subject matters, Protection in biotechnology, Protection of othe...
SACHIN R KONDAGURI
 
Supporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptxSupporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptx
Jisc
 
The Accursed House by Émile Gaboriau.pptx
The Accursed House by Émile Gaboriau.pptxThe Accursed House by Émile Gaboriau.pptx
The Accursed House by Émile Gaboriau.pptx
DhatriParmar
 
How libraries can support authors with open access requirements for UKRI fund...
How libraries can support authors with open access requirements for UKRI fund...How libraries can support authors with open access requirements for UKRI fund...
How libraries can support authors with open access requirements for UKRI fund...
Jisc
 
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXXPhrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
MIRIAMSALINAS13
 
Unit 2- Research Aptitude (UGC NET Paper I).pdf
Unit 2- Research Aptitude (UGC NET Paper I).pdfUnit 2- Research Aptitude (UGC NET Paper I).pdf
Unit 2- Research Aptitude (UGC NET Paper I).pdf
Thiyagu K
 

Recently uploaded (20)

Home assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdfHome assignment II on Spectroscopy 2024 Answers.pdf
Home assignment II on Spectroscopy 2024 Answers.pdf
 
A Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in EducationA Strategic Approach: GenAI in Education
A Strategic Approach: GenAI in Education
 
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa
 
TESDA TM1 REVIEWER FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
TESDA TM1 REVIEWER  FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...TESDA TM1 REVIEWER  FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
TESDA TM1 REVIEWER FOR NATIONAL ASSESSMENT WRITTEN AND ORAL QUESTIONS WITH A...
 
The Challenger.pdf DNHS Official Publication
The Challenger.pdf DNHS Official PublicationThe Challenger.pdf DNHS Official Publication
The Challenger.pdf DNHS Official Publication
 
Instructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptxInstructions for Submissions thorugh G- Classroom.pptx
Instructions for Submissions thorugh G- Classroom.pptx
 
Operation Blue Star - Saka Neela Tara
Operation Blue Star   -  Saka Neela TaraOperation Blue Star   -  Saka Neela Tara
Operation Blue Star - Saka Neela Tara
 
Synthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptxSynthetic Fiber Construction in lab .pptx
Synthetic Fiber Construction in lab .pptx
 
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdfAdversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
Adversarial Attention Modeling for Multi-dimensional Emotion Regression.pdf
 
The French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free downloadThe French Revolution Class 9 Study Material pdf free download
The French Revolution Class 9 Study Material pdf free download
 
Chapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptxChapter 3 - Islamic Banking Products and Services.pptx
Chapter 3 - Islamic Banking Products and Services.pptx
 
Model Attribute Check Company Auto Property
Model Attribute  Check Company Auto PropertyModel Attribute  Check Company Auto Property
Model Attribute Check Company Auto Property
 
2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...2024.06.01 Introducing a competency framework for languag learning materials ...
2024.06.01 Introducing a competency framework for languag learning materials ...
 
Digital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and ResearchDigital Tools and AI for Teaching Learning and Research
Digital Tools and AI for Teaching Learning and Research
 
"Protectable subject matters, Protection in biotechnology, Protection of othe...
"Protectable subject matters, Protection in biotechnology, Protection of othe..."Protectable subject matters, Protection in biotechnology, Protection of othe...
"Protectable subject matters, Protection in biotechnology, Protection of othe...
 
Supporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptxSupporting (UKRI) OA monographs at Salford.pptx
Supporting (UKRI) OA monographs at Salford.pptx
 
The Accursed House by Émile Gaboriau.pptx
The Accursed House by Émile Gaboriau.pptxThe Accursed House by Émile Gaboriau.pptx
The Accursed House by Émile Gaboriau.pptx
 
How libraries can support authors with open access requirements for UKRI fund...
How libraries can support authors with open access requirements for UKRI fund...How libraries can support authors with open access requirements for UKRI fund...
How libraries can support authors with open access requirements for UKRI fund...
 
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXXPhrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
Phrasal Verbs.XXXXXXXXXXXXXXXXXXXXXXXXXX
 
Unit 2- Research Aptitude (UGC NET Paper I).pdf
Unit 2- Research Aptitude (UGC NET Paper I).pdfUnit 2- Research Aptitude (UGC NET Paper I).pdf
Unit 2- Research Aptitude (UGC NET Paper I).pdf
 

NYSCOSS Conference Superintendents Training on Assessment 9 14

  • 1. From a Superintendent’s Perspective: Using data wisely Andy Hegedus, Ed.D. September 2014
  • 2. Trying to gauge my audience and adjust my speed . . . • How many of you think your literacy with assessments is “Good” or better? • How many of you have a fine tuned assessment program? • How many of you think your practical knowledge about using data for systemic improvement is “Good” or better?
  • 3. The Bottom Line First: Here’s what you must have • An adequate depth of knowledge to ask really good questions and the tenacity to do so • A person willing to “Speak the Truth, With Love, To Power” • A big lever – “What gets measured (and attended to), gets done” The measurement and reinforcement system is your responsibility
  • 4. My Purpose • Increase your understanding about various urgent assessment related topics – Ask better questions – Useful for making all types of decisions with data
  • 5. Three main topics • Assessment basics + • Improving your assessment program • Data culture
  • 6. Go forth thoughtfully with care • What we’ve known to be true is now being shown to be true – Using data thoughtfully improves student achievement and growth rates – 12% mathematics, 13% reading • There are dangers present however – Unintended Consequences Slotnik, W. J. , Smith, M. D., It’s more than money, February 2013, retrieved from http://www.ctacusa.com/PDFs/MoreThanMoney-report.pdf
  • 7. Remember the old adage? “What gets measured (and attended to), gets done”
  • 8. • NCLB An infamous example – Cast light on inequities – Improved performance of “Bubble Kids” – Narrowed taught curriculum The same dynamic happens inside your schools
  • 9. A patient’s health doesn’t change because we know their blood pressure It’s our response that makes all the difference It’s what we do that counts
  • 10. Be considerate of the continuum of stakes involved Support Compensate Terminate Increasing levels of required rigor Increasing risk
  • 11. Assessment basics (in a teacher evaluation frame)
  • 12. Three primary conditions 1. Alignment between the content assessed and the content to be taught 2. Selection of an appropriate assessment • Used for the purpose for which it was designed (proficiency vs. growth) • Can accurately measure the knowledge of all students • Adequate sensitivity to growth 3. Adjust for context/control for factors outside a teacher’s direct control (value-added)
  • 13. Two approaches we like 1. Assessment results used wisely as part of a dialogue to help teachers set and meet challenging goals 2. Use of tests as a “yellow light” to identify teachers who may be in need of additional support or are ready for more
  • 14. What question is being answered in support of using data in evaluating teachers? Is the progress produced by this teacher dramatically different than teaching peers who deliver instruction to comparable students in comparable situations?
  • 15. Marcus’ growth College readiness standard Marcus Normal Growth Needed Growth
  • 16. There are four key steps required to The Test answer this question Top-Down Model The Growth Metric The Evaluation The Rating
  • 17. Let’s begin at the beginning The Test The Growth Metric The Evaluation The Rating
  • 18. What is measured should be aligned to what is to be taught 3rd Grade ELA Teacher? 3rd Grade ELA Standards 3rd Grade Social Studies Teacher? 1. Answer questions to demonstrate understanding of text…. 2. Determine the main idea of a Elem. Art Teacher? text…. 3. Determine the meaning of general academic and domain specific words… Would you use a general reading assessment in the evaluation of a…. ~30% of teachers teach in tested subjects and grades The Other 69 Percent: Fairly Rewarding the Performance of Teachers of Nontested Subjects and Grades, http://www.cecr.ed.gov/guides/other69Percent.pdf
  • 19. What is measured should be aligned to what is to be taught • Assessments should align with the teacher’s instructional responsibility – Specific advanced content • HS teachers teaching discipline specific content – Especially 11th and 12th grade • MS teachers teaching HS content to advanced students – Non-tested subjects • School-wide results are more likely “professional responsibility” rather than reflecting competence – HS teachers providing remedial services
  • 20. The purpose and design of the instrument is significant • Many assessments are not designed to measure growth • Others do not measure growth equally well for all students
  • 21. Let’s ensure we have similar meaning Adult Reading 5th Grade x Beginning Literacy x Time 1 Time 2 Status Two assumptions: 1. Measurement accuracy, and 2. Vertical interval scale
  • 22. Accurately measuring growth depends on accurately measuring achievement
  • 23. What does it take to accurately measure achievement? Questions surrounding the student’s achievement level The more questions the merrier
  • 24. Teachers encounter a distribution of student performance Adult Reading 5th Grade x x x x x x x x x x x x x x x Beginning Literacy Grade Level Performance
  • 25. Adaptive testing works differently Item bank can span full range of achievement
  • 26. How about accurately measuring height? What if the yardstick stopped in the middle of his back?
  • 27. Items available need to match student ability California STAR NWEA MAP
  • 28. How about accurately measuring height? What if we could only mark within a pre-defined six inch range?
  • 29. 5th Grade Level Items These differences impact measurement error .12 .10 .08 .06 .04 .02 .00 Information 190 200 210 220 230 240 260 Scale Score Fully Adaptive Test Significantly Different Error 250 Constrained Adaptive or Paper/Pencil Test
  • 30. To determine growth, achievement measurements must be related through a scale
  • 31. Let’s measure height again If I was measured as: 5’ 9” And a year later I was: 1.82m Did I grow? Yes. ~ 2.5” How do you know?
  • 32. Traditional assessment uses items reflecting the grade level standards Adult Reading 6th Grade 5th Grade 4th Grade Grade Level Standards Beginning Literacy Traditional Assessment Item Bank
  • 33. Traditional assessment uses items reflecting the grade level standards Adult Reading 6th Grade 5th Grade 4th Grade Grade Level Standards Grade Level Standards Beginning Literacy Overlap allows linking and scale construction Grade Level Standards
  • 34. Error can change your life! • Think of a high stakes test – State Summative – Designed mainly to identify if a student is proficient or not • Do they do that well? • 93% correct on Proficiency determination • Does it go off design well? • 75% correct on Performance Levels determination *Testing: Not an Exact Science, Education Policy Brief, Delaware Education Research & Development Center, May 2004, http://dspace.udel.edu:8080/dspace/handle/19716/244
  • 35. Using tests in high stakes ways creates new dynamic • Tests specifically designed to inform classroom instruction and school improvement in formative ways No incentive in the system for inaccurate data
  • 36. 8.00 6.00 4.00 2.00 0.00 -2.00 -4.00 -6.00 New phenomenon when used as part of a compensation program Mean value-added growth by school 1 3 5 7 9 11 13 15 17 19 21 23 25 27 29 31 33 35 37 39 41 43 45 47 49 51 53 55 57 59 61 63 65 67 69 71 Students taking 10+ minutes longer spring than fall All other students
  • 37. Other consequence When teachers are evaluated on growth using a once per year assessment, one teacher who cheats disadvantages the next teacher
  • 38. Lessons? • What were some things you learned? • What practices do you want to reinforce? • What do you need to do differently? • Think – Pair – 2 min to make some notes – 3 min to share with a neighbor
  • 39. Testing is complete . . . What is useful to answer our question? The Test The Growth Metric The Evaluation The Rating
  • 40. The problem with spring-spring testing Teacher 1 Summer Teacher 2 4/14 5/14 6/14 7/14 8/14 9/14 10/14 11/14 12/14 1/15 2/15 3/15 4/15
  • 41. A better approach • When possible use a spring – fall – spring approach • Measure summer loss and incentivize schools and teachers to minimize it • Measure teacher performance fall to spring, giving as much instructional time as possible between assessments • Monitor testing conditions to minimize gaming of fall or spring results
  • 42. 100 90 80 70 60 50 40 30 20 10 0 Grade 2 Grade 3 Grade 4 Grade 5 Grade 6 Grade 7 Grade 8 Reading Math The metric matters - Let’s go underneath “Proficiency” Difficulty of New York Proficient Cut Score National Percentile College Readiness New York Linking Study: A Study of the Alignment of the NWEA RIT Scale with the New York State (NYS) Testing Program, November 2013
  • 43. The metric matters - Let’s go underneath “Proficiency” Dahlin, M. and Durant, S., The State of Proficiency, Kingsbury Center at NWEA, July 2011
  • 44. What actually happened? Estimated Proficiency Rates For Six NY Districts With Proficiency Cut Scores Changed 87 4th Grade Mathematics 55 100 80 60 40 20 0 -20 -40 -31 2012 2013 Reported Change
  • 45. What actually happened? Estimated Proficiency Rates For Six NY Districts 4th Grade Mathematics With Proficiency Cut Scores 87 Changed 55 100 80 60 40 20 0 -20 -40 -31 2012 2013 Reported Change 46 55 10 60 50 40 30 20 10 0 Estimated Proficiency Rates For Six NY Districts 4th Grade Mathematics With 2013 Proficiency Cut Scores Applied 2012 2013 Actual Change
  • 46. Number of Students What gets measured and attended to Mathematics Fall RIT No Change Down Up really does matter Proficiency College Readiness One district’s change in 5th grade mathematics performance relative to the KY proficiency cut scores
  • 47. Number of Students Changing from Proficiency to Growth means all kids matter Mathematics Student’s score in fall Below projected growth Met or above projected growth Number of 5th grade students meeting projected mathematics growth in the same district
  • 48. Lessons? • What were some things you learned? • What practices do you want to reinforce? • What do you need to do differently? • Think – Pair – Share – 2 min to make some notes – 3 min to share with a neighbor
  • 49. How can we make it fair? The Test The Growth Metric The Evaluation The Rating
  • 50. Without context what is “Good”? Scale Norms Study Adult Literacy Beginning Reading National Percentile College Readiness ACT Benchmarks State Test Performance Levels “Meets” Proficiency Common Core Proficient Performance Levels
  • 51. Normative data for growth is a bit different Fall Score Basic Factors Grade: 5th Subject: Reading 7 points FRL vs. non-FRL? IEP vs. non-IEP? ESL vs. non-ESL? Starting Achievement Instructional Weeks Typical growth Outside of a teacher’s direct control
  • 52. A Visual Representation of Value Added Spring 5th Grade Test Student A Spring Score 209 Score 207 (Average Spring Score for Similar Students) Value Added (+2 Score) Student A Fall Score 200 Fall 5th Grade Test
  • 53. Consider . . . • What if I skip this step? – Comparison is likely against normative data so the comparison is to “typical kids in typical settings” • How fair is it to disregard context? – Good teacher – bad school – Good teacher – challenging kids
  • 54. Value-added is science • Value added models can control for a variety of classroom, school level, and other conditions – Proven statistical methods – All attempt to minimize error – Variables outside controls are assumed as random
  • 55. A variety of errors means more stability only at the extremes • Control for measurement error – All models attempt to address this issue • Population size • Multiple data points – Error is compounded with combining two test events – Many teachers’ value-added scores will fall within the range of statistical error
  • 56. 12.00 11.00 10.00 9.00 8.00 7.00 6.00 5.00 4.00 3.00 2.00 1.00 0.00 -1.00 -2.00 -3.00 -4.00 -5.00 -6.00 -7.00 -8.00 -9.00 -10.00 -11.00 -12.00 Average Growth Index Score and Range Range of teacher value-added estimates Mathematics Growth Index Distribution by Teacher - Validity Filtered Q5 Q4 Q3 Q2 Q1 Each line in this display represents a single teacher. The graphic shows the average growth index score for each teacher (green line), plus or minus the standard error of the growth index estimate (blue line). We removed students who had tests of questionable validity and teachers with fewer than 20 students.
  • 57. With one teacher, error means a lot
  • 58. Assumption of randomness can have risk implications • Value-added models assume that variation is caused by randomness if not controlled for explicitly – Young teachers are assigned disproportionate numbers of students with poor discipline records – Parent requests for the “best” teachers are honored • Sound educational reasons for placement are likely to be defensible
  • 59. Instability at the tails of the distribution “The findings indicate that these modeling choices can significantly influence outcomes for individual teachers, particularly those in the tails of the performance distribution who are most likely to be targeted by high-stakes policies.” Ballou, D., Mokher, C. and Cavalluzzo, L. (2012) Using Value-Added Assessment for Personnel Decisions: How Omitted Variables and Model Specification Influence Teachers’ Outcomes. LA Times Teacher #1 LA Times Teacher #2
  • 60. How tests are used to evaluate teachers The Test The Growth Metric The Evaluation The Rating
  • 61. Translation into ratings can be difficult to inform with data • How would you translate a rank order to a rating? • Data can be provided • Value judgment ultimately the basis for setting cut scores for points or rating
  • 62. Decisions are value based, not empirical • What is far below a district’s expectation is subjective • What about • Obligation to help teachers improve? • Quality of replacement teachers?
  • 63. Even multiple measures need to be used well • System for combining elements and producing a rating is also a value based decision – Multiple measures and principal judgment must be included – Evaluate the extremes to make sure it makes sense
  • 64. Leadership Courage Is A Key 5 4 3 2 1 0 Ratings can be driven by the assessment Teacher 1 Teacher 2 Teacher 3 Observation Assessment Real or Noise?
  • 65. Big Message If evaluators do not differentiate their ratings, then all differentiation comes from the test
  • 66. Lessons? • What were some things you learned? • What practices do you want to reinforce? • What do you need to do differently? • Think – Pair – Share – 2 min to make some notes – 3 min to share with a neighbor – 2 min for two report outs on anything so far
  • 68. • Read the sheet and highlight anything interesting to you Let’s Define Types of Assessments
  • 69. The pursuit of compliance is exhausting because it is always a moving target. Governors move on, the party in power gets replaced, a new president is elected, and all want to put their own stamp on education. It is saner and less exhausting to define your own course and align compliance requirements to that.
  • 70. Seven standards that define the purpose driven assessment system
  • 71. The purposes of all assessments are defined and the assessments are valid 1 and useful for their purposes
  • 72. Teachers are educated in the proper administration and application of the assessments used in their classrooms 2
  • 73. Redundant, mis-aligned, or unused 3assessments are eliminated
  • 74. Assessment results are aligned to the needs of their audiences 4
  • 75. Assessment results are delivered in a timely and useful manner 5
  • 76. The metrics and incentives used encourage a focus on all learners 6
  • 77. The assessment program contributes to a climate of transparency and objectivity with a long-term focus 7
  • 78. 1. Typical assessment purposes • Identify student learning needs • Identify groupings of students for instruction • Guide instruction • Course placement • Determine eligibility for programs • Award credits and/or assign grades • Evaluate proficiency • Monitor student progress • Predict proficiency • Project achievement of a goal • Formative and summative evaluation of programs • Formative evaluation to support school and teacher improvement • Report student achievement, growth, and progress to the community and stakeholders • Summative evaluation of schools and teachers
  • 79. 1. Assessment Purpose Survey To increase value… Identify gaps between: 1. How critical is this data to your work? 2. How do you actually use this data? Take 10 min to fill this out and 5 min to pair and discuss areas of biggest gap
  • 80. Compare assessments and their purposes to find unnecessary overlaps Take 10 min to fill this out and 5 min to pair and discuss areas of redundancy 3. Eliminate waste
  • 81. Lessons? • What were some things you learned? • What practices do you want to reinforce? • What do you need to do differently? • Think – Pair – Share – 2 min to make some notes – 3 min to share with a neighbor – 2 min for two report outs on this section
  • 83. Education Organizations Mature Poor to Fair Fair to Good Good to Great Great to Excellent Achieving the basics of literacy and numeracy Getting the foundations in place Shaping the professional Improving through peers and innovation Data use does too Barber, M., Chijioke, C., & Mourshed, M. (2011). How the world’s most improved school systems keep getting better. McKinsey & Company.
  • 84. Data Use Continuum Poor to Fair Fair to Good Good to Great Great to Excellent One on One Within Teams Within the Walls Across the Walls Requires a shift in the culture
  • 85. What would you do with this?
  • 86. Reflection Time • Where are your pockets of most maturity? • Least maturity? • What is causing the differences? Think - 2 min
  • 87. Research on data use in school improvement • Education problems are “Wicked” – Problem boundaries are ill-defined – No definitive solutions – Highly resistant to change – Problem and solutions depend on perspective – Changes are consequential Data can only take you so far
  • 88. Research on data use in school improvement • Use data as a platform for deeper conversations • Define your problem well – Problem title and description – Magnitude – Location – Duration
  • 89. Research on data use in school improvement • Part of a continuous improvement process –Data conversations • Collaborative • Embedded in culture • Structured process • Love, Nancy – Using data to improve learning for all • Lipton & Wellman – Got data? Now what? • NWEA Data Coaching
  • 91. What are your biggest take-aways? • Think – 2 min • Pair – 3 min • Share – 2 min - Two people
  • 92. Last thing • Materials in your Conference App Presentation available on Slideshare.net
  • 93. More information Contacting me: E-mail: andy.hegedus@nwea.org Contacting us: Exhibit Booth or Liz Kaplan

Editor's Notes

  1. Teacher evaluations and the use of data in them can take many forms. You can use them for supporting teachers and their improvement. You can use the evaluations to compensate teachers or groups of teachers differently or you can use them in their highest stakes way to terminate teachers. The higher the stakes put on the evaluation, the more risk there is to you and your organization from a political, legal, and equity perspective. Most people naturally respond with increasing the levels of rigor put into designing the process as a way to ameliorate the risk. One fact is that the risk can’t be eliminated. Our goal – Make sure you are prepared. Understand the risk. Proper ways to implement including legal issues. Clarify some of the implications – Very complex – Prepare you and a prudent course
  2. This is the value added metric Not easy to make nuanced decisions. Can learn about the ends.
  3. Contrast with what value added communicates Plot normal growth for Marcus vs anticipated growth – value added. If you ask whether the teachers provided value added, the answer is Yes. Other line is what is needed for college readiness Blue line is what is used to evaluate the teacher. Is he on the line the parents want him to be on? Probably not. Don’t focus on one at the expense of the other NCLB – AYP vs what the parent really wants for goal setting Can be come so focused on measuring teachers that we lose sight of what parents value We are better off moving towards the kids aspirations As a parent I didn’t care if the school made AYP. I cared if my kids got the courses that helped them go where they want to go.
  4. Steps are quite important. People tend to skip some of these. Kids take a test – important that the test is aligned to instruction being given Metric – look at growth vs growth norm and calculate a growth index. Two benefits – Very transparent/Simple. People tend to use our growth norms – if you hit 60% for a grade level within a school you are dong well. Norms – growth of a kid or group of kids compared to a nationally representative sample of students Why isn’t this value added? Not all teachers can be compared to a nationally representative sample because they don’t teach kids that are just like the national sample The third step controls for variables unique to the teacher’s classroom or environment Fourth step – rating – how much below average before the district takes action or how much above before someone gets performance pay. Particular challenge in NY state right now. Law requires it.
  5. Steps are quite important. People tend to skip some of these. Kids take a test – important that the test is aligned to instruction being given Metric – look at growth vs growth norm and calculate a growth index. Two benefits – Very transparent/Simple. People tend to use our growth norms – if you hit 60% for a grade level within a school you are dong well. Norms – growth of a kid or group of kids compared to a nationally representative sample of students Why isn’t this value added? Not all teachers can be compared to a nationally representative sample because they don’t teach kids that are just like the national sample The third step controls for variables unique to the teacher’s classroom or environment Fourth step – rating – how much below average before the district takes action or how much above before someone gets performance pay. Particular challenge in NY state right now. Law requires it.
  6. Common core – very ambitious things they want to measure – tackle things on an AP test. Write and show their work. A CC assessment to evaluate teachers can be a problem. Raise your hand if you know what the capital of Chile is. Santiago. Repeat after me. We will review in a couple of minutes. Facts can be relatively easily acquired and are instructionally sensitive. If you expose kids to facts in a meaningful and engaging ways, it is sensitive to instruction.
  7. State assessment designed to measure proficiency – many items in the middle not at the ends Must use multiple points of data over time to measure this. We also believe that a principal should be more in control of the evaluation than the test – Principal and Teacher leaders are what changes schools
  8. 5th grade NY reading cut scores shown
  9. Steps are quite important. People tend to skip some of these. Kids take a test – important that the test is aligned to instruction being given Metric – look at growth vs growth norm and calculate a growth index. Two benefits – Very transparent/Simple. People tend to use our growth norms – if you hit 60% for a grade level within a school you are dong well. Norms – growth of a kid or group of kids compared to a nationally representative sample of students Why isn’t this value added? Not all teachers can be compared to a nationally representative sample because they don’t teach kids that are just like the national sample The third step controls for variables unique to the teacher’s classroom or environment Fourth step – rating – how much below average before the district takes action or how much above before someone gets performance pay. Particular challenge in NY state right now. Law requires it.
  10. NCLB required everyone to get above proficient – message focus on kids at or near proficient School systems responded MS standards are harder than the elem standards – MS problem No effort to calibrate them – no effort to project elem to ms standards Start easy and ramp up. Proficient in elem and not in MS with normal growth. When you control for the difficulty in the standards Elem and MS performance are the same
  11. Not only are standards different across grades, they are different across states. It’s data like this that helps to inspire the Common Core and consistent standards so we compare apples to apples
  12. Dramatic differences between standards based vs growth KY 5th grade mathematics Sample of students from a large school system X-axis Fall score, Y number of kids Blue are the kids who did not change status between the fall and the spring on the state test Red are the kids who declined in performance over spring – Decender Green are kids who moved above it in performance over the spring – Ascender – Bubble kids About 10% based on the total number of kids Accountability plans are made typically based on these red and green kids
  13. Same district as before Yellow – did not meet target growth – spread over the entire range of kids Green – did meet growth targets 60% vs 40% is doing well – This is a high performing district with high growth Must attend to all kids – this is a good thing – ones in the middle and at both extremes Old one was discriminatory – focus on some in lieu of others Teachers who teach really hard at the standard for years – Teachers need to be able to reach them all This does a lot to move the accountability system to parents and our desires.
  14. Be ready to coach the media – Need to be assessment literate Importance of longitudinal data systems and consistent measurement – Lots of impacts – Program evaluation for example There is a huge difference between achievement (and therefore growth) and proficiency
  15. Steps are quite important. People tend to skip some of these. Kids take a test – important that the test is aligned to instruction being given Metric – look at growth vs growth norm and calculate a growth index. Two benefits – Very transparent/Simple. People tend to use our growth norms – if you hit 60% for a grade level within a school you are dong well. Norms – growth of a kid or group of kids compared to a nationally representative sample of students Why isn’t this value added? Not all teachers can be compared to a nationally representative sample because they don’t teach kids that are just like the national sample The third step controls for variables unique to the teacher’s classroom or environment Fourth step – rating – how much below average before the district takes action or how much above before someone gets performance pay. Particular challenge in NY state right now. Law requires it.
  16. Close by noting that NWEA recognized the need for this level of precision when trying to understand student performance (and by extension, teacher performance). This is why, in NY (where we first began having this conversation with partners), we sought to partner with VARC, because of their background and experience providing these services (and because this is something that we did not want to do, even if we had the background/experience). Talk about the number of districts and students in 11-12 and 12-13, to provide context for the ability for this to be done on a broad scale.
  17. There are wonderful teachers who teach in very challenging, dysfunctional settings. The setting can impact the growth. HLM embeds the student in a classroom, the classroom in the school, and controls for the school parameters. Is it perfect. No. Is it better? Yes. Opposite is true and learning can be magnified as well. What if kids are a challenge, ESL or attendance for instance. It can deflate scores especially with a low number of kids in the sample being analyzed. Also need to make sure you have a large enough ‘n’ to make this possible especially true in small districts. Our position is that a test can inform the decision, but the principal/administrator should collect the bulk of the data that is used in the performance evaluation process.
  18. Experts recommend multiple years of data to do the evaluation. Invalid to just use two points and will testify to it. Principals never fire anyone – NY rubber room – myth If they do, it’s not fast enough. – Need to speed up the process This won’t make the process faster – Principals doing intense evaluations will
  19. Measurement error is compounded in test 1 and test 2
  20. Green line is their VA estimate and bar is the error of measure Both on top and bottom people can be in other quartiles People in the middle can cross quintiles – just based on SEM Cross country – winners spread out. End of the race spread. Middle you get a pack. Middle moving up makes a big difference in the overall race. Instability and narrowness of ranges means evaluating teachers in the middle of the test mean slight changes in performance can be a large change in performance ranking
  21. Non –random assignments Models control for various things – FRL, ethnicity, school effectiveness overall. Beyond this point assignment is random. 1st year teachers get more discipline problems than teachers who have been 30 years. Pick the kids they get. If the model doesn’t control for disciplinary record – none do have that data – scores are inflated. Makes model invalid. Principals do need to do non-random assignment – sound educational reasons for the placement – match adults for kids
  22. Steps are quite important. People tend to skip some of these. Kids take a test – important that the test is aligned to instruction being given Metric – look at growth vs growth norm and calculate a growth index. Two benefits – Very transparent/Simple. People tend to use our growth norms – if you hit 60% for a grade level within a school you are dong well. Norms – growth of a kid or group of kids compared to a nationally representative sample of students Why isn’t this value added? Not all teachers can be compared to a nationally representative sample because they don’t teach kids that are just like the national sample The third step controls for variables unique to the teacher’s classroom or environment Fourth step – rating – how much below average before the district takes action or how much above before someone gets performance pay. Particular challenge in NY state right now. Law requires it.
  23. Use NY point system as the example
  24. Be ready to coach the media – Need to be assessment literate Importance of longitudinal data systems and consistent measurement – Lots of impacts – Program evaluation for example There is a huge difference between achievement (and therefore growth) and proficiency