Richard Rothstein explains that VAM is an unproven methodology with negative consequences for the quality of education.
Rothstein says he is not surprised that Chicago teachers oppose its use. He wonders why other teachers have not gone on strike for the same reason.
It has not worked anywhere.
It narrows the curriculum.
It relies too heavily on tests that were not designed to measure teacher quality.
The teachers are being used as guinea pigs for an unproven methodology that will harm education.

Here is what NH came up with in their document that is supposed to develop effective teaching: http://www.education.nh.gov/teaching/documents/phase1report.pdf
Page 17 under teacher evaluations:
Increasing numbers of studies have failed to establish the validity of value-added growth
The primary measure of effective teaching is the impact of teaching on student growth and achievement. The challenge of evaluating teacher effectiveness is two-fold: how to measure student achievement and how to measure a teacher’s impact on that achievement. No single point-in-time measure of student learning can effectively measure all areas—social, emotional, academic, and physical—of student growth and success. Students’ depth of learning cannot be captured simply by averaging test scores. While further research is needed to establish the reliability of alternative forms of measurement (Rabinowitz, 2011), multiple measures of student achievement, such as formative classroom assessments, exhibitions, demonstrations, performances, portfolios, and self-assessments should accompany standardized tests. New Hampshire’s participation in the Smarter Balanced Assessment Consortium will help educators stay current with the ongoing research on performance assessment.
Just as multiple measures are necessary to gauge student achievement, so should they
models (VAMs) for measuring teacher impact on student achievement (Golhaber and Hansen, 2010; Braun, 2005; Rothstein, 2009; Goe, 2008). Appropriately applied, multiple measures of teacher effectiveness can include student test scores, classroom observation, analysis of classroom artifacts, portfolios, self- reports, student evaluations, and parent surveys. Measures should match the behavior evaluated;
for example, parent surveys are appropriate measures of school-community
outreach, or artifacts such as syllabi or lesson plans are appropriate measures of instructional planning (Little, Goe, and Bell, 2009).
LikeLike
Good analysis. I like Montgomery County MD Peer Assistance and Review approach
Diane Ravitch
LikeLike
VAM continues to strike me as unscientific. As a science teacher I am continually astounded by the lack of scientific integrity when these methods are employed to personally, and professionally, evaluate teachers.
The validity of such methods are the first red flag. VAM uses the philosophy of economics which should be a crushing blow right off as economics is as soft a science as they come with a myriad of “lurking variables” ready at any given time to destroy the ability to measure any one variable, i.e. teacher effect. I cannot believe more science and math teachers, alongside mathematicians (I know Ewing stepped up) and scientists, have not come out in disgust over these methods. Of course, they have somewhat, but their voices cannot overcome billions of dollars acting as carrots to states under RttT and ESEA flex to adopt these policies.
The reliability of VAM, which has come under fire, really is moot. If a process is not valid, then questions of reliability are meaningless. I laugh when I read that VAM is unreliable, because real scientists know that if a process is invalid, reliability measures are meaningless. Economists are not real scientists.
If VAM were used as a diagnostic, then it would be acceptable, but it cannot be used as a diagnostic as its only purpose is to employ punitive measures. The VAM empire knows this, including SAS. VAM is not designed to help kids learn – it is designed to fire teachers under its own invalid measures measured solely on test scores.
The “multiple measures” of teacher evaluation are meaningless. Administrators will be biased concerning a teacher’s past VAM scores and will be scolded for not matching scores under the premise that they are not good evaluators (this is already being documented in HISD and in TN). Now, we have just placed more power in the hands of administrators to relieve themselves of whatever teachers they wish.
What good is a “projected score” to teachers? Teachers can do nothing to remedy a student who may not make “growth”, because there is no way to know who will or who will not make growth. No teacher-made assessment gives teachers a clue as to what students will or will not grow. A teacher may have a students that earns an A+ on all their assignments from the teacher and may still not show growth because they may have also earned A+’s from their previous teacher also. It used to be, at least with “proficiency”, that a teacher could figure out what students needed help and zero in on those students in order to attempt to get them to reach proficiency, but what are teachers to do with “growth”? How do teachers measure whether a student will or will not show growth? It is an impossible target.
I believe that target was made impossible for a reason. Just as NCLB was designed to show schools as a failure, so too, in 10 years, will reformers look back at NAEP, PISA, and TIMSS outcomes, and declare that teachers are failures.
You can bet that state mandated tests will increase as teachers rush to teach to the test, teach THE test (for those who can see or get their hands on them), or downright cheat.
May the best test prep. teacher or cheater win and hold their job and barely feed their families to do what they love. Good luck.
LikeLike
“May the best test prep. teacher or cheater win and hold their job and barely feed their families to do what they love. Good luck.”
Can you imagine anyone rushing off towork to pass on their passion for test prep?. There will no longer be a reason to evaluate teachers. There won’t be any.
LikeLike