A reader who teaches in the District of Columbia wrote to ask my advice. He couldn’t understand his evaluation based on the District’s complex IMPACT evaluation system. I sent it to a teacher evaluation expert, Audrey Amrein-Beardsley. She wrote a post trying to make sense of the evaluation report.
Despite its recent gains on the 2013 NAEP, the District of Columbia is not a national model.
It remains the lowest performing urban district in the nation.
Its policy of test-and-punish-and-fire have produced a startlingly high attrition rate among teachers.
Churn is not good for schools or for children or for building a culture of collaboration.
Few of the principals hired by Rhee remain in the system.
The real story in D.C.: Thanks to Mayor Vincent Gray, D.C. started universal pre-K, and it is showing benefits in the early grades. A writer in the Washington Post called Gray’s work “a staggering achievement.”
As for the rest of the story, read this article that I wrote for Talking Points Memo.
It is hard to imagine that anyone would want to copy a system built on striking fear into the hearts of teachers and principals.
No successful corporation–large or small– operates in that manner. The best of our nation’s business companies boast of how they carefully select new hires, support them, and pamper them with perks to make them happy in their work.
Success in schooling grows from collaboration, love of learning, experienced teachers and principals, equitable and adequate funding, and leadership that holds itself accountable for its decisions.
Some weeks back, the media reported that the District of Columbia’s infamous teacher evaluation program–known as IMPACT–was successful, based on a paper by researchers Thomas Dee and James Wyckoff. The takeaway allegedly was that VAM (value-added measurement) works and that DC is right to judge teacher quality by student test scores.
But Audrey Amrein-Beardsley, one of the pre-eminent national experts on VAM, says “not so fast. Don’t believe the hype.”
In this post, she dissects the DC research and says it was never peer-reviewed and is deeply flawed.
She identifies many errors, but this one is the most egregious:
“Teacher Performance:” Probably the largest fatal flaw, or the study’s most major limitation was that only 17% of the teachers included in this study (i.e., teachers of reading and mathematics in grades 4 through 8) were actually evaluated under the IMPACT system for their “teacher performance,” or for that which they contributed to the system’s most valued indicator: student achievement. Rather, 83% of the teachers did not have student test scores available to determine if they were indeed effective (or not) using individual value-added scores. It is implied throughout the paper, as well as the media reports covering this study post release, that “teacher performance” was what was investigated when in fact for four out of five DC teachers their “performance” was evaluated only as per what they were observed doing or self-reported doing all the while. These teachers were evaluated on their “performance” using almost exclusively (except for the 5% school-level value-added indicator) the same subjective measures integral to many traditional evaluation systems as well as student achievement/growth on teacher-developed and administrator-approvedclassroom-based tests, instead.
Thus, it is wrong to say that the paper vindicates IMPACT or its reliance on VAM when more than four of every five teachers in the study did not have value-added scores available.
