Bruce Baker has written an important post about the inability of pundits (and journalists) to read NAEP data.
Part of the misinterpretation is the fault of the National Assessment Governing Board, which supervises NAEP. It has a tight embargo on the scores, which are widely released to reporters. It holds a press conference, where board members and one or two carefully chosen outsiders characterize the scores.
He writes:
“Nothin’ brings out good ol’ American statistical ineptitude like the release of NAEP or PISA data. Even more disturbing is the fact that the short time window between the release of state level NAEP results and city level results for large urban districts permits the same mathematically and statistically inept pundits to reveal their complete lack of short term memory – memory regarding the relevant caveats and critiques of the meaning of NAEP data and NAEP gains in particular, that were addressed extensively only a few weeks back – a few weeks back when pundit after pundit offered wacky interpretations of how recently implemented policy changes affected previously occurring achievement gains on NAEP, and interpretations of how these policies implemented in DC and Tennessee were particularly effective (as evidenced by 2 year gains on NAEP) ignoring that states implementing similar policies did not experience such gains and that states not implementing similar policies in some cases experienced even greater gains after adjusting for starting point.
“Now that we have our NAEP TUDA results, and now that pundits can opine about how DC made greater gains than NYC because it allowed charter schools to grow faster, or teachers to be fired more readily by test scores… let’s take a look at where our big cities fit into the pictures I presented previously regarding NAEP gains and NAEP starting points.
The first huge caveat here is that any/all of these “gains” aren’t gains at all. They are cohort average score differences which reflect differences in the composition of the cohort as much as anything else. Two year gains are suspect for other reasons, perhaps relating to quirks in sampling, etc. Certainly anyone making a big deal about which districts did or did not show statistically significant differences in mean scale scores from 2011 to 2013, without considering longer term shifts is exhibiting the extremes of Mis-NAEP-ery!”
But if NAGB wanted intelligent reporting of the results, it would release them not just to reporters but to qualified experts in psychometric s and statistics. Because it refuses to do this, NAEP results are reported like a horse race. Scores are up, scores are down. But most journalists never get past the trend lines and cannot find experts who have had time to review the scores and put them into context.
I have a personal beef here because I was given access to the embargoed data when I blogged at Education Week and had 4,000 readers weekly. Now, as an independent blogger with 120,000-150,000 readers weekly, I am not qualified to gain access to the data until after they are released (because i do not work for a journal like Edweek.) I don’t claim to be a statistical expert like Bruce Baker, but surely the governing board of NAEP could release the data in advance to a diverse group of a dozen qualified experts to help journalists do a better job when the scores come out.

Statistically Illiterate Media Pundits
(You do the acronym …)
LikeLike
Diane, You wrote in the above post :
“But if NAGB wanted intelligent reporting of the results, it would release them not just to reporters but to qualified experts in psychometric s and statistics. Because it refuses to do this, NAEP results are reported like a horse race. Scores are up, scores are down. But most journalists never get past the trend lines and cannot find experts who have had time to review the scores and put them into context”
With all due respect, based on you your ongoing experiences with the the NAEP governing board, the ‘educational reformers’, as well the ‘pundit corps’, what would lead you to believe that any of these groups have a need or a desire to conduct themselves in fair or transparent ‘open’ manner?
Time and again we are ‘surprised’ by the ‘ed reformers’ out right dishonesty, lack of respect for any one other than their claque and search for cheap PR in service of their ideology. We ask, ‘how could they’? why don’t they? We expect a respect for diverging opinions and for honest interpretation of data. We are chronically disappointed.
Their own behaviors give the lie to any claims they might make for either inadvertence or ignorance. Their behavior is willful and purposeful
LikeLike
When I read blogs like this, I always remember W. E. Deming’s comment: “the most important figures that one needs for management are unknown or unknowable.” In education, unlike the production sector, there are too many variables flowing in and around schools to every discover a cause and effect relationship between teaching and learning. This is not an excuse for misinterpreting what numbers we have in front of us, but the honest admission by all of us who have taught in classrooms that our effect on the social, emotional, and intellectual growth of students in our classrooms are ultimately unknowable.
LikeLike
The uses and abuses of the scores generated by high-stakes standardized tests are inevitable given the studied ignorance and casual ethics of the leading charterites/privatizers. Mixed messages, glaring inconsistencies, strange inaccuracies—all flow from the tone set from the top.
For example, the bewildering claim by “Education” President George W. Bush Jr. that “A reading comprehension test is a reading comprehension test. And a math test in the fourth grade—there’s not many ways you can foul up a test … It’s pretty easy to ‘norm the results.’” [quoted in Daniel Koretz, MEASURING UP: WHAT EDUCATIONAL TESTING REALLY TELLS US, 2008, p. 7].
But surely the folks on the ground understand, right? From a legislative hearing in March 2004 in Tennessee [land of value-added measurement/modeling] comes this immortal declaration:
[start quote] …Vernon Coffey, Director of Grainger County Schools and former Tennessee Commissioner of Education, stated that he believed TVAAS [Tennessee Value Added Assessment System] to be as reliable and valid as SAT and ACT, while admitting that “I don’t understand all the numbers, but I’m not supposed to.” [end quote] [brackets mine] [Jim Horn and Denise Wilburn, THE MISMEASURE OF EDUCATION, 2013, p. 105]
Pretty easy, not supposed to know—give me a 13th percentile and I’ll raise you to a 90th.
Rheeally!
And the assertion by another State Commissioner of Education, John King of New York, that Common Core and Montessori are practically one and the same.
Rheeally!
So it’s not surprising that the assertions of the pundits of the education establishment regarding test scores and their importance don’t add up.
Go figure…
😎
LikeLike
Dr. Ravitch, you are so right again!. The NAEP and the PISA boards really do not want their new data juried by authentic experts because it might diminish the political impact–and political impact is their bread and butter. Political impact, not authenticity governs both the professional stature and funding sources of these agencies. Further, given the “statistical ineptitude” (your words) of both the reporters and pundits in the mainstream media their “stories” about the news releases are governed by technology assisted searches of what the public is reading. Near-continuous polling on the net tells them what the public prefers to read and they find ways to keep writing it again and again. Authentic evidential refutation of the data release down the pike is dismissed. For a news pundit, “not being read” will very shortly get you a pink slip. This is why, once there is a main media mindset, no matter how fallacious, about an important issue such as local public education, it is very difficult to change with evidence and logic.
LikeLike
Some insight into the issue:
“They use statistics like a drunken man uses lamp posts – for support rather for illumination.” – Andrew Lang
“If your experiment needs statistics, you ought to have done a better experiment.” – Ernest Rutherford
“Statistics are no substitute for judgement.” – Henry Clay
And my favorite . . .
“There are three kinds of lies: lies, damned lies, and statistics.” – Mark Twain
LikeLike