Archives for category: Research

Matthew Di Carlo of the Albert Shanker Institute of the American Federation of Teachers is neither pro-TFA nor anti-TFA.

 

Here he reviews the latest study of TFA by Mathematica.

 

It has been widely reported that the study found little or no difference between the test scores of the students taught by TFA and by regular teachers. TFA saw that as a victory, since it presumably showed that no training or experience was needed to achieve the same results. Others saw it as a repudiation of TFA’s oft-repeated claims that their recruits were superior to career teachers.

 

Di Carlo parses the results and reaches this conclusion:

 

Now, on the one hand, it’s absolutely fair to use the results of this and previous TFA evaluations to suggest that we may have something to learn from TFA training and recruitment (e.g., Dobbie 2011). Like all new teachers, TFA recruits struggle at first, but they do seem to perform as well as or better than other teachers, many of whom have had considerably more experience and formal training.

 

On the other hand, as I’ve discussed before, there is also, perhaps, an implication here regarding the “type” of person we are trying to recruit into teaching. Consider that TFA recruits are the very epitome of the hard-charging, high-achieving young folks that many advocates are desperate to attract to the profession. To be clear, it is a great thing any time talented, ambitous, service-oriented young people choose teaching, and I personally think TFA deserves credit for bringing them in. Yet, no matter how you cut it, they are, at best, only modestly more effective (in raising math and reading test scores) than non-TFA teachers.

 

This reflects the fact that identifying good teachers based on pre-service characteristics is extraordinarily difficult, and the best teachers are very often not those who attended the most selective colleges or scored highly on their SATs. And yet so much of our education reform debate is about overhauling long-standing human resource policies largely to attract these high-flying young people. It follows, then, that perhaps we should be very careful not to fixate too much on an unsupported idea of the “type” of person we want to attract and what they are looking for, and instead pay a little more attention to investigating alternative observable characteristics that may prove more useful, and identifying employment conditions and work environments that maximize retention of effective teachers who are already in the classroom.

 

For me, the problem with all such studies is the assumption that the best (perhaps the only) way to identify the best teachers is by comparing changes in test scores. Great teachers supposedly get higher scores than mediocre teachers. I think that places far too much faith in standardized testing and in the assumption that education is solely measured by those tests. It makes the tests the arbiters of all things, even though most teachers do not teach tested subjects. Test-based findings are even more suspect when the children are very young.

Julian Vasquez Heilig, who recently moved from the University of Texas to California State University at Sacramento, is one of the nation’s leading authorities on Teach for America. He has studied their performance over time (see here and here), and he is not a fan. When Mathematica released its latest study of TFA, Heilig read it closely and analyzed the findings. TFA boasted that the study showed that its teachers were just as good as those who had studied education and intended to be career teachers. Some readers gleaned from this finding that “anyone can teach, no professional preparation needed,” that is, if they graduate from a highly selective college and are admitted to TFA.

 

Heilig digs deeper and has a different take on the study. The main finding, he says, is that Mathematica found no statistically significant differences in the groups of teachers they studied. However, he points out, the TFA teachers were overwhelmingly white, and few had any intention of staying in teaching as a career.

 

He notes that the test of “effectiveness” in pre-K-grade 2 is a five minute test:

 

Equally effective at what?…Mathematica utilized performance on the Woodcock Johnson III for the Pre-K-2 results— which takes 5 minutes to administer. Thus, the effectiveness of TFA teachers compared to Pre-K – 2nd grade teachers is based on a five minute administration to capture letter-word identification (Pre-K – 2) and applied problems for mathematics Pre-K – 2). Furthermore, one of the more egregious issues in the study is the aggregation of grades is that of the states that have Pre-K programs, more than half of states do not even require Pre-K teachers to have a bachelor’s degree. The report does not state that lack of a degree was an exclusion criteria and it is explicit that community preschools were included, so it appears than an aggregate that includes not only alternatively certified but also non-degreed teachers worked to TFA’s advantage. Should we really be impressed that TFA teachers outperformed a group that could have included non-degreed teachers? And they do it twice: with kindergarten and with grades K, 1, and 2.

 

What are the lessons of the study? Heilig writes:

 

So the [TFA] teachers were— on average— young, White, and from selective colleges. They had not studied early childhood in college and had very little teaching experience. They reported a similar amount of “pedagogy” (primarily the 60 hours from the five week Summer Institute), and more professional development (as we discussed above, they viewed it not very valuable). TFA teachers also reported less student teaching experience before they entered the classroom. They also were more likely to be working with a formal mentor (I mentioned David Greene’s point about the drain on mentors due to the constant carousel of Teach For America teachers in and out of schools here). As new teachers, they spent more time planning their own lessons, but were less likely to to help other teachers. Finally, TFA teachers were less satisfied “with many aspects of teaching” and less likely to “plan to spend the rest of the career as a classroom teacher….”

 

In conclusion, read at face value, here is the message Mathematica appears to promulgate with the report:

 

We do not need experienced (read: more expensive) teachers when non-experienced, less expensive teachers get the “same” —though not statistically significant— outcomes.
We do not need a more diverse workforce of teachers, again, because TFA teachers, who are overwhelmingly white, get the same outcomes.
Is TFA really in alignment with a vision for providing every student a high quality teacher? Or do they, Mathematica et al. just keep telling us that they are?

 

For myself, I have read many times that Teach for America invites young people to “make history” by serving for two years. And Wendy Kopp has frequently said that “One day,” all children in America will have an excellent teacher. I have a hard time understanding the logic of these claims. If the TFA teachers get the same results as current teachers, how is that “making history”? If most TFA recruits leave after two years, how does that lead to the conclusion that one day all children will have an excellent teacher? If TFA persuades policymakers that teachers can do a good enough job with no professional preparation, doesn’t that decimate the idea of teaching as a profession? If anyone can teach so long as they went to a selective college, how does that raise the standard for teachers? If our policymakers prefer churn, with teachers leaving every two or three years to find their real career, how is that good for students? How does TFA improve the profession? It doesn’t. It eliminates it.

 

For his fearlessness, for his willingness to stand up to those with money and power, for his willingness to present the evidence as he finds it without fear or favor, I place Julian Vasquez Heilig on the honor roll of this blog. He is an example to all researchers of the ethics of his profession. To be an outstanding researcher requires years of study, scholarship, discipline, dedication, and experience. Sort of like being a great teacher.

 

 

 

 

A new study by Mathematica Policy Research finds that young corps members in Teach for America get no better results than other teachers.

 

Normally, this would not be big news, since TFA teachers have only five weeks of training. But for years, TFA has boasted that their young people were far superior to other teachers who had gone through professional preparation programs. Now, TFA leaders are claiming to be satisfied that their five weeks of training allows them to do just as well as those who spent a year or more learning to teach. The implicit logic of their perspective is that teaching is not a profession and that no preparation is needed beyond five weeks of TFA training. However you slice it, the TFA message degrades the profession. No profession would be considered to be a profession if any bright young person could succeed with only a few weeks of preparation. One cannot even imagine doctors or lawyers or accountants boasting that they were successful with a five-week training program.

 

The Mathematica study may not end the debate about the value of TFA. Its biggest fans seem to be the Walton Foundation, the Broad Foundation, and other foundations that want to support the proliferation of non-union charter schools with low costs and high teacher turnover. Walton gave $50 million to TFA; Broad collected $100 million from a group of foundations for TFA. And Arne Duncan gave TFA $50 million. TFA’s special contributions to American education, it appears, are to staff non-union charter schools and to demonstrate that teaching is not a profession.

The New York State Education Department released educator evaluation results on February 26, and once again, the overwhelming majority of teachers received effective or highly effective ratings. State officials were deeply disappointed by the overwhelmingly positive results. They seem to operate under the assumption that poor test results must be caused by “bad” teachers, and that their evaluation program should identify them so they may be fired.

 

The SED found that:

 

The final evaluation results show more than 95 percent of teachers statewide are rated effective (54 percent) or highly effective (42 percent); 4 percent are rated as developing; 1 percent are rated ineffective. Ninety-four percent of principals are rated effective (66 percent) or highly effective (28 percent).

 

The results were somewhat different in New York City, which used a plan imposed by then-State Commissioner John King:

 

New York City, whose evaluation plan was imposed by former Commissioner King when the New York City Department of Education could not reach agreement on the terms of the evaluation plan with the teachers union, showed greater differentiation than most districts in the State. Although New York City teachers and principals were evaluated on the same overall subcomponents as the rest of the State, the three subcomponents used different scoring ranges to determine the subcomponent rating categories (i.e., Highly Effective, Effective, Developing, Ineffective). Less than 10 percent of teachers in the city are rated Highly Effective, while 83 percent are rated Effective, 7 percent are Developing and 1 percent are Ineffective.

 

The leader of the state Board of Regents expressed disappointment at the high proportion of teachers found to be effective or highly effective:

 

“The ratings show there’s much more work to do to strengthen the evaluation system,” Board of Regents Chancellor Merryl H. Tisch said. “There’s a real contrast between how our students are performing and how their teachers and principals are evaluated. The goal of the APPR process is to identify exceptional teachers who can serve as mentors and role models, and identify struggling teachers to make sure they get the help they need to improve. The evaluation system we have now doesn’t do that. The ratings from districts don’t reflect the struggles our students face to achieve college and career readiness. State law must be changed to build an evaluation system that supports teaching and learning in classrooms across the State. Our students deserve no less.”

 

Chancellor Tisch, like Governor Cuomo, assumes that the proportion of students getting low scores should somehow be matched by a similar proportion of low-rated teachers. It would be useful if Chancellor Tisch and Governor Cuomo reviewed two basic documents: the American Statistical Association statement on the uses and misuses of value-added measurement (VAM) and the joint statement of the National Academy of Education and the American Educational Research Association. It is unfortunate that the Board of Regents and the Governor proceed without regard to research on the effects of out-of-school and in-school factors that affect test scores. Were they to familiarize themselves with the two documents cited, they might develop a very different action plan, one that helps both students and teachers.

 

See state ratings here.

 

 

Experts in early childhood education are calling for the abandonment of Common Core standards in kindergarten and their replacement by developmentally appropriate, research-based practice.

Defending the Early Years (DEY), in conjunction with the Alliance for Childhood, released a new report “Reading Instruction in Kindergarten: Little to Gain and Much to Lose.”

Early childhood experts could find no solid research showing long-term educational gains for children who are taught to read in kindergarten, yet this is what the Common Core Standards require. The pressure of implementing the CC reading standard is leading many kindergarten teachers to resort to inappropriate drilling on specific skills and excessive testing. Teacher-led direct instruction in kindergarten has almost entirely replaced the active, play-based experiential learning that we know children need.

Defending the Early Years and Alliance for Childhood are calling for the withdrawal of the kindergarten standards from the Common Core so they can be rethought along developmental lines. You can read the full report and watch a video, along with calls to action on the DEY website:

Find the full report at: http://www.DEYproject.org .

The video: http://youtu.be/DVVln1WMz0g

If you want to tweet your support, use this hashtag:

#2much2soon

Here are some suggested tweets:
#EarlyEd experts @dey_project @4childhood conclude #CCSS Kinder reading requirement is #2much2soon http://youtu.be/DVVln1WMz0g

and

Why @dey_project @4childhood call for withdrawal of kinder standards from #CCSS http://youtu.be/DVVln1WMz0g

Bill Gates convened 1,000 people in Seattle, where he admitted that his big global health challenge has not produced significant gains in the Third World. Educators may recognize parallels to the Gates’ involvement in Common Core, where the foundation looked for a technological fix to complex human, social, and economic problems.
.

“When he took the stage this fall to celebrate the 10th anniversary of his signature global health research initiative, Bill Gates used the word “naive” — four times — to describe himself and his charitable foundation.
It was a surprising admission coming from the world’s richest man.

“But the Microsoft co-founder seemed humbled that, despite an investment of $1 billion, none of the projects funded under the Gates Foundation’s “Grand Challenges” banner has yet made a significant contribution to saving lives and improving health in the developing world….

“Not only did he underestimate some of the scientific hurdles, Gates said. He and his team also failed to adequately consider what it would take to implement new technologies in countries where millions of people lack access to basic necessities such as clean water and medical care….

“Among his favorite projects is an effort to eliminate Dengue fever by infecting mosquitoes with bacteria that block disease transmission. Another is a spinoff biotech working on a probiotic to cure cholera.

“But critics say projects like those demonstrate the foundation’s continuing emphasis on technological fixes, rather than on the social and political roots of poverty and disease.

“The main harm is in the opportunity cost,” said Dr. David McCoy, a public-health expert at Queen Mary University, London. “It’s in looking constantly for new solutions, rather than tackling the barriers to existing solutions.”

“The toll of many diseases could be lowered simply by strengthening health systems in developing countries, he said. Instead, programs like Grand Challenges — heavily promoted by the Gates Foundation’s PR machine — divert the global community’s attention from such needs, McCoy argues.”

Reformers have framed their narrative around the myth of “the bad teacher, without whom all children would make A’s in every subject every year. With this false narrative, they have promoted lengthy, tme-wasting evaluations to find and fire these academic frauds.

The narrative itself is the fraud. Like every profession, there are good and bad practitioners. Some teachers are excellent in some settings, not in others. We count on qualified administrators–not algorithms–to evaluate their staffs.

But now comes another reason to doubt the reformers’ narrative. A new study shows that the quality of teachers has been increasing over the past 15 years.

The abstract says:

“The relatively low status of teaching as a profession is often given as a factor contributing to the difficulty of recruiting teachers, the middling performance of American students on international assessments, and the well-documented decline in the relative academic ability of teachers through the 1990s. Since the turn of the 21st century, however, a number of federal, state, and local teacher accountability policies have been implemented toward improving teacher quality over the objections of some who argue the policies will decrease quality. In this article, we analyze 25 years of data on the academic ability of teachers in New York State and document that since 1999 the academic ability of both individuals certified and those entering teaching has steadily increased. These gains are widespread and have resulted in a substantial narrowing of the differences in teacher academic ability between high- and low-poverty schools and between White and minority teachers. We interpret these gains as evidence that the status of teaching is improving.”

This is a terrific article by civil rights attorney Wendy Lecker about the madness of our nation’s obsession with standardized testing.

 

She writes:

 

Last year, President Barack Obama committed hundreds of millions of dollars to brain research, stressing the importance of discovering how people think, learn and remember. Given the priority President Obama places on the brain in scientific research, it is sadly ironic that his education policies ignore what science says is good for children’s brains.

It is well known that play is vital in the early grades. Through play, kindergarteners develop their executive function and deepen their understanding of language. These are the cornerstones of successful reading and learning later on.

At-risk children often arrive at school having heard fewer words than more advantaged children. This deficit puts at-risk children behind others in learning to read. Scientists at Northwestern have recently shown that music training in the early years helps the brain improve speech processing in at-risk children.

Scientists at the University of Illinois have demonstrated that physical activity, coupled with downtime, improves children’s cognitive functions.

Scientists have also shown that diversity makes people more innovative. Being exposed to different disciplines broadens a student’s perspective. More importantly, working with a people from different backgrounds increases creativity and critical thinking.

These proven paths to healthy brain development are blocked by Obama’s education policies, the most pernicious of which is the overemphasis on standardized tests.

Despite paying lip service to the perils of over-testing, our leaders have imposed educational policies ensuring that standardized tests dominate schooling. Though standardized tests are invalid to measure teacher performance, the Obama administration insists that students’ standardized test scores be part of teacher evaluation systems. Both under NCLB and the NCLB waivers, schools are rated by standardized test scores. Often, a high school diploma depends at least in part on these tests. When so much rides on a standardized test scores, tests will drive what is taught and learned.

Just last month, U.S. Education Secretary Arne Duncan declared that yearly standardized testing is essential to monitor children’s progress. Setting aside the fact the new Common Core tests have not been proven to show what children learn, data shows that a child who passes a standardized test one year is overwhelmingly likely to pass the next year. Therefore, yearly standardized testing is unnecessary.

 

She adds:

 

The result? More than 10 years of high-stakes test-based education policy under NCLB and the waivers has narrowed curricula. Schools de-emphasize any subject other than language arts and math. In kindergarten, play has all but been eliminated in favor of direct instruction, and social studies, art, music, science, physical education and other subjects are disappearing. School districts at all grade levels are forced to reduce or eliminate these subjects to pay for implementation of the Common Core and its testing regime. Lansing Michigan last year eliminated art, music and physical education from elementary schools and the state of Ohio is considering the same. Recess has disappeared from many schools. The Obama administration promotes policies that increase school segregation yet have questionable educational value, like school choice. Consequently, school segregation continues to rise.

 

If we don’t end our obsession with picking the right bubble, marking the right box, we will ruin the education of a generation of children.

 

 

Professor Francesca Lopez of the University of Arizona responded to Betts and Tang’s critique of her post on the website of the National Education Policy Center.

 

 

 

She writes:

 

 

 

In September, the National Education Policy Center (NEPC) published a think-tank review I wrote on a report entitled, “A Meta-Analysis of the Literature on the Effect of Charter Schools on Student Achievement,” authored by Betts and Tang and published by the Center on Reinventing Public Education. My review examined the report, and I took the approach that a person reading the report and the review together would be in a better position to understand strengths and weaknesses than if that person read the report in isolation. While my review includes praise of some elements of the report, there is no question that the review also points out flawed assumptions and other areas of weakness in the analyses and the presentation of those analyses. The authors of the report subsequently wrote a prolonged rebuttal claiming I misrepresent their analysis and essentially reject my criticisms.

 

The rebuttal takes up 13 pages, which is considerably longer than my review. Yet these pages are largely repetitive and can be addressed relatively briefly. In the absence of sound evidence to counter the issues raised in my review, the rebuttal resorts to lengthy explanations that obscure, misrepresent, or altogether evade my critiques. What seems to most strike readers I’ve spoken with is the rebuttal’s insulting and condescending tone and wording. The next most striking element is the immoderately recurrent use of the term “misleading,” which is somehow repeated no fewer than 50 times in the rebuttal.

 

Below, I respond to each so-labeled “misleading statement” the report’s authors claim I made in my review—all 26 of them. Overall, my responses make two primary points:

 

 The report’s authors repeatedly obscure the fact that they exaggerate their findings. In their original report, they present objective evidence of mixed findings but then extrapolate their inferences to support charter schools. Just because the authors are accurate in some of their descriptions/statements does not negate the fact that they are misleading in their conclusions.

 

 The authors seem to contend that they should be above criticism if they can label their approaches as grounded in “gold standards,” “standard practice,” or “fairly standard practice.” When practices are problematic, they should not be upheld simply because someone else is doing it. My task as a reviewer was to help readers understand the strengths and weaknesses of the CRPE report. Part of that task was to attend to salient threats to validity and to caution readers when the authors include statements that outrun their evidence.

 

One other preliminary point, before turning to specific responses to the rebuttal’s long list. I am alleged by the authors to have insinuated that, because of methodological issues inherent in social science, social scientists should stop research altogether. This is absurd on its face, but I am happy to provide clarification here: social scientists who ignore details that introduce egregious validity threats (e.g., that generalizing from charter schools that are oversubscribed will introduce bias that favors charter schools) and who make inferences on their analyses that have societal implications, despite their claims of being neutral, should act more responsibly. If unwilling or unable to do so, then it would indeed be beneficial if they stopped producing research.

 

What follows is a point-by-point response to the authors’ rebuttal. For each point, I briefly summarize those contentions, but readers are encouraged to read the full 13 pages. The three documents – the original review, the rebuttal, and this response – are available at http://nepc.colorado.edu/thinktank/review-meta-analysis-effect-charter. The underlying report is available at http://www.crpe.org/publications/meta-analysis- literature-effect-charter-schools-student-achievement.

 

#1. The authors claim that my statement, “This report attempts to examine whether charter schools have a positive effect on student achievement,” is misleading because: “In statistics we test whether we can maintain the hypothesis of no effect of charter schools. We are equally interested in finding positive or negative results.” It is true that it is the null hypothesis that is tested. It is also true that the report attempts to examine whether charter schools have a positive effect on student achievement.

 

Moreover, it is telling that when the null hypothesis is not rejected and no assertion regarding directionality can be made, the authors still make statements alluding to directionality (see the next “misleading statement”).

 

#2. The authors object to my pointing out when they claim positive effects when their own results show those “effects” to not be statistically significant. There is no question that the report includes statements that are written in clear and non-misleading ways. Other statements are more problematic. Just because the authors are accurate in some of their descriptions does not negate my assertion that they make “[c]laims of positive effects when they are not statistically significant.” They tested whether a time trend was significant; it was not. They then go on to say it is a positive trend in the original report, and they do it again in their rebuttal: “We estimate a positive trend but it is not statistically significant.” This sentence is misleading. As the authors themselves claim in the first rebuttal above, “In statistics we test whether we can maintain the hypothesis of no effect.” This is called null hypothesis statistical testing (NHST). In NHST, if we reject the null hypothesis, we can say it was positive/negative, higher/lower, etc. If we fail to reject the null hypothesis (what they misleadingly call “maintain”), we cannot describe it in the direction that was tested because the test told us there isn’t sufficient support to do that. The authors were unable to reject the null hypothesis, but they call it positive anyway. Including the caveat that it is not significant does not somehow lift them above criticism. Or, to put this in the tone and wording of the authors’ reply, they seem “incapable” of understanding this fundamental flaw in their original report and in their rebuttal. There is extensive literature on NHST. I am astonished they are “seemingly unaware” of it.

 

#3. My review pointed out that the report shows a “reliance on simple vote-counts from a selected sample of studies,” and the authors rebut this by claiming my statement “insinuates incorrectly that we did not include certain studies arbitrarily.” In fact, my review listed the different methods used in the report, and it does use vote counting in a section, with selected studies. My review doesn’t state or imply that they were arbitrary, but they were indeed selected.

 

#4. The authors also object to my assertion that the report includes an “unwarranted extrapolation of the available evidence to assert the effectiveness of charter schools.” While my review was clear in stating that the authors were cautious in stating limitations, I also pointed to specific places and evidence showing unwarranted extrapolation. The reply does not rebut the evidence I provided for my assertion of extrapolation.

 

#5. My report points out that the report “… finds charters are serving students well, particularly in math. This conclusion is overstated; the actual results are not positive in reading and are not significant in high school math; for elementary and middle school math, effect sizes are very small…” The authors contend that their overall presentation of results is not misleading and that I was wrong (in fact, that I “cherry picked” results and “crossed the line between a dispassionate scientific analysis and an impassioned opinion piece”) by pointing out where the authors’ presentation suggested pro-charter results where unwarranted. Once again, just because the authors are accurate in some of their descriptions does not negate my assertion that the authors’ conclusions are overstated. I provided examples to support my statement that appear to get lost in the authors’ conclusions. They do not rebut my examples, but instead call it “cherry picking.” I find it telling that the authors can repeatedly characterize their uneven results as showing that charters “are serving students well” but if I point to problems with that characterization it is somehow I, not them, who have “crossed the line between a dispassionate scientific analysis and an impassioned opinion piece.”

 

#6. I state in my review that the report includes “lottery-based studies, considering them akin to random assignment, but lotteries only exist in charter schools that are much more popular than the comparison public schools from which students are drawn. This limits the study’s usefulness in broad comparisons of all charters versus public schools.” The rebuttal states, “lottery-based studies are not ‘akin’ to random assignment. They are random assignment studies.” The authors are factually wrong. Lottery-based charter assignments are not random assignment in the sense of, e.g., random assignment pharmaceutical studies. I detail why this is so in my review, and I would urge the authors to become familiar with the key reason lottery-based charters are not random assignment: weights are allowed. The authors provided no evidence that the schools in the study did not use weights, thus the distinct possibility exists that various students do not have the same chance of being admitted, and are therefore, not randomly assigned. The authors claim charter schools with lotteries are not more popular than their public school counterparts. Public schools do not turn away students because seats are filled; their assertion that charters do not need to be more popular than their public school counterparts is unsubstantiated. Parents choose a given charter school for a reason – oftentimes because the neighborhood school and other charter school options are less attractive. But beyond that, external validity (generalizing these findings to the broader population of charter schools) requires that over-enrolled charters be representative of charters that aren’t over-enrolled. That the authors test for differences does not negate the issues with their erroneous assumptions and flatly incorrect statements about lottery-based studies.

 

#7. The authors took issue with my critique that their statement, “One conclusion that has come into sharper focus since our prior literature review three years ago is that charter schools in most grade spans are outperforming traditional public schools in boosting math achievement” is an overstatement of their findings. In their rebuttal, they list an increase in the number of significant findings (which is not surprising given the larger sample size), and claim effect sizes were larger without considering confidence intervals around the reported effects. In addition to that, the authors take issue with my critique of their use of the word “positive” in terms of their non-significant trend results, which I have already addressed in #2.

 

#8. The authors take issue with my finding that their statement, “…we demonstrated that on average charter schools are serving students well, particularly in math” (p. 36) is an overstatement. I explained why this is an overstatement in detail in my review.

 

#9. The authors argue, “Lopez cites a partial sentence from our conclusion in support of her contention that we overstate the case, and yet it is she who overstates.” The full sentence that I quoted reads, “But there is stronger evidence of outperformance than underperformance, especially in math.” I quoted that full sentence, sans the “[b]ut.” They refer to this as “chopping this sentence in half,” and they attempt to defend this argument by presenting this sentence plus the one preceding it. In either case, they fail to support their contention that they did not overstate their findings. Had the authors just written the preceding sentence (“The overall tenor of our results is that charter schools are in some cases outperforming traditional public schools in terms of students’ reading and math achievement, and in other cases performing similarly or worse”), I would not have raised an issue. To continue with “But there is stronger evidence of outperformance than underperformance, especially in math” is an ideologically grounded overstatement.

 

#10. The authors claim, “Lopez seriously distorts our work by comparing results from one set of analyses with our conclusions from another section, creating an apples and oranges problem.” The section the authors are alluding to reported results of the meta- analysis. I pointed out examples of their consistent exaggeration. The authors address neither the issue I raise nor the support I offer for my assertion that they overstate findings. Instead, they conclusively claim I am “creating an apples and oranges problem.”

 

#11. The authors state, “Lopez claims that most of the results are not significant for subgroups.” They claim I neglected to report that a smaller sample contributed to the non-significance, but they missed the point. The fact that there are “far fewer studies by individual race/ethnicity (for the race/ethnicity models virtually none for studies focused on elementary schools alone, middle schools alone, or high schools) or other subgroups” is a serious limitation. The authors claim that “This in no way contradicts the findings from the much broader literature that pools all students.” However, the reason ethnicity/race is an important omission is because of the evidence of the segregative effects of charter schools. I was clear in my review in identifying my concern: the authors’ repeated contentions about the supposed effectiveness of charter schools, regardless of the caution they maintained in other sections of their report.

 

#12. The authors argue, “The claim by Lopez that most of the effects are insignificant in the subgroup analyses is incomplete in a way that misleads. She fails to mention that we conduct several separate analyses in this section, one for race/ethnicity, one for urban school settings, one for special education and one for English Learners.” Once again, the authors miss the point, as I explain in #11. The authors call my numerous examples that discredit their claims “cherry picking.” The points I raise, however, are made precisely to temper the claims made by the authors. If cherry-picking results in a full basket, perhaps there are too many cherries to be picked.

 

#13. The authors take issue that I temper their bold claims by stating that the effects they found are “modest.” To support their rebuttal, they explain what an effect of .167 translates to in percentiles, which I argued against in my review in detail. (The authors chose to use the middle school number of .167 over the other effect sizes, ranging from .023 to .10; it was the full range of results that I called “modest.”) Given their reuse of percentiles to make a point, it appears the authors may not have a clear understanding of percentiles: they are not interval-level units. An effect of .167 is not large given that it may be negligible when confidence intervals are included. That it translates into a 7 percentile “gain” when percentiles are not interval level units (and confidence bands are not reported) is a continued attempt to mislead by the authors. I detail the issues with the ways the authors present percentiles in my review. (This issue is revisited in #25, below.)

 

#14. The authors next take issue with the fact I cite different components of their report that were “9 pages apart.” I synthesized the lengthy review (the authors call it “conflating”), and once again, the authors attempt to claim that my point-by-point account of limitations with their report is misleading. Indeed, according to the authors, I am “incapable of understanding” a “distinction” they make. In their original 68-page report, they make many “distinctions.” They appear “incapable of understanding” that the issues I raise concerning “distinctions” is that they were reoccurring themes in their report.

 

#15. The authors next find issue with the following statement: “The authors conclude that ‘charter schools appear to be serving students well, and better in math than in reading’ (p. 47) even though the report finds ‘…that a substantial portion of studies that combine elementary and middle school students do find significantly negative results in both reading and math – 35 percent of reading estimates are significantly negative, and 40 percent of math estimates are significantly negative (p. 47)’.” This is one of the places where I point out that the report overstates conclusions notwithstanding their own clear findings that should give them caution. In their rebuttal, the authors argue that I (in a “badly written paragraph”) “[insinuate] that [they] exaggerate the positive overall math effect while downplaying the percentage of studies that show negative results.” If I understand their argument correctly, they are upset that I connected the two passages with “even though the report finds” instead of their wording: “The caveat here is”. But my point is exactly that the caveat should have reigned in the broader conclusion. They attempt to rebut my claim by elaborating on the sentence, yet they fail to address my critique. The authors’ rebuttal includes, “Wouldn’t one think that if our goal had been to overstate the positive effects of charter schools we would never have chosen to list the result that is the least favorable to charter schools in the text above?” I maintain the critique from my review: despite the evidence that is least favorable to charter schools, the authors claim overall positive effects for charter schools—obscuring the various results they reported. Again, just because they are clear sometimes does not mean they do not continuously obscure the very facts they reported.

 

#16. The authors take issue with the fact that my review included two sentences of commentary on a companion CRPE document that was presented by CRPE as a summary of the Betts & Tang report. As is standard with all NEPC publications, I included an endnote that included the full citation of the summary document, clearly showing an author (“Denice, P.”) other than Betts & Tang. Whether Betts & Tang contributed to, approved, or had nothing to do with the summary document, I did not and do not know.

 

#17. The next issue the authors have is that I critiqued their presentation and conclusions based on the small body of literature they included in their section entitled, “Outcomes apart from achievement.” The issue I raise with the extrapolation of findings can be found in detail in the review. The sentence from the CRPE report that seems to be the focus here reads as follows, “This literature is obviously very small, but both papers find evidence that charter school attendance is associated with better noncognitive outcomes.” To make such generalizations based on two papers (neither of which was apparently peer reviewed) is hardly an examination of the evidence that should be disseminated in a report entitled, “A Meta-Analysis of the Literature on the Effect of Charter Schools on Student Achievement.” The point of the meta-analysis document is to bring together and analyze the research base concerning charter schools. The authors claim that because they are explicit in stating that the body of literature is small, that their claim is not an overstatement. As I have mentioned before, just because the authors are clear in their caveats, making assertions about the effects of charter schools with such limited evidence is indeed an overstatement. We are now seeing more and more politicians who offer statements like, “I’m not a scientist and haven’t read the research, but climate change is clearly a hoax.” The caveats do little to transform the ultimate assertion into a responsible statement.

 

Go to the link to read the rest of Professor Lopez’s response to Betts and Tang, covering the 26 points they raised.

On September 30, Francesca Lopez of the University of Arizona reviewed a study of charter schools by Julian R. Betts and Y. Emily Tang of the University of California at San San Diego

Betts and Tang here respond to Lopez’ critique of their study of charter school achievement..

The critical study by Lopez was published by the National Education Policy Center and posted on this blog.

Betts and Tang say that Lopez misrepresented their study. They write:

“First, what did we find in our meta-analysis of the charter school effectiveness literature? On average, charter school studies are revealing a positive and statistically significant difference between math achievement at charter schools and traditional public schools. We also find a positive difference for reading achievement, but this difference is not statistically significant. Second, we devote much of our paper to studying not the mean effect, but the variation across studies in the effect of attending a charter school. We find that charter schools’ effectiveness compared to nearby traditional public schools varies substantially across locations.

“What is the central claim of Lopez? She writes: “The report does a solid job describing the methodological limitations of the studies reviewed, then seemingly forgets those limits in the analysis” (p. 1). She uses words like “exaggeration” and “overstated” (p. 8) to characterize our analysis of the literature, and implies that our conclusions are not “reserved,” “responsible,” (p. 7) or “honest” (p. 7 and p. 8).

“Throughout her essay, Lopez falsely projects intentions in our words that simply are not there. We encourage interested readers to review the words that we actually wrote, in their full context, in our abstract, main paper, and our conclusion. We are confident that readers will confirm for themselves that any “overstated” conclusions of which Lopez accuses us are imagined.”

“There are serious problems with Lopez’s arguments. First, she habitually quotes our work in a selective and misleading way. Such rhetorical slights, in which she quotes one of our sentences while ignoring the highly relevant adjacent sentences, or even cutting important words out of our sentences, overlook important parts of our analysis and result in a highly inaccurate presentation of our work. Second, her analysis contains six technical errors. These technical mistakes, some quite serious, invalidate many of Professor Lopez’s claims. An appendix to this essay exposes more than two dozen misleading or outright incorrect statements that Lopez makes in a mere 9-page essay. To give readers a sense of the scope and severity of these problems, consider the following examples:

“Example 1: A Partial and Misleading Quotation

“Lopez insinuates that we exaggerate the positive overall math effect while downplaying the percentage of studies that show negative results. She writes:
“The authors conclude that ‘charter schools appear to be serving students well, and better in math than in reading’ (p. 47) even though the report finds ‘…that a substantial portion of studies that combine elementary and middle school students do find significantly negative results in both reading and math—35 percent of reading estimates are significantly negative, and 40 percent of math estimates are significantly negative (p. 47)’”

“Here is what we actually wrote on page 47: “Examining all of these results as separate parts of a whole, we conclude that, overall, charter schools appear to be serving students well, and better in math than in reading. The caveat here is that a substantial portion of studies that combine elementary and middle school students do find significantly negative results in both reading and math—35 percent of reading estimates are significantly negative, and 40 percent of math estimates are significantly negative.”

“Lopez uses two rhetorical devices to lead readers to the perception that we overstated findings. First, she separates the two quotations, implying that we are somehow hiding the second result, when in fact we intentionally mention the positive overall mean math effect and the variation in the results across studies side by side. Second, she further misleads the reader by again cutting out part of our sentence. Instead of stating that we have a “caveat” to the positive mean math effect she removes that entire clause.

“What makes the approach of Lopez even more misleading is that in the paragraph above, we were bending over backwards to be fair. We cite only one type of study in that quotation: those that combine elementary and middle schools. (These account for about 1/7th of all the studies.) Why did we focus only on those studies in the above quotation? Because these studies were the exception to our conclusion—the ones that produced the highest percentage of studies with negative and significant estimates. Wouldn’t one think that if our goal had been to overstate the positive effects of charter schools we would never have chosen to list the result that is the least favorable to charter schools in the text above? For example, we could have stated that for elementary school studies, only 12% showed negative and significant reading results, compared to 71% showing positive and significant results. Or we could have stated that only 11% of elementary school studies showed negative and significant math results, while 61% showed positive and significant results in math.

“Lopez fails to list any of the more positive results from the other grade span categories studied that led us to our overall conclusion. We noted the exception above precisely because it was an exception. While it is worth noting, it does not refute the other evidence. By citing an exception as a reason to dismiss all of the other results, Lopez misses the main point of a statistical meta-analysis. This is a consistent pattern throughout her essay.”

Betts and Tang make 26 points about the flaws of Lopez’s analysis.