Archives for the month of: August, 2014

We have become accustomed in recent years to seeing films in which teachers are shown as lazy, greedy slugs. This fits nicely with the corporate reform narrative that seeks to strip all honor, dignity, and rights from teachers. Teachers don’t deserve those mean-spirited caricatures, nor the treatment they receive from legislatures.

Remembering Robin Williams’ portrayal of English teacher John Keating in “The Dead Poets’ Society” takes us back to another era, a time when the teacher might be seen as a source of wisdom and inspiration, a rebel and a non-conformist. Here is the trailer. Robin Williams represented the teacher as the best that one could hope to be: not just a man who taught language and literature but a man who changed lives.

My favorite scene in the movie occurs when Mr. Keating invites the class to read the introduction to the poetry anthology. The introduction describes a mathematical formula for judging the worth of a poem. Mr. Keating tells his students to “tear out the entire introduction! Rip! Tear! Rip!”

Now as I read about the econometricians who have developed algorithms to determine who are the best and the worst teachers, I will think of Mr. Keating–Robin Williams– telling his students “Rip!” Live life to the fullest! Dare to be yourself! He was–or he acted–the teacher of our dreams, the one who inspired us to be our own best selves, to defy authority when it is wrong, and to live lives of possibility, not lives weighted down by the routine. Now as I see the purveyors of Big Data descending upon students, teachers, principals, and superintendents alike, ready to label them, rank them, crunch their numbers and their souls, I will think of Robin Williams as the irreverent Mr. Keating. I know what he would have done with those forms and spreadsheets. “Rip!”

These images remain 25 years after seeing “The Dead Poets’ Society” because Robin Williams was that teacher. Not just in one film but in dozens of films.

In the past few days, there has been vigorous debate about whether California’s Commission on Teacher Credentialing was lowering standards; some critics went so far as to claim that future teachers in core subject areas would no longer need a four-year degree of any kind. None of this was true. I reached out to friends across California who assured me that it was untrue. Just this morning, the Sacramento Bee published an article by Linda Darling-Hammond, chair of the CCTC, explaining exactly what transpired.

D-H explained that standards were not lowered and remain high in every field.

Here are the changes that alarmed teachers of physical education:

“The commission did vote to provide an authorization within the physical education credential for ROTC teachers. Those who pass a basic skills test and a test of PE knowledge and complete 135 hours of teacher preparation can earn this authorization. If they are to teach English learners, they must also meet the requirements for an English learner authorization. Under the new authorization, these individuals can still teach only ROTC courses. They cannot teach regular PE classes. However, they will have a stronger knowledge base in physical education, which will help them serve their students.

“Some of these instructors do not hold a bachelor’s degree, but they have relevant job experience that California has long recognized in lieu of a degree for ROTC instructors and other specific groups of teachers, like those from industry who teach in Career Technical Education programs.”

In short, ROTC instructors will not replace PE teachers.

Be it noted that Governor Brown is a strong supporter of ROTC, having started a military institute charter school when he was mayor of Oakland. With crucial decisions about the Vergara trial hanging in the balance, this is not a propitious moment to pick a fight with the governor when so little is at stake. He respects teachers; he loves learning. Message to California teachers: Unite on the big issues.

Read more here: http://www.sacbee.com/2014/08/12/6620786/another-view-teacher-standards.html#storylink=cpy

Federal law states clearly that no agent of the federal government may seek to influence, direct, or control curriculum or instruction. For many decades, both parties agreed that they did not want the party in power to use federal power to control the schools of the nation. Thus, while it was appropriate for the U.S. Department of Education to use its funding to enforce Supreme Court decisions to desegregate the schools, it was prohibited from seeking to control curriculum and instruction. Both parties recognized that education is a state and local function, and neither trusted the other to impose its ideas on the schools.

 

That explains why Arne Duncan did not use federal funding to pay for the Common Core, but it does not explain why he used the power of his office to promote the CCSS or why he paid out some $350 million for tests specifically designed to test the Common Core standards. As every teacher knows, tests drive curriculum and instruction, especially when the tests are connected to high stakes.

 

In this post, Mercedes Schneider explains the battle royal in Louisiana, where Governor Jindal is fighting the State Commissioner of Education John White and the Board of Elementary and Secondary Education over the CCSS and the aligned PARCC tests. Now Jindal has decided to sue on grounds that the U.S. Department of Education acted illegally by aiding the creation of CCSS and the tests. The funniest part of the post , as Schneider writesis to see politicians accusing other politicians of acting like politicians.

 

I hope the underlying issues get a full airing. When I worked at the U.S. Department of Education in the early 1990s in the administration of President George H.W. Bush, we were much aware of the ban on federal involvement in curriculum and instruction. We funded voluntary national standards, but we kept our distance from the professional associations working to write them, and they were always described as voluntary national standards.

Lisa Woods, who has taught for 25 years, explains clearly in this post why schools will never run like businesses. It originally appeared. In the Greensboro (N.C.) News-Record.

 

She asks readers to imagine a job where one’s compensation depends one’s “job performance and value” depend on the following conditions:

 

 

 

 

 

“* You are meeting with 35 clients in a room designed to hold 20.

“* The air conditioning and/or heat may or may not be working, and your roof leaks in three places, one of which is the table where your customers are gathered.

“* Of the 35, five do not speak English, and no interpreters are provided.

“* Fifteen are there because they are forced by their “bosses” to be there but hate your product.

“* Eight do not have the funds to purchase your product.

“* Seven have no prior experience with your product and have no idea what it is or how to use it.

“* Two are removed for fighting over a chair.

“* Only two-thirds of your clients appear well-rested and well-fed.

“You are expected to:

“* Make your presentation in 40 minutes.

“* Have up-to-date, professionally created information concerning your product.

“* Keep complete paperwork and assessments of product understanding for each client and remediate where there is lack of understanding.

“* Use at least three different methods of conveying your information: visual, auditory and hands-on.

“The “criterion” for measuring your “worth and value” is that no less than 100 percent of your clients must buy and have the knowledge to assemble and use your product, both creatively and critically, and in conjunction with other products your company produces, of which you have working but limited knowledge

“Only half of the clients arrive with the necessary materials to be successful in their understanding of your product, and your presentation is disrupted at least five times during the 40 minutes.

“You have an outdated product manual and one old computer, but no presentation equipment. Your company’s budget has been cut every year for the past 10 years, the latest by a third. Does this mean you only create two-thirds of a presentation? These cuts include your mandatory training and presentation materials (current ones available to you are outdated by five years).

“You have no assistant, and you must do all the paperwork, research your knowledge deficiencies and produce professional-looking, updated materials during the 40 minutes allotted to you during the professional day. You cannot use your 30-minute lunch break. Half is spent monitoring other clients who are not your own.

“Your company cannot afford to train you in areas of its product line where you may be deficient, yet you are expected to have this knowledge and incorporate it into your product presentation in a meaningful way.

“You haven’t had a raise in eight years and your benefits have been purged, nor do you receive a commission for any product you sell. Do you purchase all the materials needed so your presentation is effective? Will you pay for the mandatory training necessary to do your job in a competent and professional manner?”

What business could succeed under those conditions?

Join educators and community leaders for an important one day Summer Institute to provide an alternate pathway to the test based accountability system. On August 21st, leaders will convene at Dowling College from 8:30am- 3:00 pm to explore the work of Michael Fullan and Andy Hargreaves (Professional Capital) as they work collaboratively to build the social capital needed to confront the de-professionalization of teaching.

For more information please contact David Gamberg at: dgamberg@southoldufsd.com, or go to the following website for details:

http://www.schoolleadership20.com/events/developing-an-alternative-path-for-school-improvement-summer-inst

The election for the Los Angeles School Board seat in District 1 is Tuesday August 12.

If you live in Board District 1, be sure to vote.

Monica Ratliff endorsed veteran educator George McKenna. So did Steve Zimmer.

Here is Steve Zimmer’s statement:

I am posting this on Facebook in the hopes that each of my friends who lives in LAUSD Board District 1, works in Board District 1 or knows folks who live in Board District 1 will urge people to vote in Tuesday’s Special Election. This special election, to fulfill the unexpired term of my friend and colleague Ms. La Motte, is expected to have an incredibly low voter turnout. I ask you first and foremost to vote on Tuesday. We must show that public education is important for every student in every community.

As many of you know, I have endorsed Dr. George McKenna in this race. I do not make endorsements lightly. I respect the rights of voters, especially when it is not my district, to make their own decisions. When I saw the attacks against Dr. McKenna, however, I decided I could not stand by and watch. As I saw the same organizations that spent over $4 million to take me out, now raising over $1 million to spread lies and cast doubts about Dr. McKenna’s career, I could not look away. TheLAUSD school board is not for sale. It is immoral to seek victory by any means necessary. It is unconscionable to use the injuries of children as a political weapon. In some of the same ways as in my election, our core democratic values hang in the balance as ballots are cast on Tuesday.

These were the reasons I endorsedDr. McKenna two weeks ago. Since that time, I have come to know George McKenna in a different way than I had working with him over the last 20 years. I have seen the man behind the Superintendent. I have seen the values and reasoning behind his decisions. I had always respected Dr. McKenna. From the moment I met him in 1992 as a first year teacher, I knew that Dr. McKenna was a consummate educator in an ongoing struggle for civil rights and education rights for all students.It is a very hard thing to do that from within public school systems, so I had always respected his skills and tenacity. But now that respect has grown into trust. I trust that Dr. McKenna is running for all the right reasons, and It rust that his understanding of the needs of children in district one and throughout LAUSD will make Dr. McKenna an outstanding board member.

Specifically,

· I trust Dr. McKenna will put the needs of students and their families first in every decision he makes as a Board Member

· I trust Dr. McKenna has absolute urgency to transform our schools so they meet needs of students and families for whom the promise of public education has yet to be realized

· I trust Dr. McKenna understands the complexity and diversity of District One and that he will serve all areas of the district; balancing the charge to concentrate resources in areas of greatest need with ensuring that all schools are well supported

· I trust Dr. McKenna is truly independent. He will not be beholden to any interest group or endorser and has shown this over and over again during the campaign. He is honest and transparent because he is an educator not a politician.

· I trust that Dr. McKenna will not blame teachers for the problems facing public education. He knows that our challenges do not have simple answers and that we all must work together. He will hold every employee to the highest expectations while respecting their dignity and honoring their service.

These are the qualities and values we need in our next school board member. Dr. McKenna has both the moral compass and the experiential direction to lead LAUSD towards a new day for children and their schools. It has been an honor to get to know a man I have admired for so long. I ask you to spread the word about this election and aboutDr. McKenna. I encourage you to visit his website and listen to one of McKenna’s speeches.

If you should have questions, feel free to email me at szdeke@hotmail.com

Please feel free to share this note and thank you for reading and caring about public education.

AFT President Randi Weingarten Calls for Full Release of Test Questions

WASHINGTON— Statement of AFT President Randi Weingarten following news that a portion of the Common Core-aligned testing questions were released in New York as teachers and community members protest the overuse of testing in Albany.

“Releasing just some of the Common Core-aligned test questions in the middle of the summer doesn’t cut it. Parents and educators repeatedly have called for the full release of the questions—even taking our call to the Pearson shareholder meeting this past spring.

“We renew our call for the full release of the test questions—in a timely manner and in a way that is most useful for parents, educators and kids—not in the middle of the summer and right before the test results are announced.”

###

District of Columbia Chancellor Kaya Henderson announced the suspension of the practice of evaluating teachers by their students’ test scores. This practice was considered the signal policy initiative of Henderson’s predecessor Michelle Rhee.

Henderson described the move “as necessary in order to allow students to acclimate themselves to new tests built around the standards established by the Common Core.”

The Gates Foundation applauded the retreat on test-based evaluation, as did Randi Weingarten of the AFT. The U.S. Department of Education released a statement expressing its disappointment. It said: “Although we applaud District of Columbia Public Schools (DCPS) for their continued commitment to rigorous evaluation and support for their teachers, we know there are many who looked to DCPS as a pacesetter who will be disappointed with their desire to slow down.” Since test-based evaluation is the centerpiece of Arne Duncan’s Race to the Top, this is surely a setback for Duncan and his theory of change. On the other hand, D.C.’s test scores have been stagnant since 2009, which does not speak well of test-based evaluation, whether pushed by Michelle Rhee or Arne Duncan.

By the way, Michelle Rhee has apparently changed her name to Michelle (Rhee) Johnson.

This is an important article, which criticizes and deconstructs the notorious VAM study by Chetty et al. I refer to it as notorious because it was reported on the first page of the New York Times before it was peer-reviewed; it was immediately presented on the PBS Newshour; and President Obama referred to its findings in his State of the Union address only weeks after it first appeared.

These miraculous events do not happen by accident. The study made grand claims for the importance of value-added measures of teacher quality, a keystone of Obama’s education policy. One of the authors told the New York Times that the lesson of the study was to fire teachers sooner rather than later. A few months ago, the American Statistical Association reacted to the study, not harshly, but made clear that the study was overstated, that the influence of teachers on the variability of test scores ranged from 1-14%, and that changes in the system would likely have more influence on students’ academic outcomes than attaching the scores of students to individual teachers.

I have said it before, and I will say it again: VAM is Junk Science. Looking at children as machine-made widgets and looking at learning solely as standardized test scores may thrill some econometricians, but it has nothing to do with the real world of children, learning, and teaching. It is a grand theory that might net its authors a Nobel Prize for its grandiosity, but it is both meaningless in relation to any genuine concept of education and harmful in its mechanistic and reductive view of humanity.

CHETTY, ET AL, ON THE AMERICAN STATISTICAL ASSOCIATION’S RECENT POSITION STATEMENT ON VALUE-ADDED MODELS (VAMs): FIVE POINTS OF CONTENTION

by Margarita Pivovarova, Jennifer Broatch & Audrey Amrein-Beardsley — August 01, 2014

Over the last decade, teacher evaluation based on value-added models (VAMs) has become central to the public debate over education policy. In this commentary, we critique and deconstruct the arguments proposed by the authors of a highly publicized study that linked teacher value-added models to students’ long-run outcomes, Chetty et al. (2014, forthcoming), in their response to the American Statistical Association statement on VAMs. We draw on recent academic literature to support our counter-arguments along main points of contention: causality of VAM estimates, transparency of VAMs, effect of non-random sorting of students on VAM estimates and sensitivity of VAMs to model specification.

INTRODUCTION

Recently, the authors of a highly publicized and cited study that linked teacher value-added estimates to the long-run outcomes of their students (Chetty, Friedman, & Rockoff, 2011; see also Chetty, et al., in press I, in press II) published a “point-by-point” discussion of the “Statement on Using Value-Added Models for Educational Assessment” released by the American Statistical Association (ASA, 2014). This once again brought the value-added model (VAM) and its use for increased teacher and school accountability to the forefront of heated policy debate.

In this commentary we elaborate on some of the statements made by Chetty, et al. (2014). We position both the ASA’s statement and Chetty, et al.’s (2014) response within the current academic literature. As well, we deconstruct the critiques and assertions advanced by Chetty, et al. (2014) by providing counter-arguments and supporting them by the scholarly research on this topic.

In doing so, we rely on the current research literature that has really been done on this subject over the past ten years. This more representative literature was completely overlooked by Chetty, et al. (2014), even though, paradoxically, they criticize the ASA for not citing the “recent” literature appropriately themselves (p. 1). With this being our first point of contention, we also discuss four additional points of dispute within the commentary.

POINT 1: MISSING LITERATURES

In their critique of the ASA statement, posted on a university-sponsored website, Chetty, et al. (2014) marginalize the current literature published in scholarly journals on the issues surrounding VAMs and their uses for measuring teacher effectiveness. Rather, Chetty et al. cite only works representing econometrician’s scholarly pieces, apparently in support of their a priori arguments and ideas. Hence, it is important to make explicit the rather odd and extremely selective literature Chetty, et al. included in the reference section of their critique, on which Chetty, et al. relied “to prove” some of the ASA’s statements incorrect. The whole set of peer-reviewed articles that counter Chetty, et al.’s arguments and ideas are completely left out of their discussion.

A search on the Educational Resources Information Center (ERIC) with “value-added” as key words for the same last five years yields 406 entries, and a similar search in Journal Storage (JSTOR, a shared digital library) returns 495. Chetty, et al., however, only cite 13 references to critique the ASA’s statement, one of which was the actual statement itself, leaving 12 external citations in total and in support of their critique. Of these 12 external citations, three are references to their two forthcoming studies and a replication of these studies’ methods; three have thus far been published in peer-reviewed academic journals, six were written by their colleagues at Harvard University; and 11 were written by teams of scholars with economics professors/econometricians as lead authors.

POINT 2: CORRELATION VERSUS CAUSATION

The second point of contention surrounds whether the users of VAMs should be aware of the fact that VAMs typically measure correlation, not causation. According to the ASA, as pointed out by Chetty, et al. (2014), effects “positive or negative—attributed to a teacher may actually be caused by other factors that are not captured in by the model” (p. 2). This is an important point with major policy implications. Seminal publications on the topic, Rubin, Stuart and Zanutto (2004) and Wainer (2004) who positioned their discussion within the Rubin Causal Model framework (Rubin, 1978; Rosenbaum and Rubin, 1983; Holland, 1986), clearly communicated, and evidenced, that value-added estimates cannot be considered causal unless a set of “heroic assumptions” are agreed to and imposed. Moreover, “anyone familiar with education will realize that this [is]…fairly unrealistic” (Rubin, et al. 2004, p. 108). Instead, Rubin, et al. suggested, given these issues with confounded causation, we should switch gears and evaluate interventions and reward incentives as based on the descriptive qualities of the indicators and estimates derived via VAMs. This point has since gained increased consensus among other scholars conducting research in these areas (Amrein-Beardsley, 2008; Baker, et al., 2010; Betebenner, 2009; Braun, 2008; Briggs & Domingue, 2011; Harris, 2011; Reardon & Raudenbush, 2009; Scherrer, 2011).

POINT 3: THE NON-RANDOM ASSIGNMENT OF STUDENTS INTO CLASSROOMS

The third point of contention pertains to Chetty, et al.’s statement that recent experimental and quasi-experimental studies have already solved the “causation versus correlation” issue. This claim is made despite the substantive research that evidences how the non-random assignment of students constrains VAM users’ capacities to make causal claims.

The authors of the Measures of Effective Teaching (MET) study cited by Chetty, et al. in their critique, clearly state, “we cannot say whether the measures perform as well when comparing the average effectiveness of teachers in different schools…given the obvious difficulties in randomly assigning teachers or students to different schools” (Kane, McCaffrey, Miller & Staiger, 2013, p. 38). VAM estimates were found to be biased for teachers who taught more relatively homogenous sets of students with lower levels of prior achievement, despite the levels of sophistication in the statistical controls used (Hermann, Walsh, Isenberg, & Resch, 2013; see also Ehlert, Koedel, Parsons, & Podgursky, 2014; Guarino et al., 2012).

Researchers repeatedly demonstrated that non-random assignment confounds value-added estimates independent of how many sophisticated controls are added to the model (Corcoran, 2010; Goldhaber, Walch, & Gabele, 2012; Guarino, Maxfield, Reckase, Thompson, & Wooldridge, 2012; Newton, Darling-Hammond, Haertel, & Thomas, 2010; Paufler & Amrein-Beardsley, 2014; Rothstein, 2009, 2010).

Even in experimental settings, it is still not possible to distinguish between the effects of school practice, which is of interest to policy-makers, and the effects of school and home context. There are many factors at the student, classroom, school, home, and neighborhood levels that would confound causal estimates that are beyond researchers’ control. Thus, the four experimental studies cited by Chetty, et al. (2014) do not provide ample evidence to refute the ASA on this point.

POINT 4: ISSUES WITH LARGE-SCALE STANDARDIZED TEST SCORES

In their position statement, ASA authors (2014) rightfully state that the standardized test scores used in VAMs should not be the only outcomes of interest for policy makers and stakeholders. Indeed, current agreement is that test scores might not even be one of the most important outcomes capturing a student’s educated self. Also, if value-added estimates from standardized test scores cannot be interpreted as causal, then the effect of “high value-added” teachers on college attendance, earnings, and reduced teenage birth rates cannot be considered causal either as opposed to what is implied by Chetty, et al. (2011; see also Chetty, et al., in press I, in press II).

Ironically, Chetty, et al. (2014) cite Jackson’s (2013) study to confirm their point that high value-added teachers also improve long-run outcomes of their students. Jackson (2013), however, actually found that teachers who are good at boosting test scores are not always the same teachers who have positive and long-lasting outcomes on non-cognitive skills acquisition. Moreover, value-added as related to test scores and non-cognitive outcomes for the same teachers were then, and have since been shown to be, weakly correlated with one another.

POINT 5: MODEL SPECIFICITY

Lastly, ASA (2014) expressed concerns about the sensitivity of value-added estimates to model specifications. Recently, researchers have found that value-added estimates are highly sensitive to the tests being used, even within the same subject areas (Papay, 2011) and the different subject areas taught by the same teachers given different student compositions (Loeb & Candelaria, 2012; Newton, et al., 2010; Rothstein, 2009, 2010). While Chetty, et al. rightfully noted that different VAMs typically yield correlations around r = 0.9, this is typical with most “garbage in, garbage out” models. These models are too often used, too often without question, to process questionable input and produce questionable output (Banchero & Kesmodel, 2011; Gabriel & Lester, 2012, 2013; Harris, 2011).

What Chetty, et al. overlooked, though, are the repeatedly demonstrated weak correlations between value-added estimates and other indicators of teacher quality, on average between r = 0.3 and 0.5 (see also Corcoran, 2010, Goldhaber et al., 2012; McCaffrey, Sass, Lockwood, & Mihaly, 2009; Broatch and Lohr, 2012; Mihaly, McCaffrey, Staiger, & Lockwood, 2013).

CONCLUSION

In sum, these are only a few “points” from this “point-by-point discussion” that would strike anyone even fairly familiar with the debate over the use and abuse of VAMs. These “points” are especially striking given the impact Chetty, et al.’s original (2011) study and now forthcoming studies (Chetty, et al., in press I, in press II) have already had on actual policy and the policy debates surrounding VAMs. Chetty, et al.’s (2014) discussion of the ASA statement, however, should cause others pause in terms of whether in fact Chetty, et al. are indeed experts in the field, or not. What certainly has become evident is that they do not have their minds wrapped around the extensive set of literature or knowledge on this topic. If they had, they may not have come off as so selective, as well as biased, citing only those representing certain disciplines and certain studies to support certain assumptions and “facts” upon which their criticisms of the ASA statement were based.

References

American Statistical Association. (2014). ASA Statement on using value-added models for educational assessment. Retrieved from http://www.amstat.org/policy/pdfs/ASA_VAM_Statement.pdf

Amrein-Beardsley, A. (2008). Methodological concerns about the Education Value-Added Assessment System (EVAAS). Educational Researcher, 37(2), 65–75. doi: 10.3102/0013189X08316420

Baker, E. L., Barton, P. E., Darling-Hammond, L., Haertel, E., Ladd, H. F., Linn, R. L., Ravitch, D., Rothstein, R., Shavelson, R. J., & Shepard, L. A. (2010). Problems with the use of student test scores to evaluate teachers. Washington, D.C.: Economic Policy Institute. Retrieved from http://www.epi.org/publications/entry/bp278

Banchero, S. & Kesmodel, D. (2011, September 13). Teachers are put to the test: More states tie tenure, bonuses to new formulas for measuring test scores. The Wall Street Journal. Retrieved from http://online.wsj.com/article/SB10001424053111903895904576544523666669018.html

Betebenner, D. W. (2009b). Norm- and criterion-referenced student growth. Education Measurement: Issues and Practice, 28(4), 42-51. doi:10.1111/j.1745-3992.2009.00161.x

Braun, H. I. (2008). Viccissitudes of the validators. Presentation made at the 2008 Reidy Interactive Lecture Series, Portsmouth, NH. Retrieved from http://www.cde.state.co.us/cdedocs/OPP/HenryBraunLectureReidy2008.ppt

Briggs, D. & Domingue, B. (2011, February). Due diligence and the evaluation of teachers: A review of the value-added analysis underlying the effectiveness rankings of Los Angeles Unified School District Teachers by the Los Angeles Times. Boulder, CO: National Education Policy Center. Retrieved from nepc.colorado.edu/publication/due-diligence

Broatch, J. and Lohr, S. (2012) “Multidimensional Assessment of Value Added by Teachers to Real-World Outcomes”, Journal of Educational and Behavioral Statistics, April 2012; vol. 37, 2: pp. 256–277.

Chetty, R., Friedman, J. N., & Rockoff, J. E. (2011). The long-term impacts of teachers: Teacher value-added and student outcomes in adulthood. Cambridge, MA: National Bureau of Economic Research (NBER), Working Paper No. 17699. Retrieved from http://www.nber.org/papers/w17699

Chetty, R., Friedman, J. N., & Rockoff, J. (2014). Discussion of the American Statistical Association’s Statement (2014) on using value-added models for educational assessment. Retrieved from http://obs.rc.fas.harvard.edu/chetty/ASA_discussion.pdf

Chetty, R., Friedman, J. N., & Rockoff, J. E. (in press I). Measuring the impact of teachers I: Teacher value-added and student outcomes in adulthood. American Economic Review.

Chetty, R., Friedman, J. N., & Rockoff, J. E. (in press II). Measuring the impact of teachers II: Evaluating bias in teacher value-added estimates. American Economic Review.

Corcoran, S. (2010). Can teachers be evaluated by their students’ test scores? Should they be? The use of value added measures of teacher effectiveness in policy and practice. Educational Policy for Action Series. Retrieved from: http://files.eric.ed.gov/fulltext/ED522163.pdf

Ehlert, M., Koedel, C., Parsons, E., & Podgursky, M. J. (2014). The sensitivity of value-added estimates to specification adjustments: Evidence from school- and teacher-level models in Missouri. Statistics and Public Policy. 1(1), 19–27.

Gabriel, R., & Lester, J. (2012). Constructions of value-added measurement and teacher effectiveness in the Los Angeles Times: A discourse analysis of the talk of surrounding measures of teacher effectiveness. Paper presented at the Annual Conference of the American Educational Research Association (AERA), Vancouver, Canada.

Gabriel, R. & Lester, J. N. (2013). Sentinels guarding the grail: Value-added measurement and the quest for education reform. Education Policy Analysis Archives, 21(9), 1–30. Retrieved from http://epaa.asu.edu/ojs/article/view/1165

Goldhaber, D., & Hansen, M. (2013). Is it just a bad class? Assessing the long-term stability of estimated teacher performance. Economica, 80, 589–612.

Goldhaber, D., Walch, J., & Gabele, B. (2012). Does the model matter? Exploring the relationships between different student achievement-based teacher assessments. Statistics and Public Policy, 1(1), 28–39.

Guarino, C. M., Maxfield, M., Reckase, M. D., Thompson, P., & Wooldridge, J.M. (2012, March 1). An evaluation of Empirical Bayes’ estimation of value-added teacher performance measures. East Lansing, MI: Education Policy Center at Michigan State University. Retrieved from http://www.aefpweb.org/sites/default/files/webform/empirical_bayes_20120301_AEFP.pdf

Harris, D. N. (2011). Value-added measures in education: What every educator needs to know. Cambridge, MA: Harvard Education Press.

Hermann, M., Walsh, E., Isenberg, E., & Resch, A. (2013). Shrinkage of value-added estimates and characteristics of students with hard-to-predict achievement levels. Princeton, NJ: Mathematica Policy Research. Retrieved form http://www.mathematica-mpr.com/publications/PDFs/education/value-added_shrinkage_wp.pdf

Holland, P. W. (1986). Statistics and causal inference. Journal of the American Statistical Association, 81(396), 945–960.

Jackson, K. C. (2012). Non-cognitive ability, test scores, and teacher quality: Evidence from 9th grade teachers in North Carolina. Cambridge, MA: National Bureau of Economic Research (NBER), Working Paper No. 18624. Retrieved from http://www.nber.org/papers/w18624

Kane, T., McCaffrey, D., Miller, T. & Staiger, D. (2013). Have we identified effective teachers? Validating measures of effective teaching using random assignment. Bill and Melinda Gates Foundation. Retrieved from http://www.metproject.org/downloads/MET_Validating_Using_Random_Assignment_Research_Paper.pdf

Loeb, S., & Candelaria, C. (2013). How stable are value-added estimates across
years, subjects and student groups? Carnegie Knowledge Network. Retrieved from http://carnegieknowledgenetwork.org/briefs/value‐added/value‐added‐stability

McCaffrey, D. F., Sass, T. R., Lockwood, J. R., & Mihaly, K. (2009). The intertemporal variability of teacher effect estimates. Education Finance and Policy, 4, 572–606.

Mihaly, K., McCaffrey, D., Staiger, D. O., & Lockwood, J.R. (2013). A
composite estimator of effective teaching. Seattle, WA: Bill and Melinda Gates Foundation. Retrieved from: http://www.metproject.org/downloads/MET_Composite_Estimator_of_Effective_Teaching_Research_Paper.pdf

Newton, X. A., Darling-Hammond, L., Haertel, E., & Thomas, E. (2010). Value added modeling of teacher effectiveness: An exploration of stability across models and contexts. Educational Policy Analysis Archives, 18(23). Retrieved from: epaa.asu.edu/ojs/article/view/810.

Papay, J. P. (2010). Different tests, different answers: The stability of teacher value-added estimates across outcome measures. American Educational Research Journal, 48(1), 163–193.

Paufler, N. A., & Amrein-Beardsley, A. (2014). The random assignment of students into elementary classrooms: Implications for value-added analyses and interpretations. American Educational Research Journal.

Reardon, S. F., & Raudenbush, S. W. (2009). Assumptions of value-added models for estimating school effects. Education Finance and Policy, 4(4), 492–519. doi:10.1162/edfp.2009.4.4.492

Rosenbaum, P., & Rubin, D. (1983). The central role of the propensity score in observational studies for causal effects. Biometrika, 17, 41–55.

Rothstein, J. (2009). Student sorting and bias in value-added estimation: Selection on observables and unobservables. Education Finance and Policy, (4)4, 537–571. doi:http://dx.doi.org/10.1162/edfp.2009.4.4.537

Rothstein, J. (2010, February). Teacher quality in educational production: Tracking, decay, and student achievement. Quarterly Journal of Economics. 175–214. doi:10.1162/qjec.2010.125.1.175

Rubin, D. B. (1978). Bayesian inference for causal effects: The role of randomization. The Annals of Statistics, 6, 34–58

Rubin, D. B., Stuart, E. A., & Zanutto, E. L. (2004). A potential outcomes view of value-added assessment in education. Journal of Educational and Behavioral Statistics, 29(1), 103–116.

Scherrer, J. (2011). Measuring teaching using value-added modeling: The imperfect panacea. NASSP Bulletin, 95(2), 122–140. doi:10.1177/0192636511410052

Wainer, H. (2004). Introduction to a special issue of the Journal of Educational and Behavioral Statistics on value-added assessment. Journal of Educational and Behavioral Statistics, 29(1), 1–3. doi:10.3102/10769986029001001

Cite This Article as: Teachers College Record, Date Published: August 01, 2014
http://www.tcrecord.org ID Number: 17633, Date Accessed: 8/10/2014 8:23:06 AM

Fred Smith worked for many years at the New York City Board of Education as a testing analyst. For all the parent groups who are upset by the over-testing of their children and concerned about the quality of the tests, Smith has become the go-to guy, who can be counted on to give a tough review of what the testing corporations are doing and what they should be doing.

 

In this post, Smith takes the New York State Education Department to task for withholding the technical report on the 2013 state tests. Just this week, responding to public outrage about its lack of transparency, the Department released 50% of the questions on the April 2014 tests. Until 2011, the SED released the entire exam with questions and answers. But no more. Since Pearson became the state’s testing agency, the state has been parsimonious in releasing questions and also technical data needed to understand the validity of the tests and the items.

 

The technical report for the 2013 tests should have been released in December 2013, but was not made public until July 2014. This is ridiculous. The information was available in Albany but was kept under wraps.

 

Smith says it is time for transparency and truth in testing. The public cannot trust the tests without seeing it and without allowing experienced experts like Smith to review its technical quality.