Archives for the year of: 2014

Arne Duncan, Raj Chetty, Eric Hanushek, John King, Kevin Huffman, John White, and Michael Johnston, and the other evaluation hawks did not think about this teacher when they said full steam ahead on evaluating teachers by student scores:

Beth writes:

“As Diane points out, teachers are already, and always have been, evaluated. Here is the problem: I am a special education teacher in an alternative high school. I teach students with severe psychological and behavioral disorders. These students are not exempt, and must take all the same state tests (including Regents) that other students take. My evaluation is based upon what percentage of my students improve their test scores by an amount estimated at the beginning of the year. Because I teach what are called 8-1-1 classes (8 students, 1 teacher, 1 teacher’s aide), my evaluation will be based upon, at most, 16 students. However, at this point in the year, it looks like only 6 of the students I began the year with will still be in our school at the end. This isn’t unusual–students move, get put in residential facilities, drop out, or become chronically truant at a high rate in my school. Of the 6 students left, one has become pregnant and has gone off her meds. Most days she can’t even make it into class because of her emotional breakdowns. One student has been hospitalized for months, not receiving instruction from me. One comes to school about once a week. Two swear at me whenever I try to get them to do work, and tell me they don’t care, they’re dropping out as soon as they’re 16. Their parents tell me they can’t help, they can’t make their sons do anything, either. One is facing incarceration, and may not be here at the end of the year. If 4 of these 6 don’t reach their goals on the end of the year test, that will “prove” I’m a bad teacher.

“I wanted to work with students who really needed me, to help students who are struggling the most. But because I work with students who are impoverished, disabled, homeless, incarcerated, and mentally unstable, I may very well be labeled as “ineffective.” Does this really mean I’m a bad teacher?”

You know Common Core is in deep trouble when Michelle Rhee’s StudentsFirst creates a group to rally round the cause of high expectations. Somehow this new organization pretends to be antagonists to the union but the teachers’ unions have been generally supportive of Common Core. The criticism of the state’s rushed rollout has been nearly universal. Exactly what the demonstrators are supporting is unclear, unless the point is to defend the startlingly high failure rates generated by the state tests. Only 3% of English learners passed. Only 5% of students with disabilities passed. Less than 20% of black and Hispanic students passed. Maybe what StudentsFirst would like best is a test that no one passed. Now, that’s high expectations!

Gerri K. Songer maintains that the Common Core standards misunderstands how students learn to read. In a previous post, she demonstrated that the reading levels of PARCC were set so high and were so unrealistic that they would cause a very high failure rate.

New Research on Text Complexity – CCSS vs. Sound Educational Practice

By: Gerri K. Songer, Education Chair – Illinois Township High School District 214

Appendix A of the Common Core State Standards (CCSS) offers a review of research asserting that it is important for students to read complex text in order to be successful in meeting college and career challenges. CCSS argues, “The research shows that while the complexity of reading demands for college, career, and citizenship have held steady or risen over the past half century, the complexity of texts students are exposed to has steadily decreased in that same interval. In order to address this gap, the CCSS emphasize increasing the complexity of texts students read as a key element in improving reading comprehension.”

The study in Appendix A evaluates six different computer programs:

ATOS by Renaissance Learning
Degrees of Reading Power (DRP) by Questar Assessment, Inc.
Flesch-Kincaid
The Lexile Framework for Reading by MetaMetrics
Reading Maturity by Pearson Education
SourceRater by Educational Testing Service
Easability Indicator by Coh-Metrix
The different qualitative dimensions include:

1. STRUCTURE – Texts of low complexity tend to have simple, well-marked, and conventional structures, whereas texts of high complexity tend to have complex, implicit, and (in literary texts) unconventional structures.

Translation: CCSS finds it more desirable for students to read text that does not follow standard convention rules (i.e. text without an identifiable pattern).

2. LANGUAGE CONVENTIONALITY AND CLARITY – Text that rely on literal, clear, contemporary, and conversational language tend to be easier to read than texts that rely on figurative, ironic, ambiguous, purposefully misleading, archaic, or otherwise unfamiliar language (such as general academic and domain-specific vocabulary).

Translation: CCSS finds it more desirable for students to read text that is unclear, misleading, old, unfamiliar, ironic, and figurative (text that doesn’t say what it means).

3. KNOWLEDGE DEMANDS – Texts that make few assumptions about the extent of readers’ life experiences and the depth of their cultural/literary and content/discipline knowledge are generally less complex than are texts that make many assumptions in one or more of those areas.

Translation: CCSS finds it more desirable for students to read text with which few people can identify in terms of life experience.

4. LEVELS OF MEANING (literary texts) OR PURPOSE (informational texts) – Literary texts with a single level of meaning tend to be easier to read than literary texts with multiple levels of meaning (such as satires, in which the author’s literal message is intentionally at odds with his or her underlying message). Similarly, informational texts with an explicitly stated purpose are generally easier to comprehend than informational texts with an implicit, hidden, or obscure purpose.

Translation: CCSS finds it more desirable for students to read text that has multiple meanings with information that is implied, hidden, or obscure.

SUMMARY: CCSS advocates utilizing text for educational purposes that follows no pattern, that is unclear and misleading, that few people can identify with, and that has multiple meanings.

REALLY? This must go against every best practice strategy in existence! Perhaps this explains why politicians seem to be clueless.
In the quotation above, therefore, CCSS must be demonstrating the following skills:

“The research shows that while the complexity of reading demands for college, career, and citizenship have held steady or risen over the past half century . . . “

Purposely Mislead the Reader (PMR): It has most likely stayed the same over the past half century.

” . . . the complexity of texts students are exposed to has steadily decreased in that same interval.”

Obscure Information Using Multiple Meanings (OIMM): What type of complexity was actually analyzed in the research? From what study was this information taken?

Conclusions Based on Lack of Experience (CBLE): It is obvious CCSS has very little to no experience in helping students become better at reading because they would never advocate for text that follows no pattern, that is unclear and misleading, that few people can identify with, and that has multiple meanings as a means for improving comprehension.

“In order to address this gap . . . “

Figurative Language Lacking Patternization (FLLP): This must be figurative language because all educators know this is not actually happening. The inferred meaning is, “in order to steal public funding and confound future generations of America.” There is no pattern to support a gap actually exists, let alone that CCSS can bridge it. (If truth be told, I made up the term ‘patternization’ as it applies to text complexity. If CCSS advocates misleading, confusing, calling something something that it isn’t, and implying misinformation, I thought I’d give it a shot.)

“. . . increasing the complexity of texts students read as a key element in improving reading comprehension.”

Satiristic Assertion (SA): I have to call this one satiristic assertion because it’s really funny, yet hysterically tragic. Reading comprehension will not be increased by increasing the complexity of text in the manner proposed by CCSS. (I made up Satiristic Assertion too; although, I don’t think CCSS gives bonus points for creativity.)

Reading comprehension is a skill; it is just like learning to throw a football, make a basket, or hit a baseball. Athletes become good at their sport because they consistently practice individual skills.

Recall the first time you learned to ride a bike, throw a ball, or swim. Imagine how, in the beginning, someone demonstrated these activities. Yet, you could only learn so much from watching someone else; eventually, you had to give it a try yourself. With repetition and time, you became proficient in these activities, and perhaps, you may have ultimately excelled at and enjoyed them.

Apply these activities to reading. In the beginning, reading is modeled and taught. But, students can only learn so much from observing others read. At some point, students need to read for themselves. If they practice, they will become proficient, and in the end, they can excel and even enjoy this skill.

In order for students to read, they must practice reading consistently (a minimum of several 20 minute intervals each day) using text that has a vocabulary they understand and a level of complexity they are cognitively able to manipulate. I advocate that students read material they connect with and enjoy, so they are motivated to read rather than turned off by it. As time goes by and students demonstrate mastery of comprehension skills (finding the main idea; identifying supporting details; recognizing sequential, comparative, and cause-effect relationships; understanding the meaning of words; and making generalizations and conclusions) the complexity of text increases, as does vocabulary. I’m not against increasing text complexity, I am against increasing it in a manner aligned with the study produced by CCSS.

As a side note, sequence is a difficult skill for students with learning disabilities to master since most neurologically-based academic deficits include a processing deficit in the area of short-term (working) memory. Sequence is a skill that requires short-term memory; therefore, students with an academic deficit in the area of reading benefit from scaffolded instruction when practicing this skill.

As an English teacher, I would define text complexity in terms of the conventions used to produce the text at sound level and word level(decoding), and at the paragraph level, multi-paragraph level, single text level, and multi-text level (comprehension). These steps apply to all types of literature, including fiction, non-fiction, and informational text.

1. A simple sound would consist of one letter: | f |.
2. A complex sound increases in complexity based on the number of letters blended together: | ph | is more complex than | f |.
3. A simple word would be one syllable: cat.
4. A complex word increases in complexity as it increases in syllables: feline is a more complex word than cat.
5. A simple sentence is one that contains a subject and a verb, and it expresses a complete thought.
6. A compound sentence contains two independent clauses joined by a coordinator.
7. A complex sentence has an independent clause joined by one or more dependent clauses. A complex sentence always has a subordinator such asbecause, since, after, although, or when (and many others) or a relative pronoun such as that, who, or which.
8. A simple paragraph would consist of applying comprehension strategies exclusively to one paragraph.
9. Multi-paragraph complexity would increase as the number of paragraphs increase.
10. Simple text would consist of applying comprehension strategies to a text promoting a single point of view.
11. Multi-text complexity would increase as the number of differing viewpoints, either within a single text or within multiple texts, increase.

An error I find that is commonly made in reading instruction is based on the lack of understanding behind how fluency should be used. I have observed that mega-corporations producing reading materials often promote fluency when soliciting comprehension materials.

Fluency is a skill needed for students who are reading at the level of decoding. When these students begin to put together sounds and to form words, fluency is important so students can hear the sounds put together to form a word. At this level, they also need to read more fluently in order to process, or manipulate, the information they read. Sounds become words, words become sentences, and sentences become paragraphs.

Yet there is a grey area between decoding and comprehension where fluency is no longer the objective, comprehension is. No teacher would tell a student struggling with comprehension to, “read faster” (fluency). It is at this point where students actually benefit from slowing down and interacting with text using strategies such as annotation and materials such as graphic organizers related to individual comprehension skills as listed above.

Once students get to this level, it is purely a matter of consistent practice and raising the level of text complexity (as I identified it above) upon mastery of individual comprehension skills, while also increasing their vocabulary. This is similar to how a judo player would advance from one level to the next, for example. Of course, teachers will also have to deal with issues such as student motivation, attendance, the availability of appropriate materials, the number of students in a class, administrative decision-making, and etc. Those issues are beyond the scope of this article. It’s my intent to merely identify the components of solid reading instruction.

As you can clearly see, in contrast to CCSS, this follows a pattern. As demonstrated in the publication, ‘Learning About Numbers With Patterns,’ best practice maintains that children learn better when they can identify patterns. Although this study cites an example related to mathematics, its example can be applied to any discipline. Students learn better when they can see patterns, connect patterns, and build on patterns. This is a complete negation of the educational information CCSS is soliciting to the public.

In this fourth installment in his series of posts criticizing PISA, Yong Zhao examines the claim that low-income children in China outperformed the children of professional in the rest of the developed world.

He begins with the shock value of the headlines, which are guaranteed to stir nationalistic fervor:

“China’s poorest beat our best pupils”—The Telegraph (UK), 2-17-2014
“Children of Shanghai cleaners better at math than kids of Israeli lawyers”—Haaretz (Israel), 2-19-2014
“Cleaners’ children in China beat kids of US, UK professionals at maths: study”—NDTV (India), 2-18-2014
“Children of Chinese janitors outscore wealthy Canadians in global exams”—The Globe and Mail (Canada), 2-19- 2014

He writes:

“These are some of the most recent sensational headlines generated by PISA with a 4-page report entitled Do parents’ occupations have an impact on student performance released in February 2014. These headlines exemplify the secret of PISA’s great success as a masterful illusionist: effective misdirection of attention by exploiting human instinct for competition.

“From the start, the entire PISA enterprise has been designed to capitalize on the intense nationalistic concern for global competitiveness by inducing strong emotional responses from the unsuspecting public, gullible politicians, and sensation-seeking media. Virtually all PISA products, particularly its signature product—the league tables, are intended to show winners and losers, in not only educational policies and practices of the past, but more important, in capacity for global competition in the future. While this approach has made PISA an extremely successful global enterprise, it has misled the world down a path of self-destruction, resulting in irrational policies and practices that are more likely to squander precious resources and opportunities than enhancing capacity for future prosperity.”

I won’t summarize his arguments but I will share his conclusion:

“The bottom line: Until OECD-PISA became the only employer in the world with PISA scores as the only qualification, I would not suggest lawyers and doctors in the U.S., U.K., or any nation to replace your children’s activities in music, arts, sports, dancing, debates, and field trips with math tutoring. For the same reason, it is not time yet for schools in developed countries to close your swimming pools, burn your musical instruments, end museums visits, or fire your art teachers.”

Karen Klein, who writes editorials for the Los Angeles Times about education (and other topics), told her 16-year-old daughter she could opt out.

Like many other parents, Klein reached the breaking point where the tests didn’t make sense any more. After years of complying with the testing regime, she realized that this test was pointless. She even envied home-schoolers, who could take their children on field trips and explore what interested them. Imagine that!

Most touching was her story about the teacher who offered poetry teas. By the time her child was old enough to take the class, the poetry teas had disappeared. Test prep.

And then there was this event: “After one of the earlier versions gave a low score to my eldest on reading comprehension, my husband and I shrugged and knew there had to be something wrong with the test. That’s the daughter who is now finishing off her dissertation for a doctorate in literature.

The Los Angeles Times has been a reliable supporter of the new era of corporate reform, with occasional deviations (I recall an editorial scoffing at the parent trigger).

High-stakes testing is one of the Golden Calves of the Corporate Reform movement.

Karen Klein’s defection, rooted in her experience as a parent, not a think tank ideologue, suggests that there is hope for the future, that the patina of certitude attached to the standardized testing regime may in time crumble as more parents realize how flawed, how subjective, and how limited these tests really are.

She says, “Take that, world of Scantron.”

We say, “Right on. Welcome to the fight against the status quo. If it’s right for your child to opt out, it’s right for other people’s children.

The world is divided into two kinds of people: those who think that the world can be divided into two kinds of people and those who don’t. No, let me try that again: those who think that everything that matters can be measured, and those who think that what matters most cannot be measured. Count me in the latter group. What matters most is love, friendship, family, imagination, joy….I see no reason to develop measures for those things. And if they were developed, I would doubt their value or accuracy.

But here comes another attempt to measure creativity. This comment was posted by Laura Chapman in response to a discussion of the PISA problem-solving test:

“The PISA examples are math and logic problems. They are not tests of creativity. Look up the tests and informed theoretical work of Joseph W. Getzels and E. Paul Torrance.

“The Torrance tests, available from Scholastic http://ststesting.com/2005giftttct.html, are most often used to identify children, adults, and “special populations” as gifted. The pictorial and verbal tests measure three strengths in thinking: fluency, flexibility, and originality. In the figural tests, participants create simple drawings and respond to images. Scores are derived from evidence of qualities such as elaboration, expressiveness, storytelling, humor, and fantasy.

“Relatively few people are aware that the Council of Chief State School Officers (CCSSO) is a developing a web-based scale for measuring creativity, one of several in the “EdSteps” project—funded by the Bill and Melinda Gates Foundation and operated by the CCSSO

“EdSteps had a low profile until July, 2010 when Newsweek announced a “The Creativity Crisis,” citing a steady decline in scores on the Torrance Tests of Creativity since 1990. The tests are widely used and respected, in part, because records have been kept on the childhood scores and the later-in-life creative accomplishments of each cohort of test takers since the late 1950s (e.g., citations in art publications, patents and awards, books and articles published).

“In response to inquiries, the CCSSO issued a press release that dismissed the Torrance tests and referred its own work on creativity, emphasizing that EdSteps is a project to “advance creativity to the highest possible international standards, and measure creativity in a way that is situated in a context of actual activity.” Creativity is defined as “the valued uses and outcomes of originality driven by imagination, invention, and curiosity.”

“The Edsteps creativity scale a work-in process. The website solicits work samples on any subject from people of all ages and abilities, “globally”…”in any form, genre, or media”…” “writing, videos, images, charts, or other graphics.” People who visit the site are asked to compare two submissions and decide which is the most “effective” (undefined, but the favorite word of Bill Gates).

“That process is carried out in multiple iterations, by multiple judges, with multiple examples. This process is supposed to result in a scale representing a progression of achievement from novice to expert, without the need for written criteria or explanations.

“The process is not different from a popularity contest, with samples of work identified by age, gender, ability level, geographic region, type of work, and the like.

“I could not discover how the EdSteps addresses this fact: Works created by children can be judged more creative than work produced by well-trained adults (e.g., a quote attributed to Picasso: ”It took me four years to paint like Raphael, but a lifetime to paint like a child”). I cannot imagine how a single scale of creative achievement can be constructed from an “anything goes” basket of work from around the world, subject to further editing by EdSteps into web-friendly snippets. The release forms for the project are horrific.

“I think this effort is a crock, but I could be wrong. My sources: Bronson, P. & Merryman, A. (2010, July 10). The creativity crisis. Newsweek. Retrieved from http://www.newsweek.com/2010/07/10/the-creativity-crisis.html .///. Council of Chief State School Officers (2010). CCSSO response to ‘The creativity crisis.’ Press release Retrieved from http://www.ccsso.org/News_and_Events/Current_News/CCSSO_Response_to_the_Creativity_Crisis.html //// EdSteps. (2010b). February 22). Developing the EdSteps continuum: Report of the EdSteps technical advisory group. Retrieved from http://www.edsteps.org/CCSSO/DownloadPopUp.aspx?url=SampleWorks/EdStepsScalingApproach_Long.pdf // EdSteps. (2011b). Creativity launches. Retrieved from http://www.edsteps.org/CCSSO/ManageContent.aspx?system_name=nP6iGdNaft7MEwLG6uDXXA==&selected_system_name=DRkDdjiObdU=”

Here is a chance to send readers to two very important blogs.

Jonathan Pelto reprints Susan Ohanian’s post comparing the Common Core to a $5 billion fully automated warship.

Ohanian sees the writing on the wall. She fears that the full package, once deployed, will strip teachers of any autonomy or professionalism.

She notes the many professional organizations, including the unions, that have taken money to be aboard the train that has allegedly left the station. Susan, Jon, and many other outspoken bloggers and educators are proof that the train is sitting in the station.

Bottom line: when you get on a train, make sure it is headed where you want to go.

You may recall that when I went to AERA and shared a session with the wonderful, dynamic Helen Gym, she managed to pick her way carefully through the crowd that lined the wall of the room, while I managed to trip over someone’s foot and fell flat on my face. No harm done, the room was carpeted, and I landed gracefully in such a fashion that I was unhurt, indeed bounced up and proceeded to the podium.

 

Well, it turns out that the fall in Philadelphia was merely practice for what happened two days later. On Saturday morning, I packed the car and drove from Brooklyn to Long Island for what I expected would be a quiet weekend. My dear partner was away for the weekend. I dropped the dog at Doggie Daycare (she is a 60-pound critter and she loves to run with playmates), then proceeded to the abode by the sea. I took the cat inside, then went to the car, thinking I would go for the mail and supplies. But something happened, I don’t know what. I tripped, landed on my left knee and couldn’t get up. I felt a snap inside my leg. There was no carpet, there was stone. At first I thought the pain would go away if I just lay there for a few minutes, but when I tried to get up, I couldn’t stand. So I dragged myself on my back up the steps and into the house, reached up to the phone and called a neighbor. She called emergency services, and within 10 minutes, there was an ambulance, a police car, and assorted other vehicles in the driveway. Literally 15 people were there to help me, and I was grateful for their kind and efficient care. I was taken away by the volunteer fire department ambulance to the local hospital in Greenport, where the doctor did an x-ray and told me I had no broken bones. As soon as he heard what happened, my son took the bus from Brooklyn so that he could take care of me and bring me home. Today, I saw a knee specialist at the Hospital for Special Surgery in NYC, who told me I had torn my ACL, which seems to be a very valuable ligament in the knee. He told me that I did not need surgery but my basketball career was over (sorry, Arne).

 

The good news is that I am alive and well. I am fortunate to have friends and family who are kind and caring.

 

I will be in Louisville, Kentucky, next week to accept the Grawemeyer award.

 

What’s the moral of the story? Be careful. Slow down. Type faster. Walk slower. Watch your step. Try not to multi-task. Live in the moment.

 

I will try to remember the moral of the story.

 

 

Jill Osler, a parent in Nebraska, wrote the following statement:

 

Parents Can Take a Stand Against Testing

Almost one year ago, I wrote an editorial stating my belief that high-stakes standardized testing is not necessary for, and even detrimental to learning. Teachers feel pressure to spend large amounts of instructional time preparing students for Ne-SA tests—which robs students of authentic learning opportunities while killing natural curiosity and passion for learning. No educational research supports standardized testing, yet districts must administer Ne-SA to comply with No Child Left Behind.

What can I do? This past year, I have contacted the Nebraska Board of Education, our former Commissioner, senators on the Education Committee, Nebraska State Education Association, our Governor, Secretary of Education, President, two school lawyers, and met with my state senator. I learned I cannot wait for the government and politicians to do what is best for kids.

Therefore, we are choosing to exercise our parental right to opt our children out of taking Ne-SA tests. I can no longer speak strongly against these tests, and then support them as I tell our children to do their best on them.

Is it possible to opt out? It is stated on page 27 of the Nebraska Department of Education’s Update: Standards, Assessment, Accountability 2013-2014, that parents or guardians can make a formal written request to the district that their student be removed from testing. No reason is required.

How will opting out affect our school? NCLB states that if fewer than 95% of a district’s students take Ne-SA, the district will not make Adequate Yearly Progress. However, I am doubtful any district will make AYP, because this year 100% of students are required to be proficient in Ne-SA Reading and Math. This is another reason I feel the time is right to opt out.

How will opting out affect our children? Our principal assures us they will never feel punished. On the Ne-SA results, our children will show a score of zero—which will not impact grades, future schooling, or graduation.

What will our children do during Ne-SA testing? We identified activities they could do independently, so direct supervision is not required. They are prepared to respond if peers ask why they are not participating.

It is up to us, as parents, to stand up for what is best for children. It is nothing personal against our school or state; it is a much larger issue. We love our school and state, however, we love our children more.

I asked Audrey Amrein-Beardsley to compile a list of the most important VAM research.

 

Here are her recommendations for the top 13 research articles about Value-Added Measurement:

 
Amrein-Beardsley, A. (2008). Methodological concerns about the Education Value-Added Assessment System (EVAAS). Educational Researcher, 37(2), 65-75. doi: 10.3102/0013189X08316420.

Amrein-Beardsley, A., & Collins, C. (2012). The SAS Education Value-Added Assessment System (SAS® EVAAS®) in the Houston Independent School District (HISD): Intended and Unintended Consequences. Education Policy Analysis Archives, 20(12), 1-36. Retrieved from http://epaa.asu.edu/ojs/article/view/1096

Berliner, D. C. (2014). Exogenous variables and value-added assessments: A fatal flaw. Teachers College Record, 116(1). Retrieved from http://www.tcrecord.org/content.asp?contentid=17293

Baker, E. L., Barton, P. E., Darling-Hammond, L., Haertel, E., Ladd, H. F., Linn, R. L., Ravitch, D., Rothstein, R., Shavelson, R. J., & Shepard, L. A. (2010). Problems with the use of student test scores to evaluate teachers. Washington, D.C.: Economic Policy Institute. Retrieved from http://www.epi.org/publications/entry/bp278

Darling-Hammond, L., Amrein-Beardsley, A., Haertel, E., & Rothstein, J. (2012). Evaluating teacher evaluation. Phi Delta Kappan, 93(6), 8-15. Retrieved from http://www.kappanmagazine.org/content/93/6/8.full.pdf+html

Haertel, E. H. (2013). Reliability and validity of inferences about teachers based on student test scores. Princeton, NJ: Education Testing Service. Retrieved from http://www.ets.org/Media/Research/pdf/PICANG14.pdf

Haut, M. & Elliott, S. W (Eds.). (2011). Incentives and test-based accountability in education. Committee on Incentives and Test-based Accountability in Public Education, National Research Council. Washington, DC: National Academies Press. Retrieved from http://nfpcar.org/Archive/Education_Evaluation_12521.pdf

Hill, H. C., Kapitula, L, & Umlan, K. (2011, June). A validity argument approach to evaluating teacher value-added scores. American Educational Research Journal, 48(3), 794-831. doi:10.3102/0002831210387916

Newton, X., Darling-Hammond, L., Haertel, E., & Thomas, E. (2010) Value-Added Modeling of Teacher Effectiveness: An exploration of stability across models and contexts. Educational Policy Analysis Archives, 18(23), 1-27. Retrieved from http://epaa.asu.edu/ojs/article/view/810

Papay, J. P. (2010). Different tests, different answers: The stability of teacher value-added estimates across outcome measures. American Educational Research Journal. doi: 10.3102/0002831210362589

Paufler, N. A. & Amrein-Beardsley, A. (2013, October). The random assignment of students into elementary classrooms: Implications for value-added analyses and interpretations. American Educational Research Journal. doi: 10.3102/0002831213508299

Rothstein, J. (2009). Student sorting and bias in value-added estimation: Selection on observables and unobservables. Education Finance and Policy, 4(4), 537-571. doi: http://dx.doi.org/10.1162/edfp.2009.4.4.537

Schochet, P. Z. & Chiang, H. S. (2010, July). Error rates in measuring teacher and school performance based on student test score gains. U.S. Department of Education. Retrieved from http://ies.ed.gov/ncee/pubs/20104004/

 

When asked for the top 25 (including the above), here is her list.

 

If your curiosity is not sated, here is an even longer list, thanks to Professor Amrein-Beardsley.

 

You too can be an expert.