Archives for category: Teacher Evaluations

Audrey Amrein Beardsley has been studying William Sanders’ value-added assessment system for a decade or so, and she is no fan of his methodology.

Here she explains why.

William Sanders has argued that his methodology is not volatile, but Beardsley and other critics say otherwise.

TVAAS or EVAAS is highly controversial, yet Arne Duncan praised it and claimed that Tennessee made great strides because it uses Sanders’ methods.

Beardsley makes the following observations (she has many more links, and I can’t copy them all, so I urge you to read her article and follow the links to understand the evidence she cites):

 

Sanders and others (including state leaders and even U.S. Secretary of Education, Arne Duncan) have lauded Tennessee’s use of accountability instruments, like the TVAAS, for Tennessee’s large gains on the most recent National Assessment of Educational Progress (NAEP) outcomes. Others, however, have cautioned against this celebration because 1) the results mask the expanding achievement gap in Tennessee; 2) the state’s lowest socioeconomic students continue to perform poorly on the test; 3) Tennessee didn’t make gains significantly different than many other states; and 4) other states with similar accountability instruments and policies (e.g., Colorado, Louisiana) did not make similar gains, while states without such instruments and policies (e.g., Kentucky, Iowa, Washington) did make similar gains (I should add that Kentucky’s achievement gap is also narrowing and their lowest socioeconomic students have made significant gains). Claiming that NAEP scores increased because of TVAAS-use and other stringent accountability policies is completely unwarranted (see also this article in Education Week).

 

John Thompson, teacher and historian, here reviews the testimony in the Vergara trial of economists Raj Chetty and Tom Kane. They are believers in economic models for judging teacher quality. Thompson concludes they are seriously out of touch with the real world of teachers.

Thompson reviews their testimony and writes:

“Chetty, Kane, and other expert witnesses are assisting in an all-out assault on teachers’ most basic rights. I disagree with them, but I can see why they would believe that their research is relevant to 3rd through 8th graders in math and, to a lesser degree, elementary reading classes. But, even though they have not studied high schools, they are participating in an effort to also destroy the rights of high school teachers.

“And, nothing in their research could possibly support the opinion that once current laws are stricken that data-driven evaluations in non-tested subjects would likely benefit students in those classes. Up to 80% of students are in classes that remain virtually unstudied by value-added researchers. Yet, they are so confident in their opinions – based on their goal of addressing the bottom 5% of teachers – that they are helping a legal campaign (based almost completely on the opinions of some like-minded persons) to strike down duly enacted laws.

“Of course, I would also like to understand why a few corporate reformers are so convinced in the righteous of their opinions that they have initiated this assault on teachers. But, I’ve already gone too far down the path of trying to speculate on why they engage in such overreach. I just hope the Vergara judge has the inclination to look deeply into both the testimony of expert witnesses and how it is very different than the evidence and logic they have presented in written documents.”

Arne Duncan, Raj Chetty, Eric Hanushek, John King, Kevin Huffman, John White, and Michael Johnston, and the other evaluation hawks did not think about this teacher when they said full steam ahead on evaluating teachers by student scores:

Beth writes:

“As Diane points out, teachers are already, and always have been, evaluated. Here is the problem: I am a special education teacher in an alternative high school. I teach students with severe psychological and behavioral disorders. These students are not exempt, and must take all the same state tests (including Regents) that other students take. My evaluation is based upon what percentage of my students improve their test scores by an amount estimated at the beginning of the year. Because I teach what are called 8-1-1 classes (8 students, 1 teacher, 1 teacher’s aide), my evaluation will be based upon, at most, 16 students. However, at this point in the year, it looks like only 6 of the students I began the year with will still be in our school at the end. This isn’t unusual–students move, get put in residential facilities, drop out, or become chronically truant at a high rate in my school. Of the 6 students left, one has become pregnant and has gone off her meds. Most days she can’t even make it into class because of her emotional breakdowns. One student has been hospitalized for months, not receiving instruction from me. One comes to school about once a week. Two swear at me whenever I try to get them to do work, and tell me they don’t care, they’re dropping out as soon as they’re 16. Their parents tell me they can’t help, they can’t make their sons do anything, either. One is facing incarceration, and may not be here at the end of the year. If 4 of these 6 don’t reach their goals on the end of the year test, that will “prove” I’m a bad teacher.

“I wanted to work with students who really needed me, to help students who are struggling the most. But because I work with students who are impoverished, disabled, homeless, incarcerated, and mentally unstable, I may very well be labeled as “ineffective.” Does this really mean I’m a bad teacher?”

I asked Audrey Amrein-Beardsley to compile a list of the most important VAM research.

 

Here are her recommendations for the top 13 research articles about Value-Added Measurement:

 
Amrein-Beardsley, A. (2008). Methodological concerns about the Education Value-Added Assessment System (EVAAS). Educational Researcher, 37(2), 65-75. doi: 10.3102/0013189X08316420.

Amrein-Beardsley, A., & Collins, C. (2012). The SAS Education Value-Added Assessment System (SAS® EVAAS®) in the Houston Independent School District (HISD): Intended and Unintended Consequences. Education Policy Analysis Archives, 20(12), 1-36. Retrieved from http://epaa.asu.edu/ojs/article/view/1096

Berliner, D. C. (2014). Exogenous variables and value-added assessments: A fatal flaw. Teachers College Record, 116(1). Retrieved from http://www.tcrecord.org/content.asp?contentid=17293

Baker, E. L., Barton, P. E., Darling-Hammond, L., Haertel, E., Ladd, H. F., Linn, R. L., Ravitch, D., Rothstein, R., Shavelson, R. J., & Shepard, L. A. (2010). Problems with the use of student test scores to evaluate teachers. Washington, D.C.: Economic Policy Institute. Retrieved from http://www.epi.org/publications/entry/bp278

Darling-Hammond, L., Amrein-Beardsley, A., Haertel, E., & Rothstein, J. (2012). Evaluating teacher evaluation. Phi Delta Kappan, 93(6), 8-15. Retrieved from http://www.kappanmagazine.org/content/93/6/8.full.pdf+html

Haertel, E. H. (2013). Reliability and validity of inferences about teachers based on student test scores. Princeton, NJ: Education Testing Service. Retrieved from http://www.ets.org/Media/Research/pdf/PICANG14.pdf

Haut, M. & Elliott, S. W (Eds.). (2011). Incentives and test-based accountability in education. Committee on Incentives and Test-based Accountability in Public Education, National Research Council. Washington, DC: National Academies Press. Retrieved from http://nfpcar.org/Archive/Education_Evaluation_12521.pdf

Hill, H. C., Kapitula, L, & Umlan, K. (2011, June). A validity argument approach to evaluating teacher value-added scores. American Educational Research Journal, 48(3), 794-831. doi:10.3102/0002831210387916

Newton, X., Darling-Hammond, L., Haertel, E., & Thomas, E. (2010) Value-Added Modeling of Teacher Effectiveness: An exploration of stability across models and contexts. Educational Policy Analysis Archives, 18(23), 1-27. Retrieved from http://epaa.asu.edu/ojs/article/view/810

Papay, J. P. (2010). Different tests, different answers: The stability of teacher value-added estimates across outcome measures. American Educational Research Journal. doi: 10.3102/0002831210362589

Paufler, N. A. & Amrein-Beardsley, A. (2013, October). The random assignment of students into elementary classrooms: Implications for value-added analyses and interpretations. American Educational Research Journal. doi: 10.3102/0002831213508299

Rothstein, J. (2009). Student sorting and bias in value-added estimation: Selection on observables and unobservables. Education Finance and Policy, 4(4), 537-571. doi: http://dx.doi.org/10.1162/edfp.2009.4.4.537

Schochet, P. Z. & Chiang, H. S. (2010, July). Error rates in measuring teacher and school performance based on student test score gains. U.S. Department of Education. Retrieved from http://ies.ed.gov/ncee/pubs/20104004/

 

When asked for the top 25 (including the above), here is her list.

 

If your curiosity is not sated, here is an even longer list, thanks to Professor Amrein-Beardsley.

 

You too can be an expert.

 

 

This superintendent posted a request for help. I will be posting a summary of research on value-added-measurement later today. I think it is fair to say that while economists like VAM (they measure productivity), education researchers overwhelmingly oppose VAM because they know that most of the factors affecting test scores are beyond the control of the teacher.

 

 

I am a Superintendent in Texas and I’m looking for some insight into a connection I just became aware of. The state of Texas has begun the process of revamping principal and teacher evaluations. Recently (in the last few months) the Commissioner of Education reached a compromise with the USDE about NCLB requirements. Part of the compromise required Texas to include test scores in the teacher evaluation tool.

Now I see, taken from the SEDL website ( http://txcc.sedl.org/our_work/), that the states’ work on both the Principal and Teacher Evaluation systems are based on the priorities of the USDE. Unless I’m mistaken, the USDE priorities have been in place for several years. That would make the Commissioner’s “compromise” essentially a lie. He planned all along to implement a system like this. The best remedy to this kind of “in the dark” activity is sunlight.

Can anyone help explain these connections? I realize my explanation is short on details, best I believe the answers could be very enlightening when you consider the following points:
-Texas, especially our governor, has made a point of opposing EVERYTHING Washington
-Texas filed a waiver from NCLB and then pretended the result was the best it could do
– Educators are about to have an evaluation system imposed on them that will for all practical purposes, reestablish High Stakes Testing as a priority in this state by requiring student test scores be a SIGNIFICANT (emphasis TEA) portion of their evaluation

This stuff is not a coincidence, just look at the pattern of reform initiatives in other states. Its only just begun here in Texas.

My email is bendeancarson@gmail.com

THE BELOW INFORMATION IS FROM THE SEDL WEBSITE REFERENCED ABOVE

This project relates to the following USDE Priorities:

Identifying, recruiting, developing, and retaining highly effective teachers and leaders
Identifying and scaling up innovative approaches to teaching and learning that significantly improve student outcomes

 

I came across an article in the Washington Post by Michelle Rhee, in which she chastised parents who opted their children out of state tests. This article made me happy, because it shows that the Queen Bee of high-stakes testing is worried. She is worried that the opt out movement is gaining traction. She is worried that parents are sick of the Status Quo of the past dozen years. If parents opt out, there won’t be enough data to fire teachers, to give bonuses, and to close schools. The Status Quo might collapse. How will we know how students are doing if we don’t test them? How will we know if their teachers are any good without standardized tests? How will we know if their school should be closed?

I must say that I was brought to a sharp halt in my reading of this article when Rhee spoke of what happened when her daughter came home from public school, relieved that the last test was over. This puzzled me because Rhee lives in Sacramento, and her daughters live in Nashville. I wondered, was she visiting Nashville that day? Then I remembered that one of her daughters goes to a public school, and the other goes to an elite private school that does not give standardized tests. How does she know how the daughter in the private school is doing? How can she judge her teachers? How will the principals in that school know if the teachers are doing a good job if the kids don’t take standardized tests? It is very puzzling.

And I wondered about one other thing: Michelle Rhee is a fierce advocate for charters and vouchers because she believes in choice. Why doesn’t she believe that parents should be able to choose to say no to state testing? Many voucher schools are exempt from state testing but I haven’t heard her demand that legislators include them. How will they know how their children are doing?

I wasn’t going to write about Rhee, because she seems so yesterday, but then Peter Greene sent me this hilarious post, and I realized I had to write too. But he is so funny! he calls it: “The WaPo Wastes Space on That Woman.”

Kenneth Mitchell, a school superintendent in the Lower Hudson Valley of New York, has been concerned about the costs imposed on school districts by Race to the Top. He previously estimated that six districts in his region would spend $11 million to comply with the mandates of Race to the Top, which paid these districts $400,000.

In this comment, he describes a recent meeting with lawyers about possible lawsuits that will be brought because of New York’s flawed Educator Evaluation System.

 

On Friday, March 14, The Lower Hudson Council of School Superintendents hosted a panel of education attorneys to address the following topic:

LEGAL IMPLICATIONS OF NEW YORK STATE’S NEW TEACHER AND PRINCIPAL EVALUATION SYSTEMS

Supervision, Evaluation & Tenure Decisions

• What is the effect of Education Law §3012-c on a school district’s ability to terminate probationary teachers and principals?

• How might overly prescriptive, rigid statutory and regulatory policy frameworks, such as §3012-c, regarding teacher evaluation, tenure, and employment decisions withstand teacher and principal appeals?

Statistical Reliability and Validity of Data in Supervision, Evaluation & Tenure Decisions

• How might the statistical reliability and validity of measures of teaching effectiveness – state assessments, VAM, SLO’s, school-wide assessment scores – affect teacher evaluation, tenure, and employment decisions?

• How will the metric of ‘confidence intervals’ be considered in a legal decision about a teacher’s effectiveness?

• How will the number of years of value-added assessment data to determine teacher quality be a factor in a teacher or principal appeal?

• In what ways will the use of locally-developed assessments, such Student Learning Objectives (SLOs), be challenged in an appeal?

• How will the individual evaluation of a teacher based on school-wide data, such as the 4th grade math assessment, withstand an appeal?

Implementation, Professional Development, and Resources

• How will such factors as consistency, training, and quality be considered in observations and evaluations developed by supervisors?

• How will equity issues, such as the access to materials (e.g., Common Core units) or technology, be a factor in an appeal?

• Experts in child and adolescent development have asked for a review of the Common Core to ensure that all of the standards are developmentally appropriate.
Since assessments are being developed on the basis of Common Core and teachers and principals being assessed accordingly, how will the aforementioned concerns be considered?

Other References

“Evaluation Law Could Limit Ability to Terminate Probationary Teachers”; Warren Richmond III (Harris Beach), New York Law Journal, (May 2013)

“Legal Issues in the Use of Student Test Scores and VAM to Determine Educational Quality”; Diana Pullin, Education Policy Analysis (2010 Manuscript)

In addition to these references, we have posted other related legal articles on the main page of our website: http://www.lhcss.org. We have also raised other concerns about the model that we have shared with state legislators, members of the Board of Regents, officials at the State Education Department and with representatives of the governor’s office. There are many other questions that will need to be answered once this enters the legal arena.

We shared that many in our organization have concerns that a) the design of reform model is flawed on multiple levels; b) the expedited and unsupported implementation will further contribute to inevitable legal challenges; c)the weak technical basis and very limited or no research behind elements of the model will not withstand legal challenges. These are just a few of our concerns. As a result, school districts will be wasting even more time and money on legal costs. Unless significant changes are made on the basis of substantive evidence, New York’s reform model is headed for trouble that will move beyond the anxiety and frustration of over-tested students, angry parents, weary teachers, and harried administrators.

Peter Greene just keeps writing hit after hit. This
one explains
what VAM means and why it works well in
manufacturing but not in dealing with human beings.

He explains how
Pennsylvania measures teacher quality: PVAAS uses a
thousand points of data to project the test results for students.
This is a highly complex model that three well-paid consultants
could not clearly explain to seven college-educated adults, but
there were lots of bars and graphs, so you know it’s really good. I
searched for a comparison and first tried “sophisticated guess;”
the consultant quickly corrected me—“sophisticated prediction.” I
tried again—was it like a weather report, developed by comparing
thousands of instances of similar conditions to predict the
probability of what will happen next? Yes, I was told. That was
exactly right. This makes me feel much better about PVAAS, because
weather reports are the height of perfect prediction.

Here’s how it’s supposed to work.

The magic formula will
factor in everything from your socio-economics through the trends
over the past X years in your classroom, throw in your pre-testy
thing if you like, and will spit out a prediction of how
Johnny would have done on the test in some neutral universe where
nothing special happened to Johnny. Your job as a teacher is to get
your really Johnny to do better on The Test than Alternate Universe
Johnny would.
The only thing that goes wrong is that it
doesn’t work. Students are not inanimate objects like pieces of
steel. So he concludes: This is one more example of a
feature of reformy stuff that is so top-to-bottom stupid that it’s
hard to understand.

But whether you skim the surface, look at the
philosophical basis, or dive into the math, VAM does not hold up.
You may be among the people who feel like you don’t quite get it,
but let me reassure you– when I titled this “VAM for Dummies,” I
wasn’t talking about you. VAM is always and only for dummies; it’s
just that right now, the dummies are in charge.

See? All that’s required for VAM to work is believing
that the state can accurately predict exactly how well your
students would have done this year if you were an average teacher.
How could anything possibly go wrong??

As is well known, the U. S. Department of Education zealously believes–like Michelle Rhee–that low test scores are caused by “bad” teachers. The way to find these ineffective teachers, the theory goes, is to see whose students get higher scores and whose don’t. That’s known as value-added measurement (VAM), and the DOE used Race to the Top to persuade or bribe most states to use it to discover who should be terminated.

As we also know, things have not worked out too well, as some Teachers of the Year were fired; some got a bonus one year, then got fired the next year. In many states, teachers are rated by the scores of students they never taught. The overall effect of VAM has been demoralization, even among those with high scores because they know the ratings are arbitrary.

For some reason, teachers don’t like to “win ” at the expense of their colleagues and they can spot a phony deal a mile away.

But the U.S. DOE won’t give up, so they released a research brief attempting to show that VAM does work!

But Audrey Amrein Beardsley deconstructs the brief and shows that it is a mix of ho-hum, old-hat and wrong-headed assumptions.

It’s true (but not new) that disadvantaged students have less access to the best teachers (e.g., NBCT, advanced degrees, expertise in content areas (although as Beardsley says, the brief doesn’t suggest such things matter).

It is true, that “Students’ access to effective teaching varies across districts. There is indeed a lot of variation in terms of teacher quality across districts, thanks largely to local (and historical) educational policies (e.g., district and school zoning, charter and magnet schools, open enrollment, vouchers and other choice policies promoting public school privatization), all of which continue to perpetuate these problems.”

She writes:

“What is most relevant here, though, and in particular for readers of this blog, is that the authors of this brief used misinformed approaches when writing this brief and advancing their findings. That is, they used VAMs to examine the extent to which disadvantaged students receive “less effective teaching” by defining “less effective teaching” using only VAM estimates as the indicators of effectiveness, and as relatively compared to other teachers across the schools and districts in which they found that such grave disparities exist. All the while, not once did they mention how these disparities very likely biased the relative estimates on which they based their main findings.

Most importantly, they blindly agreed to a largely unchecked and largely false assumption that the teachers caused the relatively low growth in scores rather than the low growth being caused by the bias inherent in the VAMs being used to estimate the relative levels of “effective teaching” across teachers. This is the bias that across VAMs is still, it seems weekly, becoming more apparent and of increasing concern.”

VAM in the real world is Junque Science.

This letter arrived from:

Douglas McGuirk

English Teacher

Dumont High School Dumont, NJ

My Testimony about the AchieveNJ Act:

The AchieveNJ Act is certainly doing its part to make a convoluted mess out of the art of teaching our children.

In this testimony, I will address the most readily apparent of its many problems: data collection, Student Growth Objectives, Student Growth Percentiles, PARCC tests, and the new observation system. The AchieveNJ Act, and all of its affiliated changes, is simultaneously stretching the education profession in two different directions, most likely to the point of snapping it in half. I am no longer certain about what my job description is these days; am I a teacher, one who attempts to engage students and help them understand subject matter and their world, or am I a data collector, one who keeps statistics on all manner of measurables in a theoretical attempt to improve the process of teaching in which I am often not engaged because I am busy collecting the data?

AchieveNJ seems to operate on the fallacious principle that there is an infinite amount of time. During my day, this humble English teacher will collect data, analyze data, send students out for standardized tests, be observed by an administrator, and, somewhere in and among all of that, plan lessons, grade papers, and teach students. When do all of these things happen? How do they get done? How do I prioritize if each of these items is now considered crucial?

Most days only allow for one to two hours of time not spent in front of a class. Allow me to recount a personal story of how I spent two weeks in October of 2013. Every moment I worked, excluding those during which I was contractually obligated to actually teach students, was spent doing something related to my Student Growth Objectives (SGOs). I had previously administered a benchmark assessment or pre-test (no staff member in my school is sure about what terminology to use, so we have alternately used both, to the point that the students are not sure whether they are being benchmarked, or pre-tested, or, to put in plainly, harassed into doing something they do not wish to do), so I had a stack of essays that needed scoring. To start work on my SGO, I graded the essays according to the soon-to-be obsolete NJ Holistic Scoring rubric. Then I created and organized a spreadsheet to sort and organize my data. Then I entered all of the scores into the spreadsheet. Then I read through all the emails sent by district administrators about how to create my SGO. Following that, I formally wrote my SGO and submitted it to my supervisor.

The next day, the SGO was rejected, and my supervisor told me that all SGOs had been done incorrectly and that our staff would need training. We held a department meeting to review SGO policies. We then held an after school training session to discuss the writing of SGOs. I attended both of these. After two weeks of writing and rewriting my SGO, complete with all of the Core Curriculum Content Standards pasted from the web site, I finally had an acceptable SGO. I managed to accomplish absolutely no lesson planning during this period of time. I graded no papers. I am a veteran teacher with nine years in the profession. I understand how to manage my workload, overcome setbacks, and complete my responsibilities. In short, I am a professional who maintains a diligent work ethic.

But nothing could prepare me for the amount of time I had just spent on a new part of my job that basically exists so that I can continue to prove that I should be entitled to do the other parts of my job. After I completed my SGO, my principal told our staff to make sure we save all of the data, paperwork, and student work relating to our SGO, just in case people from the State want to review the integrity of the data. Seriously? This is the most egregious assumption that there is an infinite amount of time.

When will State reviewers go back and reread mountains upon mountains of SGO data to make sure that my essay scores (which suffer from an inherent subjectivity anyway) are accurate? The real goal of the SGO process seems to be to take teachers so far out of their comfort zones, and so far from working directly with students, that they may begin to question what kind of work they are doing anyway. Wouldn’t this time spent collecting mountains of dust-collecting data be better spent planning more interesting lessons? Offering students more feedback on work they understand and view as necessary? Researching content to make myself more knowledgable and helpful to my students? I guess not.

I have to teach my students the content needed to improve on the SGO so I can keep my job, which apparently consists of collecting even more SGO data. Just in case the SGO process is not intimidating and distracting enough, many of us (myself included) now have the threat of Student Growth Percentiles (SGPs) looming as well. The fact that these SGPs only apply to certain disciplines is inequitable and unfair to begin with, but that does not even address the fact that the correlation between my SGP score and my actual effectiveness is non-existent. Every article that I have read on this issue shows that the data produced by SGPs is statistically insignificant in its ability to determine my actual teaching effectiveness. I might as well determine a sizable portion of my evaluation by rolling dice or, to draw upon history, releasing doves and watching which way they fly. I have no control over how hard the students will work on these tests. I have no control over how thoroughly they will prepare.

From what I have read, these PARCC tests do not even have any actual effect on student grades or promotion. They are only used to evaluate me. In that case, allow me to hand-select the students who will be used to determine my effectiveness. Or better yet, the most fair thing to do would be to allow me to take the test myself, so at least I can have complete control over my own evaluation. Beyond just potentially affecting me in a random (and possibly absurd) way, the PARCC tests further reinforce the current contradictory nature of education rhetoric. What do policymakers want for our children? I consistently hear, from the mouths of our politicians, that our students are falling behind (falling behind whom?) in their critical thinking skills. (May we at least ask, how are these critical thinking skills measured? By bubble tests?) If that is the case, then shouldn’t we, as professionals, seek to introduce more critical thinking tasks, like project-based learning, into our curricula? Aren’t multiple choice standardized tests anathema to critical thinking tasks? Why is anyone promoting them, then? Where is the emphasis? Do we want students to legitimately be able to assess and evaluate on their own? Or do we want illogical measures to make sure that our teachers are, well, doing what exactly? If (some) teachers’ jobs are contingent on whether or not they achieve a high SGP score, then those teachers will, for the sake of their own self-preservation, certainly spend a great deal of time and energy trying to prepare students for those very tests, even though they cannot do the one thing that will ensure satisfactory scores, which is make the students put forth their best effort.

No students dislike learning. But many dislike education, because education consists of misguided and needlessly enervating tasks like standardized tests. Instead of spending this time engaged in critical thinking, students will be responding to questions that will be used to make sure their teachers are doing their jobs. Ironically enough, teachers will again be doing less of their jobs, as I assume we will be called upon increasingly to babysit computer labs full of children clicking vapidly through PARCC assessments. (As a side note, I am sure international test production companies like Pearson stand to profit from this arrangement immeasurably, probably at the expense of my own paycheck, most of which would have been spent in the local New Jersey economy.)

The final issue I will address in the AchieveNJ Act is the inconsistent new observation system. For starters, the public school districts across the state use two different evaluation systems: Danielson or McRel. If we are striving for consistency, why can we not agree on a single, unified observation system, so that all teachers are theoretically evaluated in the same fashion? Still, even if we achieved such uniformity, all observations would continue to suffer from the same inherent bias as the grades on students’ essays: each observer (or teacher, as is the case with the essays) has a different viewpoint (yes, even using a rubric). The administrators who serve as observers in my school have wildly varying interpretations about what constitutes an effective lesson. Even worse, some administrators are offering critiques to teachers about “how the lesson should have been conducted,” and providing less than satisfactory ratings to teachers who choose to do something in a different way.

The biggest source of all of this uncertainty and inconsistency has been the use of technology. Some of our administrators have said that we are to use technology in every single lesson, no exceptions. Others have been more lax about this requirement. I make this point to further illuminate the backwards nature of many these evaluative changes. If we must use technology, then technology is the starting point for each and every lesson. Previously, student learning was my starting point. What tools will help my students learn? Am I there to teach them or to show off the latest and greatest tech toys in my classroom? Are observers looking for critical thinking? Are they looking at my rapport with students? Or are they there to make sure that I go through the motions (according to one person’s rubric of what constitutes effective teaching) of reaching all of my supposed requirements? The inherent subjectivity of trying to quantify the unquantifiable is of course the same issue with which I wrestle when trying to score the essays that will make up my SGO. We all now must worship at the altar of data, even though, at best, the data is fickle and, at worst, it is fraudulent.

In the end I am not quite sure how to proceed under the AchieveNJ system. To paraphrase Plato, a single part of one’s soul cannot be engaged in two contradictory actions at the same time. So the only thing I can do is to default back to the ways in which I have always taught. I will try to help my students learn. I will try to reinforce material that I think is of value. I will provide as many insights from my own experiences as I can. I will focus on the human side of teaching and learning, my AchieveNJ ratings be damned. If this system says that an intelligent and dedicated individual like me is not fit to teach the students of New Jersey, then it is even more broken than my testimony could ever hope to convey.