FairTest released the following fact sheet about the Common
Core tests:
Common Core Assessment Myths and Realities:
Moratorium Needed From More Tests, Costs, Stress
Behind (NCLB), each state set its own learning standards and
developed tests to measure them. But NCLB’s failure to spur overall
test score gains or close racial gaps led “reformers” to push for
national, or “common,” standards. With millions in federal Race to
the Top money and NCLB “waivers” as incentives, all but a few
states agreed to adopt Common Core standards. Two multi-state
consortia — the Smarter Balanced Assessment Consortium (SBAC) and
the Partnership for Assessment of Readiness for College and Careers
(PARCC) — won federal grants to develop Common Core tests, which
are due to be rolled out in 2014-15.
Here are the realities behind major Common Core myths.
Myth: Common Core tests will be much
better than current exams, with many items measuring higher-order
skills. Reality: New tests will largely
consist of the same old, multiple-choice
questions.
assessments that they said would measure – and help teachers
promote – critical thinking. In fact, the exams will
remain predominantly
multiple choice. Heavy reliance on such items continues
to promote rote teaching and learning. Assessments will generally
include just one session of short performance tasks per subject.
Some short-answer and “essay” questions will appear, just as on
many current state tests. Common Core math items are often simple
computation tasks buried in complex and sometimes confusing “word
problems” (PARCC, 2012; SBAC, 2012). The
prominent Gordon
Commission of measurement and education experts
concluded Common Core tests are currently “far from what is
ultimately needed for either accountability or classroom
instructional improvement purposes” (Gordon Commission, 2013).
Myth:
Adoptionof Common Core exams will end NCLB testing
overkill.
Reality: Under
Common Core, there will be many more tests and the same
misuses. NCLB triggered a testing tsunami
(Guisbond, et al., 2012); the Common Core will flood classrooms
with even more tests. Both consortia keep mandatory annual
English/language arts (ELA) and math testing in grades 3-8 and once
in high school, as with NCLB. However, the tests will be longer
than current state exams. PARCC will test
reading and math in three high school grades instead of
one; SBAC moves
reading and math tests from 10th grade to 11th. In PARCC states,
high schoolers will also take a speaking and listening test. PARCC
also offers “formative” tests for kindergarten through second
grade. Both consortia produce and encourage additional interim
testing two to three times a year (PARCC, 2012; SBAC, 2012). As
with NCLB, Common Core tests will be used improperly to make
high-stakes decisions, including high
school graduation (Gewertz, 2012), teacher
evaluation, and school accountability.
Myth: New multi-state assessments will
save taxpayers money. Reality: Test costs
will increase for most states. Schools will spend even more for
computer infrastructure upgrades.
have been a big concern, especially for the five states that
dropped out of a testing consortium as of August 2013. PARCC
acknowledges that half its member states will spend more than they
do for current tests. Georgia pulled out when PARCC announced costs
of new, computer-delivered summative math and ELA tests alone
totaled $2.5 million more than its existing state assessment
budget. States
lack resources to upgrade equipment, bandwidth and
provide technical support, a cost likely to exceed that of the
tests themselves (Herbert, 2012). One analysis indicates that Race
to the Top would provide districts with less than ten cents on the
dollar to defray these expenses plus mandated
teacher evaluations (Mitchell, 2012).
Myth: New assessment consortia will
replace error-prone test manufacturers.
Reality: The same, incompetent, profit-driven
companies will make new exams and prep
materials.
Pearson, Educational Testing Service and CTB/McGraw-Hill, are
producing the tests. These firms have long histories of mistakes
and incompetence. The multi-national Pearson,
for example, has been responsible for poor-quality items, scoring
errors, computer system crashes and missed deadlines (Strauss,
2013). Despite these failures, Pearson
shared $23 million in contracts to design the first
18,000 PARCC test items (Gewertz, 2012).
Myth: More rigor means more, or
better, learning. Reality: Harder tests
do not make kids smarter.
teachers witnessed students
brought to tears (Hernandez & Baker, 2013),
faced with confusing instructions and unfamiliar material on Common
Core tests. New York tests gave fifth graders questions written at
an 8th grade level (Ravitch, 2013). New York and Kentucky showed
dramatic drops in proficiency and wider achievement gaps. Poor
results hammer students’ self-confidence and disengage them from
learning. They also bolster misperceptions about public school
failure, place urban schools in the cross hairs and lend ammunition
to privatization schemes. If a child struggles to clear the high
bar at five feet, she will not become a “world class” jumper
because someone raised the bar to six feet and yelled “jump
higher,” or if her “poor” performance is used to punish her coach.
Myth:
CommonCore assessments are designed to meet the needs of all
students.
Reality: The
new tests put students with disabilities and English language
learners at risk. Advocates
for English
language learners (Maxwell, 2013) have raised
concerns about a lack of appropriate accommodations. A U.S.
Education Department’s technical review assessed the consortia’s
efforts in July 2013 and issued a stern warning, saying that
attempts to accommodate students with disabilities and
ELLs need
more attention(Gewertz, 2013).
Myth:
CommonCore “proficiency” is an objective measure of college- and
career-readiness.
Reality: Proficiency
levels on Common Core tests are subjective, like all performance
levels. Recent disclosures demonstrate
that New York State set
passing scores arbitrarily (Burris, 2013). There is
no evidence that these standards or tests are linked to the skills
and knowledge students need for their wide range of college and
career choices (Ravitch, 2013). In addition, school officials have
often yielded to the temptation to cheat and manipulate test
results to bolster the credibility of their favored
reforms. Examples include
Atlanta, New York, Washington, DC, Indiana, Florida, and more
(FairTest, 2012).
Myth:
Stateshave to implement the Common Core assessments; they have no other
choice.
Reality: Yes
they do. Activists should call for an indefinite moratorium on
Common Core tests to allow time for implementation of truly better
assessments. High-quality assessment
improves teaching and learning and provides useful information
about schools. Examples of better assessments include
well-designed formative
assessments (FairTest, 2006), performance
assessments that are part of the curriculum (New
York Performance Standards Consortium), and portfolios
or Learning
Records (FairTest, 2007) of actual student
work. Schools
can be evaluated using multiple sources of evidence
that includes limited, low-stakes testing, school quality reviews,
and samples of ongoing student work (Neill, 2010). It’s time to
step back and reconsider what kinds of assessments will help our
students and teachers succeed in school and life.
References
- Burris, C. 2013. “How come officials could predict test
score results?” Blog post, Answer
Sheet.http://www.washingtonpost.com/blogs/answer-sheet/wp/2013/08/12/how-come-… - FairTest. 2006. “The Value of Formative
Assessment.” http://www.fairtest.org/value-formative-assessment - FairTest. 2007. “The Learning Record.”
http://www.fairtest.org/learning-record\ - FairTest. 2012. Confirmed Cases of Test Cheating
(2008-2012).http://fairtest.org/sites/default/files/CheatingReportsList.pdf - Gewertz, C. 2012. “Questions Dog Design of
Tests,” Education Week.http://www.edweek.org/ew/articles/2012/07/31/37act.h31.html - Gewertz, C. 2012. “Will the Common Assessments Be Used as
a Graduation Requirement?” Education
Week.http://blogs.edweek.org/edweek/curriculum/2012/09/common_core_tests_to_r… - Gewertz, C. 2013. “Ed. Dept. Panel Says Test Consortia
Need Sharper Focus on Accessibility,” Education
Week.http://blogs.edweek.org/edweek/curriculum/2013/07/ed_dept_technical_revi… - Gordon Commission on the Future of Assessment in
Education. 2013. “A Public Policy Statement.”http://www.gordoncommission.org/rsc/pdfs/gordon_commission_public_policy… - Guisbond, L., Neill, M., and Schaeffer, R. 2012.
NCLB’s Lost Decade for Educational Progress: What
Can We Learn from this Policy Failure? Boston:
FairTest, http://www.fairtest.org/NCLB-lost-decade-report-home. - Herbert, M. 2012, July/August. “Common Core Testing
Online Without Constant Connectivity?” District
Administration. http://www.districtadministration.com/article/common-core-testing-online… - Hernandez, J. and Baker, A. “A Tough New Test Spurs
Protest and Tears,” New York Times.http://www.nytimes.com/2013/04/19/education/common-core-testing-spurs-ou… - Maxwell, L.A. 2013, August 5. “ELL Advocates Call for
PARCC Tests in Spanish,” Education
Week.http://blogs.edweek.org/edweek/learning-the-language/2013/08/ell_advocat… - Mitchell, Kenneth. 2012. “Federal Mandates on Local
Education: Costs and Consequences – Yes, it’s a Race, but is it in
the Right Direction?” CRREO Discussion Brief #8. http://www.newpaltz.edu/crreo/brief_8_education.pdf - Neill, M. 2010. “A Better Way to Assess Students and
Evaluate Schools,” Education Week.http://www.fairtest.org/better-way-assess-students-and-evaluate-schools - Partnership for Assessment of Readiness for College and
Careers. Item and Task Prototypes. 2012.http://www.parcconline.org/samples/item-task-prototypes#6. - Ravitch, D. 2013. “Punishing kids for adult
failures,” Daily News. http://www.nydailynews.com/opinion/punishing-kids-adult-failures-article… - Ravitch, D. 2013. “The Biggest Fallacy of the Common Core
Standards: No Evidence.”https://dianeravitch.net/2013/08/24/the-biggest-fallacy-of-the-common-cor… - Singer, A. 2013. “What Does a Common Core/Danielson
Lesson Plan Look Like?” Huffington
Post.http://www.huffingtonpost.com/alan-singer/what-does-a-common-coreda_b_38… - Smarter Balanced Assessment Consortium. A Summary of Core
Components. 2012.http://www.smarterbalanced.org/wordpress/wp-content/uploads/2011/12/Smar… - Strauss. V. 2013. “A brief history of Pearson’s problems
with testing,” Answer Sheet.http://www.washingtonpost.com/blogs/answer-sheet/wp/2013/04/24/a-brief-h… - Ujifusa, A. 2013. “Tests Linked to Common Core in
Critics’ Cross Hairs,” Education
Week.http://www.edweek.org/ew/articles/2013/08/02/37testside.h32.html
| Attachment | Size |
|---|---|
| CommonCoreTestsMyths&RealitiesFactSheet.pdf | 576.04 KB |

Reblogged this on Transparent Christina.
LikeLike
Outstanding scholarship! Thank you for posting this. I will share it with many, especially those who think they have been fully informed about the Common Core “Package”.
LikeLike
Unfortunately it appears that Fair Test believes that there is such a thing as a “fair” test outside those made by teachers for specific classes. That belief belies the reality that any standardized test has any validity whatsoever-they don’t. Wilson has identified 13 logical errors in the making of educational educational standards, the giving of standardized tests and the dissemination of the results that render the whole process completely invalid in his “Educational Standards and the Problem of Error” found at:
http://epaa.asu.edu/ojs/article/view/577/700
Brief outline of Wilson’s “Educational Standards and the Problem of Error” and some comments of mine. (updated 6/24/13 per Wilson email)
1. A quality cannot be quantified. Quantity is a sub-category of quality. It is illogical to judge/assess a whole category by only a part (sub-category) of the whole. The assessment is, by definition, lacking in the sense that “assessments are always of multidimensional qualities. To quantify them as one dimensional quantities (numbers or grades) is to perpetuate a fundamental logical error” (per Wilson). The teaching and learning process falls in the logical realm of aesthetics/qualities of human interactions. In attempting to quantify educational standards and standardized testing we are lacking much information about said interactions.
2. A major epistemological mistake is that we attach, with great importance, the “score” of the student, not only onto the student but also, by extension, the teacher, school and district. Any description of a testing event is only a description of an interaction, that of the student and the testing device at a given time and place. The only correct logical thing that we can attempt to do is to describe that interaction (how accurately or not is a whole other story). That description cannot, by logical thought, be “assigned/attached” to the student as it cannot be a description of the student but the interaction. And this error is probably one of the most egregious “errors” that occur with standardized testing (and even the “grading” of students by a teacher).
3. Wilson identifies four “frames of reference” each with distinct assumptions (epistemological basis) about the assessment process from which the “assessor” views the interactions of the teaching and learning process: the Judge (think college professor who “knows” the students capabilities and grades them accordingly), the General Frame-think standardized testing that claims to have a “scientific” basis, the Specific Frame-think of learning by objective like computer based learning, getting a correct answer before moving on to the next screen, and the Responsive Frame-think of an apprenticeship in a trade or a medical residency program where the learner interacts with the “teacher” with constant feedback. Each category has its own sources of error and more error in the process is caused when the assessor confuses and conflates the categories.
4. Wilson elucidates the notion of “error”: “Error is predicated on a notion of perfection; to allocate error is to imply what is without error; to know error it is necessary to determine what is true. And what is true is determined by what we define as true, theoretically by the assumptions of our epistemology, practically by the events and non-events, the discourses and silences, the world of surfaces and their interactions and interpretations; in short, the practices that permeate the field. . . Error is the uncertainty dimension of the statement; error is the band within which chaos reigns, in which anything can happen. Error comprises all of those eventful circumstances which make the assessment statement less than perfectly precise, the measure less than perfectly accurate, the rank order less than perfectly stable, the standard and its measurement less than absolute, and the communication of its truth less than impeccable.”
In other word all the logical errors involved in the process render any conclusions invalid.
5. The test makers/psychometricians, through all sorts of mathematical machinations attempt to “prove” that these tests (based on standards) are valid-errorless or supposedly at least with minimal error [they aren’t]. Wilson turns the concept of validity on its head and focuses on just how invalid the machinations and the test and results are. He is an advocate for the test taker not the test maker. In doing so he identifies thirteen sources of “error”, any one of which renders the test making/giving/disseminating of results invalid. As a basic logical premise is that once something is shown to be invalid it is just that, invalid, and no amount of “fudging” by the psychometricians/test makers can alleviate that invalidity.
6. Having shown the invalidity, and therefore the unreliability, of the whole process Wilson concludes, rightly so, that any result/information gleaned from the process is “vain and illusory”. In other words start with an invalidity, end with an invalidity (except by sheer chance every once in a while, like a blind and anosmic squirrel who finds the occasional acorn, a result may be “true”) or to put in more mundane terms crap in-crap out.
7. And so what does this all mean? I’ll let Wilson have the second to last word: “So what does a test measure in our world? It measures what the person with the power to pay for the test says it measures. And the person who sets the test will name the test what the person who pays for the test wants the test to be named.”
In other words it measures “’something’ and we can specify some of the ‘errors’ in that ‘something’ but still don’t know [precisely] what the ‘something’ is.” The whole process harms many students as the social rewards for some are not available to others who “don’t make the grade (sic)” Should American public education have the function of sorting and separating students so that some may receive greater benefits than others, especially considering that the sorting and separating devices, educational standards and standardized testing, are so flawed not only in concept but in execution?
My answer is NO!!!!!
One final note with Wilson channeling Foucault and his concept of subjectivization:
“So the mark [grade/test score] becomes part of the story about yourself and with sufficient repetitions becomes true: true because those who know, those in authority, say it is true; true because the society in which you live legitimates this authority; true because your cultural habitus makes it difficult for you to perceive, conceive and integrate those aspects of your experience that contradict the story; true because in acting out your story, which now includes the mark and its meaning, the social truth that created it is confirmed; true because if your mark is high you are consistently rewarded, so that your voice becomes a voice of authority in the power-knowledge discourses that reproduce the structure that helped to produce you; true because if your mark is low your voice becomes muted and confirms your lower position in the social hierarchy; true finally because that success or failure confirms that mark that implicitly predicted the now self evident consequences. And so the circle is complete.”
In other words students “internalize” what those “marks” (grades/test scores) mean, and since the vast majority of the students have not developed the mental skills to counteract what the “authorities” say, they accept as “natural and normal” that “story/description” of them. Although paradoxical in a sense, the “I’m an “A” student” is almost as harmful as “I’m an ‘F’ student” in hindering students becoming independent, critical and free thinkers. And having independent, critical and free thinkers is a threat to the current socio-economic structure of society.
LikeLike
I do think some standardized tests are valid and useful, but I am very dubious about the Smarter Balanced Assessments. What do they purport to be measuring? What, if anything, do they really measure? Will their influence lead to a new era of crappy, useless skills-drills teaching? I’d love to see a detailed analysis of these new Frankentests.
LikeLike