About two years ago, I was first introduced to the Program for International Student Assessment (PISA), when taking a class on education and globalization. PISA is a worldwide study by the Organization for Economic Co-operation and Development (OECD) in member and non-member nations of 15-year-old school pupils' scholastic performance on mathematics, science, and reading. First carried out in 2000, PISA assessments have been following a triennial cycle, testing students from a growing number of countries (in 2000, 43 countries participated; the number rose to 65 in 2012) and producing reports after a-year-and-a-half analysis period. Last December, educators, policy makers and academics around the world anxiously waited for OECD to release its report on PISA 2012 results. The ranking, once again, ignited heated debates and discussions in the international education community.
I have remained skeptical about PISA since the first
encounter. Personally, I have had detrimental experiences with standardized
tests. Although I believe that testing could be used as a method to help
teachers monitor student learning, I tend to take any standardized tests that go beyond the
purpose of understanding individual students’ development with a grain of salt. Test results analyzed and reported in a collective manner (like what OECD
is doing) take education systems and the discourses around them out of their
social, political and economic contexts and lead to inaccurate and dehumanized
understanding and policies talks around education reform. Over the years, I have
kept an eye on the PISA data and discussions revolving around it. What I have
found is that more and more educators and policy makers have spoken about
the troubling interpretation of international tests data.
In Daniel Koretz’s book, Measuring
Up: What Educational Testing Really Tells Us, he discussed the trickiness
of interpreting international test data. First off, the term “international
mean” used by many media outlets and policy makers to compare American students’
performance to a seemingly “objective” standard is a fallacy. “International
mean” varies from test to test depending on the participating nations.
Secondly, more specific comparison between countries could still be
inconsistent. Depending on the test content and ranking system, countries
perform differently on different test. It is dangerous to draw conclusions
based on an individual source. Koretz warned people not to “treat either one of
these assessments as the definitive answer.”
Koretz did however make the point that it is relative safe
to conclude that countries like Japan ranking above the U.S. with a large
margin consistently perform better. But many people take such conclusions too
far by claiming that East Asian countries have a better education system or
better math/science curriculum without shedding a critical light on those tests.
When the 2012 PISA report came out last December, Shanghai impressed the world
with its students ranking number one again (after its initial appearance on
PISA 2009 report) in all three areas PISA tests. PISA has turned Shanghai into
this magic land whose superb student achievement educators around the world
envy. When I was studying abroad in France last spring semester, PISA test was briefly
discussed in my Comparative Education class. The professor was absolutely
impressed by Shanghai’s scores and commented that “amazing things are happening
there.” Not only did I have to disagree with her comment, I was also alarmed by
such a misleading impression. Back then, I was not informed enough to explain
to my professor why her interpretation of the data was incomprehensive and misguided.
A year later, when the new report comes out, I was able to send my professors articles written by educators around the
world who are critically examining PISA results and speaking up against the misleading
information propelled by PISA report.
Tom Loveless enlightens the public of OECD-PISA’s negligenceof the migrant student population in Shanghai, which makes the Shanghai data a
skewed depiction of the reality. Due to China’s kukou system that ties people
to their geographic origins, many children of families that have moved to the
city from rural areas of China are excluded from Shanghai school system. Loveless
claims that those children were abandoned by OECD-PISA assessment. Their plight
is brushed under the rug while Shanghai enjoys international praise for its
education success. Benjamin Riley, a
Fulbright public policy fellow, also sheds doubt on the PISA data. Coming from a
more technical perspective, he questioned the methodology used by PISA to
analyze its data. He pointed us to the critique of PISA by Svend Kreiner, one
of the world’s leading experts on the Rasch model, who argues that the OECD is
using the Rasch model incorrectly. I am incapable of psychometrician analysis,
but such critique should not be taken lightly.
I want to close this blog with a quote from Daniel Koretz’s
book: “Scores describe some of what students can do, but they don’t describe
all they can do, and they don’t explain
why they can or cannot do it.” With that said, I believe standardized tests
(both international and state-mandated ones) have enjoyed overly-exaggerated validity
and attention over the past few decades. It is time to refocus and reframe our
discussion on education reform.
No comments:
Post a Comment