Professional Documents
Culture Documents
Introduction
I wish to thank John Meyer, Francisco Ramirez, Evan Schofer, Michelle Jackson, Patricia Bromley,
and members of Stanford’s Comparative Education workshop for valuable comments on the article.
Received May 12, 2020; revised May 30, 2021, and December 18, 2020; accepted June 1, 2021; elec-
tronically published April 1, 2022
Comparative Education Review, volume 66, number 2, May 2022.
q 2022 Comparative and International Education Society. All rights reserved. Published by The University of
Chicago Press for the Comparative and International Education Society. https://doi.org/10.1086/718829
1
Rationalization, following Max Weber’s classic work on the topic, refers broadly to the ongoing
process of formally “systematiz[ing] social life around cultural schemes that explicitly differentiate and
then seek to link social means and social ends,” as well as “[formal] efforts to reconstruct all social
organization . . . as a means for the pursuit of collective purposes, these purposes themselves subject to
increasing simplification and systematization” ( Jepperson 2002, n. 6; see also Meyer et al. 1987).
H1: Countries with higher levels of education-related research output more likely to
adopt a national assessment.
H2: Countries with more linkages to INGOs are more likely to adopt a national
assessment.
H3: Countries with more projects funded by the World Bank are more likely to
adopt a national assessment test.
typically reserved for socializing a small group of elites. This reflects an in-
stitutional context of limited individual personhood. In the post-World War II
era, however, these older logics of schooling were replaced by an underlying
assumption of the empowered personhood of students: the fundamental
ontological equality of individuals, both in terms of their basic human rights
and their ability to be educated and contribute to a nation’s human capital. In
this context, schooling became an important way to cultivate students’ agentic
capacities to act autonomously, and other goals of education developed
around this assumption (Mundy 2010, 336).2 The ongoing construction of
these norms and conceptions of individual personhood culminated in a series
of conferences on “Education for All” in the 1990s and formal agreements by
nearly every country in the world to commit to the educational goals estab-
lished there (Chabbott 2003).3 The first conference on Education for All in
1990 included explicit commitments to develop national assessments in order
to formally measure a country’s equality of educational opportunity (see ar-
ticle 4 of the World Declaration on Education for All).
These institutional developments therefore suggest that the explosive
growth of national assessments over time is closely related to the cultural
processes that also drive longer-term trends in other national educational
institutions like high stakes exams and tracking. The rise of national high
stakes exams in the nineteenth and early twentieth centuries reflects the ra-
tionalization of educational stratification around formally “meritocratic” pro-
cesses of selection of individual students; between-school tracking was also
common in many education systems around the world during this period
(Furuta 2020). The use of high stakes exams and tracking at early stages of the
educational process reflects constricted conceptions of individual personhood.
During the postwar period, both high stakes exams and tracking have been
shown to decline at early stages of schooling in national education systems
around the world;4 this was a result of the aforementioned global expansion of
conceptions of equal personhood and the idea that all students are capable of
receiving an education (Furuta 2020, 2021). Thus, over time, students take
2
School textbooks around the world increasingly emphasized the empowered rights of equal in-
dividuals along several identities (e.g., children, women, minorities) and a variety of different types of
rights that these individuals possessed (e.g., educational, cultural, health rights); they also increasingly
incorporated a number of pedagogical approaches that facilitate learning in an empowered manner (e.g.,
open-ended questions, role-playing activities, and questions that encourage student to develop their own
opinions; Lerch, Bromley, et al. 2017).
3
During this period, enrollments expanded more rapidly in many countries around the world
(Schofer and Meyer 2005), former members of the Soviet Union were integrated into liberal world society
(Heyneman 2010), and international assessment tests like PISA and TIMSS were formally established by the
OECD and the IEA as part of the consolidation of a more global education regime (Ramirez et al. 2018).
4
In 1960, roughly 63 percent of countries in the world administered high stakes exams at the end
of the primary level of schooling; by 2010, only 43 percent of countries in the world did so at the level
(Furuta et al. 2021). Likewise, nearly 70 percent of countries in the world tracked students into different
school types at the junior secondary level in 1960; by 2010, however, fewer than 30 percent of countries
in the world did so (Furuta 2020).
fewer high stakes exams at early ages (which prevent a large number of students
from moving on to higher levels of schooling,) and students are less likely to be
tracked into different types of schools that drastically limit their future
educational opportunities (Furuta et al. 2021). The rise of “low stakes” national
assessments reflects a further step in the rationalization of educational insti-
tutions around these expanding conceptions of equal personhood: in order to
promote educational equality at early ages among formally equal individuals,
assessment tests shift the locus of responsibility for a student’s educational
performance from the individual to the school or state and measures educa-
tional equality in formally standardized ways (Smith 2014, 6).
The foregoing discussion suggests two additional sets of hypotheses. First,
national assessments are expected to have increased on a global basis after
1990 (after the first Education for All conference in Jomtien, Thailand, took
place) and at the primary school level. The explicitly global consensus on
these norms of Education for All, formed in 1990, suggests that this has been a
global, rather than a regional or idiosyncratically national, trend (Chabbott
2003). In the context of the global Education for All movement, universal
basic education is prioritized as a fundamental human right, and national
assessments reflect a way to fulfill this vision of empowered personhood; these
universalistic conceptions of education at the primary level can be contrasted
with later stages of the educational process, where educational opportunities
continue to be allocated in more unequal ways (e.g., Furuta 2020).
H5a: The use of national assessments increases on a global basis, especially after
1990.
Second, countries that continue to use high stakes exams and between-
school tracking to select and stratify students at early ages of the educational
process are less likely to adopt a national assessment. Tracking at early ages
and a higher number of high stakes exams reflect the retention of older and
more relentless logics of sorting and selecting children into unequal oppor-
tunities within a country’s educational institutions: national assessments there-
fore exist on one end of the spectrum of expanded conceptions of student
personhood, while educational tracking at the junior secondary level and a
greater number of high stakes exams in the schooling process exist on the op-
posite end (Furuta 2020, 2021).
H6a: Countries that have a higher number of high stakes examinations are less
likely to adopt a national assessment.
H6b: Countries that track students into different types of schools at lower levels
of the educational process are less likely to adopt a national assessment.
Alternative Explanations
5
In theory, high stakes exams measure student knowledge or academic skills in a standardized way
like assessment tests do; unlike assessment tests, however, they stratify and select individual students, and
they also restrict access to future educational opportunities.
Data
Dependent Variable
To identify the first year in which a country adopted a national assessment,
I relied on data from UNESCO’s EFA Monitoring Report (2015, 304–11),
which provides the most comprehensive cross-national and historical infor-
mation on the use of national assessments to date; these data were collected
from a variety of sources, including “printed material, websites, experts,
contacts through UNESCO regional offices” (UNESCO 2015, 304). The data
set includes the year, grade level, name, and subjects of every national as-
sessment conducted over the period identified. For each country listed in the
source, I identified the first year in which a country is listed as administering a
Independent Variables
A country’s linkages to the international community is measured as its total
number of memberships in international nongovernmental organizations
(INGOs) in a given year (logged, to account for the variable’s right-skewed
6
One change was made to the data from UNESCO (2015), based on information available in the
literature on testing: Singapore’s Primary School Leaving Exam, as well as the O-Level and A-Level
exams, were not considered “national assessments.” As high stakes exams, these tests have substantial
effects on stratification and individual educational opportunities in the schooling process (see Ramirez
et al. 2018). These exams are also part of the Cambridge Overseas Examination Syndicate, which are
used by many other countries around the world. Given that these exams were only reported as assess-
ment tests by Singapore, they were omitted in order to apply a consistent definition across the sample. All
other assessments in the data, to the author’s knowledge, were not high stakes exams.
7
The UIA’s Yearbook of Organizations identifies a wide range of organizations from many societal
domains. In the literature on world society theory, INGO memberships are intentionally measured in
order to include a diffuse array of organizations, because this reflects the more general principles of
universalism and rationalism that structure institutional models at the world level (Boli and Thomas
1999).
8
See Clarivate Analytics (2018), “Web of Science Databases.” https://clarivate.com/products/web
-of-science/databases/.
9
Additional analyses, not shown here, yield identical results when observations before 1966 were
excluded from the analyses.
dichotomous variable, where 1 indicates that students are tracked at this level,
and 0 indicates that students are not tracked at this level; given extensive cross-
national variation in the severity of tracking in the educational process, this
variable identifies “between school” tracking (e.g., academic schools versus
technical or vocational schools), rather than “within school” tracking by courses
(Furuta 2020). Both of these variables were coded manually using a wide range
of primary sources that describe the structure of educational institutions from
1960 to 2011 (e.g., UNESCO reports, international encyclopedias, interna-
tional conference reports; see Furuta 2020 for a more extended discussion).
Given that these measures reflect a similar underlying concept, they are mildly
correlated with each other (r p 0:28).
Finally, I include a variable, measured at the global level, that identifies
the cumulative number of countries that have conducted a national assess-
ment up to a given year. This measure captures the global institutionalization
of national assessments: neo-institutional theories of diffusion emphasize that
countries experience greater social pressures to adopt an innovation as other
countries increasingly adopt it (e.g., Hironaka 2014). As expected, this vari-
able is also highly correlated with linear time (r p 0:87) and captures the
dramatic growth of national assessments over the time period of this study.
Control Variables
A country’s GDP per 10,000 capita (logged) is a standard measure of
economic development and is coded using data on real GDP from the Penn
World Table (Feenstra et al. 2015).
A country’s primary enrollment ratio is coded as the gross proportion of
students at the primary school age-level that are enrolled in a given year;
this variable is coded using several data sources, including the World Devel-
opment Indicators (World Bank 2013) and other sources described more fully
in Meyer et al. (1992) and Benavot and Riddle (1988). Given the large num-
ber of overage students who are enrolled in mass schooling in many devel-
oping countries over the time period of this study, this variable is top-coded at
a value of 100 to avoid potential interpretive issues this may present in the
analyses.
A country’s level of democratization is coded using standard data from the
Polity IV project, which measures the degree to which a country’s political
system enables political participation and civil liberties, has institutionalized
checks and balances on the executive power, and allows citizens to choose
freely between alternative policies and leader (Marshall et al. 2014); a score
of 210 indicates a completely autocratic government, and a 10 indicates a
completely democratic government.
A country’s participation in international assessment tests (e.g., TIMSS,
PISA, PIRLS) is coded as the cumulative number of international assessment
tests it has participated in up to a given year (starting from 1960), using data
TABLE 1
DESCRIPTIVE STATISTICS OF DEPENDENT ANDINDEPENDENT VARIABLES (ALL YEARS POOLED)
Standard
Variable Observation Mean Deviation Minimum Maximum
Primary enrollment 4,202 84.92 22.85 2.80 100.00
GDP per 10,000 capita (log) 4,202 3.57 1.18 .43 6.72
Democracy score 4,202 .16 7.59 210.00 10.00
Number of high stakes exams 3,614 2.05 1.03 .00 3.00
Tracking, junior secondary level (d ) 4,202 .46 .50 .00 1.00
Cumulative number of international
assessments 4,202 1.07 2.32 .00 15.00
Cumulative countries ever conducted
national assessment 4,202 21.74 34.19 1.00 137.00
INGO memberships (log) 4,190 5.81 1.08 1.91 8.23
Educational publications per capita
(5-year intervals, log) 4,202 27.29 3.24 210.79 2.07
Cumulative World Bank educational
projects funded (log) 4,202 2.55 1.72 22.30 3.47
Method
To test the argument presented above, I use event history models that
identify the rate at which countries adopt its first national assessment. Event
history models are a standard statistical approach for measuring a country’s
risk of adopting a particular policy over time (e.g., Allison 2014). These
models measure the duration until an event takes place and its relationship
with other time-varying explanatory variables. In the analyses below, I follow
previous literature and present constant-rate models, which assume that the
baseline hazard rate is constant in the absence of other independent and
control variables included in the analyses (e.g., Frank et al. 2000; Longhofer
and Schofer 2010). The models presented below draw on data from 1960 to
2011, which are structured as country-year observations, and the estimates are
presented as untransformed coefficients (rather than hazard ratios). The first
event occurs in 1969 (when the US adopts the National Assessment of Edu-
cational Progress). All independent and control variables are lagged by 1 year
to ensure that these characteristics precede the event’s occurrence.
Results
countries adopt national assessment tests from 1960 to 1990, but the cumu-
lative hazard of the adoption of national assessments explodes especially after
1990 and the first World Education for All conference in Jomtien, Thailand.
When the adoption rates of national assessments are disaggregated across
different world regions (fig. 3), a clear finding emerges that assessments dif-
fuse at a near uniform rate across regions after 1990. As this figure shows, at
least half of all countries in every region have adopted a national assessment by
2011; furthermore, in most regions, at least 60–80 percent of countries have
done so by the end of the time period of the study. These descriptive findings
provide evidence that support hypothesis 5a of my argument that the rise of
national assessments is a worldwide phenomenon that grows from the con-
solidation of global norms of Education for All in the 1990s.
The argument of this article emphasizes the striking amount of global
isomorphism that shapes the rise of national assessments. Within this more
global view of these trends over time, however, some more specific inter-
regional patterns are worth noting. The pace of diffusion in South/Central
America, for example, accelerates earlier and more rapidly than the other
regions but then slows after this; as Kamens and McNeely (2010, n15) note,
the early acceleration in this region may be attributable to support from
nongovernmental organizations. Central/Eastern Europe also sees a burst of
national assessments in the early 2000s, as a result of the rapid dismantling of
Soviet institutions that were replaced with liberal standards and curricula
(e.g., Bethell and Mihael 2005). Finally, it is notable that Sub-Saharan Africa
has a slightly lower proportion of countries that adopt national assessments by
2010; as other studies have shown, this region of the world is by far the most
dependent on high stakes exams at all ages of schooling (Furuta 2021).
Figure 4 presents descriptive evidence of the specific grade levels in which
national assessments are administered; as noted above, each single grade level
for each national assessment in the data set is treated as a distinct test. These
findings are striking, and also support the argument advanced above (hy-
pothesis 5b): over 60 percent of the tests administered over the period from
1960 to 2011 are administered at the primary level (which typically comprises
the first 6–7 grades of the educational process); national assessment tests are
rarely administered at the senior secondary cycle (the last cycle of schooling
before higher education, typically around grades 9–12). Thus, in conjunction
with findings from prior research (Furuta et al. 2021), it is clear that low-stakes
national assessments that emphasize the importance of expanded access and
equality to education have emerged to replace high stakes exams that sort and
stratify students at the primary level.
Table 2 depicts a series of constant rate event history models that examine
the factors that predict the hazard rate of national assessment adoption. Given
FIG. 4.—All national assessments by school level, 1960–2011 (pooled). NOTE.—National assess-
ments were coded by name and grade-level for a given country and year; a national assessment that
was administered at several different grade-levels in a given year was coded as a distinct test for each
grade level within that year, for a total of 3,221 national assessments from 1960 to 2011.
that the independent variables for both hypotheses 1–4 and hypothesis 6
reflect similar underlying concepts, these variables are tested separately in
models 1–6. As expected by the argument presented above (hypotheses 2 and
3), a country’s memberships in international nongovernmental organizations
in a given year positively shapes its hazard rate of adopting a national assess-
ment (p < :05 in all models), while a country’s cumulative number of edu-
cation projects funded by the World Bank also positively and significantly in-
fluences its adoption of a national assessment (p < :01 in all models). At the
same time, a country’s commitments to more scientized conceptions of the
educational process is also positively associated with its likelihood of adopting a
national assessment test (hypothesis 1), although this variable is only signifi-
cant at the 10 percent level in model 5. Furthermore, as neo-institutional
theories of diffusion would expect, it is not surprising that the global number
of countries that have conducted a national assessment is positively and sig-
nificantly related with a country’s hazard rate of national assessment adoption
(p < :001 in all models).
Surprisingly, the effect of a country’s cumulative number of interna-
tional assessments on the hazard rate of national assessment adoption is not
statistically significant (hypothesis 4). When alternative specifications of this
1 2 3 4 5 6 7 8
Control variable:
Primary enrollmenta .00 .00 .00 .01 .01 .01 .01 .01
(.01) (.01) (.01) (.01) (.01) (.01) (.01) (.01)
GDP per 10,000 capita (log) 2.25∗ 2.27∗ .18 2.17 2.14 2.221 .241 2.08
(.13) (.13) (.14) (.13) (.11) (.12) (.13) (.12)
Democracy score .03∗ .02 .04∗ .031 .04∗∗ .031 .04∗∗ .04∗
(.02) (.02) (.02) (.02) (.02) (.02) (.02) (.02)
Independent variable:
Cumulative number of countries ever conducted .02∗∗∗ .02∗∗∗ .02∗∗∗ .02∗∗∗ .02∗∗∗ .02∗∗∗ .02∗∗∗ .02∗∗∗
national assessment (.00) (.00) (.00) (.00) (.00) (.00) (.00) (.00)
Number of high stakes examinations 2.34∗∗ 2.31∗∗ 2.39∗∗∗ 2.34∗∗
(.11) (.12) (.12) (.11)
J. sec tracking (d) 2.67∗∗ 2.67∗∗ 2.64∗∗ 2.69∗∗
(.24) (.24) (.23) (.24)
Educational publications per capita (5-year intervals log) .09∗ .081
(.04) (.04)
INGO memberships (log) .42∗ .50∗∗
(.19) (.17)
Cumulative World Bank educational projects funded (log) .31∗∗∗ .25∗∗
(.09) (.09)
Cumulative number of international assessment tests .04 .04
(.04) (.04)
Constant 22.72∗∗ 25.72∗∗∗ 24.56∗∗∗ 23.87∗∗∗ 23.83∗∗∗ 26.93∗∗∗ 25.56∗∗∗ 24.74∗∗∗
(.92) (1.07) (.81) (.72) (.79) (.93) (.70) (.63)
N 3,216 3,216 3,216 3,216 3,683 3,683 3,683 3,683
Countries 131 131 131 131 137 137 137 137
Number of failures 94 94 94 94 105 105 105 105
l2 172.33 163.42 171.01 158.84 170.73 154.81 130.55 150.68
df 6 6 6 6 6 6 6 6
NOTE Robust standard errors in parentheses. All independent and control variables lagged 1 year.
a
Variable capped at 100% enrollment ratio.
1
p ! .10.
∗
p !.05.
∗∗
p ! .01.
∗∗∗
p ! .001.
FURUTA
variable are tested (e.g., a dichotomous variable that indicates whether an in-
ternational assessment has ever been conducted, or a variable that captures
the number of years since an international assessment was first adopted) this
coefficient’s significance does not change. This variable strongly predicts a
country’s use of high stakes exams at the primary level and tracking at the junior
secondary level (Furuta 2020), although the magnitude of its correlation with
these variables is relatively weak (r p 20:29 for number of high stakes exams,
r p 20:21 for tracking at the junior secondary level); it is also moderately
correlated with the variable that identifies the cumulative number of countries
that have adopted a national assessment (r p 0:41). These results may suggest
that international assessments do not provide countries with as much of an
institutional infrastructure to conduct their own national assessments on its own
(see Kijima and Leer 2016), or it may also suggest that countries that participate
in international assessments have less of an incentive to create their own nation-
specific assessment system (see Kamens and McNeely 2010, n. 7).
My emphasis on the importance of expanded conceptions of student per-
sonhood in the educational process, which is reflected in a country’s educational
institutions like high stakes exams and tracking, is also supported by these results
(hypotheses 6a and 6b): the number of high stakes exams in a country’s edu-
cation system is negatively and significantly associated with a country’s likeli-
hood of adopting a national assessment (p < :01 in all models), and tracking at
the junior secondary level is also negatively and significantly associated with a
nation’s likelihood of adopting an assessment (p < :01 in all models).
By and large, the control variables in the models yield few significant
results: the coefficients for primary enrollments and democratization are
mostly insignificant, and the coefficient of a country’s level of economic de-
velopment is only weakly significant in some cases. The lack of consistently
significant results for these control variables suggests that a country’s do-
mestic institutional capacity to conduct a national assessment does not
strongly shape its adoption of one, when the independent variables are in-
cluded in the models; instead, the social processes that shape the adoption of
national assessments are more global.
Discussion
1990 and in every region of the world (hypothesis 5a), and these assessment
tests are predominately administered at the primary level (hypothesis 5b). A
country’s level of education-related research output (hypothesis 1), social ties to
international nongovernmental organizations (hypothesis 2), and eco-
nomic ties to World Bank funding for education projects all significantly
predict the adoption of a national assessment (hypothesis 3), while countries
that implement more high stakes exams or track students at the junior sec-
ondary level are much less likely to adopt one (hypotheses 6a and 6b). A
country’s participation in international assessments, however, does not
strongly predict when a national assessment test is adopted (hypothesis 4).
These results are observed net of several variables that capture a country’s
internal institutional capacity to conduct a national assessment, which gen-
erally have minimal effects on whether a national assessment is adopted.
The cultural processes emphasized in my argument are crucial because
they create the institutional conditions that enable and legitimate the remark-
able growth of national assessments over the past few decades. What policy-
related themes of monitoring and evaluation, accountability, and standards-
based reform have in common is that they are all part of this more general
process of rationalization that sweeps through formal institutions around the
world (Meyer et al. 1997). International institutions like the World Bank create
relationships of coercion and economic dependence, as power-dependence
arguments emphasize, but in doing so they enact a particular institutional model.
An alternative set of institutional conditions, legitimated by different cultural
principles, would lead to a different set of policy prescriptions: an institutional
model that emphasizes scientized conceptions of achievement and learning
without expanded conceptions of individual personhood, for example, could
legitimate a more extensive use of high stakes exams instead of assessment tests
(see n. 5; Kellaghan 1996). These more general cultural principles of equal in-
dividual personhood also shape long-term global trends in other crucial ed-
ucational institutions like high stakes exams and tracking (e.g., Furuta et al.
2021); in drawing attention to this point, my argument has emphasized how
global trends in these educational institutions are linked through an under-
lying cultural process, rather than shaped by entirely separate processes.
The macro-level, cultural perspective advanced in this article also implies
a distinctive framework for understanding the role of national assessments in
shaping educational change. In a world where formal policies and institutions
are often loosely coupled from practice and implementation, social change is
often not directly produced by a single policy or institution that delivers im-
mediate results; from this perspective, it is not surprising that existing studies
have shown limited or conflicting evidence of the effectiveness of national
assessments on a variety of normative outcomes (e.g., Ramirez et al. 2018).
However, national assessments are only one dimension of a more general
attempt to structure schooling around more rationalized processes: they reflect
Appendix
TABLE A1
YEAR OF A COUNTRY’S ADOPTION OF A NATIONAL ASSESSMENT TEST
TABLE A1 (Continued )
Year Country Year Country
2000 Belize 2004 Qatar
2000 Ethiopia 2005 Barbados
2000 Jordan 2005 Cambodia
2000 Korea 2005 Montenegro
2000 Luxembourg 2005 Solomon Islands
2000 Mongolia 2006 Azerbaijan
2000 Mozambique 2006 Denmark
2000 Slovak Republic 2006 Egypt
2000 Slovenia 2006 Norway
2000 Trinidad and Tobago 2007 Bulgaria
2001 Burkina Faso 2007 China
2001 Comoros 2007 Croatia
2001 Kyrgyz Republic 2007 Cyprus
2001 Macedonia 2007 Fiji
2001 Mauritania 2007 Japan
2001 Vietnam 2007 Mali
2002 Albania 2007 Myanmar
2002 Algeria 2007 Russia
2002 India 2007 Vanuatu
2002 Poland 2008 Latvia
2002 Tajikistan 2008 Maldives
2002 Turkey 2008 The Gambia
2002 Yemen 2008 Uzbekistan
2003 Australia 2009 Bahrain
2003 Bhutan 2009 Lithuania
2003 Bosnia and Herzegovina 2009 Seychelles
2003 Germany 2009 Sudan
2003 Guyana 2010 Armenia
2003 Pakistan 2010 Kenya
2003 Sri Lanka 2010 Tanzania
2003 Thailand 2010 United Arab Emirates
2004 Bangladesh 2011 Benin
2004 Georgia 2011 Cameroon
2004 Kiribati 2011 Rwanda
2004 Philippines 2011 Tonga
References
Abdul-Hamid, Husein, Namrata Saraogi, and Sarah Mintz. 2017. “Lessons Learned
from World Bank Education Management Information System Operations.” World
Bank Group, Washington, DC. http://documents.worldbank.org/curated/en/607
441491551866327/pdf/114096-PUB-PUBLIC-PUBDATE-4-6-17.pdf.
Allison, Paul. 2014. Event History Analysis: Regression for Longitudinal Event Data.
2nd ed. Thousand Oaks, CA: Sage.
Benavot, Aaron, and Nihan Koseleci. 2015. “Seeking Quality in Education: The
Growth of National Learning Assessments, 1990–2013.” Background paper for
Education for All Global Monitoring Report, UNESCO, Paris.
Benavot, Aaron, and Phyllis Riddle. 1988. “The Expansion of Primary Education,
1870–1940: Trends and Issues.” Sociology of Education 61 (3): 191–210.
Benveniste, Luis. 2002. “The Political Structuration of Assessment: Negotiating State
Power and Legitimacy.” Comparative Education Review 46 (1): 89–118.
Bethell, George, and Roxana Mihael. 2005. “Assessment and Examinations in
Romania.” Assessment in Education 12 (1): 77–96.
Boli, John, and George Thomas, eds. 1999. Constructing World Culture: International
Nongovernmental Organizations since 1875. Stanford, CA: Stanford University Press.
Bromley, Patricia, John Meyer, and Francisco Ramirez. 2011. “Student-Centeredness
in Social Science Textbooks, 1970–2008: A Cross-National Study.” Social Forces
90 (2): 547–70.
Chabbott, Colette. 2003. Constructing Education for Development: International Orga-
nizations and Education for All. New York: Routledge.
Diaz Rios, Claudia Milena. 2020. “The Role of Policy Legacies in the Alternative Tra-
jectories of Test-Based Accountability.” Comparative Education Review 64 (4): 619–41.
Drori, Gili, John Meyer, and Hokyu Hwang, eds. 2006. Globalization and Orga-
nization: World Society and Organizational Change. New York: Oxford University
Press.
Drori, Gili, John Meyer, Francisco Ramirez, and Evan Schofer. 2003. Science in the
Modern World Polity. Stanford, CA: Stanford University Press.
Feenstra, Robert, Robert Inklaar, and Marcel Timmer. 2015. “The Next Gener-
ation of the Penn World Table.” American Economic Review 105 (10): 3150–82.
Frank, David John, Ann Hironaka, and Evan Schofer. 2000. “The Nation-State
and the Natural Environment over the Twentieth Century.” American Sociological
Review 65 (1): 96–116.
Frank, David John, and John Meyer. 2007. “University Expansion and the Knowl-
edge Society.” Theory and Society 36:287–311.
Furuta, Jared. 2017. “Rationalization and Student/School Personhood in U.S.
College Admissions: The Rise of Test-Optional Policies, 1987 to 2015.” Sociology of
Education 90 (3): 236–54.
Furuta, Jared. 2020. “Liberal Individualism and the Globalization of Education as
a Human Right: The Worldwide Decline of Early Tracking, 1960–2010.” Soci-
ology of Education 93 (1): 1–19.
Furuta, Jared. 2021. “Western Colonialism and World Society in National Edu-
cation Systems: Global Trends in the Use of High-Stakes Exams at Early Ages,
1960 to 2010.” Sociology of Education 94 (1): 84–101.
Furuta, Jared, Evan Schofer, and Shawn Wick. 2021. “The Effects of High Stakes
Educational Testing on Enrollments in an Era of Hyper-Expansion: Cross-
National Evidence, 1960–2010.” Social Forces 99 (4): 1631–57.
Greaney, Vincent, and Thomas Kellaghan. 2008. Assessing National Achievement Levels
in Education. Washington, DC: World Bank.
Heyneman, Stephen. 2010. “A Comment on the Changes in Higher Education in
the Former Soviet Union.” European Education 42 (1): 76–87.
Hironaka, Ann. 2014. Greening the Globe: World Society and Environmental Change.
New York: Cambridge University Press.
Jepperson, Ronald. 2002. “Political Modernities: Disentangling Two Underlying
Dimensions of Institutional Differentiation.” Sociological Theory 20 (1): 61–85.
Kamens, David, and Aaron Benavot. 2011. “National, Regional, and International
Learning Assessments: Trends among Developing Countries, 1960–2009.” Globali-
sation, Societies, and Education 9 (2): 285–300.
Kamens, David, and Connie McNeely. 2010. “Globalization and the Growth of In-
ternational Educational Testing and National Assessment.” Comparative Education
Review 54 (1): 5–25.
Ramirez, Francisco, Evan Schofer, and John Meyer. 2018. “International Tests,
National Assessments, and Educational Development (1970–2012).” Comparative
Education Review 62 (3): 344–64.
Schilt, Kristen, and Danya Lagos. 2017. “The Development of Transgender Studies
in Sociology.” Annual Review of Sociology 43:425–43.
Schofer, Evan. 1999. “Science Associations in the International Sphere, 1875–
1990: The Rationalization of Science and the Scientization of Society.” In Con-
structing World Culture: International Nongovernmental Organizations since 1875, ed.
John Boli and George Thomas. Stanford, CA: Stanford University Press.
Schofer, Evan, and John Meyer. 2005. “The Worldwide Expansion of Higher Edu-
cation in the Twentieth Century.” American Sociological Review 70 (6): 898–920.
Smith, William. 2014. “The Global Transformation toward Testing for Account-
ability.” Education Policy Analysis Archives 22 (116). https://doi.org/10.14507/epaa
.v22.1571.
Smith, William, ed. 2016. The Global Testing Culture: Shaping Education Policy, Perceptions,
and Practice. Oxford: Symposium.
UIA (Union of International Associations). 1950–2014. Yearbook of International Orga-
nizations. Munich: UIA.
UNESCO. 1994. “World Declaration on Education for All and Framework for
Action to Meet Basic Learning Needs.” Adopted by the World Conference on
Education for All Meeting Basic Learning Needs ( Jomtien, Thailand), UNESCO,
New York.
UNESCO. 2015. “Education for All 2000–2015: Achievements and Challenges.” EFA
Global Monitoring Report, UNESCO, Paris.
Verger, Antoni, Lluis Parcerisa, and Clara Fontdevila. 2018. “The Growth and
Spread of Large-Scale Assessments and Test-based Accountabilities: A Political
Sociology of Global Education Reforms.” Educational Review 71 (1): 5–30.
World Bank. 2013. “World Development Indicators (Data Set).” World Bank, Wash-
ington, DC.http://data.worldbank.org/data-catalog/world-development-indicators.
World Bank. 2017. “Projects and Operations.” World Bank, Washington, DC. http://
projects.worldbank.org/.
World Bank. 2018. “World Development Report: Learning to Realize Education’s
Promise.” World Bank, Washington, DC.