Professional Documents
Culture Documents
Math Common Core
Math Common Core
net/publication/281272769
CITATIONS READS
3 2,580
2 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Yang Gao on 15 August 2017.
Article
Educational Policy
2017, Vol. 31(3) 275–302
Mathematical © The Author(s) 2015
Reprints and permissions:
Communication in State sagepub.com/journalsPermissions.nav
DOI: 10.1177/0895904815595723
Standards Before the journals.sagepub.com/home/epx
Common Core
Abstract
Mathematical communication has been an important feature of standards
documents since National Council of Teachers of Mathematics’ (NCTM)
(1989) Curriculum and Evaluation Standards. Such an emphasis has influenced
content standards of states from then to present. This study examined how
effective the prevalence of various forms of mathematical communication in
2009 state standards documents was in regard to National Assessment of
Educational Progress (NAEP) 2009 achievement scores for Grade 4. Analysis
suggests mixed results with potential implications as states move toward
fully implementing the Common Core State Standards in Mathematics.
Specifically, although including language requiring mathematical descriptions
from students had a positive effect on Grade 4 NAEP 2009 achievement
scores, including language requiring rationales and justifications was not
found to have a statistically significant effect.
Keywords
mathematics education, standards, achievement, common core
Corresponding Author:
Karl Wesley Kosko, Kent State University, 401 White Hall, P.O. Box 5190, Kent, OH 44242,
USA.
Email: kkosko1@kent.edu
276 Educational Policy 31(3)
Introduction
Many education reformers continue to advocate engaging students in math-
ematical communication through discussion and writing. First gaining trac-
tion as a reform-oriented position with the publication of Curriculum and
Evaluation Standards for School Mathematics (National Council of
Teachers of Mathematics [NCTM], 1989), mathematical communication
has been observed to engage students in developing deeper understandings
of mathematics (e.g., Hufferd-Ackles, Fuson, & Sherin, 2004; Staples,
2007; Yackel & Cobb, 1996). Furthermore, student engagement in mathe-
matical communication has been found to statistically predict students’
mathematics achievement (Hiebert & Wearne, 1993; Koichu, Berman, &
Moore, 2007; Kosko, 2012; Mercer & Sams, 2006). While evidence exists
to support teacher practices that engage students in mathematical commu-
nication, no such examination has been done from a policy perspective.
Specifically, the importance of mathematical communication was first
argued successfully from a standards document (Lampert & Cobb, 2000;
McLeod, 2003; NCTM, 1989). Therefore, it is important to evaluate the
effectiveness of various standards documents in advocating student engage-
ment in mathematical communication. As individual states transition away
from their own standards to the new Common Core State Standards in
Mathematics (CCSSM), the importance of such an evaluation is more per-
tinent than ever. It is important to evaluate the differences in the “old”
standards and the “new” CCSSM ones to determine if we are moving for-
ward effectively. Furthermore, there is a need to examine what ways differ-
ent content standards may be improved by including expectations that
students describe and justify their mathematics. The purpose of the present
study is to provide a baseline of research for future efforts in the field to
review and recommend revisions for standards-based language. Specifically,
we chose to examine the general effectiveness of including such expecta-
tions as integrated in mathematics content standards. To accomplish these
goals, we examined the effectiveness of states’ standards expectations for
mathematical communication in regard to students’ National Assessment of
Education Progress (NAEP) 2009 mathematics scores. Comparing states by
way of students’ NAEP 2009 mathematics achievement provides a common
metric between states, as well as a metric that is similar in form (i.e., stan-
dardized test) to how most states tend to assess the effectiveness of teach-
ers’ instruction in accordance with standards. In a similar manner, this study
examines the effectiveness of the various 2009 standards documents (as
opposed to teachers’ instruction) regarding conveyance of expectations for
students to communicate mathematically.
Kosko and Gao 277
The communication standard was the second standard listed in each grade
band (NCTM, 1989). A useful example of expectations comes from the
Grades K-4 band in which NCTM (1989) advocated communication as a
means for children to make connections among multiple means of represent-
ing mathematical ideas (pictorial, graphical, symbolic, etc.). Discussion and
writing were encouraged as ways for students to develop better understand-
ings of mathematics. Furthermore, the Curriculum and Evaluation Standards
for School Mathematics encouraged teachers at all grade levels to engage
students in communication through the posing of questions, facilitating dis-
cussions among students, and engaging students in writing for mathematics
(NCTM, 1989).
The legacy of NCTM’s 1989 document can be seen in Principles and
Standards for School Mathematics (NCTM, 2000), a revision of their 1989
standards document. The Principles and Standards refined the ideas posed at
various grade bands in 1989 and provided a more comprehensive set of
expectations for what they referred to as the mathematical process standard
of communication. As a process standard, communication was viewed as dis-
tinct, but integrated with the mathematics content standards as a “[way] of
acquiring and using content knowledge” (NCTM, 2000, p. 29). Specifically,
278 Educational Policy 31(3)
NCTM (2000) suggested that from prekindergarten through Grade 12, stu-
dents should organize and consolidate their mathematical thinking through
communication; communicate their mathematical thinking coherently and
clearly to peers, teachers, and others; analyze and evaluate the mathematical
thinking and strategies of others; and use the language of mathematics to
express mathematical ideas precisely (NCTM, 2000). Among the specific
actions that support these usages of communication are students describing,
explaining, justifying, conjecturing, writing, questioning, arguing, listening,
and speaking about their mathematics. More recently, the Common Core State
Standards for Mathematics (CCSSM) have included expectations for various
aspects of mathematical communication within their grade-specific content
standards and their more general Standards for Mathematical Practice
(Common Core State Standards Initiative [CCSSI], 2010). These expectations
include the mathematical practice to construct viable arguments and critique
the reasoning of others. Like the NCTM communication process standard that
preceded it, this mathematical practice includes expectations for students to
provide descriptions and rationales in various formats, as well as interpreting
them via reading and listening to others’ mathematical arguments.
This brief history of mathematical communication in mathematics stan-
dards documents shows that mathematical communication has consistently
been considered an important expectation since the first serious standards
document was written. Yet, the general effect of including such expectations
in standards documents on students’ mathematics achievement has not been
examined. We argue that such an examination is necessary to determine how
effective language used to convey such expectations of mathematical
communication are in their most current form.
(H. C. Hill, 2001, p. 310). A separate study by H. C. Hill (2005) found differ-
ences between teachers’ and researchers’ assessment of teaching practices in
the classroom in regard to “mathematical communication and representa-
tion.” Examining secondary mathematics teachers interpretation of the
Common Core Standard for Mathematical Practice, construct viable argu-
ments and critique the reasoning of others, Kosko, Rougee, and Herbst
(2014) also found many teachers had interpretations of such practices that did
not match expectations of researchers.
Observations as those presented in the prior paragraph may help explain
findings from Kosko and Miyazaki (2012) who found that there was signifi-
cant variance in the effect of discussion frequency on mathematics achieve-
ment between teachers and schools. The variance was so large that in many
schools, engaging in discussion had a negative effect and in others, it was
generally positive. Yet, the cause of such variation may go beyond the teacher.
In examining various state standards for the usage of primary verbs in content
standards, Larnell and Smith (2011) noted that verbs such as “explain” and
“justify” (as well as others that would be classified here as related to com-
munication) were used relatively infrequently. The verb “explain,” which
was used more frequently than other communication-related verbs, was used
in varying ways. “Though explain is commonly considered a relatively high-
level verb in Bloom’s Taxonomy, this analysis suggests that states may be
using it to identify low-level cognitive processes” (pp. 111-112) as well as
high-level cognitive processes. Thus, Larnell and Smith’s (2011) findings
suggest that misinterpreting of standards, and mathematical communication
in standards, may generate at the state level, and not merely at the teacher
level. It also seems entirely plausible that teachers’ misinterpretations (or cor-
rect interpretations) may be influenced by the state policy documents they are
beholden.
Although it is not a purpose of this study to investigate teachers’ interpre-
tations of standards documents, they are discussed here because such factors
may influence the effectiveness of states’ standards. As useful as it would be
to incorporate teachers’ interpretations of standards in the present study, the
nature of data (as will be discussed later) does not allow for such a compari-
son at the scale of the present study.
Research Question
Research suggests that standards documents may affect how teachers incor-
porate mathematical communication in their classrooms (e.g., Loeber, 2008;
Young, 2007). In turn, students who are taught by teachers incorporating
mathematical communication appropriately tend to have higher mathematics
280 Educational Policy 31(3)
achievement (Hiebert & Wearne, 1993; Koichu et al., 2007; Kosko, 2012;
Mercer & Sams, 2006). Yet, some evidence exists that suggests teachers may
not interpret standards as intended, when it comes to mathematical commu-
nication (H. C. Hill, 2001, 2005; Kosko et al., 2014), possibly because some
states’ standards may not include consistent language in providing expecta-
tions in content standards for mathematical communication (Larnell & Smith,
2011). In essence, these various studies suggest engaging students in mathe-
matical discussions and writing in particular ways can have a statistically
significant and meaningful effect on their mathematics achievement, but it is
unclear whether or not state standards as policy documents are effective in
ensuring students engage in such actions.
The purpose of this study is to provide an initial examination of whether
and to what degree states’ incorporation of mathematical communication in
their mathematics standards documents affects students’ mathematics achieve-
ment. Our focus on mathematical communication is interpreted from NCTM’s
(2000) descriptions, in accordance with McLeod’s (2003) description of the
historical incorporation of mathematical communication in standards docu-
ments. While the term effectiveness can be interpreted in a number of ways, the
present study uses the term in regard to students’ mathematics achievement.
The reason for such an interpretation is simple: Those who drive policy (e.g.,
politicians and the public) typically interpret effectiveness of education reforms
in terms of test scores. Whether or not such usage is appropriate is a worthwhile
topic, but not one which we discuss here. Rather, we use this indicator of effec-
tiveness because of its present currency in policy decision making. Therefore,
the present study uses that metric to determine how “effective” policy docu-
ments in the form of state content standards are. To assess this study’s objec-
tive, a secondary analysis of NAEP 2009 data was conducted. NAEP 2009
provides an assessment common among all 50 states and at the last measured
point before adoption of CCSSM beginning in 2010. As will be described in the
forthcoming sections, state mathematics content standards effective in the
2008-2009 academic year were collected and examined for elements of math-
ematical communication. These data were then merged with NAEP 2009
Grades 4 restricted-use data to provide an indicator of the effectiveness of vari-
ous states’ standards emphasizing mathematical communication on students’
mathematics achievement.1 Merging of the data and the analysis that followed
allowed for the examination of the following research question:
Nature of Data
Two sources of data were used in the present study. The first were mathemat-
ics content standards documents from all 50 U.S. states. These data are pub-
licly available on request from the various governments of each of those
states. The second source of data was restricted-use data for Grade 4 mathe-
matics in the NAEP 2009 assessment (National Center for Education Statistics
[NCES], 2009b). The NAEP mathematics assessment uses a common set of
items to examine the status of mathematics learning across the United States.
Additional data related to the students, teachers, schools, and districts are
collected as well to provide for a general indication of the status of education
in the nation. The nature of NAEP’s common assessment across states, which
can have very different state-level content tests, was a necessary requirement
for the present study’s goals. Although the nature of NAEP 2009 provided for
a useful and necessary source of data for this study, use of such data came
with certain limitations. Specifically, neither teachers’ interpretations of stan-
dards nor their teaching practices were observed in NAEP 2009, and such
effects are not assessed in the present study. This limits the interpretability
and available range of focus of the present study to the effectiveness of
the standards documents alone. So, inferences regarding teaching practices
cannot be made.
Sample
The sample used for NAEP 2009 incorporated a multistage cluster sampling
design to obtain a sample where “estimates of population and subpopulation
characteristics could be obtained with reasonably high precision” (NCES,
2009a, p. 68). Therefore, certain subpopulations were oversampled (e.g.,
Native Americans) to insure that appropriate estimates for those subpopula-
tions could be obtained. To adjust for this oversampling in certain clusters,
NCES (2009a) recommends use of certain sample weights, depending on the
analysis performed. In accordance with these recommendations, the sample
weight identified in the data set with the designation ORIGWT was used in
the present study. Using this weight, the sample in this study can be inter-
preted as representative of students in those enrolled in Grade 4 in a public
school in one of the 50 U.S. states.2
While NAEP 2009 collected data from private schools, the goals of this
study were to examine how state mathematics content standards influenced
student achievement, and thus only public school data were used. In addition,
only students who resided in a U.S. state were included in the study. Other
reductions in sample were due to various independent variables included as
282 Educational Policy 31(3)
Measures
Dependent variable. Students’ mathematics achievement was used as the out-
come in this study. According to NCES (2009a), the NAEP 2009 mathemat-
ics test assessed five content strands: number properties and operations;
measurement; geometry; data analysis, statistics, and probability; and alge-
bra. The assessment used multiple-choice, short constructed-response and
extended constructed-response items in such a manner as to cover all content
strands. NAEP 2009 tests were assessed and scored using item response the-
ory (IRT). “IRT models the probability of answering a question in a certain
way as a mathematical function of proficiency or skill” (NCES, 2009a,
p. 26). NAEP created a composite score based on IRT estimations with a
range of 0 to 500. For the present analysis, the mean of variables MRPCM1
to MRPCM5, which represent composite score estimates as previously
described, were used to create the outcome measure (MathScore). MathScore
represents the mathematics composite score for students across all content
strands (M = 240.05, SD = 26.96, range = 120.60-332.77).
contexts, become embedded in a clause such that the meaning is not initially
at the fore. This process, referred to as nominalization, causes the transitive
process to serve as a participant in the clause. In other words, such processes
are acted on by other processes, thereby providing multiple layers of meaning
in a grammatical text. While others have looked at the difference between
nominalized and nonnominalized processes in content standards (i.e., Kosko
& Herbst, 2011), we looked only at the presence of processes eliciting expec-
tations for communication (i.e., description and rationale). This presence was
counted whether the process acted as a main transitive process in the standard
or if it was embedded in the clause as a participant. Throughout the descrip-
tions of our coding, we provide examples with italicized processes and nomi-
nalized processes to help illustrate their presence in standards language.
Description refers to standards that require students to describe mathemat-
ical procedures, strategies, steps, or concepts. According to the Principles
and Standards (NCTM, 2000), communication helps students to “gain
insights into their thinking when they present their methods for solving prob-
lems” (p. 60), and students should “learn new mathematical concepts”
through giving “verbal accounts and explanations” (p. 61). Maryland (2004)
requires students to
explain the need to use standard units for measuring. (Grade 4, Measurement,
4.4.1)
The transitive process justify solicits the action to engage students in mathe-
matical communication and arguments, which has the potential to help them
to better understand mathematical concepts or procedures. Therefore, exam-
ples such as the ones provided in this paragraph were coded as rationale
standards.
Grade 4 mathematics content standards were collected for analysis. To
align the standards with NAEP mathematics achievement scores, standards
documents that were in effect in the academic year of 2008-2009 were col-
lected for analysis (as this coincided with the year NAEP 2009 data were
collected). These documents were collected from a variety of sources
including the websites from the Department of Education in each U.S.
state, and directly contacting each of these offices for verification we had
identified the standards document effective in the 2008-2009 academic
year. After standards documents were collected, a portion of the data
(10%) was examined by both authors, as is typical in such analyses, to
refine the coding rubric (see Hruschka et al., 2004). The remaining 90% of
data were coded by both authors individually. Afterward, both authors
compared and reconciled results. Weighted Kappa was calculated to assess
interrater reliability or prereconciled data, as it allows for ratings closer in
magnitude to be given more weight in terms of agreement than ratings
further away in magnitude (Landis & Koch, 1977). Our weighted Kappa
was .62 for description and .60 for rationale. Landis and Koch (1977) sug-
gested that kappa coefficients .41 to .60 have moderate reliability and .61
to .80 have substantial reliability. This indicates that the coding of 2009
standards was sufficiently reliable for the present study. Descriptive statis-
tics suggest that description standards (M = .16; SD = .10; range = .00-.47)
may be more prevalent than rationale standards (M = .05; SD = .05; range
= .00-.24). However, there was substantial variance in the frequency of
standards that required either form of communication. While most states
had relatively low frequencies of standards requiring descriptions or ratio-
nales from students, there was a great deal of variation in the degree to
which different states incorporated such expectations of students, with
some states having no such expectations and others requiring descriptions
for nearly half of their math content standards and rationales for nearly a
quarter.
Kosko and Gao 285
Covariates. While the focus of the present study is on how language in 2009
state mathematics standards affected students’ mathematics achievement in
Grade 4, other factors have historically affected mathematics achievement.
Namely, gender, race/ethnicity, and socioeconomic status (SES) have each
been consistently associated with differences in mathematics achievement.
Tate (1997) reviewed literature and found that between 1973 and 1992, the
achievement gap attributable to race closed in Grades 4, 8, and 12. This was
also found to be true of SES. However, Lubienski (2002) examined NAEP
data from 1990, 1996, and 2000 and found that the achievement gap due to
race increased between 1990 and 2000 for Grade 8 Black students, but
remained the same at Grade 4. SES did not adequately explain the achieve-
ment gaps. A later study of NAEP data found that, while small, the achieve-
ment gap due to gender was consistently present for data between 1990 and
2003 (McGraw, Lubienski, & Strutchens, 2006). This finding is consistent
with prior research (Tate, 1997). Given the consistent effects associated with
gender, race, and SES, these factors were included at the student level of
analysis in the present study.
Gender was included by taking the variable SEX from the NAEP data set
and dummy coding it (dFemale) so that the effect would associate with being
female (0 = male; 1 = female). For race/ethnicity, the variable SRACE, which
indicated several designations of race for students, was separated into 5
dummy coded variables so that the designation White/Caucasian was the ref-
erence group: dBlack (0 = not Black; 1 = Black), dHispanic (0 = not Hispanic;
1 = Hispanic), dAsian (0 = not Asian; 1 = Asian), dNatAmer (0 = not Native
American; 1 = Native American), and dOther (0 = not Other non-White
ethnicity; 1 = Other non-White ethnicity). SES was assessed using the variable
SLUNCH1, which designated a student’s eligibility for the National School
Lunch Program (i.e., free and/or reduced lunch). As the provided variable
included multiple designations, it was dummy coded into dLunch (0 = not
eligible for free/reduced lunch; 1 = eligible for free/reduced lunch). Descriptive
statistics for these student-level covariates are displayed in Table 1.
Additional covariates were included in the analysis to account for
teacher factors influential to mathematics achievement. According to find-
ings by Clotfelter, Ladd, and Vigdor (2007), the more years of experience
286 Educational Policy 31(3)
Unconditional Model
The initial model examined in all HLM analyses is the unconditional model,
or the model without conditions (i.e., variables). Analysis of the uncondi-
tional model allows for the researchers to determine the variance in the out-
come measure and how much of this variance is due to the respective levels
in the analysis. It also allows for a generalized baseline interpretation of the
outcome measure across the sample, without adjusting for other factors. In
the Equation Set 1 below, the mathematics score (MathScore) of student
i taught by teacher j in state k is represented as the outcome at Level 1. The
variable π0 jk represents the mean achievement of students taught by teacher
j in state k, and eijk represents the random variance from this mean that is
associated with the individual student i. At Level 2, the mean score per
teacher, π0 jk, is the outcome variable with β00k representing the mean
achievement for state k and r0 jk is the random variance associated with the
individual teacher j. At Level 3, β00k , or the mean score per state, is the out-
come variable with γ 000 symbolizing the grand mean, or overall average, of
math achievement scores in all 50 states. The variable u00 k represents the
deviation of state k’s mean achievement from the grand mean.
Equation Set 1.
Level 1:
( MathScore )ijk = π0 jk + eijk .
Level 2:
π0 jk = β00 k + r0 jk .
Level 3:
β00 k = γ 000 + u00 k .
Results from the unconditional model suggest the typical student in the
United States, residing in 1 of 50 states and not accounting for any other fac-
tors, has a mathematics achievement score of 239.50, which is statistically
significant from zero at the .001 level.
288 Educational Policy 31(3)
Coefficient SE
Intercept, γ 000 251.33*** 0.77
dFemale, γ100 −1.12** 0.17
dLunch, γ 200 −12.69*** 0.21
dBlack, γ300 −16.86*** 0.29
dHispanic, γ 400 −12.14*** 0.32
dAsian, γ 500 1.64 0.49
dNatAmer, γ 600 −13.75*** 0.77
dOther, γ 700 −10.34** 0.74
are male, and who are either White or Asian (students reported as Asian were
found not to have statistically significant math scores from White students).
These results generally follow findings from previous research and merit
their inclusion in the present analysis.
Coefficient SE
Intercept, γ 000 246.46*** 0.93
SES_Mean, γ 010 14.33*** 0.69
MathScore_Mean, γ 020 0.96*** 0.01
dFemale, γ100 −1.04** 0.32
dLunch, γ 200 −10.57*** 0.66
dBlack, γ300 −9.46*** 0.58
dHispanic, γ 400 −6.44*** 0.52
dAsian, γ 500 1.16 1.04
dNatAmer, γ 600 −7.69*** 0.98
dOther, γ 700 −6.90** 2.53
this particular finding is outside of the scope of the present study, it is worth
noting that prior research has indicated use of eligibility for free or reduced
lunch may not be a reliable indicator of poverty, primarily because many
students are eligible for such lunch programs for 1 or 2 years and then rise out
of such indicators of poverty (M. S. Hill & Jenkins, 2001). However, a less
reliable indicator would likely result in nonstatistically significant findings,
suggesting that our small finding here needs further investigation. We
acknowledge the possible limitation of eligibility for free or reduced lunch as
a sole measure for SES, but include it due to a lack of additional information
for SES of the student population in the sample. As expected, MathScore_
Mean, β02k , was found to be a statistically significant and positive predictor
of student achievement ( γ 020 = 0.96, p < .001).
π1 jk = β10 k .
π2 jk = β20 k .
π3 jk = β30 k .
π4 jk = β40 k .
π5 jk = β50 k .
π6 jk = β60 k .
π7 jk = β70 k .
Level 3:
β00 k = γ 000 + γ 001 ⋅ ( Description )k + γ 002 ⋅ ( Rationale )k + u00 k .
β01k = γ 010 .
β02 k = γ 020 .
β10 k = γ100 .
β20 k = γ 200 .
β30 k = γ 300 .
β40 k = γ 400 .
292 Educational Policy 31(3)
β50 k = γ 500 .
β60 k = γ 600 .
β70 k = γ 700 .
Final model results. Final model results for Grade 4 are presented in Table 4.
As can be seen from Table 4, Level 1 coefficients saw little change from pre-
vious models, and Level 2 covariates were also found to be similar to find-
ings from Model 2. For Level 3, description was found to have a positive and
statistically significant effect ( γ 001 = 22.41, p < .01). However, rationale was
found not to be statistically significant ( γ 002 = −2.12, p = .90). These coef-
ficients should be interpreted in reference to the percentage of such standards
a state has included. For example, in a state with 16% of its standards including
requirements for description (the national average), the average student would
typically see an increase in his or her scores by 3.59 points. Such changes are not
meaningful in magnitude. Rather, to improve the average student’s mathematics
achievement score by just half a standard deviation, a state would need approxi-
mately 60% of its content standards to include expectations for mathematical
description. As indicated previously, no state had such high frequencies of math-
ematics standards requiring description. Therefore, these results suggest that
although the inclusion of expectations for mathematical description in content
standards does have a statistically significant effect on students’ mathematics
achievement, for most states this effect is not meaningful in size. Furthermore,
inclusion of standards requiring rationales from students was not found to have
a statistically significant effect on their mathematics achievement. Essentially,
various states’ incorporation of mathematical communication in their Grade 4
Kosko and Gao 293
Grade 4
Coefficient SE
Intercept, γ 000 243.21*** 1.81
Description, γ 001 22.41*** 7.81
Rationale, γ 002 −2.12 16.94
SES_Mean, γ 010 14.32*** 0.68
MathScore_Mean, γ 020 0.96*** 0.01
dFemale, γ100 −1.04** 0.32
dLunch, γ 200 −10.56*** 0.66
dBlack, γ300 −9.47*** 0.58
dHispanic, γ 400 −6.44*** 0.51
dAsian, γ 500 1.19 1.03
dNatAmer, γ 600 −7.68*** 0.98
dOther, γ 700 −6.95** 2.53
Discussion
Results of the present study should be interpreted with great care. Specifically,
these findings are indicative of the quality of state standards that were in
effect in 2009 for Grade 4, but of which most are no longer in effect.
Furthermore, the results are indicative of mathematics content standards spe-
cifically and not of mathematics teaching practices generally. It is of utmost
importance that such facts be considered prior to interpreting the findings
presented and discussed here.
The present study presents a complicated view of how various states’ stan-
dards conveyance of mathematical communication affected students’ mathe-
matics achievement. Results revealed that higher frequencies of description
standards in Grade 4 documents did have a statistically significant effect on
mathematics scores. However, the size of these effects were relatively low in
magnitude. Interestingly, higher frequencies of rationale standards were found
to not have a statistically significant effect associated with mathematics achieve-
ment. On one hand, these findings could be interpreted that efforts to include
expectations for mathematical communication are not worthwhile. On the other
hand, there is a large body of research demonstrating young children’s ability to
294 Educational Policy 31(3)
describe and argue about their mathematics (e.g., Hufferd-Ackles et al., 2004;
Yackel & Cobb, 1996). Therefore, expectations for Grade 4 children to
explain and justify their mathematics are reasonable. In addition, prior
research on the effect of NCTM’s (2000) Principles and Standards document
suggest this is not the case (Hamilton & Berends, 2006; Loeber, 2008; Young,
2007). Such documents can be effective, but the majority of U.S. state’s stan-
dards documents simply were not effective, or were not effective enough.
Rather than conclude that standards do not matter, it is more likely that the
incorporation of communication expectations in content standards by various
states has not been done in an effective manner. Such a conclusion would
align with studies such as that conducted by Larnell and Smith (2011), who
found that various states’ inclusion of mathematics communication-oriented
expectations was vague, inconsistent, and not always aligning with the higher
expectations described by NCTM (2000). Although the findings of the pres-
ent study did not report on the quality of such expectations in various state
content standards, findings from this study do suggest that such expectations,
regardless of quality, were included in relatively few of the content standards
by state. Therefore, findings from the present study provide additional evi-
dence that language used in state standards is ineffective at communicated
expectations for students (particularly in regard to providing mathematical
description and rationales).
If the reason for the lackluster effects found in this study rests in the man-
ner in which such expectations were included in content standards, then it is
logical that either low frequency, the vague nature of inclusion, or a combina-
tion of both would influence conditions such that these standards would not
have a sizable effect on student achievement. Although findings from the
present study are general in nature, these findings can inform potential
improvements regarding inclusion of communication expectations in content
standards. Practical implications from this study suggest that mathematics
content standards should include expectations for mathematical communica-
tion in higher frequencies and in language more explicit and supportive of
NCTM (2000) recommendations. While approximately half of states did
include reference to communication in a separate set of expectations, there
appears to be little effect on students’ mathematics achievement whether this
separate section is included in a document or not.5 Therefore, it may be that
for expectations for mathematical communication to be conveyed effectively,
they should be embedded in content standards. However, such expectations
must be conveyed in a meaningful manner. For example, Larnell and Smith’s
(2011) analysis suggests that standards including expectations for mathemat-
ical descriptions should do so in a way that elicits higher reasoning and
engagement in the content. One solution to this mismatch is for states to
Kosko and Gao 295
Figure 1. NAEP extended constructed-response item that includes the questions
“why or why not.”
Source. Obtained from NAEP questions tool (NCES, 2014).
Note. NAEP = National Assessment of Education Progress; NCES = National Center for
Education Statistics.
these particular skills. Although NAEP does include some prompts that solicit
explanations, accessible items do not provide a clear assessment of rationales
on Grade 4 items. Figure 1 provides an example item from the NAEP
Question Tool that seemingly solicits a rationale. However, the scoring rubric
evaluates the correct or incorrect answer along with whether the description
of procedures is appropriate. No mention of rationale or justification is pro-
vided. A similar phenomenon was observed by Kosko and Wilkins (2011) in
which open-response items from various assessments of quantitative literacy
were examined. Specifically, while several items included requests for
descriptions and rationales, the rubrics often did not assess such data from
test takers. If a similar phenomenon is present in NAEP items, it may explain
why the effect sizes obtained in this study were not larger. However, various
studies have documented the positive association between these mathemati-
cal communication behaviors and achievement scores, where achievement
was not measured with items soliciting description or justification (e.g.,
Hiebert & Wearne, 1993; Kosko, 2012; Kosko & Miyazaki, 2012). Therefore,
it is more likely that the findings of the present study are due mainly to the
manner in which various states incorporated communication expectations in
their content standards.
Funding
The author(s) received no financial support for the research, authorship, and/or publi-
cation of this article.
Notes
1. Although National Assessment of Education Progress (NAEP) 2009 also
assessed mathematics in Grades 8 and 12, we limit our focus on Grade 4 for
simplicity of discussion.
2. Although NAEP data exist for the District of Columbia and certain U.S. territo-
ries, we restricted the scope of analysis to U.S. states. States as political entities,
though different from one another, have more similarities in their organiza-
tion than various territories. Thus, a focus on states and excluding District of
Columbia and territories was meant to simplify the focus of this study.
3. Per guidelines for reporting analysis results from restricted-use NAEP 2009 data,
all sample sizes, and indicators of sample size were rounded to the nearest 10.
4. The NAEP 2009 variable used for Experience was YRSEXP.
5. We conducted a similar analysis of expectations for communication (descrip-
tion and rationale) separate from content standards (e.g., inclusion of
National Council of Teachers of Mathematics [NCTM] Process Standards and
variations of such). Results suggest no statistically significant effects of such
expectations.
References
Common Core State Standards Initiative. (2010). Common Core State Standards for
Mathematics. Retrieved from http://corestandards.org/assets/CCSSI_Math%20
Standards.pdf
Clotfelter, C. T., Ladd, H. F., & Vigdor, J. L. (2007). Teacher credentials and stu-
dent achievement: Longitudinal analysis with student fixed effects. Economics of
Education Review, 26(6), 673-682.
Confrey, J., Nguyen, K. H., Lee, K., Panorkou, N., Corley, A. K., & Maloney, A. P.
(2012). Turn-on common core math: Learning trajectories for the Common Core
State Standards for Mathematics. Available from www.turnonccmath.net
300 Educational Policy 31(3)
de Fraine, B., van Damme, J., van Landeghem, V., Opdenakker, M., & Onghena,
P. (2003). The effect of schools and classes on language achievement. British
Educational Research Journal, 29, 841-859.
de Jong, R., Westerhof, K. J., & Kruiter, J. H. (2004). Empirical evidence of a com-
prehensive model of school effectiveness: A multilevel study in mathematics in
the 1st year of junior general education in the Netherlands. School Effectiveness
and School Improvement, 15, 3-31.
Halliday, M. A. K., & Matthiessen, C. M. I. M. (2004). An introduction to functional
grammar (3rd ed.). London, England: Hodder Education.
Hamilton, L. S., & Berends, M. (2006). Instructional practices related to standards
and assessments. Pittsburgh, PA: RAND Education.
Hiebert, J., & Wearne, D. (1993). Instructional tasks, classroom discourse, and stu-
dents’ learning in second-grade arithmetic. American Educational Research
Journal, 30, 393-425.
Hill, H. C. (2001). Policy is not enough: Language and the interpretation of state stan-
dards. American Educational Research Journal, 38, 289-318.
Hill, H. C. (2005). Content across communities: Validating measures of elementary
mathematics instruction. Educational Policy, 19, 447-475.
Hill, M. S., & Jenkins, S. P. (2001). Poverty among British children: Chronic or
transitory? In B. Bradbury, S. P. Jenkins, & J. Micklewright (Eds.), The dynam-
ics of child poverty in industrialised countries (pp. 174-195). Cambridge, UK:
Cambridge University Press.
Hruschka, D. J., Schwartz, D., St. John, D. C., Picone-Decaro, E., Jenkins, R. A., &
Carey, J. W. (2004). Reliability in coding open-ended data: Lessons learned from
HIV behavioral research. Field Methods, 16, 307-331.
Hufferd-Ackles, K., Fuson, K. C., & Sherin, M. G. (2004). Describing levels and com-
ponents of a math-talk learning community. Journal for Research in Mathematics
Education, 35, 81-116.
Koichu, B., Berman, A., & Moore, M. (2007). The effect of promoting heuristic
literacy on the mathematical aptitude of middle-school students. International
Journal of Mathematical Education in Science and Technology, 38, 1-17.
Kosko, K. W. (2012). Student enrollment in classes with frequent mathematical dis-
cussion and its longitudinal effect on mathematics achievement. The Mathematics
Enthusiast, 9(1&2), 111-148.
Kosko, K. W., & Herbst, P. G. (2011). Where’s the proof? Proof in U.S. high school
geometry content standards. In L. R. Wiest & T. Lamberg (Eds.), Proceedings
of the 33rd annual meeting of the North American Chapter of the International
Group for the Psychology of Mathematics Education (pp. 329-337). Reno:
University of Nevada, Reno.
Kosko, K. W., & Miyazaki, Y. (2012). The effect of student discussion frequency on
fifth-grade students’ mathematics achievement in U.S. schools. The Journal of
Experimental Education, 80, 173-195.
Kosko, K. W., Rougee, A., & Herbst, P. (2014). What actions do teachers envi-
sion when asked to facilitate mathematical argumentation in the classroom?
Mathematics Education Research Journal, 26, 459-476.
Kosko and Gao 301
Newton, J. (2009). The relationship between the written and enacted curricula: The
mathematical routine of questioning. In S. L. Swars, D. W. Stinson, & S. Lemons-
Smith (Eds.), Proceedings of the 31st annual meeting of the North American
Chapter of the International Group for the Psychology of Mathematics Education
(pp. 337-345). Atlanta: Georgia State University.
Opdenakker, M., van Damme, J., de Fraine, B., van Landeghem, G., & Onghena, P.
(2002). The effect of schools and classes on mathematics achievement. School
Effectiveness and School Improvement, 13, 399-427.
Porter, A. C. (1994). National standards and school improvement in the 1990s: Issues
and promise. American Journal of Education, 102, 421-449.
Raudenbush, S. W., & Bryk, A. S. (2002). Hierarchical linear models: Applications
and data analysis methods (2nd ed.). Thousand Oaks, CA: SAGE.
Staples, M. (2007). Supporting whole-class collaborative inquiry in a secondary
mathematics classroom. Cognition and Instruction, 25, 161-217.
Tate, W. F. (1997). Race-ethnicity, SES, gender, and language proficiency trends
in mathematics achievement: An update. Journal for Research in Mathematics
Education, 28, 652-679.
Yackel, E., & Cobb, P. (1996). Sociomathematical norms, argumentation, and
autonomy in mathematics. Journal for Research in Mathematics Education, 27,
458-477.
Young, L. (2007). Determining teachers’ behaviors concerning the NCTM standards
in low and high performing rural high schools in Kansas (Doctoral dissertation).
Kansas State University, Manhattan.
Author Biographies
Karl Wesley Kosko is an Assistant Professor at Kent State University in mathematics
education. His program of research centers on mathematical communication with a
focus on student engagement in and teacher facilitation of whole class discussion, and
students’ mathematical writing.
Yang Gao is currently a Doctoral Candidate and Graduate Assistant at Kent State
University. His research interests are English as a second/foreign language (ESL/
EFL) teacher beliefs and practice, and ESL/EFL reading instruction.