Professional Documents
Culture Documents
98 SCIENCE EDUCATOR
in association with the U.S. Department a teacher plans a pedagogical strategy inform instruction is depicted. Here, the
of Education simply call it the data use based on his/her learning objectives or results of the assessment were analyzed
cycle (Institute of Education Sciences, goals (Plan). As a note, we favor the with consideration to the original teach-
2009). For a graphical example, Figure term “goals” over “objectives” as the lat- ing strategy employed. In the final ex-
1 shows the data-driven decision-making ter can hold a connotation (particularly ample (Table 3), data-driven inquiry is
process from the U.S. Department of Ed- among teachers) as learning objectives used to help identify the specific content
ucation (2010; 2011). when data-driven inquiry additionally area in which students are struggling
Data-driven inquiry frameworks re- calls for instructional objectives. Thus, the most. It highlights the importance
sembles scientific inquiry in process, “goals” entails both. Then, s/he imple- of the process that can be used in order
namely, defining a problem, collect- ments the teaching method (Implement), to isolate the detailed, specific learning
ing data, analyzing and interpreting the designs an assessment related to the objective not met by the students. Infor-
data, and then making and assessing learning objective, and collects/organizes mation in these tables is then referred to
a decision. Although the ideas behind the assessment results (Assess). The throughout the description of the indi-
the various processes are similar, the analysis and interpretation (Analyze) of vidual steps.
name is not, which explains the dis- the data can get complicated as s/he can
crepancy between the labels in Figure 1 analyze and interpret both in terms of the Steps of Data-Driven Inquiry
and those we use throughout this paper. learning goals and/or problems or ques- Defining a problem – Step 1. There is
We describe the process for using data tions different from those learning goals a semantic difference that identifies the
to inform teaching with the terms “data (i.e., how effective was the teaching unit for which analysis takes place. Gen-
use process” and “data-driven inquiry” strategy, other compounding factors to erally when the word “goal” or “prob-
throughout this review. As an anecdotal performance, impact of educational set- lem” is used, it refers to a student out-
example for how these terms could be ting, etc.). Finally, a pedagogical action come, a learning objective, or a problem
used in an educational context, we begin is hypothesized through reflection on the with students’ understandings and is set
the cycle depicted in Figure 1 with plan- results and other relevant contextual in- prior to data collection in order to guide
ning. As a note, several authors com- formation (Reflect), and then the assess- the design of assessments. The original
ment that a teacher can start anywhere ment process begins anew with the new goal in Table 2 was to assess students’
on the cycle (Brunner, 2005; U.S. De- pedagogical strategy being used. understandings of atomic structure. Al-
partment of Education, 2010; Institute ternatively, when “hypothesizing” or
of Education Sciences, 2009), but this The Process of Data-Driven Inquiry “question posing” appears, it refers to
example is presented in the chronologi- as Exemplified in Chemistry the attempt to explain or address results
cal order typically seen in teaching. First To aid the understanding of data- of the designed assessments and there-
driven inquiry processes, three examples fore occurs after data collection. These
from the reviewed literature were cho- are hypotheses about the effect of factors
sen and presented in Tables 1-3. These such as educational contexts, individual
examples include scenarios in which and class-level history, and even word-
the data-driven inquiry cycle led the ing of items (as just a few examples) on
instructors to modify the assessment student performance assessments. In Ta-
(Table 1), inform instructional decisions ble 2, the teacher hypothesized that the
(Table 2), and identify content difficul- teaching strategy used may be having a
ties in order to refine assessment goals large impact on the outcome. Analysis
(Table 3). In each table, the first column of both types of questions are important
identifies which step of the data-driven in instructional improvement because in
inquiry process is being exemplified and order to know how to adjust instruction,
the second column contains an example teachers need to know where problems
from the literature. The last column il- exist in students’ understandings, but
lustrates how these examples might exist also need to understand how confound-
in a high school chemistry context, thus ing factors impact the results from which
Figure 1. Data-Driven Decision-Making addressing the first research question. one draws conclusions (Institute of Edu-
(Department of Education, 2010) This shows Table 1 highlights how an original goal cation Sciences, 2009; U.S. Department
a representative data use process, although is modified after the teacher gets assess- of Education, 2011).
it does not use the same nomenclature as we ment results and also illustrates the im- In the data-driven decision making
do in this review. This cycle includes defining
a problem (plan), collecting data (implement
portance of considering the alignment of model, Knapp (2006), Cuban (1998),
and assess), analyzing and interpreting the the assessment task and the learning ob- and Copland (2003) detail the impor-
data (analyze), and making decisions (reflect), jectives. In the constitution example (Ta- tance of the ability to reframe potential
similar to scientific inquiry. ble 2), the use of assessments to directly interpretations of data from multiple
Step in Data-Driven Inquiry 4th Grade Vocabulary (Calfee and Masuda, 1997) Categorizing Reaction Types
Defining a problem One hypothetical 4th grade boy (Sam) may have a poor In high school chemistry, a student (older Sam) may not
vocabulary. understand chemical reactions.
Designing/Collecting assessment An assessment is designed and implemented to determine An assessment that requires Sam to identify reactions as
his grade-level in vocabulary. synthesis, decomposition, etc.
Interpretation and analysis Assessment reveals his vocabulary grade level to be On the assessment, Sam cannot adequately identify the
2.4 (in between 2nd and 3rd grade) so his teacher deems types of chemical reactions.
him not very smart.
Making instructional decisions Teacher places him in a low ability group in order to give Figuring Sam doesn’t understand chemical reactions, the
him slower pace. teacher goes over the definitions and how to identify
them multiple times.
Defining an alternative Teacher thinks that Sam possesses an adequate Being able to identify types of chemical reactions is not
problem/hypothesis to describe results vocabulary, but does not perform well on the skill of the only aspect of understanding them, and he may
vocabulary recall, which is only one aspect of understand other aspects of reactions.
understanding vocabulary.
Designing/Collecting assessment The teacher tasks Sam to define words such as The teacher develops an alternative assessment that asks
“petroleum,” use it in an original sentence, and match Sam to predict the products including states of matter
the word to a definition in different assessments. and balancing of various types of reactions.
Interpretation and analysis If Sam performs differentially on these tasks and others If this assessment yields different results, then Sam
like it, then he probably understands the word but the probably understands one but not all aspects of
type of assessment affects his performance and thus chemical reactions.
only the pure recall skill required by that assessment
type may be what Sam struggles with.
Making instructional decisions No instructional decision was provided with this example, With additional information, the teacher can either target
however, they do say the following: The teacher needs specifically the identification aspect or make curricular
to consider the consequences that (a) short-term change regarding chemical reactions if that teacher
assessment(s) has/have on long-term decisions, such as decides identification is not as highly valued as other
the decision to place Sam in a low ability group for what aspects of chemical reactions.
could simply be due to the assessment design or context
(such as time).
perspectives. These multiple interpreta- the analysis and interpretation of the re- 2010). In its simplest definition, instruc-
tions, formed as questions or hypotheses sulting data. These suggestions are best tional sensitivity is the extent to which
(Calfee & Masuda, 1997), give teachers illustrated in Table 3 where the teacher students’ performance reflects the qual-
the opportunity to access a wide variety modifies a complex goal (understanding ity of instruction received (Koscoff &
of information about their students and of stoichiometry) to a simpler goal (un- Klein, 1974).
their own teaching from one item, one derstanding of molar ratios) thereby as- To demonstrate instructional sensitiv-
assessment, or a group of assessments. In sessing that goal in a more valid manner. ity, imagine a chemistry teacher wants to
the U.S. Department of Education large Designing assessments and collect- evaluate the effectiveness of a simulation
scale study of elementary, middle, and ing data – Step 2. Teachers frequently of particles responding to increasing tem-
high school teachers (2008), only 38% design, collect, and analyze students’ perature and therefore administers some
of teachers reported having profession- data using formative assessments such as form of assessment. If the a) content as-
al development that focused on how to quizzes, homework, in-class activities, sessed by the assessment aligns with the
formulate questions to answer with data and tests. Many of these contain items learning objectives, b) items are not be-
from assessments. To address this, we that are designed (or at least chosen) ing misinterpreted by the students, and c)
refer teachers to the IES’s guidelines for by the teachers themselves. A constant format of the response allows the teacher
a good hypothesis: 1) identifies a prom- concern for these items is the extent to to validly determine students’ thought
ising intervention, 2) ensures that out- which the results can be used to deter- processes, the results can be interpreted
comes can be measured accurately, and mine the instructional effectiveness. This to make conclusions about the effect of
3) lends itself for comparison study (pre- has been referred to as consequential the simulation on student learning. Factors
post or treatment-control designs) (Insti- (Linn & Dunbar, 1991; Messick, 1989), a-c are aspects of instructional sensitiv-
tute of Education Sciences, 2009). Ad- instructional (Yoon & Resnick, 1998), ity, and without them being considered
ditionally Suskie (2004) warns against or pedagogical validity (Moran & in some way, any conclusion about
having too many learning goals or inap- Malott, 2004), but has recently been the student responses would be suspect.
propriate (too complex or too simple) called instructional sensitivity (Ruiz- For example, an assessment item that
learning goals as this negatively affects Primo, 2012; Popham, 2007; Polikoff, simply asks “predict if the volume of
Step in Data-Driven Inquiry U.S. Constitution (Calfee and Masuda, 1997) Atomic Structure
Defining a problem A teacher wants students to have a fundamental A teacher wants students to have fundamental
understanding of the U.S. Constitution. understanding of atomic structure.
Designing/Collecting assessment In order to gain information on what students already The teacher asks the students to tell them anything they
know about the Constitution, the teacher asks his know about the structure of an atom. Some might
students to tell him something about the Constitution. volunteer something about protons, neutrons, and
After total silence, he follows up with open questions electrons but provide little follow-up. The teacher
about the federal government in Washington D.C. After could then ask about plum pudding versus solar
another long silence, he writes keywords on the board system models, periodicity, or sub-atomic particles,
such as “President,” “Congress,” and “Supreme Court” but gets little response.
hoping to elicit dialogue, but still no response.
Interpretation and analysis One possible conclusion is that the students actually Considering the lull in discussion, the teacher assumes
know nothing about these topics, informing the they know very, very little about atomic structure.
teacher that he will have to start from scratch.
Making instructional decisions The teacher begins at the absolute basics, tailoring As a result, the teacher begins talking about charges of
his instruction to include all aspects of the sub-atomic particles and basic construction of the
government and the Constitution. atoms.
Defining an alternative problem/hypothesis A hypothesis is that the students do not normally A hypothesis is that the students do not normally
to describe results participate in such open discussions and for lack of participate in such open discussions and for lack of
knowing how to react, students remain silent. knowing how to react, students remain silent.
Designing/Collecting assessment To test this right on the spot, he shifts gears and asks To test this, the teacher asks more convergent questions,
for his students to tell him something about weather such as what charges do the proton, neutron, and
because he knows they understand weather. electron have?
Interpretation and analysis If the response is no longer silence, this may be an If the students can answer these questions, then it could
indication that it is the pedagogical style (or other have been the open-ended discussion, not the lack
contextual aspects) that is yielding the silence, not in content knowledge that cause students to remain
necessarily a lack of content knowledge regarding silent.
the U.S. Constitution.
Making instructional decisions No instructional decision was provided with this example. Further, if the teacher’s interpretation is that students
don’t do well with open-ended discussions, that
teacher may revise his prompts to elicit more from
the students.
gas will increase or decrease at higher and what is assessed by the items leads standards for evaluation of instructional
temperature,” has limited sensitivity to to misinterpretations and missed oppor- sensitivity have been published. Outside
instruction in this case for three reasons: tunities to gather valuable information of “typical” sources of assessment data
1) it assesses prediction, whereas the about teaching and learning. Project (i.e., homework, quizzes, etc.), Calfee
simulation focuses on explanation (related 2061 of the American Association for and Masuda (1997) would argue that, in
to factor a); 2) “volume of gas” can be the Advancement of Science provides light of a classroom assessment as ap-
misinterpreted by students as “volume of alignment criteria of this nature that plied social science research (Cronbach,
gas particle” (factor b); and 3) the item is may prove helpful for chemistry teach- 1988), teachers should be open to data
susceptible to guessing given its current ers (Roseman, Kesidou, & Stern, 1996; collection complete with observations
format (factor c). A more instructionally Stern & Ahlgren, 2002). Chemistry ex- and interviews.
sensitive item might be “using drawings amples of this alignment (or the lack Interpretation and analysis of data –
of gas particles, explain why an increase thereof) can be found in Table 1 and the Step 3. Even when an assessment of any
in temperature will cause an increase in implications section. kind has been designed so that learning
volume” because it minimizes factors Instructional sensitivity is an important objectives and assessed concepts are
a-c, meaning that results can more read- consideration in the data use process be- aligned, the task of interpreting results
ily be used to inform instruction. cause if teachers use items that are sen- in a way that is meaningful to instruc-
It is commonly stated that the assess- sitive to their instruction, they can use tors is daunting. Textbooks on classroom
ment items must align with the learning data from those items to adjust instruc- assessment (McMillan, 2011; Popham,
objectives of that particular unit (Irons, tion (Ruiz-Primo, 2012). Judging both 2002; Witte, 2012) frequently discuss
2008; Anderson; 2003; Taylor, 2003). In quantitatively and qualitatively the degree the importance of:
regards to a data-driven inquiry process, to which items are instructionally sensi- • understanding validity, reliability,
this alignment is crucial. Lack of align- tive has been examined (Popham, 2007; descriptive and (minimal amount
ment between goals of the assessment Burstein, 1989; Polikoff, 2010), but no clear of) inferential statistics
• ethics of assessment teachers generally want to see that stu- (NCLB, 2002), the U.S. Department of
• absence of bias when evaluating dents have learned what they have in- Education launched a series of large-
students so as to inform an tended the students learn. This method, scale studies to assess teachers’ abilities
instructor, often associated with a performance- to interpret quantitative assessment data
• means of summarizing and repre- oriented learning approach, is one of two (U.S. Department of Education, 2008;
senting results for oneself as an general approaches. The other interpre- 2010; 2011). First, the U.S. Department
instructor, other teachers, parents, tation strategy is a growth approach, or of Education found that teachers wanted
and other decision-makers. student-referenced assessment (Harlen more opportunities for professional de-
However, even an understanding of the & James, 1996), meaning that teachers velopment specific to the interpretation
previously mentioned psychometric as- reference students’ previous assessments and incorporation of data from standard-
pects does not describe how it applies to a in interpreting the current assessments. ized tests, and that they were unlikely
teacher’s particular content, context, and It is generally accepted that these two to do this if they were not confident in
pedagogical style and ability. The previ- methods should be performed in con- their ability to do so (U.S. Department
ously mentioned interpretation strategies junction with each other in interpreting of Education, 2008). A few years later,
generally align with the criterion-refer- and delivering the results of assess- a different sample revealed that teachers
enced era of assessment, meaning the ments (Bell & Cowie, 2002). As shown preferred to interact with colleagues on
results are interpreted based on predeter- in Table 1, Sam is assessed by a criteria common assessments to interpret data
mined criteria defined by the discipline (understanding of chemical reactions), rather than participate in formal profes-
(Linn, Baker, & Dunbar, 1991). In sci- but a growth model could easily be in- sional development (U.S. Department of
ence classrooms, Bell and Cowie (2002) corporated by repeated measurements of Education, 2010). With the most recent
note that teachers frequently rely on cri- understanding of chemical reactions. sample, teachers seemed to express diffi-
teria (where the criterion is a scientific Some years following the implemen- culties in fundamental data interpretation
understanding of phenomena) because tation of No Child Left Behind Act skills, such as differentiating bar graphs