Professional Documents
Culture Documents
https://doi.org/10.1007/s11229-018-01995-9
Abstract
Representationalism—the view that scientific modeling is best understood in rep-
resentational terms—is the received view in contemporary philosophy of science.
Contributions to this literature have focused on a number of puzzles concerning the
nature of representation and the epistemic role of misrepresentation, without consid-
ering whether these puzzles are the product of an inadequate analytical framework.
The goal of this paper is to suggest that this possibility should be taken seriously.
The argument has two parts, employing the “can’t have” and “don’t need” tactics
drawn from philosophy of mind. On the one hand, I propose that representationalism
doesn’t work: different ways to flesh out representationalism create a tension between
its ontological and epistemological components and thereby undermine the view. On
the other hand, I propose that representationalism is not needed in the first place—a
position I articulate based on a pragmatic stance on the success of scientific research
and on the feasibility of alternative philosophical frameworks. I conclude that repre-
sentationalism is untenable and unnecessary, a philosophical dead end. A new way
of thinking is called for if we are to make progress in our understanding of scientific
modeling.
1 Introduction
Many contributions to the philosophy of science literature highlight the central role
played by mediated or indirect forms of investigation. Rather than directly intervening
upon the various real-world phenomena they are interested in, scientists often build
and manipulate models that simulate those phenomena. Direct interventions are some-
times impractical, dangerous or even unethical. There are moral limits to how scientists
123
Synthese
can use human subjects and animals in the lab, just as there are practical barriers to
direct experimentation on global climate change due to the phenomenon’s complex-
ity and spatio-temporal scale. In these and other cases, modeling enables scientists
in all disciplines to indirectly advance their understanding of real-world phenomena.
And not only are there different reasons for modeling, there are also many different
ways to do it, utilizing distinct methods and modalities. Diagrams, graphs, mathe-
matical equations, and concrete scale models have been used in science for centuries.
More recently, technological advances have led to the widespread use of computer
simulations and robotic models.
But how is it possible to learn through modeling? It is easy to see how mathe-
matical equations, computer simulations, and robotic agents could help advance our
understanding of mathematics, computing, and robotics, respectively. How is it that
building and operating models enables scientists to learn something not only about the
models themselves, but also about the real-world phenomena scientists are ultimately
interested in? In short, why is modeling epistemically valuable? The answer seems
obvious and intuitive: models can give us knowledge of their targets because they
represent those targets. That is, models are related to real-world target phenomena
in such a way that they can stand in for those targets in empirical research and that
the outcomes of modeling generate knowledge of the phenomena being modeled. It is
because they are representations of certain phenomena that mathematical equations,
computer simulations and concrete models are viable indirect routes to understanding
and explaining those phenomena.
This intuitive answer has, over the past couple of decades, shaped the literature on
scientific modeling and brought to the forefront of philosophers’ attention a number of
problems relating to the nature of representation and the role of misrepresentation in
science. Explaining what representation is and how it works turns out to be no trivial
matter. Philosophers generally agree that modeling is a legitimate means to knowledge,
and they also generally agree that the intuitive answer above is right, i.e., modeling
is epistemically valuable because models stand in a special relation to their targets
(namely, one of representation). But how to support this representationalist intuition?
What is it about representation that makes it epistemically valuable in the way we take
it to be and that justifies attempting to analyze model-based science in representational
terms? Some theories of representation stand out as the most influential, but none has
escaped criticism unscathed. And the continued development of revamped or brand-
new theories to correct what was wrong with their predecessors seems to only multiply
the disagreement about what representation is and how it works.
In a situation like this, when an intuitive view gives rise to problems no one appears
to be able to solve, it is wise to at least entertain the possibility that our intuitive
view was mistaken. This is what I do in this paper. I begin, in Sect. 2, by giving a
brief overview of recent controversies in order to make explicit the representationalist
assumptions that pervade the literature. I identify representationalism as a method-
ological stance that is based on a twofold assumption comprised of an ontological
component and an epistemological component, and I argue that these components
create serious challenges for making sense of the role that misrepresentation (e.g.,
idealization and abstraction) is said to play in science. I then provide a two-part argu-
ment against representationalism inspired by the “can’t have” and “don’t need” tactics
123
Synthese
used in a different debate about content in philosophy of mind (see Hutto and Myin
2013; Myin and Hutto 2015). My goal is to show, first, that we cannot have a repre-
sentationalist approach to scientific modeling, and, second, that we do not even need
one. The “can’t have” argument, presented in Sect. 3, proposes that representational-
ism does not work because, in any way of fleshing out the view, its ontological and
epistemological components undermine one another. The “don’t need” argument, in
Sect. 4, proposes that, regardless of whether representationalism is tenable or not, rep-
resentationalism is unnecessary: pragmatic and conceptual considerations suggest that
we can address what is philosophically interesting about scientific modeling without
getting into the representationalist quagmire. Representationalism is a familiar way of
thinking about model-based science, but it is a philosophical dead end, and we have
good reason to pursue alternatives.
How can building and manipulating models generate knowledge of some target phe-
nomenon? What makes this question particularly puzzling is the fact that, as many
philosophers point out, models always contain idealizations and abstractions that
make them imperfect copies of their targets. Models abstract away many of their tar-
gets’ complexities, neglecting details where possible so as not to complicate matters
unnecessarily. Models also include intentional distortions. They often posit processes,
elements and properties that are absent in the target, or that cannot even be found
anywhere in nature, as in the case of infinite populations, frictionless planes, and the
rational self-interested agent. These lies “by omission” and “by commission,” as it
were, pose the challenge of elucidating the contribution of “falsehoods” to scientific
modeling, directing philosophers to the question how can scientists learn through
misrepresentation?
A considerable amount of attention in recent years has been devoted to explaining
the status of misrepresentation in model-based science. Abstractions, simplifications,
approximations and idealizations are recognized as useful means to the future develop-
ment of “truer theories” (Wimsatt 1987), and as serving at the very least as temporary
placeholders for more accurate descriptions. But some in the literature go so far as to
claim that falsehoods are not defects of a model, but are often instrumental to its suc-
cess: “fictions can be genuinely explanatory” (Bokulich 2012, p. 736); “false models
can explain, and (...) they often do so in virtue of their idealizations” (Kennedy 2012,
p. 332); and “idealizations aid in representation not simply by what they eliminate,
such as noise or noncentral influences, but in virtue of what they add, that is, their
positive representational content” (Potochnik 2017, p. 50). The options seem clear
enough: either we hold that explanatory success requires truth/accuracy and accord-
ingly see the false parts of models as explanatorily superfluous, or we agree with the
philosophers just quoted in holding that, at least sometimes, the false parts of models
are themselves required for models to be successful. Bokulich explains the situation
as follows:
123
Synthese
The field has largely split into two camps on this issue: those who think it is
only the true parts of models that do explanatory work and those who think
the falsehoods play an essential role in the model explanation. Those in the
former camp rely on things like de-idealization and harmless analyses to show
that the falsehoods do not get in the way of the true parts of the model that do
the real explanatory work. Those in the latter camp have the challenging task
of showing that some idealizations are essential and some fictions yield true
insights. (Bokulich 2017, p. 108)
123
Synthese
123
Synthese
123
Synthese
stand why modeling is successful: this is, most fundamentally, what is communicated
in claims that “we need to know the variety of ways models can represent the world
if we are to have faith in those representations as sources of knowledge” (Morrison
2015, p. 97, emphasis added) and that “if we want to understand how models allow us
to learn about the world, we have to come to understand how they represent” (Frigg
and Nguyen 2017a, p. 49, emphasis added). In short, EC motivates representation-
alism as a methodological stance for understanding scientific modeling because it
delineates a general way of thinking about why models are epistemically valuable
(namely, because they represent), which can then be developed into different views of
how modeling works given particular accounts of the nature of representation.
To conclude this section I want to briefly indicate an initial difficulty attending
representationalism, namely that the role that many philosophers assign to idealization
and abstraction in scientific modeling creates puzzles connected to each component
of the twofold representationalist assumption. Following the ontological component
of representationalism, for some object (say, a pair of first-order differential equations
or a robotic agent) to be a model of some real-world target, a minimal requirement is
that it represent that target. That is, OC entails that at least part of what makes X a
model of some target Y is the fact that X represents Y, even if inaccurately (or partially,
if you use ‘representation’ as a success term). Yet, this seems incompatible with the
common idea, reviewed above, that introducing idealizations and abstractions does
not necessarily stand in the way of a model’s success and sometimes even enhances
the model. Put simply, the puzzle is: if what makes X a model of Y is that X represents
Y, then how can X still be a model of Y when X misrepresents Y, or falls short from
representing Y accurately?
A similar problem is associated with the epistemological component of representa-
tionalism. EC frames the epistemic value of modeling in terms of the representational
relation between model and target. In other words, EC says that modeling advances
scientific knowledge of target systems and phenomena because models represent those
targets. The problem, however, is that EC seems to be in tension with the purported
role of misrepresentation in science. If, as suggested by EC, models are epistemically
valuable in investigations of some target because they represent the target, then how
can models sometimes be more epistemically valuable when they misrepresent, i.e.,
when the representational relationship between the two is faulty?
These two puzzles show that popular views concerning the productive role of
misrepresentation (or partial representation) in science are in tension with the represen-
tationalist stance they presume and its ontological and epistemological commitments.
One might take this as indicating that our understanding of idealization and abstraction
is in need of revision, and I think this is right. But perhaps a deeper lesson is that the
representationalist intuition we took for granted just pushed our questions to another
level instead of answering them. The assumption that models are best understood,
ontologically and epistemologically, in representational terms gives rise to controver-
sies surrounding the nature and status of misrepresentation. This assumption creates
the need to account for how the parts of a model that do not represent accurately are
themselves epistemically valuable or, at least, how they do not get in the way of the
parts that are. But these problems only arise if we accept representationalism in the
first place.
123
Synthese
Representation is typically thought of, broadly speaking, as a relation. But exactly what
kind of relation is it, and between what sorts of entities does it hold? Traditional views
treat representation as a mind-independent relation between model and target, in which
the two are related by virtue of some property such as isomorphism or similarity. In
views like these, the representation relation exists whenever the appropriate objective
correspondence holds between model and target. This traditional conception of repre-
sentation as mind-independent has been dubbed “informational” (Chakravartty 2010)
because it sees models as objectively containing information about target phenom-
ena, and it has also been described as “dyadic” (Giere 2004; Suarez 2004; Knuuttila
2011) in that it sees the representation relation as one that holds between two entities
only, i.e., model and target. The designations “informational” and “dyadic” are con-
trasted with, respectively, “functional” views which see representation as a relation that
depends on human insight or use and is established by the activities of cognitive agents
(Chakravartty 2010), or “triadic” views which see representation as necessarily hold-
ing between three entities: agents, models and targets (Knuuttila 2011). In what follows
I will focus first on the mind-independent (or informational, or dyadic) conception of
representation and then on the mind-dependent (or functional, or triadic) conception to
show how each fails with representationalism in different ways. It is worth noting from
the outset that most participants in the literature have moved toward adopting a view of
representation as a mind-dependent agential accomplishment—including even those
123
Synthese
123
Synthese
r
M odel T arget
Fig. 1 Representation as mind-independent: the model represents the target, and this objective two-place
representation relation (r) is defined in terms of, e.g., isomorphism (van Fraassen 1980) or similarity (Giere
1988)
have also been revised (or re-explained) by their own authors. But our level of analysis
is different. What matters for present purposes is to understand the theory-type (i.e.,
the conception of representation as mind-independent) and to determine how it fares
with the twofold representationalist assumption. In short, our question is not whether
these views get representation right, but rather: is representationalism, as composed
of ontological and epistemological intuitions (i.e., OC and EC), supported by a view
of representation as a mind-independent, informational, two-place relation?
At first blush, views of representation as mind-independent seem highly promising.
They provide a simple way to flesh out OC: saying that models stand in a representa-
tional relation to their targets means that there is some actual correspondence between
them, be it one-to-one structural identity or some other form of similarity. Views of
representation as mind-independent can thus point to objective features of a model
and of a target that make the one a representation of the other. And this answer to
OC, in turn, lends support to EC. Representation is not something we make up; and
the existence of an objective correspondence between a model and its target inspires
confidence in that model as a source of knowledge about the target. This makes EC
seem perfectly unproblematic: model-based science is epistemically valuable because
successful models stand in the appropriate mind-independent representational relation
to their targets and can, for this reason, act as sources of knowledge about those targets.
However, the apparent success of mind-independent representation with EC is
undermined by its failure with OC. Note how EC relies on OC: the idea that mod-
eling is epistemically valuable because of its representational nature (i.e., EC) relies
on the idea that modeling actually has this representational nature (i.e., OC). But
mind-independent views notoriously fail as explanations of how scientific represen-
tation works, that is, as attempts to flesh out OC. And, by failing with OC, views of
representation as mind-independent make representationalism untenable.
A first difficulty attending mind-independent views of representation is the problem
of the asymmetry of modeling. For any account of representation as mind-independent
to be successful in elucidating how model-based science works, it needs to accom-
modate the inherent asymmetry that characterizes modeling. Scientists build and
manipulate models to learn about real-world systems, but usually not the other way
around: scientists do not, for example, intervene on the global climate as a means to
understanding how computer simulations work. To say the same using more explicitly
representational terms, models represent their targets, but it does not seem right to say
that target systems also represent their models. Because representation is an asymmet-
ric relation, accounts of representation need to respect this asymmetry if they are to
be plausible. Isomorphism is a symmetric relation and therefore fails in this respect:
if A is isomorphic to B, then B is isomorphic to A as well (Suarez 2003; Knuuttila
2010). The same is true for similarity: if a model is similar to its target to some degree
and in some respects, then the target will necessarily be similar to the model as well,
to the same degree and in the same respects (Suarez 2003).
123
Synthese
123
Synthese
According to the views considered so far, the correspondence between a model and
a target exists purely because of the features that the model and the target have in
common—purely by virtue of what the model is like and what the target is like. But,
consider that, while two entities may relate to one another in a number of mind-
independent ways, ordinary usage seems to suggest that one represents the other only
for humans, that is, only if humans use the one as a representation of the other. Suppose
you witness a traffic accident and later you tell a friend about it, using your hands to
show how the two cars in front of you collided. You might find it unproblematic to
say that in this context your hands represent the two cars. And you might grant that
there are mind-independent correspondences between hands and cars, some degree of
similarity, such as their roughly approximate width-to-length ratio. Yet, it would seem
wrong to say that hands always represent cars: they can represent cars and they do so
in situations like the one described; still, what makes hands represent cars is not any
objective correspondence between the two, but the fact that we intentionally use them
that way, if and when we do so.
The recognition that traditional views failed to account for the role that human
intention plays in representation has spurred alternative ways of understanding the
representation relation. Rather than framing it as an objective relation between two
entities, more recent accounts treat representation as a mind-dependent relation that
123
Synthese
Scientists
M odel r
T arget
Fig. 2 By using a model to represent some target, scientists establish (e) that the model represents the target
(r). Strictly speaking, the representation relation is mind-dependent and therefore not reducible to r: it is
necessarily constituted by both r and e
A model can (be used to) represent a given phenomenon accurately only if it
has a substructure isomorphic to that phenomenon. (That structural relationship
to the phenomenon is of course not what makes it a representation, but what
makes it accurate: it is its role in use that bestows the representational role.) (van
Fraassen 2008, p. 309, italics original)
The formula is: Agents (1) intend; (2) to use model, M; (3) to represent a part
of the world, W; (4) for some purpose, P. So agents specify which similarities
are intended, and for what purpose. This conception eliminates the problem of
multiple similarities and introduces the necessary asymmetry. I propose to call
this “The Intentional Conception of Scientific Representation.” (Giere 2010, p.
274)
123
Synthese
these are perfectly legitimate two-place relations that may be of scientific interest. But,
in a functional view of representation as mind-dependent, neither isomorphism nor
similarity (nor any other two-place relation) entails a relation of representation. To be
sure, we often take advantage of correspondences between different entities when we
use one to represent the other, as in the case of hands and cars discussed above. Still,
in views of representation as mind-dependent, it is our use of these correspondences
(rather than the correspondences themselves) that creates the representation relation.
Is representation in fact a mind-dependent relation? This is not a question I will
attempt to answer here. Instead, what matters for this paper’s argument is to deter-
mine how views of representation as mind-dependent fare with OC and EC. This
decides whether mind-dependent views can support the representationalist assump-
tions philosophers of science by and large take for granted. We have seen that views
of representation as mind-independent fail with OC. In this respect, mind-dependent
views appear to fare better. Indeed, the main motivation for the formulation of mind-
dependent views in the first place was precisely to give a better account of the
representation relation—in the current terminology, to flesh out OC in a more appro-
priate way. And that they do. Mind-dependent views accommodate the intuitions many
people have about how representation works, most importantly intuitions about how
representation involves intention and agency. Mind-dependent views seem to tell a
more realistic story about what makes a model a representation of its target: models
represent their targets, not independently from human activity, but as a result of sci-
entific practice and the purposes guiding research; i.e., models represent their targets
because they are used as representations of those targets.
The problem, however, is that mind-dependent views strengthen OC at the cost of
weakening EC. Granting for the sake of the argument that the functional conception
is correct and representation is a mind-dependent relation, this secures OC in a way
that undermines EC. Mind-dependent views dissociate the question of what makes
something represent well from the question of what makes it represent at all (see, e.g.,
Callender and Cohen 2006; Chakravartty 2010; Knuuttila 2011). This means that, in a
mind-dependent view of representation, anything can by definition represent anything
else in some context as long as someone is willing to establish the relevant repre-
sentational mapping: intention and stipulation are not sufficient to make something
represent successfully or accurately, but they are enough to make it a representation.
The problem is that even if anything can represent anything else for someone, plausi-
bly not anything can be informative about anything else. Scientists do not use any old
representation to model their targets of investigation because not any representation
would be explanatory, illuminating, or epistemically valuable in some way. If repre-
sentation entails only usage, not success or accuracy, then it no longer makes good
sense to also hold that the model-target representational relation is what enables us
to learn about a target through modeling. Accepting the conception of representation
as mind-dependent would render the representationalist stance absurd: in this view
we can paraphrase EC and OC, respectively, as stating that scientists can use mod-
els to learn about target phenomena because models represent their targets, and that
models represent their targets because scientists use them as representations of those
targets—in short, this would mean that the reason scientists can use models to study
real-world phenomena is that they do use them to study real-world phenomena.
123
Synthese
123
Synthese
123
Synthese
value with a non-representational criterion is not a way to resolve the tension between
EC and OC in the mind-dependent conception of representation because that would,
instead, necessitate rejecting EC: in this view, models would be epistemically valuable
due to some non-representational relation they bear to their targets, rather than due
to the representational relation between the two. And by failing with EC, even this
hypothetical substantive mind-dependent account of representation would undermine
representationalism.
The advocate of representationalism might recognize that she cannot rescue EC but
still resist the conclusion that the representationalist methodological stance is unjus-
tified: she could do so by claiming that she was never committed to EC as framed
earlier but rather to an alternative formulation (say, EC*) according to which mod-
els are epistemically valuable because of their representational accuracy (rather than
because of their representational status simpliciter). But this move would not avoid
the criticisms just raised. Instead, the same problem would arise of whether accuracy
should be defined in mind-dependent or mind-independent terms. Not only that, how-
ever, this move would also raise additional problems. One problem is related to our
earlier discussion about misrepresentation. As seen in Sect. 2, an increasingly popular
philosophical view is that idealizations, abstractions and other misrepresentations can
actually make models better—i.e., that sometimes models are epistemically valuable
not despite misrepresenting their targets but precisely because of misrepresentation.
Adopting something like EC* and making accuracy the criterion for epistemic value
would entail that idealization and abstraction (i.e., inaccurate representation) can-
not make models better epistemic tools. This would not only go against a growing
philosophical view, but would also fly in the face of scientific practice and make the
widespread reliance on idealization and abstraction into a mystery—something that,
somehow, is more and more used for epistemic purposes but is not epistemically valu-
able. And lastly, adopting something like EC* would in fact motivate abandoning
the methodological stance of representationalism. If fleshing out our view of models
as representations (in a mind-dependent conception) forces us to hold that the epis-
temic value of modeling is not due to its representational nature (but rather to some
notion of accuracy), then we no longer need an account of scientific models in rep-
resentational terms because the real epistemological heavy-lifting would be done by
the model-target correspondence or informational, two-place relation (which is not
representational). Accordingly, if our goal was to find out why scientific modeling is
successful and how it leads to knowledge of real-world target phenomena, figuring out
what the right theory of representation is would not help because the answer to that
question will be some non-representantional mind-independent feature that models
and targets have in common.
The previous section proposed a “can’t have” argument, showing that the methodolog-
ical stance of representationalism does not work: the different ways to flesh out the
representationalist approach, based on views of representation as mind-independent
or as mind-dependent, all undermine representationalism by failing with at least one
123
Synthese
of its two components (i.e., OC or EC). Naturally, there is no alternative way to flesh
out representationalism because there are no alternative conceptions of representation:
either representation is mind-independent or it is mind-dependent. What then? In this
section I offer a “don’t need” argument, claiming that, whether or not representation-
alism is something we can have, a representationalist approach is not actually required
for making sense of model-based science. I discuss two reasons why representation-
alism is uncalled for, and along the way I indicate promising directions for advancing
our understanding of scientific modeling without stepping into the representationalist
minefield.
The first reason philosophy of science does not need representationalism is what I call
the pragmatic reason. Scientists build and manipulate models of many different sorts
to study a wide range of phenomena, and in this process, they learn many different
things about how the world works. Now, most scientists clearly do not have a theory
of representation. Some chemists use simulations to study chemical reactions, some
biologists use mathematical equations to understand population dynamics, and some
economists have used a hydraulic analog computer like the Phillips Machine to learn
about financial systems—and they all manage to learn about their objects of study via
modeling without having an account of what it means for their simulations, equations
and machines to represent chemical reactions, populations and financial systems. Of
course, scientists have standards, both explicit and tacit, that guide the procedures and
tools they use in their research. But these are standards for making “good models,”
not necessarily criteria for making (accurate) representations of target phenomena.
In practice, the precise meaning of “good model” depends on a number of factors.
Crucial factors implicated in what makes a “good model” typically include the research
project’s disciplinary and theoretical context, the methodological and technological
background (i.e., which prior “successful” models a given project builds on), and what
sorts of questions the project is meant to address (e.g., if it aims to generate predictions
of future events, explain past events, guide real-world interventions, etc). Modeling
is always informed by these disciplinary, theoretical, methodological, technological,
erotetic and purposive aspects of particular research projects. And, taking scientific
practice at face value, success in following these various kinds of standards is what
justifies the epistemic claims coming out of model-based research. That is, models
are constrained in their construction and usage by all of these theoretical and practical
standards that have come to be accepted by the scientific community as justified paths
to knowledge, and, as a result, models have their epistemic justification built in: models
are meant “to solve [certain] tasks, and their success or failure at the task at hand is
both the measure of their value and the justification of their design” (Isaac 2013, p.
3622).
The representationalist methodological stance is predicated on the assumption that
the single most important factor in model-based science, and the factor to be elucidated
philosophically, is the (representational) relation between a model and a target system:
models are taken to be “about” their targets, and to be evaluated by scientists in terms
123
Synthese
of the success of this relation. While I will have more to say concerning “aboutness”
in Sect. 4.2, what matters for the present point is to see how it is more adequate, prag-
matically speaking, to bring to the forefront other factors, such as the disciplinary,
theoretical, methodological, technological, erotetic and purposive aspects I have indi-
cated here. An example might help illustrate this point. Climate modeling, for instance,
typically involves the use of multiple models: no single model is accepted as “the right
model” but, together, a number of competing models are used to support explanatory
or predictive practices (see, e.g., Lloyd 2010; Parker 2011). In these typical cases, the
various climate models are representationally incompatible (i.e., if analyzed represen-
tationally, each model “describes” the climate in ways that are known to be false and
which contradict the other models). Yet, following theoretical and practical (method-
ological, technological, etc) standards, scientists evaluate their model-based climate
research as more or less successful and justified to the degree that it enables them to
accomplish relevant goals. The pragmatic stance I am proposing here takes this feature
of climate modeling to be the rule in science rather than the exception. The primacy of
pragmatic value (understood broadly to include the various theoretical, methodolog-
ical, technological, and other practical aspects) explains why scientists can advance
their knowledge through model-based research without having a theory of represen-
tation and without needing to wait for philosophers to provide the correct account of
representation that explains the success of good models and the failure of bad ones.
In practice, scientists care about whether or not their models “work”—whether the
models fruitfully connect with and advance current knowledge, methods, technology,
and research goals—not whether the models represent real-world target phenomena
in the sense of meeting some formal definition of “representation.”
What I have said up to now highlights one way in which this is a “pragmatic”
reason, namely in that it takes scientific practice seriously and recognizes the practi-
cal epistemic justification of modeling as independent of representational status and
representational accuracy. But this reason is also “pragmatic” in the way it connects
to the philosophical approach of the American pragmatist tradition. The reader might
be willing to grant that scientists do not need a theory of representation in order to
be epistemically successful, but still be inclined to hold that philosophers need one
in order to adequately make sense of the scientists’ epistemic success. Here, though,
the pragmatist response is that this would only lead to bad philosophy. William James
(1907) describes the now famous anecdote of a man who goes around a tree to try to
see a squirrel while the squirrel is, at the same time, also going around the same tree
to try to hide from the man: in going around the tree, does the man also go around
the squirrel? James concludes that the disagreement between competing answers to
this question is pointless: “If no practical difference whatever can be traced, then the
alternatives mean practically the same thing, and all dispute is idle” (p. 45). Later on
in the same lecture and generalizing the lesson from the anecdote to various philo-
sophical debates, James claims: “It is astonishing how many philosophical disputes
collapse into insignificance the moment you subject them to this simple test of tracing
a concrete consequence. There can be no difference anywhere that doesn’t make a
difference elsewhere” (pp. 49–50, emphasis original).
James’ quote could perhaps be read as condemning any kind of abstract
philosophizing—but that is more than I will ask the reader to accept here. My point,
123
Synthese
123
Synthese
123
Synthese
Agent
pA
pB
ObjectA ObjectB
Fig. 3 Practical engagement with one object (pA) can facilitate the development of skills useful for engaging
with another object in a different context (pB): e.g., playing with a beach ball or a toy guitar can help develop
the motor skills needed for participating in a real soccer match or for performing with a band. Object A
can be seen as a mediator or surrogate for B, but in this pragmatic, developmental view ‘mediation’ and
‘surrogacy’ do not entail a representational relation and do not require analysis in representational terms
conventional instruments, and go from, say, a toy guitar to a Gibson Les Paul or from
a rubber soccer ball to the Adidas Telstar 18 (the official ball of the 2018 FIFA World
Cup). These simpler, toy instruments can be said to act as mediators and surrogates for
more advanced, conventional instruments. One might wish to describe the surrogates as
representations of conventional instruments, but this is not required. In fact, if our goal
is to make sense of the expert performance of professional athletes or musicians, it is
arguably more illuminating to analyze their performance in terms of skill development
through practice (see Fig. 3).
The same view of surrogate or mediated engagement applies to skills that are more
abstract than sports and music, such as mathematics. Children usually begin to learn
mathematics by reasoning about concrete objects, for example, learning to count using
oranges and learning fractions with slices of a cake; facility with abstract, symbolic
operations can be seen as on a continuum with, and emerging from, these simpler
types of concrete engagement. And while it is always possible to interpret surrogate
reasoning in representational terms, in a case like this assuming that surrogates always
represent what they stand in for leads to the strange conclusion that concrete quantities
represent numbers rather than the other way around. That is, one could say that the
five oranges represent the number 5 or that the half cake represents the fraction 1/2,
but this would go against normal usage, according to which it is the symbols that are
abstract representations of various equivalent concrete quantities such as five oranges
and five apples or half a cake and half a pizza.
And even if some philosophers use ‘mediation’ and ‘surrogate reasoning’ in
accounts of model-based representation, these notions do not by themselves entail
a representational relation and do not require analysis in representational terms. In a
representationally-neutral fashion, ‘mediation’ and ‘surrogacy’ can help explain skill
development and learning transfer not only across contexts in everyday practices like
sports, music, and mathematics, as seen above, but also in the domain of model-based
scientific inquiry.
The scientific practice of using concrete models as ‘mediators’ or tools for ‘surro-
gate reasoning’ about concrete systems (e.g., the San Francisco Bay model) should
not be seen as distinct in kind and independent from what kids do when they play with
makeshift toys, using a cardboard box as a fort, a pen as a sword, and so on. As kids
grow up, and through formal training from grade school to grad school and beyond,
their surrogate reasoning skills gain complexity and are directed to novel domains
of practical engagement, yet the different applications are outgrowths of the same
cultural and psychological developmental context. To be sure, I am not suggesting
123
Synthese
that simulating different hydrological conditions in the San Francisco Bay model and
playing with Legos are entirely identical. Yet, the crucial difference between the two
seems to be a difference in the practical and theoretical constraints and goals involved,
rather than a difference in mechanism: depending on what one wants to accomplish
and what practice one aims to contribute to (play or science, say), different criteria
will define ‘success’ and ‘failure’ in each case, but they remain instances of mediated
or surrogate practical engagement.
This way of thinking is not limited to concrete models, but also applies to suppos-
edly “abstract” varieties such as mathematical equations and diagrams. These varieties
of modeling co-opt skills at work in mathematical reasoning (from ordinary instances
such as the ones with oranges and cakes described above) as well as in reading and
writing more generally, applying these skills to a new domain where engaging with
equations or diagrams on paper or on a computer screen mediates the development
of strategies for practical engagement with and reasoning about some other object or
entity (the “target”). Given a complex set of background skills and practices, inter-
acting with a pair of differential equations, for example, is then used as a scaffold for
thinking about various biological or physical systems, and for generating understand-
ing, forming hypotheses, guiding interventions, and so on. The point is that, whether
concrete or abstract, a model need not be analyzed in representational terms as a
description of some target: in the view sketched here, a model is instead understood as
a distinct, autonomous object that is used for developing skills that are also useful for
engaging with those other objects in specified ways (including, of course, reasoning
about those other objects).
Skill development and learning transfer in model-based science are often indirect:
typically modelers cannot intervene in the systems they model, but instead they have
to communicate their findings to others (e.g., policy makers) who are in a position to
directly manipulate the target system in question (e.g., regulating carbon emissions
to avoid negative outcomes predicted by climate models). This indirect character of
engagement and application does not contradict the account I am sketching here: on
the contrary, the widespread difficulties with translating scientific findings into public
policy can, at least in some cases, be understood as a resistance individuals have to
surrogate reasoning processes they are not themselves skillful in due to lack of relevant
training. The focus in recent years on active learning in science education points in
the same direction. Arguably, non-experts such as school-aged children and grown-up
policy-makers become capable of making epistemic use of models not by learning
how a model meets criteria that make it a representation of some target, but through
direct engagement and the development of relevant practical (motor and reasoning)
skills.
Working out further details of a non-representationalist account of modeling in
terms of skill development and learning transfer is beyond the scope of this article,
but this pragmatic and representationally-neutral rendering of ‘surrogate reasoning’
and ‘mediation’ indicates possible future directions. Focusing on scientific practice in
this way gives us a useful lens through which to think about how models can act as
mediators and surrogates for skill development and for transferring insights to novel
objects. Through modeling, scientists explore new ways of thinking and acting, and
understanding how this happens is crucial for making sense of the epistemic outcomes
123
Synthese
123
Synthese
5 Conclusion
123
Synthese
Section 4 concluded the discussion by offering a “don’t need” argument for rep-
resentationalism about scientific models: whether or not representationalism works,
analysis of models in representational terms is not required, and this for reasons having
to do with the success of scientific practice as well as with the fact that key philosoph-
ical notions are not inherently representational. To be clear, the account sketched in
Sect. 4 is representationally-neutral in that it does not affirm that models do not repre-
sent their targets nor that models will never meet formal definitions of representation:
rather, the claim is that modeling is best understood in terms of pragmatic engagement
mediated by skill development and learning transfer, and that meeting formal defini-
tions of representation is incidental to the purpose and epistemic value of modeling.
These arguments show that representationalism is both untenable and unnecessary;
together they suggest that it should be abandoned. Our representationalist intuitions
appear to work fine so long as we do not confront them directly. As soon as we try
to figure out what supports them, it becomes clear that nothing does—that is, noth-
ing other than habit. There is nothing terribly wrong with thinking about models as
representations, but that can’t tell us why models are epistemically valuable. Repre-
sentationalist approaches to model-based science have reached a dead end, and the
only reasonable way to move forward is to take one step back and change directions.
The hard but interesting philosophical question about models concerns how we can
learn through modeling: representationalism cannot answer this question, and we
don’t need representationalism to answer it, so we might as well begin working on
developing alternative ways of thinking.
Acknowledgements I have presented ideas related to this paper at various conferences over the past couple
of years, and I have benefited from questions and objections raised by more people than I can hope to name.
I am grateful for all of these interactions and recognize the crucial role they have played in helping me
develop my thinking. Very special thanks go to Angela Potochnik for her extensive and insightful comments
on multiple drafts of this paper. My research was supported by a dissertation fellowship from the Charles
Phelps Taft Research Center at the University of Cincinnati.
References
Batterman, R. W., & Rice, C. C. (2014). Minimal model explanations. Philosophy of Science, 81(3), 349–
376.
Bokulich, A. (2012). Distinguishing explanatory from nonexplanatory fictions. Philosophy of Science,
79(5), 725–737.
Bokulich, A. (2017). Models and explanation. In Springer handbook of model-based science (pp. 103–118).
Springer.
Callender, C., & Cohen, J. (2006). There is no special problem about scientific representation. Theoria
Revista de Teoría, Historia y Fundamentos de la Ciencia, 21(1), 67–85.
Chakravartty, A. (2010). Informational versus functional theories of scientific representation. Synthese,
172(2), 197–213.
Contessa, G. (2007). Scientific representation, interpretation, and surrogative reasoning. Philosophy of
Science, 74(1), 48–68.
Elgin, C. Z. (2004). True enough. Philosophical Issues, 14(1), 113–131.
Elgin, C. Z. (2017). True enough. Cambridge: MIT Press.
Frigg, R., & Nguyen. J. (2017a). Models and representation. In Springer handbook of model-based science
(pp. 49–102). Springer.
Frigg, R., & Nguyen, J. (2017b). Scientific representation is representation-as. In H. K. Chao & J. Reiss
(Eds.), Philosophy of science in practice (pp. 149–179). Berlin: Springer.
123
Synthese
Gelfert, A. (2017). The ontology of models. In Springer handbook of model-based science (pp. 5–23).
Springer.
Giere, R. (2010). An agent-based conception of models and scientific representation. Synthese, 172(2),
269–281.
Giere, R. N. (1988). Explaining science: A cognitive approach. Chicago: University of Chicago Press.
Giere, R. N. (2004). How models are used to represent reality. Philosophy of Science, 71(5), 742–752.
Giere, R. N. (2006). Scientific perspectivism. Chicago: University of Chicago Press.
Godfrey-Smith, P. (2006a). The strategy of model-based science. Biology and Philosophy, 21, 725–740.
Godfrey-Smith, P. (2006b). Theories and models in metaphysics. The Harvard Review of Philosophy, 14(1),
4–19.
Hughes, R. I. (1997). Models and representation. Philosophy of Science, 64, S325–S336.
Hutto, D., & Myin, E. (2013). Radical enactivism: Basic minds without content. Cambridge, MA: MIT
Press.
Isaac, A. M. (2013). Modeling without representation. Synthese, 190, 3611–3623.
James, W. (1907). Pragmatism, a new name for some old ways of thinking: Popular lectures on philosophy.
Harlow: Longmans, Green and Co.
Kennedy, A. G. (2012). A non representationalist view of model explanation. Studies in History and Phi-
losophy of Science, 43, 326–332.
Knuuttila, T. (2010). Not just underlying structures: Towards a semiotic approach to scientific representation
and modeling. In Bergman, M., Paavola, A.P.S., & Rydenfelt, H. (Eds.) Ideas in action: Proceedings
of the applying peirce conference (pp. 163–172).
Knuuttila, T. (2011). Modelling and representing: An artefactual approach to model-based representation.
Studies in History and Philosophy of Science Part A, 42(2), 262–271.
Lloyd, E. A. (2010). Conrmation and robustness of climate models. Philosophy of Science, 77(5), 971–984.
MacBride, F. (2016). Relations. In E. N. Zalta (Ed.), The Stanford encyclopedia of philosophy, Winter 2016
edition. Stanford: Metaphysics Research Lab, Stanford University.
Morgan, M., & Morrison, M. (1999). Models as mediators: Perspectives on natural and social science.
Ideas in context. Cambridge: Cambridge University Press.
Morrison, M. (2015). Reconstructing reality: Models, mathematics, and simulations. Oxford: Oxford Uni-
versity Press.
Myin, E., & Hutto, D. D. (2015). Rec: Just radical enough. Studies in Logic, Grammar and Rhetoric, 41(1),
61–71.
Parker, W. S. (2011). When climate models agree: The significance of robust model predictions. Philosophy
of Science, 78(4), 579–600.
Pincock, C. (2012). Mathematics and scientific representation. Oxford: Oxford University Press.
Potochnik, A. (2015). The diverse aims of science. Studies in History and Philosophy of Science Part A,
53, 71–80.
Potochnik, A. (2017). Idealization and the aims of science. Chicago: The University Chicago Press.
Suarez, M. (2003). Scientific representation: Against similarity and isomorphism. International Studies in
the Philosophy of Science, 17(3), 225–244.
Suarez, M. (2004). An inferential conception of scientific representation. Philosophy of Science, 71(5),
767–779.
van Fraassen, B. C. (1980). The scientific image. Oxford: Clarendon Press.
van Fraassen, B. C. (2008). Scientific representation: Paradoxes of perspective. Oxford: Oxford University
Press.
Weisberg, M. (2012). Simulation and similarity: Using models to understand the world. Oxford: Oxford
University Press.
Wimsatt, W. C. (1987). False models as means to truer theories. In N. Nitecki & A. Hoffman (Eds.), Neutral
models in biology (pp. 23–55). Oxford: Oxford University Press.
123