Professional Documents
Culture Documents
net/publication/306012498
CITATIONS READS
38 362
2 authors, including:
Xiang Wan
The Ohio State University
24 PUBLICATIONS 451 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Xiang Wan on 05 November 2017.
by
Philip T. Evers
(Corresponding Author)
Associate Professor
University of Maryland
Robert H. Smith School of Business
Logistics, Business and Public Policy
College Park, MD 20742-1815
Phone: 301-405-7164
Email: pevers@rhsmith.umd.edu
Xiang Wan
Assistant Professor
Marquette University
College of Business Administration
Milwaukee, WI 53201-1881
Phone: 414-288-3095
Email: xiang.wan@marquette.edu
Systems Analysis Using Simulation
Abstract
Page 1 of 30
INTRODUCTION
The purpose of research is to address significant and substantive questions. The wealth of
existing research methods represents the means available for researchers who attempt to
methodology must be employed. Use of the wrong tool can lead either to an unsatisfactory
The focus of this paper is on one particular research methodology: simulation. This
method will be discussed within a broader framework of systems analysis and specifically as it
relates to supply chain management. The power, as well as the limitations, of simulation will
be presented along with a description of how a simulation study is performed. Issues arising
during a study will also be discussed as will the future of simulation as a methodology. All of
SYSTEMS ANALYSIS
Whether the topic is medicine, sociology, mathematics, agriculture, history, or any other
discipline, the question being posed should always be the driver of the research and
consequently the methodology, not vice versa. In many of these same disciplines, a primary
topic of interest is the understanding of how entities (these could be organizations, individuals,
machines, organisms, etc.) relate to and interact with one another—in other words, the analysis
of systems. Systems analysis is especially integral to business research, in general, and supply
chain management research, in particular, as firms attempt to coordinate their activities such
that customer demands are met in both an efficient and an effective manner.
Page 2 of 30
Systems analysis within supply chain management can be found in a multitude of areas,
including inventory, transportation, and warehousing systems, to name a few (Carter and
Jennings 2004; Evers 1996; Waller et al. 2008; Waller et al. 2010). In the case of inventory
management, firms in a supply chain attempt to size their inventories at levels that will satisfy
customer demand at the lowest cost. However, if each organization sets its inventory level
independently of one another, especially when there is a lack of information transfer, the result
is the well-known bullwhip effect—large fluctuations in inventory levels and order quantities
across the supply chain stemming from small variations in customer demands (Lee et al. 1997;
Wan and Evers 2011). Likewise, stock-keeping locations at the same echelon of a supply chain
will need to carry more inventory to maintain the same level of customer service if they operate
independently of one another than if they work together—a result of the portfolio effect (Zinn
et al. 1989; Evers 1996). These two examples highlight the importance of systems analysis as
the nature of inventory management within a supply chain is such that individual stock-keeping
With regard to transportation management, the movement of goods from one location to
another is quite obviously another instance of systems analysis (Stank and Roath 1998;
Daugherty et al. 2001). From the user’s perspective, systems-related questions can range from
spatial concerns, such as the allocation of product to various locations, to temporal concerns,
such as the determination of shipment size on a specific route. From the carrier’s perspective,
their operating networks are, by their very nature, systems, and any attempt to examine only a
portion of their activities could very well lead to sub-optimization of the larger structure.
Even within a facility, systems analysis is important. For example, the activities within a
Page 3 of 30
warehouse can be viewed as a system of handling and storage processes. A determination of
the degree of product variety most appropriate for a particular distribution center cannot be
made without considering the interactions of demand forecasting, unit price, service level,
sales quantity, and so on (Wan et al. 2011). Of course, similar arguments could be made for
other warehousing questions such as a comparison of fixed slotting (where stock-keeping units
are assigned predetermined storage locations) versus variable slotting (where stock-keeping
In general then, viewing supply chains as a system of links (movement activities) and
nodes (storage and/or processing activities) highlights the underlying structure on which the
supply chain operates (Coyle et al. 2008). Consequently, research that investigates operating
relationships amongst supply chain members must recognize the importance of taking a
system-wide perspective. Having established this connection between supply chains and
A system can be fully described by its elements. These elements include the structure, the
variables or attributes, and the state. The structure of the system is the underlying relationship
among entities. Starting with the links and nodes, the system structure for a transportation
carrier network typically includes the origin and destination points, as well as the terminal
locations. Other resources could also be considered a part of the structure; in the carrier
network case, an important resource is the number of vehicles. Of course, changes typically
occur within a system even if its underlying structure remains constant—measures of these
system changes are variables. For example, the transportation carrier may see shipment
Page 4 of 30
volumes fluctuate over time within its network, and this volume could represent a variable of
interest. Moreover, each shipment may have particular characteristics such as size,
origin-destination points, handling requirements, etc. Associated with specific entities in the
system, these individual differences are termed attributes. The particular values that the
variables and attributes take on at any given time within the system represent its state. Thus,
the state of the carrier’s system on day 123 at time 14:00 is 400 shipments on 200 vehicles
How the state changes is very important in terms of analyzing the system. If the changes in
a system are distinct and discrete, this is referred to as a discrete-event system. On the other
hand, if changes within a system occur such that they cannot be precisely attributed to a distinct
pipeline operations typically fall into this category. Various products may constantly flow
through the pipeline in varying volumes, but the precise transition from one product to another
or from one volume level to another may not be clearly identified. Generally speaking, the
discrete-event system applies, or at least suffices, in most supply chains; the continuous-event
system is more suitable for flow processes (things like pipelines, refineries and chemical plants,
With the state and its manner of change defined, methods for analyzing the system itself
can now be enumerated. Law and Kelton (2000) provide a concise hierarchy of methods to
analyze systems. As shown in Figure 1, this paper expands their categorization to include an
even wider range of methods. At the highest level of systems analysis, the first distinction is on
Page 5 of 30
the basis of systems control. Many research methodologies are available that allow the analyst
to investigate a system without requiring control over it. Perceptual analyses, such as the use of
surveys and panel discussions, and factual analyses, such as the use of archival data and
objective observations, afford the opportunity to understand the system without actually
affecting it. In essence, the analyst can act as an outside observer and glean pertinent
information that can be examined using such tools as econometric models and structural
equations to understand the system. While there is an obvious advantage to these approaches
in that the actual system is being directly (in the case of factual analysis) or indirectly (in the
case of perceptual analysis) observed, there is a major disadvantage with these approaches, too.
Since the analyst has no control over the system, it can be difficult to separate out the effects of
confounding exogenous factors and to perform sensitivity analyses when relevant changes in
----------------------------------------------------
----------------------------------------------------
While methods not requiring control abound, there are also methods that allow the analyst
to be more directly involved with the system. These techniques afford the analyst the ability to
actually dictate system operations and observe their effects. They can be broken down into
three categories (though, as will be discussed in short order, only two are shown in Figure 1):
Physical analysis involves manipulating either the actual system or a scale replica of it.
Consider first the manipulation of an actual system: say, an existing warehouse which employs
Page 6 of 30
a set of docks used for both loading and unloading. Operations could be altered such that some
of the docks are used only for loading and others are used only for unloading. A comparison of
the two could then take place. While a big advantage of this approach is that effects on actual
operations can be witnessed firsthand, there are a couple of equally big disadvantages. First,
experimenting with the actual system could result in enormous costs and cause irreparable
damage—changing dock assignments might be extremely costly and create disruptions that
adversely affect customer service. Second, while the effects on actual operations can be seen,
they may still be confounded with other factors—the change in dock assignments might occur
just as an increase in demand occurs and so it may be hard to unambiguously attribute adverse
Consider now the manipulation of a scale replica: a small version of the warehouse could
be built and different layouts tried in an attempt to analyze the impact of changing dock
assignments. The difficulty here lies in making the scale replica actually operate and reflect the
intricate nuances of a system. Consequently, it is not surprising that scale replicas are rarely
used to analyze supply chain systems, but they do, nevertheless, represent an option in the
appropriate situation and can be commonly found in other disciplines (for example,
engineering studies using scale models in wind tunnels and medical studies using lab mice to
performed, this category is not shown in Figure 1. On the other hand, the two
Behavioral analysis includes experiments and games. In these exercises, participants take
Page 7 of 30
on certain roles within a system. One of the best known business applications of an exercise is
the Beer Game, where individuals act as members of a beer supply chain (Forrester 1958;
Sterman 1989). For instance, Croson and Donohue (2006) employed a set of experiments with
human participants to study the behavioral causes of the bullwhip effect. Note that while the
Beer Game requires participants to follow steps specified by the analyst, they have the freedom
to make their own decisions, and so while experiments and games can be influenced by the
analyst they usually cannot be completely controlled. Moreover, participants often learn from
their involvement and respond to incentives, whether perceived or real, so they are apt to
modify their own actions over time. And in many cases, the behavioral exercise itself is
significantly removed from reality and, as a result, may lead to fundamental questions about
the research. Even with the Beer Game case, concerns may arise regarding the likelihood of
finding an organization where one person is aware of both customer demands and inventory
Offering complete control to the analyst, logical analysis involves mathematical models.
These can be either analytical or simulation models. An analytical model is one or more
equations incorporating factors solely based on the analyst’s discretion that are solved to
understand a system. For example, Lee et al (1997) analyzed a series of analytical models in
order to study the bullwhip effect in supply chains. If the model cannot be solved, as is often the
case, a heuristic result may still be available or limiting mathematical assumptions may need to
be made to allow for a solution. For instance, Evers and Beier (1993) extended the square root
law stemming from the portfolio effect by relaxing some of the restrictive assumptions. For the
warehouse example mentioned earlier, a series of equations reflecting vehicle arrivals and
Page 8 of 30
departures as well as forklift movements could be developed to reflect dock operations. If an
optimal solution can be determined, then the analytical method has great appeal. Of course,
this assumes that the underlying assumptions do not stray too far from reality; otherwise, the
heuristic solution may still be available (DeCroix and Zipkin 2005). However, a heuristic
solution relies on even more assumptions and is no longer guaranteed to find the optimal result
When all else fails, there is simulation! Simulation is also based on a series of logical steps
that may include equations to represent the actual system as accurately as possible. As all
decisions and parameters are specified by the analyst, the analyst has complete control over the
system without having to make potentially restrictive assumptions and can isolate the results by
avoiding confounding factors. In the warehouse case, a simulation model could be built to
reflect the individual warehouse operations and then run to ultimately generate inferences
regarding the system, without having to worry about exogenous influences such as electrical
outages or supplier disruptions (unless of course these things are actually of interest). Even in
the case of simulation, however, there is a downside as optimality is not guaranteed and
simulation modeling skills are necessary (Mentzer 1989)…it is not easy to build a simulation
SIMULATION
The previous discussion outlined various methods for performing systems analysis and
highlighted some of the advantages and disadvantages. Too be sure, there are many more pros
and cons than those listed above, so with regard to simulation, a more detailed discussion
Page 9 of 30
follows. A useful way of looking at simulation modeling is in terms of what it can, and cannot,
do.
First, what can simulation do? A lot! One of the biggest benefits of simulation is that it
allows the analyst to investigate complex systems without making limiting assumptions. A
simulation can be crafted to the appropriate precision to account for as many interactions as
deemed necessary. Indeed, anyone familiar with modern video games has no doubt seen this
stunning level of complexity and realism ranging from sports games to war and fantasy games
and everything in between (and beyond). Research simulations also span the gamut of
applications and the corresponding complexity continues to increase as computing power and
simulations can be run in a fraction of real time and changes to the system can be monitored
accordingly. And as alluded to before, perhaps the most powerful aspect of simulation is that it
provides the researcher with complete control of the system. No changes to the system can
occur unless and until the researcher says they can. In so doing, confounding influences can be
Other features of simulation include the ability to incorporate random occurrences into the
system to estimate their effects. Similarly, changes to the underlying simulation model or
inputs can be made to allow for sensitivity analysis. In other words, simulations can be used to
answer “what if” questions. For example, Closs et al. (2010) simulate the random daily demand
of computers and perform sensitivity analysis on the inputs to investigate the impacts of
Page 10 of 30
Now, what can’t simulation do? Everything! Though numerous methods exist to help
researchers reduce output variation in simulation results, simulations do not provide specific
answers. Simulations are stochastic processes and, as such, rely upon statistical
inferences—they do not give exact results. And though numerous methods exist to help
researchers find near optimal results, they cannot guarantee optimality (Vashi et al. 1995). If
Because simulation models are a representation of an actual system, they also cannot
ensure generalizeability. Simulation results are only as generalizeable as the system under
consideration (Meredith 1998). For example, results from a simulation model of a continuous
review inventory management system may be very transferrable from one industry to another;
however, results from a simulation model of a humanitarian aid system may not. It all depends
models are built on logic. And while they typically incorporate randomness, there is still a
systematic process to it. Human behavior, on the other hand, can be erratic and based on
factors other than logic, such as love, fear, opportunism, etc. Traits such as these can be very
difficult to define and describe in logical terms. In these situations, participant exercises may
be more appropriate.
How it works
Simulation modeling has three general aspects: the model, its inputs, and its outputs.
Regarding the model itself, this is the “black box” at the heart of the study. As noted earlier, a
simulation model is the set of logical steps that reflects the system of interest. In other words, it
Page 11 of 30
reflects the relationships between the various entities (i.e., the underlying structure of the
system).
As shown in Figure 1, there are different types of simulation models. Law and Kelton
(2000) provide a succinct classification of them based on time, randomness (or lack thereof),
and change characteristics found in the system of interest. The first dichotomy is time.
Simulations models that reflect system operations over time are referred to as dynamic models,
whereas those that do not involve time are called static models. Most supply chain
interested in how the system changes over time (Cattani et al. 2011; Klassen and Menor 2007).
For example, the analysis of continuous and periodic review systems is, by its very nature, the
analysis of inventory fluctuations over time. However, static models do exist. The analysis of
problems, which includes the very well-known newsvendor problem, time is typically of little
Deterministic simulation models have no probabilistic elements (Dhouib et al. 2008; Vashi et al.
1995). These models are often used when the underlying system is just too complex to solve.
discrete lot sizes in a materials requirements planning environment. For dependent demand
items where the requirements are known (i.e., based on parent items), the important question of
production and/or order quantities to minimize costs remains. In this situation, even though
randomness is not an issue, a simulation may still be necessary to ascertain the impact of
Page 12 of 30
different lot sizing techniques on total costs, especially in products having many levels.
Simulations that incorporate randomness are called stochastic models. Again, the underlying
stochastic process may simply be too complex to solve. Independent demand items typically
fall into this category as they are difficult, if not impossible, to forecast with complete accuracy.
(To be sure, many other supply chain systems also fall in this category as supply chains tend to
The final distinction was mentioned earlier and reflects how system changes occur: either
inventory levels are checked after specific intervals, is an example of a discrete system; a
continuous review system, where inventory levels are constantly monitored, is an example of a
discrete system.
While the model itself represents the foundation and is typically the first thing that comes
to mind when referring to simulation, it is by no means the only thing. Viewing simulation
analysis in terms of a process, it is clear that the inputs and outputs are equally important. The
inputs represent the parameters that the simulation model will assume. The inputs also
influence the relationships between entities as the model is run. For example, in a simulation of
materials handling systems within a cross-docking facility, while the interactions between the
conveyors, forklifts, and other resources are represented by the simulation model itself, the
influences on this underlying structure come from the model inputs: factors such as the precise
number of forklifts, the speed of the conveyors, the hours of operations, and the probability of
breakdowns. Inputs drive the simulation and can be considered the independent variables in
the process.
Page 13 of 30
At the other end of the process are the outputs. Simulation outputs are the genesis of the
results that must first be appropriately extracted and analyzed. In the materials handling
example, outputs might include item throughput times, forklift movement distances, and
capacity utilization. Outputs can be considered the dependent variables in the process.
A SIMULATION STUDY
Even though simulation may be a less costly way of conducting research when compared
with many other methodologies (Bowersox and Closs 1989), simulation still has a cost. Aside
from the simplest of systems, computer software is needed to construct the simulation model.
Simulations can be developed in a wide range of software applications ranging from common
spreadsheet packages (with add-ons if desired) and standard programming languages (such as
Visual Basic or C++) to simulation-specific packages (such as Arena and Simul8). While
spreadsheets and programming languages are widely available and easy to understand, the
simulation-specific packages are becoming very easy to work with as well. Of course, these
latter packages tend to be more expensive to acquire, but their simulation capabilities are
tremendous.
Beyond the expense of software, a related cost is the heretofore mentioned learning cost.
To some extent, the cost of learning how to build and run simulation models using any
particular package is a one-time upfront cost. However, like any computer software, continual
updates of simulation packages require continual learning. For example, many of those who
learned simulation using the SIMAN language have transitioned over time to the Arena system
Page 14 of 30
In addition to the expenses directly associated with the software, other simulation costs
involve data collection and output analysis. These costs are not unlike those associated with
other standard data collection and analysis techniques. The cost of time is also a factor. While
researchers are able to speed up time through simulation (as opposed to, say, survey research
where significant time elapses between the sending of questionnaires and the receiving of
responses), a substantial amount of time is required to develop the model. Thus, a discussion
Modeling issues
It is important to recognize upfront that modeling issues are more than just computer
coding. Indeed, the first modeling issue arises long before the task of coding is to commence.
The system of interest must first be framed and its boundaries established.
Any arbitrary system is a component of a larger system and, at the same time, encompasses
many smaller systems. Hence, accurately scoping the system is critical. This can be done by
referring to the research question—what exactly is being asked? If the research problem deals
with on-line retailers and the fulfillment of customer orders at a distribution center, then the
system should be confined to those activities (Yan et al. 2010). In this case, only supplier
shipment receipts are of interest—their manufacturing operations are not. Likewise, only
In a similar vein, the level of system detail should also be defined at the appropriate level
order arrival processes, and demand fulfillment procedures are all fundamental to analysis.
The routing of order pickers within the facility may not be very important. As a consequence,
Page 15 of 30
the system should be defined to the level necessary; anything more is overkill and results in
needless and time-consuming effort in terms of such tasks as understanding and coding the
process, data collection and analysis, and simulation run times. One of the most important tips
The next step is to map out the process. Drawing a flowchart of the system is extremely
useful in understanding how the different entities and activities relate and examples abound
(Closs et al. 2010; Evers 1994). This diagram can help to identify where gaps in the knowledge
of the process lie. In addition, it can help to determine whether the system boundaries have
been established correctly and whether its level of complexity has been set properly. Clearly,
these first two steps—framing and flowcharting—are typically done in an iterative fashion.
Once the system has been suitably framed and appropriately defined, and the process
mapping has been completed, attention can be turned to the inputs and outputs. What
parameters will be varied in the simulation and what data will be needed for input purposes?
On the output side, what variables will be collected for further analysis? For example, in order
to investigate the impacts of information sharing in supply chains, Angulo et al. (2004) used
stochastic demand and lead time values as input and measured average inventory levels, fill
rates, and inventory costs as output. A well-described flowchart helps tremendously with this
endeavor.
After all of this upfront work has taken place, simulation coding can begin. For years,
those unfamiliar with simulation often viewed this activity as the be-all and end-all of
simulation. The preceding discussion on inputs and outputs should put this notion to rest.
Moreover, coding was viewed as the primary task of simulation research, and it was considered
Page 16 of 30
the major hurdle because it required programming knowledge. The new wave of simulation
packages with their practically non-existent programming requirements puts to rest this latter
concern. To be sure, coding is extraordinarily important, but it can be done. A solid flowchart
Once the simulation model is in place, the next step is arguably the most important:
verification and validation. Verifying a simulation model means ensuring that the simulation
logic works as desired. Manually walking through the model is a brute force method of doing
this. Test running the simulation with no probabilistic elements is another way. In both cases,
the researcher should be able to identify any errors in coding or logic. In the distribution center
order fulfillment example, running the simulation without randomness is a straightforward way
of making sure that orders are properly made when reorder points are reached and that lead
times are working correctly, too. Validating a simulation model means ensuring that the
simulation logic reflects the system of interest. Beyond mere face validity which might be
comparison with actual system data can help to establish content validity. For example, Torres
and Maltz (2010) built a simulation model using the supply chain characteristics reported by an
actual firm and validated the simulation results by comparing them with actual financial data
provided by the company. If verification and validation cannot be corroborated, then the
simulation is nothing more than some arbitrary and meaningless set of procedures. On the
other hand, if they can be adequately established, the simulation process is finally about to take
place.
Input issues
Page 17 of 30
The simulation model is set to go, but what will it run? The inputs still need to be
addressed. This is an extremely important step; if not carefully carried out, the research runs
One issue is data collection. Since the simulation is intended to replicate a system, it is
imperative to collect data on that process, if possible. Standard data collection methods can be
used to do this, ranging from the use of historical data to time and motion studies and the like.
If the simulation deals with shipping dock utilization, for example, data on vehicle inter-arrival
times, loading times, etc. will need to be collected, some of which might come from company
Once the data has been pulled together, it needs to be incorporated into the simulation.
Some of the data might be used to set system constraints. This typically does not require much
effort as system resources are rather straightforward to identify. For example, in the dock
utilization case, the number of docks at the facility may represent a constraint. Indeed, it may
either be set at a fixed level if the goal is to examine utilization within the current system or be
allowed to vary if the goal is to determine the impact of changes to the facility itself. In either
case though, the determination of the number of docks is fairly simple as system constraints are
usually deterministic.
The more challenging task involves stochastic inputs. In most simulations, much of the
data used as input will be probabilistic in nature. In the dock utilization example, the vehicle
interarrival times will likely be random, the amount of product to be shipped will be as well. In
this case, the data will need to be appropriately converted. There are three common methods of
using data: theoretical distributions, empirical distributions, and the data themselves.
Page 18 of 30
A theoretical distribution involves fitting a known distribution to the data and then
sampling from this fitted distribution as input to the simulation model. Take the vehicle
interarrival times mentioned above: when aggregated, this data might have the general shape of
an exponential distribution. The amount of product to be shipped may have a very different
shape, perhaps looking a bit like a normal distribution; however, since the normal distribution
can take on negative values, and the amount of product shipped must be at least zero (ignore the
possibility of returns or, better yet, consider the returns process separately), a different
distribution such as the lognormal might be better suited. The point is that the researcher may
fully defined by one parameter, its mean (the inverse of its rate), while a normal distribution is
defined by two, its mean and standard deviation. As part of the distribution fitting process,
both the theoretical distribution and its parameters must be set. Suffice it to say that various
statistical procedures exist to help ensure the goodness-of-fit between the aggregated data and
the theoretical distribution. Indeed, many software packages, including the major
simulation-specific packages, offer comprehensive data fitting abilities involving a wide range
Try as one might, sometimes no theoretical distribution will work. The aggregate data is
so oddly shaped that nothing fits. In cases like these, an empirical distribution is necessary. An
empirical distribution is merely a distribution that is constructed directly from the aggregate
data. For example, if 45% of the interarrival times are 10-20 minutes, 10% are 20-30 minutes,
and 45% are 30-40 minutes, no theoretical distribution may adequately fit; an empirical
Page 19 of 30
distribution taking on these proportions can be used instead.
Comparing the use of theoretical and empirical distributions, there are pros and cons to
each. Regarding theoretical distributions, there is a vast array of distributions out there to
choose from (anywhere from well-known continuous ones like exponential, normal, and
uniform, to lesser known ones like Pearson Type V, Dagum, and Rayleigh, plus all of the
discrete distributions [Poisson, negative binomial, etc.]), though they do not cover all possible
shapes and situations. A good example of this latter possibility is found in transit times: most
transit times between two locations are typically within a close window. If there are a
significant number of delays, however, the distribution tail could be long and even have a
sizeable bump in it. Unfortunately, theoretical distributions do not usually look like this. This
big drawback for theoretical distributions is a big advantage for empirical distributions. On the
contrary, the big disadvantage of an empirical distribution is that it is entirely data driven and
dependent upon the sample collected. Thus, while it might reflect the data collected, it might
not accurately reflect the underlying process. A fitted theoretical distribution, on the other
hand, relies upon statistics to recognize that the distribution is an adequate fit, not an exact fit.
In general, a rule of thumb is to rely upon theoretical distributions wherever possible, but to use
The third method is to directly use the data values themselves. Interarrival times
throughout the day can be directly input into the simulation in the same sequence that they were
collected. Of course, this removes the randomness from the input process, but this approach
can be very useful for verification and validation purposes. As the process has been observed
and the data collected, plugging the data directly into the simulation should lead to outputs
Page 20 of 30
identical to those observed in the actual process. If not, then either the model coding was not
performed correctly (verification is lacking) or the model mapping was not done correctly
(validation is lacking).
In some instances, little data is available. Even in these situations, input distributions can
be created. For example, if a range can be identified (i.e, minimum and maximum values), then
a uniform distribution can be used to provide at least some randomness. If the most frequently
occurring value can also be identified (i.e., the mode) along with the range, then a triangular
distribution can be used. A beta distribution is another useful option when data is lacking; only
the minimum and maximum values, along with the mode and the mean are needed. The point
is that even if only a modicum of data exists simulation can still be employed. In fact, in many
cases no data exists whatsoever and yet simulation remains the preferred method. In order to
provide a comprehensive data input, factorial experimental designs are generally employed to
generate input data sets (Cattani et al. 2011; Pagell and Melnyk 2004; Nachtmann et al. 2010;
Yang et al. 2010). Since simulation can be used to answer “what-if” questions, it is often the
case that “what-if” involves the consideration of different, unobserved inputs. The lack of data
does not mean that simulation cannot be used; it really boils down to whether simulation
Output issues
Now that the simulation model is built and the input parameters have been set, the
simulation can be run and, as a consequence, the output generated. However, the work is not
complete; indeed, it is just beginning. All of the tasks performed up to this point are
meaningless if the output is not properly analyzed. The key to output analysis is the
Page 21 of 30
recognition that a single simulation run is merely one realization of the underlying process. In
as much as one could not claim that the transit time between two locations can be determined
by only making one actual shipment (there could be weather, congestion, etc.), one also cannot
claim that the transit time can be determined by only making one simulated shipment.
for randomness. This is typically performed by running the same simulation multiple times
using different random realizations of the input parameters and then making statistical
inferences based on the results. Note that this is no different than collecting actual data based
on multiple observations and then making statistical inferences based on them. Simulation is
The output issues that arise with simulation generally revolve around how the simulation
model is run. Since a simulation model is usually stochastic and dynamic, the state of the
system randomly varies over time. Therefore, it is often the case that data collected from a
simulation run will not be independent of time, so standard statistical techniques that assume
independence are not appropriate. In an inventory system for example, the goal may be to
identify the average on-hand inventory level of a particular reorder policy, but the inventory
level at day 2 is related to the level at day 1, making the output analysis more difficult to
perform. There are three standard approaches to dealing with this problem.
The first approach is to make multiple runs of the simulation using the same input
parameters (Ross et al. 2009; Wan and Evers 2011). If each run is independent of one another
(i.e., the random realizations of the inputs within the simulation are unrelated across runs), then
the observations across the runs are independent as well and standard statistical methods that
Page 22 of 30
assume independence can be used. This is referred to as the replication approach. There are
two main factors that must be accounted for: 1) the number of replications, and 2) the length of
each replication.
With regard to the number of replications, this is based on standard statistics: the more
replications, the more precision in terms of statistical inferences. With regard to the length of
each replication, the primary concern revolves around the simulation starting point. Note that
the starting point for a simulation is often quite arbitrary. For example, in an inventory system,
the simulation could start with the on-hand inventory level at zero, at the reorder point, at the
reorder point plus order quantity, at two times the order quantity, or at any other random point.
As the process evolves over time, the impact of the starting point will fade away. Once this
opening period has passed, the process has reached a level of stability termed steady-state. The
data in this opening period is commonly referred to as initial transient data and is discarded.
All data generated after this can be collected as output data and analyzed accordingly. Thus,
the primary concern with regard to the length of each replication is the elimination of the
The second approach is to make one long run in steady-state and then, after deleting the
initial transient data, average adjacent outputs until any autocorrelation has been eliminated
(Feinberg et al. 1992; Rabinovich and Evers 2003). For example, if on-hand inventory levels
are recorded at the end of each day, every two days of data are combined and their means
calculated. If correlation still exists, then every three days of data are combined and their
means calculated. If correlation still exists, then every four days of data are combined and so
on. At some point, correlation will be mitigated and the means can be analyzed using standard
Page 23 of 30
statistical methods. The combining of data is referred to as batching and this technique is
called batch means. When using this method, a couple of issues arise: 1) the number of
observations per batch, 2) the length of the run, and 3) the smoothing of variation. While the
first two issues have already been alluded to, the third one bears further mention. One
drawback of batch means is that taking an average of individual observations dampens the
variability witnessed within the system; if the concern is understanding the underlying
variability of the process, batch means is usually not a preferred method (on the other hand, the
After either of the first two approaches is performed, standard statistical techniques that
assume independence can be applied and inferences made. The inferences made, or statistical
tests employed, should address the questions posed, or initial hypotheses generated, at the
beginning of the research project and thus complete the analysis. Of course, sensitivity
analysis may be undertaken as a result of the findings from the initial output analysis, but this
The third approach is to simply analyze the simulation output of a long steady-state run
using appropriate methods after deleting the initial transient data. By doing so, the hypotheses
can be directly tested without performing batch means or making multiple replications.
Though the output might be highly correlated, advanced econometric techniques exist to
account for autocorrelation. Perhaps the biggest concern with this approach is that many
researchers may not be familiar with using statistical techniques that do not assume
independence.
Page 24 of 30
Practically everything that has been presented up to this point regarding simulation has
been said before (Law and Kelton 2001). The fundamentals have been around for many years
and those who conduct research with the assistance of simulation know it. What has changed
simulation-specific software packages now exist that make modeling very easy. Consequently,
In the past, a researcher needed programming ability (and desire) to build a simulation
model—not anymore! Drop-and-drag packages like Arena and Simio have completely
changed the way models are developed, and a few hours spent learning the tool is all that is
processors are readily available with most of these packages to assist in finding input
distributions and analyzing output data. In some ways, simulation has become more of a black
box then every before as the user does not need to know all of the intricate details. And therein
While the technical tasks of building and running the model have been significantly
simplified, the underlying cautions remain. Those using simulation must recognize, for
example, that the output is not the answer—it must be appropriately analyzed—or that
simulation is not the right tool for every situation—it must make sense to employ it. As
simulation becomes easier to do, it also becomes easier to do incorrectly. This admonition is no
different than for those employing high level statistical analyses software packages. It is easy
to run almost any arbitrary statistical model on a set of data, it is also easy to violate the
Page 25 of 30
Assuming that the simulation is carried out correctly, the prospects are endless. For
example, advances are being made to help researchers identify near optimal solutions using
simulation (Alrefaei and Andradottir 1999). Likewise, the generation of correlated input data
is now easier then every before (Cezik and L’Ecuyer 2008). For practitioners in particular, the
software packages allow the user to represent the system with realistic-looking pictures instead
of sleepy-looking diagrams. This capability makes presentations even more powerful when
selling the results of the simulation analysis. Moreover, as computer processing power
continues to improve, the size and complexity of models that the simulation-specific packages
can handle will continue to grow. And as previously indicated, the ease of use of these
methods for systems analysis. These various methods, however, can be quite complimentary
and they can be used in conjunction with one another to tackle a problem from several
perspectives. Indeed, it is often the case that analytical solutions are found and then
simulations run to cross-validate the results. There is no reason to expect that simulation
modeling could not be combined with other forms of analysis (say, survey or archival data
approach is certainly something to be encouraged in the future as all of the various systems
The purpose of research is to address significant and substantive questions. This paper
presented the details of one such methodology to assist researchers in their attempt to address
Page 26 of 30
such questions. Simulation was discussed as one method for addressing questions related to
systems analysis, specifically supply chain management in business research. It was suggested
that simulation is a very useful tool and is not that hard to use anymore. However, it was also
noted that with this newfound ease it can also be easy to take shortcuts that lead to a
fundamentally flawed analysis. Ultimately, like any research method, simulation requires a
great deal of diligence to ensure that it is done correctly. If this effort is taken seriously, the
extent of the research contribution is limited only by the extent of the research question.
Page 27 of 30
REFERENCES
Alrefaei, M. H., and Andradottir, S. 1999. “A Simulated Annealing Algorithm with Constant
Temperature for Discrete Stochastic Optimization.” Management Science 45(5):748–64.
Angulo, A., Nachtmann, H., and Waller, M. A. 2004. “Supply Chain Information Sharing in A
Vendor Managed Inventory Partnership.” Journal of Business Logistics 25(1):101–20.
Bowersox, D. J., and Closs, D. J. 1989. “Simulation in Logistics: A Review of Present Practice
and A Look to the Future.” Journal of Business Logistics 10(1):133–48.
Carter, C. R., and Jennings, M. M. 2004. “The Role of Purchasing in Corporate Social
Reponsibility: A Structural Equation Analysis.” Journal of Business Logistics 25(1):145–
86.
Cattani, K. D., Jacob, F. R., and Schoenfelder, J. 2011. “Common Inventory Modeling
Assumptions that Fall Short: Arborescent Networks, Poisson Demand, and Single-Echelon
Approximations.” Journal of Operations Management 29(5):488–99.
Cezik, M. T., and L’Ecuyer, P. 2008. “Staffing Multiskill Call Centers via Linear Programming
and Simulation.” Management Science 54(2):310–23.
Closs, D. J., Jacobs, M. A., Swink, M., and Webb, G. S. 2008. “Toward A Theory of
Competencies for The Management of Product Complexity: Six Case Studies.” Journal of
Operations Management 26(5):590–610.
Coyle, J. J., Langley, C. J., Gibson, B. J., Novack, R. A., and Bardi, E. J. 2008. Supply Chain
Management: A Logistics Perspective. 8th ed., South-Western.
Croson, R., and Donohue, K. 2006. “Behavioral Causes of the Bullwhip Effect and the
Observed Value of Inventory Information.” Management Science 52(3):323–36.
Daugherty, P. J., Autry, C. W., and Ellinger, A. E. 2001. “Reverse Logistics: The Relationship
Between Resource Commitment and Program Performance.” Journal of Business Logistics
22(1):107–23.
DeCroix, G. A., and Zipkin, P. H. 2005. “Inventory Management for An Assembly System
with Product or Component Returns.” Management Science 51(8):1250–65.
Dhouib, K., Gharbi, A., and Ayed, S. 2008. “Availability and Throughput of Unreliable,
Unbuffered Production Lines with Non-Homogeneous Deterministic Processing Times.”
Page 28 of 30
International Journal of Production Research 46(20):5651–77.
Evers, P. T., and Beier, F.J. 1993. “The Portfolio Effect and Multiple Consolidation Points: A
Critical Assessment of the Square Root Law.” Journal of Business Logistics 14(2):109–25.
Feinberg, F. M., Kahn, B. E., and McAlister, L. 1992. “Market Share Response When
Consumers Seek Variety.” Journal of Marketing Research 29(2):227–37.
Forrester, J.W. 1958. “Industrial Dynamics: A Major Breakthrough for Decision Makers.”
Harvard Business Review 36(4):37–66.
Klassen, R. D., and Menor, L. J. 2007. “The Process Management Triangle: An Empirical
Investigation of Process Trade-Offs.” Journal of Operations Management 25(5):1015–34.
Law, A.M., and Kelton, W. D. 2000. Simulation Modeling and Analysis. 3rd ed., McGraw-Hill.
Lee, H.L., Padmanabhan, V., and Whang, S. 1997. “Information Distortion in a Supply Chain:
The Bullwhip Effect.” Management Science 43(4):546–58.
Meredith, J. 1998. “Building Operations Management Theory Through Case and Field
Research.” Journal of Operations Management 16(4):441–54.
Nachtmann, H., Waller, M. A., and Rieske, D. W. 2010. “The Impact of Point-Of-Sale Data
Inaccuracy and Inventory Record Data Errors.” Journal of Business Logistics 31(1):149–
58.
Pagell, M., and Melnyk, S. A. 2004. “Assessing the Impact of Alternative Manufacturing
Layouts in A Service Setting.” Journal of Operations Management 22(4):413–29.
Rabinovich, E., and Evers, P. 2003. “Product Fulfillment in Supply Chains Supporting
Internet-Retailing Operations.” Journal of Business Logistics 24(2):205–236.
Ross, A. D., Twede, D., Clarke, R. H., and Ryan, M. 2009. “A Framework for Developing
Implementation Strategies for A Radio Frequency Identification (RFID) System in A
Distribution Center Environment.” Journal of Business Logistics 30(1):157–83.
Page 29 of 30
Stank, T. P., and Roath, A. S. 1998. “Some Propositions on Intermodal Transportation and
Logistics Facility Development: Shipper’ Perspective.” Transportation Journal 37(3):13–
24.
Torres, O. C., and Maltz, A. B. 2010. “Understanding the Financial Consequences of the
Bullwhip Effect in A Multi-Echelon Supply Chain.” Journal of Business Logistics 31(1): 23
–41.
Vashi, V. H., Bienstock, C. C., and Mentzer, J. T. 1995. “The Use of Response Surface
Methodology to Optimize Logistics Simulation Models.” Journal of Business Logistics
16(2): 197–225.
Waller, M. A., William, B. D., Tangari, A., and Burton, S. 2010. “Marketing at the retail shelf:
An Examination of Moderating Effects of Logistics on SKU Market Share.” Journal of the
Academy of Marketing Science 38(1): 105–17.
Waller, M. A., William, B. D., and Eroglu, C. 2008. “Hidden Effects of Variable Order Review
Intervals in Inventory Control.” International Journal of Physical Distribution & Logistics
Management 38(3): 244–58.
Wan, X., and Evers, P. T. 2011. “Supply Chain Networks with Multiple Retailers: A Test of the
Emerging Theory on Inventories, Stockouts, and Bullships.” Journal of Business Logistics
32(1): 27–39.
Wan, X., Evers, P. T., and Dresner, M. E. forthcoming. “Too Much of A Good Thing: The
Impact of Product Variety on Operations and Sales Performance.” Journal of Operations
Management.
Yan, T., Rabinovich, E., Dooley, K., and Evers, P. T. 2010 “Managing Backlog Variation in
Order Fulfillment: The Case of Internet Retailers.” International Journal of Production
Economics 128 (1): 261 - 68.
Yang, K. K., Balakrishnan, J., and Cheng, C. H. 2010. “An Analysis of Factors Affecting Cross
Docking Operations.” Journal of Business Logistics 31(1):121–48.
Zinn, W., Levy, M., and Bowersox, D. J. 1989. “Measuring the Effect of Inventory
Centralization/Decentralization on Aggregate Safety Stock: The 'Square Root Law'
Revisited.” Journal of Business Logistics 10 (1): 1-14.
Page 30 of 30
FIGURES
System Analysis
Analyst Analyst
Uncontrolled Controlled
Note: The logical analysis branch is based on Law and Kelton (2000).
SHORT BIOGRAPHIES