Professional Documents
Culture Documents
This module is an introduction to the field of statistics and how engineers use
statistical methodology as part of the engineering problem-solving process. Statistics is a
science that helps us make decisions and draw conclusions in the presence of
variability.
Learning Objectives:
After studying this module the student should
Identify the role that statistics can play in the engineering problem-solving
process
Discuss how variability affects the data collected and used for making
engineering decisions
Explain the difference between enumerative and analytical studies
Discuss the different methods that engineers use to collect data
Identify the advantages that designed experiments have in comparison to other
methods of collecting engineering data
Explain the differences between mechanistic models and empirical models
Discuss how probability and probability models are used in engineering and
science.
The field of statistics deals with the collection, presentation, analysis, and use of
data to make
decisions, solve problems, and design products and processes. In simple terms, statistics
is the science
of data. Because many aspects of engineering practice involve working with data,
obviously knowledge of statistics is just as important to an engineer as are the other
engineering sciences. Specifically, statistical techniques can be powerful aids in
designing new products and systems, improving existing designs, and designing,
developing, and improving production processes.
Variability
Example 1 gasoline mileage performance of your car. Do you always get exactly the same
Consider the
mileage performance on every tank of fuel?
Answer:
Of course not — in fact, sometimes the mileage performance
varies considerably. This observed variability in gasoline mileage
depends on many factors, such as:
the type of driving that has occurred most recently (city versus highway),
the changes in the vehicle’s condition over time (which could include factors such as
tire inflation, engine compression, or valve wear),
the brand and/or octane number of the gasoline used, or possibly even
the weather conditions that have been recently experienced.
These factors represent potential sources of variability in the system. Statistics provides a
framework for describing this variability and for learning about which potential sources of
variability are the most important or which have the greatest impact on the gasoline mileage
performance.
Example 2
Another example of variability, but this time is use in dealing engineering problems.
Eight prototype units are produced and their pull-off forces measured, resulting in the following
data (in pounds):
1) 12 .6
2) 12.9
3) 13.4
4) 12.3
5) 13.6
6) 13.5
7) 12.6
8) 13.1
As we anticipated, not all of the prototypes have the same pull-off force. We say that there is
variability in the pull-off force measurements. Because the pull-off force measurements exhibit
variability, we consider the pull-off force to be a random variable. A convenient way to think of
a random variable, say X, that represents a measurement is by using the model
X= μ+ ∈
where: μ = a constant
∈ = random disturbance
The constant remains the same with every measurement, but small changes in the
environment, variance in test equipment, differences in the individual parts themselves, and so
forth change the value of e. If there were no disturbances, ∈ would always equal zero and X
would always be equal to the constant μ. However, this never happens in the real world, so the
actual measurements X exhibit variability. We often need to describe, quantify, and ultimately
reduce variability.
Figure 1-2 presents a dot diagram of these data. The dot diagram is a very useful plot for
displaying a small body of data—say, up to about 20 observations. This plot allows us to easily
see two features of the data: the location, or the middle, and the scatter or variability. When
the number of observations is small, it is usually difficult to identify any specific patterns in the
variability, although the dot diagram is a convenient way to see any unusual data features.
The need for statistical thinking arises often in the solution of engineering
problems. Consider the engineer designing the connector. From testing the
prototypes, he knows that the average pull-off force is 13.0 pounds. However, he thinks
that this may be too low for the intended application, so he decides to consider an
alternative design with a thicker wall, 1/8 inch in thickness. Eight prototypes of this
design are built, and the observed pull-off force measurements are
The average is 13.4. Results for both samples are plotted as dot diagrams in Fig.
1-3. This display gives the impression that increasing the wall thickness has led to an
increase in pull-off force.
Often, physical laws (such as Ohm’s law and the ideal gas law) are applied to help
design products and processes. We are familiar with this reasoning from general laws to
specific cases. But it is also important to reason from a specific set of measurements to
more general cases to answer the previous questions. This reasoning comes from a
sample (such as the eight connectors) to a population (such as the connectors that will
be in the products that are sold to customers). The reasoning is referred to as statistical
inference.
See Fig. 1-4. Historically, measurements were obtained from a sample of people and
generalized to a population, and the terminology has remained. Clearly, reasoning
based on measurements from some objects to measurements on all objects can result
in errors (called sampling errors). However, if the sample is selected properly, these risks
can be quantified and an appropriate sample size can be determined.
The importance of proper sampling revolves around the degree of confidence with
which the analyst is able to answer the questions being asked. Let us assume that only a
single population exists in the problem. Simple random sampling implies that any
particular sample of a specified sample size has the same chance of being selected as
any other sample of the same size. The term sample size simply means the number of
elements in the sample. Obviously, a table of random numbers can be utilized in
sample selection in many instances. The virtue of simple random sampling is that it aids
in the elimination of the problem of having the sample reflect a different (possibly more
confined) population than the one about which inferences need to be made.
Example 3
For example, a sample is to be chosen to answer certain questions regarding political
preferences in a certain state in the United States. The sample involves the choice of, say,
1000 families, and a survey is to be conducted. Now, suppose it turns out that random
sampling is not used. Rather, all or nearly all of the 1000 families chosen live in an urban
setting. It is believed that political preferences in rural areas differ from those in urban areas. In
other words, the sample drawn actually confined the population and thus the inferences
need to be confined to the “limited population,” and in this case confining may be
undesirable. If, indeed, the inferences need to be made about the state as a whole, the
sample of size 1000described here is often referred to as a biased sample.
Experimental Design
The concept of randomness or random assignment plays a huge role in the area of
experimental design, and is an important staple in almost any area of engineering or
experimental science. A set of so-called treatment or treatment combinations becomes
the populations to be studied or compared in some sense.
Example 4
Solution:
The experiment involves four treatment combinations that are listed in the table that
follows. There are eight experimental units used, and they are aluminum specimens
prepared; two are assigned randomly to each of the four treatment combinations. The data
are presented in table below. The corrosion data are averages of two specimens. A plot of
the averages is pictured in below. A relatively large value of cycles to failure represents a
small amount of corrosion. As one might expect, an increase in humidity appears to make
the corrosion worse. The use of the chemical corrosion coating procedure appears to reduce
corrosion.
In this experimental design illustration, the engineer has systematically selected
the four treatment combinations. In order to connect this situation to concepts with
which the reader has been exposed to this point, it should be assumed that the
conditions representing the four treatment combinations are four separate
populations and that the two corrosion values observed for each population are
important pieces of information.
The foregoing example illustrates these concepts:
(1) Random assignment of treatment combinations (coating, humidity) to
experimental units (specimens)
(2) The use of sample averages (average corrosion values) in summarizing
sample information
(3) The need for consideration of measures of variability in the analysis of any
sample or sets of samples
There are other measures of central tendency that are discussed in detail in future
chapters. One important measure is the sample median. The purpose of the sample
median is to reflect the central tendency of the sample in such a way that it is
uninfluenced by extreme values or outliers.
As an example, suppose the data set is the following: 1.7, 2.2, 3.9, 3.11, and 14.7. The
sample mean and median are, respectively,
𝑥̅ = 5.12, 𝑥̃ = 3.9.
Example 5
Two samples of 10 northern red oak seedlings were planted in a greenhouse, one
containing seedlings treated with nitrogen and the other containing seedlings with no
nitrogen. All other environmental conditions were held constant. All seedlings contained
the fungus Pisolithus tinctorus. The stem weights in grams were recorded after the end of
140 days. The data are given below.
0.26, 0.43, 0.46, 0.47, 0.49, 0.52, 0.62, 0.75, 0.79, 0.86
0.28 + 0.32 + 0.36 + 0.37 + 0.38 + 0.42 + 0.43 + 0.43 + 0.47 + 0.53
𝑥̅ (𝑛𝑜 𝑛𝑖𝑡𝑟𝑜𝑔𝑒𝑛) =
10
0.2 + 0.43 + 0.46 + 0.47 + 0.49 + 0.52 + 0.62 + 0.75 + 0.79 + 0.86
𝑥̃(𝑛𝑖𝑡𝑟𝑜𝑔𝑒𝑛) =
10
One example is the contrast of two data sets below. Each contains two samples and
the difference in the means is roughly the same for the two samples, but data set B
seems to provide a much sharper contrast between the two populations from which
the samples were taken. If the purpose of such an experiment is to detect differences
between the two populations, the task is accomplished in the case of data set B.
However, in data set A the large variability within the two samples creates difficulty. In
fact, it is not clear that there is a distinction between the two populations.
Just as there are many measures of central tendency or location, there are many
measures of spread or variability. Perhaps the simplest one is the sample range Xmax
−Xmin. The sample measure of spread that is used most often is the sample standard
deviation. We again let x1, x2, xn denote sample values.
Example 6
It should be apparent that the variance is a measure of the average squared deviation
from the mean 𝑥̅ . We use the term average squared deviation even though the
definition makes use of a division by degrees of freedom n − 1 rather than n. Of course,
if n is large, the difference in the denominator is inconsequential. As a result, the sample
variance possesses units that are the square of the units in the observed data whereas
the sample standard deviation is found in linear units.
The reflux rate should be held constant for this process. Consequently, production
personnel change this very infrequently.
A retrospective study would use either all or a sample of the historical process data
archived over some period of time. The study objective might be to discover the
relationships among the two temperatures and the reflux rate on the acetone
concentration in the output product stream. However, this type of study presents some
problems:
1. We may not be able to see the relationship between the reflux rate and acetone
concentration because the reflux rate did not change much over the historical period.
2. The archived data on the two temperatures (which are recorded almost
continuously) do not correspond perfectly to the acetone concentration
measurements (which are made hourly). It may not be obvious how to construct an
approximate correspondence.
3. Production maintains the two temperatures as closely as possible to desired targets or
set points.
Because the temperatures change so little, it may be difficult to assess their real
impact on acetone concentration.
4. In the narrow ranges within which they do vary, the condensate temperature tends
to increase with the reboil temperature. Consequently, the effects of these two process
variables on acetone concentration may be difficult to separate.
As you can see, a retrospective study may involve a significant amount of data, but
those data may contain relatively little useful information about the problem.
Furthermore, some of the relevant data may be missing, there may be transcription or
recording errors resulting in outliers (or unusual values), or data on other important
factors may not have been collected and archived. In the distillation column, for
example, the specific concentrations of butyl alcohol and acetone in the input feed
stream are very important factors, but they are not archived because the
concentrations are too hard to obtain on a routine basis. As a result of these types of
issues, statistical analysis of historical data sometimes identifies interesting phenomena,
but solid and reliable explanations of these phenomena are often difficult to obtain.
Example 7
Consider the problem involving the choice of wall thickness for the nylon connector.
This is a simple illustration of a designed experiment. The engineer chose two wall
thicknesses for the connector and performed a series of tests to obtain pull-off force
measurements at each wall thickness. In this simple comparative experiment, the
engineer is interested in determining whether there is any difference between the 3 32-
and 1 8-inch designs. An approach that could be used in analyzing the data from this
experiment is to compare the mean pull-off force for the 3 32 -inch design to the mean
pull-off force for the 1 8-inch design using statistical hypothesis testing.
Models play an important role in the analysis of nearly all engineering problems. Much
of the formal education of engineers involves learning about the models relevant to
specific fields and the techniques for applying these models in problem formulation and
solution. As a simple example, suppose we are measuring the flow of current in a thin
copper wire. Our model for this phenomenon might be Ohm’s law:
We call this type of model a mechanistic model because it is built from our underlying
knowledge of the basic physical mechanism that relates these variables. However, if
we performed this measurement process more than once, perhaps at different times, or
even on different days, the observed current could differ slightly because of small
changes or variations in factors that are not completely controlled, such as changes in
ambient temperature, fluctuations in performance of the gauge, small impurities
present at different locations in the wire, and drifts in the voltage source.
Sometimes engineers work with problems for which there is no simple or well understood
mechanistic model that explains the phenomenon. The type of model that uses our
engineering and scientific knowledge of phenomenon, but is not directly developed
from our theoretical or first-principles understanding of the underlying mechanism is
called the empirical model
It was mentioned that decisions often need to be based on measurements from only a
subset of objects to conclusions for a population of objects was referred to as statistical
inference. Probability models help quantify the risks involved in statistical inference, that
is, the risks involved in decisions made every day.
Activity 01.
Direction: Check our Google classroom.
Reference
Compiled by: