Professional Documents
Culture Documents
1. To test the Goodness of fit for the given quality characteristic using Uniform distribution
2. To test the Goodness of fit for the given quality characteristic using Binomial distribution
3. To test the Goodness of fit for the given quality characteristic using Poisson distribution
4. To test the Goodness of fit for the given quality characteristic using Normal distribution
5. Construction of control chart for attribute quality characteristic ( P,NP,C, X - R chart)
6. Construction of control charts using Minitab Software
7. Conduction of Repeatability and Reproducibility studies for appraiser
8. Assessing Process Capability of the given manufacturing process using Normal Probability
paper method and process capability indices
10. Conduction of Design of Experiments–Full Factorial approach for the given quality
characteristic for Catapult and Golf Experiment
11. Simple Linear Regression model and find the dependent variable with help of independent
variable
REFERENCE BOOKS
Exercise No. 1
UNIFORM DISTRIBUTION
The simplest of all discrete probability distributions is one where the random variable assumes all
its values with equal probability.
If the random variable X assumes all the values X1,X2,X3-------XN with equal probability, then the
discrete uniform distribution is given by
Goodness – of - fit test: A test to determine if a population has a specified theoretical distribution
.The test is based on how good a fit. We have between the frequency of occurrence of
observation in an observed sample and the expected frequencies obtained from the hypothesized
distribution
To illustrate, consider the tossing of a die, we hypothesize that die is honest, which is equivalent to
testing the hypothesis that the distribution of outcomes is uniform. Suppose that the die is tossed
120 times and each outcome is recorded.
Theoretically if the die is balanced, we would expect each face to occur 20 times.
By comparing the observed frequencies with the corresponding expected frequencies, we must
decide whether these discrepancies are likely to occur as a result of sampling fluctuations and the
die is balanced or the die is not honest and the distribution of outcomes is not uniform .It is
common practice to refer to each possible outcomes of an experiment as a cell .Hence in our
illustration we have six cells. The appropriate statistic on which we base our decision criteria for an
experiment involving k cells is defined by the following theorem.
A goodness – of - fit test between observed k expected frequencies is based on the quantity.
Aim: To check whether the distribution of outcomes of tossing a dice is a uniform distribution or
not.
Formulae :
= 1/6×120 = 20
5)
Calculation :
fo = 21, fe = 20
= 0.05
Where is value of the random variable whose sampling distribution is approximated very closely
by the Chi-square distribution .The symbols oi and ei represents the observed and expected
frequencies, respectively for the i cell.
If the observed frequencies are close to the corresponding expected frequencies, the 2 value will
be small indicating a good fit. If the observed frequencies differ considerably from the expected
frequencies, the value will be large and the fit is poor. A good fit leads to the acceptance of Ho,
where as a poor fit leads to its rejections. The critical region will, therefore fall in the right tail of
the chi-square distribution. For level of significance equal to .We find the critical value chi –
square value from table and then > alpha constitutes the critical region. The decision criterion
described here should not be used unless of the expected frequencies is at least equal to 5.
The number of degrees of freedom associated with the chi-square distribution used here depends on
two factors; the no. of cells in the experiments and the no. of quantities obtained.
Result: the value calculated is 4.4 and it is less than the table value i.e, 11.07 for 95% confidence
level.
Exercise No. 2
BINOMIAL DISTRIBUTION
An experiment often consists of repeated trails, each with two possible outcomes, which may be
labeled as success or failure. Which is true in testing items as they come off an assembly line.
Where each test or trail may indicates a defective or a non – defective item,
Aim: To conduct goodness fit for five dice which follow binomial distribution.
Formulae used:
P(r) = nCr Pr qn - r
P = success = 0.5
q = failure = 0.5
n = No. of trails
TABULAR COLUMN:
∑ = 3.5199
P(r) 120
Calucaltion :
fo = 4, fe = 3.75
= 0.0167
Result: the value calculated is 3.5199 and it is less than the table value i.e, 11.07 for 95%
confidence level, The fit is good.
Exercise No. 3
POISSON DISTRIBUTION
Experiments yielding numerical values of a random variable X, the number of success occurring
during a given time, interval or in a specified region are often called “Poisson experiments”.
The given time intervals may be of any length. Such as a minute a day, a week, a month or even a
year. Hence a poisson experiments might generate observation for the random Variable X
representing the number of telephone calls per hour received by an office, the number of days
school is closed due to snow during the winter or the number of postponed games due to rain during
a baseball season.
The specified region could be a line segment ,an area, a volume or perhaps a Piece of material. In
this case, X might represent the number of field mice per acre. The number of bacteria in a given
culture, or the number of typing errors per page.
1. The number of successes occurring in one time interval or specified region are independent of
those occurring in any other disjoint time interval or region of space.
2. The probability of a single success occurring during a very short time interval or in a small
region is proportional to the length of the time interval or the size of the region and does not
depend on the number of success occurring outside this time interval or region.
3. The probability of more than one success occurring in such a short time interval or failing in such
a small region is negligible.
Aim: To check whether the distribution of the outcomes of defective marbles is a poisson
distribution.
Formula used:
where µ = n p (mean)
p = % of defective = 5% = 0.05
n = sample size = 20
Tabular column:-
7 6 1 1 0.003 0.51
∑ = 6.26
Observation :-
Degree of freedom, ν = k- m -1
Calculation:
(1) µ = n p = 20 0.05 = 1
(2)
= 0.3678
P(1) = 0.3678
P(2) = 0.1839
P(3) = 0.0613
P(4) = 0.0153
= 0.3678 100
= 36.78
= 0.3885
Result: the value calculated is 6.26 and it is less than the table value i.e, 11.07 for 95% confidence
level, The fit is good and hence the null hypothesis is accepted.
Exercise No. 4
NORMAL DISTRIBUTION
The variation of the pattern obtained for a set of quality characteristic data that has been produced
by a process subjected to a chance cause alone is known as normal distribution.
AIM: - To Conduct a goodness of fit study for a quality characteristic and to ascertain
35.7 32.3 32.9 31.8 32.4 33.8 32.7 33.3 33.8 31.3
4 2 7 3 9 0 7 6 1 4
33.5 30.2 30.5 29.8 32.5 31.9 35.1 31.9 30.9 30.9
6 4 7 1 9 3 6 5 7 1
33.3 31.7 31.6 31.1 31.6 33.3 32.2 33.4 29.7 29.9
2 6 1 9 9 4 1 1 4 5
28.9 36.0 35.3 34.8 34.1 33.9 29.1 29.1 33.8 28.7
6 1 9 6 6 1 0 2 1 9
29.2 28.5 33.5 31.5 31.5 30.6 31.5 31.3 30.5 30.5
0 9 6 9 3 7 9 6 3 9
TABULAR COLUMN:-
35.3919 – 36.0103 11 2 4 16 8 32
36.2684
50 36 150
TABULAR COLUMN
Calculations:
A = 31.0635,
Specimen calculation
= = 0.0207
= = 0.0808
= 2.985
Result: the value calculated is 5.6572 and it is less than the table value i.e, 9.48 for 95% confidence
level, The fit is good
Exercise No 5
Aim: To conduct 100% inspection on manufacturing washer and the data is summarized by the
hour. To plot a P chart for the data and interpret it.
Theory: The terminology defective or non defective is often used to identify the two classification
of product. More recently the terminology conforming and non conforming has become popular.
Quality characteristics of this type are called attribute. The first of these relates to the fraction of
non conforming or defective product produced by a manufacturing process and is called the control
chart for non conforming fraction or P chart
The actual operation of this chart would consist of rating subsequent samples of ‘n’ unit,
computing the sample fraction non conforming ‘p’, and plotting the statistic P on the chart. As long
as P remains within the control limits and the sequence of plotted points does not exhibit any
systematic non random pattern, we can conclude that the process is in control at the level ‘P’ if a
plot is outside the control limits, we can conclude that the process fraction non conforming has
most likely shifted to a new level and process is out of control.
p
Di
mn
p1 p
UCL p p 3
n
CL p p
p1 p
LCL p p 3
n
FORMULAE:
d
1. i = n , where d ---- no of defectives
p
d p
p1 p
UCL p p 3
n
CL p p
p1 p
LCL p p 3
n
OBSERVATION:
(n) (i)
10 10 20 0 0
11 11 20 3 0.15
12 12 20 2 0.1
13 13 20 3 0.15
14 14 20 0 0
15 15 20 2 0.1
16 16 20 0 0
17 17 20 3 0.15
18 18 20 1 0.05
19 19 20 1 0.05
20 20 20 1 0.05
n 400 d 33
CALCUATION:
p
d
n = = 0.0825
2) Fraction defective
d
i = n = = 0.05
CL p p = 0.0825
Procedure :
1. The group with a constant sample size is inspected for 20 hours
2. At the end of each hour, the number of unit is inspected for defectives value are using Go &
no go gauge noted down.
3. 20 bowls of washer are given in which a random amount of washer are taken and inspected.
4. The fraction defective is calculated as the ratio of number of defective unit to the total no of
units checked.
5. The total number of inspected washer are calculated as and total number of defective are
called .
6. A graph of fraction defective v/s the number of hours is to be plotted to determine whether
the process is in control or not.
Result:
It is seen from the graph that the points are within control limit, there for the process is in
state of statistical control.
UCLp = 0.267
CLp = 0.0825
LCLp = -0.1021
Exercise No 6
Aim : To plot an np- chart and determine the trail control limits and to ascertain whether the
production of washers is in statistical control or not.
THEORY: It is always possible to base a control chart on the number non conforming rather than
the fraction non conforming. This is often caked ab number forming control chart or np-chart.
The parameter of this chart is as follows,
UCLnp np 3 np 1 p
CLnp np
LCLnp np 3 np 1 p
p
Di
where, n
Observation:
n 300 d 126
If a standard value for P is un available, then p can be used to estimate P. many non statistically
trained personnel find the np chart easier to interpret then the usual fraction non conforming control
chart.
Procedure:
Department of IEM, BMSCE Bangalore - 19Page 18
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
CALCUATION:
p
d
n = = 0.42
np = 20 × 0.42 = 8.4
Result:
It is seen form the graph that the points are within control limits, therefore the process is in
the state of statistical control.
The UCL is 15.02, LCL is 1.78 and the control limit CL is 8.4.
Exercise No 7
Theory: C chart are use for number of non conformities with constant sample size. The of
conformities ‘x’ follows the poisson distribution with
where, x = 1,2,…………n
Tabular column
Procedure:
1. A set of 10 sheet of page from a certain article is taken
2. The spelling mistake from each such page is found and counted. They are considerd as
defects.
3. The number of errors/defectives in each page is found and counted.
4. The total number of defects and the number of subgroups is calculated and hence c is found
using the formula.
5. The control limit is found a group is plotted to determine if the process is in control.
CALCUATION:
c
c
k = = 12.4
Department of IEM, BMSCE Bangalore - 19Page 20
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
Result: It is seen form the graph that the points are within control limits, therefore the process is
in the state of statistical control.
The UCL is 22.96, LCL is 1.835 and the control limit CL is 12.4
Exercise No 8
AIM:-To Plot a control chart for variable for the random size of bolt head and to ascertain whether
the process of manufacturing of bolts is a state of statistical control.
_
CL _ X
X
_ _
LCL _ X A2 R
X
The process variability maybe monitored by plotting value of the sample range R on a control chart.
The centre line and control limits for R-chart are as follows
_
UCL _ D4 R
R
_
CL _ R
R
_
LCL _ D3 R
R
Observation:
_
X = 619.49 _
R = 36.04
Procedure :
1. The 100 bolts are place in 20 group, each group consisting of 5 bolts.
2. The first group is selected and with the help of vernier caliper length is measures the
obtained value are noted down which provides X – values.
_
3. The average of these 5 values gives X and the difference of max value gives the range.
4. The above procedure is repeated for all 20 sub groups.
_
5. Using the value, X and R are calculated.
6. Then, the upper control limit and the lower control limit are calculated for both X -R
charts.
7. The control limits and the values of X is plotted on a graph. Similarly the value of R
and the control limits are plotted on another graph to determine whether, the process is
in control or not.
CALCULATION:
_
_
X
X
k = = 30.974
_
R
R
k = = 1.802
R 1.802
= d 2 = 2.326 = 0.774
Cp = 6 = 6× 0.774 = 4.648
USL LSL
Pr ocess Capability
6 = = 0.43 < 1
Result:
It is seen from the graph that all the points are within the control limits, thre fore the process
is in a state of statistical control
UCL _ CL _ LCL _
R 3.804, R = 1.802, R =0
Since the process capability index is less than 1, the process is not capable.
Inference: it is seen that all the points are within the control limit. Therefore the process is not in the
state of statistical control.
Exercise No 9
Theory:
Reproducibility is defined as the variability due to different operation using the gauge,[I,e due to
different time period, different environment, or different condition]
Repeatability is defined as the inherent precision of the gauge itself. Range average method
computes the total measurement system variability to be separated in to repeatability,
reproducibility and part variation. Repeatability is given by,
R _
R
repeatability =
d2 = average of ranges for all appropriates parts
Reproducibility is given by
Rx _
R
reproducibility =
d2 = average of ranges for all appropriates parts
2
gage = 2repeatability + 2reproducibility
Observation:
SL
No Operators
A B C
TR 1 TR2 X R TR1 TR X R TR TR2 X R
2 1
1 7.82 7.80 7.8 0.0 7.92 7.86 7.8 0.0 7.80 7.78 7.79 0.02
1 2 9 6
2 7.90 7.86 7.8 0.0 7.90 7.92 7.9 0.0 7.80 7.80 7.8 0
8 4 1 1
3 7.86 7.82 7.8 0.0 7.90 7.90 7.9 0 7.9 7.86 7.88 0.04
4 4
SI Operators
No A B C
TR 1 TR2 X R TR1 TR X R TR TR2 X R
2 1
4 7.8 7.86 7.8 0.0 7.96 7.90 7.9 0.0 7.88 7.90 7.89 0.02
3 6 5 6
5 7.82 7.84 7.8 0.0 7.86 7.80 7.8 0.0 7.86 7.90 7.89 0.02
3 4 3 6
Department of IEM, BMSCE Bangalore - 19Page 26
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
6 7.90 7.7 7.8 0.2 7.9 7.9 7.9 0 7.88 7.9 7.89 0.02
7 7.88 7.8 7.8 0.0 7.8 7.9 7.8 0.0 7.86 7.84 7.85 0.02
4 8 5 1
8 7.86 7.80 7.8 0.0 7.86 7.90 7.8 0.0 7.9 7.92 7.91 0.02
3 6 8 4
9 7.8 7.7 7.7 0.1 7.88 7.70 7.7 0.0 8 8 8 0
5 9 2
10 7.9 7.88 7.8 0.0 7.7 7.88 7.7 0.0 7.86 7.88 7.87 0.02
9 2 9 2
11 7.8 7.78 7.7 0.0 8 8 8 0 7.9 7.8 7.85 0.01
9 2
12 7.8 7.78 7.7 0.0 7.80 7.86 7.8 0.0 7.86 7.88 7.87 0.02
9 2 3 6
13 7.8 7.78 7.7 0.0 7.86 7.86 7.8 0 7.86 7.88 7.87 0.02
9 2 6
14 7.86 7.8 7.8 0.0 7.9 7.86 7.8 0.0 7.8 7.92 7.86 0.12
3 6 8 4
15 7.7 7.84 7.7 0.1 7.8 7.86 7.8 0.0 7.86 7.88 7.87 0.02
7 4 3 6
16 7.8 7.9 7.8 0.1 7.86 7.9 7.8 0.0 7.86 7.86 7.86 0
5 8 6
17 7.7 7.72 7.7 0.0 7.7 7.8 7.7 0.0 7.86 7.86 7.86 0
1 2 5 1
18 7.8 7.9 7.8 0.1 7.88 7.86 7.8 0.0 7.76 7.7 7.73 0.06
5 7 2
19 7.82 7.84 7.8 0.0 7.86 7.9 7.8 0.0 7.86 7.82 7.84 0.04
3 2 8 6
20 8 8 8 0 7.9 7.88 7.8 0.0 7.86 7.84 7.82 0.04
9 2
TR = Trail
_ _ _
X 1 7.827 X 2 7.867 X 3 7.85
_ _ _
R1 = 0.06 R 2 = 0.04 R3 = 0.03
Procedure:
Calculation:
_ _ _ _
R ( R1 + R 2 + R3 ) = = 0.0433
_ _ _ _
X max max( X 1, X 2, X 3 ) = max (7.827,7.867,7.85) = 7.867
_ _ _ _
X min = min ( X 1, X 2, X 3 ) = min (7.827,7.867,7.85) = 7.827
R_ _ _
X X max - X min = 7.867 – 7.827 = 0.04
From table d2 = 1.128, for n = 2
R
repeatability =
d 2 = = 0.0384
Rx
reproducibility =
d2 = = 0.0236
2
gage = 2repeatability + 2reproducibility
= (0.0384)2+(0.0236)2 = (0.002)
gage = 0.045
Inference: Since repeatability is more than reproducibility and the P/T ratio is 0.67 which is greater than
0.1, we can conclude that the gauge capability is not good.
Exercise No 10
Probability plotting is an informal way of conducting a process capability analysis especially with
a small number of data points, say 30 or fewer. As in the precision case, the process must be in
control to use the technique.
Monographs by Nelson (1979) and Shapiro (1980), both sponsored by the a SQC, describe
probability plotting in considerable detail. The basic idea behind the method is that a special
transformation has been applied to the vertical scale of a graph of the assumed cumulative
distribution function. This transforms all cumulative distribution function of the assumed type into
straight lines.
If the assumed distribution is correct, the sample order statistics will be nearly linear when plotted
on probability paper. There will usually be random deviations from linearity, but the larger the
sample size, the greater the tendency toward line rarity of the distributional assumption is incorrect
the plotted points will be nonlinear in a systematic manner.
Shapiro (1980) gives the basic steps in the preparation of probability plots:
Since we do not need 100 data points for probability plotting, we have randomly selected 20 points,
without replacement from table 9.2 and listed them (in order selected) as follows:
2 94.59 12 104.45
3 95.89 13 104.58
4 97.05 14 104.71
5 99.51 15 105.19
6 100.12 16 106.81
7 100.57 17 106.94
8 102.13 18 107.41
9 102.90 19 108.12
10 104.13 20 113.56
The ordered observations are then plotted against P[i]. The special graph paper for this example has a
relabeled horizontal axis for plotting the P[i] values. The drawn to accommodate the smallest and
largest values. The smallest sample value x[i] = 93.21, is plotted against P[i] = 50/20 = 2.5, the second
smallest; x[2] = 94.59, is plotted against P[2] = 2.5 + 100/20 = 7.5; and so on, until finally, the highest
value, x[20] =113.56, is plotted against P[20] = 97.5. The plotted values are shown in Fig 9.6.
Using the “eyeball” method, a straight line has been drawn through the plotted points. In
evaluation the linearity of the plot, the following should be considered.
1. The observed values are random and will never fall exactly on a straight line.
2. The ordered values are not independent, since they have been ranked Hence, if one point is
above the line. it is unlikely that the points will be scattered about the line.
3. The variances of the extremes (largest and smallest values) are much higher than the
variances in the middle of the plot. Greater discrepancies can be accepted at the extremes.
The linearity of the points in the middle of the plot is more important than the linearity of
the points in the middle of the plot is more important than the linearity at the extremes.
Outliers in the data should be treated with caution. Nelson (1979) states the all too often
“inexperienced analysts tend to over interpret plots and expect them be more orderly than
they usually are.” If a data point is suspected of being a outlier, investigate the causes
behind it rather than discarding it without a second. Remember the first of Shapiro’s three
points presented earlier. The data will some variability, since they are random values.
The plotted data points in Fig 9.6 fall close to a straight line. Informally the analysis of
normality is not rejected. Now and can be estimated form the probability plot. We
approximate = 102.9 hours as the point at which P[i] = 84.1 “percent under” P[2] = 50 “percent
under.” Thus.
= 108.2 – 102.9 =5.3 hours
Terms that is rater far from the value determined using all of the data in Eq. but the
estimated standard deviation is quite close.
80
70
Percent
60
50
40
30
20
10
1
85 90 95 100 105 110 115 120
percent(2)
Observation:
Procedure:
1. A set of 30 ms plates are taken and weighed..
2. These weight are noted down from the electronic weighing machine.
3. The weight are ranked in the descending order and the median ranking is called using
the formula.
4. Using the weights, the tolerance limit for the m.s. plate are set.
5. The values are plotted on the normal probability paper with weight on x- axis and values
of probability on i.e median ranking y- axis.
6. The mean µ and standard deviation ( ) is found from the graph
7. The value of process capability is found from the graph
8. The value of process capability is found using the formula.
CALCULATION:
= = 1.593
USL LSL
Pr ocess Capability
6
= = 1.04
RESULT:
Exercise No 11
Aim: To Construct operating characteristic cure and determine consumers risk(α) and producer’s
risk (β) for the given sampling plan
Theory: Acceptances sampling is the process of evaluating a portion of the product in a lot for the
propose of accepting or rejecting the lots as either conforming or non conforming a quality
specification. Inspection for acceptance purpose is carried out at many stages in manufacturing.
There are generally two ways in which inspection is carried out: 1) 100% inspection ii) Sampling
inspection. Sampling plans may be grouped in to three categories they are i) Single sampling plan,
ii) Double sampling plan and iii) multiple sampling plan when a decision on acceptance or rejection
of the lot is made on the basis of only one sample, the acceptance plan is known as single sampling
plan. In single sampling plans three number are specified. N= Lot size and C= acceptance number.
The OC Curve for an attribute sampling is a graph of fraction defective in a lot against the
acceptance Pa. for any fraction defective p’ in a submitted lot, the OC Curve shows the probability
Pa that such a lot will be accepted by the sampling plan, by using sampling plan and respective OC
Curve for the given AQL and LTPD producer’s risk and consumer’s risk can be calculated.
For simulating the manufacturing situations in these exercise plastic beads are represented by
brown colour beads and defective are represented by white beads. A mixture of brown and white
beads will constitute a real time lot containing good ones and bad ones. For preparing a lot size of
1000 containing 1% defective take 990 brown colour bead and 10 white beads. Similarly up to 9%
defective lots can be prepared by mixing appropriate defective beads in the lot. Selection of a
sampling plan: N = 100, n= 25, c=2, conduction of the experiment:
The lot size of 1000 with different proportion defectives is prepared as per the procedure explained
above
1. The number of the cavities in the scoop determines the sample size, we selected scoop
having 25 cavities, i.e. n=25
2. Start with 1% defective lot, the scoop is dipped in the bin containing the beads and taken out
3. Note down the number of defective in the observation sheet. If the number of white beads
are less than or equal to the acceptance number i.e., 2, accept the lot or else reject the lot.
4. Beads collected in the scoop are put back in to bin the mixture of bead is stirred well to
ensure uniformity of distribution of defectives in the lot. Repeat the above steps for 25 times
and note down the number of times the lot is accepted or rejected
5. Step no. 5 is repeated for up to 9% defective lots.
Department of IEM, BMSCE Bangalore - 19Page 34
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
6. Check whether the obtained data follows Poisson distributing using goodness of fit chi
square test.
7. For the different value percent defective(p’) and probability acceptance Pa, draw the OC
Curve.
8. For AQL = 1.5% and LTPD= 6.5%. Draw the horizontal line corresponding to points.the
probability values represents the producer’s risk and consumer’s risk
Tabular column:
= 4.1627
<
Conclusion: since chi square calculated value is less than chi square critical value, the observed
data follows passion distribution.
Exercise No 12
Theory: Design of experiment (DOE) is a powerful technique used for discovering a set of process
variables (or factors) which are most important to the process(or system) and then determine at
what levels these factors must be kept to optimize the process (or system ) performance. It provides
a quick and cost-effective method of understand and optimize any manufacturing processes. It is a
direct replacement of the hit or miss approach of experimentation which requires a lot of guess
work and luck for its success in real life situations. Moreover, the hit or miss approach does not take
into account interactions among the factors (or variables) and therefore there is always a risk of
arriving a false optimum conditions for the process under investigation.
In the past decade or so, DOE has gained increasing importance in the reduction of variability in
core processes, whereby consistent product quality can be achieved. Moreover, companies striving
for a six-sigma approach to achieving quality treat DOE as the key player. The author believes that
DOE must be a key element of the management strategy in the 21st century in many manufacturing
companies during new product and process introduction, so that robust and consistent performance
of the product can be achieved in the user environment. The purpose of this experiment is to
provide an insight into the process of understanding the role of DOE on the part of a group of
engineers and managers with the help of golf experiment. The results of the experiment have been
extracted from real golf experiment performed by a group of engineers in a company during a
training program on DOE. The first experiment was performed to identify the key variables or
factors which affect the response (in-flight distance) of interest. This is called a screening
experiment and the objective is to separate out the key factors (or variables) from the trivial. Having
identified the key factors, the team performed a second experiment with object of understanding the
nature of interaction (if present) among the key factors. The results of the experiment were analyzed
using Minitab software for rapid and easier understanding of the results.
Factors –
The following are the factors that are considered in this case are
Stop Angle
Hook attachment point
Type of Ball
Levels – These factors are each tested at only two levels, high and low. A complete set of tests is
run as shown in Table 1. In Table 1, a plus sign indicates a level and a minus sign indicates a low
level. Thus, in test #1 the club, the swinging angle and the ball are all at the low level. In the test
#8, all the levels are high setting.
Interactive effects: you notice three other columns in Table 1, AB,BC, and AC. AB is the products
of the level values for A and B (Ball and Strings) . The other columns also products. These products
will be indicators of interactive effects between the factors.
1 -1 -1 -1 +1 +1 +1
2 -1 -1 +1 +1 -1 -1
3 -1 +1 -1 -1 -1 +1
4 -1 +1 +1 -1 +1 -1
5 +1 -1 -1 +1 +1 -1
6 +1 -1 +1 -+1 -1 +1
7 +1 +1 -1 -1 -1 -1
8 +1 +1 +1 -1 +1 +1
For each test number, three repetitions were made. In other words, the ball was hit three times at
each setting and the travel distance was recorded. From these an average value for each test is
determined.
Important variables (screening) – If we see a small difference we can say that the factors is not so
important. Thus, the difference defines the relative importance of each factor. This also works for
the combined factors so we can see the relative importance of the interactive effects.
A B C
AB BC AC
Avg Y- 51.8 50.1 83.8 64.4 70.4 64.5
Avg Y+ 81.1 82.8 49.1 68.5 62.5 68.4
Y 29.4 32.7 -34.7 4.1 -7.9 3.9
From the Y values we can see that the three factors are of relatively equal importance to the
distance travelled by the ball. The interactive effects are of relatively small importance, however.
ANALYSIS OF RESULTS:
Predicting with the Empirical Model:
The information in Table 3 can be combined to produce a predictive equation for travel distance.
Ŷ=Ȳ+ (eqn. 2)
Where Ŷ is the predicted travel distance of the ball. A is the Y value for factor A. A is the scaled for
type of ball ( scaled between -1 and +1) and the other terms follow from those definitions.
Ŷ = 66.4+
Ŷ = 52.4 inches.
52.4 inches is close to the 51 inches that was average for the experimental value. For the #8,
equation 2 predicts a distance of 80.1 inches, less than 2 inches away from the 82 inch experimental
value. Thus we can see that the equation is a reasonable predicator of the past. Now we need to see
if it can predict intermediate values of the different factors.
Predicting the future: To keep things simple, let’s set the string lower position (A=-1), the stop
angle at high position (C= +1), and the arm length halfway between high and low (B=0). Therefore,
AB=0,BC=0 and AC=-1. Equation 2 predicts a value of 32.4
It may be noted that there is an assumption of linearity for the effect of each of the factors. If
expectations are not be true, there is a need for more runs for reduction of variability of results.
With only two levels, a linear model is the most complex possible
Conclusion:
Department of IEM, BMSCE Bangalore - 19Page 38
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
The goal was to introduce you to the field of Design of Experiments and to convince you that it has
value for engineers and scientists. The catapult experiment shows how DOE can be used to build
predictive models and show interactive effects.
Exercise No 13
EXPERIMENTAL WORK:
Factors –
The following are the factors that are considered in this case are
Swinging Angle
Length of club
Type of Ball
Levels – These factors are each tested at only two levels, high and low. A complete set of tests is
run as shown in Table 1. In Table 1, a plus sign indicates a level and a minus sign indicates a low
level. Thus, in test #1 the club, the swinging angle and the ball are all at the low level. In the test
#8, all the levels are high setting.
Experimentation on Equipment
For each test number, three repetitions were made. In other words, the ball was hit three times at
each setting and the travel distance was recorded. From these an average value for each test is
determined.
®
5 7 45 Gol 150 156 175 160.33
® f
6 1 45 TT 100 100 100 100
®
7 4 45 Gol 125 125 125 125
® f
8 3 45 TT 100 100 100 100
®
Ȳ = 165.87
Data Reduction - The interesting thing is how the data is analyzed. If we average all the distances
from tests in which the hook is at the high level, then compare that with the average from all the
tests in which the hook is at the low level, we would expect to see some difference. Indeed, this
difference between high and low setting is the key to this technique. Table 3 shows the results for
those calculations.
Important variables (screening) – If we see a small difference we can say that the factors is not so
important. Thus, the difference defines the relative importance of each factor. This also works for
the combined factors so we can see the relative importance of the interactive effects.
From the Y values we can see that the three factors are of relatively equal importance to the
distance travelled by the ball. The interactive effects are of relatively small importance, however.
ANALYSIS OF RESULTS:
Predicting with the Empirical Model:
The information in Table 3 can be combined to produce a predictive equation for travel distance.
Ŷ=Ȳ+ (eqn. 2)
Department of IEM, BMSCE Bangalore - 19Page 41
Quality Assurance and Reliability Lab Sub Code: 20IM5DLQAR
Where Ŷ is the predicted travel distance of the ball. A is the Y value for factor A. A is the scaled
for type of ball ( scaled between -1 and +1) and the other terms follow from those definitions.
Ŷ = 165.87+
It may be noted that there is an assumption of linearity for the effect of each of the factors. If
expectations are not be true, there is a need for more runs for reduction of variability of results.
With only two levels, a linear model is the most complex possible
Conclusion:
The goal was to introduce you to the field of Design of Experiments and to convince you that it has
value for engineers and scientists. The catapult experiment shows how DOE can be used to build
predictive models and show interactive effects.
Exercise No 14
Theory: Regression analysis is a technique for the modeling and analysis of numerical data
consisting of values of a dependent variable (also called response variable or measurement) and of
one or more independent variables (also known as explanatory variable or predictors). The
dependent variable in the regression equation is modeled as a function of the independent variable,
corresponding parameter (constant). The parameters are estimated so as to give a best fit of the data.
Most commonly the best fit is evaluated by using the least squares method.
Regression can be used for prediction inference, and hypothesis testing, and modeling of causal
relationships.
Procedure:
1. Clean the slip gauge
2. Measure the slip gauge with the help of LVDT.
3. Tabulated the LVDT reading for 15 slip gauges
4. Build the simple regression model.
5. Plot the graph.
Observation:
Calculation
SI
X Y (X)2 (XY) (Y)2
No
Regression Equation
X Slop