Professional Documents
Culture Documents
of
Reliability
and
Maintainability
Second Edition
Steven Davis
This book was written by a U.S. Government employee as part of his
normal duties; therefore, it is in the public domain.
ii
Dedication
The second edition of this book is dedicated to the memory of my good
friend and roommate at Georgia Tech, Joseph Lester “Joey” Wilson
(Textiles, 1978), Feb. 27, 1956 — Jan. 3, 2004. Rest in peace, Joey.
iii
iv
Dr. Robert Abernethy begins his New Weibull Handbook with an excerpt
from Oliver Wendall Holmes’ “The Deacon’s Masterpiece.” I thought it
appropriate to begin with the entire poem—a classic 19th century work on
reliability.
vi
Do! I tell you, I rather guess
She was a wonder, and nothing less!
Colts grew horses, beards turned gray,
Deacon and deaconess dropped away,
Children and grandchildren—where were they?
But there stood the stout old one-hoss shay
As fresh as on Lisbon-earthquake-day!
vii
FIRST OF NOVEMBER,—the Earthquake-day,—
There are traces of age in the one-hoss shay,
A general flavor of mild decay,
But nothing local, as one may say.
There couldn't be,—for the Deacon's art
Had made it so like in every part
That there wasn't a chance for one to start.
For the wheels were just as strong as the thills,
And the floor was just as strong as the sills,
And the panels just as strong as the floor,
And the whipple-tree neither less nor more,
And the back-crossbar as strong as the fore,
And spring and axle and hub encore.
And yet, as a whole, it is past a doubt
In another hour it will be worn out!
viii
Acknowledgements
I would like to thank Tracy Jenkins, our co-op at the time; she performed
the research and wrote the original version of Case Study 3: Fire Truck
Depot Overhaul.
I would like to thank Glenn Easterly, Director, Georgia College & State
University at Robins AFB, for his kind review of the first edition of this
book.
ix
x
Fundamentals of Reliability & Maintainability (R&M)
Table of Contents
Introduction ............................................................................................... 1
1: Reliability.............................................................................................. 2
2: Maintainability..................................................................................... 6
3: Availability............................................................................................ 8
5: Reliability Analyses............................................................................ 30
xii
Fundamentals of Reliability & Maintainability (R&M)
Introduction
However, the need for R&M did not go away. Poor reliability and
maintainability cause mission aborts, increase maintenance costs, and
reduce end item availability, leading to both operator and maintainer
frustration.
1
Reliability Toolkit: Commercial Practices Edition. Rome NY:
Reliability Analysis Center (RAC) (now the Reliability Information
Analysis Center (RIAC)), 1995, p. i.
Chapter 1: Reliability
The classic definition of reliability is “the probability that an item can
perform its intended function for a specified interval under stated
conditions.”2 The reliability of repairable items is traditionally measured
by mean time between failure (MTBF): “A basic measure of reliability for
repairable items. The mean number of life units during which all parts of
the item perform within their specified limits, during a particular
measurement interval under stated conditions.”3 This is calculated by:
T
MTBF = ,
n
where:
2
MIL-HDBK-470A, “Designing and Developing Maintainable Products
and Systems, Volume I and Volume II,” 4 Aug 1997, p. G-15, definition
(2).
3
Ibid., p. G-11.
4
MIL-HDBK-338B, “Electronic Reliability Design Handbook,” 1 Oct.
1998, p. 3-17.
5
Ibid., p. 3-13.
2
T
MTTF = ,
n
where:
Note that before we can calculate either MTBF or MTTF, we must know
the number of failures; therefore, we must define “failure.”
Failure
6
MIL-HDBK-470A, op. cit., p. G-5.
7
MIL-STD-721C, “Definition of Terms for Reliability and
Maintainability”, 12 Jun 1981 (since cancelled), p. 4.
3
A dependent failure (also known as a secondary failure) is “a failure of
one item caused by the failure of an associated item(s). A failure that is
not independent.”8
T
MTBCF = ,
nc
8
MIL-HDBK-338B, op. cit., p. 3-6.
9
MIL-STD-721C, op. cit., p. 4.
10
MIL-HDBK-338B, op. cit.
4
where:
11
Ibid., p. 3-16.
12
Reliability Toolkit: Commercial Practices Edition, op. cit., p. 41.
5
Chapter 2: Maintainability
The classic definition of maintainability is:
Note that this definition includes “retained in” as well as “restored to.”
13
MIL-HDBK-470A, op. cit., p. G-8, definition (1).
14
Ibid., p. G-14.
15
Ibid., p G-15.
16
Ibid., p. G-3.
17
Ibid., p. G-17.
6
The corrective portion of maintainability is typically measured by mean
time to repair (MTTR) (also known as mean repair time (MRT)), defined
as: “the sum of corrective maintenance times at any specific level of
repair, divided by the total number of failures within an item repaired at
that level during a particular interval under stated conditions.”18 MTTR is
calculated by:
CMT
MTTR = ,
n
where:
Corrective maintenance time, also known as repair time, is “the time spent
replacing, repairing, or adjusting all items suspected to have been the
cause of the malfunction, except those subsequently shown by interim test
of the system not to have been the cause.”19
18
Ibid., p. G-11.
19
Ibid., p. G-15.
20
Ibid., p. G-9.
21
Ibid.
7
Chapter 3: Availability
The classic definition of availability is “a measure of the degree to which
an item is in an operable and committable state at the start of a mission
when the mission is called for at an unknown (random) time.”22 It is
calculated by:
Uptime
Availability = ,
Uptime + Downtime
where:
Note that the formula above is only a notional equation for availability:
depending on the exact definitions used for Uptime and Downtime, there
are four distinct measures of availability, and several additional variations
of the most frequently used variety, operational availability.
22
Ibid., p. G-2.
23
Reliability Toolkit: Commercial Practices Edition., op. cit., p. 11.
24
MIL-HDBK-470A, op. cit., p. G-7.
8
MTBF
Ai = ,
MTBF + MTTR
where:
MTBM
Aa = ,
MTBM + MTTRactive
where:
25
Reliability Toolkit: Commercial Practices Edition., op. cit., p. 12.
26
Ibid.
27
Ibid.
9
MTBM
Ao = ,
MTBM + MDT
where:
T
MTBM = ,
m
where:
28
Ibid.
29
MIL-HDBK-470A, op. cit., p. G-11.
30
Ibid., p. G-10.
10
definition includes only corrective maintenance downtime, while
preventive maintenance downtime should also be included in the
operational availability equation above. The MIL-HDBK-338B definition
of downtime is “that element of time during which an item is in an
operational inventory but is not in condition to perform its required
function,”31 which includes preventive maintenance downtime, as the item
“is not in condition to perform its required function” while it is undergoing
preventive maintenance. Active time is “that time during which an item is
in an operational inventory.”32
DT
MDT = ,
m
where:
T T
MTBM m m T
Ao = = = = .
MTBM + MDT T DT T + DT T + DT
+
m m m
31
MIL-HDBK-338B, op. cit., p. 3-5.
32
MIL-HDBK-470A, op. cit., p. G-1.
11
T + ST
Ao = ,
T + ST + DT
where:
Uptime
Availability = ,
Uptime + Downtime
where:
In the first equation for operational availability, the item is credited with
Uptime only when it is being operated; in the second equation, the item is
credited with Uptime when it is operated and also when it is in standby—
ready to operate, but not in operation. Note that this increases both the
numerator and the denominator by the same amount; therefore, operational
availability including standby time is always greater than operational
availability without standby time. Also, note that for the case that ST=0
both equations for operational availability are equivalent.
The upper limit of standby time is reached if we define standby time as all
time other than operating time and downtime:
33
Reliability Toolkit: Commercial Practices Edition., op. cit., p. 11.
12
STmax = CT − (T + DT ) ,
where:
CT is calendar time,
T is the number of life units, and
DT is the total downtime.
CT − DT DT
Ao = = 1− ,
CT CT
where:
T
UR = ,
CT
34
MIL-HDBK-470A, op. cit., p. G-17.
13
where:
The first equation for operational availability without standby time is:
MTBM
Ao = .
MTBM + MDT
If we assume that, for any given item, neither MTBM nor MDT is a
function of the utilization rate (a common assumption, approximately true
in many cases), it follows that Ao without standby time is also not a
function of the utilization rate.
DT
Ao = 1 − .
CT
T T DT
Since UR = , CT = . Also, since MDT = , DT = m × MDT .
CT UR m
T T
Finally, since MTBM = , m = . Substituting in the previous
m MTBM
equation for DT:
T T × MDT
DT = m × MDT = × MDT = .
MTBM MTBM
Substituting the above expressions for CT and DT into the equation for Ao
with maximum standby time:
T × MDT
DT MDT
Ao = 1 − = 1 − MTBM = 1 − UR × .
CT T MTBM
UR
14
Thus, if neither MTBM nor MDT is a function of the utilization rate,
operational availability with maximum standby time (e.g., VIC rate or MC
rate) is a function of the utilization rate, decreasing in proportion with
increasing utilization rate.
If, however, you were to examine the Air Force vehicle reliability,
maintainability, and availability data as reported in OLVIMS, you would
find that that the above equation does not hold. While, in general,
increasing the utilization rate decreases the VIC rate, the reported VIC rate
ranges from 85% to 95% with rare exception. Why? In practice, when the
utilization rate of a particular type of vehicle is high, that type of vehicle
will receive priority at the maintenance shop; it will move to the front of
the line, so to speak. Therefore, when the utilization rate is high, the mean
downtime will be reduced, resulting in an almost constant VIC rate.
Therefore, our assumption that MDT is not a function of the utilization
rate is not valid in practice.
MTBCF
Do = ,
MTBCF + MTTRS
where:
35
Ibid., pp. G-3 – G-4.
15
period of time. (Excludes time for off-product maintenance
and repair of detached components.)36
36
MIL-HDBK-338B, op. cit., p. 3-13.
16
4. Another approach the contractor can take to expediting parts is to
air freight everything, regardless of urgency. Of course, this also
comes at an additional cost.
By falling into the “availability trap,” the customer has lost, because:
Thus, the contractor has ultimately been rewarded for poor performance.
17
Chapter 4: Introduction to Reliability Math: The
Exponential Distribution
Reliability as a Probability
MIL-HDBK-338B states:
t
F (t ) = ∫ f (t )dt
−∞
37
Ibid., p. G-15, definition (2).
38
MIL-HDBK-338B, op. cit., p. 5-1.
39
Ibid., p. 5-2.
18
The Reliability Function
MIL-HDBK-338B continues:
∞
R(t ) = 1 − F (t ) = ∫ f (t )dt
t
∞ ∞
MIL-HDBK-338B continues:
R(t1 ) − R(t 2 )
λ (t ) =
(t 2 − t1 ) R(t1 )
40
Ibid.
19
The hazard rate, h(t), or instantaneous failure rate, is
defined as the limit of the failure rate as the interval length
approaches zero, or
1 ⎡ − dR (t ) ⎤
h(t ) = K = (5
R(t ) ⎢⎣ dt ⎥⎦
Figure 4-1 shows a typical time versus failure rate curve for
equipment. This is the "bathtub curve," which, over the
years, has become widely accepted by the reliability
community. It has proven to be particularly appropriate for
electronic equipment and systems. The characteristic
pattern is a period of decreasing failure rate (DFR)
followed by a period of constant failure rate (CFR),
followed by a period of increasing failure rate (IFR).
41
Ibid.
42
Ibid., p. 5-28.
20
The Exponential Distribution
MIL-HDBK-338B states:
R(t ) = e − λt
There is one more factor that makes the exponential distribution the single
most important in reliability math.46 As MIL-HDBK-338B states, “The
43
Ibid., p. 5-5.
44
Ibid., p. 5-29.
45
Kailash C. Kapur and Leonard R. Lamberson, Reliability in Engineering
Design. John Wiley and Sons, Inc., 1977, p. 235.
46
MIL-HDBK-338B, op. cit., p. 5-17.
21
simplicity of the approach utilizing the exponential distribution, as
previously indicated, makes it extremely attractive.”47 Not only is the
exponential distribution a good model for electronic parts and complex
systems in general, it also involves the simplest math.
1
MTBF = Θ = .48
λ
n
λtotal = ∑ λi .
i =1
Confidence Levels
47
Ibid., p. 5-29.
48
Ibid., p. 5-18.
22
T
MTBF = ,
n
where:
This is known as the observed reliability. Note that this is a point estimate
of the true reliability; since, as we have discussed, reliability is a
probability, there is a possibility that the true reliability is somewhat
better, or somewhat worse, than the observed reliability. We would like to
be able to calculate confidence levels, so that we could state, for example,
that the widget has a reliability of X hours MTBF measured at the 90%
confidence level. This would mean that we are 90% confident that the
MTBF of the widget is at least X hours, or, equivalently, that there is only
a 10% chance that the MTBF of the widget is less than X hours.
In general, formal reliability tests are time truncated (terminated), that is,
successful completion of the test is defined as completing T hours of
49
RADC Reliability Engineer’s Toolkit. Griffiss Air Force Base, NY:
Systems Reliability and Engineering Division, Rome Air Development
Center, Air Force Systems Command (AFSC), 1988, p. A-47.
23
testing with no more than N failures. The results could be analyzed as a
failure truncated test if the test were terminated early due to excessive
failures or if one or more failures were discovered during the post-test
inspection.
24
Table 4-1: Summary of χ2 Models50
50
Ibid.
51
Ibid., pp. A-48 – A-50. This table has been abridged to include only the
10% and 20% upper and lower confidence levels (those most commonly
used in reliability calculations) and to delete the odd-numbered degrees of
freedom, which are not used in confidence level calculations. It has been
expanded to include more degrees of freedom and more significant digits.
25
Degrees of Probability in Percent
Freedom (f)
10.0 20.0 80.0 90.0
20 12.443 14.578 25.038 28.412
22 14.041 16.314 27.301 30.813
24 15.659 18.062 29.553 33.196
26 17.292 19.820 31.795 35.563
28 18.939 21.588 34.027 37.916
30 20.599 23.364 36.250 40.256
32 22.271 25.148 38.466 42.585
34 23.952 26.938 40.676 44.903
36 25.643 28.735 42.879 47.212
38 27.343 30.537 45.076 49.513
40 29.051 32.345 47.269 51.805
42 30.765 34.157 49.456 54.090
44 32.487 35.974 51.639 56.369
46 34.215 37.795 53.818 58.641
48 35.949 39.621 55.993 60.907
50 37.689 41.449 58.164 63.167
52 39.433 43.281 60.332 65.422
54 41.183 45.117 62.496 67.673
56 42.937 46.955 64.658 69.919
58 44.696 48.797 66.816 72.160
60 46.459 50.641 68.972 74.397
62 48.226 52.487 71.125 76.630
64 49.996 54.337 73.276 78.860
66 51.770 56.188 75.425 81.085
68 53.548 58.042 77.571 83.308
70 55.329 59.898 79.715 85.527
72 57.113 61.756 81.857 87.743
26
Degrees of Probability in Percent
Freedom (f)
10.0 20.0 80.0 90.0
74 58.900 63.616 83.997 89.956
76 60.690 65.478 86.135 92.166
78 62.483 67.341 88.271 94.374
80 64.278 69.207 90.405 96.578
82 66.076 71.074 92.538 98.780
84 67.876 72.943 94.669 100.98
86 69.679 74.813 96.799 103.18
88 71.484 76.685 98.927 105.37
90 73.291 78.558 101.05 107.57
100 82.358 87.945 111.67 118.50
1000 943.13 962.18 1037.4 1057.7
27
Table 4-3: Factors for Calculating Confidence Levels52
Factor
80%
Failures Two- 80%
60% Two-Sided
Sided Two-
80% One-Sided
90% One- Sided
Sided
Time
All Other Lower Lower Upper Upper
Terminated
Cases Limit Limit Limit Limit
Lower Limit
0 1 0.43429 0.62133 4.4814 9.4912
1 2 0.25709 0.33397 1.2130 1.8804
2 3 0.18789 0.23370 0.65145 0.90739
3 4 0.14968 0.18132 0.43539 0.57314
4 5 0.12510 0.14879 0.32367 0.41108
5 6 0.10782 0.12649 0.25617 0.31727
6 7 0.09495 0.11019 0.21125 0.25675
7 8 0.08496 0.09773 0.17934 0.21477
8 9 0.07695 0.08788 0.15556 0.18408
9 10 0.07039 0.07988 0.13719 0.16074
10 11 0.06491 0.07326 0.12259 0.14243
11 12 0.06025 0.06767 0.11073 0.12772
12 13 0.05624 0.06290 0.10091 0.11566
13 14 0.05275 0.05878 0.09264 0.10560
14 15 0.04968 0.05517 0.08560 0.09709
52
The Rome Laboratory Reliability Engineer’s Toolkit. Griffiss Air Force
Base, NY: Systems Reliability Division, Rome Laboratory, Air Force
Materiel Command (AFMC), 1993, p. A-43. This table has been adapted
and abridged to include only the 10% and 20% upper and lower
confidence levels (those most commonly used in reliability calculations).
It has been expanded to include more failures and more significant digits.
Note that The Rome Laboratory Reliability Engineer’s Toolkit is in the
public domain; it can, therefore, be freely distributed.
28
Factor
80%
Failures Two- 80%
60% Two-Sided
Sided Two-
80% One-Sided
90% One- Sided
Sided
Time
All Other Lower Lower Upper Upper
Terminated
Cases Limit Limit Limit Limit
Lower Limit
15 16 0.04697 0.05199 0.07953 0.08980
16 17 0.04454 0.04917 0.07424 0.08350
17 18 0.04236 0.04664 0.06960 0.07799
18 19 0.04039 0.04437 0.06549 0.07314
19 20 0.03861 0.04231 0.06183 0.06885
20 21 0.03698 0.04044 0.05855 0.06501
21 22 0.03548 0.03873 0.05560 0.06156
22 23 0.03411 0.03716 0.05292 0.05845
23 24 0.03284 0.03572 0.05048 0.05563
24 25 0.03166 0.03439 0.04825 0.05307
25 26 0.03057 0.03315 0.04621 0.05072
26 27 0.02955 0.03200 0.04433 0.04856
27 28 0.02860 0.03093 0.04259 0.04658
28 29 0.02772 0.02993 0.04099 0.04475
29 30 0.02688 0.02900 0.03949 0.04305
30 31 0.02610 0.02812 0.03810 0.04147
31 32 0.02536 0.02729 0.03681 0.04000
32 33 0.02467 0.02652 0.03559 0.03863
33 34 0.02401 0.02578 0.03446 0.03735
34 35 0.02338 0.02509 0.03339 0.03615
39 40 0.02071 0.02212 0.02890 0.03111
49 50 0.01688 0.01791 0.02274 0.02428
499 500 0.00189 0.00193 0.00208 0.00212
29
Chapter 5: Reliability Analyses
Reliability Modeling and Prediction
Reliability Modeling
For basic (logistics) reliability, all items, including those intended solely
for redundancy and alternate modes of operation, are modeled in series.
As we saw in chapter 4, if the exponential distribution (constant failure
rate) is assumed, to calculate the reliability of items in series, add their
failure rates:
n
λtotal = ∑ λi .
i =1
53
MIL-HDBK-338B, op. cit., p. 6-20.
54
Ibid.
30
Table 5-1: Redundancy Equation Approximations Summary55
With Repair Without Repair
All units
are active
on-line Equation 4
with equal Equation 1
unit λ
failure n!(λ ) q +1 λ( n − q ) / n =
λ( n − q ) / n = n
1
∑i
rates. (n-
q) out of (n − q − 1)!( µ ) q
n required i=n−q
for
success.
Two
active on-
line units
with
different Equation 2 Equation 5
failure
and repair λ A λ B [( µ A + µ B ) + (λ A + λ B )] λ A 2 λB + λ A λ B 2
rates. λ1 / 2 = λ1 / 2 = 2
One of ( µ A )( µ B ) + ( µ A + µ B )(λ A + λ B ) λ A + λ B + λ A λB
2
two
required
for
success.
One
standby
off-line
unit with
n active
on-line
units
required
for Equation 3 Equation 6
success.
n[nλ + (1 − P) µ ]λ
Off-line
nλ
spare
λn / n+1 = λn / n+1 =
assumed
to have a
µ + n( P + 1)λ P +1
failure
rate of
zero. On-
line units
have
equal
failure
rates.
55
The Rome Laboratory Reliability Engineer’s Toolkit, op. cit., p. 90.
31
Key:
λx/y is the effective failure rate of the redundant configuration where x
of y units are required for success
n = number of active on-line units. n! is n factorial (e.g.,
5!=5x4x3x2x1=120, 1!=1, 0!=1)
λ = failure rate of an individual on-line unit (failures/hour) (note that
this is not the more common failures/106 hours)
q = number of on-line active units which are allowed to fail without
system failure
µ = repair rate (µ=1/Mct, where Mct is the mean corrective maintenance
time in hours)
P = probability switching mechanism will operate properly when
needed (P=1 with perfect switching)
Notes:
1. Assumes all units are functional at the start
2. The approximations represent time to first failure
3. CAUTION: Redundancy equations for repairable systems should
not be applied if delayed maintenance is used.
Reliability Prediction
GB Ground Benign
GF Ground Fixed
GM Ground Mobile
NS Naval Sheltered
NU Naval Unsheltered
AIC Airborne Inhabited Cargo
AIF Airborne Inhabited Fighter
AUC Airborne Uninhabited Cargo
AUF Airborne Uninhabited Fighter
ARW Airborne Rotary Wing (i.e., Helicopter) (Both Internal
and External Equipment)
SF Space Flight
56
MIL-HDBK-338B, op. cit., p. 6-44.
57
The Rome Laboratory Reliability Engineer’s Toolkit, op. cit., p. 92.
33
n
λ EQUIP = ∑ ( N i )(λ gi )(Π Qi )
i =1
where
[End of the Parts Count Reliability Prediction article reprinted from The
Rome Laboratory Reliability Engineer’s Toolkit.]
Failure rate data for use in reliability predictions can be difficult to obtain.
The single best source for failure rate data for electronic components is
MIL-HDBK-217F(2), “Reliability Prediction of Electronic Equipment;”
for non-electronic components, use NPRD-95, “Nonelectronic Parts
Reliability Data 1995,” available from the Reliability Information
Analysis Center, 201 Mill Street, Rome NY 13440.
34
Table 5-2: Top-Level Reliability Prediction (Feasibility Analysis) for
the BPU
Mean
Failure Rate Time
Description in NPRD-95
(Failures/10E6 Between Page
PD Description
Hours) Failure
(hours)
Internal Engine,
Combustion Diesel 14.2389 70,230 2-88
Engine (Summary)
Heat
Engine Cooling Exchangers, 2-
7.8829 126,857
System Radiator 112
(Summary)
Brushless AC Generator, 2-
0.7960 1,256,281
Generator AC 105
Voltage Regulator,
2-
Regulator/Exciter Voltage 5.5527 180,093
166
System (Summary)
Starter, 2-
Cranking Motor 0.0212 47,169,811
Motor 192
(assumed to
be included
Controls 0 N/A N/A
in other
items)
(included in
Governor Engine, 0 N/A N/A
Diesel)
(assumed to
be
Other Devices negligible 0 N/A N/A
for initial
analysis)
35
Chapter 6: Reliability Growth Test (RGT)
[The following article is reprinted from Appendix 6 of RADC Reliability
Engineer’s Toolkit58, which is in the public domain, and can, therefore, be
freely distributed.]
Who pays for the RGT? Does the government end up paying more?
The usual case is that the government pays for the RGT as an additional
reliability program cost and in stretching out the schedule. The savings in
support costs (recurring logistics costs) exceed the additional initial
acquisition cost, resulting in a net savings in [Life Cycle Cost (LCC)]. The
amount of these savings is dependent on the quantity to be fielded, the
maintenance concept, the sensitivity of LCC to reliability, and the level of
development required. It is the old "pay me now or pay me later situation"
58
RADC Reliability Engineer’s Toolkit, op. cit., pp. A-63 – A-68.
59
“TAAF” stands for Test Analyze And Fix.
36
which in many cases makes a program manager's situation difficult
because his performance is mainly based on the "now" performance of
cost and schedule.
Does RGT allow contractors to "get away with" a sloppy initial design
because they can fix it later at the government's expense?
It has been shown that unforeseen problems account for 75% of the
failures due to the complexity of today's equipment. Too low an initial
reliability (resulting from an inadequate contractor design process) will
necessitate an unrealistic growth rate in order to attain an acceptable level
of reliability in the allocated amount of test time. The growth test should
be considered as an organized search and correction system for reliability
problems that allows problems to be fixed when it is least expensive. It is
oriented towards the efficient determination of corrective action. Solutions
are emphasized rather than excuses. It can give a nontechnical person an
appreciation of reliability and a way to measure its status.
The answer to this question is yes in that all programs should analyze and
correct failures when they occur in prequalification testing. A distinction
should be in the level of formality of the growth program. The less
challenge there is to the state-of-the-art, the less formal (or rigorous) a
reliability growth program should be. An extreme example would be the
case of procuring off-the-shelf equipment to be part of a military system.
In this situation, which really isn't a development, design flexibility to
correct reliability problems is mainly constrained to newly developed
interfaces between the "boxes" making up the system. A rigorous growth
program would be inappropriate but a [failure reporting and corrective
action system (FRACAS)] should still be implemented. The other extreme
is a developmental program applying technology that challenges the state-
of-the-art. In this situation a much greater amount of design flexibility to
correct unforeseen problems exists. Because the technology is so new and
challenging, it can be expected that a greater number of unforeseen
problems will be surfaced by growth testing. All programs can benefit
from testing to find reliability problems and correcting them prior to
deployment, but the number of problems likely to be corrected and the
cost effectiveness of fixing them is greater for designs which are more
complex and challenging to the state-of-the-art.
37
How does the applicability of reliability growth testing vary with the
following points of a development program?
39
How much validity/confidence should be placed on the numerical
results of RGT?
6.3 Duane model. Because the Duane model is the one most commonly
used, it will be further explained. The model assumes that the plot of
MTBF versus time is a straight line when plotted on log-log paper. The
main advantage of this model is that it is easy to use. The disadvantage of
the model is it assumes that a fix is incorporated immediately after a
failure occurs (before further test time is accumulated). Because fixes are
not developed and implemented that easily in real life, this is rarely the
case. Despite this problem, it is still considered a useful planning tool.
Below is a brief summary of the Duane model.
∆MTBF
a. Growth Rate α=
∆TIME
1 α
b. Cumulative MTBF MTBFC = T
K
40
MTBFC
c. Instantaneous MTBF MTBFI =
1−α
1
d. Test Time T = [(MTBFI )(K )(1 − α )]α
1
TPC = (MTBFPRED )
2
where
Tα
K= .
MTBFC
α
⎛ MTBF pred ⎞
⎜⎜ ⎟⎟ α −1
K= ⎝ 10 ⎠ = ⎛⎜ MTBF pred ⎞
⎟⎟ .
MTBF pred ⎜ 10
⎝ ⎠
10
60
MIL-STD-1635, “Reliability Growth Testing,” 3 Feb 1978 (since
cancelled), p. 27.
42
Expected Number of Failures
The Duane Model can be used to calculate the number of failures expected
during an RGT.
Note that, by definition, the cumulative reliability, MTBFC, is the test time,
T, divided by the number of failures, N:
T
MTBFC = .
N
T Tα
MTBFC = = ,
N K
T K
N= α = .
T T α −1
K
The above analysis can be extended to predict the expected time to each
failure. Let i be the i-th failure and ti be the time at which the i-th failure is
expected to occur. By substituting i for N and ti for T in the above
equation, we have:
K
i= α −1 .
ti
1
K
ti = ( ) α −1 .
i
43
This can be used, for example, to predict how many failures will occur
during each week of an RGT so as to estimate the number of engineers
that will be required for the FRACAS needed to achieve the planned
growth rate. It can also be used for a Monte Carlo simulation of an RGT.
61
Ibid., p. 32.
44
3. It cannot be used prior to the N-th failure. However, there is a
work-around for this: use a 1-point moving average at the first
failure, a 2-point moving average at the second failure, etc., until
the N-th failure occurs.
With these disadvantages, why would anyone use the moving average
method?
It is this last point that makes the moving average method so attractive.
Recall the discussion of confidence levels in chapter 4. If a system has a
constant failure rate, we can easily calculate the confidence levels using
the χ2 distribution and the number of failures or the table of confidence
level factors. We know the number of failures during the period of interest
with the moving average method. The assumption of a constant failure rate
is not strictly true, as the point of an RGT is to introduce improvements,
and reduce the failure rate, as the RGT progresses. However, if the period
is relatively short in comparison to the entire RGT, the failure rate can be
assumed to be constant during the period. Further, if the failure rate is
decreasing, as it should in an RGT, the confidence levels would be
conservative. Thus, the moving average method provides an easy method
of calculating confidence levels for the current MTBF. We can calculate
the current MTBF at any point with a subtraction and a division; we can
calculate the two-sided confidence levels with two additional
multiplications.
62
Molly Statham and I developed this method while working in the
Loader Program Office in 1996. It is so simple, yet so powerful, I cannot
imagine that no one had previously developed it; however, I have never
seen it mentioned in the literature.
45
AMSAA Method
The Block Modification (or Block Mod) Method differs from the methods
previously described in that, rather than implementing corrective actions
as soon as they are available, they are held and implemented in groups as
Block Mods. The result is that a graph of the instantaneous or current
reliability will appear as a stair-step rather than a continuous line or curve.
There are also two basic disadvantages to the Block Mod Method:
Recommended Approach
63
RADC Reliability Engineer’s Toolkit, op. cit., pp. A-64.
46
technically more rigorous than the moving average method, unless the
RGT is being used in lieu of an RQT—which is never recommended—the
moving average method is adequate. Therefore, the recommended
approach is to use the Duane Model for planning and the moving average
method for assessment and tracking.
Failure Purging
The hard fact is that fixes do not always fix; and, therefore,
the attitude of the government must be to defer judgment
until further testing is conducted. However, even after the
effectiveness of a design fix has been established, failures
47
associated with eliminated failure modes should not be
purged. The reason is—if there has been sufficient testing
to establish the effectiveness of a design fix, then an
appropriate reliability model will, by then, have sufficient
data to reflect the effect of the fix in the current reliability
estimate.
A variation on the failure purging theme is to purge all but the first
occurrence of a particular failure mode—essentially, once a particular
failure mode has been experienced, any recurrence of that failure mode is
ignored—even before a corrective action for that failure mode has been
identified. Obviously, this form of failure purging is also unacceptable.
64
MIL-HDBK-189, “Reliability Growth Management,” 13 Feb. 1981, pp.
87-88.
65
MIL-STD-1635, op. cit., p. 19.
66
RADC Reliability Engineer’s Toolkit, op. cit., pp. A-66.
48
interpolate the test time on each test item at which any particular test item
failed, this is unnecessary extra effort and adds little, if any, accuracy to
the result. Cumulative test time should be calculated using the test time for
each unfailed item as recorded prior to each failure.
67
“R&M Design in System Acquisition,” Air Force Institute of
Technology (AFIT) QMT 335, 16 – 26 Oct 1989.
49
Table 6-1: Moving Average Method Example68
68
MIL-STD-1635, op. cit., p. 31.
50
Chapter 7: Reliability Qualification Test (RQT)
Reliability qualification test (RQT) is defined as “A test conducted under
specified conditions, by, or on behalf of, the government, using items
representative of the approved production configuration, to determine
compliance with specified reliability requirements as a basis for
production approval.”69
69
MIL-STD-785B, “Reliability Programs for Systems and Equipment
Development and Production,” 15 Sep. 1980 (since cancelled), p. 3.
70
MIL-HDBK-781A, “Handbook for Reliability Test Methods, Plans, and
Environments for Engineering, Development Qualification, and
Production,” 1 Apr 1996, p. 17.
71
Ibid., p. 6.
72
Ibid.
51
“The discrimination ratio (d) is one of the standard test plan parameters; it
is the ratio of the upper test MTBF (θ0) to the lower test MTBF (θ1) that is,
Θ
d = 0 .”73
Θ1
There are two types of standard test plans that are of interest here:
Probability Ratio Sequential Test (PRST) plans, summarized in Table 7-1,
and fixed-duration test plans, summarized in Table 7-2. The PRST plans
have a variable length. MIL-HDBK-781A provides the following
guidance for choosing between a fixed-duration test plan and a PRST plan.
73
Ibid.
74
Ibid., p. 7. Actually, the MIL-HDBK-781A definition reads, ““Lower
test MTBF (θ1) is that value which is unacceptable.” However, this is
confusing.
75
Ibid.
76
Ibid.
52
demonstrated by the test, as well as an accept-reject
decision, or when total test time must be known in advance.
77
Ibid., p. 17.
78
Ibid., p. 37.
79
Ibid., p. 19.
53
percent of that for test plan IX-D. Increasing the discrimination ratio to 3.0
(test plan XV-D) reduces the test duration to 9.3 times the lower test
MTBF (θ1), only 21.7 percent of that for test plan IX-D. Thus, doubling
the discrimination ratio, from 1.5 to 3.0, reduces the test duration by a
factor of 4.84. Therefore, the highest discrimination ratio that results in a
feasible upper test MTBF (θ0) should be selected.
Review of Table 7-2 also demonstrates the effect of decision risks on test
time. The test plans with a discrimination ratio of 2.0 and balanced
decision risks (that is, the nominal consumer’s risk is equal to the nominal
producer’s risk) are XII-D, XIV-D, and XX-D. Test plan XII-D, with 10
percent nominal decision risks, has a duration of 18.8 times the lower test
MTBF (θ1). Increasing the nominal decision risks to 20 percent (test plan
XIV-D) reduces the test duration to 7.8 times the lower test MTBF (θ1),
only 41.5 percent of that for test plan XII-D. Increasing the nominal
decision risks to 30 percent (test plan XX-D) reduces the test duration to
3.7 times the lower test MTBF (θ1), only 19.7 percent of that for test plan
XII-D. Thus, tripling the decision risks, from 10 percent to 30 percent,
reduces the test duration by a factor of 5.1. This is a classic case of trading
cost and schedule for risk reduction.
Another point to note is that, in order to reach an “accept” decision for any
of the test plans, the observed reliability must be significantly greater than
the required reliability, the lower test MTBF (θ1). Dividing the test
duration by the maximum number of failures to accept (except for test
plan XXI-D, of course) reveals that the minimum acceptable observed
reliability is 1.196 times the required reliability (test plan X-D). At the
other extreme, test plan XVII-D requires an observed reliability 2.15 times
the required reliability. Although this may seem to be unfair, it is the
nature of dealing with probabilities.
54
Table 7-1: Summary of MIL-HDBK-781A PRST Test Plans 80
80
Ibid., p. 36.
55
Table 7-2: Summary of MIL-HDBK-781A Fixed-Duration Test Plans
81
Test
Producer’s Consumer’s Discrim- Maximum
Test Duration
Risk (α) Risk (β) ination Failures
Plan (multiples
(%) (%) Ratio (d) to Accept
of θ1)
81
Ibid., p. 131.
56
Case Study 1: Integrated Suitability Improvement Program (ISIP)
57
Next Generation Small Loader (NGSL)
Integrated Suitability Improvement Program (ISIP)
Purpose
The Draft AMC Operational Requirements Document (ORD) for the Next
Generation Small Loader (NGSL) (AMC 020-93 I, 7 Mar 1997) requires a
Mission Completion Success Probability (MCSP) of at least 86% for a 60
hour mission at Follow-On Operational Test and Evaluation (FOT&E); by
assuming a constant failure rate, this can be translated to a requirement for
400 hours mean time between critical failure (MTBCF). The ORD
Operational Availability (Ao) requirement is based on 100 hours mean
time between maintenance (MTBM) at FOT&E. MTBM is essentially
mean time between workorder. On-Line Vehicle Interactive Management
System (OLVIMS) data for the current Southwest Mobile Systems (SMS)
25K Loader shows an average of 1.36 to 1.64 failures per workorder
58
(FY93 through FY96); for this period, with the exception of FY94, the
average varied between 1.53 to 1.64. A ratio of 1.50 failures per
workorder is reasonable; this results in a 66.7 hour mean time between
failure (MTBF) requirement.
The first step is to select or develop appropriate RDT plans for both
MTBF and MTBCF. This involves selecting decision risks (consumer’s
risk, the probability of accepting loaders which do not meet the reliability
requirement, and producer’s risk, the probability of rejecting loaders
which do meet the reliability requirement), the discrimination ratio (a
measure of the power of the test to reach an accept/reject decision
quickly), and whether a fixed length or variable length (probability ratio
sequential test (PRST)) is to be used. For the first iteration, only fixed
length MIL-HDBK-781 test plans with 10% and 20% decision risks are
considered. These are summarized below.
59
MTBF MTBCF
Discrim- Test Test
Producer’s Consumer’s Duration Duration
Test ination
Risk Risk (hours) (hours)
Plan Ratio
(α) (β) θ1 = 66.7 θ1 = 400
(d)
hours hours
Note that only one test plan, XVIID, can verify the 400 hour MTBCF
requirement in less than 2,000 hours. This test plan has a discrimination
ratio, d, of 3.0. The discrimination ratio is the ratio of the upper test limit
(the design-to reliability) to the lower test limit (the required reliability);
therefore, if this test plan is selected, the contractor should design to a
predicted 1,200 hour MTBCF, which may not be feasible. Selecting a test
plan with a discrimination ratio of 1.5 would reduce the upper test limit to
600 hours MTBCF. However, this would require a test length of 8,600
hours, which is too long to be practical. A reasonable compromise would
be test plan XIVD, with a discrimination ratio of 2.0, resulting in an upper
test limit of 800 hours. This test plan would require a test length of 3,120
hours.
60
The next step is to determine the RGT length needed. The minimum
recommended test length is 5.0 times the predicted MTBF or MTBCF, or
4,000 hours for MTBCF. Test time can be calculated by:
1
T = [ MTBFinstantaneous × K × (1 − α )]α
where:
1
MTBFcumulative = ×T α ,
K
where:
Tα
K= .
MTBFcumulative
82
Note that this is incorrect; MIL-STD-1635 actually states that the
initial reliability is approximately 10% of the predicted reliability, which
61
If MTBFinstantaneous = MTBFpredicted = 133.3 hours MTBF or 800
hours MTBCF, K is as follows.
Reliability
Growth Rate K for MTBF K for MTBCF
(α)
Reliability
MTBF Test Time MTBCF Test Time
Growth Rate
(hours) (hours)
(α)
means that both T and MTBFcumulative for the initial failure would be
MTBFpred
. Of course, this affects all of the subsequent RGT calculations
10
in this case study. For the correct equation for calculating K, see chapter 6.
62
It is obvious that, even at the extremely aggressive 0.50 growth rate
(which may be unachievable in any case), a 10,000 hour RGT is too long
to be practical.
The Draft AMC ORD addresses the reliability and availability required at
FOT&E. This allows for a System Maturity Matrix (SMM) which
provides for lesser values at measurement periods prior to FOT&E, such
as at QOT&E. In fact, this approach has been applied for the 60K Loader,
which requires 144 hours MTBCF and 22 hours MTBF at IOT&E. Since
the 60K MTBCF requirement at FOT&E is 400 hours, the same as the
NGSL, 144 hours would be a reasonable interim requirement for the
NGSL at QOT&E. However, the ratio of failures to critical failures has
been reduced from 6.67 for the 60K to 6.0 for the NGSL. Therefore, the
corresponding interim requirement would be 24 hours MTBF at IOT&E.
It would be reasonable to select the same test plans previously chosen for
the “textbook approach” for RDT2. Therefore, the design-to/predicted
reliability levels would remain at 133.3 hours MTBF and 800 hours
MTBCF. RGT1 is essentially the same as the RGT in the “textbook
approach,” terminated early, and RGT2 is simply the continuation of
RGT1. The interim reliability requirements, 24 hours MTBF and 144
hours MTBCF, are verified by RDT1. For consistency, a discrimination
ratio of 2.0 should be selected. The fixed length MIL-HDBK-781 test
plans with 10% and 20% discrimination ratios are summarized below.
63
MTBCF Test
MTBF Test
Producer’s Consumer’s Duration
Test Duration
Risk Risk (hours)
Plan (hours)
(α) (β) θ 1 = 144
θ1 = 24 hours
hours
XIID 10% 10% 451 2,707
XIIID 10% 20% 298 1,786
XIVD 20% 20% 187 1,123
Test plan XIVD is a reasonable choice for verifying MTBCF, while test
plan XIID is a reasonable choice for MTBF.
The test time required for RGT1 is the time predicted to achieve an
instantaneous reliability of 48 hours (2.0 times 24 hours) and an
instantaneous critical reliability of 288 hours (2.0 times 144 hours). Test
time is calculated as follows.
Reliability
MTBF Test Time MTBCF Test Time
Growth Rate
(hours) (hours)
(α)
At a nominal 0.35 growth rate, the test time required for MTBCF is
excessive. However, if priority is given to developing corrective actions
for critical failures, and a 0.40 growth rate can be achieved, a 2,743 hour
RGT would be sufficient. To be prudent, this should be rounded up to a
3,000 hour fixed length test.
The test time required for RGT2 is the difference between the test time
required for the “textbook approach” (at this growth rate) and the test time
64
for RGT1. The RGT2 time for MTBCF is 32,273 hours. If a growth rate of
0.35 can be maintained for MTBF for RGT1 and RGT2, the RGT2 for
MTBF would be approximately 18,000 hours (since the MTBCF
requirement drove a 3,000 hour RGT1). Therefore, the MTBCF RGT2
time would be the longer. However, a dedicated RGT in excess of 30,000
hours is not feasible.
65
There are several benefits to conducting the RGT in an operational
environment. The most obvious is the exact correlation between the test
hours and operational hours. Secondly, a more reasonable pace can be
maintained. During the OA, the three loaders tested averaged
approximately 5.0 hours per loader per day. A reasonable FRACAS
should be able to keep pace with the test, even if it includes several
loaders operated seven days a week. Finally, the additional cost to the Air
Force associated with accumulating test hours should be significantly less.
The cargo moved is cargo that would have to be moved in any case;
therefore, the cost of the operators, personnel to lock and unlock pallets,
fuel, etc., would have been spent even if the RGT were not performed.
66
By broadening the scope of the program to include maintainability and
T.O.s, the benefits of the RGT can be significantly increased, with no
appreciable degradation to the reliability growth process.
67
Case Study 2: Basic Expeditionary Airfield Resources
(BEAR)
Power Unit (BPU)
Reliability Feasibility Analysis
1. PURPOSE
2. BACKGROUND
3. METHODOLOGY
NPRD-95 contains the failure rate for one item that appears to be directly
comparable to the BPU engine-generator itself.
Generator,
Diesel,
31.4131 31,834 2-106
Packaged,
Continuous
Mean Time
Failure Rate
NPRD-95 Between
Description in PD (Failures/10E6 Page
Description Failure
Hours)
(hours)
Internal
Engine, Diesel
Combustion 14.2389 70,230 2-88
(Summary)
Engine
Heat
Engine Cooling Exchangers, 2-
7.8829 126,857
System Radiator 112
(Summary)
Brushless AC 2-
Generator, AC 0.7960 1,256,281
Generator 105
Voltage Regulator,
2-
Regulator/Exciter Voltage 5.5527 180,093
166
System (Summary)
2-
Cranking Motor Starter, Motor 0.0212 47,169,811
192
(assumed to be
Controls included in 0 N/A N/A
other items)
(included in
Governor 0 N/A N/A
Engine, Diesel)
(assumed to be
Other Devices negligible for 0 N/A N/A
initial analysis)
70
However, this margin will decrease as additional components are added to
increase the accuracy of the prediction.
The BPU is the replacement for the existing MEP-012. During the MEP-
012 First Article Test Endurance Run, two units were tested for 500 hours
each. Results are summarized in the following table:
71
Segment AW0001 AW0002 Total
Failures
0 – 100 hours 0 4 4
101 – 200 hours 1 1 2
201 – 500 hours 0 0 0
Total 1 5 6
MTBF (hours)
Observed 500 100 166.7
80 Percent 128.5 ≤ MTBF ≤ 53.5 ≤ MTBF ≤ 94.8 ≤ MTBF ≤
Confidence Level 441 158.5 257
37.8 ≤ MTBF ≤
1st 400 Hours N/A N/A
102.8
259.8 ≤ MTBF ≤
Last 600 Hours N/A N/A
5,677
Note that all six failures occurred in the first 400 hours of the test; two
failure modes were observed (fuse failures—four occurrences, and
improperly installed hose end on the fuel line—two occurrences). These
failure modes were identified, corrective actions were implemented, and
the units completed the remaining 600 hours with no further failures.
While the overall results were much less than the 1,500 hours MTBF
required for the BPU, the 600 hour failure-free period at the end of the test
indicates that the requirement may be feasible.
The required reliability is the lower test value, 500 hours MTBF in each
case. This is only 33.3 percent of that required by the RCM. According to
The Rome Laboratory Reliability Engineer’s Toolkit, p. 16, “A 10-20%
reliability improvement factor is reasonable for advancement of
technology.” This would suggest that a feasible reliability requirement for
the BPU would be 550 to 600 hours MTBF, and indicates that the 1,500
hour MTBF threshold in the RCM is unfeasible. Due to the age of the
MEP-012 and the technology advancements that have been made in diesel
engines, generators, control systems since that time, a 50% reliability
improvement factor may be reasonable.
To demonstrate the requirement in the RCM, 1,500 hours MTBF, the draft
PD requires the contractor test eight BPUs for a total of 13,950 hours with
a maximum allowable five (5) failures; this will demonstrate the MTBF at
90 percent one-sided confidence limits (ref. 4.6.20.2 of the draft BPU PD
73
and MIL-HDBK-781A Fixed-Duration Test Plans, Test Plan XV-D). This
will result in a minimum observed MTBF = 13,950/5 = 2,790 hours.
Comparing this 2,790 hours observed MTBF to the observed DPGDS
MTBF for RAII, 1,200 hours, and observed MTBF for the ODR, 374
hours, the requirement of 1,500 hours MTBF is not feasible. Refer to the
table below to see how the numbers change if the requirement in the RCM
is changed from 1,500 hours MTBF to the suggested 750 hours MTBF. At
the end of the RQT, with only 5 failures allowed, the minimum observed
MTBF equals 1,395 hours. The requirement of 750 hours MTBF is
feasible.
74
Percent of
Projected Feasibility and
Methodology RCM
MTBF (hours) Risk
Requirement
Generator Set Feasible, Low
31,834 2,122
Reliability Data Risk
Top-level parts
Feasible, Low
count reliability 35,098 2,340
Risk
prediction
DPGDS ODR Not Feasible,
374 24.9
results High Risk
166.7 (observed, Not Feasible,
11.1
entire test) High Risk
MEP-012 First 17.3 to 378 (80
Article Test 260 ≤ MTBF ≤ percent Feasible,
5,677 confidence Moderate Risk
level)
MEP-012/MEP-
012A Not Feasible,
500 33.3
Specification High Risk
Requirements
5. RECOMMENDATIONS
This analysis indicates that the RCM reliability requirement is not feasible
and would, therefore, result in a high risk of program failure. The BPU
reliability requirement should be reduced to a level that is feasible; 750
hours MTBF would appear to be an appropriate value, as this represents a
50 percent improvement over the MEP-012 requirements.
Note, however, that even a 750 hour MTBF requirement would not be low
risk. Both the DPGDS ODR results and the MEP-012 First Article Test
75
results indicate that achieving 750 hours MTBF will necessitate significant
engineering oversight. Recommend that the program include, at a
minimum:
Note that this feasibility analysis highlights a basic flaw with the DPGDS
design, which utilizes two engine-generators for each MEP-810 Prime
Power Unit (PPU). This doubles the predicted failure rate, and reduces the
predicted MTBF by half, for a PPU. Such a design would typically be
utilized to increase the mission (critical) reliability of the item, with
redundant components in parallel so that failure of one component would
not result in inability to complete the mission. However, in this
application, one engine-generator cannot supply sufficient power to
complete the mission; therefore, failure of either engine-generator will
result in mission failure. The DPGDS design suffers from the typical costs
of redundancy: increased cost, increased weight, increased package space,
increased complexity, and reduced basic (logistics) reliability, without
benefiting from the normal benefit of redundancy, increased mission
(critical) reliability.
76
Case Study 3: Fire Truck Depot Overhaul Study
This case study consists of the final report for Tracy Jenkin’s co-op project
from the summer of 2004. In a Mercer Engineering Research Center
(MERC) report entitled “Vehicle Depot Study,” dated 24 Aug 1998,
MERC provided the results and recommendations for a project:
83
“Vehicle Depot Study Final Engineering Report,” Contract No. F09603-
98-F-0019, Document No. 080600-98062-F1, 24 Aug 1998, Mercer
Engineering Research Center (MERC), 135 Osigian Boulevard, Warner
Robins, GA 31088.
84
Ibid., p. 18.
77
Table 13 below shows a comparison of the average cost to
operate the equipment on a per hour basis. This is probably
the most important data collected relating to maintenance
costs, since it includes the amount of utilization and the
cost of maintenance. The cost of maintenance in this data
includes a sum of the parts costs, labor costs and fuel costs.
All vehicle groups show marked improvement in FY97.
Overall, there is a 37% drop in cost per hour to operate the
vehicles.85
Since the MERC study only considered data from three vehicles of each
type one year prior and one year after depot overhaul, the data analyzed
was limited and the confidence in the results was low.
85
Ibid., p. 20.
78
Fire Truck Depot Overhaul Study
by Tracy Jenkins
Edited by Steven Davis
1. Abstract
2. Background
The intent of the overhaul is to increase the performance of the fire truck
fleet and extend the life expectancy of the current trucks. While in depot,
the truck is completely dismantled, defective parts are repaired or
replaced, and the vehicle is reassembled.
3. Methodology
This depot overhaul study was a continuation of the MERC task consisting
only of Air Force fire trucks. The vehicles used include the P-19 Aircraft
Rescue and Fire Fighting (ARFF) trucks and the P-22/P-24 structural
pumpers. These trucks have over ten years of service with the Air Force
and have been sent to depot within the last six years. Ten years of data was
collected from the Crash Rescue and ATAP depot programs. This data
includes 154 trucks from Crash Rescue and 34 trucks from ATAP. Crash
Rescue is the current depot contractor.
4. Results
The effects of performance before and after depot were analyzed. The data
were examined to determine trends between the two depot overhaul
programs and trends between the two types of trucks studied.
Table 1 shows the effect depot overhaul had on the 154 vehicles from
Crash Rescue. MTBF increased 20.2 percent, a significant improvement;
however, O&S cost increased 16.9 percent, a significant degradation. This
is an unexpected result, as a reduction in unscheduled maintenance
tasks—indicated by the increase in MTBF—should result in a
corresponding reduction in O&S costs. The 3.0 percent reduction in UR is
insignificant and does not explain the results. With conflicting results for
MTBF and O&S cost, the effectiveness of the Crash Rescue depot
overhaul is inconclusive.
O&S Cost
Per
MTBF UR VIC Rate
Operating
(hours) (%) (%)
Hour
($/hour)
Percent
20.2 % -3.0 % -3.0 % 16.9 %
Change
80
Table 2 shows the effects that the depot program had on the 34 ATAP
vehicles. MTBF increased 10.3 percent, again, a significant improvement.
However, O&S cost increased 23.2 percent, again, a significant
degradation. The 13.9 percent reduction in UR is significant, but does not
explain the results. With conflicting results for MTBF and O&S cost, the
effectiveness of the ATAP depot overhaul is also inconclusive.
O&S Cost
Per
MTBF UR VIC Rate
Operating
(hours) (%) (%)
Hour
($/hour)
Percent
10.3 % -13.9 % -0.4 % 23.2 %
Change
The data from both Crash Rescue and ATAP demonstrated similar results,
with a significant increase in MTBF along with a similar significant
increase in O&S costs. The effects of the depot overhaul, therefore, do not
appear to depend on the depot overhaul contractor.
The data were then analyzed according to ARFF and structural trucks to
determine if the depot overhaul program produced different results on
different types of vehicles.
Table 3 shows the effects that the depot program had on the P-19 ARFF
vehicles. MTBF increased 16.4 percent, again, a significant improvement.
However, O&S cost increased 2.6 percent, a slight degradation. The UR
81
did not change. With conflicting results for MTBF and O&S cost, the
effectiveness of the P-19 depot overhaul is also inconclusive.
O&S Cost
Per
MTBF UR VIC Rate
Operating
(hours) (%) (%)
Hour
($/hour)
Percent
16.4 % 0% 0.1 % 2.6 %
Change
Table 4 shows the effects that the depot program had on the P-22 and P-24
structural pumpers. MTBF decreased 4.7 percent, again, a slight
degradation. However, O&S cost decreased 12.2 percent, a significant
improvement. The UR decreased 16.7 percent; however, this does not
explain the results. With conflicting results for MTBF and O&S cost, the
effectiveness of the P-22/P-24 depot overhaul is also inconclusive.
82
Table 4: All P-22/P-24 (Crash Rescue and ATAP)
O&S Cost
Per
MTBF UR VIC Rate
Operating
(hours) (%) (%)
Hour
($/hour)
Percent
-4.7 % -16.7 % -0.4 % -12.2 %
Change
The data for the P-19 ARFF vehicles and P-22/P-24 structural pumpers
demonstrated contrasting results, with a significant increase in MTBF for
the P-19 versus a slight reduction in MTBF for the P-22/P-24, and a
significant increase in O&S costs for the P-19 compared to a significant
reduction in O&S costs for the P-22/P-24. Again, the O&S costs were
directly related to the MTBF when an inverse relationship was expected.
With conflicting results for MTBF and O&S cost, the effectiveness of the
fire truck depot overhaul is also inconclusive.
5. Conclusions
The direct relationship between MTBF and O&S costs was unexpected
and is unexplained. In three of the four cases analyzed, O&S costs
increased after the depot overhaul, which eliminates any possible
justification of depot overhaul on the basis of reducing overall life cycle
cost (LCC).
In summary, the depot overhaul essentially allows the user to keep the
truck in operational status for several more years.
83
Appendix 1: Developing a Textbook Reliability Program
3. The system engineer develops the contractual requirement from the user
requirement. Since it is impossible to test for true reliability, the system
engineer selects appropriate confidence levels and selects or develops an
appropriate reliability demonstration test plan. The contractual
requirement can be stated in a number of equivalent ways (for example,
stating any two of the values for lower test limit, upper test limit, and
discrimination ratio is equivalent to stating all three). In order to have a
“good chance” of satisfying the user requirement, the lower test limit
should be set equal to the user requirement. When the contractor passes
the selected RQT, the system engineer is (1 − β ) × 100% confident that the
items meet the user reliability requirement.
86
The Rome Laboratory Reliability Engineer’s Toolkit, op. cit., p. 16.
84
Appendix 2: Example R&M Requirements Paragraphs
3.13.2 Maintainability.
85
d. The BPU shall be designed so the correct oil and coolant levels
can be checked while the unit is running.
50 24
75 24
25 24
100 24
MTBFc
MTBFi = ,
1−α
where
T
MTBFc = ,
n
where
Failure purging, either the removal of a failure from the RGT tracking
process after the corrective action for that failure has been implemented
and its effectiveness verified or the removal of all but the first occurrence
of a failure mode, shall not be allowed.
Cumulative test time at any failure shall be calculated by adding the test
time of the failed BPU with that of each of the other BPUs as recorded on
their data logs prior to the time of the failure.
In the event that the instantaneous reliability is less that 90 percent of the
planned reliability, accumulation of RGT hours shall cease until the
contractor has incorporated sufficient corrective actions so that the
projected reliability is greater than the planned growth curve.
87
4.6.20.2 RQT. After successful completion of the RGT, a 6,975 hour
fixed-duration RQT shall be performed to demonstrate compliance with
3.13.1. Nominal consumer’s and producer’s risks shall be 10 percent; the
discrimination ratio shall be 3.0, and no more than five failures shall be
allowed. (Ref. Test Plan XV-D of MIL-HDBK-781A). Configuration
changes shall not be made during the RQT without approval of the
procuring activity.
88
Example Reliability Appendix from the Draft BPU PD
(PD05WRLEEG11, dated 23 Apr 2007)
APPENDIX B
RELIABILITY
B.1 SCOPE
B.3 DEFINITIONS
89
B.3.1 Discrimination ratio. (d) is one of the standard test plan parameters;
it is the ratio of the upper test MTBF (θ0) to the lower test MTBF (θ1) that
Θ
is, d = 0 . (Ref. MIL-HDBK-781A)
Θ1
B.3.2 Failure. The event, or inoperable state, in which any item or part of
an item does not, or would not, perform as previously specified. (Ref.
MIL-HDBK-470A)
B.3.8 MTBF, lower test. (θ1) is that value which is the minimum
acceptable. The standard test plans will reject, with high probability,
equipment with a true MTBF that approaches (θ1). The lower test MTBF
is the required MTBF. (Ref. MIL-HDBK-781A)
90
B.3.9 MTBF, upper test. (θ0) is an acceptable value of MTBF equal to the
discrimination ratio times the lower test MTBF (θ1). The standard test
plans will accept, with high probability, equipment with a true MTBF that
approaches (θ0). This value (θ0) should be realistically attainable, based on
experience and information. The upper test MTBF is also known as the
“design to” MTBF. (Ref. MIL-HDBK-781A)
B.4 REQUIREMENTS
91
B.4.1.1.1 Problem and failure reporting. A failure report shall be initiated
at the occurrence of each problem or failure of the contractor hardware or
software, and Government-furnished equipment (GFE). The report shall
contain the information required to permit determination of the origin and
correction of failures. The existing failure report forms may be used with
minimum changes necessary to conform to the requirements of this
specification and shall include the information specified in a through c:
B.4.1.1.6 Problem and failure tracking and closeout. The closed loop
failure reporting system shall include provisions for tracking problems,
failures, analyses, and corrective actions. Status of corrective actions for
all problems and failures shall be reviewed at scheduled test status
reviews. Problem and failure closeout shall be reviewed to assure their
adequacy.
93
B.5.1 Reliability test requirements. The reliability tests shall be conducted
in accordance with the reliability test procedures which have been
approved by the procuring activity. Testing shall be continued until a
reject decision has been reached or the total required test time has been
completed, whichever comes first.
94
Example Paragraphs from the Draft BPU SOW
(dated 15 Feb 2007)
3.6.2.1 Reliability.
95
basic reliability prediction shall be documented in the design analysis (see
3.6.4) and reviewed at the design reviews.
96
Appendix 3: Summary of χ2 Models87
Notes:
87
RADC Reliability Engineer’s Toolkit, op. cit., p. A-47.
97
Appendix 4: Fractiles of the χ2 Distribution88
88
Ibid., pp. A-48 – A-50. This table has been abridged to include only the
10% and 20% upper and lower confidence levels (those most commonly
used in reliability calculations) and to delete the odd-numbered degrees of
freedom, which are not used in confidence level calculations. It has been
expanded to include more degrees of freedom and more significant digits.
98
Degrees of Probability in Percent
Freedom (f)
10.0 20.0 80.0 90.0
42 30.765 34.157 49.456 54.090
44 32.487 35.974 51.639 56.369
46 34.215 37.795 53.818 58.641
48 35.949 39.621 55.993 60.907
50 37.689 41.449 58.164 63.167
52 39.433 43.281 60.332 65.422
54 41.183 45.117 62.496 67.673
56 42.937 46.955 64.658 69.919
58 44.696 48.797 66.816 72.160
60 46.459 50.641 68.972 74.397
62 48.226 52.487 71.125 76.630
64 49.996 54.337 73.276 78.860
66 51.770 56.188 75.425 81.085
68 53.548 58.042 77.571 83.308
70 55.329 59.898 79.715 85.527
72 57.113 61.756 81.857 87.743
74 58.900 63.616 83.997 89.956
76 60.690 65.478 86.135 92.166
78 62.483 67.341 88.271 94.374
80 64.278 69.207 90.405 96.578
82 66.076 71.074 92.538 98.780
84 67.876 72.943 94.669 100.98
86 69.679 74.813 96.799 103.18
88 71.484 76.685 98.927 105.37
90 73.291 78.558 101.05 107.57
100 82.358 87.945 111.67 118.50
1000 943.13 962.18 1037.4 1057.7
99
Appendix 5: Factors for Calculating Confidence Levels89
Factor
80%
Two-
Failures 80%
Sided 60% Two-Sided
Two-
90% 80% One-Sided
Sided
One-
Sided
Time
All Other Lower Lower Upper Upper
Terminated
Cases Limit Limit Limit Limit
Lower Limit
0 1 0.43429 0.62133 4.4814 9.4912
1 2 0.25709 0.33397 1.2130 1.8804
2 3 0.18789 0.23370 0.65145 0.90739
3 4 0.14968 0.18132 0.43539 0.57314
4 5 0.12510 0.14879 0.32367 0.41108
5 6 0.10782 0.12649 0.25617 0.31727
6 7 0.09495 0.11019 0.21125 0.25675
7 8 0.08496 0.09773 0.17934 0.21477
8 9 0.07695 0.08788 0.15556 0.18408
9 10 0.07039 0.07988 0.13719 0.16074
10 11 0.06491 0.07326 0.12259 0.14243
11 12 0.06025 0.06767 0.11073 0.12772
12 13 0.05624 0.06290 0.10091 0.11566
13 14 0.05275 0.05878 0.09264 0.10560
14 15 0.04968 0.05517 0.08560 0.09709
89
The Rome Laboratory Reliability Engineer’s Toolkit, op. cit., p. A-43. This table
has been adapted and abridged to include only the 10% and 20% upper and
lower confidence levels (those most commonly used in reliability
calculations). It has been expanded to include more failures and more
significant digits. Note that The Rome Laboratory Reliability Engineer’s Toolkit is
in the public domain; it can, therefore, be freely distributed.
100
Factor
80%
Two-
Failures 80%
Sided 60% Two-Sided
Two-
90% 80% One-Sided
Sided
One-
Sided
Time
All Other Lower Lower Upper Upper
Terminated
Cases Limit Limit Limit Limit
Lower Limit
15 16 0.04697 0.05199 0.07953 0.08980
16 17 0.04454 0.04917 0.07424 0.08350
17 18 0.04236 0.04664 0.06960 0.07799
18 19 0.04039 0.04437 0.06549 0.07314
19 20 0.03861 0.04231 0.06183 0.06885
20 21 0.03698 0.04044 0.05855 0.06501
21 22 0.03548 0.03873 0.05560 0.06156
22 23 0.03411 0.03716 0.05292 0.05845
23 24 0.03284 0.03572 0.05048 0.05563
24 25 0.03166 0.03439 0.04825 0.05307
25 26 0.03057 0.03315 0.04621 0.05072
26 27 0.02955 0.03200 0.04433 0.04856
27 28 0.02860 0.03093 0.04259 0.04658
28 29 0.02772 0.02993 0.04099 0.04475
29 30 0.02688 0.02900 0.03949 0.04305
30 31 0.02610 0.02812 0.03810 0.04147
31 32 0.02536 0.02729 0.03681 0.04000
32 33 0.02467 0.02652 0.03559 0.03863
33 34 0.02401 0.02578 0.03446 0.03735
34 35 0.02338 0.02509 0.03339 0.03615
39 40 0.02071 0.02212 0.02890 0.03111
49 50 0.01688 0.01791 0.02274 0.02428
499 500 0.00189 0.00193 0.00208 0.00212
101
Appendix 6: Redundancy Equation Approximations Summary90
With Repair Without Repair
All units
are active
on-line Equation 4
with equal Equation 1
unit λ
failure n!(λ ) q +1 λ( n − q ) / n =
λ( n − q ) / n = n
1
∑i
rates. (n-
q) out of (n − q − 1)!( µ ) q
n required i=n−q
for
success.
Two
active on-
line units
with
different Equation 2 Equation 5
failure
and repair λ A λ B [( µ A + µ B ) + (λ A + λ B )] λ A 2 λB + λ A λ B 2
rates. λ1 / 2 = λ1 / 2 = 2
One of ( µ A )( µ B ) + ( µ A + µ B )(λ A + λ B ) λ A + λ B + λ A λB
2
two
required
for
success.
One
standby
off-line
unit with
n active
on-line
units
required
for Equation 3 Equation 6
success.
n[nλ + (1 − P) µ ]λ
Off-line
nλ
spare
λn / n+1 = λn / n+1 =
assumed
to have a
µ + n( P + 1)λ P +1
failure
rate of
zero. On-
line units
have
equal
failure
rates.
Notes:
4. Assumes all units are functional at the start
5. The approximations represent time to first failure
6. CAUTION: Redundancy equations for repairable systems should
not be applied if delayed maintenance is used.
103
Appendix 7: Summary of MIL-HDBK-781A PRST Test
Plans 91
Test
Producer’s Consumer’s Discrim- Maximum
Test Duration
Risk (α) Risk (β) ination Failures
Plan (multiples
(%) (%) Ratio (d) to Accept
of θ1)
92 Ibid., p. 131.
105
Appendix 9: Glossary
106
Discrimination Ratio (d) – “One of the standard test plan parameters; it is
the ratio of the upper test MTBF (θ0) to the lower test MTBF (θ1) that is,
Θ
d = 0 .” (MIL-HDBK-781A, p. 6)
Θ1
Downtime – “That element of time during which an item is in an
operational inventory but is not in condition to perform its required
function.” (MIL-HDBK-338B, p. 3-5)
Duane Model – An RGT model that is based on the observation that a log-
log plot of the cumulative reliability versus cumulative test time will be a
straight line. It is named after its developer, J. T. Duane. (MIL-HDBK-
781A, p. 19)
Failure – “The event, or inoperable state, in which any item or part of an
item does not, or would not, perform as previously specified.” (MIL-
HDBK-470A, p. G-5)
Failure Purging – The removal of a failure from the RGT tracking process
after the corrective action for that failure has been implemented and its
effectiveness verified.
Failure Rate (λ(t)) – “The ratio of probability that failure occurs in the
interval, given that it has not occurred prior to t1, the start of the interval,
divided by the interval length.” (MIL-HDBK-338B, p. 5-2)
Failure Reporting And Corrective Action System (FRACAS) – “A closed
loop system … used to collect data on, analyze, and record timely
corrective action for all failures that occur during reliability tests. The
system should cover all test items, interfaces between test items, test
instrumentation, test facilities, test procedures, test personnel, and the
handling and operating instructions.” (MIL-HDBK-781A, p. 11)
Failure, Critical – “A failure or combination of failures that prevents an
item from performing a specified mission.” (MIL-HDBK-338B, p. 3-6)
Failure, Dependent – “A failure of one item caused by the failure of an
associated item(s). A failure that is not independent.” (MIL-HDBK-338B,
p. 3-6)
Failure, Non-Chargeable – A non-relevant failure; a failure induced by
Government furnished equipment (GFE); or a failure of parts having a
specified life expectancy and operated beyond the specified replacement
time of the parts (e.g., wear out of a tire when it has exceeded its specified
life expectancy). (Based upon MIL-STD-721C, p. 4)
107
Failure, Non-Relevant – A failure caused by installation damage; accident
or mishandling; failures of the test facility or test-peculiar instrumentation;
caused by an externally applied overstress condition, in excess of the
approved test requirements; normal operating adjustments (non-failures)
specified in the approved technical orders; dependent failures within the
test sample, which are directly caused by non-relevant or relevant primary
failures; or caused by human errors. (Based upon MIL-STD-721C, p. 4)
Failure, Relevant, Chargeable – Any failure other than a non-chargeable
failure.
Failure, Secondary – Another term for dependent failure.
Hazard Rate (h(t)) – “The limit of the failure rate as the interval length
approaches zero.” Also known as the instantaneous failure rate. (MIL-
HDBK-338B, p. 5-2)
Life Cycle Cost (LCC) – The sum of acquisition, logistics support,
operating, and retirement and phase-out expenses. (MIL-HDBK-470A, pp.
G-8)
Lower Test MTBF (θ1) – The lowest value of MTBF which is acceptable.
“The standard test plans will reject, with high probability, equipment with
a true MTBF that approaches (θ1).” (derived from MIL-HDBK-781A, p.
7)
Maintainability – “The relative ease and economy of time and resources
with which an item can be retained in, or restored to, a specified condition
when maintenance is performed by personnel having specified skill levels,
using prescribed procedures and resources, at each prescribed level of
maintenance and repair.” (MIL-HDBK-470A, p. G-8, definition (1))
Maintenance Action – “An element of a maintenance event. One or more
tasks (i.e., fault localization, fault isolation, servicing, and inspection)
necessary to retain an item in or restore it to a specified condition.” (MIL-
HDBK-470A, p. G-9)
Maintenance Event – “One or more maintenance actions required to effect
corrective and preventative maintenance due to any type of failure or
malfunction, false alarm, or scheduled maintenance plan.” (MIL-HDBK-
470A, p. G-9)
Mean Downtime (MDT) – The average time during which an item is in an
operational inventory but is not in condition to perform its required
function. (derived from MIL-HDBK-338B, p. 3-5)
Mean Repair Time (MRT) – Another term for Mean Time To Repair
(MTTR).
Mean Time Between Critical Failure (MTBCF) – A measure of mission
(critical) reliability.
108
Mean Time Between Failure (MTBF) – “A basic measure of reliability for
repairable items. The mean number of life units during which all parts of
the item perform within their specified limits, during a particular
measurement interval under stated conditions.” (MIL-HDBK-470A, p.
G-11)
Mean Time To Failure (MTTF) – “A basic measure of reliability for non-
repairable items. The total number of life units of an item population
divided by the number of failures within that population, during a
particular measurement interval under stated conditions.” (MIL-HDBK-
338B, p. 3-12)
Mean Time To Repair (MTTR) – “The sum of corrective maintenance
times at any specific level of repair, divided by the total number of failures
within an item repaired at that level during a particular interval under
stated conditions.” (MIL-HDBK-470A, p. G-11)
Mean Time To Restore System (MTTRS) – “A measure of the product
maintainability parameter, related to availability and readiness: The total
corrective maintenance time, associated with downing events, divided by
the total number of downing events, during a stated period of time.
(Excludes time for off-product maintenance and repair of detached
components.)” (MIL-HDBK-338B, p. 3-13)
Predicted MTBF (θp) – “That value of MTBF determined by reliability
prediction methods; it is a function of the equipment design and the use
environment. (θp) should be equal to or greater than (θ0) in value, to ensure
with high probability, that the equipment will be accepted during the
reliability qualification test.” (MIL-HDBK-781A, p. 7)
Preventive Maintenance (PM) – “All actions performed in an attempt to
retain an item in specified condition by providing systematic inspection,
detection, and prevention of incipient failures.” (MIL-HDBK-470A, p. G-
14)
Producer’s Risk (α) – “The probability of rejecting equipment which has a
true MTBF equal to the upper test MTBF (θ0). The probability of rejecting
equipment with a true MTBF greater than the upper test MTBF will be
less than (α).” (MIL-HDBK-781A, p. 6)
Redundancy – “The existence of more than one means for accomplishing
a given function. Each means of accomplishing the function need not
necessarily be identical. The two basic types of redundancy are active and
standby.” (MIL-HDBK-338B, p. 3-16)
Redundancy, Active – “Redundancy in which all redundant items operate
simultaneously.” (MIL-HDBK-338B, p. 3-16)
109
Redundancy, Standby – “Redundancy in which some or all of the
redundant items are not operating continuously but are activated only upon
failure of the primary item performing the function(s).” (MIL-HDBK-
338B, p. 3-16)
Reliability – “The probability that an item can perform its intended
function for a specified interval under stated conditions.” (MIL-HDBK-
470A, p. G-15, definition (2))
Reliability Demonstration Test (RDT) – Another term for Reliability
Qualification Test (RQT).
Reliability Growth Test (RGT) – “A series of tests conducted to disclose
deficiencies and to verify that corrective actions will prevent recurrence in
the operational inventory. (Also known as “TAAF” testing.)” (MIL-STD-
785B, p. 3, definition for “Reliability development/growth test (RD/GT)”)
Reliability Qualification Test (RQT) – “A test conducted under specified
conditions, by, or on behalf of, the government, using items representative
of the approved production configuration, to determine compliance with
specified reliability requirements as a basis for production approval. (Also
known as a “Reliability Demonstration,” or “Design Approval” test.)”
(MIL-STD-785B, p. 3, definition for “Reliability qualification test
(RQT)”)
Reliability, Basic – “Measure of system’s ability to operate without
logistics support.” (Rome Laboratory Reliability Engineer’s Toolkit, p. 11)
Reliability, Critical – Another term for mission reliability.
Reliability, Logistics – Another term for basic reliability. (Rome
Laboratory Reliability Engineer’s Toolkit, p. 11)
Reliability, Mission – “Measure of system’s ability to complete mission.
Consider[s] only failures that cause [or would cause] a mission abort.”
(Rome Laboratory Reliability Engineer’s Toolkit, p. 11)
Reliability, Observed – “A point estimate of reliability equal to the
probability of survival for a specified operating time, t, given that the
equipment was operational at the beginning of the period.” (MIL-HDBK-
781A, p. 7)
Repair Time – Another term for Corrective Maintenance Time.
Repairable Item – “An item which, when failed, can be restored by
corrective maintenance to an operable state in which it can perform all
required functions.” (MIL-HDBK-338B, p. 3-17)
Scheduled Maintenance – “Periodic prescribed inspection and/or servicing
of products or items accomplished on a calendar, mileage, or hours
operation basis.” (MIL-HDBK-470A, p. G-15)
110
Test Analyze And Fix (TAAF) – Another term for Reliability Growth Test
(RGT).
Time, Active – “That time during which an item is in an operational
inventory.” (MIL-HDBK-470A, p. G-1)
Time, Standby – Time during which an item is ready to operate, but not in
operation.
Unscheduled Maintenance – “Corrective maintenance performed in
response to a suspected failure.” (MIL-HDBK-470A, p. G-17)
Upper Test MTBF (θ0) – “An acceptable value of MTBF equal to the
discrimination ratio times the lower test MTBF (θ1). The standard test
plans will accept, with high probability, equipment with a true MTBF that
approaches (θ0). This value (θ0) should be realistically attainable, based on
experience and information.” (MIL-HDBK-781A, p. 7)
Uptime – “Hours that product is in customer’s possession and works”
(Reliability Toolkit: Commercial Practices Edition, p. 11)
Utilization Rate (UR) – “The planned or actual number of life units
expended, or missions attempted during a stated interval of calendar time.”
(MIL-HDBK-470A, p. G-17)
111