131

Australasian Physical & Engineering Sciences in Medicine Volume 28 Number 2, 2005
EDUCATIONAL NOTE
A standard approach to measurement uncertainties for
scientists and engineers in medicine*
K. Gregory
1,2
, G. Bibbo
1,3
and J. E. Pattison
1
1
School of Electrical and Information Engineering (Applied Physics), University of South Australia, Mawson
Lakes, Australia
2
Radiation Protection Division, Environment Protection Authority, Kent Town, Australia
3
Division of Medical Imaging, Women’s and Children’s Hospital, North Adelaide, Australia
Abstract
The critical nature of health care demands high performance levels from medical equipment. To ensure these
performance levels are maintained, medical physicists and biomedical engineers conduct a range of measurements on
equipment during acceptance testing and on-going quality assurance programs. Wherever there are measurements, there
are measurement uncertainties with potential conflicts between the measurements made by installers, owners and
occasionally regulators. Prior to 1993, various methods were used to calculate and report measurement uncertainties. In
1993, the International Organization for Standardization published the Guide to the Expression of Uncertainty in
Measurement (GUM). The document was jointly published with six international organizations principally involved in
measurements and standards. The GUM is regarded as an international benchmark on how measurement uncertainty
should be calculated and reported. Despite the critical nature of these measurements, there has not been widespread use
of the GUM by medical physicists and biomedical engineers. This may be due to the complexity of the GUM. Some
organisations have published guidance on the GUM tailored to specific measurement disciplines. This paper presents
the philosophy behind the GUM, and demonstrates, with a medical physics measurement example, how the GUM
recommends uncertainties be calculated and reported.
Key words uncertainty, error, accuracy, GUM,
measurement
Introduction
The performance tolerances of medical equipment are
generally more stringent than domestic and industrial
equipment due to the critical nature of health care. In the
case of medical X-ray units where patients are exposed to
ionising radiation, the performance levels may be set by
legal instruments, such as is done in South Australia
1
, or by
published recommendations
2
. Other medical equipment,
such as those relating to the administration of drugs
(eg balances and flow rate meters), and patient monitoring
*Presented in part at the 29th annual conference of the
Australasian Radiation Protection Society, Adelaide, SA,
Australia, 24-27 October 2004.
Corresponding author: Kent Gregory, Radiation Protection
Division - EPA, PO Box 721, Kent Town, SA 5071, Australia
Tel: 08 8130 0713, Fax: 08 8130 0777
Email: kent.gregory@state.sa.gov.au
Received: 14 February 2005; Accepted: 14 March 2005
Copyright © 2005 ACPSEM/EA
(eg thermometers and blood analysers), also need to
maintain high performance levels. There are examples in
the literature showing that the health care industry makes
widespread use of quality assurance (QA) programs
3-5
to
monitor performance of equipment. These QA programs
usually include a variety of measurements that are
conducted at regular intervals. Performance measurements
may also be conducted on new equipment as part of
acceptance testing. As medical equipment is subject to so
many measurements, there is potential for conflicts to arise
between the various stakeholders, such as installers,
operators, purchasers, manufacturers and regulators. Any
measurements involved in such conflicts are likely to be
closely scrutinised, and the question of measurement
uncertainty may arise.
Prior to 1993, there was no international consensus on
how measurement uncertainties were to be calculated and
reported
6
. The variation in philosophies and calculation
methods caused three main problems, especially for
laboratories conducting comparisons with other laboratories
in different countries
6
. The first problem was that
laboratories had to give extensive explanations of the
measurement uncertainty calculation method they used
whenever results were reported, so that other laboratories
could make comparisons. The second problem was that it
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
132
was common to keep random and systematic uncertainties
separated, both during calculations and in the final
reporting, even though they were both uncertainties. The
final problem was that some uncertainties were deliberately
overstated in the belief that this was acting conservatively.
Such practise has the capacity to increase type I and type II
errors
7
in hypothesis testing (a type I error is when a good
result is erroneously rejected, while a type II error is when a
poor result is erroneously accepted). A desire to overcome
these three problems was the driver for what eventually
became the Guide to the Expression of Uncertainty in
Measurement
6
(GUM).
In 1993, the GUM was jointly published by the
International Organization for Standardization and six
international organisations
a
primarily concerned with
measurements and standards. The document is therefore
recognised by organisations internationally, including the
Australian National Measurement Institute (NMI). NMI
holds most of Australia’s First Level Standards, with the
exception of those relating to ionising radiation, which NMI
has delegated to the Australian Nuclear Science and
Technology Organisation and the Australian Radiation
Protection and Nuclear Safety Agency.
The GUM is underpinned with an extensive
mathematical basis and has the capacity to deal with most
measurement uncertainty problems. This generalist
approach comes at a cost; the GUM concepts and
methodologies are non-trivial, and may be difficult to
comprehend without continuous involvement in
measurement science at the highest levels (e.g. working at
institutions such as NMI). Some organisations
8-10
have
published specific guidance on the GUM for various
measurement disciplines. Others
11,12
have published
explanatory texts on the GUM in general. In particular, the
National Measurement Laboratory (part of NMI) has
published a document
12
(the NMI Guide), which explains in
simple terms the philosophy, methods and reporting style of
the GUM for use in applied environments. The NMI Guide
also offers a number of pragmatic approaches that are in
keeping with the philosophy of the GUM, but substantially
reduce the complexity of the calculations.
The NMI Guide is still somewhat complex and requires
some introduction and further explanation before engineers
and scientists adopt the GUM more widely. Indeed, there
appears to be very little application of the GUM by
scientists and engineers in medicine within published
literature. These workers may be making use of simpler
methods
13,14
for dealing with uncertainties that were
introduced to them during their undergraduate university
courses.
This paper presents a simplified theoretical
background, and uses a medical physics example to
illustrate how to apply the GUM to measurements made by
scientists and engineers in medicine.
a
International Bureau of Weights and Measures, International
Electrotechnical Commission, International Federation of
Clinical Chemistry, International Union of Pure and Applied
Chemistry, International Union of Pure and Applied Physics, and
the International Organization of Legal Metrology.
Theoretical background
Review of statistics
All measurements have an uncertainty. If a large
number of measurements are taken and each measurement
value is plotted against the number of times a measurement
value occurs (i.e. a histogram), the end product would be
the measurement’s uncertainty distribution. Uncertainty
distributions are typically scaled (standardised) so that the
area under a curve defining the histogram is equal to 1. This
is achieved by changing the x-axis to represent the
difference between each measurement value and the mean
value in terms of the number of standard deviations, and
changing the y-axis to represent probability density. Thus,
the probability of a measurement lying between two values
is equal to the area under the curve between these two
points. Most scientists and engineers will be familiar with
the Normal distribution, also known as the Gaussian
distribution and depicted in figure 1(a), which has a bell-
shaped curve. The Normal distribution is commonly
encountered in many types of measurements. The reason
why the Normal distribution is so common is explained by
the Central Limit Theorem. The ubiquitous nature of
Normal distributions is utilised by the GUM, and is
discussed later. Another uncertainty distribution is the
rectangular distribution as depicted in figure 1(b), which
applies to measurements that are rounded or truncated (such
as by a digital display). Occasionally, some measurements
are encountered that have uncertainty distributions that are
neither Normal nor rectangular, such as those depicted in
figures 1(c) and 1(d).
The aim of uncertainty analysis is to estimate the
uncertainty distribution for a final measurement result. The
final result’s uncertainty distribution is entirely dependent
on the uncertainty distributions of all factors that influence
Figure 1. Uncertainty distributions are like frequency histograms,
except they are standardised so that the area under an uncertainty
distribution curve is equal to 1. This is achieved by changing the
x-axis to represent the difference between measurement values
and the mean in terms of the number of standard deviations
(instead of measurement values), and changing the y-axis to
represent probability density (instead of frequency). Uncertainty
distributions can take any form, but the most common is the
Normal distribution (a). The rectangular distribution (b) is
associated with rounding or truncation of readings. Any
uncertainty distribution, including (c) and (d), may be assigned to
an uncertainty component.
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
133
the final result. This may include other measured quantities
(e.g. the measurement of mass and volume to determine
density), rounding of readings, or variations in atmospheric
or other conditions. Note that all factors that influence the
final result are collectively referred to as uncertainty
components (a term derived from the NMI Guide). Thus a
major task in uncertainty analysis is to characterise the
uncertainty distributions of the uncertainty components, so
that they can be combined.
In order to characterise an uncertainty distribution
exactly, the entire population of measurements would need
to be plotted as a histogram. As this is not possible for
infinite populations of measurements, a finite number of
measurements are sampled instead. From this sample,
inferences can be made about the infinite population. In the
case of the common Normal distribution of measurements,
the distribution of a finite sample of such measurements is
represented by a t-distribution (also called the Student’s
t-distribution). A t-distribution is virtually identical to a
Normal distribution when a sufficiently large number of
measurements are sampled (such as more than 20
measurements). However, if fewer measurements are
sampled, the shape of the t-distribution is flatter and
broader (see figure 2). This shape change reflects the fact
that with fewer measurements, less is known about the
infinite population. The shape of the t-distribution is
dependent on the degrees of freedom, ν. The value ν equals
the number of measurements sampled minus the number of
quantities calculated from the measurements. In most cases,
the measurements are used to calculate a mean, so ν will
equal one less than the number of measurements sampled.
For rectangular distributions, measurement sampling is
usually not required to ascertain the distribution shape.
Instead, its shape is obtained intuitively; the distribution is
flat with upper and lower bounds defined by, say, the
resolution of the instrument.
Figure 2. A comparison of the Normal distribution and t-
distributions with 1, 3 and 10 degrees of freedom. When the
degrees of freedom equals 20 or more, the t-distribution and the
Normal distribution may be considered equivalent for the
purposes of calculating uncertainties.
The uncertainty distribution of the final result is in
most cases, a known shape, due to the Central Limit
Theorem (CLT). The CLT states that when uncertainty
distributions are combined, the resulting distribution
approximates a Normal distribution. Knowing the shape of
the final result’s uncertainty distribution reduces the
amount of information needed regarding the distributions of
all the uncertainty components. In fact, only three quantities
for each uncertainty component are needed by the GUM to
estimate the uncertainty distribution of the final result.
The first quantity used to describe an uncertainty
distribution is the standard uncertainty, u (or estimated
standard deviation). This is a measure of the spread of the
distribution, with small u values indicative of repeatable
measurements (i.e. close agreement of successive
measurements). The second is the degrees of freedom, ν.
This quantity reflects the reliability of the value u. For
example, a value for u derived from many measurements
should have greater influence on the final result than
another u derived from a smaller number of measurements.
Thus, the higher the value of ν, the more reliable the value
of u. The third quantity is the sensitivity coefficient, c.
Values for c are used to take into account the relative
sensitivity of the final result to small changes in uncertainty
components. Thus if c = 1, an increase or decrease in the
value of an uncertainty component causes an equivalent
change in the final result. When c > 1 or c < 1, the
uncertainty component has, respectively, a greater and
lesser effect on the final result.
Background to the GUM and NMI Guide
The GUM and NMI Guide use methods and
terminology that may be unfamiliar to users of other
uncertainty analysis methodologies. For example, the term
measurand, which describes the quantity or parameter
subject to measurement, is used extensively in the GUM
and NMI Guide. While a detailed justification of why
certain procedures are used is documented in the GUM,
together with derivations of the equations, a brief
background is provided here, together with essential
terminology.
It is important to note that the GUM is not completely
regimented; there remain steps in the process requiring
professional judgment. As such, uncertainty analysis
remains an inexact science, since professional judgment is
subjective. This leads to uncertainty in the uncertainty
estimates. The NMI Guide plays on this fact, and proposes
a number of pragmatic approaches to the GUM process that
theoretically increase the uncertainty in the uncertainty
estimate. However, the NMI Guide argues that in practice
these increases are insignificant compared to the
uncertainty introduced through professional judgment.
There are four main steps in the GUM process, and
these are described in more detail below.
Step 1: Modelling the measurement
The first step in the process is to make what the GUM
refers to as a model of the measurement. This is a simple
task that involves writing an equation that relates the
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
134
(a)
(b)
Figure 3. (a) For a rectangular uncertainty distribution, the value
a is selected such that the range ±a includes the true value with
100% confidence; (b) For the Normal distribution, the value a is
selected such that the range ±a includes the true value with
approximately 95% probability. The estimated standard
uncertainty is half the value of a. If a very conservative value is
chosen (a′ ) such that the range ±a′ almost certainly includes the
true value (nearly 100%), half the value of a′ will be a poor
estimate of the standard uncertainty.
various uncertainty components to the final result (see
equation 9 in the example). Modelling the measurement
serves two purposes; it helps to identify uncertainty
components, and it assists in finding values of c for the
uncertainty components.
Step 2: Calculating values for u, ν and c
The GUM process involves characterising all
uncertainty components with these 3 descriptors; u, ν and c.
Values for the standard uncertainty u and the degrees of
freedom ν are calculated or assigned following what the
GUM refers to as either a Type A or Type B evaluation. It
may be helpful to think of Type A and Type B evaluations
as those uncertainties commonly described as ‘random’
13,14
and ‘systematic’
13,14
respectively (although this comparison
is not strictly true). In Type A evaluations, the value u is
equal to the estimated standard deviation of the population
of measurements, while ν is equal to its degrees of freedom.
Type B evaluations are performed when Type A
evaluations cannot be performed, such as when there is
only one value for an uncertainty component and additional
measurements cannot be obtained (e.g. a single value from
a table of reference data). With only a single value,
statistics cannot be used to calculate u and ν. Hence, u and
ν are derived using non-statistical approaches (i.e.
professional judgment), some of which have been
simplified by the NMI Guide. The first step in deriving u
and ν in a Type B evaluation is to assign an uncertainty
distribution representative of how the value of the
uncertainty component might vary if more values were
available. The GUM proposes that one of the four
distributions in figure 1 be assigned, with the triangular and
trapezoidal distributions assigned when there is insufficient
confidence to suggest the data do not follow a Normal
distribution, but sufficient information to suggest more
certainty than a rectangular distribution. The NMI Guide
suggests that uncertainty components, such as rounding or
truncation of readings, be assigned a rectangular
distribution, while almost all others may be assigned a
Normal distribution. The NMI Guide suggests that
triangular and trapezoidal distributions are rarely
encountered in the real world, and that more information
should be obtained before assigning such uncertainty
distributions to any measurement.
Having assigned an uncertainty distribution, the next
step is to estimate the range, ±a, (figure 3a and 3b) within
which the true value lies, and the likelihood that the true
value is within this range. If a rectangular distribution was
assigned, a value for a can be chosen such that the range ±a
represents exactly 100% of the uncertainty distribution. The
standard uncertainty can then be calculated;
3
a
u =
for a rectangular distribution (figure 3a) (1)
If a Normal distribution was assigned, the NMI Guide
suggests choosing a value for a such that the range ±a
represents 95% of the measurements. The standard
uncertainty can then be calculated;
2
a
u =
for a Normal distribution (see figure 3b) (2)
It is important not to deliberately increase a beyond
what is necessary, such as a′ in figure 3b. By deliberately
inflating the value a, the uncertainty calculated for the final
result may be unnecessarily large.
In order to assign values of ν, the NMI Guide argues that
the final result is not particularly sensitive to the ν values
chosen. If that is the case, then it is sufficient to categorise
each evaluation broadly as either poor, reasonable, good or
excellent (see table 1), and to assign the corresponding
value for ν. (Note that the third column in table 1, k, is
explained in Step 4.)
Values of the sensitivity coefficient c may be obtained
in one of three ways; (i) by making a small change in the
value of the uncertainty component in the model of the
measurement and calculating the consequent change in the
final result, (ii) experimentally changing the value of the
uncertainty component slightly and observing the
consequent change in the final result, or (iii) using calculus
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
135
Table 1. Relationship between measurement quality (as judged
through experience), degrees of freedom (ν) and coverage factor
(k) to generate a 95% confidence interval.
Measurement
quality
ν
k for 95% confidence
interval
Poor 2 4.30
Reasonable 8 2.31
Good 20 2.09
Excellent 1000 1.96
to differentiate the model of the measurement with respect
to the uncertainty component (this is a common uncertainty
analysis technique
14
). While it may not be possible to use
all three methods to calculate each value of c, at least one
method will be.
Step 3: Calculation of u
comb
and ν
eff
Using the u, ν and c values, equations are then used to
calculate the combined standard uncertainty, u
comb
and the
effective degrees of freedom of the final result, ν
eff
.
Given two conditions specified below, the equation to
calculate u
comb
is the square root of the sum of the squares
of the standard uncertainties, weighted with the respective
sensitivity coefficients. Thus for a final result with N
uncertainty components, the combined standard uncertainty
of the final result is given by:

=
=
N
i
i i comb
u c u
1
2
) (
(3)
where u
i
is the i
th
standard uncertainty, and c
i
is the i
th
sensitivity coefficient.
Equation (3) is valid if two conditions are met; (a),
each uncertainty component is not correlated with any other
uncertainty component, and (b), an approximately linear
relationship exists between each uncertainty component and
the final result for the range of values the uncertainty
components are likely to have. For most measurements, if
both the above conditions are not met, it is usually possible
to alter the way the measurement is made so that both
conditions are satisfied. If not, the mathematics becomes
more complex.
With regard to condition (a), the degree of correlation
is determined by calculating the correlation coefficient,
which can have any value from –1 (anti-correlated) to +1
(fully correlated). Values close to 0 are indicative of
uncorrelated uncertainty components. To account for
correlation, equation (3) becomes:
( )

=
+ =
N
i
i i
comb u c u
1
2 2
( )
∑ ∑

= + =
1
1 1
2
N
i
N
i j
j i j i j i
x x r u u c c
(4)
where r(x
i
x
j
) is the correlation coefficient of the two
uncertainty components x
i
and x
j
.
In the special case where all uncertainty components
are fully correlated, i.e. r(x
i
x
j
) =1, equation (4) reduces to;

=
=
N
i
i i comb
u c u
1
) (
(5)
Note that equation (5) is simply the linear sum of the
weighted standard uncertainties of the uncertainty
components. If any uncertainty components happen to be
correlated, it may be possible to alter the measurement
procedure to obtain uncorrelated uncertainty components,
and thus retain the simpler equation (3). However,
correlation may be a desirable effect, especially since anti-
correlated uncertainties can reduce the combined standard
uncertainty. The NMI Guide provides a pragmatic approach
to correlations to reduce the mathematical process.
With regard to condition (b), equations (3) and (4) are
the first terms in a Taylor series expansion, and are good
approximations if an approximately linear relationship
exists between the final result and each of the uncertainty
components. If this is not the case, the next term in the
Taylor series will need to be added to equations (3) and (4);
2 2
2
3
2
2
1 1
2
1
j i
j i i j i
N
j
N
i
u u
x x
f
x
f
x x
f

∂ ∂



+
|
|
.
|

\
|
∂ ∂

∑ ∑
= =
(6)
where f is the model of the measurement with N
uncertainty components.
Non-linearity may occur when the value of the final
result is very small compared to the value of the uncertainty
components (such as calculating the area of a small
rectangle by measuring two sides using a ruler with a large
uncertainty), or the uncertainty components vary the final
result in a non-linear manner (such as an uncertainty
component raised to a power in the model of the
measurement). The GUM provides more detail on non-
linearity.
As mentioned earlier, the CLT suggests that the final
result’s uncertainty distribution is a Normal distribution,
but this is only an approximation. However, with the
combined standard uncertainty, u
comb
, and the effective
degrees of freedom, ν
eff
, obtained from the Welch-
Satterthwaite formula
6
;

=

=
N
i
i
i i
comb
eff
u c
u
1
4
4
) (
ν
ν
(7)
the final result’s uncertainty distribution can be treated
mathematically like a t-distribution. In practice, ν
eff
will
typically have a value greater than 20.
Step 4: State the final result
The final result’s uncertainty may be expressed simply
as ±u
comb
, which represents a range of values within which
the true value is expected to lie with approximately 68%
probability (assuming ν
eff
≥20). This is the traditional one
standard deviation or one sigma level. However, the GUM
reports the expanded uncertainty, U, calculated using
equation (8);
U = u
comb
×k (8)
where U is the expanded uncertainty, and k is the coverage
factor.
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
136
Table 2. An uncertainty analysis table. Each row represents a different uncertainty component, with values for
standard uncertainty (u), degrees of freedom (ν) and sensitivity factor (c).
Uncertainty components u ν c
Reading 0.0127 mGy 24 1.017 mGy/mGy
Rounding 0.000289 mGy 1000 1.017 mGy/mGy
Cal. Factor 0.025 units 20 1.953 mGy/unit
Temperature 0.50 K 20 0.00668 mGy/K
Pressure 50 Pa 20 -0.0000196 mGy/Pa
Position (vertical plane) 2 mm 20 0.00592 mGy/mm
Position (anode-cathode plane) 2 mm 20 0.00272 mGy/mm
Position (lateral plane) 2 mm 20 0 mGy/mm
The combined standard uncertainty
and effective degrees of freedom:
u
comb
= 0.0523 mGy
Values for U can be calculated such that ±U has any
desired probability of containing the true value, including
the increasingly common
12
95% probability. The value k is
selected from tabulated data
6,12
, given ν
eff
and the desired
probability (known as confidence interval, CI). Some
typical values of k appear in table 1.
Both the final result and U need to be rounded such that
three objectives are met; namely U contains only significant
digits, the final result and U are rounded to the same least
significant digit, and the rounding process itself does not
introduce a significant uncertainty. Rounding is described
in more detail in the NMI Guide, but the basic steps appear
below.
1. Round the final result and U once only.
2. Round U to 1 significant digit if the first significant
digit is 5 or more, otherwise round to 2 significant
digits. (eg 0.0078 becomes 0.008, while 0.013 is
unchanged).
3. Round U upwards unless rounding down will only
change U by a few percent (eg 0.0078 becomes
0.008, while 0.0071 becomes 0.007)
4. Round the final result to match the significant digits
in U.
5. Round a 5 in the final result to the nearest even
number of the next significant digit (eg 10.65
becomes 10.6, while 10.75 becomes 10.8).
Summary of the GUM method
The GUM process is to;
Step 1: Make a model of the measurement. For
measurements with many uncertainty
components, it may become difficult to manage
the model. In such cases, it may be appropriate
to separate the model into smaller parts, and
analyse these separately.
Step 2: Identify and characterise each uncertainty
component. The NMI Guide suggests creating
an uncertainty analysis table for recording the
results of this step, such as table 2.
Step 3: Calculate the values of u
comb
and ν
eff
. For N
uncorrelated uncertainty components, the
values of u
comb
and ν
eff
are given by equations
(3) and (7), respectively.
Step 4: State the final result. The statement should
mention that the uncertainty was calculated in
accordance with the GUM, give the values of
U and k, indicate the CI used, and ensure the
final result and U are rounded appropriately.
Example
Consider the measurement of radiation dose from a
mammographic X-ray unit with exposure settings of 28
kVp, 20 mAs, with large focus selected, molybdenum
anode, 30 µm molybdenum filtration, at a position 40 mm
from the chest wall in the anode-cathode plane, 45 mm
above the imaging plate in the vertical plane, and centred
on the remaining axis (i.e. centred laterally with respect to
the patient).
All data for the example are summarised in table 2.
Step 1. Make a model of the measurement
Suppose that a calibrated, vented ionisation chamber is used
to measure the dose. The temperature and pressure of the
ambient atmosphere are also measured and used to
compensate the ionisation chamber. Given this information,
the measurement model is:

×

× × =
pressure
e temperatur
factor cal reading result final
101325
15 . 295
.
(9)
where: final result is the dose to be determined, reading is
the value displayed by the electrometer, cal. factor is the
calibration factor for the ionisation chamber and
electrometer (a unitless value), temperature is the
temperature inside the ionisation chamber at the time of
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
137
measurement (in K), pressure is the pressure inside the
ionisation chamber at the time of measurement (in Pa).
The final result is 1.97254 mGy, based on a reading of
1.940 mGy, a calibration factor of 1.01, pressure of 100 650
Pa, and temperature of 295.15 K.
Step 2. Identify and characterise each uncertainty
component
Four uncertainty components were identified when
modelling the measurement and appear in equation 9. Four
more uncertainty components have been identified;
rounding of the reading by the display, and variation of the
chamber’s position from the desired position in the three
spatial planes.
Calculation of c
For the purpose of demonstration, values of c have
been acquired using the three different methods for three
groups of uncertainty components.
(i) For the reading, the rounding of the displayed
reading, and the calibration factor, the differentiation
method is used. Differentiating the model with respect to
‘reading’:

×

× =


=
pressure
e temperatur
factor cal
reading
result
c
101325
15 . 295
.
) (
) (
c = 1.017 calculated using measured values for the
other uncertainty components.
For rounding, as the rounding of the reading is
essentially a change in the value of the reading, again,
c = 1.017.
For the calibration factor,

×

× =


=
pressure
e temperatur
reading
factor cal
result
c
101325
15 . 295 ) . (
) (
c = 1.953 mGy/unit
(ii) For the temperature and pressure, values for c are
derived by changing the value of these uncertainty
components by a small amount in the model equation (9),
and observing the change in the final result.
Increasing the temperature by 1 K to 296.15 K
increases the final result by 0.00668 mGy to 1.97922 mGy.
Thus
00668 . 0
1
00668 . 0
= =


=
e temperatur
result
c mGy/K
Similarly, increasing the pressure by 100 Pa to 100750
Pa decreases the final result by 0.00196 mGy to 1.97058
mGy. Thus
0000196 . 0
100
00196 . 0
− =

=


=
pressure
result
c mGy/Pa
(iii) For positioning, values for c were obtained using
the empirical method.
Dose varies with position in the 3 spatial planes;
vertically (inverse square law), the anode-cathode plane
(‘heel’ effect) and laterally. Note that while modelling the
inverse square law is simple, modelling exposure variation
in the other planes is non-trivial, which makes these
uncertainty components well suited to an empirical
approach to evaluate c.
Vertically, a 10 mm position change altered the final
result by 0.0592 mGy.
Therefore c = 0.00592 mGy/mm.
In the anode-cathode plane, a 10mm position change
altered the final result by 0.0272 mGy.
Therefore c = 0.00272 mGy/mm.
Laterally, a 10 mm position change did not alter the
final result.
Therefore c = 0 mGy/mm.
Calculation of u and ν
Both the Type A and Type B evaluation methods will
be demonstrated in the evaluation of u and ν for each
uncertainty component.
(i) When the reading value was recorded, a further 24
individual readings were also taken to obtain information
about the distribution of these values. Thus, a Type A
evaluation can be used. The standard uncertainty of the
reading is equal to the estimated standard deviation of the
25 readings, i.e. 0.0127 mGy. The value ν in this case
equals one less than the number of data used. Thus ν = 24.
(ii) The rounding of a displayed reading could be any
value in the range ±0.0005 mGy, as the reading is displayed
to 3 decimal places. Furthermore, there is equal probability
of the rounding being any value in this range, and zero
probability of the rounding being outside the range. This is
an example of a rectangular distribution. Thus from
equation 1;

3
0005 . 0
3
= =
a
u
= 0.000289 mGy
As the boundaries of the distribution are precisely
known, this evaluation would be judged as ‘excellent’ in
accordance with table 1, thus ν = 1000.
Assume the calibration factor was provided with
only a limited explanation, such as “…the calibration
factor = 1.01, with an uncertainty of 5%…”. To evaluate u
and ν, a number of assumptions will need to be made. The
first assumption is that the uncertainty distribution of the
calibration factor is a t-distribution. This is a reasonable
assumption as it is consistent with the CLT. The next
assumption is that it would be expected that coming from a
laboratory with traceable standards, the calibration was at
least ‘good’. This translates to ν=20 from table 1. Finally, a
decision needs to be made with regard to what confidence
interval was used. If the stated uncertainty of 5% was
assumed to represent a 95%CI, then from equation 2;
025 . 0 % 5 . 2
2
% 5
2
= = = =
a
u
.
(iii) Atmospheric conditions were measured using a
portable temperature and pressure meter. The meter’s
specifications stated that temperature measurements were
accurate to ±1°C, and pressure measurements were accurate
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
138
to ±100 Pa. To gain a greater insight into the uncertainty of
temperature and pressure readings from the meter,
comparisons were made at ambient temperature and
pressure with a calibrated thermometer (±0.2°C, 95%CI)
and a calibrated barometer (±50 Pa, 95% CI). For all
measurements taken over a period of several weeks, the
difference between the portable meter and the calibrated
instruments was less than the stated ‘accuracy’. It is
therefore reasonable to assume that uncertainties of ±1°C
and ±100 Pa represent 95%CIs, and that the evaluation for
both is at least ‘good’. Hence, from equation 2, u for
temperature = 0.50 K, u for pressure = 50 Pa, and from
table 1, ν = 20 for temperature and pressure.
(iv) The uncertainty in placing the ionisation chamber
in the specified location of the three spatial planes was
assessed by repeatedly placing the chamber in position for
measurement and measuring its actual position with digital
vernier calipers. The measurement with calipers is not done
routinely; it is too time consuming. However, the data
regarding positioning uncertainty enabled values for u and
ν to be estimated for the three positioning uncertainty
components. It was estimated that for all three planes, the
ionisation chamber was within a = 4 mm of the correct
location 95% of the time, and the measurements followed a
Normal distribution. From equation 2, u = 2 mm. The
evaluation was regarded as good (ν = 20).
Step 3. Combine the uncertainties
This step is most simply performed using a spreadsheet
program. Using equations (3) and (7) gives, u
comb
= 0.0523
mGy, and ν
eff
= 26.
Step 4. State the final result.
Using tabulated data
6,12
, the appropriate coverage
factor, k, for the desired confidence interval can be found.
If a 95%CI is chosen, and given ν
eff
= 26, then k = 2.056.
Hence U = 0.1074 mGy for a 95%CI. For Excel
spreadsheet users, U can be calculated using the formula
‘=TINV(0.05,ROUND(ν
eff
,0))× u
comb
’. Given the final
result of 1.97254 mGy, the final result and uncertainty need
to be rounded so they are harmonised. Firstly, round U in
accordance with the rounding method, yielding ±0.11.
Next, round the final result to the same decimal place,
yielding 1.97 mGy. The following statement can now be
made:
“Under the conditions prescribed for the measurement
the exposure was found to be 1.97 mGy. The uncertainty
was calculated in accordance with the ISO GUM, and was
found to have a 95% confidence interval of ±0.11 mGy and
26 degrees of freedom.”
The uncertainty analysis data for this example is shown
in table 2.
Discussion
A detailed description of the quantity to be measured is
an important step in reducing measurement uncertainty.
The omission of any detail, such as whether or not the
X-ray unit should be warmed up prior to testing, can cause
different (yet valid) final results.
Note that for simplicity of the example, the final result
was based on the value of one individual reading, while a
further 24 readings were taken but only used to estimate the
standard deviation of the reading. However, it would be
more sensible to use the mean of the 25 individual readings
as the value of ‘reading’. If this were done, the standard
uncertainty of ‘reading’ would take the smaller value of the
experimental standard deviation of the mean (ESDM)
6,11,12
.
Since the ESDM equals the standard deviation of the
individual data divided by the square root of the number of
readings, the standard uncertainty of ‘reading’ would be
five times smaller.
Also note from the example that equation 3 was used,
despite one of the uncertainty components (vertical position
of the ionisation chamber) varying with the final result in
accordance with the inverse square law, not linearly as
required by equation 3. However, given the measurement
was conducted approximately 600mm from the anode, and
that the standard uncertainty was 2 mm, the variation of
vertical position and dose is approximately linear over this
range.
From the data in the uncertainty analysis table, table 2,
an extra column for the result c
2
u
2
for each uncertainty
component can be generated. Uncertainty components with
larger values of c
2
u
2
contribute the largest proportion to the
overall measurement uncertainty, and so effort may be
directed to improving these measurements. Conversely,
there is little to be gained by directing effort towards
improving the measurements of uncertainty components
with low values for c
2
u
2
.
It is only necessary to include uncertainty components
that are significant in the uncertainty analysis table. If it is
unknown as to whether or not an uncertainty component is
significant, it should be included.
Conclusions
Those who are familiar with other measurement
uncertainty methods, especially in the field of
radiotherapy
15
, may recognise elements of the GUM
process. The advantage of using the GUM process is
that it provides a uniform approach to the determination
and expression of uncertainty in measurement
internationally. This unified approach allows measurements
to be compared more easily than beforehand,
particularly if there is disagreement of two independently
measured results. The NMI Guide provides useful
background information on the GUM, as well as
pragmatic solutions for measurements in applied
environments, without deviating from the GUM
philosophy. While there will initially be some work in
applying the protocols to the measurements used to test
medical equipment, these can easily be programmed into
spreadsheets, after which further calculations are
straightforward.
Australas. Phys. Eng. Sci. Med. Vol. 28, No 2, 2005 Gregory et al • A standard approach to uncertainties
139
References
1. South Australian Government, Radiation Protection and
Control (Ionising Radiation) Regulations 2000, Government
Gazette, 2000.
2. Craig, A. R., Heggie, J. C. P., McLean, I. D., Coakley, K. S.
and Nicoll, J. J., Recommendations for a mammography
quality assurance program, Australas. Phys. Eng. Sci. Med.,
24(3):107-131, 2001.
3. McCollough, C. H. and Zink, F. E., Performance evaluation of
a multi-slice CT scanner, Med. Phys., 26:2223-2230, 1999.
4. Luketina, I. A., Greig, L., Linear accelerator output
variability, Australas. Phys. Eng. Sci. Med., 27(3):155-159,
2004.
5. Nobels, F., Beckers, F., Bailleul, E., De Schrijver, P., Sierens,
L., Van Crombrugge, P., Feasibility of a quality assurance
programme of bedside blood glucose testing in a hospital
setting: 7 years' experience, Diabet. Med., 21(12):1288-91,
2004.
6. International Organization for Standardization, Guide to the
expression of uncertainty in measurement, International
Organization for Standardization, Geneva, 1995.
7. Hinkle, D., Wiersma, W., and Jurs, S., Applied Statistics
for the Behavioural Sciences, Houghton Miffin, Boston,
2002.
8. EURACHEM, Quantifying Uncertainty in Analytical
Measurement, Laboratory of the Government Chemist,
London, 1995.
9. International Atomic Energy Agency, Quantifying uncertainty
in nuclear analytical measurements, TECDOC 1401, IAEA,
Vienna, 2004.
10. Bentley R. E., Applying the ISO Guide to the calculation of
uncertainty: temperature, National Measurement Laboratory,
Sydney, 2001.
11. Taylor, B. N., Kuyatt, C. E., Guidelines for evaluating and
expressing the uncertainty of NIST measurement results,
National Institute of Standards and Technology, Gaithersburg,
1994.
12. Bentley R. E., Uncertainty in measurement: The ISO Guide,
6
th
ed., National Measurement Laboratory, Sydney, 2003.
13. Pentz, M., Shott, M., Handling Experimental Data, Open
University Press, Milton Keynes, 1988.
14. Kirkup, L., Experimental methods: an introduction to the
analysis and presentation of data, John Wiley & Sons, Milton,
1994.
15. International Atomic Energy Agency, Absorbed Dose
Determination in External Beam Radiotherapy: An
International Code of Practice for Dosimetry based on
Standards of Absorbed Dose to Water, TRS-398, IAEA,
Vienna, 2000.

Master your semester with Scribd & The New York Times

Special offer for students: Only $4.99/month.

Master your semester with Scribd & The New York Times

Cancel anytime.