Professional Documents
Culture Documents
đánh giá dữ liệu phân tích
đánh giá dữ liệu phân tích
chapter
J. Scott Smith
Food Science Institute, Kansas State University,
Manhattan, KS 66506-1600, USA
jsschem@ksu.edu
S.S. Nielsen, Food Analysis, Food Science Texts Series, DOI 10.1007/978-1-4419-1478-1_4, 53
c Springer Science+Business Media, LLC 2010
Chapter 4 • Evaluation of Analytical Data 55
of other labs to determine how well they agree, assum- If we have many samples, then the standard devi-
ing the other labs are accurate. ation is designated by the Greek letter sigma (σ ). It is
A term that is much easier to deal with and deter- calculated according to Equation [3], assuming all of
mine is precision. This parameter is a measure of how the food products were evaluated (which would be an
reproducible or how close replicate measurements infinite amount of assays).
become. If repetitive testing yields similar results, then
we would say that the precision of that test was good. ∑ ( xi − μ ) 2
From a true statistical view, the precision often is called σ= [3]
n
error, when we are actually looking at experimental
variation. So, the concepts of precision, error, and where:
variation are closely related.
σ = standard deviation
The difference between precision and accuracy can
be illustrated best with Fig. 4-1. Imagine shooting a xi = individual sample values
rifle at a target that represents experimental values. μ = true mean
The bull’s eye would be the true value and where the n = total population of samples
bullets hit would represent the individual experimen-
tal values. As you can see in Fig. 4-1a, the values can Because we do not know the value for the true
be tightly spaced (good precision) and close to the mean, the equation becomes somewhat simplified so
bull’s eye (good accuracy), or, in some cases, there that we can use it with real data. In this case, we now
can be situations with good precision but poor accu- call the σ term the standard deviation of the sample
racy (Fig. 4-1b). The worst situation, as illustrated in and designate it by SD or σ. It is determined accord-
Fig. 4-1d, is when both the accuracy and precision are ing to the calculation in Equation [4], where x̄ replaces
poor. In this case, because of errors or variation in the the true mean term μ and n represents the number of
determination, interpretation of the results becomes samples.
very difficult. Later, the practical aspects of the various ∑ (xi − x̄)2
types of error will be discussed. SD = [4]
n
When evaluating data, several tests are commonly
If the number of replicate determinations is small
used to give some appreciation of how much the
(about 30 or less), which is common with most assays,
experimental values would vary if we were to repeat
the n is replaced by the n − 1 term, and Equation [5]
the test (indicators of precision). An easy way to look
is used. Unless you know otherwise, Equation [5] is
at the variation or scattering is to report the range of
always used in calculating the standard deviation of a
the experimental values. The range is simply the dif-
group of assays.
ference between the largest and smallest observation.
This measurement is not too useful and thus is seldom
used in evaluating data. ∑ (xi − x̄)2
SD = [5]
Probably the best and most commonly used sta- n−1
tistical evaluation of the precision of analytical data is
the standard deviation. The standard deviation mea- Depending on which of the equations above is
sures the spread of the experimental values and gives used, the standard deviation may be reported as SDn
a good indication of how close the values are to each or σn and SDn−1 or σn−1 . (Different brands of soft-
other. When evaluating the standard deviation, one ware and scientific calculators sometimes use different
has to remember that we are never able to analyze labels for the keys, so one must be careful.) Table 4-1
the entire food product. That would be difficult, if not shows an example of the determination of standard
impossible, and very time consuming. Thus, the calcu- deviation. The sample results would be reported to
lations we use are only estimates of the unknown true average 64.72% moisture with a standard deviation of
value. 0.293.
Comparison of accuracy and precision: (a) Good accuracy and good precision, (b) Good precision and poor
4-1 accuracy, (c) Good accuracy and poor precision, and (d) Poor accuracy and poor precision.
figure
Chapter 4 • Evaluation of Analytical Data 57
Deviation
from
Observed % the Mean
Measurement Moisture (xi − x̄ ) (xi − x̄ )2
Once we have a mean and standard deviation, we 4-2 Values for Z for Checking Both Upper and
next must determine how to interpret these numbers. table Lower Levels
One easy way to get a feel for the standard deviation is
to calculate what is called the coefficient of variation Degree of Certainty Z value
(Confidence) (%)
(CV), also known as the relative standard deviation.
This calculation is shown below for our example of the 80 1.29
moisture determination of uncooked hamburger. 90 1.64
95 1.96
SD 99 2.58
Coefficient of variation (CV) = × 100% [6] 99.9 3.29
x̄
0.293
CV = × 100% = 0.453% [7]
64.72 interval around the mean using the statistical param-
eter called the Z value. We do this calculation by first
The CV tells us that our standard deviation is only looking up the Z value from statistical tables once we
0.453% as large as the mean. For our example, that
have decided the desired degree of certainty. Some Z
number is small, which indicates a high level of pre- values are listed in Table 4-2.
cision or reproducibility of the replicates. As a rule, The confidence limit (or interval) for our moisture
a CV below 5% is considered acceptable, although it
data, assuming a 95% probability, is calculated accord-
depends on the type of analysis. ing to Equation [8]. Since this calculation is not valid
Another way to evaluate the meaning of the stan-
for small numbers, assume we ran 25 samples instead
dard deviation is to examine its origin in statistical
of four.
theory. Many populations (in our case, sample values
or means) that exist in nature are said to have a normal Confidence interval (CI) = x̄ ± Z value
distribution. If we were to measure an infinite num-
standard deviation (SD)
ber of samples, we would get a distribution similar to × √
n
that represented by Fig. 4-2. In a population with a nor-
mal distribution, 68% of those values would be within [8]
±1 standard deviation from the mean, 95% would be
within ±2 standard deviations, and 99.7% would be 0.2927
CI (at 95%) = 64.72 ± 1.96 × √
within ±3 standard deviations. In other words, there 25
is a probability of less than 1% that a sample in a popu- = 64.72 ± 0.115% [9]
lation would fall outside ±3 standard deviations from
the mean value. Because our example had only four values for the
Another way of understanding the normal distri- moisture levels, the confidence interval should be cal-
bution curve is to realize that the probability of finding culated using statistical t tables. In this case, we have
the true mean is within certain confidence intervals as to look up the t value from Table 4-3 based on the
defined by the standard deviation. For large numbers degrees of freedom, which is the sample size minus
of samples, we can determine the confidence limit or one (n − 1), and the desired level of confidence.
58 Part I • General Information
To interpret this number, we can say that, with 95% Absolute error = Eabs = x − T [15]
confidence, the true mean for our moisture will fall where:
within 64.72 ± 0.465% or √ between 65.185 and 64.255%.
The expression SD/ n is often reported as the x = experimentally determined value
standard error of the mean. It is then left to the T = true value
reader to calculate the confidence interval based on the
desired level of certainty. The absolute error term can have either a positive
Other quick tests of precision used are the rela- or a negative value. If the experimentally determined
tive deviation from the mean and the relative average value is from several replicates, then the mean (0)
deviation from the mean. The relative deviation from would be substituted for the x term. This is not a
the mean is useful when only two replicates have been good test for error, because the value is not related
performed. It is calculated according to Equation [12], to the magnitude of the true value. A more useful
with values below 2% considered acceptable. measurement of error is relative error.
Eabs x−T
xi − x̄ Relative error = Erel = = [16]
Relative deviation from the mean = × 100 T T
x̄
[12] The results are reported as a negative or positive value,
where: which represents a fraction of the true value.
If desired, the relative error can be expressed as
xi = individual sample value percent relative error by multiplying by 100%. Then
x̄ = mean the relationship becomes the following, where x can
be either an individual determination or the mean (0)
If there are several experimental values, then the of several determinations.
relative average deviation from the mean becomes
a useful indicator of precision. It is calculated simi- Eabs x−T
%Erel = × 100% = × 100% [17]
larly to the relative deviation from the mean, except T T
the average deviation is used instead of the individual Using the data for the percent moisture of
deviation. It is calculated according to Equation [13]. uncooked hamburger, suppose the true value of
Chapter 4 • Evaluation of Analytical Data 59
the sample is 65.05%. The percent relative error Blunders are easy to eliminate, since they are so
is calculated using our mean value of 64.72% and obvious. The experimental data are usually scattered,
Equation [17]. and the results are not close to an expected value. This
type of error is a result of using the wrong reagent
x̄ − T or instrument or of sloppy technique. Some people
%Erel = × 100%
T have called this type of error the “Monday morn-
64.72 − 65.05 ing syndrome” error. Fortunately, blunders are easily
= × 100% = −0.507% [18] identified and corrected.
65.05
where:
xi and yi = individual values
x̄ and ȳ = means of the individual values
Low-cost calculators and computer spreadsheet soft-
ware can readily calculate regression equations so no
attempt is made to go through the mathematics in the
formulas.
The formulas give what is known as the line of
regression of y on x, which assumes that the error
occurs in the y direction. The regression line repre-
sents the average relationship between all the data
points and thus is a balanced line. These equations also
assume that the straight line fit does not have to go
through the origin, which at first does not make much
sense. However, there are often background inter-
A typical standard curve plot showing the data
4-3 points and the generated best fit line. The data ferences so that even at zero concentration, a weak
figure signal may be observed. In most situations, calculating
used to plot the curve are presented on the
graph. the origin as going through zero will yield the same
results.
Using the data from Fig. 4-3, calculate the concen-
Figure 4-3 illustrates a typical standard curve used tration of caffeine in the unknown and compare with
in the determination of caffeine in various foods. the graphing method. As you recall, the unknown had
Caffeine is analyzed readily in foods by using HPLC an area at 272 m of 4000. Linear regression analysis
coupled with an ultraviolet detector set at 272 nm. of the standard curve data gave the y-intercept (b) as
The area under the caffeine peak at 272 nm is directly 89.994 and the slope (a) as 90.727 (r2 = 0.9989).
proportional to the concentration. When an unknown y = ax + b [22]
sample (e.g., coffee) is run on the HPLC, a peak area is
or
obtained that can be related back to the sample using
the standard curve. y−b
x= [23]
The plot in Fig. 4-3 shows all the data points and a
a straight line that appears to pass through most of 4000 − 89.994
x (conc) = = 43.0964 ppm caffeine
the points. The line almost passes through the origin, 90.727
which makes sense because zero concentration should [24]
produce no signal at 272 nm. However, the line is not
perfectly straight (and never is) and does not quite The agreement is fairly close when comparing the
pass through the origin. calculated value to that estimated from the graph.
To determine the caffeine concentration in a sam- Using high-quality graph paper with many lines could
ple that gave an area of say 4000, we could extrapolate give us a line very close to the calculated one. How-
to the line and then draw a line down to the x-axis. ever, as we will see in the next section, additional
Following a line to the x-axis (concentration), we can information can be obtained about the nature of the
estimate the solution to be at about 42–43 ppm of line when using computer software or calculators.
caffeine.
We can mathematically determine the best fit of 4.4.2 Correlation Coefficient
the line by using linear regression. Keep in mind the
equation for a straight line, which is y = ax + b, where In observing any type of correlation, including lin-
a is the slope and b is the y-intercept. To determine ear ones, questions always surface concerning how to
the slope and y-intercept, the regression equations draw the line through the data points and how well the
shown below are used. We determine a and b and thus, data fit to the straight line. The first thing that should
for any value of y (measured), we can determine the be done with any group of data is to plot it to see if
concentration (x). the points fit a straight line. By just looking at the plot-
ted data, it is fairly easy to make a judgment on the
linearity of the line. We also can pick out regions on
∑ (xi − x̄)(yi − ȳ) the line where a linear relationship does not exist. The
Slope a = [20]
∑ (xi − x̄)2 figures below illustrate differences in standard curves;
y-intercept b = ȳ − ax̄ [21] Fig. 4-4a shows a good correlation of the data and
62 Part I • General Information
of the assay, and thus we strive to report a meaningful However, adding a decimal point and another
value, be it a mean, standard deviation, or some other zero gives the number 7000.0, which has five
number. The next three sections discuss how we can significant figures.
evaluate experimental values so as to be precise when
reporting results. A good way to measure the significance of zeros, if
the above rules become confusing, is to convert the
number to the exponential form. If the zeros can be
omitted, then they are not significant. For example,
4.5.1 Significant Figures 7000 expressed in exponential form is 7 × 103 and con-
The term significant figure is used rather loosely to tains one significant figure. With 7000.0, the zeros are
describe some judgment of the number of reportable retained and the number becomes 7.0000 × 103 . If we
digits in a result. Often, the judgment is not soundly were to convert 0.007 to exponent form, the value is
based, and meaningful digits are lost or meaningless 7 × 10−3 and only one significant figure is indicated.
digits are retained. Exact rules are provided below As a rule, determining significant figures in arith-
to help determine the number of significant figures metic operations is dictated by the value having the
to report. However, it is important to keep some least number of significant figures. The easiest way to
flexibility when working with significant figures. avoid any confusion is to perform all the calculations
Proper use of significant figures is meant to give an and then round off the final answer to the appropri-
indication of the sensitivity and reliability of the ana- ate digits. For example, 36.54 × 238 × 1.1 = 9566.172,
lytical method. Thus, reported values should contain and because 1.1 contains only two significant figures,
only significant figures. A value is made up of signif- the answer would be reported as 9600 (remember, the
icant figures when it contains all digits known to be two zeros are not significant). This method works fine
true and one last digit that is in doubt. For example, for most calculations, except when adding or subtract-
a value reported as 64.72 contains four significant fig- ing numbers containing decimals. In those cases, the
ures, of which three digits are certain (64.7) and the last number of significant figures in the final value is deter-
digit is uncertain. Thus, the 2 is somewhat uncertain mined by the numbers that follow the decimal point.
and could be either 1 or 3. As a rule, numbers that are Thus, when adding 7.45 + 8.725 = 16.175, the sum is
presented in a value represent the significant figures, rounded to 16.18 because 7.45 has only two numbers
regardless of the position of any decimal points. This after the decimal point. Likewise, 433.8 − 32.66 gives
also is true for values containing zeros, provided they 401.14, which rounds off to 401.1.
are bounded on either side by a number. For exam- A word of caution is warranted when using the
ple, 64.72, 6.472, 0.6472, and 6.407 all contain four simple rule stated above, for there is a tendency
significant figures. Note that the zero to the left of to underestimate the significant figures in the final
the decimal point is used only to indicate that there answer. For example, take the situation in which we
are no numbers above 1. We could have reported the determined the caffeine in an unknown solution to be
value as 0.6472, but using the zero is better, since we 43.5 ppm (see Equation [24]). We had to dilute the sam-
know that a number was not inadvertently left off our ple 50-fold using a volumetric flask in order to fit the
value. unknown within the range of our method. To calculate
Special considerations are necessary for zeros that the caffeine in the original sample, we multiply our
may or may not be significant. result by 50 or 43.5 μg/ml × 50 = 2, 175 μg/ml in the
unknown. Based on our rule above, we then would
1. Zeros after a decimal point are always signif- round the number to one significant figure (because
icant figures. For example, 64.720 and 64.700 50 contains one significant figure) and report the value
both contain five significant figures. as 2000. However, doing this actually underestimates
2. Zeros before a decimal point with no other pre- the sensitivity of our procedure, because we ignore
ceding digits are not significant. As indicated the accuracy of the volumetric flask used for the dilu-
before, 0.6472 contains four significant figures. tion. A Class-A volumetric flask has a tolerance of
3. Zeros after a decimal point are not significant 0.05 ml; thus, a more reasonable way to express the
if there are no digits before the decimal point. dilution factor would be 50.0 instead of 50. We now
For example, 0.0072 has no digits before the have increased the significant figures in the answer by
decimal point; thus, this value contains two sig- two, and the value becomes 2180 mg/ml.
nificant figures. In contrast, the value 1.0072 As you can see, an awareness of significant fig-
contains five significant figures. ures and how they are adopted requires close inspec-
4. Final zeros in a number are not significant tion. The guidelines can be helpful, but they do not
unless indicated otherwise. Thus, the value always work unless each individual value or number
7,000 contains only one significant figure. is closely inspected.
Chapter 4 • Evaluation of Analytical Data 65
4.5.2 Rejecting Data The example below shows how the test is used for
the moisture level of uncooked hamburger for which
Inevitably, during the course of working with experi-
four replicates were performed giving values of 64.53,
mental data we will come across a value that does not
64.45, 64.78, and 55.31. The 55.31 value looks as if it is
match the others. Can you reject that value and thus
too low compared to the other results. Can that value
not use it in calculating the final reported results?
be rejected? For our example, x1 is the questionable
The answer is “sometimes,” but only after care-
value (55.31) and x2 is the closest neighbor to x1 (which
ful consideration. If you are routinely rejecting data to
is 64.45). The spread (W ) is the high value minus the
help make your assay look better, then you are mis-
low measurement, which is 64.78 − 55.31.
representing the results and the precision of the assay.
If the bad value resulted from an identifiable mistake 64.45 − 55.31 9.14
Q-value = = = 0.97 [27]
in that particular test, then it is probably safe to drop 64.78 − 55.31 9.47
the value. Again, caution is advised because you may
From Table 4-4, we see that the calculated Q-value
be rejecting a value that is closer to the true value than
must be greater than 0.76 to reject the data. Thus, we
some of the other values.
make the decision to reject the 55.31% moisture value
Consistently poor accuracy or precision indicates
and do not use it in calculating the mean.
that an improper technique or incorrect reagent was
used or that the test was not very good. It is best
to make changes in the procedure or change meth-
4.6 SUMMARY
ods rather than try to figure out ways to eliminate
undesirable values.
This chapter focuses on statistical methods to measure
There are several tests for rejecting an aberrant
data variability, precision, etc. and on basic mathe-
value. One of these tests, the Q-Test, is commonly
matical treatment that can be used in evaluating a
used. In this test, a Q-value is calculated as shown
group of data. For example, it should be almost second
below and compared to values in a table. If the cal-
nature to determine a mean, standard deviation, and
culated value is larger than the table value, then the
CV when evaluating replicate analyses of an individ-
questionable measurement can be rejected at the 90%
ual sample. In evaluating linear standard curves, best
confidence level.
line fits should always be determined along with the
x2 − x1 indicators of the degree of linearity (correlation coef-
Q-value = [26]
W ficient or coefficient of determination). Fortunately,
where: most computer spreadsheet and graphics software will
readily perform the calculations for you. Guidelines
x1 = questionable value are available to enable one to report analytical results
x2 = next closest value to x1 in a way that tells something about the sensitivity and
confidence of a particular test. A section is included
W = total spread of all values, obtained by subtracting
which describes sensitivity and LOD as related to var-
the lowest value from the highest value
ious analytical methods and regulatory agency poli-
Table 4-4 provides the rejection Q-values for a 90% cies. Additional information includes the proper use of
confidence level. significant figures, rules for rounding off numbers and
use of the Q-test to reject grossly aberrant individual
values.
4-4
table Q-Values for the Rejection of Results
4. Using linear regression we get analytical chemists alike. It contains a fair amount of
Group A: y = 0.0504x − 0.0029, r2 = 0.9990 detail, sufficient for most analytical statistics, yet works
Group B: y = 0.0473x + 0.0115, r2 = 0.9708 through the material starting at a basic introductory level.
The assay evaluation parameters LOD, MDL, and LOQ
are thoroughly discussed. The authors also discuss the Q-
The Group A r2 is closer to 1.000 and is more linear and
test used for rejecting data. A big plus of this book is the
thus the better standard curve.
realistic approach taken throughout.
Sodium in the sample using group A standard curve is:
3. Miller JN, Miller JC (2005) Statistics and chemometrics
for analytical chemistry, 5th edn. Pearson Prentice Hall,
0.555 = 0.0504x − 0.0029, x = 11.1μg/ml
Upper Saddle River, NJ. This is another excellent intro-
ductory text for beginner analytical chemists at the under-
Sodium in the sample using group B standard curve is: graduate and graduate level. It contains a fair amount of
detail, sufficient for most analytical statistics, yet works
0.555 = 0.0473x + 0.0115, x = 11.5 μg/ml through the material starting at a basic introductory level.
The authors also discuss the Q-test used for rejecting data.
4. Skoog DA, West DM, Holler JF, Crouch SR (2000) Ana-
4.9 RESOURCE MATERIALS lytical chemistry: an introduction, 7th edn. Brooks/Cole,
Pacific Grove, CA. Section I, Chaps. 5–7 do an excel-
1. Garfield FM, Klestra E, Hirsch J (2000) Quality assur- lent job of covering most of the statistics needed by an
ance principles for analytical laboratories, AOAC Inter- analytical chemist in an easy-to-read style.
national, Gaithersburg, MD. This book covers mostly 5. Code of Federal Regulations (2010) Environmental Pro-
quality assurance issues but has a good chapter (Chap. 3) tection Agency. 40 CFR, Vol. 22. Chapter I part 136 –
on the basics of statistical applications to data. Guidelines Establishing Test Procedures for the Analysis
2. Meier PC, Zund RE (2000) Statistical methods in ana- of Pollutants. Appendix B to Part 136 – Definition and
lytical chemistry, 2nd edn. Wiley, New York. This Procedure for the Determination of the Method Detection
is another excellent text for beginner and advanced Limit – Revision 1.11.