Professional Documents
Culture Documents
VICTOR DAVID
University of Bucharest, Bucharest, Romania
Elsevier
Radarweg 29, PO Box 211, 1000 AE Amsterdam, Netherlands
The Boulevard, Langford Lane, Kidlington, Oxford OX5 1GB, United Kingdom
50 Hampshire Street, 5th Floor, Cambridge, MA 02139, United States
Copyright © 2021 Elsevier B.V. All rights reserved.
No part of this publication may be reproduced or transmitted in any form or by any means, electronic or
mechanical, including photocopying, recording, or any information storage and retrieval system, without
permission in writing from the publisher. Details on how to seek permission, further information about the
Publisher’s permissions policies and our arrangements with organizations such as the Copyright Clearance
Center and the Copyright Licensing Agency, can be found at our website: www.elsevier.com/permissions.
This book and the individual contributions contained in it are protected under copyright by the Publisher
(other than as may be noted herein).
Notices
Knowledge and best practice in this field are constantly changing. As new research and experience broaden
our understanding, changes in research methods, professional practices, or medical treatment may become
necessary.
Practitioners and researchers must always rely on their own experience and knowledge in evaluating and
using any information, methods, compounds, or experiments described herein. In using such information or
methods they should be mindful of their own safety and the safety of others, including parties for whom
they have a professional responsibility.
To the fullest extent of the law, neither the Publisher nor the authors, contributors, or editors, assume any
liability for any injury and/or damage to persons or property as a matter of products liability, negligence or
otherwise, or from any use or operation of any methods, products, instructions, or ideas contained in the
material herein.
ISBN: 978-0-12-821405-3
The most utilized techniques in analytical previously reported ones to a specific goal.
chemistry are undoubtedly those based on This book provides guidance to this effort.
chromatographic separations and detection. Besides solutions to more and more difficult
In spite of the unmatched separation capa- problems in current applications, the
bility of chromatography, and exceptional research in this field is also focused on the
sensitivity of chromatographic detectors, it is use of new materials and new techniques in
common that many chromatographic ana- order to improve the analytical outcome or
lyses cannot be performed directly on the to diminish the impact of certain procedures
sample to be analyzed. In some instances, on the environment or human safety. The
part of the sample components affects unfa- continuous increase in the need for more
vorably the chromatographic separation or types of analyses, more precise results, and
alters the quality of separation media, and in measurement of lower levels of particular
other instances the level of the compounds to compounds led to further progress in chro-
be analyzed in a sample is too low, even for matographic instrumentation and in devel-
the most sensitive detectors used in chro- opment of new methods of analysis, most of
matography. Sample preparation for chro- them having sample preparation as an integral
matography is designed to solve these part. This process keeps accelerating and more
problems, and makes a sample amenable for sample preparation techniques and new ma-
analysis, or improves the performance of terials with better performance in sample
analysis by involving steps such as dissolu- preparation practice have been introduced.
tion, cleanup of the initial sample, enrich- The new edition of this book attempts to cap-
ment of the compounds of interest, or ture as much as possible from these aspects.
changes of their structure in order to bring The presentation of sample preparation
them in the range of the detector capability. for chromatography as a coherent subject,
Sample preparation is a topic frequently including the description of new de-
included in the published materials in peer- velopments in the field, has been the goal of
reviewed journals such as Journal of Chro- the previous edition of this book that was
matography A, and B, Journal of Separation published by Elsevier in 2015. This new
Science, Journal of Chromatographic Science, edition of the book “Modern Sample Prepara-
Chromatographia, Analytica Chimica Acta, tion for Chromatography” includes an update
Talanta, Trends in Analytical Chemistry of the literature and the new achievements in
(TrAC), etc., this being a proof of the this domain. The second edition of this book
importance of this subject in separation sci- contains several new chapters and some new
ences. Seen as a common part of developing theoretical aspects related to sample prepa-
an analytical method, many times the ana- ration, which are useful for the chemists
lysts spent time and energy to develop studying or working in the field of chemical
sample preparation procedures or to adapt analysis.
xv
xvi PREFACE
The authors wish to thank the editorial thanks to Dr. Steven Alderman, Mrs. Karen
team from Elsevier, Ms. Emerald Li, Mrs. Kilby, and Dr. Ching-En Chou for reviewing
Kathryn Eryilmaz, and Mrs. Sruthi Satheesh, the manuscript and suggesting valuable
for their contribution to the publication of corrections.
this book. Also, the authors express their
C H A P T E R
1
Preliminaries to sample preparation
1.1 Collection of information and planning for a chromatographic analysis
General comments
The most utilized procedures today in analytical chemistry are undoubtedly those based
on chromatographic separations and detection. In spite of the unmatched separation capa-
bility of chromatography, and exceptional sensitivity of chromatographic detectors, it is
common that some chromatographic analyses cannot be performed directly on the sample
to be analyzed. In some instances part of the sample components affects unfavorably the
chromatographic separation, and in other, the level of the compounds to be analyzed in a
sample is too low, even for the most sensitive detectors used in chromatography. Chemical
and/or physical modification of a sample in order to make it amenable for a chemical anal-
ysis or to improve the performance of analysis is indicated as sample preparation, or sample
treatment. Samples can be considered as made from two distinct parts, the analytes and the
matrix. The analytes are the chemical species of interest in the sample, and the matrix is
the rest of the sample components. Sample preparation can be performed for sample disso-
lution, matrix modification and/or simplification, analytes concentration, preliminary sepa-
ration of the analytes from the sample, chemical changes of the analytes or interferences,
etc. All these operations are performed before the core analysis with the purpose to improve
the analysis results. The specific core analysis for which the sample preparation is discussed
in this material is chromatography.
sample processing must ensure integrity of samples, prevent deterioration and cross contam-
ination, maintain sample tracking and the chain of custody. Preservation of part of the sample
for potential needs to be reanalyzed as well as its disposal after the analytical process is
completed should also be done in a correct manner. In some procedures, sampling and sam-
ple preparation are performed simultaneously. Such situations occur, for example, in gaseous
sampling when the process requires the analyte trapping from a flow. Sample preparation is
always performed correlated to the selected core analysis. This book presents various aspects
of sample preparation done with the purpose of using chromatography as a core analytical
procedure.
References
[1] S.C. Moldoveanu, V. David, Sample Preparation in Chromatography, Elsevier, Amsterdam, 2002.
[2] Infodent, n.d. http://www.info.dent.nu.ac.th/chemistry/files/chemical%20grade%20diagnosis.pdf.
[3] International Standards Organization (ISO), General Requirements for Competence of Reference Material
Producers, ISO Guide 17034, 2016.
[4] S.A. Wise, What is novel about certified reference materials? Anal. Bioanal. Chem. 410 (2018) 2045e2049.
[5] S.A. Wise, K.W. Phinney, L.C. Sander, M.M. Schantz, Role of chromatography in the development of standard
reference materials for organic analysis, J. Chromatogr. A 1261 (2012) 3e22.
[6] S.A. Wise, D.L. Poster, J.R. Kucklick, J.M. Keller, S.S. Vanderpol, L.C. Sander, M. Schantz, Standard reference ma-
terials (SRMs) for determination of organic contaminants in environmental samples, Anal. Bioanal. Chem. 386
(2006) 1153e1190.
General aspects
Sample preparation is an integral part of most analytical methods, and is included with the
purpose to improve the analysis results. Although sample preparation is performed at the
beginning and data analysis is performed after the analysis is completed, it is important to
have clear criteria for the evaluation of the quality of an analytical method (including its sam-
ple preparation part). Besides descriptors such as the number of operations, higher
throughput for the laboratory, and lower cost of a sample preparation procedure, the quality
of the analytical results must be evaluated and potentially compared with other procedures
used for achieving the same goal. Sample preparation can be applied for improving both
qualitative and quantitative results. The role of sample preparation for improving the qual-
itative results will be presented in more details in Chapter 2, as well as in other places in
this book. This section describes criteria for the evaluation of quantitative results from an
analytical method, as well as procedures for comparing different methods regarding the qual-
ity of the quantitative results they generate. This comparison can be useful for the selection of
a method to be adopted in solving a given analytical task.
Chemical analysis may be designed to provide qualitative information (e.g., compound na-
ture, structure, presence of organic functional groups, etc.) or quantitative information related
to the amount or the concentration of one or more components in the sample. Both qualitative
and quantitative information must be accurate and representative for the material analyzed.
The characterization of qualitative information depends on the purpose of analysis and the
procedure utilized to obtain it. For quantitative information the results are typically evalu-
ated using statistical concepts. This is possible because related to any sample it is common
to perform more than one measurement. The most common characteristics for quantitative
data are the precision and the accuracy. Precision is used to describe the reproducibility of
the results obtained in identical fashion and can be defined as the agreement between the nu-
merical values of two or more measurements. Accuracy denotes the nearness of the results to
the true value of the measured quantity (if known) or to its generally accepted value. The
comparison of different analytical methods and implicitly of different sample preparation
procedures is frequently done using precision and accuracy as key criteria for their selection.
Some statistical aspects related to a quantitative analytical method are presented in this
section.
y ¼ FðxÞ (1.2.1)
When more than one measurement is performed, the results for x are typically scattered
around a specific value, the measurements being always affected by errors. The errors of mea-
surements are classified as systematic (determinate) or random (see, e.g., Ref. [1]). Systematic
errors are generated by a specific cause, and they affect the accuracy of the results. They are
typically detected by the comparison of the results from the analysis to be evaluated with the
results from different methods of analysis or with known values for the measured analyte.
Systematic errors can be eliminated only when their source is identified. There are two
main types of systematic errors, the constant errors and the proportional errors. The constant
errors are independent of the measured value. The proportional errors depend on the
measured value. The source of constant errors can be (a) insufficient selectivity and the mea-
surement of signals from other components together with the analyte, (b) loss of analyte,
(c) interference of the matrix, (d) inadequate blank corrections, (e) contamination,
(f) problems with the analytical instrumentation, etc. The source of proportional errors can
be (a) incorrect slope of the calibration line, (b) incorrect assumption of linearity,
(c) changes in time of the sensitivity of the analytical instrumentation, etc.
Random errors do not have an assigned cause, and they affect the precision of the mea-
surement. Precision refers to the reproducibility of measurement within a set, indicating
the scatter or dispersion of a set of measurements about their central value [2]. It can be
assumed that random errors are scattered within a continuous range of values. Therefore,
the measurement of one variable x can generate any values in this continuous range, with
x indicated as a random variable. Any obtained set of measurements {x1, x2 . xn} is defined
in statistics as a sample of the random variable in this continuous range of values. Only all
measurements (which must be infinite in number to cover the whole range) would generate
the ideal set, which is called population. (The statistical term sample can easily be confused in
analytical chemistry with the term sample ¼ specimen, as the term population may also have a
different meaning. To avoid this confusion, the statistical terms sample and population are
italicized in this book.) Statistical treatment of random variables is described in detail in
many dedicated books [3e8].
Repeated determination of the amount or concentration of an analyte generates for each
measurement j a value xj. If the number of measurements is n, they will generate the sample
{x1, x2 . xn}. The average (or the mean) of these measurements is m and the standard deviation
(SD) is s. Standard deviation shows the distribution of measurements about the mean. The
expressions for m and s are given by the following formulas:
P
n
xj
j¼1
m¼ (1.2.3)
n
and
The mean m and the standard deviation s are the most common values to express analyt-
ical results. Standard deviation is the measure characterizing precision (of a set of measure-
ments), a small standard deviation indicating good precision for the set of measurements.
Both m and s are expressed in the same units. A relative standard deviation % (RSD%) is
frequently used instead of standard deviation s for the characterization of precision, and it
is expressed by the following formula:
s
RSD% ¼ 100 % (1.2.5)
m
Instead of standard deviation s, it is also common to use for precision characterization the
value s2 called variance.
In statistics, any obtained set of measurements {x1, x2 . xn} represents a sample of values
taken by a random variable x. This sample is a part of a continuous infinite set of values that
variable x may have. This infinite set of values is known as a population. Similar to the average
m and standard deviation s for one sample, the population is characterized by the mean m of the
population and the standard deviation s of the population which are given by the following
formulas:
m ¼ lim mn (1.2.6)
n/N
s ¼ lim sn (1.2.7)
n/N
Since an infinite number of measurements cannot be performed, the true values for m and
for s are not known, and they are approximated with m and s for large number of measure-
ments n. The value s2 is also indicated as variance. The true (ideal) value for m represents the
ideal result for all the measurements, but does not necessarily represent the true value of the
measured quantity. The difference between m and the true value (if known) of a quantity is
called the bias. This true value can be, for example, the level of an analyte in a standard
(regardless the fact that making of the standard can also be affected by errors). Another
convention for a true value is the assumption that it is the one “generally accepted.” Accept-
ing mo as a true value of the measurement and taking the population mean m z m, the bias is
approximated by the following formula:
bias z m mo (1.2.8)
A method is considered accurate when the bias is very small (or even zero). For one mea-
surement xi the difference ximo is a combination of the systematic errors and random errors
and can be considered a total error of the measurement xi.
In expression 1.2.9, s is the standard deviation for all measurements for all sets and n is the
number of measurements in a set. Formula 1.2.9 can be used even for k ¼ 1. The standard
error of the mean sk does not evaluate the precision, and it is only a measure of the confidence
in a result indicated by the mean.
One question regarding the distribution of measurements about their mean is the expected
frequency of occurrence of an error as a function of the error magnitude. It is expected that
larger random errors will be less frequent than small errors. The most commonly utilized
function, which describes well the relative frequency of occurrence of random errors in large
sets of measurements, is given by the Gauss formula:
2
1
f ðxÞ ¼ 1=2
exp x m =2s 2
(1.2.10)
ð2ps2 Þ
This frequency function f(x) (known as Gaussian density function) shows that the point of
maximum frequency is obtained for the mean (when x ¼ m), the distribution of positive
and negative errors is symmetrical, and as the magnitude of the deviation from the mean
increases, an exponential decrease in the frequency takes place. The errors with the relative
frequency of occurrence given by eq. 1.2.10 have a so-called normal distribution N (m,s).
Besides Gaussian density functions, other frequency functions are known (e.g., Student
frequency function).
For x with a normal distribution N (m,s), the following substitution
y ¼ x m =s (1.2.11)
The cumulative frequency distribution is equal to the probability P for x to have a value
below a in any measurement. The integral of f(x) over the whole space gives P ¼ 1. The values
of function F(a) (for f(x) a Gaussian function) are known and tabulated (e.g., Ref. [9]) or given
in computer statistical packages (e.g., Ref. [5,10]).
The mean mk of a sample of n values {x1, x2 . xn} is itself a value of a random variable m.
Assuming that x has a normal distribution N (m, s), the random variable m(mk is one of
the possible values of m) takes a continuous range of values with a normal distribution
N (m, s/n1/2).
A mean-centered standardized variable defined by the formula
z ¼ m m = s=n1=2 (1.2.13)
will have an N (0,1) distribution. With the help of variable z it is possible to evaluate how
close the values of m and mk are for a certain population. For the variable z given by eq.
1.2.13, two values za/2 and z1a/2 can be found such that the probability for z of being
outside the interval (za/2, z1a/2) is equal to a (areas a/2 indicated in Fig. 1.2.1). The
interval (za/2, z1a/2) is indicated as confidence interval. For z being inside the interval
(za/2, z1a/2) or
za=2 < m m = s = n1=2 < z1a=2 (1.2.14)
FIGURE 1.2.1 Gaussian curve N (0,1), showing two values za/2 and z1a/2 such that the probability for z of
being outside the interval (za/2, z1a/2) is equal to a (and area under the curve is P ¼ 1a showing probability
inside the same interval).
nþ1
G ðnþ1Þ=2
1 2 t2
f ðt; nÞ ¼ pffiffiffiffiffiffi n 1þ (1.2.16)
np G n
2
are also known and tabulated [4] or are given by computer packages. The variable t is equiv-
alent with variable z for Gaussian distribution. For a selected probability P (and a ¼ 1P)
chosen for the decision, two values ta/2,n and t1a/2,n (with ta/2,n ¼ t1a/2,n) can be found
such that the probability for t of being inside the interval (ta/2,n, t1a/2,n) is equal with P
and outside the interval is a. This interval and the associated probability are indicated as
two sided (both sides of the interval (ta/2,n, t1a/2,n) are finite). A one-sided probability can
also be defined such that N < t < ta/2,n. The variation of t1a/2,n values for several n and
at three different values for P is shown in Fig. 1.2.2 for a two-sided probability. The graphs
indicate that for any specified probability, an increase in the number of measurements
n ¼ n þ 1 leads to a decrease in the value of t1a/2,n. Using for Student distribution an expres-
sion similar to eq. 1.2.15 to evaluate the maximum possible difference between m and m, it can
be seen from the variation shown in Fig. 1.2.2 that a larger interval is expected for a small
number of determinations compared to a larger number.
Several other frequency (density) functions are known and utilized for describing the
occurrence of random errors. Among these, binomial distribution, c2 distribution, and F-dis-
tribution (Snedecor) are used in data processing in analytical chemistry [4].
FIGURE 1.2.2 Variation of t1a/2,n as a function of n for a Student distribution for four selected probabilities P
(99%, 98%, 95%, and 90%).
A random variable x that has an F-distribution has the frequency function (density func-
tion) that depends on two parameters d1 and d2 and has the following form:
d1
d
d1 þd2
1 d1 2
2 1
1 d1 2
f ðx; d1 ; d2 Þ ¼ x 1þ x (1.2.19)
Bðd1 =2; d2 =2Þ d2 d2
Similar to other distributions, values for a cumulative frequency F(d1, d2, a) given by the
formula
Z f
Fðd1 ; d2 ; aÞ ¼ f ðt; d1 ; d2 Þdt (1.2.21)
N
are known and are tabulated [4] for given values d1, d2 and confidence probabilities P ¼ 1a.
A collection of statistical models used to analyze data is offered by ANOVA (analysis of
variance) package that is available in Excel. One of the parameters calculated in ANOVA
is, for example, the variance of a given dataset.
dFðxÞ
S ¼ (1.2.22)
dx
FðxÞ ¼ a þ bx (1.2.23)
In this case, the sensitivity S is equal to the constant b. Sensitivity can therefore be deter-
mined from the calibration curve for the method. For nonlinear dependencies the expression
(1.2.22) still can be applied, but the sensitivity is not constant for all concentrations. Many
methods have only a range of linear dependence, with an upper region and a lower region
that are not linear. In the nonlinear regions, the sensitivity varies, and for this reason sensi-
tivity is not necessarily a convenient way to characterize an analytical method. Two examples
are given in Figs. 1.2.3 and 1.2.4.
Curve A in Fig. 1.2.3 shows a linear dependence with the slope 0.4, and curve B shows a
linear range with the slope 0.75. However, for concentrations below 0.2 or higher than 0.8, the
curve B is not linear. At a low concentration, the slope indicated by C shows lower sensitivity
than for curve A. It is also possible that the calibration is not linear and better correlation
between the level of the analyte and the response of the analytical instrument is quadratic
or even of a higher degree polynomial. Also, it is common that concentrations below a certain
value show a sensitivity decrease. The deviation from linearity at lower concentrations in
chromatographic analysis is probably due to a combined effect of the loss of a small amount
of sample that is decomposed or adsorbed irreversibly and the overall modification of the
chromatographic process. The loss of sample in itself is expressed in the dependence
y ¼ a þ bx by the negative value of a. The losses may depend on concentration and are
smaller at lower concentrations. This changes the slope of the calibration curve as shown
in Fig. 1.2.4. For nonlinear dependencies (e.g., quadratic) the slope is always a function of
concentration. However, at low concentrations the dependence is typically considered to
be linear.
0.7
0.6
0.5
0.4
Response
B A
0.3
0.2
0.1
C
0
0 0.2 0.4 0.6 0.8 1
Conc.
FIGURE 1.2.3 Calibration curves, one showing constant sensitivity (A) and the other (B) with constant sensitivity
for a limited interval (0.2e0.8 arbitrary units). Curve C shows the slope at a low concentration for curve B.
2000
1800
1600
ideal
1400 dependence
y = bx
1200
a
signal y
1000
800
200
0
0 non- 5 10 15 20
linear
linear concentration x
FIGURE 1.2.4 The calibration line for an ideal linear dependence and one for a dependence that is linear only for
a specific concentration interval.
The concentration of an analyte below which the detection is not possible is typically indi-
cated as limit of detection. For a quantitative characterization, the detection limit can be dis-
cussed in terms of signal and transformed in terms of amount or concentration using the
calibration function. The signal y for an analytical measurement is in fact the difference
between the signal for a sample ys and that of a blank yb, both affected by errors. The signals
for the blank (or the noise) are assumed to have a mean mb and the signals from the sample to
have a mean ms. The standard deviations for both the blank and the sample can be considered
equal to the same value sy ¼ s. A specific value of the signal can be selected (arbitrarily) and
considered as not generated by the noise. This value is indicated as decision limit L [12] where
the signal is noticeable and
L ¼ ms (1.2.24)
Assuming that the errors in the signal of the blank and sample have a normal distribution,
the probability to obtain signals from the blank higher than the decision limit L is given by the
following expression:
Z N
a¼ f ðyb Þdyb (1.2.25)
L
In expression 1.2.25, the function f is given by formula 1.2.10 (Gaussian density function)
that has m ¼ mb (s, the standard deviation is the same for both the blank and the sample).
L ¼ m b þ ks (1.2.26)
The probability to consider a noise as being signal from the analyte for a value higher than
L is given by P ¼ a. For k ¼ 2.33 the resulting value for the one-sided normal distribution
gives a ¼ 0.01 (or 1% if expressed in percent). Therefore, if the signal is higher than mb þ
2.33 s, the probability of false positives (signal from the background to be considered analyte)
is 1%.
A signal y with ms ¼ L has, however, the problem of possibly generating false negatives.
The probability of false negatives is given by the following expression:
Z L
b¼ f ðys Þdys (1.2.27)
N
In expression 1.2.27, f(ys) is a Gaussian with m ¼ ms. This probability is P ¼ .5 (or 50%)
because the normal curve is symmetrical around the mean. Therefore, the possibility of false
negatives is very high, and 50% of the true positives will be considered noise. This situation is
depicted in Fig. 1.2.5.
A higher signal (higher values for ms) and maintaining L ¼ ms will continue to diminish the
chances for false positives since the area a will become smaller and smaller as the second
Gaussian moves to higher values of y. However, maintaining L ¼ ms does not modify the
FIGURE 1.2.5 Graph showing on the horizontal axis the signal y, the values mb and ms as well as the decision limit
L ¼ ms. The vertical axis gives the frequencies f(y) of occurrence for the value of a measurement, where f is given by
eq. 1.2.10 and describes a normal distribution.
probability of false negatives. For a chosen probability P ¼ .01 of obtaining a false negative,
the corresponding value of the signal can be calculated. This value is noted by D, should be
smaller than ms, and has the following expression:
D ¼ ms k0 s (1.2.28)
For the probability b ¼ 0.01, the resulting value for the one-sided normal distribution is
k0 ¼ 2.33.
For L ¼ D, the subtraction of eq. 1.2.28 from eq. 1.2.26 leads to the following expression:
ms ¼ mb þ ðk þ k0 Þs (1.2.29)
The overlapping of frequency functions for the two probabilities is shown in Fig. 1.2.6.
The result of these considerations is that in order to have a probability of 1% for a false-
negative signal and a probability of 1% for a false-positive signal requires k þ k0 ¼ 4.66,
and in this case it can be written the following formula:
ms ¼ mb þ 4:66s (1.2.30)
If the signal is considered S ¼ msemb and the noise is taken as N ¼ s, formula 1.2.30 is
equivalent with the following formula:
S = N z4:66 (1.2.31)
For the detection to be possible in analytical practice, a value of S=N z3:33 that is slightly
smaller than 4.66 was considered to be sufficient (with lower probability than 1% for false
negatives, and higher than 1% to give false positives). The value S=N ¼ 3:33 has been
FIGURE 1.2.6 Graph showing on the horizontal axis the signal y, the values mb and ms, as well as the decision
limit D ¼ msk0 s. The vertical axis gives the frequencies f(y) of occurrence for the value of a measurement where f is
given by eq. 1.2.10 and describes a normal distribution.
m ¼ 3:33s (1.2.32)
Expression 1.2.32 indicates that for detection to be possible, a concentration (or level) at
least equal to 3s for a set of measurements (at low concentration) is necessary. Based on
expression 1.2.32 it results that for a set of measurements at low analyte concentration the
definition of LOD should be based on the following expression [17e19]:
FIGURE 1.2.7 Example of measurement of two signals and two selections for the noise leading to two different
values N 1 and N 2 .
The convention of taking LOD equal with 3.33 versus 4.66 increases the probability P of
obtaining a false positive and decreases the probability P of obtaining a false negative
(in Fig. 1.2.3 a smaller coefficient for s moves the point D toward lower values).
Various other formulas were developed for the evaluation of LOD. For example, for a
dependence of the form y ¼ a þ b$C, when the average signal of the blanc mb and the standard
deviation of the signal s are known, the LOD for the concentration of the analyte can be
obtained from the expression (ms is the average signal of the analyte):
ms a 1
LOD ¼ ¼ ðm a þ 3:33sÞ (1.2.34)
b b b
An additional parameter indicated as limit of quantitation LOQ has been introduced, with
the goal of decreasing the probability of false positives. The definition of LOQ is the
following:
LOQ indicates the minimum amount (or concentration) that can be quantitated by a spe-
cific method. The choice of 10s for LOQ decreases the probability P for obtaining a false pos-
itive but increases the probability P of obtaining a false negative. To prevent the problem of
generating false negatives, the measurements with results below 10s are not labeled as “an-
alyte absent” but “analyte below LOQ.”
One problem related to the determination of LOQ and LOD using the standard deviation s
is caused by its variation depending on the magnitude of xi (and m), which in analytical
chemistry are concentrations or amounts. Not only the absolute magnitude of s depends
on that of measured concentration or amount but it was empirically shown that even the rela-
tive standard deviation (RSD%) depends on it. An empirical formula relating RSD% on the
concentration level C as dimensionless mass fraction known as Horwitz’ function is given
by the following expression [20]:
Because of the variation of s with the magnitude of measured values, in expressions 1.2.32
and 1.2.35 the s is evaluated for the lowest possible or a very low concentration (or amount) of
the analyte. To avoid the problem of s variability, the LOQ can be substituted by the lowest
concentration used in the calibration curve used for the quantitation (e.g., expression 1.2.23),
with the condition that the back calculation of this concentration obtained using the calibra-
tion curve generates a value within the RSD% obtained by expression 1.2.36.
More elaborate formulas were also developed for the calculation of LOD and LOQ, such
that to eliminate the problems of selection of specific range of measuring the noise, or of the
selection of the concentration and number of measurements to be selected for the evaluation
of s. One such formula for LOD is indicated below for a dependence given by expression
y ¼ a þ bC where C is the concentration of the calibration solution [19]:
3:33 Sy pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
LODC ¼ 1 þ h0 þ 1=N (1.2.37)
b
where
I. General concepts in sample preparation
1.2 Statistical evaluation of quantitative data 23
vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
uN 2
uP
u y i yi;calc
ti ¼ 1 m2C
Sy ¼ h0 ¼ (1.2.38)
N2 P
N
2
ðCi mC Þ
i¼1
with N being the number of measurements, mC the average of calibration values Ci, and yi,calc
the values calculated for yi corresponding to Ci. For a formula for LOQ, the coefficient 3.33 in
formula 1.2.37 is replaced with 10.
The establishing of LOD and LOQ for an analytical method is very important. Some vari-
ation in the procedure for the establishing of LOD and LOQ are described with each analyt-
ical method. Although it is common to measure s (or SD) for a set of low standards, there is a
known fact that, it is possible for the analysis of standards to generate cleaner chromatograms
than the analysis of the same analytes in a matrix. Even when sample preparation offers pro-
cedures for matrix cleanup, in many instances only a partial cleanup is performed, and
besides the analytes some additional matrix compounds may remain in the processed sample
(residual matrix). The residual matrix can be tolerated when it does not interfere significantly
with the analytes measurement, but this does not necessarily imply that at very low level of
the analyte the residual matrix has no effect on sensitivity. For this reason, in certain analyses,
although the measurement is accurate for the analytes above a certain level, the determina-
tions close to the LOQ (LOQ determined with standards) is difficult or even not possible.
For this reason, for some procedures instead of reporting LOD and LOQ obtained with stan-
dards, (practical) PLOD and PLOQ are reported. In these cases, the measurement of standard
deviation s or the signal to noise S=N generated by a low level of analyte in the sample with
residual matrix components (possibly after partial cleanup) are used in the calculation. This is
not always possible since samples with low levels of analyte may not be available. The con-
centration of the analyte in the sample can be measured as usual, using standards or standard
addition procedures. The values of PLOD and PLOQ are more useful for practical application
of a method when they are different from LOD and LOQ.
ri ¼ yi ða þ bxi Þ (1.2.39)
The values ri are called residuals, and for the minimization of all ri (i ¼ 1,2,.n), the
following expression must be minimal:
X
n
2
X
n
Eða; bÞ ¼ ðyi a bxi Þ ¼ r2i ¼ min (1.2.40)
i¼1 i¼2
Since the procedure involves the minimization of the errors in the calculated values for y, it
is known as linear minimum mean-square (least square) error fitting. For expressing further
results related to the minimization of expression E(a, b), several notations must be introduced.
The average of xi values will be noted with mx, and the average of yi values will be noted with
my. Also several partial sums are noted as follows:
X
n
Sxx ¼ ðxi mx Þ2 (1.2.41a)
i¼1
X
n
Syy ¼ ðyi my Þ2 (1.2.41b)
i¼1
X
n
Sxy ¼ ðxi mx Þðyi my Þ (1.2.41c)
i¼1
The minimum is achieved for E(a,b) when its partial derivative with respect to a and b are
set to zero. This is expressed by the following formulas:
vEða; bÞ vEða; bÞ
¼0 ¼ 0 (1.2.42)
va vb
Expression 1.2.42 leads to a pair of equations in a and b. The solution of these two equa-
tions leads to the following results for the minimization of E(a,b):
The same minimization generates results for the standard deviations for the residuals sr,
for the slope sb, and for the intercept sa.
Syy b2 Sxx
s2r ¼ (1.2.44)
n2
s2r
s2b ¼ (1.2.45)
Sxx
P
n
s2r x2i
i¼1
s2a ¼ (1.2.46)
nSxx
Perfect fit of experimental data {y1, y2 . yn} and calculated data {aþbx1, aþbx2,. aþbxn} is
Pn
indicated by R2 ¼ 1 when r2i ¼ 0. For larger residuals, the value of R2 decreases and can
i¼1
have values between 1 and 0. The calculation of a and b for the regression line as well as of sr,
sb, sa, and R2 is available in Excel computer package (see, e.g., Ref. [21]).
One procedure for evaluating the results of a calibration is the use of ANOVA package.
This package generates the variance of the set {y1, y2 . yn} of instrument responses and
back calculated data using the calibration curve {aþbx1, aþbx2,. aþbxn}. For a selected prob-
ability (a value) the ANOVA generates the results of an F-test (see expression 1.2.21) by
comparing the ratio F of variance within the sets by that between the sets with a Fcritical value
generated by the F-distribution. For F > Fcritical, the hypothesis that there is no difference be-
tween the sets (null hypothesis) is rejected and if F < Fcritical the two sets are not statistically
different. The ANOVA package also calculates a parameter P. For a P < a, the null hypothesis
should be rejected and for P > a the null hypothesis cannot be rejected (the two sets are not
statistically different). The ANOVA package has several other options such as the comparison
based on two variables. The comparison of experimental instrument responses {y1, y2 . yn}
and back calculated data using the calibration curve it can be assessed the correctness of the
calibration.
One additional aspect of comparison of two sets {x1, x2 . xn} and {y1, y2 . yn} expected as
having a linear correlation is their scatter from the values y ¼ a þ bx for the range of variation.
A measure of the joint variability of the two variables is given by covariance. The covariance
for x and y is defined by the following expression:
1X n
covðx; yÞ ¼ ðxi mx Þðyi my Þ (1.2.48)
n i¼1
where mx and my are the averages of the results for the two variables. If the scatter of the two
variables is the same, they are indicated as homoscedastic. It is not uncommon that the scatter
varies and is larger for larger values of x and y. In this case the variables are indicated as
heteroscedastic.
The linearity of a calibration curve of the form y ¼ a þ bC where H concentration values
{C1, C2,.CH} are considered as not affected by errors can also be tested using a statistical pro-
cedure. For this purpose, the set of data generated by variable Fexp is first obtained by
analyzing samples of each concentration a number of K times such that to generate a set of
N ¼ H$K results. The results are compared for a selected a ¼ 1P to the critical level for
an F-distribution Fcritical (d1,d2,a) where d1 ¼ Ne2 and d2 ¼ NeH. The expression for Fexp is
given by the following formula [19]:
2
Fexp ¼ ðSy =Sz Þ (1.2.49)
For Fexp larger than Fcritical (d1,d2,a), the calibration is not linear, and for Fexp smaller
(for the selected a), the calibration can be considered linear.
The minimum mean-square error fitting can be applied not only to linear dependences but
also to other types of dependence such as the quadratic one. The value for R2 is identical for
the dependence y ¼ FðxÞ and the dependence x ¼ F1 ðyÞ when F expresses a linear depen-
dence, but the R2 values are not identical for other types of dependence.
Also, the minimum mean-square error fitting can be applied to a set of data using a
weighted least square fitting. In a dataset obtained, for example, for the calibration content
in analyte versus instrumental response, some {x,y} values are small, while others are large.
Assuming, for example, that the errors in yi are about 5%, a much larger error will occur for
the larger yi values as compared to the smaller yi values. Since the least square errors apply to
the differences ri (see expression 1.2.39), the impact of larger {xi,yi} values will be larger than
those for small {x,y} values in generation of the regression line (or curve). This problem can be
corrected by generating a weighed least square curve. For this reason, a set of “weights” {wi}
is utilized and the minimization from expression 1.2.40 takes the following form:
X
n
wi r2i ¼ min (1.2.51)
i¼1
Common selections for {wi} are {1/xi) or {1/x2i }, although other selections can be made, if
the importance of some points is higher than for other.
The least square regression can also be applied to nonlinear dependencies such as
quadratic dependence of the form y ¼ a þ bx þ cx2. For the specific dependence, the optimum
parameters that minimize the residuals ri can be obtained. For quadratic dependence, for
example, the residuals are of the form ri ¼ yi (a þ bxi þ cx2i ). The capability to calculate
parameters a, b, c, as well as R2 are commonly available in many computer packages such
as Excel either using the graphics capabilities or LINEST function (the calculations are per-
formed based on relatively complicated expressions not adequate for manual calculation).
These packages also have the capability to calculate the optimum parameters for the best
fit with experimental data for other nonlinear dependencies such as higher power polyno-
mials, logarithmic, power, etc., weighed or not. Also, the data processing capabilities of
some analytical instruments provide various procedures for linear, nonlinear, and weighed
or not least square calibrations (e.g., MassHunter package from Agilent).
Propagation of uncertainty
In any analytical measurement, the random error affecting a result xi may be caused by the
errors in different steps in the analytical process. It is important to understand how these
The total derivative of x obtained from eq. 1.2.52 can be expressed by the following
formula:
vx vx vx
dx ¼ dw þ dy þ dz (1.2.53)
vw y;z vy w;z vz w;y
vx vx
where the cross terms are of the type 2 dwdy, etc. An expression similar to eq. 1.2.54 can
vw vy
be assumed to exist when the differentials dx, dw, dy, and dz are replaced by small deviations
dx, dw, dy, and dz from the mean value for any measurement “i.” In this case, eq. 1.2.54 can be
written in the following form [22]:
2 2 2
vx vx vx
ðdxÞ2 ¼ ðdwÞ2 þ ðdyÞ2 þ ðdzÞ2 (1.2.55)
vw y;z vy w;z vz w;y
In eq. 1.2.55, the cross terms are neglected since dxi, dwi, dyi, and dzi representing random
deviations can be either positive or negative, and can be considered that the cross terms are
canceling one another. Taking the sum over all “i” values for a number n of measurements
in eq. 1.2.55 and dividing by n1 the following formula is obtained:
P
n
2 P
n
2 P
n
2 P
n
2
ðdxi Þ 2 ðdwi Þ 2 ðdyi Þ 2 ðdzi Þ
1 vx 1 vx 1 vx 1
¼ þ þ (1.2.56)
n1 vw y;z n1 vy w;z n1 vz w;y n1
P
n P
n P
n P
n
ðdxi Þ2 ðdwi Þ2 ðdyi Þ2 ðdzi Þ2
In eq. 1.2.56 the ratios n1 ; n1 ; n1 , n1 represent the variances s2, and for n
1 1 1 1
high enough s2 for each variable. From relation 1.2.56 it can be concluded that the depen-
dence of the variance of x on the variance from other operational steps is given by the
following general expression:
2 2 2
vx vx vx
s2x ¼ s2w þ s2y þ s2z (1.2.57)
vw y;z vy w;z vz w;y
First carve them entirely into joints, then remove the bones,
beginning with the legs and wings, at the head of the largest bone;
hold this with the fingers, and work the knife as directed in the
receipt above. The remainder of the birds is too easily done to
require any instructions.
TO ROAST A TURKEY.
Cradle Spit.
Prepare as for boiling a fine well-kept hen turkey; wipe the inside
thoroughly with a dry cloth, but do not wash it; throw in a little salt to
draw out the blood, let it remain a couple of hours or more, then
drain and wipe it again; next, rub the outside in every part with about
four ounces of fine dry salt, mixed with a large tablespoonful of
pounded sugar; rub the turkey well with these, and turn it every day
for four days; then fill it entirely with equal parts of choice sausage-
meat, and of the crumb of bread soaked in boiling milk or cream, and
wrung dry in a cloth; season these with the grated rind of a large
lemon and nutmeg, mace, cayenne, and fine herbs, in the same
proportion as for veal forcemeat (No. 1, Chapter VIII). Sew the turkey
up very securely, and when trussed, roll it in a cloth, tie it closely at
both ends, put it into boiling water, and boil it very gently between
three and four hours. When taken up, sprinkle it thickly with fine
crumbs of bread, mixed with plenty of parsley, shred extremely
small. Serve it cold, with a sauce made of the strained juice and
grated rind of two lemons, a teaspoonful of made mustard, and one
of pounded sugar, with as much oil as will prevent its being more
than pleasantly acid, and a little salt, if needed; work these together
until perfectly mixed, and send them to table in a tureen.
This receipt was given to us abroad, by a Flemish lady, who had
had the dish often served with great success in Paris. We have
inserted it on her authority, not on our own experience; but we think it
may be quite depended on.
TO ROAST A TURKEY POULT.
Season the inside with a little pepper and salt, and roast the goose
at a brisk fire from forty to fifty minutes. Serve it with good brown
gravy only. To this sorrel-sauce is sometimes added at not very
modern English tables, Green geese are never stuffed.
TO ROAST A FOWL.
(A French Receipt.)
Fill the breast of a fine fowl with good forcemeat, roast it as usual,
and when it is very nearly ready to serve take it from the fire, pour
lukewarm butter over it in every part, and strew it thickly with very
fine bread-crumbs; sprinkle these again with butter, and dip the fowl
into more crumbs. Put it down to the fire, and when it is of a clear,
light brown all over, take it carefully from the spit, dish, and serve it
with lemon-sauce, and with gravy thickened and mixed with plenty of
minced parsley, or with brown gravy and any other sauce usually
served with fowls. Savoury herbs shred small, spice, and lemon-
grate, may be mixed with the crumbs at pleasure. Do not pour gravy
over the fowl when it is thus prepared.
TO ROAST A GUINEA FOWL.
Let the bird hang for as many days as the weather will allow; then
stuff, truss, roast, and serve it like a turkey, or leave the head on and
lard the breast. Send gravy and bread-sauce to table with it in either
case: it will be found excellent eating.
3/4 to 1 hour.
FOWL À LA CARLSFORS. (ENTRÉE.)
Bone a fowl without opening the back, and restore it to its original
form by filling the vacant spaces in the legs and wings with
forcemeat; put a roll of it also into the body, and a large sausage
freed from the skin on either side; tie it very securely at both ends,
truss it with fine skewers, and roast it for a full hour, keeping it basted
plentifully with butter. When appearance is not regarded, the pinions
may be taken off, and the legs and wings drawn inside the fowl,
which will then require a much smaller proportion of forcemeat:—that
directed for veal will answer quite well in a general way, but for a
dinner of ceremony, No. 17 or 18 of the same Chapter, should be
used in preference. The fowl must be tied securely to the spit, not
put upon it. Boned chickens are excellent when entirely filled with
well-made mushroom forcemeat, or very delicate and nicely
seasoned sausage-meat, and either roasted or stewed. Brown gravy,
or mushroom sauce should then be sent to table with them.
BOILED FOWLS.
Either of these, when merely split and broiled, is very dry and
unsavoury eating; but will be greatly improved if first boiled gently
from five to ten minutes and left to become cold, then divided, dipped
into egg and well seasoned bread-crumbs, plentifully sprinkled with
clarified butter, dipped again into the crumbs, and broiled over a
clear and gentle fire from half to three quarters of an hour. It should
be served very hot, with mushroom-sauce or with a little good plain
gravy, which may be thickened and flavoured with a teaspoonful of
mushroom-powder mixed with half as much flour and a little butter;
or with some Espagnole. It should be opened at the back, and
evenly divided quite through; the legs should be trussed like those of
a boiled fowl; the breast-bone, or hat of the back may be removed at
pleasure, and both sides of the bird should be made as flat as they
can be that the fire may penetrate every part equally: the inside
should be first laid towards it. The neck, feet and gizzard may be
boiled down with a small quantity of onion and carrot, previously
browned in a morsel of butter to make the gravy; and the liver, after
having been simmered with them for five or six minutes, may be
used to thicken it after it is strained. A teaspoonful of lemon-juice,
some cayenne, and minced parsley should be added to it, and a little
arrow-root, or flour and butter.
1/2 to 3/4 hour.