Professional Documents
Culture Documents
Validación de Métodos Analíticos
Validación de Métodos Analíticos
CONTENTS SUMMARY
This system is based on a few fundamental a measurement process enables a user to make
concepts: technically and administratively correct decisions for
1. The laboratories that produce analytical a stated purpose”, is more related to the results of
information must operate under quality assurance the analytical method than to the method itself. In a
principles. The best way to guarantee that they do is very recent joint project between ISO and AOAC
to ensure that they are accredited by the ISO/IEC International, IUPAC states that fitness-for-purpose
Guide 17025 (1999). is not only related to statistically-based performance
2. The laboratory must participate in proficiency criteria but also to practicability and suitability criteria
testing schemes, which should be designed and (IUPAC 1999a). So, it is important to evaluate,
conducted in accordance with the “International among other characteristics, the ease of operation or
harmonized protocol for proficiency testing of the cost. In fact, method validation is the means that
analytical laboratories” (Thompson 1993). analysts have of demonstrating that the method is
3. The laboratory must use internal quality fit-for-purpose, showing to the customers that they
control procedures which comply with the can trust the reported results.
“Harmonized guidelines for internal quality control in This present concept of method validation
analytical chemistry” (Thompson 1995). introduces the need for versatility in the laboratory
4. The laboratory must use validated methods and, therefore, compels analysts to adapt the
of analysis. analytical methodologies to the different problems to
In other words, if the laboratories that produce the be solved. For instance, there is no need to calculate
results are compared to other similar laboratories the limit of detection for a method used to determine
and can show that they use good tools properly oleic acid in olive oils but there is a need when
(validated methods under quality assurance another method is used to determine trace residues
conditions), that they internally control their in the same oil sample. On the other hand, the need
processes and demonstrate proficiency by for versatility introduces some complexity. The
comparing their performance to similar laboratories, requirements should be defined by first taking into
there are reasons to think that the general quality of account the needs of the end user (something that is
the laboratories is translated to their individual not evident in all cases) and they should then be
results. confirmed by using calibrated instruments, trained
In this paper, we shall discuss the concept of analysts and all other quality assurance conditions.
validated methods of analysis and explain their
different elements after we establish the close 3. ANALYTICAL REQUIREMENTS AND
relationship between validation and fitness for PERFORMANCE CRITERIA
purpose. We shall also describe the analytical
requirements and performance characteristics and Once the specific analytical problem has been
provide useful practical information such as which clearly defined, the producer of the analytical
basic norms to follow in a validation process or how information should ideally be able to agree with the
to conduct validation studies in house or by using customer about the specific analytical requirements.
inter-laboratory comparison studies. For instance, if a laboratory carries out the quality
control of a product, legislation often determines
2. VALIDATION AND FITNESS-FOR-PURPOSE particular limits for certain parameters. The analyst
must ensure that the test results will reflect the
Validation is defined as the confirmation by characteristics of the analysed sample and not those
examination and provision of objective evidence that of the performance method used to obtain the
the particular requirements for a specified intended results. Therefore, the estimated value must not only
use are fulfilled (ISO 1994b). be traceable but also be associated to an appropriate
This definition implies that analytical methods level of uncertainty so as to ensure the compliance
should be validated taking into account the or not of that product with the regulations. Analysts
requirements of specific applications. Therefore, it is also know that they must use a method whose limit
misleading to think that there is a single process of of quantification for that parameter is well below the
method validation that will demonstrate that a legal limit. At the same time, the rapidity with which
method fulfils a list of requirements and that it can be results are provided, the cost of the analysis and the
used in a general way for many different applications. availability of trained staff should also be evaluated.
On the contrary, the definition implies that the In fact, the end user is often not well defined. The
specific requirements for an intended use should be customer may take different forms: it could be an
defined before the performance capabilities of the internal or external client to the own company, it
method are specified. could be the administration that is usually guided by
Fitness for purpose, defined by IUPAC in the the current legislation or even a written standard that
Orange Book as “degree to which data produced by must be followed. It is not unusual to meet customers
130 Grasas y Aceites
are not available. This has the disadvantage that bias The t value calculated is then compared with the
may be incorrectly estimated if the spiked analyte two-sided t tabulated value, ttab, for α =0.05 and the
has a different behaviour than the native analyte. On effective degrees of freedom, νeff, obtained with the
the other hand, a spike can be a good representation Welch-Satterthwaite approach (1941).
of the native analyte in case of liquid samples and in 2
analytical methods that involve the total dissolution s2
u(c ref ) 2 +
or destruction of the samples (IUPAC, 1999b).
n (2)
Otherwise, the different behaviour of the native ν eff =
analyte can be investigated using a less u(c ref ) 4 s4
+
representative reference material (Barwick 1999). ν eff n −1
The assessment of trueness depends on whether
the method is intended to be used in a restricted where νref are the degrees of freedom associated
concentration range or in large concentration ranges. with u(cref). If this information is not available, we can
In the following sections, we describe how to assess assume that ν=n-1. The trueness of the method is
trueness in both situations. assessed if t ≤ ttab.
3.1.5. Assessment of trueness in restricted Case b. Trueness is assessed against a reference
concentration ranges method
Trueness is assessed by assuming that the bias is Trueness is assessed by analysing a test sample
the same in the whole concentration range. This bias both with the reference method and the method to be
is calculated using a reference sample similar to the validated. This test sample should be analysed with
routine samples. This reference sample should be both methods in different runs (i.e. in different days,
analysed n times with the method to be validated in different operators, etc.). An average value, cref, and
different runs, i.e. in different days, with different a variance, s2ref, can be obtained from the nref results
operators, etc. It is recommended to analyse the obtained with the reference method. Likewise, an
reference sample at least 7 times (preferably n ≥ 10) average value, c, and a variance, s2, can be obtained
to obtain an acceptable reliability (IUPAC 1999a). An from the n results obtained with the method to be
average value, cmethod, and a standard deviation, s, validated. The t test used to assess trueness
can be calculated from these n measurements. depends on whether the difference between the
Trueness is then assessed by statistically variances of both methods is statistically significant.
comparing the mean value obtained with the This is checked with an F test:
analytical method, cmethod, with the reference value,
cref. This comparison is done with a t test. The type of s12
t test applied depends on the reference used. We can F= (3)
s 22
distinguish two different cases:
a) Trueness is assessed against a reference
value (here we consider CRMs, reference materials, in which s21 is the larger of the two variances. The F
proficiency testing and spiked samples) value calculated is compared with the two-sided F
b) Trueness is assessed against a reference tabulated value, F(α,ν1,ν2), for α =0.05 and the
method. degrees of freedom ν1=n1-1 and ν2=n2-1.
If F ≤ F(α,ν1,ν2), the difference between both
Case a. Trueness is assessed against a reference variances is not statistically significant. Therefore,
value trueness is verified with the following t test:
The following t test is used to assess trueness:
cref − c
t=
c ref − c s ·( n − 1) + s ⋅ (nref − 1) 1
2 2
1 (4)
t= (1) ref
⋅ +
2
s n + nref − 2 n nref
u(c ref ) 2 +
n
The t value calculated is then compared with the
where u(cref) is the standard uncertainty of the two-sided t tabulated value, ttab, for α=0.05 and the
reference value. For instance, if the certified value of degrees of freedom n+nref-2. The trueness is
a CRM is expressed as cref ± U(cref) (with k=2), u(cref) assessed if t ≤ ttab.
is calculated as U(cref)/k (e.g. if the certified If F>F(α,ν1,ν2), the difference between both
concentration of copper is 10.1± 0.1 mg/l (with k=2), variances is statistically significant. Therefore,
u(cref)=0.1/2=0.05 mg/l). trueness is assessed with the following t test:
132 Grasas y Aceites
1
2 *
* R=
c O+ S − c O (8)
* cref
·100
Youden Method (Youden 1947). This method consists estimated by interlaboratory studies (ISO 5725
of estimating constant bias by analysing different 1994a). On the contrary, repeatability gives the
amounts of a sample. smallest value of precision because the results are
For each reference sample, the “apparent obtained by the same operator, with the same
recovery” should be calculated at least 7 times equipment and within short intervals of time.
(preferably n ≥ 10). An average apparent recovery, R, Intermediate precision relates to the variation in
and a standard deviation, s(R), can be obtained from results when one or more factors, such as time,
the n apparent recoveries calculated. Trueness is equipment and operator, are varied within a
then assessed by statistically comparing the mean laboratory. Different types of intermediate precision
apparent recovery, R, with the 100%. This estimates are obtained depending on factors
comparison can be done with the following t test: changed between the measurements (ISO 5725
1994a). For instance, the time-different intermediate
precision is obtained when the test results are
100 − R obtained in different days. If results are obtained in
t= (9)
s(R ) different days and by different operators, we obtain
the [time+operator]-different intermediate precision.
n Finally, if the laboratory obtains the results by
changing all the factors that may affect the results
This t test considers that recoveries are (i.e. day, operator, instrument, calibration, etc.) we
expressed as a percentage and that the uncertainty obtain the [time + operator + instrument + calibration
of the reference value is negligible. The t value +...] different intermediate precision. This
calculated is then compared with the two-sided t intermediate precision is also known as the
tabulated value, ttab, for α=0.05 and n-1 degrees of run-different intermediate precision. For a laboratory,
freedom. There can be two possibilities: this is the most useful precision because it gives an
a) t ≤ ttab: the apparent recovery does not differ idea of the sort of variability that the laboratory can
significantly from 100%. The trueness of the method expect of its results and, also, because it is an
is assessed. Therefore, there is no reason to correct essential component of the overall uncertainty of the
results by the apparent recovery. results.
b) t > ttab: the apparent recovery differs The within-laboratory precision of an analytical
significantly from 100%. Therefore, the trueness of method should be characterised by the repeatability
the method can not be assessed. In this situation, and the run-different intermediate precision. This
there is a current debate focused on whether results latter precision is an estimate of the “internal
should be corrected by the apparent recovery. reproducibility” of the laboratory. Precision should be
IUPAC, ISO and EURACHEM recommend to correct calculated with test samples that have been taken
the results. However, AOAC international does not through the whole analytical procedure. These test
agree that results should be corrected as a general samples may be reference materials, control
policy (IUPAC 1999b). materials or actual test samples. It is important to
Table I
ANOVA table and calculation of variances for the design proposed in Figure 2
Replicate p n
∑∑ (xij − xi ) 2
MSe =
i j=1 =1 p·(n-1) σ e2
p ⋅ (n − 1)
Table II
Calculation of the precision estimates for the experimental design proposed in Figure 2
2
Between-run variance, s run
MSrun − MSe
n
Run-different intermediate variance, sI(run) 2
s run + s r2
note that we should not estimate precision using the run-different intermediate precision and the
synthetic samples that do not represent the test between-run precision (Table I). Table II shows the
samples (IUPAC 1999a). Moreover, if the analytical expressions for calculating the precision estimates.
method is intended to be used over a large It is recommended that the number of replicates
concentration range, precision should be estimated per day be equal to 2 (n=2) and to focus the effort on
at several concentration levels across the working the number of runs in which the test sample is
range. For instance, precision can be estimated at analysed (e.g. if we want to estimate precision using
low, medium and high concentration levels. a series of 20 measurements, it is better to analyse
twice the sample in 10 runs than to analyse 4 times
3.2.1. Calculation of the within-laboratory the sample in 5 runs) (Kuttatharmmakul 1999). It is
precision also recommended to analyse the sample in at least
7 different runs (preferably p ≥ 10) to obtain good
The simplest method for estimating precision
precision estimates. The advantage of this
within one laboratory consists of calculating the
methodology is that it provides better precision
standard deviation of a series of n measurements. It
estimates than the simplest method.
is recommended that at least 7 measurements
(preferably n ≥ 10) should be carried out to obtain 3.2.2. Calculation of the repeatability limit
good precision estimates (IUPAC 1999a). The
repeatability standard deviation is obtained when the The repeatability limit, r, is the value below which
n measurements are obtained under repeatability the absolute difference between two single test
conditions and the run-different intermediate results obtained under repeatability conditions may
precision is obtained when the measurements are be expected to lie with a probability of 95%. This limit
obtained by changing the factors that may affect the is obtained as (ISO 5725 1994a):
analytical results (i.e. operator, day, equipment, etc.).
An alternative procedure consists of performing n r = 2.8.sr (10)
replicates in each of the p runs in which the test
sample is analysed. This experimental design 3.2.3. Calculation of the “internal reproducibility”
corresponds to a two-factor fully nested design (ISO limit
5725 1994a). Here the factors studied are the run
and the replicate (Figure 2). The use of the Analysis An “internal reproducibility” limit, R, can be
of the Variance (ANOVA) provides the repeatability, calculated from the run-different intermediate
standard deviation. This limit is the value below which
Vol. 53. Fasc. 1 (2002) 135
the absolute difference between two single test In the assessment of trueness it is checked
results obtained under run-different intermediate whether the bias of the analytical method is
conditions may be expected to lie with a probability of significant. If bias is not statistically significant, the
95%. This limit is obtained as: trueness of the method is assessed. However, it
remains an uncertainty due to the estimation of the
R = 2.8.sl(run) (11)
operator, day, material, etc.). The uncertainty can the working range. For instance, in the determination
then be calculated as: of calcium in edible oils using atomic absorption
spectroscopy, the working range comprises any oil
u pretreatments = s 2pretreatments − sconditions
2 (14) containing calcium while the linear range is normally
between 1 and 10 ppm.
An initial estimation of the working and linear
where s2pretreatments is the variance of the results ranges may be obtained using blank and reference
obtained after analysing the different portions. materials (or a blank and fortified sample blanks) at
s2conditions depends on the conditions in which we several concentration levels. Ideally, at least six
analyse the different portions subsampled and/or concentration levels plus blank are needed. One
pre-treated. If they are analysed under repeatability replicate at every concentration level is enough to
conditions, it corresponds to the variance of the visually identify the linear range and the boundaries
repeatability, i.e. s2conditions= s2r. If they are analysed of the working range. After this primary estimation, a
under intermediate precision conditions, it more accurate plot can be obtained using reference
corresponds to the run-different intermediate materials or fortified sample blanks. At least, six
precision, i.e. s2conditions= s2I(run). different concentration levels, using at least three
replicates of each level, are necessary. The linearity
3.3.3. Uncertainty of other terms of the calibration line can be assessed using the
correlation coefficient and the plot of the residual
This term contains all the sources of uncertainty
values. The residual values (ei, with i=1… n, where n
that have not yet been considered in the former
is the number of concentration levels used) in a
terms. Most of these sources correspond to factors
straight-line model are calculated by subtracting the
not representatively varied in the estimation of the
model estimates from the experimental data
intermediate precision. The uncertainty of these
(ei=yi-b0-b1xi, where b0 and b1 are respectively the
factors can be calculated using information from
intercept and the slope of the straight line).
robustness studies (Barwick 2000), from
A high value of the correlation coefficient and a
bibliography or from other sources such as previous
good plot of the residual values are normally enough
knowledge.
to assess linearity. It is important to point it out that
the sole use of the correlation coefficient does not
3.4. Linear and working ranges always guarantee the linearity of the calibration line,
even if the correlation coefficient has value near to
In all the quantitative methods it is necessary to one. A good plot of the residual values has to fulfil the
determine the range of analyte concentrations or following conditions:
property values for which the analytical method can – the residual values must not show tendencies
be applied. The lower end of the working range is – the residual values must be more o less
limited by the values of the limits of detection and randomly distributed
quantification. At the upper end, several effects limit – the number of negative and positive residual
the working range. values has to be approximately the same the
Within the working range there may exist a linear residual values should have approximately the
response range in which the method gives results same absolute value.
that are proportional to the concentration of analyte. Figure 4 shows some examples of plots of
Obviously, the linear range can never be wider than residual values. Figure 4a shows a plot with a good
ei a) ei b) distribution of residual values, where all the residual
values fulfil the aforementioned requisites. Figure 4b
shows a plot where the residuals do not have the
x x same absolute value, probably showing the need for
a weighted calibration method. In Figure 4c the
residual values clearly show tendencies (probably
xi xi indicating the presence of non-linearities). Figure 4d
ei c) ei d) shows the presence of a probable outlier in the data
set.
x Experimental points with an abnormally high
x residual value may be outliers, and they should be
carefully examined using a test for the detection of
outliers (for instance, the test of Cook (Meloun 2002).
xi xi If a more accurate check of the linearity is
desired, a test for the lack of fit based on the analysis
Figure 4 of variance (ANOVA) can be used alongside with the
Different plots of residual values.
Vol. 53. Fasc. 1 (2002) 137
correlation coefficient and the plot of the residual where z1-α is the upper- αpercentage point of a
values (Massart 1997). normal distribution and σo is the standard deviation of
It is important to note that the calibration line is the net concentration when the analyte is not present
usually found using the ordinary least squares in the sample (i.e. the true value is zero). The Lc is
method, but if the variance of the replicates at each defined in order to mark a minimum value for which a
concentration level is not constant through all the predicted concentration is considered as being
linear range (giving rise for instance to the plot of caused by the analyte. By doing so, there exists a
residual values shown in Figure 4b), then a better risk α of committing a type I error, i.e., a false positive
option is to use the weighted least squares method, decision, that means stating that the analyte is
which takes into account the individual variance present when in fact it is not. However, if we want to
values in each calibration point (Massart 1997). keep the risk of a false negative decision (called β)
low, the LOD of the method, LD, must be higher by
3.5. Limit of detection (LOD) taking into account both probabilities of error:
implies experimental errors (in the measurement of equivalent to twenty times the width at half weight of
the blank sample and the analysed sample), which the peak of the analyte. The LOD is expressed as:
are assumed to be random and normally distributed.
At zero concentration level, the standard deviation of LD = 2z1-αhnoiseR (22)
the net concentration is expressed in a general way where hnoise is the average of the maximum
as: amplitudes of the noise and R is the response factor
σ 0 = σ Bφ (20) (concentration/peak height). This procedure is valid
only when the chromatographic peak heights are
used for quantification purposes. If peak areas are
σ B is the standard deviation of the blank and φ =
used instead, then the LOD has to be estimated by
√ 1/m + 1/n, where m and n are the number of other means, such as the described above.
replicates on the analysed sample and on the blank
Finally, it is highly recommended when reporting
sample, respectively. φ = 1 ( σ0 = σ B) in the special
the detection limit of a given analytical method, to
case when m=1 and n is high. When the net
specify the approach used for its calculation.
concentration is calculated in a paired experiment
(i.e. as the concentration in the sample minus the 3.6. Limit of quantification (LOQ)
blank) then φ.= √ 2 . If σ B is not known, then it has to
be replaced by its corresponding estimate, sB, in Eq. The limit of quantification, LOQ, is a performance
(20). For a reliable estimation of the detection limit, a characteristic that marks the ability of a chemical
minimum suggested number of 10 determinations measurement process to adequately quantify an
should be carried out on the blank sample. The analyte and it is defined as the lowest amount or
precision conditions (repeatability, intermediate concentration of analyte that can be determined with
precision) in which the blank is analysed should be an acceptable level of precision and accuracy
clearly specified. (EURACHEM 1998). In practice, the quantification
limit is generally expressed as the concentration that
3.5.2. Alternatives for LOD calculation can be determined with an specified relative
standard deviation (usually 10%). Thus:
If the calibration line used to check the linearity of
the analytical method has been built in a restricted LQ = k Q σ Q (23)
concentration interval close to the expected limit of
detection, then this curve can be used to calculate where kQ = 10 if the required RSD=10% (this is the
the LOD of the method. The expression for LOD in recommended value by IUPAC, and σQ is the standard
this case is: deviation of the measurements at the level of the limit of
quantification. To estimate σQ a number of independent
determinations (n > 10) must be carried out on a
sy / x 1 1 x2 sy / x 1 1 (L − x ) 2
L D = t1−α ,ν + + N
+ t1− β ,ν + + N D sample which is known to contain the analyte at a
b1 m N b1 m N
∑ (x
i =1
i − x)2 ∑ ( xi − x ) 2
i =1
concentration close to LQ. Since LQ is not known,
some guides recommend to analyse a sample with a
(21) concentration between 2 and 5 times the estimated
where: detection limit (IUPAC 2001). Other guidelines (EURA-
sy/x: is the standard deviation of the residuals of the CHEM 1998) suggest to perform the determinations
calibration line on a blank sample. Such a procedure, however,
N: is the number of calibration standards is discouraged, unless there is a strong evidence that
xi: is each of the concentrations of the calibration the precision is constant between c = 0 and c = LQ.
standards When determining σQ, the sample has to be taken
x: is the mean value of the concentrations of the cali- to the whole analytical procedure. Also, the precision
bration standards. conditions in which the sample is analysed must be
The LOD can be calculated by solving Eq (21) or clearly specified.
by an iterative procedure. Usually, the LOQ estimation is carried out as a
In chromatographic methods of analysis, in which part of the study to determine the working range of
the peak corresponding to the analyte has to be the analytical method and it is a common practice to
distinguished from the chromatographic baseline, fix the LOQ as the lowest standard concentration of
the detection limit can be calculated by measuring the calibration range.
the noise of a chromatogram of several (usually
three) sample blanks, which have been submitted to 3.7. Selectivity (specificity)
the whole analytical process. The procedure
recommended by the SFSTP (STP Pharma In the majority of analytical methods it is
Practiques, 1992), consists on determining the necessary to ensure that the signal produced in the
maximum amplitude of the baseline in a time interval measurement stage is only due to the analyte of
Vol. 53. Fasc. 1 (2002) 139
interest, and not to the presence of interferences in sensitivity is simply the slope of the calibration
the sample. Hence it is necessary to test the straight-line.
selectivity (or specificity) of the analytical method.
The terms selectivity and specificity are often 3.9. Ruggedness (or robustness)
interchanged. Selectivity (or specificity) can be
defined as the ability of a method to accurately and The robustness/ruggedness of an analytical
specifically determine the analyte of interest in the procedure is a measure of its capacity to remain
presence of other components in a sample matrix unaffected by small, but deliberate variations in
under the stated conditions of the test (EURACHEM method parameters and provides an indication of its
1998). Some authors consider specificity to be 100% reliability during normal usage (ICH 1994). ISO and
selectivity, and hence make a difference between the IUPAC have not yet provided definitions and
two terms: specificity is referred to a method that only recommendations to evaluate ruggedness, but
gives response for a single analyte, while selectivity Vander Heyden et al (2001) has extensively covered
refers to a method that gives responses for a number the subject.
of chemical entities that may or may not be Where different laboratories use the same
distinguished from each other. This situation creates method, they inevitably introduce small variations in
unnecessary confusions and can be avoided by the procedure, which may or may not have a
authors by giving preference for the use of selectivity significant influence on the performance of the
(IUPAC 2001). IUPAC clarified this overlap method. The ruggedness of a method is tested by
expressing the view that ‘specificity is the ultimate of deliberately introducing small changes to a number
selectivity’ (den Boef 1983). of parameters of the method and examining the
From a practical point of view, the selectivity of a effect of these changes on the accuracy and
method may be checked studying its ability to measure precision of the final results.
the analyte of interest when other specific interferences The evaluation of ruggedness is normally carried
have been introduced in the sample. In a similar way, out by the individual laboratory, before this
taking into account the matrix effect, which can participates in a collaborative trial, and should be
enhance or suppress the sensitivity of the method. This considered at an earlier stage during the
is usually carried out analysing reference materials and development and/or optimisation of the analytical
samples containing the analyte of interest and various method.
suspected interferences with both the candidate and A large number of parameters may need to be
other independent analytical methods. From the considered, but because most of these will have a
comparison of the results using all the methods, we can negligible effect, it will normally be possible to vary
assess the ability of the candidate method to analyse several at once. An established technique for
the required analyte in presence of other substances. ruggedness testing is covered in detail by Youden and
Another option, if no independent analytical Steiner (1975) and uses Plackett-Burman
methods are available, is to analyse a series of three experimental designs. Such designs allow the
matrix blanks (or samples containing the analyte of investigation of the influence of several parameters in
interest if matrix blanks are not available), three a limited number of experiments. Typical parameters
method blanks and the lowest concentration can be subdivided in continuous (i.e. extraction time,
standard (or a standard corresponding to the analyte mobile phase composition, temperature, flow rate,
concentration in the positive sample if matrix blanks etc.) or non-continuous (i.e. type of chromatographic
are not available), and assessing their background column, brands, chemicals, etc.).
signals and the significance of any interferences The results from ruggedness testing can also be
relative to the lowest analyte level specified in the used to estimate components of uncertainty which
working range (IUPAC, 1999a). Responses from have not been considered in the assessment of
interfering substances should be less than 1% of the trueness and precision of the method (Barwick 2000).
response of the lowest concentration measured.
Table III
3.8. Sensitivity
Desing to calculate the ruggedness of a method
Sensitivity is defined as the change in the
response of a measuring instrument divided by the Parameter
corresponding change in the stimulus (being the
stimulus for instance the amount of the measurand) Experiment A B C Result
(EURACHEM 1988). Although it clearly applies to the
measuring instrument, the sensitivity can also be 1 + + + y1
applied to the method as a whole. From a practical 2 - + - y2
point of view, it corresponds to the gradient of the 3 + - - y3
response curve. Inside the linear range, the 4 - - + y4
140 Grasas y Aceites
The ruggedness test should be carried out on a 5. PERFORMING THE METHOD VALIDATION
representative sample, preferably a reference
material. The first step consists on identifying and When and who should validate the analytical
selecting the parameters that are going to be tested methodology to be used are also aspects of method
and defining their nominal and extreme levels (i.e. validation that are of practical importance.
flow rate at 1.0 ± 0.2 ml/min). After this, the There are many laboratory situations in which a
experimental design is has to be selected; for method needs to be validated. For instance, the
instance, if three parameters were to be evaluated, a validation required for a new method is clearly
design with only four experiments would be sufficient different from the validation required for a method
as indicated in Table III. The + and – signs denote the that has previously been validated by a method
coded levels for the maximum and minimum values performance study. Validating modifications
of the parameter under study. For each parameter, introduced into a test procedure is different from
there are two experiments at each level. To study the validating a well-known method for testing samples
effect of a given parameter, the mean of the results that incorporate new matrices. In all cases, some
obtained at the - level have to be subtracted from the sort of method validation should be performed.
mean of the results at the + level. In the example, the However, the extent of the validation varies from
effect of parameter A would be calculated as: case to case. This topic is further developed in
section 12.7
Effect A = [(y1 + y3)/2 – (y2 + y4)/2] Clearly, the laboratory that is going to use a
method for analysing test samples is responsible for
guaranteeing that the method is fit-for-purpose.
Once the effects have been calculated for every Therefore, this laboratory should demonstrate that is
parameter, a statistical and/or graphical analysis of valid to analyse the test samples adequately.
the effects must be carried out in order to draw
chemically relevant conclusions and, if necessary, to 6. INTER-LABORATORY COMPARISON AS A
take actions to improve the performance of the MEANS OF METHOD VALIDATION
method.
When a method is going to be used by many
4. THE BASIC PRINCIPLES OF METHOD laboratories throughout the world, it is advisable that
VALIDATION it be validated by means of a collaborative or method
performance study involving a group of laboratories.
There are three basic principles of any process of In this type of studies, all the laboratories follow
method validation: the same written analytical protocol and analyse the
1. The validation must include the complete same samples in order to estimate the performance
analytical procedure. The measurement of the criteria of the method being studied. Usually the
instrumental signal is often considered to be the within-laboratory and among-laboratory precision
main step in the analytical procedure and are estimated and, whenever necessary and
considerable effort is made to establish the linear possible, so are other criteria such as the systematic
range and to calculate the uncertainty derived from error, recovery, sensitivity or limit of detection.
this step. However, nowadays, due to advances in After the consensus reached under the auspices
instrumentation, this step usually contributes little to of the IUPAC, William Horwitz prepared the “Protocol
the bias or to the expanded uncertainty of the final for the Design, Conduct and Interpretation of
result. Special care has to be taken with the sample Method-Performance Studies” (Horwitz 1995). In this
pre-treatment process where the intervention of the document, the design of the method performance
analysts and the non-automated steps has a crucial study is described in considerable detail. For
effect on the performance parameters. instance, at least five different materials are needed
2. The validation must comprise the whole for a single type of substance, a minimum of eight
range of concentrations in which the analyte may be laboratories should participate in the study and the
present in the test samples. designs used to estimate the repeatability and
3. Representativity is an essential factor. The reproducibility are described. The statistical analysis,
method should be validated by considering the range the final estimation of the parameters and the final
of matrices in which the analyte has to be report are also described.
determined. Moreover, the different effects However, collaborative studies have not been
influencing the results during the normal application undertaken for many analytical methods. Among
of the method should be considered during the other problems, they are too time-consuming and
method validation process. This concept will be require expensive resources, particularly for the
further developed when the uncertainty values are laboratory in charge of the organisation. Moreover,
calculated. the general guidelines for method validation are too
Vol. 53. Fasc. 1 (2002) 141
impractical to be used in industrial laboratories or particular analytical problem that we have in our
laboratories that monitor, for instance, trace-level laboratory. (EURACHEM 1998) It is always needed
organic compounds. some degree of validation, although this degree may
Because of these unfavourable factors, the significantly change depending in each particular
Association of Official Analytical Chemists, AOAC situation. Some of these particular situations are
International, has introduced the “Peer Verified (AOAC/FAO/IAEA/IUPAC 2000):
Method Program” (AOAC 1998) so that methods – The laboratory wants to use as a routine
used by laboratories working with only one or two method a previously validated one in another
others can be validated. The aim of this program is to laboratory or group of laboratories: Here a full
provide a class of tested methods whose validation is not necessary, since the absence of
performance has been checked in at least one the method bias has already been assessed.
independent laboratory. The laboratory only needs to verify that it is
capable of achieving the published performance
7. IN-HOUSE VALIDATION characteristics of the method by checking the
absence of laboratory bias, and by calculating its
We have seen that collaborative trials give rise to own precision (repeatability and intermediate
a sound validation of methodologies. But they also precision) estimates. In case of measuring at
present some drawbacks: low concentration values, the laboratory also
– They require to gather enough laboratories needs to calculate the limits of detection and
that can devote the necessary resources to quantification, since these values depend, among
achieve proper validation. others, on the laboratory instrumentation and
– There are not enough scientists willing to on the analysts.
undertake the organisation of the very large – The laboratory wants to use a validated
number of method trials that are required. method, but the analytes to be determined either
– Problems with shipping regulations, import are new or they are found in new sample
restrictions, analyte stability. matrices: In this case the validation has to be
– This type of studies are uneconomic, too slow, extended to the new analytes or matrices.
restricted in scope Parameters like trueness, precision, calibration,
For these and other reasons, there is currently a analytical ranges, and selectivity have to be
move towards giving more importance to the use of checked for every new analyte or matrix.
methods of analysis that have only been validated – The method is published in the scientific
‘in-house’, provided that the laboratory implements literature together with some analytical
appropriate quality assurance procedures. In-house characteristics: The laboratory has to undertake
validation is the process of checking, inside a verification and a limited validation in order to
laboratory, that a method gives correct results. It can meet the requirements.
provide a practical and cost-effective alternative – The method is published in the scientific
approach to other validation systems. literature but not characteristics given : The
Many analysts think that a previously validated laboratory has to undertake a full validation of
standard or reference method may be directly the method and a self-verification in order to
applied in their laboratory. The previous validation of achieve the stated characteristics.
the method is erroneously considered as a sufficient – The method has been developed in-house:
warranty to achieve quality results. This is not true, The laboratory has to arrange a full validation
and when a new method is introduced in a laboratory of the method.
(even if it is a previously validated one by an
Apart from these situations, some other checks
international organism) it has always to be validated,
are routinely used to test the whole analytical
although in this latter case a full validation is not
procedure before starting a new set of analysis. For
needed. As there is some parallelism between the
instance, checking the sensitivity of the method
laboratory of analysis and the cuisine, we can
(slope of the calibration straight line), or the
compare a validated analytical procedure and a
selectivity in some chromatographic assays. These
recipe in a good cookbook. There are many good
latter are often referred as to suitability checks.
previously tested recipes that are very carefully
Another different kind of in-house validation is the
described. Is it enough to get a good dish in our
retrospective validation. The retrospective validation
cuisine? We all know that unfortunately this is not
is the validation of a method already in use based
enough. The same happens with validated methods.
upon accumulated production, testing and control
We can get good results using them, but we need to
data. In some cases an analytical method may have
prove that they adequately work in our laboratory.
been in use without sufficient validation. In these
Hence a method always has to be validated to check
cases, it may be possible to validate, to some extent,
that their quality parameters are valid for the
the adequacy of the method by examination of
142 Grasas y Aceites
STP Pharma Practiques (1992) Guide de validation Vander-Heyden, Y., Nijhuis A., Smeyers-Verbeke, J.,
analitique: Report SFTP. Methodologie and examples 2 Vandeginste, B.G.M., Massart, D.L. (2001) Guidance
(4), 205-226. for robustness/ruggedness tests in method validation.
Thompson, M. Wood, R. (1993) The International J. Pharmaceutical and Biomedical Analysis 24,
Harmonised Protocol for the Proficiency Testing of 723-753.
(Chemical) Analytical Laboratories. Pure Appl. Chem., Youden, W.J. (1947) Technique for testing the accuracy of
65, 2123-2144. Also published in J. AOAC Intl. 76, analytical data. Anal. Chem. 19, 946-950.
926-940. Youden, W.J., Steiner, E.H. (1975) Statistical Manual of the
Thompson, M. Wood, R., (1995) Harmonised Guidelines Association of Official Analytical Chemists. AOAC,
for Internal Quality Control in Analytical Chemistry Arlington, VA.
Laboratories. Pure Appl. Chem., 67, 49-56.