Professional Documents
Culture Documents
net/publication/223735068
CITATIONS READS
17 4,708
1 author:
Joel Tellinghuisen
Vanderbilt University
332 PUBLICATIONS 7,976 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Joel Tellinghuisen on 04 February 2018.
a r t i c l e i n f o a b s t r a c t
Article history: The least-squares analysis of data with error in x and y is generally thought to yield best results when the
Received 10 May 2010 quantity minimized is the sum of the properly weighted squared residuals in x and in y. As an alternative to
Received in revised form 6 July 2010 this “total variance” (TV) method, “effective variance” (EV) methods convert the uncertainty in x into an
Accepted 10 July 2010
effective contribution to that in y, and though easier to use are considered to be less reliable. There are at
Available online 15 July 2010
least two EV methods, differing in how the weights are treated in the optimization. One of these is identical
Keywords:
to the TV method for fits to a straight line. The formal differences among these methods are clarified, and
Least-squares Monte Carlo simulations are used to examine the statistical properties of each on the widely used straight-
Total variance line model of York, a quadratic variation on this, Orear's hyperbolic model, a nonlinear binding (Langmuir)
Effective variance model, and Wentworth's kinetics model. The simulations confirm that the EV and TV methods are
Monte Carlo statistically equivalent in the limit of small data error, where they yield unbiased, normally distributed
Data analysis parameter estimates, with standard errors correctly predicted by the a priori covariance matrix. With
Uncertainty in x and y increasing data error, these properties fail to hold; and the TV method is not always statistically best.
Nonetheless, the method differences should seldom be of practical significance, since they are likely to be
small compared with uncertainties from incomplete information about the data error in x and y.
© 2010 Elsevier B.V. All rights reserved.
1. Introduction which was proposed long ago by Deming [6] and implemented in
iterative form as early as 1972 [7–12]. In Eq. (1) the δ′is are residuals in
Much has been written on the least-squares (LS) analysis of data the uncertain variables, here just x and y. A complicating factor in
having uncertainty in more than one variable, as comprehensively specifying and minimizing STV is the need for “calculated” or
reviewed through 1991 by Macdonald and Thompson [1]. Most such “adjusted” values (designated yi and xi) for both y and x, because
works have dealt with just two uncertain variables, x and y, usually the residuals are Yi − yi and Xi − xi, where upper-case letters represent
assumed to fit a straight line, with the model of York [2] being a the observed values. Minimum-variance estimates of the model
frequent target. This problem has been revisited recently by York et parameters are assumed to result from the use of weights inversely
al. [3], and is the subject of at least two reviews [4,5]. In the present proportional to variance—wxi = Cσ−2 −2
xi , wyi = Cσyi , etc. (C a propor-
work I show that the “prevailing wisdom” on this subject is not tionality constant, the same for all variables)—as can be proven to
completely correct. In particular the accepted “best” method does hold for linear LS, in which there is a single uncertain variable (y).
not always yield statistically best results. Most prior studies have The subscript TV in Eq. (1) is short for “total variance,” to distinguish
dealt just with specific problems and data sets. In the present work, I it from similar quantities labeled “effective variance,” in which the
use Monte Carlo methods to produce typically 105 simulated data variance in x is converted into an effective contribution to the variance
sets for a given model, thereby obtaining what I believe are the first in y = f(x) using error propagation, to yield weights [13–16]
valid statistical comparisons of several approaches to this problem.
−1 2 2 2
The comparisons are limited to the simplest case of random normal w ∝ varðyÞeff = σy + ðdy=dxÞ σx : ð2Þ
(Gaussian) error added to true models.
It is generally thought that the best solution is obtained by For y = a + bx, σy,eff 2 = σy2 + b2σx2, yielding wi ∝ (σy2 + b2σx2)−2. The
minimizing the quantity STV, minimization target is
2
2 2
SEV = ∑wi δyi : ð3Þ
STV = ∑wxi δxi + wyi δyi + …; ð1Þ
This approach attempts to take uncertainty in x into account, but still
treats x as error-free; thus only the y residuals, δyi = Yi − yi, need be
E-mail address: joel.tellinghuisen@vanderbilt.edu. evaluated.
0169-7439/$ – see front matter © 2010 Elsevier B.V. All rights reserved.
doi:10.1016/j.chemolab.2010.07.003
Author's personal copy
J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169 161
The EV method is easy to use with standard programs for 2. Theory and computational methods
weighted linear (LLS) and nonlinear LS (NLS), but it has been
criticized for some shortcomings [17,18]: (1) The minimum in SEV 2.1. Linear least-squares
is not the same as that in STV, and thus the solutions are not
“optimal.” (2) When the weights in x and y are constant (but not The mathematics and procedures of linear and nonlinear least-
necessarily the same), the solution to the straight-line fit becomes squares are readily available from many sources [24,33–35]; here I
just the unweighted or ordinary LS (OLS) solution for regression of will repeat just enough to provide continuity to the modified
y upon x and converges in one iteration. This is true even when the expressions needed for the EV and TV methods. A least-squares
uncertainties in x greatly exceed those in y, making regression of x problem is linear when there is a single uncertain variable (y), and the
upon y more sensible. (3) It follows that the results depend upon relations among the measurements Y, the independent variables (x, t,
the manner in which the fit relation is expressed. In contrast, the u,..., all error-free), the adjustable parameters β, and the fit residuals δ
TV method yields a single solution for all mathematically can be expressed as
equivalent ways of writing the fit relation [19].
With a slight change in the minimization procedures, we obtain Y = Uβ + δ; ð5Þ
162 J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169
must be of finite variance. Then if they are unbiased, the LLS parameter with all quantities evaluated at the observed points (Xi,Yi). Eq. (15) is
estimates β will be unbiased, with variances given exactly by Vprior. If extendable in an obvious way for more than two uncertain variables.
further the data error is normal (Gaussian), the β estimates will This treatment assumes the x- and y-errors are independent from
similarly follow the Gaussian distribution, each other and from those at all other points. With these wi, the
" # computations proceed as in Eqs. (13) and (14), except that through
2
ðu−μ Þ Eq. (15), the weights generally depend upon the parameters and must
PG ð μ; σ; uÞ = C exp − ; ð10Þ
2σ 2 be adjusted in the iterations. This is the EV treatment, and it leads to
minimization of SEV in Eq. (3).
with C a normalization constant and u representing an estimate of βi The weights can be handled differently: Include a factor of w1/2i =
having true value μ and SE σ. This distribution will hold even with 1 / σFi with each element of J and F and replace W by the identity
incorrect weighting, but unless wi ∝ 1 / σ2yi, the estimates of β will not matrix. With one further change, we obtain the EV2 method. Replace
be minimum-variance. With properly weighted, normal data, LLS is the fit relation in Eq. (12) by
also a maximum-likelihood method. Under these conditions, SLLS will
be χ2 distributed (mean ν, variance 2ν) and SLLS/ν will follow the Gðx; y; βÞ = Fðx; y; βÞ = σF : ð16Þ
reduced χ2 distribution (mean 1, variance 2/ν),
Then the elements of the J matrix—now the partial derivatives of G wrt
ðv–2Þ = 2 βj—include the dependence of the weights on the parameters. [These
Prcs ðzÞ = C z expð–vz = 2Þ; ð11Þ
derivatives are commonly computed by finite difference, evaluating G
where z = χ2ν (=χ2 / ν) and C is a normalization constant. Similarly for the ith point at βj + ε and βj − ε and dividing the difference by 2ε;
Vpost-based variance estimates s2βi follow a scaled χ2 distribution. And here this also requires evaluating Eq. (15) at βj ± ε. Again, all
the quantities (βi − βi,true) / sβi follow the t-distribution for ν degrees quantities are evaluated at (Xi,Yi).]
of freedom. In the large-ν limit, both the χ2 and the t distributions
become Gaussian. 2.5. Nonlinear LS: total variance method
From the properties of the χ2 distribution, sampling estimates of
variance have relative uncertainty (2/ν)1/2, and the SEs have relative Most of the foregoing for NLS still applies, but with two changes:
uncertainty (2ν)−1/2. This “uncertainty in the uncertainty” is We now also adjust all uncertain variables, and we evaluate all
generally ignored in data analysis texts; for small ν, it can significantly derivatives on the adjusted curve. Deming [6] expressed the residuals
increase the effective parameter uncertainty—by 25% for a 10-point δx and δy in terms of Lagrangian multipliers λ, given by
model, for example. Formally, the uncertainty in the SE leads to the
λ = F0 −JΔβ ð17Þ
need for the t-distribution to assess parameter confidence limits.
where F0 still designates evaluation with β0 at the observed points X
2.3. Nonlinear LS with a single uncertain variable
and Y, F0i = F(Xi,Yi;β0). (On the calculated curve, convergence should
achieve F = 0.) Then
The equations must now be solved by successive approximation.
Let the fit relation be λi ∂F
δxi = ð18Þ
wxi ∂xi
Fðx; y; βÞ = y−f ðx; βÞ = 0; ð12Þ
and similarly for δyi (and all other uncertain variables). Numerical
such that exactly fitting data give F = 0. Eq. (12) is explicit in the
tests show that the results from this procedure depend weakly on the
dependent variable, so correct weighting is again wi ∝ 1 / σ2yi. Starting
way F(x,y;β) is expressed. However, the modifications of Powell and
with an approximate solution β0, we obtain a new estimate,
Macdonald [7], Britt and Luecke [8], and Jefferys [11,12] remove this
β1 = β0 −Δβ ð13Þ dependence. In place of F0 in Eqs. (14) and (17), we use F0′, having
elements
where, in what is sometimes known as the inverse Hessian method
[35], the corrections Δβ are obtained by solving equations formally ∂F ∂F
F0′ i = Fi + δ + δ +… ð19Þ
the same as Eq. (6), ∂xi xi ∂yi yi
T T where all quantities on the rhs are evaluated on the calculated curve
J W J Δ β = A Δ β = J W F0 : ð14Þ
(giving Fi = 0 on convergence). STV (Eq. (1)) is also correctly given by
F0 contains the n values from Eq. (12), and the elements of the STV = FT0 W F0 (Deming) or F0′T W F0′. [Numerical checks show that
Jacobian matrix J are Jij = (∂Fi / ∂βj). All quantities are evaluated for the Deming's expressions yield F(x,y;β) = 0 on the calculated curve only
ith point using the current parameter estimates β0. These equations in certain cases, like the straight-line fit. Use of F0′ from Eq. (19) gives
become the same as Eqs. (6) and (7) for linear functions f(x;β). this result in all cases.]
The relation between V and A is the same as for LLS, except that
here A (hence V) may include dependence on Y and β, thus may vary 2.6. Properties of NLS solutions
for different data sets. We define the “exact” V as the Vprior for exactly
fitting data. Monte Carlo simulations have demonstrated that NLS solutions for
single uncertain variables behave like LLS in the limit of small data
2.4. Nonlinear LS with implicit and multiple uncertain variables error: unbiased, normally distributed parameter estimates with SEs
given by the exact Vprior, and χ2-distributed S values [24]. With
The fit relation is still expressed as F(x,y;β) = 0, but one or more increasing data error these properties fail to hold. The parameter
uncertain variables do not appear explicitly. For each point i, wi ∝ 1 / σ2Fi, estimates become nonnormal and biased, with confidence limits
where error propagation gives the effective variance for two uncertain exceeding those given by Vprior. A good guideline for the reliability of
variables (x and y) as the SEs is a “10% rule”: If the SE is less than 10% of the magnitude of the
2 2 2 2 2
parameter, it should yield confidence limits reliable within 10%. Most
σF = ð∂F =∂xÞ σx + ð∂F =∂yÞ σy ; ð15Þ previous efforts to check these properties for NLS with multiple
Author's personal copy
J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169 163
uncertain variables, including the TV method, have been flawed, as is 3. Results and discussion
discussed below.
With nonlinear functions f(x;β) there may be multiple minima in 3.1. The York model
S, or none at all; and convergence on a minimum may require quite
close initial values β0. Use of the Marquardt algorithm can greatly York [2] devised strongly heteroscedastic data weights for the
improve the convergence properties for NLS with a single uncertain, variables of a 10-point straight-line model originally considered by
explicit variable [38]. With it, the diagonal elements of A are increased Pearson [40], yielding a model that has served as a touchstone for x-
any time SNLS rises on iteration, enhancing a more robust gradient and-y-error fitting methods for over 40 years [1,4,5]. York and
mode of optimization [34,35]. However, with multiple uncertain colleagues included this example among others in their more recent
variables, there are problems using this method, because convergence study [3], where they used MC methods to check the reliability of
can occur from below (see below). predicted uncertainties for the parameters. They also reported that
histograms of the parameters from all these tests “almost perfectly
matched Gaussian distributions.” My results contradict that claim and
2.7. Monte Carlo simulations support a different interpretation of the reliability of the parameter
SEs.
Simulated data are generated by adding random normal error in x For simplicity here, the true model was taken to have the rounded
and y to points on the true curve, using standard methods [24,35]. The values, α = 5.5 and β = −0.48. York's weights were converted to data
resulting data are then fitted to the model under study by the TV and σ values for x and y using, e.g., wxi = 1 / σ2xi, yielding the model
EV methods. This cycle is repeated for typically 105 simulated data illustrated in Fig. 1. Results from MC simulations run as a function of a
sets, with evaluation of means and variances for key quantities. The scaling factor f for the data errors (Table 1) show that the parameter
MC precisions are governed by the estimated standard errors from the biases increase as the square of the data error (hence as f2) and
MC statistics, e.g., sa/N1/2 for the parameter a when it and its standard become statistically significant beyond f = 0.25. However, at the
error are estimated from N analyzed data sets. The uncertainty in sa is precision level of 105 simulated data sets, the sampling SEs follow
governed by the statistics of χ2, so it is uncertain by 1 part in (2N)1/2, predictions until f = 1. Fits of histograms of the parameters (Fig. 2)
which is 0.224% for N = 105. Also, when the MC results are binned to were Gaussian for f = 0.1 but clearly not for f = 0.25 and larger. (For
yield histograms and the latter are fitted, the bin counts are weighted ν = 28, values of χ2ν as large as 2.0 occur less than 0.1% of the time
in accord with Poisson statistics [34], meaning they are assigned [34].) On the other hand, the reduced χ2 distribution (Eq. (11)) holds
uncertainties = count1/2. for the STV values even for f = 2.3, but fails for the EV results at f = 1.
This behavior by the TV results—S being χ2 distributed for data errors
much larger than needed to give non-Gaussian parameter distribu-
2.8. Program tests tions—is similar to that observed previously for single uncertain
variable NLS [24].
The MC computations were done with programs written in The behavior of sa and sb in Table 1—increasing excess with
Microsoft FORTRAN for PCs, with spot checking of results using the increasing f—is not the apparent behavior in Table II of York et al. [3], in
KaleidaGraph program [39]. Under tight convergence, the TV program which some of the comparisons are positive, some negative. However,
yielded 10-figure agreement with “exact” parameter values for York's these authors took the MC estimate as the reference, changing the signs
model [4,5] and 12-figure agreement with STV [4,12] The standard of the discrepancies from my convention. Thus, for the original York
errors (SEs) agreed with those of Reed (Vprior-based, evaluated at model, the results for data set 3 in Table II of Ref. [3] mean that the MC
calculated points) [32], which are given to 7–8 significant figures in estimates of σa and σb were higher than predicted by 0.251% and
his Ref. [15]. Agreement was obtained also for the examples given by 0.464%, respectively (with uncertainty 0.022% for N = 107). My excesses
Powell and Macdonald [7] and Britt and Luecke [8]. And I have verified for f = 1 in Table 1 (TV analysis) are 0.28% and 0.46% for σa, 0.66% and
that the TV code gives identical results for parameters and SEs in all 0.60% for σb (±0.22% for N = 105)—in adequate agreement. There is not
ways of expressing the fit relation among the variables and enough information to check other results in Table II of Ref. [3].
parameters. Deming's equations for the TV method gave exact However, the largest discrepancies exceed −1%, meaning the MC
agreement with Eq. (19) for straight-line fits, but gave very small results show excesses in these amounts; and these occur for a case
differences for other models (see below). Interestingly, they also gave
slightly smaller STV values in the cases checked so far (though they do
not satisfy F = 0 on the calculated curve).
In the present MC computations and others in the past, I have
observed that convergence on SEV and STV can occur from below,
evidently as a consequence of the iterative changes in the weights (EV
method) and in the residuals (TV). This behavior would seem to rule
out use of the Marquardt procedure. However, in the EV method it is
still possible to use this tool for a given set of estimated weights wi,
using Eq. (15) to reassess these only after Δβ ≈ 0 for the current
weights. A similar procedure should work to achieve Δβ ≈ 0 before
the weights and variables are readjusted in the TV method. I have not
observed convergence from below in the EV2 method.
Codes for carrying out TV computations have been available since
1972; in fact the program of Powell and Macdonald [7] is still
downloadable from Macdonald's web site. Their paper includes a
number of test problems, on which they first used “Deming's method”
(actually the EV method) to obtain starting estimates of parameters,
for subsequent TV computations. Those choosing to check their own
routines on these examples should be aware of a minor flaw: The Ref. Fig. 1. Straight-line model of Pearson and York, as adapted for MC tests. Error bars are
[7] parameter SEs were incorrect but were later fixed [9,10]. 1σ for scaling factor f = 1. Dashed lines show the 1 − σ errors on the fitted function.
Author's personal copy
164 J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169
Table 1
Monte Carlo results (105 10-point data sets) for York model. a
b b
f a σa sa −10 b σb sb χ2v (v)c
b b
(MC) (MC) (MC) (MC)
In an extension of the York model, I have added a quadratic term 3.3. Orear's model
0.03 x2, for which the quadratic coefficient (c = 0.03) has large
relative uncertainty, σc = 0.017035, for f = 1. The TV and EV2 Orear presented results for all three methods under study here for
methods no longer yield identical results, though their solutions are a model having the fit function, y = ax − b / x [16]. His corrected results
very close. Results are given for all three methods in Table 2. for a specific 5-point data set showed the TV and EV2 solutions to be
Surprisingly, for this model the EV method outperforms TV and EV2 very close, with identical S values. However, on pushing the precision
on both bias and precision. However, as before, these results are of the calculations, I find the S values to differ slightly—STV = 2.13377,
more of fundamental than of practical concern. For example, the and SEV2 = 2.13318. The closeness persists when the σxi values are
Author's personal copy
J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169 165
abx
“y” ≡ x0 −x = ð20Þ
1 + bx
166 J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169
Table 2
Monte Carlo results (105 10-point data sets) for quadratic model.a
0.025 5.5000 0.12156 0.1217 4.7999 0.046305 0.04630 2.9998 0.042586 0.04257 1.11 (30) 0
5.5000 0.1216 4.8001 0.04630 3.0000 0.04256 1.14 (30) 0
5.5000 0.1217 4.7999 0.04630 2.9998 0.04257 1.10 (30) 0
0.05 5.4999 0.24313 0.2434 4.7993 0.092610 0.09265 2.9990 0.085172 0.08521 1.60 (30) 0
5.5000 0.2433 4.7999 0.09260 2.9997 0.08514 1.50 (30) 0
5.4999 0.2434 4.7993 0.09265 2.9990 0.08521 1.59 (29) 0
0.10 5.4992 0.48625 0.4878 4.7952 0.18522 0.1865 2.9940 0.17034 0.1720 3.94 (30) 0
5.4996 0.4872 4.7977 0.1860 2.9968 0.1714 3.42 (30) 0
5.4992 0.4877 4.7950 0.1864 2.9938 0.1719 3.82 (30) 0
0.25 5.4958 1.2156 1.2270 4.7737 0.46305 0.4722 2.9663 0.42586 0.4400 18.0 (32) 0
5.4992 1.2181 4.7938 0.4655 2.9888 0.4309 13.8 (32) 0
5.4958 1.2264 4.7737 0.4717 2.9666 0.4393 17.7 (32) 0
0.50 5.4816 2.4313 2.526 4.6844 0.92610 1.0106 2.8517 0.85172 0.9824 280
5.5003 2.463 4.7960 0.9604 2.9816 0.9089 32.4 (33) 2
5.4831 2.547 4.6929 1.0260 2.8617 1.0028 7
1.00 5.4386 4.8625 5.315 4.4047 1.8522 2.288 2.4764 1.7035 2.404 3832
5.5177 5.034 4.9053 2.051 3.0908 2.051 80.1 (42) 614
5.4685 5.498 4.5613 2.419 2.6530 2.580 58
a
True coefficients 5.5, − 0.48, and 0.03, and data error structure of York scaled by factor f. In each trio of results lines, the first is from analysis with the TV method, the second EV,
and the third EV2.
b
Reduced χ2 values for fitting histograms of the estimated c values to 3-parameter Gaussians.
c
Number of divergent data sets.
above predictions, which is well within the 10% guideline, since two of literature. It is explicit in t, so we can treat that as dependent, re-
the parameters have predicted RSEs N 0.17. It is possible that their expressing as
anomalously small MC confidence limits stemmed from the afore-
mentioned problem of convergence from below, for which their t = ½ð2P0 −PÞ
1−n
−P0
1−n
ðn−1Þ
−1 −1
k ; ð22Þ
Marquardt algorithm might have effectively frozen the parameters
near their starting values, preventing true optimization. [In private and computing the effective variance contributions from P to t as
correspondence, Alper has acknowledged this possibility and noted (dt/dP)2 σ2P. Alternatively, we can solve for P, obtaining
that later checks showed general agreement between MC and
predicted SEs.] −1 q
P = 2P0 −exp½q lnðP0 −qktÞ; ð23Þ
f 103 a (MC) 104 σa 104 sa (MC) 10−5 b (MC) 10−5 σb 10−5 sb (MC)
J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169 167
Table 5
Monte Carlo results (105 10-point data sets) for binding model.a
0.1 (1,1) 2.1006 0.049472 0.04952 0.04956 1.2007 0.036955 0.03695 0.03701
2.1002 0.04950 0.04955 1.2009 0.03695 0.03703
2.1006 0.04952 0.04955 1.2007 0.03695 0.03702
0.2 (1,1) 2.1028 0.098945 0.09927 0.09937 1.2024 0.073910 0.07405 0.07421
2.1010 0.09913 0.09930 1.2032 0.07407 0.07431
2.1025 0.09924 0.09935 1.2025 0.07404 0.07421
0.4 (1,1) 2.1115 0.19789 0.2004 0.2009 1.2093 0.14782 0.1494 0.1499
2.1042 0.1994 0.2002 1.2124 0.1496 0.1507
2.1104 0.2002 0.2007 1.2097 0.1493 0.1499
1.0 (1,1) 2.1773 0.49472 0.5408 0.5450 1.2583 0.36955 0.3980 0.4017
2.1292 0.5227 0.5332 1.2783 0.4017 0.4156
2.1691 0.5363 0.5411 1.2604 0.3970 0.4019
0.4 (1,2) 2.1154 0.22051 0.2260 0.2255 1.2271 0.23530 0.2455 0.2466
2.1077 0.2245 0.2244 1.2338 0.2466 0.2489
2.1142 0.2257 0.2253 1.2276 0.2455 0.2468
0.8 (1,2) 2.1687 0.44102 0.4881 0.4979 1.3131 0.47061 0.5637 0.5731
2.1357 0.4737 0.4836 1.3421 0.5747 0.5947
2.1637 0.4853 0.4955 1.3151 0.5638 0.5740
0.4 (2,1) 2.1091 0.18328 0.1855 0.1854 1.2298 0.23145 0.2444 0.2468
2.1281 0.1873 0.1911 1.1856 0.2330 0.2346
2.1136 0.1853 0.1860 1.2190 0.2397 0.2422
0.8 (2,1) 2.1374 0.36656 0.3834 0.3837 1.3332 0.46290 0.5935 0.6233
2.2187 0.4019 0.4388 1.1408 0.4720 0.4875
2.1526 0.3829 0.3876 1.2828 0.5455 0.5600
0.6 (2,2) 2.1044 0.093202 0.09439 0.09435 1.2299 0.24000 0.2522 0.2536
2.1152 0.09525 0.09702 1.1797 0.2375 0.2392
2.1064 0.09431 0.09451 1.2192 0.2471 0.2488
1.2 (2,2) 2.1188 0.18640 0.1974 0.1972 1.3357 0.47999 0.6176 0.6614
2.1646 0.2061 0.2239 1.1191 0.4624 0.4715
2.1248 0.1974 0.1977 1.2886 0.5576 0.5579
a
In each trio of results lines, the values come respectively from analysis with the TV, EV, and EV2 methods.
b
Data error structures of Table 4 (σxi, σyj), scaled by factor f.
c
RMS Vprior estimates of the parameter SEs.
Table 6
Analysis of Wentworth kinetics data.a
168 J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169
methods perform comparably in all cases, with differences that would for his conscientious, constructive criticism of an earlier version of this
seldom be important in practical work. Only in one statistical property work, which led to important realizations about the interrelations
does the TV method always excel: The STV values closely follow the χ2 among the various methods that have been used to treat uncertainty
distribution for all data error scaling, while the corresponding EV sums in multiple variables.
do not. The Deming equations for the TV method can be considered a
near statistical equivalent of the Powell–Macdonald–Britt–Luecke–
Jefferys equations, though Deming's yield neither F = 0 on the Appendix A. Supplementary data
calculated curve nor invariance with respect to different ways of
expressing the fit relation. Supplementary data associated with this article can be found, in
While this study had to employ specific models for the MC the online version, at doi:10.1016/j.chemolab.2010.07.003.
comparisons, the models are representative ones. Thus I anticipate
that the behavior observed for the three methods will be found to
hold approximately in other analysis problems. In particular, it is References
likely that the close similarity in the TV and EV2 results (rigorously [1] J.R. Macdonald, W.J. Thompson, Least-squares fitting when both variables contain
identical for the straight line) is general. It also appears that the EV2 errors: pitfalls and possibilities, Am. J. Phys. 60 (1992) 66–73.
iterations always converge from above. Together these properties [2] D. York, Least-squares fitting of a straight line, Can. J. Phys. 44 (1966) 1079–1086.
[3] D. York, N.M. Evenson, M.L. Martinez, J. DeB Delgado, Unified equations for the
mean that the EV2 computations can be carried out with some slope, intercept, and standard errors of the best straight line, Am. J. Phys. 72
standard nonlinear programs that utilize the Marquardt algorithm (2004) 367–375.
(like KaleidaGraph [39]). [4] J. Riu, F.X. Rius, Univariate regression models with errors in both axes, J.
Chemometr. 9 (1995) 343–362.
A requirement for successful use of any of these methods is good [5] C.A. Cantrell, Technical note: review of methods for linear least-squares fitting of
information about the data error in both variables. In most applications, data and application to atmospheric chemistry problems, Atmos. Chem. Phys.
this need will likely limit the reliability of the analysis more than the 8 (2008) 5477–5487.
[6] W. E. Deming, Statistical Adjustment of Data, (Wiley, New York, 1938, 1943;
choice of method. The usual approaches for getting this information Dover, New York, 1964).
entail analysis of the statistics of replicates or residuals [43–45]. [7] D.R. Powell, J.R. Macdonald, A rapidly convergent iterative method for the solution
Reassuringly, the data variance function does not need to be precisely of the generalized nonlinear least squares problem, Comput. J. 15 (1972)
148–155.
known to yield near-optimal results in the LS fit of the data to the model
[8] H.I. Britt, R.H. Luecke, The estimation of parameters in nonlinear, implicit models,
[46]. Technometrics 15 (1973) 233–247.
The need for information about the data error in multiple variables [9] W.H. Southwell, Fitting data to nonlinear functions with uncertainties in all
measurement variables, Comput. J. 19 (1976) 69–73.
also means that reported SEs should normally be based on Vprior
[10] D.R. Powell, J.R. Macdonald, Comments on ‘Fitting data to nonlinear functions
rather than on Vpost. The latter (in which the prefactor S/ν is included) with uncertainties in all measurement variables’, Comput. J. 20 (1977)
is the default choice in much physical science work. It is appropriate in 285–286.
situations where a single uncertain variable has constant unknown [11] W.H. Jefferys, On the method of least squares, Astron. J. 85 (1980) 177–181.
[12] M. Lybanon, A better least-squares method when both variables have uncertain-
uncertainty or some simple form of heteroscedasticity (e.g., error ties, Am. J. Phys. 52 (1984) 22–26.
proportional to signal or to signal1/2) of unknown scale, and this [13] M. Clutton-Brock, Likelihood distributions for estimating functions when both
extends to cases where weight changes are needed when such data variables are subject to error, Technometrics 9 (1967) 261–269.
[14] D.R. Barker, L.M. Diana, Simple method for fitting data when both variables have
are transformed. However, with multiple uncertain variables, all need uncertainties, Am. J. Phys. 42 (1974) 224–227.
to be known within a single scale factor (C after Eq. (1)) to obtain [15] D.S. Riggs, J.A. Guarnieri, S. Addelman, Fitting straight lines when both variables
minimum-variance results. This requirement will normally require are subject to error, Life Sci. 22 (1978) 1305–1360.
[16] J. Orear, Least squares when both variables have uncertainties, Am. J. Phys. 50
absolute information about the data error in all variables, making (1982) 912–916 erratum 52 (1984) 278.
Vprior the appropriate choice. A benefit of this choice is the availability [17] J.P. Chandler, On an iterative procedure for estimating functions when both
of the χ2 test as a simple indicator of statistical inadequacy. variables are subject to error, Technometrics 14 (1972) 71–76.
[18] M. Lybanon, Comment on ‘Least-squares when both variables have uncertainties’,
About terminology: Physical scientists have often referred to
Am. J. Phys. 52 (1984) 276–278.
treatments of problems with uncertainty in x and y as “generalized [19] J. Tellinghuisen, C.H. Bolster, Weighting formulas for the least-squares analysis of
least-squares.” That label unfortunately means something different binding phenomena data, J. Phys. Chem. B 113 (2009) 6151–6157.
[20] J.H. Williamson, Least-squares fitting of a straight line, Can. J. Phys. 46 (1968)
among statisticians, namely the estimation of both the variance
1845–1847. This treatment is essentially Exercise 3 on page 145 in Deming's book.
function and the response function for a data set, with the former [21] S.D. Christian, E.E. Tucker, LINGEN—a general linear least squares program, J.
usually obtained by fitting squared residuals or their logarithms [43– Chem. Educ. 61 (1984) 788.
45]. Statisticians have used “errors in variables (EIV) models” and [22] M.L. Johnson, L.M. Faunt, Parameter estimation by least-squares methods, Meth.
Enzymol. 210 (1992) 1–37.
“Model II regression,” but both of these also seem inadequately [23] E. Di Cera, Use of weighting functions in data fitting, Meth. Enzymol. 210 (1992)
descriptive. “Total least-squares” (TLS) is also used, but seems 68–87.
presumptuous in that in many applications of TLS (as well as EIV) [24] J. Tellinghuisen, A Monte Carlo study of precision, bias, inconsistency, and non-
Gaussian distributions in nonlinear least squares, J. Phys. Chem. A 104 (2000)
there is no attention given to the relative magnitudes of the 2834–2844.
adjustments of x and y needed to minimize STV [47]. Sometimes the [25] A.H. Kalantar, R.I. Gelb, J.S. Alper, Biases in summary statistics of slopes and
approach has been to minimize “perpendicular” residuals (also called intercepts in linear regression with errors in both variables, Talanta 42 (1995)
597–603.
“orthogonal regression”), whereupon results can vary with arbitrary [26] R.G. Krutchkoff, Classical and inverse regression methods of calibration, Techno-
scale changes in the variables. In the TLS community there also appears metrics 9 (1967) 425–439.
to be little recognition of Deming's early contributions on this problem. [27] R.G. Krutchkoff, Classical and inverse regression methods of calibration in
extrapolation, Technometrics 11 (1969) 605–608.
At the same time “Deming regression” has incorrectly come to refer to
[28] J. Berkson, Estimation of a linear function for a calibration line; consideration of a
just the straight-line model with two uncertain variables, when in fact recent proposal, Technometrics 11 (1969) 649–660.
Deming's treatment was not so limited. Suitable terminology can [29] G.K. Shukla, On the problem of calibration, Technometrics 14 (1972) 547–553.
[30] J. Tellinghuisen, Inverse vs. classical calibration for small data sets, Fresenius J.
remain a subject for further discussion.
Anal. Chem. 368 (2000) 585–588.
[31] J. Tellinghuisen, Least squares in calibration: dealing with uncertainty in x, Analyst
Acknowledgments 135 (2010) 1961–1969.
[32] B.C. Reed, Least-squares fits with errors in both coordinates. II: comments on
parameter variances, Am. J. Phys. 60 (1992) 59–62.
I thank Carl Bolster and David Parrish for rekindling my interest in [33] W.C. Hamilton, Statistics in Physical Science: Estimation, Hypothesis Testing, and
this topic, Joe Alper for helpful correspondence, and Ross Macdonald Least Squares, The Ronald Press Co., New York, 1964.
Author's personal copy
J. Tellinghuisen / Chemometrics and Intelligent Laboratory Systems 103 (2010) 160–169 169
[34] P.R. Bevington, Data Reduction and Error Analysis for the Physical Sciences, [42] W.E. Wentworth, Rigorous least squares adjustment: application to some non-
McGraw-Hill, New York, 1969. linear equations, J. Chem. Educ. 42 (1965) 96–103.
[35] W.H. Press, B.P. Flannery, S.A. Teukolsky, W.T. Vetterling, Numerical Recipes, [43] M. Davidian, R.J. Carroll, Variance function estimation, J. Am. Stat. Assoc. 82 (1987)
Cambridge Univ. Press, Cambridge, U. K., 1986. 1079–1091.
[36] J. Tellinghuisen, Weighted least-squares in calibration: what difference does it [44] R.J. Carroll, D. Ruppert, Transformation and Weighting in Regression, Chapman
make? Analyst 132 (2007) 536–543. and Hall, New York, 1988.
[37] J. Tellinghuisen, Least squares with non-normal data: estimating experimental [45] M. Davidian, P.D. Haaland, Regression and calibration with nonconstant error
variance functions, Analyst 155 (2008) 161–166. variance, Chemom. Intell. Lab. Syst. 9 (1990) 231–248.
[38] D.W. Marquardt, An algorithm for least-squares estimation of nonlinear [46] J. Tellinghuisen, Variance function estimation by replicate analysis and general-
parameters, J. Soc. Ind. Appl. Math. 11 (1963) 431–441. ized least squares: a Monte Carlo comparison, Chemom. Intell. Lab. Syst. 99 (2009)
[39] J. Tellinghuisen, Nonlinear least squares using microcomputer data analysis 138–149.
programs: KaleidaGraph in the physical chemistry teaching laboratory, J. Chem. [47] I. Markovsky, S. Van Huffel, Overview of total least squares methods, Signal
Educ. 77 (2000) 1233–1239. Process. 87 (2007) 2283–2302.
[40] K. Pearson, On lines and planes of closest fit to systems of points in space, Philos.
Mag. 2 (1901) 559–572.
[41] J.S. Alper, R.I. Gelb, Standard errors and confidence intervals in nonlinear
regression: comparison of Monte Carlo and parametric statistics, J. Phys. Chem.
94 (1990) 4747–4751.