Professional Documents
Culture Documents
Automatica
journal homepage: www.elsevier.com/locate/automatica
under binary-valued sensors (Wang, Li, Yin, Guo, & Xu, 2011), or study identifiability, identification algorithms, and input
PWM (Pulse Width Modulation)-based sampling (Wang, Feng, & design directly on the original parameters. The main contributions
Yin, 2013). Under irregular or random sampling, the sampled sys- of this paper are in the following aspects. (1) We show that under
tem of a linear time-invariant system becomes time-varying, for any inputs of exponential polynomial types, the continuous-time
which system conversion is complicated and computational com- system is identifiable if the sampling points are sufficiently dense
plexity is much higher. When sampling is slower, irregular, or ran- in a given time interval. The bound on the density of the sampling
dom, system identification formulation, identifiability, algorithms, points is tight, revealing an interesting connection, in terms of
accuracy, and convergence will be fundamentally impacted. This identification information complexity, to Shannon’s sampling the-
paper will explore related issues in this paradigm. orem for signal reconstruction (Proakis & Manolakis, 2007) and our
In Johansson (2010), the original differential equation is first recent results on state estimation (Wang et al., 2011). Note that the
converted to an algebra equation with respect to time by using input used in this paper is continuous, while the existing literature
filtered input and output signals. Then the parameters of the (Ding et al., 2009; Gillberg & Ljung, 2010) commonly uses piece-
algebra equation are estimated at the irregular sampling points, wise constant inputs by zero-order hold. (2) Robustness of sys-
and the original system parameters are recovered by a one- tem identifiability under a given sampling density is established.
to-one mapping from the estimated parameters. One possible (3) Under noise-free observations, a convergent iterative algorithm
methodology is to identify the original system parameters without is introduced, which is valid for any input signals satisfying certain
conversion to its sampled system (Gillberg & Ljung, 2010; gradient conditions. (4) Under noisy observations, suitable iden-
Larsson & Söderström, 2002; Marelli & Fu, 2010; Vajda, Valko, & tification algorithms are proposed, which are shown to converge
Godfrey, 1987). The parameters are directly identified by using strongly and carry properties of the CLT (central limit theorem)
a continuous-time frequency domain identification method in types if certain ergodicity conditions are satisfied. (5) Persistent
Gillberg and Ljung (2010). Similar to its discrete-time counterpart, excitation (PE) conditions are derived that ensure convergence of
the continuous-time system can also be expressed by a linear the developed algorithms. Departing from the traditional PE condi-
regression equation in a differential operator form (Larsson & tions that rely only on input values, it is shown that under irregular
Söderström, 2002), in which the regressor involves input and or random sampling, both sampling time sequences and the input
output signals and their derivatives. Since the derivatives are values impact on convergence. These results provide guidance for
unavailable under sampled data, higher-order derivatives of the input design in identification experiments. (6) Consistency of the
input and output signals are approximated by their related algorithms is proved without requiring fast sampling.
differences (Larsson & Söderström, 2002), introducing errors The rest of the paper is arranged as follows. The system setting
as a consequence. The resulting discrete-time system is then and several key properties are presented in Section 2. System iden-
identified by batch or recursive algorithms (Ljung, 1999; Ljung & tifiability is investigated in Section 3. The parameter estimation
Vicino, 2005). To reduce approximation errors, fast sampling is algorithms and their convergence properties under noise-free
required. An instrumental variable approach is used to enhance observation are discussed in Section 4. When observations are
estimation accuracy for continuous-time autoregressive processes noise corrupted, identification algorithms are significantly differ-
in Mossberg (2008), which demonstrates improved computational ent from noise-free cases. Two kinds of estimation algorithms (it-
efficiency in comparison to the least squares approach (Larsson, erative algorithms and recursive algorithms) are introduced and
Mossberg, & Soderstrom, 2007). their convergence conditions are established in Section 5. Section 6
Synchronization between the input sampling and output sam- is focused on input design problems. The related persistent excita-
pling is also a significant factor. Typical schemes for the indirect tion conditions are obtained. In Section 7 some numerical exam-
method assume that the input and output are sampled at the same ples are given to verify the effectiveness of the proposed algorithms
sampling points (Larsson & Söderström, 2002; Yuz, Alfaro, Agüero, of this paper. Section 8 concludes the paper with some further re-
& Goodwin, 2011). The estimation method in Yuz et al. (2011) rep- marks. The main proofs of the assertions in the paper are placed in
resents the original continuous-time state space model by an in- the Appendix.
cremental approximation under nonuniform but fast sampling and
employs the maximum likelihood approach. Zhu, Telkamp, Wang, 2. Preliminaries
and Fu (2009) propose a two-time scale sampling scheme: fast
uniform input sampling, but slow and irregular output sampling, This section describes the system setting and establishes several
with assumption that the output sampling time is a multiple of important properties to be used in the subsequent sections.
the input sampling time. Under an output error structure, the sys-
tem parameters are then estimated by minimizing a suitable loss
function. In contrast, in Gillberg and Ljung (2010) the input is a 2.1. Systems
uniformly spaced piece-wise constant function (zero-order hold),
while the output is sampled irregularly. The main technique is to We are concerned with identification of a single-input–single-
use B-spline approximation to achieve uniformly distributed knots output, linear, time-invariant, stable, finite dimensional system
from the non-uniformly sampled output. The method in Gillberg in the continuous-time domain, represented by a strictly proper
and Ljung (2010) is restricted to the noise-free sampled output transfer function
and its estimation accuracy enhancement requires fast sampling. b1 sn−1 + · · · + bn−1 s + bn b(s)
Despite extensive research effort in this area, some fundamental G(s) = , , (1)
sn + a1 s n −1 + · · · + an − 1 s + an a(s)
questions remain un-answered: (1) How fast and under what types
of sampling schemes, is the continuous-time system identifiable? where a(s) is stable, i.e., all the roots of a(s) lie on the open left-
(2) What types of inputs will imply system identifiability? (3) What half complex plane; a(s) and b(s) are coprime, i.e., they do not have
modifications must be made to identification algorithms? (4) To common roots. The impulse response of G(s) is denoted by g (t ) =
achieve convergence, how should the input be designed? L−1 (G(s)), where L−1 is the inverse Laplace transform. Let the
This paper investigates these questions from a new angle. In- system parameters be expressed as θ = [a1 , . . . , an , b1 , . . . , bn ]′ .
stead of focusing on parameter mappings between the continuous- We use G(s, θ ) and g (t , θ ) to indicate their dependence on the
time system and its sampled system, we view irregularly or parameters. R and C are the fields of real and complex numbers,
randomly sampled values as the available information set and respectively.
102 B. Mu et al. / Automatica 60 (2015) 100–114
Under the zero initial condition, the input–output relationship 3.1. Identifiability
of the system (1) is
t The simplest input is perhaps the step input U (s) = u0 /s.
y(t ) = g (t , θ ) ⋆ u(t ) = g (t − τ , θ )u(τ )dτ , f (t , θ ), (2) The following results allow more general inputs of rational types.
0
The resulting identifiability condition implies that the system is
where the symbol ⋆ denotes the continuous-time convolution. In identifiable as long as the number of the sampling points is greater
this paper, u(t ) is uniformly bounded by ∥u∥∞ ≤ umax < ∞, than some positive integer in a given interval, regardless how
where ∥ · ∥∞ is the standard L∞ norm. In the frequency domain, the sampling times are generated. Consequently, the results are
the relationship (2) has the expression applicable to different sampling schemes, such as the traditional
Y (s) = G(s, θ )U (s), (3) periodic sampling (Åström & Wittenmark, 1997), nonuniform
sampling (Ding et al., 2009), event-triggered sampling (Åström &
where U (s) and Y (s) are the Laplace transforms of the input and Bernhardsson, 1999; Persson & Gustafsson, 2001), or the PWM-
output, respectively. In this paper, u(t ) is selected in identification based sampling introduced in Wang et al. (2013).
experimental design and always assumed to be known. The output
y(t ) is sampled at the nonuniform sampling times t1 , t2 , . . . .
Assumption 1. U (s) is a nonzero proper rational coprime function,
System identification aims to identify θ from the data set
(t1 , y(t1 )), (t2 , y(t2 )), . . . . i.e., U (s) = d(s)/c (s), where c (s) and d(s) are coprime polynomials,
For convenience of statement, we define the set A on θ : (1) deg d(s) ≤ deg c (s) = m. This implies that u(t ) is an exponential
The elements in A consist of column vectors θ ∈ R2n ; (2) For polynomial and hence the output f (t , θ ) is also an exponential
any θ = [a1 , . . . , an , b1 , . . . , bn ]′ ∈ A, the polynomials a(s) = polynomial.
sn + a1 sn−1 + · · · + an−1 s + an and b(s) = b1 sn−1 + · · · + bn−1 s + bn
Denote µ∗T = (2n + m − 1) + δ2πT , δ ∗ , max1≤i,j≤2n+m {|ℑ(λi −
∗
generated by θ are coprime. Since the roots of a polynomial depend
continuously on its coefficients, A is an open set. λj )|} where {λi , i = 1, . . . , 2n + m} are the roots of a2 (s)c (s).
Since u(t ) is uniformly bounded, the order of partial derivatives
(with respect to θ ) and the integral in the following equation is
Lemma 3. Under Assumption 1 and for θ ∈ A, the N × 2n Jacobian
exchangeable if u(t ) is continuous almost everywhere, namely,
matrix of f (t , θ ) at the sampling points 0 < t1 < t2 < · · · < tN ≤ T
t
∂ f (t , θ ) ∂ 0 g (t − τ , θ )u(τ )dτ
∂ f (t1 , θ ) ∂ f (t2 , θ ) ∂ f (tN , θ )
′
h(t , θ) ,
=
∂θ ∂θ JN (θ ) = , ,...,
∂θ ∂θ ∂θ
∂ g (t − τ , θ ) ∂ g (t , θ )
t
= u(τ )dτ = ⋆ u(t ). (4) has full column rank if N > µ∗T .
0 ∂θ ∂θ
If a ̸= 2, then h(t , θ ) in (9) is linearly independent. If a = 2, then Lemma 4. Suppose that A, B ∈ Rl×l , A is nonsingular, and ∥A−1 ∥ ≤
∂c ∂c
c1 = 0, but ∂ a1 ̸= 0 and ∂ a2 ̸= 0. This indicates that h(t , θ ) in (9) ς . If ∥A − B∥ ≤ ν̄ and ς ν̄ < 1, then B is nonsingular and ∥B−1 ∥ ≤
ς
is still linearly independent even if a = 2. In other words, if we use .
1−ς ν̄
the same Jacobian matrix in our algorithms, it will still give local
convergence for the parameter estimation of the plant.
Theorem 3. Under the conditions of Theorem 1, there exists ϵ > 0
such that for any initial value θ0 in the neighborhood O(θ ∗ , ϵ) of θ ∗
4. Identification algorithms and convergence properties under with radius ϵ , the iterative sequence {θk } defined in (11) converges to
noise-free observations the true parameter θ ∗ and has the following convergence rate
θk+1 = θk + (JN′ (θk )JN (θk ))−1 JN′ (θk )r (θk ), (11) y(tk ) = f (tk , θ ∗ ) + ek ,
where where θ ∗ is the true parameter vector and ek is the noise. In the case
∂ f (t , θ ) ∂ f (t , θ ) ∂ f (tN , θk ) ′ of noisy observations, the plant is naturally required to be stable
1 k 2 k
JN (θk ) = , ,..., , and it is reasonable to consider asymptotic convergence properties
∂θ ∂θ ∂θ of parameter estimation algorithms as the number of the sampling
r (θk ) = [y(t1 ) − f (t1 , θk ), . . . , y(tN ) − f (tN , θk )] .
′
points grows. The following condition on ek is assumed throughout
the rest of the paper.
Remark 2. Under Assumption 1, the Laplace transform Y (s) of the
output y(t ) is a rational polynomial: Y (s) = b(s)d(s)/(a(s)c (s)), Assumption 2. The noise {ek } is a sequence of zero mean i.i.d.
and hence the Laplace transform L(h(t , θ )) of the gradient vector random variables with finite variance σ 2 .
of y(t ) is still a rational polynomial vector
5.1. Nonlinear least squares estimators and convergence properties
∂ Y (s, θ) b(s)d(s)[sn−1 , . . . , s, 1] d(s)[sn−1 , . . . , s, 1] ′
= − , .
∂θ a2 (s)c (s) a(s)c (s) Define the sample objective function
As a result, taking inverse Laplace transform on both sides gives
N
b(s)d(s)[sn−1 , . . . , s, 1] 1
QN (θ ) = (y(tk ) − f (tk , θ ))2 . (14)
h(t , θ) = L −1
− , N k=1
a2 (s)c (s)
d(s)[sn−1 , . . . , s, 1] ′ The vector θN which minimizes (14) on a compact subset Θ of R2n
.
a(s)c (s) containing θ ∗ is called the nonlinear least-squares (NLS) estimate
for θ ∗ based on the sampled outputs {y(t1 ), . . . , y(tN )}.
Since the inverse Laplace transform of a rational polynomial is an
We impose the following conditions as assumptions for now.
exponential polynomial, each component of h(t , θ ) is still an ex-
They will be verified by input design in Section 6.
ponential polynomial, which is uniquely determined by a finite
number of coefficients of the polynomials a(s), b(s), c (s), d(s). For
f (tk , θ ) exists for
N 2
1
Assumption 3. (i) The limit limN →∞
example, let p(s) be a proper rational polynomial and its factoriza- N k=1
N
2
vli any θ ∈ Θ and Q (θ ) = limN →∞ N1 (tk , θ ) − f (tk , θ ∗ )
p̄ nl
tion be p(s) = l=1 i=1 (s−λ )i . Thus the inverse Laplace trans- k =1 f
l
p̄ nl i−1 has a unique minimum at θ = θ ∗ .
form of p(s) is L−1 (p(s)) = i=1 vli (i−1)! exp(λl t ). This is
t
l=1 (ii) The gradient vector h(tk , θ ) and Hessian matrix χ (tk , θ )
an accurate expression over the variable t. As a result, the Jacobian of f (tk , θ ) exist and are continuous on Θ , and the limits
matrix given in the algorithm (11) can be accurately calculated for
k=1 ρk (θ )ϖk (θ ), with ρk (θ ), ϖk (θ ) replaced by
N
limN →∞ N1
any θ ∈ A if the input is available and the computation process f (tk , θ ), h(tk , θ ), χ (tk , θ ) respectively, exist. This implies that
will not introduce systematic errors.
the limit M (θ ) = limN →∞ N1 i=1 h(tk , θ )h (tk , θ ) exists for
N ′
The assertions below follow closely from the conditions Now, consider the limit ODE
imposed in Jennrich (1969), that render the NLS estimator θN a
consistent estimator of θ ∗ and asymptotical normality via applying θ̇ = W (θ ). (20)
the corresponding convergent results of the NLS estimator in We note that W (θ ) = 0, implying
∗
that θ is an equilibrium∗
Jennrich (1969). The details are omitted here. ∂ W (θ )
−M (θ ∗ ). By the ODE
point of (20). In addition, ∂θ
=
θ=θ ∗
Theorem 4. Let θN be the NLS estimates of (14). Under Assump- method (Kushner & Yin, 2003), we have the following convergence
tions 2 and 3(i), we have θN −
→ θ ∗ with probability one as N tends to conclusion.
infinity. Further, if Assumptions 3(ii) and (iii) also hold, then
√ Theorem 6. If M (θ ∗ ) is full rank, then
N (
θN − θ ∗ ) −
→ N (0, σ 2 M −1 (θ ∗ )) as N → ∞. (15)
θk → θ ∗ locally w.p.1 as k → ∞.
with j being the imaginary unit (j2 = −1) and O(µt ) with 0 < µ <
Let θk be the estimate of θ ∗ at the time instance tk . The
1 are the steady-state and transient components of the system (1),
parameter algorithm is updated recursively
respectively. An experiment is said to be repeated of period T0 if in
θk+1 = θk + εk h(tk , θk )(y(tk ) − f (tk , θk )) the subsequent disjoint intervals (iT0 , (i + 1)T0 ], i = 1, 2, . . . , the
= θk + εk h(tk , θk )(f (tk , θ ∗ ) − f (tk , θk ) + ek ), (17) sampling points are repeated as that in (0, T0 ], i.e., tiN0 +l = iT0 + tl
with t0 = 0 for i ≥ 1, 1 ≤ l ≤ N0 , where N0 is the number
where εk is the scalar step
∞size satisfying the typical conditions of the sampling points in (0, T0 ]. This is a typical scenario in the
(εk > 0, εk → 0, and k=1 εk = ∞). This is a stochastic ap- TDMA (time division multiple access) protocol of communication
proximation algorithm. For convergence analysis, we will employ systems, in which each frame is of duration T0 and frames are
the ODE (ordinary differential equation) approach (Kushner & Yin, repeated. Define the 2n × 2L matrix
2003). The main conditions on the sampling times {t1 , t2 , . . .} and ∂|G(jω, θ )| ∂ ̸ G(jω, θ )
the input signals will be stated here first as assumptions and then G(θ ) , A1 , −A1 |G(jω, θ )| ,...,
elaborated later. ∂θ ∂θ
∂|G(jωL, θ )| ∂ ̸ G(jωL, θ )
Ergodicity conditions: Suppose that the following limits exist AL , −AL |G(jωL, θ )| .
locally. That is, there exists m0 such that for all m > m0 , and each ∂θ ∂θ
θ , there exist continuous functions M (θ ) and W (θ ) such that
Theorem 7. Suppose that the input is harmonic of L ≥ n, the
m+N −1
1 sampling scheme is repeated with period T0 , G(θ ∗ ) is full rank, and
lim h(tk , θ )h (tk , θ ) , M (θ ),
′
(18) δ ∗ T0
N →∞ N N0 > µ∗T = (2n + 2L − 1) + 2π
, where δ ∗ = max{δ̄, 2Lω} and
k=m
m+N −1
δ̄ = max1≤i,j≤n {|ℑ(λi − λj )|} and {λi , i = 1, . . . , n} are the roots
1 of a(s). Then the limits (18) and (19) are valid, M (θ ∗ ) is full rank,
lim h(tk , θ )(f (tk , θ ) − f (tk , θ )) , W (θ ).
∗
(19)
N →∞ N and Assumption 3 is valid.
k=m
106 B. Mu et al. / Automatica 60 (2015) 100–114
Table 1
Parameter estimation of a stable system under different input signals.
True values k=1 k=2 k=3 k=4 k=5
U (s) = 1/s
a1 = 1.5 0.0000 0.8747 1.5032 1.5000 1.5000
a2 = 1 0.0000 −0.7109 0.9931 1.0000 1.0000
b1 = 2.5 2.3563 2.5000 2.5003 2.5000 2.5000
b2 = −1 −3.3994 −2.6904 −1.0010 −1.0000 −1.0000
U (s) = 1/(s + 1)
a1 = 1.5 0.0000 0.9891 1.5331 1.5001 1.5000
a2 = 1 0.0000 −0.8529 1.0788 0.9997 1.0000
b1 = 2.5 2.1758 2.5000 2.5008 2.5000 2.5000
b2 = −1 −2.8185 −2.5986 −0.9323 −0.9998 −1.0000
U (s) = 1/(s − 0.4)
a1 = 1.5 0.0000 0.8739 1.5032 1.5000 1.5000
a2 = 1 0.0000 −0.7101 0.9928 1.0000 1.0000
b1 = 2.5 2.3582 2.5000 2.5003 2.5000 2.5000
b2 = −1 −3.4032 −2.6907 −1.0009 −1.0000 −1.0000
U (s) = s/(s2 + 12 ) Fig. 1. Recursive estimation under repeatedly periodic sampling for different SNRs.
a1 = 1.5 0.0000 0.8746 1.5025 1.5000 1.5000
a2 = 1 0.0000 −0.7105 0.9919 1.0000 1.0000 to demonstrate the impact of the different noise levels to the
b1 = 2.5 2.3556 2.5000 2.5003 2.5000 2.5000
parameter estimation.
b2 = −1 −3.4014 −2.6912 −1.0027 −1.0000 −1.0000
U (s) = (s2 + 1.5s + 1)/((s2 + 12 )(s − 2))
Example 5. In the noisy observation scenario, the effectiveness
a1 = 1.5 0.0000 0.8669 1.5020 1.5000 1.5000
a2 = 1 0.0000 −0.7024 0.9877 1.0000 1.0000 of the iterative algorithm (16) and the recursive algorithm (17)
b1 = 2.5 2.3730 2.5000 2.5003 2.5000 2.5000 is verified by the second-order system given in Example 3. The
b2 = −1 −3.4393 −2.6943 −1.0035 −1.0000 −1.0000 input signal is u(t ) = cos(t ) + 2 cos(2t ) (its Laplace trans-
form is U (s) = s/(s2 + 12 ) + 2s/(s2 + 22 )) with the period
Table 2 T0 = π /2, and the sampling points in the interval [0, π /2] are
Parameter estimation of an unstable system under different input signals. 0.2013, 0.2380, 0.3114, 0.4642, 0.4669, 0.5641, 0.6657, 0.7422,
True values k=1 k=2 k=3 k=4 k=5 1.0959, 1.1651, which are uniformly produced from the interval
[0, π /2]. The successive sampling points are periodically repeated
U (s) = 1/s
a1 = −1.5 0.0000 0.3996 −1.9253 −1.4878 −1.5000 with the period π /2. In order to show the effectiveness of the al-
a2 = 0 0.0000 −2.3030 0.5364 −0.0137 0.0000 gorithms under different noise levels, the observation noise {ek } is
b1 = 2.5 2.2762 2.4978 2.4993 2.5000 2.5000 chosen to be a sequence of Gaussian random variables with zero
b2 = −1 4.5792 3.7181 −2.0400 −0.9690 −1.0000 mean for different variances: 0.12 , 0.42 , 0.82 , and the resulting
U (s) = 1/(s + 1) SNRs are 529.09, 33.07, 8.26 since the variance of signal is 5.2909.
a1 = −1.5 0.0000 0.3943 −1.9157 −1.4869 −1.5000 In the following, all the estimates are based on the average of the
a2 = 0 0.0000 −2.2996 0.5255 −0.0147 0.0000
estimates for ten runs. Table 3 summarizes the estimates and the
b1 = 2.5 2.2687 2.4977 2.4993 2.5000 2.5000
b2 = −1 4.5953 3.7040 −2.0155 −0.9668 −1.0000 corresponding squared sums of the estimation errors (SSEE) at the
sample size N = 600, 1200, 1800, 2400, 3000 based on the it-
U (s) = 1/(s − 0.4)
a1 = −1.5 0.0000 0.4017 −1.9303 −1.4882 −1.5000 erative algorithm (16) with the initial value [0, 0, 0, 0]′ under dif-
a2 = 0 0.0000 −2.3042 0.5420 −0.0132 0.0000 ferent SNRs, where the values in the parentheses are the resulting
b1 = 2.5 2.2795 2.4978 2.4993 2.5000 2.5000 standard deviations based on ten runs. While the recursive esti-
b2 = −1 4.5732 3.7238 −2.0525 −0.9700 −1.0000 mate is displayed in Fig. 1 in terms of the recursive algorithm (17)
U (s) = s/(s2 + 12 ) with the initial value [0, 0, 0, 0]′ .
a1 = −1.5 0.0000 0.4006 −1.9207 −1.4876 −1.5000
a2 = 0 0.0000 −2.3051 0.5309 −0.0139 0.0000
b1 = 2.5 2.2752 2.4978 2.4993 2.5000 2.5000 Example 6. The estimation effectiveness of the iterative algorithm
b2 = −1 4.5751 3.7199 −2.0283 −0.9684 −1.0000
(16) is illustrated in this example in the case of i.i.d. random
U (s) = (s2 − 1.5s)/((s2 + 12 )(s − 2)) sampling scheme. Consider the second-order transfer function
a1 = −1.5 0.0000 0.4012 −1.9321 −1.4883 −1.5000 given in Example 3. The input signal is u(t ) = 2, t ≥ 0 (its Laplace
a2 = 0 0.0000 −2.3031 0.5440 −0.0130 0.0000
b1 = 2.5 2.2801 2.4993 2.4993 2.5000 2.5000 transform is U (s) = 2/s.) The sampling sequence {tk } comes from
b2 = −1 4.5745 3.7227 −2.0570 −0.9703 −1.0000 the uniform distribution over the interval [0, 4]. The observation
noise {ek } is set to be a sequence of the Gaussian random variables
with zero mean for different variances: 0.12 , 0.42 , 0.82 , and the
The sampling points and the input signals are the same as those in
resulting SNRs are 145.51, 9.09, 2.27 since the variance of signal is
Example 3 except for the last input signal U (s) = (s2 − 1.5s)/((s2 +
1.4551. The initial value of the algorithm is set to be [0, 0, 0, 0]′ . All
12 )(s − 2)) canceling the poles of plant. Table 2 gives the iterative
the estimates are based on the average of ten runs. The estimates,
results of the algorithm (11) with the initial value [0, 0, 0, 0]′ .
the corresponding standard deviations in the parentheses, and the
The following examples involve noisy observations. We intro- SSEE at the sampling sizes N = 600, 1200, 1800, 2400, 3000 are
duce the signal-to-noise ratio (SNR) shown in Table 4 under different SNRs.
variance of signal
SNR = Example 7. The estimation effectiveness of the iterative algorithm
variance of noise (16) and the recursive algorithm (17) will be examined under
N N 2
the independent increment random sampling scheme. Consider
1
f (ti , θ ∗ ) − 1
f (ti , θ ∗ )
N −1
i =1
N
i=1
the second-order transfer function given in Example 3. The input
= signal is u(t ) = cos(t ) + 2 cos(1.5t ) (its Laplace transform is
σ2
108 B. Mu et al. / Automatica 60 (2015) 100–114
Table 3
Iterative estimation under repeatedly periodic sampling for different SNRs.
N 600 1200 1800 2400 3000
SNR = 529.09
a1 = 1.5 1.5021 1.5017 1.5011 1.5019 1.5013
(0.0078) (0.0071) (0.0060) (0.0045) (0.0036)
a2 = 1 0.9999 1.0009 0.9998 0.9993 0.9992
(0.0088) (0.0052) (0.0055) (0.0048) (0.0049)
b1 = 2.5 2.5052 2.5036 2.5020 2.5028 2.5019
(0.0110) (0.0083) (0.0066) (0.0055) (0.0048)
b2 = −1 −1.0016 −0.9994 −1.0010 −1.0005 −1.0007
(0.0179) (0.0097) (0.0088) (0.0077) (0.0073)
SSEE 3.36E−05 1.69E−05 6.20E−06 1.25E−05 6.24E−06
SNR = 33.07
a1 = 1.5 1.4773 1.4910 1.4915 1.4946 1.4973
(0.0217) (0.0159) (0.0185) (0.0220) (0.0154)
a2 = 1 1.0040 0.9995 0.9983 1.0002 1.0002
(0.0241) (0.0258) (0.0267) (0.0203) (0.0191)
b1 = 2.5 2.4737 2.4929 2.4913 2.4921 2.4969
(0.0310) (0.0240) (0.0282) (0.0307) (0.0221)
b2 = −1 −0.9975 −1.0043 −1.0029 −1.0024 −1.0011
(0.0489) (0.0395) (0.0431) (0.0344) (0.0339)
SSEE 1.23E−03 1.49E−04 1.59E−04 9.65E−05 1.83E−05
SNR = 8.26
a1 = 1.5 1.5397 1.5073 1.4941 1.4917 1.4987
(0.1008) (0.0667) (0.0415) (0.0378) (0.0318)
a2 = 1 0.9009 1.0119 0.9992 0.9949 1.0000
(0.3138) (0.0449) (0.0275) (0.0285) (0.0319)
b1 = 2.5 2.5663 2.5117 2.4975 2.4898 2.5024
(0.1582) (0.0958) (0.0650) (0.0552) (0.0419)
b2 = −1 −1.1488 −0.9777 −1.0002 −1.0017 −0.9946
(0.4724) (0.0671) (0.0544) (0.0478) (0.0491)
SSEE 3.79E−02 8.28E−04 4.14E−05 2.01E−04 3.64E−05
8. Concluding remarks
Fig. 2. Recursive estimation under Poisson sampling process for different SNRs.
Under the framework of irregular and random sampling,
identifiability, identification algorithms, convergence properties,
and input design of linear time-invariant continuous systems There are still several significant open questions. Applications
have been investigated in this paper. The identifiability conditions of the algorithms and input design developed in this paper to
obtained in this paper have established that the parameters practical systems are important steps to verify true values and also
of coprime rational systems can be uniquely determined from limitations of the algorithms. Extension to nonlinear systems is
noise-free input–output data if the number of the sampling possible and worth investigation.
points exceeds some fixed positive integer in a given time
interval regardless of the sampling schemes (periodic, irregular,
or random). The strongly convergent iterative and recursive Appendix
algorithms have been developed to identify system parameters
under noisy observations. The convergence properties of these
algorithms depend critically on input richness, sampling schemes, In the following proofs, we encounter frequently the partial
and sampling sizes. The persistent excitation conditions under derivatives (gradients) of g (t , θ ) and f (t , θ ) with respect to θ . It
irregular and random sampling have been established. is straightforward to verify that in the region of convergence of the
B. Mu et al. / Automatica 60 (2015) 100–114 109
Table 4
Iterative estimation under i.i.d. random sampling for different SNRs.
N 600 1200 1800 2400 3000
SNR = 145.51
a1 = 1.5 1.5003 1.5005 1.5034 1.5030 1.5043
(0.0246) (0.0165) (0.0127) (0.0151) (0.0139)
a2 = 1 1.0049 1.0022 0.9963 0.9962 0.9961
(0.0190) (0.0112) (0.0090) (0.0102) (0.0102)
b1 = 2.5 2.5003 2.4977 2.5013 2.5022 2.5041
(0.0199) (0.0159) (0.0111) (0.0135) (0.0112)
b2 = −1 −0.9988 −0.9981 −1.0025 −1.0028 −1.0036
(0.0159) (0.0107) (0.0073) (0.0093) (0.0079)
SSEE 2.60E−05 1.41E−05 3.37E−05 3.62E−05 6.32E−05
SNR = 9.09
a1 = 1.5 1.5436 1.5308 1.4996 1.5026 1.4986
(0.1558) (0.1140) (0.0821) (0.0791) (0.0626)
a2 = 1 0.9483 0.9639 0.9771 0.9761 0.9840
(0.0804) (0.0625) (0.0482) (0.0395) (0.0347)
b1 = 2.5 2.5338 2.5249 2.4996 2.4979 2.4932
(0.1414) (0.1062) (0.0814) (0.0789) (0.0727)
b2 = −1 −1.0340 −1.0286 −1.0089 −1.0073 −1.0029
(0.0971) (0.0642) (0.0485) (0.0454) (0.0393)
SSEE 6.87E−03 3.69E−03 6.05E−04 6.37E−04 3.11E−04
SNR = 2.27
a1 = 1.5 1.2593 1.3085 1.3589 1.3809 1.4043
(0.1894) (0.1568) (0.1581) (0.1680) (0.1555)
a2 = 1 1.0779 1.0490 1.0308 1.0345 1.0394
(0.1533) (0.1346) (0.1344) (0.1115) (0.0906)
b1 = 2.5 2.2796 2.3283 2.3816 2.4004 2.4264
(0.2163) (0.1589) (0.1403) (0.1475) (0.1494)
b2 = −1 −0.8645 −0.8988 −0.9282 −0.9347 −0.9453
(0.1417) (0.1159) (0.1143) (0.1130) (0.1044)
SSEE 1.31E−01 7.88E−02 4.00E−02 2.96E−02 1.91E−02
Laplace transform where h′ (t , θ ) is the transpose of h(t , θ ). By (23), (24), and the
′ linearity property of the Laplace transform, we have
∂ G(s, θ ) b(s)[sn−1 , . . . , s, 1] [sn−1 , . . . , s, 1]
= − , U (s)
L h′ (t , θ )γ = (a(s)β(s) − b(s)α(s)) = 0 ∀s ∈ C,
∂θ a2 ( s ) a( s )
a2 (s)
and its inverse Laplace transform is where β(s) , β1 sn−1 + · · · + βn−1 s + βn and α(s) , α1 sn−1 + · · · +
αn−1 s + αn . It follows from u(t ) ̸≡ 0 that
∂ g (t , θ ) ∂ G(s, θ )
= L−1 ,
∂θ ∂θ a(s)β(s) − b(s)α(s) = 0, ∀s ∈ C . (25)
Apparently, if α(s) = 0 (or β(s) = 0), then (25) implies β(s) = 0
which remains a stable exponential polynomial. It is clear that
∂ g (t ,θ) (or α(s) = 0). Therefore, suppose that neither β(s) nor α(s) is the
g (t , θ) and ∂θ are continuous in the region (0, tf ) × A for any
zero function. Then b(s)/a(s) = β(s)/α(s). Since deg β(s) ≤ n − 1
0 < t < tf and θ ∈ A. and deg α(s) ≤ n − 1, the right-hand side is a rational function of
order at most n − 1 and the left-hand side is a rational function of
Proof of Lemma 1. By (4), we have order n. But this contradicts the fact that a(s) and b(s) are coprime
(so that gcd(a(s), b(s)) = 1 and the order of b(s)/a(s) cannot be
∂ f (t , θ ) ∂ g (t , θ ) reduced). As a result, the only solution is β(s) = 0 and α(s) = 0,
h(t , θ ) = = ⋆ u(t ). (22)
∂θ ∂θ which implies that the vector γ = [α1 , . . . , αn , β1 , . . . , βn ]′ = 0.
Therefore, the lemma is proved.
Its equivalent representation in the frequency domain is given by
Proof of Lemma 3. By (23), we have
∂ G(s, θ )
L (h(t , θ )) = U ( s) L(h(t , θ ))
∂θ
′ −b(s)d(s)[sn−1 , . . . , s, 1] d(s)[sn−1 , . . . , s, 1] ′
b(s)[sn−1 , . . . , s, 1] [sn−1 , . . . , s, 1]
= − , U (s) = , . (26)
a2 (s) a(s) a2 (s)c (s) a(s)c (s)
Table 5
Iterative estimation under Poisson sampling process for different SNRs.
N 600 1200 1800 2400 3000
SNR = 630.65
a1 = 1.5 1.4958 1.4973 1.4975 1.4984 1.4978
(0.0066) (0.0052) (0.0057) (0.0039) (0.0042)
a2 = 1 0.9953 0.9987 1.0003 1.0001 0.9998
(0.0093) (0.0049) (0.0051) (0.0047) (0.0043)
b1 = 2.5 2.4952 2.4966 2.4970 2.4980 2.4967
(0.0100) (0.0070) (0.0086) (0.0063) (0.0070)
b2 = −1 −1.0081 −1.0031 −0.9998 −1.0008 −1.0011
(0.0174) (0.0087) (0.0081) (0.0078) (0.0079)
SSEE 1.28E−04 3.04E−05 1.56E−05 6.87E−06 1.73E−05
SNR = 39.42
a1 = 1.5 1.5138 1.5057 1.5033 1.5028 1.5047
(0.0360) (0.0259) (0.0148) (0.0151) (0.0143)
a2 = 1 0.9983 0.9993 1.0094 1.0003 1.0026
(0.0380) (0.0400) (0.0242) (0.0229) (0.0175)
b1 = 2.5 2.5236 2.5085 2.5052 2.5071 2.5084
(0.0503) (0.0401) (0.0234) (0.0224) (0.0195)
b2 = −1 −0.9925 −0.9987 −0.9848 −0.9967 −0.9912
(0.0729) (0.0639) (0.0402) (0.0387) (0.0279)
SSEE 8.05E−04 1.07E−04 3.57E−04 6.93E−05 1.78E−04
SNR = 9.85
a1 = 1.5 1.5451 1.5435 1.5121 1.5034 1.4907
(0.1358) (0.1217) (0.0435) (0.0375) (0.0310)
a2 = 1 0.9339 0.9072 1.0083 1.0043 0.9960
(0.3217) (0.3073) (0.0271) (0.0301) (0.0200)
b1 = 2.5 2.5575 2.5742 2.5176 2.5010 2.4834
(0.2055) (0.2033) (0.0655) (0.0517) (0.0396)
b2 = −1 −1.0841 −1.1281 −0.9721 −0.9873 −1.0008
(0.4737) (0.4783) (0.0543) (0.0593) (0.0400)
SSEE 1.68E−02 3.24E−02 1.30E−03 1.92E−04 3.78E−04
d(s)si
which implies that each element of h(t , θ ) is either L−1 a(s)c (s)
Proof of Theorem 3. The theorem is proved inductively. We first
show that (12)–(13) hold for k = 0. Since JN (θ ) is Lipschitz
b(s)d(s)si
or L−1 a2 (s)c (s)
, 0 ≤ i ≤ n − 1. Assume that M is the set that continuous on some convex and compact subset Θ of R2n
contains all the modes of the exponential polynomials generated containing the true value θ ∗ , there exists a constant η > 0 such
by a2 (s)c (s), i.e., M = {pi,j (t ), i = 1, . . . , lw , j = 1, . . . , ni } with that
j−1
ni = 2n + m and pi,j (t ) = (jt−1)! exp(λi t ), and {λi , i =
lw
i=1 ∥JN (x) − JN (y)∥ ≤ η∥x − y∥, ∀x, y ∈ Θ . (28)
1, . . . , lw } are the lw distinct roots of a2 (s)c (s) with multiplicity
ni . Then each element of h(t , θ ) is a weighted sum of the modes Furthermore, we have ∥JN (x)∥ ≤ ν, ∀x ∈ Θ , where ν is a positive
{pi,j (t )}. As a result, h′ (t , θ ) can be expressed as the following linear real number. Since JN (θ ∗ ) is full column rank by Lemma 3, the
combination minimum eigenvalue λ of JN′ (θ ∗ )JN (θ ∗ ) is positive. Selecting any
λ
initial value θ0 ∈ O(θ ∗ , ϵ) with ϵ = 4νη , then
h′ (t , θ) = [p1,1 (t ), . . . , p1,n1 (t ), . . . , plw ,nlw (t )]Λ, (27)
∥JN′ (θ ∗ )JN (θ ∗ ) − JN′ (θ0 )JN (θ0 )∥
where Λ is the corresponding coefficient matrix. Λ is full column
rank since h(t , θ ) is linearly independent by Lemma 1. ≤ ∥JN′ (θ ∗ )(JN (θ ∗ ) − JN (θ0 ))∥ + ∥(JN′ (θ ∗ ) − JN′ (θ ∗ ))JN (θ0 )∥
Under the sampling points (t1 , . . . , tN ), denote ≤ 2νη∥θ ∗ − θ0 ∥ < λ/2.
p (t ) p1,n1 (t1 ) plw ,nlw (t1 )
−1
··· ··· Since ∥ JN′ (θ ∗ )JN (θ ∗ ) ∥ ≤ 1/λ, the conditions
1 ,1 1
of Lemma 4 hold,
p1,1 (t2 ) ··· p1,n1 (t2 ) ··· plw ,nlw (t2 )
−1
. and hence JN (θ0 )JN (θ0 ) is nonsingular and JN′ (θ0 )JN (θ0 )
′
≤
..
PN = .. ..
. . . 2/λ whenever θ0 ∈ O(θ ∗ , ϵ) with ϵ = 4νη λ
.
p1,1 (tN ) ··· p1,n1 (tN ) ··· plw ,nlw (tN )
As JN (θ ) is Lipschitz continuous on Θ and r (θ ∗ ) = 0, we have
Apparently, JN (θ ) = PN Λ. Suppose that there exists a vector γ =
∥ − r (θ0 ) + JN (θ0 )(θ ∗ − θ0 )∥
[α1 , . . . , αn , β1 , . . . , βn ]′ ̸= 0 such that JN (θ )γ = PN Λγ = 0. 1
Define a function w(t ) = [p1,1 (t ), . . . , p1,n1 (t ), . . . , plw ,nlw (t )]Λγ .
= −JN (θ0 + s(θ ∗ − θ0 ))(θ ∗ − θ0 )ds + JN (θ0 )(θ ∗ − θ0 )
Then, JN (θ )γ = [w(t1 ), . . . , w(tN )]′ = 0 implies that w(t ) has at 0
least N zeros in [0 T ]. However, by Lemma 2, the number of zeros 1
f (t , θ) = φ(t , θ ) + O(µt ),
p2
(29)
= f 2 (t , θ )p(t )dt w.p.1. (31)
p1
where φ(t , θ ) = i=1 Ai |G(jω i, θ )| cos(iω t +
̸ G(jωi, θ )) is a
L
periodic function with period T0 , i.e., φ(t + T0 , θ ) = φ(t , θ ). It Since f (t , θ ) is a bounded exponential polynomial, the integral (31)
follows from (29) that exists if p2 is finite. When p2 = ∞, noting that p(t ) is a density
function, for any A2 > A1 > 0,
∂ f (tk , θ ) ∂φ(tk , θ )
h(tk , θ ) = = + O(µtk ), A2
A2
∂θ ∂θ
f 2 (t , θ )p(t )dt ≤ M
µt + cos(ωt ) + 1 p(t )dt
where A1 A1
A2
A1 →∞
∂φ(tk , θ ) ∂|G(jωi, θ )| p(t )dt −−−→ 0,
L
≤ M
= Ai cos(iωtk + ̸ G(jω, θ )) A1
∂θ i =1
∂θ
L where M is a positive constant and 0 < µ < 1. This means that
− Ai |G(jωi, θ )| sin(iωtk the integral (31) also exists under this setting. Similarly, we have
i =1 N
1
∂ ̸ G(jωi, θ ) Q (θ ) = lim (f (tk , θ ) − f (tk , θ ∗ ))2
G(jωi, θ ))
+̸ N →∞ N k=1
∂θ
= G(θ )ζk , = E (f (tk , θ ) − f (tk , θ ∗ ))2
p2
is also a periodic function with period T0 , and
= (f (t , θ ) − f (t , θ ∗ ))2 p(t )dt w.p.1 (32)
p1
ζk = [cos w tk + ̸ G(jω, θ ) , sin w tk + ̸ G(jω, θ ) , . . . ,
which is well defined. It follows that the Hessian matrix of Q (θ ) at
cos w Ltk + ̸ G(jωL, θ ) , sin w Ltk + ̸ G(jωL, θ ) ]′ .
the point θ ∗
Therefore, for any fixed m > 0 we have p2
Qθ θ (θ ∗ ) = 2 h(t , θ ∗ )h′ (t , θ ∗ )p(t )dt .
m+N −1 p1
1
h(tk , θ )h′ (tk , θ )
N k =m
In what follows, we will show that the Hessian matrix Qθ θ (θ ∗ ) is
positive definite. Let β = [β1 , β2 , . . . , β2n ]′ ̸= 0 be a column
0 −1
N
∂φ(tm+l , θ ) ∂φ(tm+l , θ ) ′
1 N 1 vector such that
= +O
N N0 l=0 ∂θ ∂θ N p2
β ′ Qθ θ (θ ∗ )β = 2 β ′ h(t , θ ∗ )h′ (t , θ ∗ )β p(t )dt = 0.
0 −1
N
∂φ(tm+l , θ ) ∂φ(tm+l , θ )
′
N →∞ 1 p1
−−−→
N 0 l =0 ∂θ ∂θ Then we have β ′ h(t , θ ∗ ) = 0 since p(t ) is a density function.
N By Lemma 1, h(t , θ ∗ ) is linearly independent, which derives that
0
∂φ(tl , θ ) ∂φ(tl , θ ) ′
1 β = 0. As a result, we have proved that Qθ θ (θ ∗ ) is positive definite.
= , M (θ ),
N0 l=1 ∂θ ∂θ Hence, Q (θ ) has a unique minimum at θ = θ ∗ . It follows that
Assumption 3(i) is true. Assumption 3(ii) can be proved similarly.
where ⌊a⌋ represents the integer part of a positive real number a. As the set A defined in Section 2.1 is open, the true value θ ∗ is an
As a result, we have interior point of some Θ containing θ ∗ . Since
m+N −1 N
1 1
h(tk , θ ∗ )h′ (tk , θ ∗ ) M (θ ) = lim h(tk , θ )h′ (tk , θ ) = Eh(tk , θ )h′ (tk , θ )
N k =m
N →∞ N i=1
N0
p2
∂φ(tl , θ ∗ ) ∂φ(t , θ ∗ )
′
1 = h(t , θ )h′ (t , θ )p(t )dt w.p.1,
−−−→
N →∞ N0 l=1 ∂θ ∂θ p1
Lemma 5. Under Assumption 6, for any initial phase ξ , the following It follows from (38) that for any ϵ > 0,
assertions hold at the frequency ω satisfying 0 < |ϕ(ω)| < 1 and
∞ 1 m+N −1
|ϕ(2ω)/ϕ(ω)| < 1:
P exp(jωtk ) > ϵ
N =1
N k=m
m+N −1
1 m+ N −1
cos(ωtk + ξ ) −−−→ 0 w.p.1,
2
(33)
N N →∞ ∞ E N1 exp(jωtk )
k=m k=m
≤
1
m+N −1
N =1
ϵ2
sin(ωtk + ξ ) −−−→ 0 w.p.1, (34)
N k=m
N →∞ ∞
1
= O 2
< ∞. (39)
N
while at the frequency ω satisfying 0 < |ϕ(2ω)| < 1 and N =1
|ϕ(4ω)/ϕ(2ω)| < 1 there hold Hence, by the Borel–Cantelli lemma (Chow & Teicher, 2003), one
derives that
m+N −1
1
cos2 (ωtk + ξ ) −−−→ 1/2 w.p.1, (35) 1
m+N −1
N k=m
N →∞ exp(jωtk ) −
→ 0 w.p.1 as N → ∞.
N k=m
m+N −1
1
sin2 (ωtk + ξ ) −−−→ 1/2 w.p.1. (36) Similarly, we can prove (34). Noting cos2 (ωtk + ξ ) = 1/2 +
N N →∞
k=m
1/2 cos(2ωtk + 2ξ ), by (33) we have
m+N −1
1
Proof. Because cos2 (ωtk + ξ )
N k=m
which is equivalent to Proof of Theorem 10. Under Assumption 5, the noise-free output
is
m+N −1
1
f (t , θ ) = φ(t , θ ) + O(µt ),
exp(jωtk ) −−−→ 0 w.p.1. (37) (40)
N k=m
N →∞
Using tk =
k
τi , the characteristic function of tk equals µ < 1.
i=1
Similar to the derivations in Theorem 7, under Assumptions 5
k
and 6, the gradient of (40) is given by
E exp(jωtk ) = E exp jω τi
∂ f (tk , θ ) ∂φ(tk , θ )
i=1 h(tk , θ ) = = + O(µtk )
k ∂θ ∂θ
G(θ )
ζk + O(µtk ),
= E exp(jωτi ) = ϕ k (ω). =
i =1
where
Since {τk } is an i.i.d. sequence, we have
ζk = [cos w1 tk + ̸ G(jω1 , θ ) , sin w1 tk + ̸ G(jω1 , θ ) , . . . ,
cos wL tk + ̸ G(jωL , θ ) , sin wL tk + ̸ G(jωL , θ ) ]′ .
+N −1
1 m 2
E exp(jωtk )
N
Denote ζ N = ζk
ζk′ . Then the diagonal elements of ζ N
k=m 1
m+N −1
N k=m
m+N −1 m+N −1 are either
1
= E exp(jωtk ) exp(jωtl )
N2 k=m l=m 1
m+N −1
cos2 wi tk + ̸ G(jωi , θ ) ,
1≤i≤L
ϕ(2ω) k ϕ k+1 (ω) − ϕ m+N (ω)
m+N −1 N k=m
2
=
N2 k=m
ϕ(ω) 1 − ϕ(ω) or
m+N −1
1 ϕ m (2ω) − ϕ m+N (2ω)
1
1
sin2 wi tk + ̸ G(jωi , θ ) 1 ≤ i ≤ L,
+ =O . (38) N
N2 1 − ϕ(2ω) N2 k=m
B. Mu et al. / Automatica 60 (2015) 100–114 113
which tend to 1/2 via (35) and (36). On the other hand, the non- Granas, A., & Dugundji, J. (2003). Fixed point theory. New York: Springer-Verlag.
diagonal elements of ζ N have the forms: Istratescu, V. I. (1981). Fixed point theory: an introduction. Vol. 7. Netherlands:
Springer.
m+N −1 Jennrich, R. I. (1969). Asymptotic properties of non-linear least squares estimators.
1
cos(wp tk + ̸ G(jωp , θ )) sin(wq tk + ̸ G(jωq , θ )), The Annals of Mathematical Statistics, 40, 633–643.
N k =m Johansson, R. (2010). Continuous-time model identification and state estimation
using non-uniformly sampled data. In Proceedings of the 19th international
1 ≤ p, q ≤ L; symposium on mathematical theory of networks and systems–MTNS. Vol. 5.
m+N −1
1 Kushner, H. J., & Yin, G. (2003). Stochastic approximation algorithms and applications
sin(wp tk + ̸ G(jωp , θ )) sin(wq tk + ̸ G(jωq , θ )), (2nd ed.). New York: Springer-Verlag.
N k =m Larsson, E. K., Mossberg, M., & Soderstrom, T. (2007). Identification of continuous-
1 ≤ p, q ≤ L and p ̸= q; time ARX models from irregularly sampled data. IEEE Transactions on Automatic
Control, 52, 417–427.
or Larsson, E. K., & Söderström, T. (2002). Identification of continuous-time AR
processes from unevenly sampled data. Automatica, 38, 709–718.
m+N −1
1 Ljung, L. (1999). System identification: theory for the user. Englewood Cliffs: Prentice-
cos(wp tk + ̸ G(jωp , θ )) cos(wq tk + ̸ G(jωq , θ )), Hall.
N k =m Ljung, L., & Vicino, A. (Eds.) (2005). IEEE Transactions on Automatic Control, 50,
1 ≤ p, q ≤ L and p ̸= q. Special Issue on Identification.
Marelli, D., & Fu, M. (2010). A continuous-time linear system identification method
It follows that the non-diagonal elements of ζ N approach to for slowly sampled data. IEEE Transactions on Signal Processing, 58, 2521–2533.
zeros via (33), (34), and the triangular formulas: cos(α) sin(β) = Mas-Colell, A. (1979). Homeomorphisms of compact, convex sets and the Jacobian
matrix. SIAM Journal on Mathematical Analysis, 10, 1105–1109.
1/2(sin(α + β) − sin(α − β)), sin(α) sin(β) = −1/2(cos(α + β) −
Mossberg, M. (2008). High-accuracy instrumental variable identification of
cos(α − β)), and cos(α) cos(β) = 1/2(cos(α + β) + cos(α − β)). continuous-time autoregressive processes from irregularly sampled noisy data.
As a result, we have ζ N −−−→ 21 I w.p.1. IEEE Transactions on Signal Processing, 56, 4087–4091.
N →∞
Persson, N., & Gustafsson, F. (2001). Event based sampling with application to
i=1 τi and the sequence {τk } is i.i.d, we
k
Noting that tk = vibration analysis in pneumatic tires. In 2001 IEEE International conference on
k
i=1 τi = E τi
= (E µτ1 )k . Since E µτ1 = acoustics, speech, and signal processing, 2001. proceedings. Vol. 6, (ICASSP’01).
µ µ i =1 µ
tk
k
have E = E (pp. 3885–3888). IEEE.
2τ1 τ1
µ p(t )dt < 1 and E µ /E µ < 1 for any 0 < µ < 1, where
t
Phillips, C. L., & Nagle, H. T. (2007). Digital control system analysis and design. Prentice
2 Hall Press.
p(t ) is the density function of τk , we have E N1 µ
m+N −1 tk
k=m = Proakis, J. G., & Manolakis, D. G. (2007). Digital signal processing. Pearson Prentice
Hall.
1
O N2
and Rudin, W. (1973). Functional analysis. New York: McGraw-Hill.
Sanderberg, I. W. (1980). Global inverse function theorems. IEEE Transactions on
Circuits and Systems, 27(11), 998–1004.
∞ +N −1
1 m ∞
1
P µ >ϵ =
tk
O <∞ Vajda, S., Valko, P., & Godfrey, K. R. (1987). Direct and indirect least squares methods
N N 2 in continuous-time parameter estimation. Automatica, 23, 707–718.
N =1 k=m N =1
Wang, L. Y., Feng, W., & Yin, G. G. (2013). Joint state and event observers for linear
similar to the derivations in (38) and (39). By the Borel– switching systems under irregular sampling. Automatica, 49, 894–905.
Wang, L. Y., Li, C., Yin, G. G., Guo, L., & Xu, C.-Z. (2011). State observability
µk
m+N −1 t
Cantelli lemma (Chow & Teicher, 2003), we get N1 k=m and observers of linear-time-invariant systems under irregular sampling and
−−−→ 0. It follows that sensor limitations. IEEE Transactions on Automatic Control, 56, 2639–2654.
N →∞
Wang, L. Y., Yin, G. G., Zhang, J., & Zhao, Y. (2010). System identification with quantized
m+N −1 observations. Boston: Birkhäuser.
1
Wu, F. F., & Desoer, C. A. (1972). Global inverse function theorem. IEEE Transactions
h(tk , θ )h′ (tk , θ ) on Circuit Theory, 19, 199–201.
N k =m
Yuz, J., Alfaro, J., Agüero, J. C., & Goodwin, G. C. (2011). Identification of continuous-
m +N −1 time state-space models from non-uniform fast-sampled data. IET Control
1
G(θ )ζ N
= G′ (θ ) + O(µtk ) Theory & Applications, 5, 842–855.
N k=m Zhu, Y., Telkamp, H., Wang, J., & Fu, Q. (2009). System identification using slow and
irregular output samples. Journal of Process Control, 19, 58–67.
1
−−−→ G(θ )
G′ (θ ) = M (θ ) w.p.1. (41)
N →∞ 2
It is obtained that M (θ ∗ ) is positive definite if the matrix
G(θ ∗ ) has Biqiang Mu was born in Sichuan, China, 1986. He received
the Bachelor of Engineering degree in Material Forma-
full rank. Similarly, the limit in (19) also exists as that used in the tion and Control Engineering from Sichuan University in
proof above. 2008, and the Ph.D. degree in Operations Research and Cy-
bernetics from the Institute of Systems Science, Academy
of Mathematics and Systems Science, Chinese Academy
References of Sciences in 2013. He was a postdoctoral fellow at the
Wayne State University from 2013 to 2014. He is currently
Åström, K. J., & Bernhardsson, B. (1999). Comparison of periodic and event based a postdoctoral fellow at the University of Western Sydney.
sampling for first-order stochastic systems. In Proceedings of the 14th IFAC world His research interests include recursive identification for
congress. Vol. 11 (pp. 301–306). Block-oriented nonlinear systems, identifiability of multi-
Åström, K. J., & Wittenmark, B. (1997). Computer-controlled systems: theory and variable linear systems, identification of continuous-time linear systems, and vari-
design (3rd ed.). New Jersey: Prentice Hall. able selection for high-dimensional nonlinear systems.
Berenstein, C. A., & Gay, R. (1995). Complex analysis and special topics in harmonic
analysis. Vol. 262. New York: Springer.
Jin Guo received the B.S. degree in Mathematics from
Chen, T., & Francis, B. A. (1995). Optimal sampled-data control systems. Springer.
Shandong University, China, in 2008, and Ph.D. degree in
Chow, Y. S., & Teicher, H. (2003). Probability theory: independence, interchangeability,
System Modeling and Control Theory from Academy of
martingales (3rd ed.). New York: Springer-Verlag.
Mathematics and Systems Science, Chinese Academy of
Ding, F., Qiu, L., & Chen, T. (2009). Reconstruction of continuous-time systems from
Sciences in 2013. Now, he is a postdoctoral researcher
their non-uniformly sampled discrete-time systems. Automatica, 45, 324–332.
in the School of Automation and Electrical Engineering,
Garnier, H., & Wang, L. (2008). Identification of continuous-time models from sampled
University of Science and Technology Beijing. His current
data. Springer.
research interests are identification and control of set-
Gillberg, J., & Ljung, L. (2010). Frequency domain identification of continuous-time
valued output systems and systems biology.
output error models, part II: Non-uniformly sampled data and b-spline output
approximation. Automatica, 46, 11–18.
114 B. Mu et al. / Automatica 60 (2015) 100–114
Le Yi Wang received the Ph.D. degree in Electrical En- Lijian Xu received the B.S. degree from the University
gineering from McGill University, Montreal, Canada, in of Science and Technology Beijing, China in 1998, M.S.
1990. Since 1990, he has been with Wayne State Univer- degree from the University of Central Florida in 2001,
sity, Detroit, Michigan, where he is currently a professor Ph.D. degree from Wayne State University in 2014. He
in the Department of Electrical and Computer Engineer- worked for AT&T, Florida, USA and Telus Communication,
ing. His research interests are in the areas of complex- Calgary, Alberta, Canada as an engineer and engineering
ity and information, system identification, robust control, manager from 2001 to 2007. He is currently an Assistant
H-infinity optimization, time-varying systems, adaptive Professor in the Department of Electrical and Computer
systems, hybrid and nonlinear systems, information pro- Engineering Technology at Farmingdale State College, the
cessing and learning, as well as medical, automotive, com- State University of New York. His research interests are in
munications, power systems, and computer applications the areas of networked control systems, digital wireless
of control methodologies. He was a keynote speaker in several international confer- communications, consensus control, vehicle platoon and safety. He received the
ences. He serves on the IFAC Technical Committee on Modeling, Identification and Best Paper Award from 2012 IEEE International Conference on Electro/Information
Signal Processing. He was an Associate Editor of the IEEE Transactions on Automatic Technology. He is a member of IEEE.
Control and several other journals, and currently is an Associate Editor of Journal of
Control Theory and Applications. He was a Visiting Faculty at University of Michi-
gan in 1996 and Visiting Faculty Fellow at University of Western Sydney in 2009 Wei Xing Zheng received the B.Sc. degree in Applied
and 2013. He is a member of a Foreign Expert Team in Beijing Jiao Tong University Mathematics in 1982, the M.Sc. degree in Electrical
and a member of the Core International Expert Group at Academy of Mathematics Engineering in 1984, and the Ph.D. degree in Electrical
and Systems Science, Chinese Academy of Sciences. He is a Fellow of IEEE. Engineering in 1989, all from Southeast University,
Nanjing, China. He is currently a Professor at University of
Western Sydney, Australia. Over the years he has also held
George Yin received the B.S. degree in Mathematics from various faculty/research/visiting positions at Southeast
the University of Delaware in 1983, M.S. degree in Elec- University, China, Imperial College of Science, Technology
trical Engineering, and Ph.D. in Applied Mathematics from and Medicine, UK, University of Western Australia, Curtin
Brown University in 1987. He joined Wayne State Univer- University of Technology, Australia, Munich University of
sity in 1987, and became a professor in 1996. His research Technology, Germany, University of Virginia, USA, and
interests include stochastic systems and applications. He University of California-Davis, USA. His research interests are in the areas of systems
severed on the IFAC Technical Committee on Modeling, and controls, signal processing, and communications.
Identification and Signal Processing, and many conference He is a Fellow of IEEE. Previously, he served as an Associate Editor for IEEE
program committees; he was Co-Chair of SIAM Confer- Transactions on Circuits and Systems-I: Fundamental Theory and Applications,
ence on Control & Its Application, 2011, and Co-Chair of IEEE Transactions on Automatic Control, IEEE Signal Processing Letters, and IEEE
a couple AMS-IMS-SIAM Summer Research Conferences; Transactions on Circuits and Systems-II: Express Briefs, and as a Guest Editor
he also chaired a number SIAM prize selection committees. He is Chair SIAM Activ- for IEEE Transactions on Circuits and Systems-I: Regular Papers. Currently, he
ity Group on Control and Systems Theory, and serves on the Board of Directors of is an Associate Editor for Automatica, IEEE Transactions on Automatic Control
American Automatic Control Council. He is an associate editor of SIAM Journal on (the second term), IEEE Transactions on Fuzzy Systems, IEEE Transactions on
Control and Optimization, and on the editorial board of a number of other journals. Cybernetics, IET Control Theory & Applications, and other scholarly journals. He
He was an Associate Editor of Automatica (2005–2011) and IEEE Transactions on is also an Associate Editor of IEEE Control Systems Society’s Conference Editorial
Automatic Control (1994–1998). He is a Fellow of IEEE and a Fellow of SIAM. Board.