Professional Documents
Culture Documents
1 s2.0 S0005109899001946 Main PDF
1 s2.0 S0005109899001946 Main PDF
Abstract
In this paper, iterative learning control (ILC) based on a quadratic performance criterion is revisited and generalized for
time-varying linear constrained systems with deterministic, stochastic disturbances and noises. The main intended area of application
for this generalized method is chemical process control, where excessive input movements are undesirable and many process variables
are subject to hard constraints. It is shown that, within the framework of the quadratic-criterion-based ILC (Q-ILC), various practical
issues such as constraints, disturbances, measurement noises, and model errors can be considered in a rigorous and systematic
manner. Algorithms for the deterministic case, the stochastic case, and the case with bounded parameter uncertainties are developed
and relevant properties such as the asymptotic convergence are established under some mild assumptions. Numerical examples are
provided to demonstrate the performance of the proposed algorithms. ( 2000 Elsevier Science Ltd. All rights reserved.
Keywords: Learning control; Predictive control; Robust control; Constraint satisfaction
1. Introduction
Iterative learning control (ILC) was originally proposed in the robotics community (Arimoto, Kawamura
& Miyazaki, 1984) as an intelligent teaching mechanism
for robot manipulators. The basic idea of ILC is to
improve the control signal for the present operation cycle
by feeding back the control error in the previous cycle.
Even though the mainstream ILC research has thus far
been carried out with mechanical systems in mind, chemical and other manufacturing processes could also bene"t
signi"cantly from it. Batch chemical processes such as the
batch reactor, batch distillation, and heat treatment processes for metallic or ceramic products are good examples. Traditionally, operations of these processes have
relied exclusively on PID feedback and logic-based controllers. Re"nement of input bias signals based on the
general concept of ILC can potentially enhance the
0005-1098/00/$ - see front matter ( 2000 Elsevier Science Ltd. All rights reserved.
PII: S 0 0 0 5 - 1 0 9 8 ( 9 9 ) 0 0 1 9 4 - 6
642
(3)
algorithm, thus improving the robustness. Their realtime algorithm can be viewed as a combination of the
popular model predictive control (Lee, Morari & Garcia,
1994a) and the iterative learning control.
The objective of this paper is to provide a more general
and comprehensive framework for quadratic-criterionbased ILCs that is capable of addressing all the issues
that were mentioned to be important for process control
applications. We focus on the iterative learning control
implementation rather than the feedback implementation, keeping in mind the fact that the conversion to the
latter type of implementation can always be done in
a straightforward manner. We "rst introduce an error
transition model that represents the transition of tracking error trajectories between two adjacent batches. We
also discuss how the e!ects of disturbances of various
types can be integrated into the transition model. Based
on this model, one-batch-ahead quadratic optimal control algorithms are derived for both the unconstrained
and constrained cases. In addition, a robust ILC algorithm that minimizes the worst-case tracking error for
the next batch is proposed. For each algorithm, relevant
mathematical properties such as the convergence, robustness, and noise sensitivity are investigated.
The rest of the paper is organized as follows: In
Section 2, the static gain representation of the dynamic
system is introduced and is converted into an error
transition model. The ILC design objective is de"ned
based on the model description. In Section 3, the quadratic-criterion-based iterative learning control (Q-ILC)
algorithm is derived for the unconstrained case, and the
analysis of the relevant properties such as the convergence, noise sensitivity, and robustness follows. A realtime output feedback implementation of the algorithm is
discussed and a comprehensive comparison with Amann
et al.'s state-feedback/error-feedforward algorithm is
made. The constrained Q-ILC algorithm is presented
with the convergence proof in Section 4. In Section 5, the
robust Q-ILC algorithm is proposed with the convergence proof. Numerical examples are given in Section 6
and conclusions are drawn in Section 7.
2. Process description and problem statement
We assume that the underlying system is described by
the following static map between input and output sequences de"ned over the time domain of a batch operation of interest:
y "Gu #G d #b,
k
k
$ k
where
yT"[yT(1)
k
k
uT"[uT(0)
k
k
dT"[dT(1)
k
k
yT(2)2yT(N)],
k
k
uT(1)2uT(N!1)],
k
k
dT(2)2dT(N)].
k
k
(6)
(7)
643
(11)
e$"Cx% #t ,
k
k
k
w and t are zero-mean (batch-index-wise) independent
k
k
identically distributed random vector variable sequences.
In most cases, it su$ces to work with the particular
model of
e6 $ "e6 $#w ,
k`1
k
k
(12)
e$"e6 $#t .
k
k
k
Note that e6 $ can be interpreted as the part of e$ that will
tend to repeat itself in the subsequent batches (e.g., the
error due to the bias in the input trajectory) while t can
be thought of as the part that is speci"c to the particular
batch. This simpli"ed representation makes sense since
the two types have entirely di!erent rami"cations in
iterative learning control.
Similarly, we can de"ne e6 as [e6 $#G(u( !u)], which
$
represents the part of the error trajectory e that will repeat
itself in the next batch with no further change in the disturbance or input trajectory. Writing the expression for e6
for two consecutive time steps and di!erencing yields
644
e6
"G(u( !u
)#e6 $ ,
k`1
k`1
$
k`1
(13)
e6 "G(u( !u )#e6 $ ,
k
$
k
k
(14)
q
e6
"e6 !G*u
#w ,
k`1
k
k`1
k
e "e6 #t ,
k
k
k
(15)
where *u
"u
!u .
k`1
k`1
k
Given the model in the form of (15), our objective is to
design a learning algorithm in the form of *u
"f (I ),
k`1
k
where I denotes the information available after the
k
k batch. We require that the learning algorithm has the
5)
following property:
eTQe Pmin eTQe
k k
u
as kPR when w "t "0 k.
k
k
(16)
Remark.
f When G does not have a full row-rank, the
above system has uncontrollable modes at (1,0). This
means that it is impossible to make eTQe zero in
general. In this case, the requirement of (16) is equivalent to
eTQ1@2U UTQ1@2e P0 as kPR
k
k
# #
3. Unconstrained Q-ILC
(17)
We consider solving the following quadratic subproblem upon the completion of the k batch to update the
5)
input trajectory for the k#1 batch:
5)
1
N,
(18)
min MeT Qe
#*uT R*u
k`1
k`1
2 k`1 k`1
u
* k`1
where Q and R are PD (positive-de"nite) matrices. We
may need an expectation operator for the quadratic cost
function depending on whether we are dealing with the
deterministic case or the stochastic case. Note the cost
function has a penalty term on the input change between
two adjacent batches. This term does not cause output
o!set since it vanishes if the input trajectory converges to
a limit.
3.1.1. Direct-error-based algorithm
For the deterministic case (for which w "t "0, i),
i
i
(15) simpli"es to
e "e !G*u
.
(19)
k`1
k
k`1
By solving a standard least-squares problem, we obtain
the following quadratic-criterion-based (Q-ILC) algorithm with direct error feedback:
u
"u #HQe ,
k`1
k
k
where
(20)
HQ"(GTQG#R)~1GTQ.
(21)
(25)
D C DC D C D
e6 (t#1)
I 0
k
"
e (t#1)
0 I
k
e6 (t)
G(t)
k
!
*u (t),
k
e (t)
G(t)
k
C D
e6 (t)
e (t)"[0 H(t)] k
, t"0,2, N!1,
k
e (t),
k
(26)
I
0
].
0
hij hij hij
ny C(t~1)>ny ny Cny ny C(N~t)>ny
(27)
e6
e
k`1
k`1
D C DC D C D
(0)
e6 (N)
I
k
# w
e (N)
I k
k
I 0
"
(0)
I 0
CD
0
k`1
(28)
645
(29)
e (t)"e (t)!C*x
(t), t"0,2, N!1.
k`1
k
k`1
The optimal state feedback law for the quadratic objective (18) can easily be derived by the application of
dynamic programming. The familiar Riccati recursion
for solving a "nite-horizon LQ tracking problem results.
e (t) is a given signal (just like the reference signal in the
k
LQ tracking problem) and does not complicate the LQ
controller design.
In comparing their algorithm with ours, we note the
following:
f In their algorithm, the e!ect of input represented by
y"Gu#b in our formulation is expressed through
a time-invariant state-space system,
x(t#1)"Ax(t)#Bu(t),
(30)
y(t)"Cx(t)#b.
Note that, if the model is indeed given in the above
form, it can easily be converted into the equivalent
choice of G simply by computing the pulse response
coe$cients. G can, in fact, be chosen to match any
time-varying linear state-space system, which would be
more appropriate for chemical batch processes.
f In their algorithm, they feed-forward the error signal
for the previous cycle directly. Our algorithm works
with the "ltered error signal e6 . This makes no di!erk
ence in the purely deterministic case or the case when
t "0, i. However, there invariably are disturbances
i
(such as measurement noises) that are speci"c to each
batch and will not repeat themselves in the future
cycles. Hence, the direct use of the error signal can
result in an over-compensation. The feedback will be
able to get rid of the e!ect of this over-compensation
only to a limited degree.
f Their algorithm, when implemented as a real-time
state feedback, certainly gives some extra robustness
against disturbances when compared to the pure
learning algorithms given in Section 3.1. However, the
disturbance model implicitly assumed in their deterministic algorithm, while it may be "ne for most mechanical systems, is thought to be too limiting for most
chemical process applications. It is straightforward to
see that their state feedback algorithm would be equivalent to our real-time output feedback algorithm (presented in Section 3.2) if the disturbance ewects can be
described through white noise added to the state variables in the state space model of (29). However, this
646
(31)
C D
C D
UT
# Q1@2G
# OUTQ1@2G"
UT
0
6#
with U~1"UT.
(32)
Also de"ne
C D
e
# OUTQ1@2e.
6#
(33)
(34)
where p
and p
denote maximum and minimum
.!9
.*/
singular values, respectively.
On the other hand, the "lter gain of the inverse modelbased ILC (I-ILC) algorithm represented by
u
"u #HIe where HI"G~1
k`1
k
k
(assuming G is invertible) is
(35)
(40)
where a is a scalar factor representing structured multiplicative uncertainty. The error evolution equation for
the true system is written as
e "(I!GrHQ)e "(I!L!aL)e .
(41)
k`1
k
k
For simplicity, let us also assume that Gr has a full
row-rank.
647
(42)
I!L!aL
g2
i . (43)
"U diag(l )UT where l "1!(1#a)
i
i
g2#r2
i
i
For convergence, Dl D(1 i, which is satis"ed when
i
!1(a(J1#2(r2/g2) i.
i i
It is interesting to note the lower bound of a is "xed at
!1 while the upper bound can be increased inde"nitely
by increasing r2. Even when a is less than !1, the rate of
i
divergence can be controlled through r2. On the other
i
hand, I-ILC does not o!er this #exibility ("xing r to be
i
0 (as in I-ILC) "xes the upper bound).
The above observation may remain valid for more
general uncertainty structures. Generalization of the
above analysis to more complex uncertainty structures,
however, appears to be a formidable task. See Section 5
for a more rigorous way of handling known model uncertainties at the design stage.
4. Constrained Q-ILC
4.1. Derivation of algorithms
4.1.1. Direct-error-feedback algorithm
In many industrial process control applications,
certain restrictions need to be placed on the process
variables in order to ensure safe, smooth operations.
Commonly, constraints imposed on the inputs (its raw
values as well as changes with respect to the time and
batch index) and the outputs. These constraints are often
expressed in the form of mathematical inequalities.
f For the raw values of inputs,
u-084u
4u)*.
(44)
k`1
f For the rate of input changes with respect to the time
index,
du-084du 4du)*,
(45)
k`1
where d denotes the di!erence along the discrete-time
index.
f For the rate of input changes with respect to the batch
index,
*u-084*u
4*u)*.
(46)
k`1
The above constraint may be useful to con"ne
the input adjustments to the linear regime when a
successive linear approximation (around u ) is used
k
to represent a nonlinear batch process.
648
where
CD C D
I
!I
Cu"
!J
G
!G
u-08!u 4*u
4u)*!u .
(48)
k
k`1
k
Under the assumption that du(0)"u(0), du
can be
k`1
written as
u
(0)
k`1
u
(1)!u
(0)
k`1
k`1
F
du "
k`1
"Ju
,
k`1
(49)
u
(N!1)!u
(N!2)
k`1
k`1
where
!I
J"
2
!I }
0 .
2 !I
(50)
(51)
and
*u-08H
k
!*u)*H
k
du-08!Ju
k
C
"
k`1
!du)*#Ju
k
y-08!y !e
k
k`1
!y)*#y !e
k
k`1
*u-08H"max(u-08!u , *u-08),
k
k
(57)
*u)*H"min(u)*!u , *u)*).
k
k
Now by substituting (19) into (18) and adding the penalty
term for the slack variable in the soft constraints, we
obtain
1
min
M*uT (GTQG#R)*u
k`1
k`1
2
u
k`1
k`1
* ,e
N.
(58)
#eT Se
!2eTQG*u
k`1 k`1
k
k`1
Since the feasible region is de"ned by the linear inequalities, the above minimization together with the constraints equation of (55) constitutes a standard quadratic
programming (QP) problem.
4.1.2. Observer-based algorithm
In the observer-based algorithm, we need to replace
e in (58) with e( and y in (55) with y( "y !e( . The
k
k@k
k
k@k
$
k@k
algorithm is summarized as follows:
y-08!y !e
4G*u
4y)*!y #e
.
(52)
k
k`1
k`1
k
k`1
The above constraints can be combined together into the
following linear inequality:
1
min
M*uT (GTQG#R)*u
k`1
k`1
2
*uk`1 ,ek`1
!2e( T QG*u
#eT Se
N
k@k
k`1
k`1 k`1
subject to
CM u*u
5CM
,
k`1
k`1
where
C6*u
5C
,e
50,
k`1
k`1@k k`1
where
e
50,
k`1
(53)
CD C D
I
!I
J
CM u"
!J
I
CM
"
k`1
!I
u-08!u
k
!u)*#u
k
du-08!Ju
k
!du)*#Ju
k
*u-08
!I
Cu"
!*u)*
!J
!G
y-08!y !e
k
k`1
!G
!y)*#y !e
k
k`1
The constraints on the input values and the rate of input
changes with respect to the batch index can be easily
combined together:
e
50,
k`1
Cu*u
5C
,
k`1
k`1
CD C
I
(54)
(55)
(56)
(59)
(60)
*u-08H
k
!*u)*H
k
du-08!Ju
k
C
"
.
k`1@k
!du)*#Ju
k
y-08!y #e( !e
$
k@k
k`1
!y)*#y !e( !e
$
k@k
k`1
(61)
well as the constraint vector can demand excessive computation time and lead to numerical complications. In
ILC, since the computation is performed between two
batches, there is su$cient time for carrying out numerically intensive computations such as the QP formulated
above. On the other hand, if the real-time output feedback implementation is desired, one would probably
have to reduce the computational load by adopting strategies such as the receding horizon control (Lee et al.,
1994a). Such on-line optimization based control strategies have already been used quite successfully in vast
number of continuous process applications (Qin &
Badgwell, 1997). On the other hand, these types of algorithms would not be suitable for applications involving
mechanical systems, where sampling intervals tend to be
very short and the breaks between the cycles are also
relatively insigni"cant.
4.3. Convergence proofs
In proving the asymptotic zeroing of the tracking error
under the constrained Q-ILC algorithm, we make the
following assumptions:
A.1. G has a full row-rank.
A.2. Input constraints, reference trajectory y , and
$
steady-state disturbance d are such that the zero=
ing of the error is possible with an input in the
feasible set, i.e., &u= such that y !Gu=!
$
G d #b"0, u-084u=4u)*
and
du-084
$ =
du=4du)*. In addition, *u-08(0 and *u)*'0.
A.3. Output constraints (with e"0) are satis"ed when
e"0.
A.4. Q, R, and S are positive de"nite.
The "rst assumption is made for simplicity. If G is rowrank de"cient, we can introduce the projection operator
UT and consider the convergence of the controllable part
of e as we did in Section 3.4.1. The assumptions A.2
k
and A.3 are clearly necessary in order for any algorithm to achieve zero tracking error while satisfying
the constraints. The last assumption is quite reasonable.
The convergence of the constrained Q-ILC is shown for
the direct-error-feedback algorithm and then for the observer-based algorithm.
Theorem 2. Under assumptions A.1}A.4 and t "w "0,
k
k
system (19) converges to the origin under the constrained
Q-ILC of (55), (58), i.e., e P0 and *u P0 as kPR.
k
k
Proof. Let us de"ne
J(e , u )
k k
1
"
min
'
O MeT Qe
k`1
2 k`1 k`1
(ek`1 ,*uk`1 ,ek`1 )|)k`1
N 50,
#eT Se
#*uT R*u
k`1 k`1
k`1
k`1
(62)
649
where )
is a convex set de"ned by the constraints in
k`1
(55) as well as the model equation e "e !G*u
.
k`1
k
k`1
Note that (e , 0, e )3)
since e
"e and e
"e
k
k
k`1
k`1
k
k`1
k
when we "x *u
"0. The cost at (e , 0, e ) is always
k`1
k
k
greater than or equal to the optimal cost, which implies
J(e , u )4'
D
k k
k`1 (ek ,0,ek )
"J(e
,u
)!1*uTR*u .
2 k
k
k~1 k~1
From this, it follows that
(63)
k 1
04J(e , u )# + *uTR*u 4J(e , u )(R.
(64)
k k
i
0 0
2 i
i/1
Hence, *u P0 as kPR.
k
Since *u P0, *u-08(0 and *u)*'0, the constraint
k
*u-084*u
4*u)* becomes inactive for large enough
k`1
k. Then, from the assumptions, (0, *u=, 0), where
k
*u="u=!u , is a feasible point. Moreover, it is clear
k
k
that the optimal solution (e
, *u
,e
) of J(e , u ) is
k`1
k`1 k`1
k k
also a feasible point. By the convexity of )
, any point
k`1
that lies between these two feasible points is feasible.
Now consider the directional derivative of cost function
'
from (e
, *u
,e
) to (0, *u=, 0).
k
k`1
k`1
k`1 k`1
!e
k`1
+'T D ek`1 uk`1 k`1 *u=!*u
k`1 ( ,* ,e )
k
k`1
!e
k`1
!e
k`1
"[eT Q *uT R eT S] *u=!*u
k`1
k`1
k
k`1
k`1
!e
k`1
)
"!eT Qe !eT Se #*uT R(*u=!*u
k
k`1
k`1
k`1 k`1
k`1 k`1
!*uT R*u
!eT Se
"!eT Qe
k`1
k`1
k`1 k`1
k`1 k`1
(65)
#*uT R*u=50.
k
k`1
The last inequality comes from the optimality of
(e
, *u
,e
). It follows that
k`1
k`1 k`1
*uT R*u 5eT Qe
#eT Se
k`1
=
k`1 k`1
k`1 k`1
50.
(66)
#*uT R*u
k`1
k`1
Since *u P0 as kPR, e P0 as kPR follows. h
k
k
In the operation of batch processes, time-varying constraints may be given to prevent abnormal process action
or to maintain the process within linear region. The
above result still holds for this more general case, given
that the above assumptions are satis"ed.
The theorem can be extended to the constrained observer-based algorithm. If we de"ne e8
Oe6
!
k`1
k`1
e(
, the dynamics of the observer error becomes
k`1@k
e8 "(I!K)e8
.
(67)
k
k~1
650
4(JJ(e(
,u
)!1*uTR*u
k~1@k~1 k~1
2 k
k
#J1mTQm #J1gTSg )2.
2 k k
2 k k
(72)
Then
JJ(e( , u )4JJ(e(
,u
)#J1mTQm
2 k k
k@k k
k~1@k~1 k~1
#J1gTSg
2 k k
4JJ(e(
,u
)#aDDKDD DDe8 DD,
k~1@k~1 k~1
i k
(73)
Fact.
1
[(a#b)TQ(a#b)#(c#d)TS(c#d)]
2
4[J1(aTQa#cTSc)#J1bTQb#J1dTSd]2, (70)
2
2
2
where Q and S are positive dexnite. Application of the
Schwarz inequality after expansion of the left-hand side
yields the above.
Theorem 3. Under assumptions A.1}A.4 and t "w "0,
k
k
system (15) converges to the origin under the observerbased constrained Q-ILC algorithm of (59), (60), i.e., e P0
k
and *u P0 as kPR.
k
Proof. Let
J(e( , u )
k@k k
"
min
['K
O1
k`1 2
e(
u
K
k`1@k
k`1
k`1
k`1
,* ,e )|)
(
N]
Qe(
#*uT R*u #eT Se
]Me( T
k`1 k`1
k`1
k`1
k`1@k k`1@k
50,
(71)
where )K
is a convex set de"ned by the constraints in
k`1
(60) as well as the observer equation in (24).
When *u
"0, e(
"e( . From e( "e(
#
k`1
k`1@k
k@k
k@k
k@k~1
m , we can see (e(
#m , 0, e #g ) is a feasible soluk
k@k~1
k
k
k
tion of (71) where
m
k
g O
k
!m
when m '0
k
elementwise.
when m 40
k
k
Through the same reasoning as in the proof of
Theorem 2 and using (70), we have
J(e( , u )4'K
D
k@k k
k`1 (e( k@k~1 `mk ,0,ek `gk )
1
#m )TQ(e(
#m )
" [(e(
k
k@k~1
k
2 k@k~1
#(e #g )TS(e #g )]
k
k
k
k
Qe(
#eTSe ]#J1mTQm
4(J1[e( T
k k
2 k@k~1 k@k~1
2 k k
#J1gTSg )2
2 k k
k
JJ(e( , u )4JJ(e( , u )#aDDKDD + DDe8 DD.
(74)
k@k k
0@0 0
i
j
j/1
From the exponential stability of e8 , there exists J.!9'0
k
such that
J(e( , u )4J.!9(R k
k@k k
Hence, expansion of (72) using the above leads to
J(e( , u )4J(e(
,u
)
k@k k
k~1@k~1 k~1
!1*uTR*u #aDDe8 DD#bDDe8 DD2,
2 k
k
k
k
where
(75)
(76)
a"2aJJ.!9DDKDD ,
i
b"(aDDKDD )2.
i
To this end,
k 1
J(e( , u )# + *uTR*u
k@k k
j
2 j
j/1
k
4J(e( , u )# + [aDojDDe8 DD#(cDoj)2DDe8 DD2](R.
0 0
0
0
j/1
(77)
Hence, *u P0 as kPR. Through the same arguments
k
as in the proof of Theorem 2, this implies e(
P0 as
k@k~1
kPR. Since m P0 as kPR, e( P0 as kPR. Since
k
k@k
e8 P0 as kPR, e P0 as kPR. h
k
k
5. Robust Q-ILC
5.1. Derivation of algorithm
In the previous sections, we have not incorporated
any model uncertainty information explicitly into the
algorithm. In this section, we propose a robust Q-ILC
algorithm that guarantees convergence and provides
optimal performance for a certain class of model
uncertainty.
(78)
B.4
In the above, B.1 and B.2 are needed for the subsequently
presented robust Q-ILC algorithm to be a convex
program and therefore computationally feasible. In the
robust control literature, this linear a$ne uncertainty
description is quite popular and is not considered to be
very limiting (Morari & Za"riou, 1989; Campo &
Morari, 1987; Lee & Cooley, 1997). Assumption B.3 is
needed in order to prove the convergence (convergence
cannot be achieved without this assumption since the
loss of rank means the loss of controllability), but the
algorithm can be implemented even when this assumption is not satis"ed.
Under the above assumptions, one sensible criterion
for determining *u
is the following one-batch-ahead
k`1
worst-case error minimization:
min
* |
u
U
k`1
k`1
(h)
max M(
O1[eT (h)Qe
k`1
k`1 2 k`1
h|#
]N
#eT Se
#*uT R*u
k`1 k`1
k`1
k`1
(79)
with
e (h)"e !G(h)*u ,
k`1
k
k`1
y-08#e !G(h)*u !y
k
k`1
d
if y !e #G(h)*u
(y-08,
d
k
k`1
0
e
"
k`1
if y-084y !e #G(h)*u
4y)*,
d
k
k`1
!y)*!e #G(h)*u
#y
k
k`1
d
if y !e #G(h)*u
'y)*,
d
k
k`1
651
hH(*u
)"arg max (
.
(83)
k`1
k`1
h|#
One immediate consequence of the convexity is that
hH(*u) lies on the boundary of #. If # is a polytope, the
maximum occurs at one of the vertices.
In addition, it is easy to show that max (
is
h|# k`1
a convex function of *u
. Note that, for any
k`1
*u , *u 3U
and a3[0,1], we have
1
2
k`1
a( D u1 H u1 #(1!a)(
D
k`1 (* ,h (* ))
k`1 (*u2 ,hH(*u2 ))
5a(
D
#(1!a)(
D
k`1 (*u1 ,hH(*ua ))
k`1 (*u2 ,hH(*ua ))
,
(84)
5(
D
k`1 (*ua ,hH(*ua ))
where *u "a*u #(1!a)*u . The "rst inequality is
a
1
2
because hH(*u ) and hH(*u ) are the worst-case param1
2
eters. The second is due to the convexity of (
with
k`1
respect to *u. Since U
is a convex set, the min}max
k`1
optimization is a convex programming problem for
which global optimum can be found. There exist e$cient
algorithms for these types of problems, such as the cutting-plane method discussed in Boyd and Barratt (1991).
Once again, with su$cient time allowed between batches,
it is conceivable to use the min}max algorithm to update
the input trajectory for an upcoming batch.
(80)
5.3. Convergence
(81)
where U
is a convex set de"ned by the inequality
k`1
constraints in (55) excluding the output constraints. Because the gain matrix depends on h, output prediction
also depends on h. To distinguish the model prediction
from the actual output, we use e(h) for the predicted value
and e for the true value.
min max M(
]N
(h)#*uT R*u
O1[eT (h)Qe
k`1
k`1
k`1
k`1 2 k`1
*uk`1 h|#
(85)
with the model equation constraint e
(h)"e !
k`1
k
G(h)*u .
k`1
Theorem 4. Under assumptions B.1}B.4, system (78) converges to the origin under the robust Q-ILC algorithm of
(85), i.e., e P0 and *u P0 as kPR.
k
k
652
(
"1[*uT (G(h)TQG(h)#R)*u
k`1
k`1
k`1 2
!2eTQG(h)*u #eTQe ].
k k
k
k`1
We de"ne
6. Numerical illustrations
(86)
(91)
(92)
653
(93)
654
derived from
0.8
G (s)"
1
15s2#8s#1
(94)
Example 3 (Robust Q-ILC). In robust Q-ILC, we assumed that the true process model is described by
1
0.8e~s
G (s)"
#h
1
(5s#1)(3s#1)
5s#1
with h"1. For !0.854h40.16, G(h) has a full rowrank. Outside this range, there exist h's which lead G(h) to
lose its row-rank. h for the true process was assumed to
vary according to the sequence, M!0.8, 0.1, !0.6, !0.2,
!0.7, !0.5, !0.3, !0.2, !0.7, !0.3N for the "rst 10
batches. In order to reduce computations in the min}max
optimization, h was assumed to have discrete values from
!0.8 to 0.1 with an interval of 0.1.
Fig. 7 shows the performance of robust Q-ILC. In this
"gure, we plot *u instead of u since the process is
k
k
represented as in (96). As has been proved, the output
is observed to converge to the desired trajectory as the
batch number increases.
(95)
e (h)"e !G(h)*u ,
k`1
k
k`1
where G(h) is derived from
(96)
(97)
(0)" ,
I
(98)
655
Fig. 8. Test inputs and resulting outputs for identi"cation of the nonlinear batch reactor in Example 4.
dC
A "!k e~E@RTC2 ,
0
A
dt
C (0)"C .
A
AI
(99)
656
7. Conclusions
In this paper, it was argued that the existing ILC
algorithms, despite their successes in controlling mechanical systems, are not well-suited for process control applications. Motivated by this, we presented new modelbased iterative learning control algorithms that were
tailored speci"cally for this type of application. The algorithms were based on quadratic performance criteria and
were designed to consider the issues relevant to process
control, such as disturbances, noises, nonlinearities, constraints, and model errors. We proved the convergence of
the error signal under the proposed algorithms. Investigation of other relevant properties such as the noise
sensitivity and robustness along with some numerical
studies indicated that these algorithms should perform as
intended but at the expense of added computational
requirements.
Acknowledgements
The "rst author (JHL) gratefully acknowledges the
"nancial support from the National Science Foundation's Young Investigator Program (USA) under the
Grant CTS d9357827. The second author would like to
acknowledge LG Yonam Foundation (Korea), the Automation Research Center (Korea) at POSTECH and
Korea Science and Engineering Foundation for "nancial
support.
References
Amann, N., Owens, D. H., & Rogers, E. (1996). Iterative
learning control for discrete-time system with exponential rate of
convergence. IEE Proceedings Control Theory Applications, 143(2),
217}224.
Arimoto, S., Kawamura, S., & Miyazaki, F. (1984). Bettering operation of robots by learning. Journal of Robotic Systems, 1(1),
123}140.
Bien, Z., & Huh, K. M. (1989). Higher-order iterative learning control
algorithm. IEE Proceedings Part D, 136(3), 105}112.
Bondi, P., Casalino, G., & Gambradella, L. (1988). On the iterative
learning control theory for robotic manipulators. IEEE Journal of
Robotics Automation, 4(1), 14}22.
Boyd, S. P., & Barratt, C. H. (1991). Linear controller design}limits of
performance. Englewood Cli!s: Prentice-Hall.
Chen, Y. (1998). A bibliographical library for ILC research.
http://shuya.ml.org:888/&yqchen/ILC/ilcref.html.
Campo, P. J., & Morari, M. (1987). Robust model predictive control.
Proceedings of ACC, Minneapolis, MN (pp. 1021}1026).
Lee, J. H., & Cooley, B. (1997). Stable min}max control of statespace systems with bounded input matrix. Proceedings of ACC,
Alberquerque, NM, June (pp. 2945}2949). A full version will appear
in Automatica in 2000.
Lee, J. H., Morari, M., & Garcia, C. E. (1994a). State space
interpretation of model predictive control. Automatica, 30(4),
707}717.
657
Won Cheol Kim was born in Yeosu, Koea
in 1963. He received MSc and PhD degrees
in Chemical Engineering from Sogang
University, Seoul, Korea, in 1990 and
1997, respectively. He is currently with
Conwell Co.,Ltd. His main research interests are iterative learning control, model
predictive control, and statistical process
control.