Professional Documents
Culture Documents
3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
C.1 Introduction
Nonlinear control is a mature subject with a variety of powerful methods and a long history of
successful industrial applications ⇒ Why so many researchers have recently showed an active
interest in the development and applications of nonlinear control methodologies ?
Ex: pendulum
• Design simplicity
Good nonlinear control designs may be simpler and more intuitive than their linear counterparts.
Ex: x& = Ax + Bu
x& = f + gu
___________________________________________________________________________________________________________
Chapter 1 Introduction 1
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Phase plane analysis is a graphical method for studying The nature of the system response corresponding to various
second-order systems. This chapter’s objective is to gain initial conditions is directly displayed on the phase plane. In
familiarity of the nonlinear systems through the simple the above example, we can easily see that the system
graphical method. trajectories neither converge to the origin nor diverge to
infinity. They simply circle around the origin, indicating the
2.1 Concepts of Phase Plane Analysis marginal nature of the system’s stability.
2.1.1 Phase portraits A major class of second-order systems can be described by the
The phase plane method is concerned with the graphical study differential equations of the form
of second-order autonomous systems described by
&x& = f ( x, x& ) (2.3)
x&1 = f1 ( x1 , x 2 ) (2.1a)
x& 2 = f 2 ( x1 , x 2 ) (2.1b) In state space form, this dynamics can be represented
with x1 = x and x 2 = x& as follows
where
x1 , x 2 : states of the system x&1 = x2
f1 , f 2 : nonlinear functions of the states x& 2 = f ( x1 , x 2 )
Geometrically, the state space of this system is a plane having 2.1.2 Singular points
x1 , x 2 as coordinates. This plane is called phase plane. The A singular point is an equilibrium point in the phase plane.
Since an equilibrium point is defined as a point where the
solution of (2.1) with time varies from zero to infinity can be
represented as a curve in the phase plane. Such a curve is system states can stay forever, this implies that x& = 0 , and
called a phase plane trajectory. A family of phase plane using (2.1)
trajectories is called a phase portrait of a system.
f1 ( x1 , x 2 ) = 0
(2.4)
Example 2.1 Phase portrait of a mass-spring system_______
f 2 ( x1 , x 2 ) = 0
x& For a linear system, there is usually only one singular point
although in some cases there can be a set of singular points.
-6 -3 3 6 x
The governing equation of the mass-spring system in Fig. 2.1
is the familiar linear second-order differential equation
unstable
-3
&x& + x = 0 (2.2)
-6
Assume that the mass is initially at rest, at length x0 . Then the divergence
area
to infinity
solution of this equation is -9
x(t ) = x0 cos(t )
Fig. 2.2 A mass-spring system and its phase portrait
x& (t ) = − x0 sin(t )
Consider the system &x& + 0.6 x& + 3x + x 2 = 0 whose phase
Eliminating time t from the above equations, we obtain the portrait is plot in Fig. 2.2.
equation of the trajectories
The system has two singular points, one at (0,0) and the other
x + x& 2 = x02
2 at (−3,0) . The motion patterns of the system trajectories in the
vicinity of the two singular points have different natures. The
This represents a circle in the phase plane. Its plot is given in trajectories move towards the point x = 0 while moving away
Fig. 2.1.b. from the point x = −3 .
__________________________________________________________________________________________ __________________________________________________________________________________________
___________________________________________________________________________________________________________
Chapter 2 Phase Plane Analysis 1
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Why an equilibrium point of a second order system is called a f ( x1 , x2 ) = f ( x1 ,− x 2 ) ⇒ symmetry about the x1 axis.
singular point ? Let us examine the slope of the phase portrait. f ( x1 , x2 ) = − f ( x1 ,− x2 ) ⇒ symmetry about the x2 axis.
The slope of the phase trajectory passing through a point
( x1 , x 2 ) is determined by f ( x1 , x2 ) = − f (− x1 ,− x 2 ) ⇒ symmetry about the origin.
Example 2.3 A first-order system_______________________ The second technique, on the other hand, involves directly
dx f (x , x )
Consider the system x& = −4 x + x there are three singular 3 eliminating the time variable, by noting that 2 = 2 1 2
dx1 f1 ( x1 , x 2 )
points, defined by − 4 x + x 3 = 0 , namely, x = 0, − 2, 2 . The and then solving this equation for a functional relation
phase portrait of the system consists of a single trajectory, and between x1 and x2 . Let us use this technique to solve the mass-
is shown in Fig. 2.3. spring equation again.
x&
Example 2.4 Mass-spring system_______________________
stable unstable By noting that &x& = (dx& / dx) /( dx / dt ) , we can rewrite (2.2) as
unstable
dx&
x x& + x = 0 . Integration of this equation yields x& 2 + x 2 = x02 .
-2 0 2
dx
__________________________________________________________________________________________
The mathematical model of the satellite is θ&& = u , where u is The method of isoclines (ñöôø ng ñaú ng khuynh)
the torque provided by the thrusters and θ is the satellite angle. The basic idea in this method is that of isoclines. Consider the
dynamics in (2.1): x&1 = f1 ( x1 , x 2 ) and x& 2 = f 2 ( x1 , x 2 ) . At a
Let us examine on the phase plane the behavior of the control point ( x1 , x2 ) in the phase plane, the slope of the tangent to the
system when the thrusters are fired according to the control trajectory can be determined by (2.5). An isocline is defined to
law be the locus of the points with a given tangent slope. An
isocline with slope α is thus defined to be
−U if θ >0
u (t ) = (2.7)
U if θ <0 dx2 f (x , x )
= 2 1 2 =α
dx1 f1 ( x1 , x 2 )
which means that the thrusters push in the counterclockwise
direction if θ is positive, and vice versa. This is to say that points on the curve
When the thrusters provide a negative torque −U , the phase Let us explain the isocline method on the mass-spring system
in (2.2): &x& + x = 0 . The slope of the trajectories is easily seen
trajectories are similarly found to be θ& 2 = −2U x + c1 , with the
to be
corresponding phase portrait as shown in Fig. 2.5.c.
dx2 x
=− 1
θ x& x& dx1 x2
antenna
x1 + α x 2 = 0
i.e., a straight line. Along the line, we can draw a lot of short
u u =U u = −U
line segments with slope α . By taking α to be different values,
Fig. 2.5 Satellite control using on-off thrusters a set of isoclines can be drawn, and a field of directions of
tangents to trajectories are generated, as shown in Fig. 2.7. To
The complete phase portrait of the closed-loop control system obtain trajectories from the field of directions, we assume that
can be obtained simply by connecting the trajectories on the the tangent slopes are locally constant. Therefore, a trajectory
left half of the phase plane in 2.5.b with those on the right half starting from any point in the plane can be found by
of the phase plane in 2.5.c, as shown in Fig. 2.6. connecting a sequence of line segments.
x x
α =∞
u = +U u = −U
switching line
Fig. 2.7 Isoclines for the mass-spring system
Fig.2.6 Complete phase portrait of the control system
Example 2.6 The Van der Pol equation__________________
The vertical axis represents a switching line, because the
control input and thus the phase trajectories are switched on For the Van der Pol equation
that line. It is interesting to see that, starting from a nonzero
initial angle, the satellite will oscillate in periodic motions &x& + 0.2( x 2 − 1) x& + x = 0
under the action of the jets. One can concludes from this phase
portrait that the system is marginally stable, similarly to the an isocline of slope α is defined by
mass-spring system in Example 2.1. Convergence of the
system to the zero angle can be obtained by adding rate dx& 0.2( x 2 − 1) x& + x
feedback. =− =α
__________________________________________________________________________________________ dx x&
___________________________________________________________________________________________________________
Chapter 2 Phase Plane Analysis 3
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Therfore, the points on the curve where x corresponding to time t and x0 corresponding to
time t 0 . This implies that, if we plot a phase plane portrait
0.2( x 2 − 1) x& + x + α x& = 0 with new coordinates x and (1 / x& ) , then the area under the
resulting curve is the corresponding time interval.
all have the same slope α .
2.4 Phase Plane Analysis of Linear Systems
By taking α of different isoclines can be obtained, as plot in
Fig. 2.8. The general form of a linear second-order system is
x2 α = 0 α = −1
x&1 = a x1 + b x2 (2.9a)
α = −5
α =1 x& 2 = c x1 + d x 2 (2.9b)
equation
2.3 Determining Time from Phase Portraits
s 2 + as + b = ( s − λ1 )( s − λ2 ) = 0
Time t does not explicitly appear in the phase plane having
x1 and x 2 as coordinates. We now to describe two techniques The roots λ1 , λ2 can be explicitly represented as
for computing time history from phase portrait. Both of
techniques involve a step-by step procedure for recovering
time. − a + a 2 − 4b − a − a 2 − 4b
λ1 = and λ2 =
2 2
Obtaining time from ∆t ≈ ∆x / x&
In a short time ∆t , the change of x is approximately For linear systems described by (2.10), there is only one
singular point (b ≠ 0) , namely the origin. However, the
∆x ≈ x& ∆t (2.8) trajectories in the vicinity of this singularity point can display
quite different characteristics, depending on the values of
where x& is the velocity corresponding to the increment ∆x . a and b . The following cases can occur
• λ1 , λ2 are both real and have the same sign (+ or -)
From (2.8), the length of time corresponding to the
• λ1 , λ2 are both real and have opposite sign
increment ∆x is ∆t ≈ ∆x / x& . This implies that, in order to
• λ1 , λ2 are complex conjugates with non-zero real parts
obtain the time corresponding to the motion from one point to
another point along the trajectory, we should divide the • λ1 , λ2 are complex conjugates with real parts equal to 0
corresponding part of the trajectory into a number of small We now briefly discuss each of the above four cases
segments (not necessarily equally spaced), find the time
associated with each segment, and then add up the results. To Stable or unstable node (Fig. 2.9.a -b)
obtain the history of states corresponding to a certain initial The first case corresponds to a node. A node can be stable or
condition, we simply compute the time t for each point on the unstable:
phase trajectory, and then plots x with respects to t and x& λ1 , λ2 < 0 : singularity point is called stable node.
with respects to t . λ1 , λ2 > 0 : singularity point is called unstable node.
There is no oscillation in the trajectories.
Obtaining time from t ≈ (1 / x& ) dx
∫
Since x& = dx / dt , we can write dt = dx / x& . Therefore, Saddle point (Fig. 2.9.c)
The second case ( λ1 < 0 < λ2 ) corresponds to a saddle point.
x Because of the unstable pole λ2 , almost all of the system
∫
t − t 0 ≈ (1 / x& ) dx
x0 trajectories diverge to infinity.
___________________________________________________________________________________________________________
Chapter 2 Phase Plane Analysis 4
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
x
r = x12 + x22 θ (t ) = tan −1 2
x1
dr dθ
= − r (r 2 − 1) = −1
dt dt
1 1
r (t ) = and θ (t ) = θ 0 − t , where c0 = −1
1 + c0 e − 2t r02
Similarly, we can find that the system (b) has an unstable limit
cycle and system (c) has a semi-stable limit cycle.
__________________________________________________________________________________________
Example 2.8________________________________________
Consider the nonlinear system
x&1 = g ( x 2 ) + 4 x1 x 22
x& 2 = h( x1 ) + 4 x12 x2
∂ f1 ∂ f 2
Since + = 4( x12 + x 22 ) , which is always strictly
∂x1 ∂x 2
positive (except at the origin), the system does not have any
limit cycles any where in the phase plane.
__________________________________________________________________________________________
___________________________________________________________________________________________________________
Chapter 2 Phase Plane Analysis 6
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
The objective of this chapter is to present Lyapunov stability has a single equilibrium point (the origin 0) if A is nonsingular.
theorem and illustrate its use in the analysis and the design of If A is singular, it has an infinity of equilibrium points, which
nonlinear systems. contained in the null-space of the matrix A, i.e., the subspace
defined by Ax = 0. A nonlinear system can have several (or
3.1 Nonlinear Systems and Equilibrium Points infinitely many) isolated equilibrium points.
θ
where
n
f ∈R : nonlinear vector function
x ∈ Rn : state vectors Fig. 3.1 Pendulum
n : order of the system
Consider the pendulum of Fig. 3.1, whose dynamics is given
The form (3.1) can represent both closed-loop dynamics of a by the following nonlinear autonomous equation
feedback control system and the dynamic systems where no
control signals are involved. MR 2θ&& + bθ& + MgR sin θ = 0 (3.5)
A special class of nonlinear systems is linear system. The where R is the pendulum’s length, M its mass, b the friction
dynamics of linear systems are of the from x& = A(t ) x with coefficient at the hinge, and g the gravity constant. Leting
x1 = θ , x2 = θ& , the corresponding state-space equation is
A ∈ R n×n .
x&1 = x2 (3.6a)
Autonomous and non-autonomous systems
Linear systems are classified as either time-varying or time- g b
x& 2 = − x − sin x1
2 2
(3.6b)
invariant. For nonlinear systems, these adjectives are replaced MR R
by autonomous and non-autonomous.
Therefore the equilibrium points are given by x2 = 0,
Definition 3.1 The nonlinear system (3.1) is said to be
sin( x1 ) = 0, which leads to the points (0 [2π ], 0) and
autonomous if f does not depend explicitly on time, i.e., if the
system’s state equation can be written (π [2π ], 0) . Physically, these points correspond to the
pendulum resting exactly at the vertical up and down points.
x& = f (x) (3.2) __________________________________________________________________________________________
Since both x * (t ) and x(t ) are solutions of (3.2): x& = f (x) , we ∀R > 0, ∃r > 0, x(0) < r ⇒ ∀t ≥ 0, x(t ) < R
have or, equivalently
x& * = f (x * ) x * (0) = x 0
∀R > 0, ∃r > 0, x(0) ∈ B r ⇒ ∀t ≥ 0, x(t ) ∈ B r
x& = f (x) x ( 0) = x 0 + δ x 0
Essentially, stability (also called stability in the sense of
then e(t ) satisfies the following non-autonomous differential Lyapunov, or Lyapunov stability) means that the system
equation trajectory can be kept arbitrarily close to the origin by starting
sufficiently close to it. More formally, the definition states that
e& = f (x * + e, t ) − f (x * , t ) = g (e, t ) (3.8) the origin is stable, if, given that we do not want the state
trajectory x(t ) to get out of a ball of arbitrarily specified radius
with initial condition e(0) = δ x(0) . Since g (0, t ) = 0 , the new B R . The geometrical implication of stability is indicated in
dynamic system, with e as state and g in place of f, has an Fig. 33.
equilibrium point at the origin of the state space. Therefore, 3 curve 1 - asymptotically stable
instead of studying the deviation of x(t ) from x * (t ) for the 1 curve 2 - marginally stable
original system, we may simply study the stability of the 2 curve 3 - unstable
perturbation dynamics (3.8) with respect to the equilibrium
point 0. However, the perturbation dynamics non-autonomous 0
system, due to the presence of the nominal trajectory x * (t ) on x(0) S r
the right hand side.
Example 3.2________________________________________
SR
Consider the autonomous mass-spring system Fig. 3.3 Concepts of stability
Clearly, this is a non-autonomous system. Asymptotic stability means that the equilibrium is stable, and
__________________________________________________________________________________________
in addition, states start close to 0 actually converge to 0 as
3.2 Concepts of Stability time goes to infinity. Fig. 3.3 shows that the system
trajectories starting form within the ball B r converge to the
Notation origin. The ball B r is called a domain of attraction of the
B R : spherical region (or ball) defined by x ≤ R equilibrium point.
S R : spherical itself defined by x = R
In many engineering applications, it is still not sufficient to
∀ : for any know that a system will converge to the equilibrium point
∃ : there exist after infinite time. There is a need to estimate how fast the
∈ : in the set system trajectory approaches 0. The concept of exponential
⇒ : implies that stability can be used for this purpose.
⇔ : equivalent
Definition 3.5 An equilibrium points 0 is exponential stable if
Stability and instability there exist two strictly positive number α and λ such that
Definition 3.3 The equilibrium state x = 0 is said to be stable
if, for any R > 0 , there exist r > 0 , such that if x(0) ≤ r then ∀t > 0, x(t ) ≤ α x(0) e −λt (3.9)
x(t ) ≤ R for all t ≥ 0 . Otherwise, the equilibrium point is
unstable. in some ball B r around the origin.
Using the above symbols, Definition 3.3 can be written in the (3.9) means that the state vector of an exponentially stable
form system converges to the origin faster than an exponential
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 8
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
function. The positive number λ is called the rate of A similar procedure can be applied for a controlled system.
exponential convergence. Consider the system &x& + 4 x& 5 + ( x 2 + 1) u = 0 . The system can
For example, the system x& = −(1 + sin 2 x) x is exponentially be linearly approximated about x = 0 as &x& + 0 + (0 + 1) u = 0 or
convergent to x = 0 with the rate λ = 1 . Indeed, its solution is &x& = u . Assume that the control law for the original nonlinear
− ∫0t −[1+sin 2 x (τ )] dτ system has been selected to be u = sin x + x 3 + x cos 2 x , then
x(t ) = x(0) e , and therefore x(t ) ≤ x(0) e −t .
the linearized closed-loop dynamics is &x& + x& + x = 0 .
__________________________________________________________________________________________
Note that exponential stability implies asymptotic stability.
But asymptotic stability does not implies guarantee The following result makes precise the relationship between
exponential stability, as can be seen from the system the stability of the linear system (3.2) and that of the original
nonlinear system (3.2).
x& = − x 2 , x(0) = 1 (3.10)
Theorem 3.1 (Lyapunov’s linearization method)
whose solution is x = 1 /(1 + t ) , a function slower than any • If the linearized system is strictly stable (i.e., if all
exponential function e − λt . eigenvalues of A are strictly in the left-half complex plane),
then the equilibrium point is asymptotically stable (for the
Local and global stability actual nonlinear system).
Definition 3.6 If asymptotic (or exponential) stability holds • If the linearizad system is un stable (i.e., if at least one
for any initial states, the equilibrium point is said to be eigenvalue of A is strictly in the right-half complex plane),
asymptotically (or exponentially) stable in the large. It is also then the equilibrium point is unstablle (for the nonlinear
called globally asymptotically (or exponentially) stable. system).
3.3 Linearization and Local Stability • If the linearized system is marginally stable (i.e., if all
Lyapunov’s linearization method is concerned with the local eigenvalues of A are in the left-half complex plane but at
stability of a nonlinear system. It is a formalization of the least one of them is on the jω axis), then one cannot
intuition that a nonlinear system should behave similarly to its conclude anything from the linear approximation (the
linearized approximation for small range motions. equilibrium point may be stable, asymptotically stable, or
unstable for the nonlinear system).
Consider the autonomous system in (3.2), and assumed that
f(x) is continuously differentiable. Then the system dynamics Example 3.5________________________________________
can be written as
Consider the equilibrium point (θ = π ,θ& = 0) of the pendulum
∂f
x& = x + f h.o.t . (x) (3.11) in the example 3.1. Since the neighborhood of θ = π , we can
∂ x x =0 write
where f h.o.t . stands for higher-order terms in x. Let us use the sin θ = sin π + cos π (θ − π ) + h.o.t. = π − θ + h.o.t.
constant matrix A denote the Jacobian matrix of f with respect ~
∂f thus letting θ = θ − π , the system’s linearization about the
to x at x = 0: A = . Then, the system
equilibrium point (θ = π ,θ& = 0) is
∂ x x =0
x& = A x (3.12) ~
&& ~& g ~
b
θ + θ − θ =0
2
MR R
is called the linearization (or linear approximation) of the
original system at the equilibrium point 0.
Hence its linear approximation is unstable, and therefore so is
In practice, finding a system’s linearization is often most the nonlinear system at this equilibrium point.
__________________________________________________________________________________________
easily done simply neglecting any term of order higher than 1
in the dynamics, as we now illustrate. Example 3.5________________________________________
Example 3.4________________________________________ Consider the first-order system x& = a x + b x 5 . The origin 0 is
one of the two equilibrium of this system. The linearization of
Consider the nonlinear system this system around the origin is x& = a x . The application of
Lyapunov’s linearization method indicate the following
x&1 = x22 + x1 cos x 2 stability properties of the nonlinear system
x& 2 = x 2 + ( x1 + 1) x1 + x1 sin x 2
• a < 0 : asymptotically stable
Its linearized approximation about x = 0 is • a > 0 : unstable
• a = 0 : cannot tell from the linearization
x&1 = 0 + x1.1
x& 2 = x 2 + 0 + x1 + x1 x2 ≈ x2 + x1 In the third case, the nonlinear system is x& = b x 5 . The
1 0 linearization method fails while, as we shall see, the direct
The linearized system can thus be written x& = x . method to be described can easily solve this problem.
1 1 __________________________________________________________________________________________
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 9
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
3.4 Lyapunov’s Direct Method 3.4.1. Positive definite functions and Lyapunov functions
Definition 3.7 A scalar continuous function V (x) is said to be
The basic philosophy of Lyapunov’s direct method is the locally positive definite if V (0) = 0 and, in a ball B R0
mathematical extension of a fundamental physical
observation: if the total energy of a mechanical (or electrical)
system is continuous dissipated, then the system, whether x≠0 ⇒ V ( x) > 0
linear or nonlinear, must eventually settle down to an
equilibrium point. Thus, we may conclude the stability of a If V (0) = 0 and the above property holds over the whole state
system by examining the variation of a single scalar function. space, then V (x) is said to be globally positive definite.
V = V2
Fig. 3.6 A nonlinear mass-damper-spring system
V = V1
Total mechanical energy = kinetic energy + potential energy x2
1 2 x 1 1 1 0
∫ (k x + k x )dx = 2 mx& x1
3 2
V ( x) = mx& + k 0 x 2 + k1 x 4
0 1 + V3 > V2 > V1
0 2 2 4
(3.14)
Comparing the definitions of stability and mechanical energy, Fig. 3.7 Typical shape of a positive definite function V ( x1 , x 2 )
we can see some relations between the mechanical energy and
the concepts described earlier: The 2-dimesional geometrical representation can be made as
follows. Taking x1 and x2 as Cartesian coordinates, the level
• zero energy corresponds to the equilibrium point
(x = 0, x& = 0) curves V ( x1 , x 2 ) = Vα typically present a set of ovals
• assymptotic stability implies the convergence of surrounding the origin, with each oval corresponding to a
mechanical energy to zero positive value of Vα .These ovals often called contour curves
• instability is related to the growth of mechanical energy may be thought as the section of the cup by horizontal planes,
projected on the ( x1 , x 2 ) plane as shown in Fig. 3.8.
The relations indicate that the value of a scalar quantity, the
mechanical energy, indirectly reflects the magnitude of the
V = V2 x2 V = V1
state vector, and furthermore, that the stability properties of
the system can be characterized by the variation of the
mechanical energy of the system. x1
0
The rate of energy variation during the system’s motion is
obtained by differentiating the first equality in (3.14) and
using (3.13) V = V3 V3 > V2 > V1
3
V& (x) = m x& &x& + (k 0 x + k1 x 3 ) x& = x& (−b x& x& ) = −b x& (3.15) Fig. 3.8 Interpreting positive definite functions using contour
curves
(3.15) implies that the energy of the system, starting from
some initial value, is continuously dissipated by the damper Definition 3.8 If, in a ball B R0 , the function V (x) is positive
until the mass is settled down, i.e., x& = 0 . definite and has continuous partial derivatives, and if its time
derivative along any state trajectory of system (3.2) is
The direct method of Lyapunov is based on generalization of
negative semi-definite, i.e., V& (x) ≤ 0 then, V (x) is said to be a
the concepts in the above mass-spring-damper system to more
complex systems. Lyapunov function for the system (3.2).
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 10
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
A Lyapunov function can be given simple geometrical Obviously, this function is locally positive definite. As a mater
interpretations. In Fig. 3.9, the point denoting the value of of fact, this function represents the total energy of the
V ( x1 , x2 ) is seen always point down an inverted cup. In Fig. pendulum, composed of the sum of the potential energy and
3.10, the state point is seen to move across contour curves the kinetic energy. Its time derivative yields
corresponding to lower and lower value of V .
V& (x) = θ& sin θ + θ&θ&& = −θ& 2 ≤ 0
V = V2 x2 V = V1 x&1 = x1 ( x12 + x 22 − 2) − 4 x1 x 22
x& 2 = 4 x12 x2 + x 2 ( x12 + x22 − 2)
x1
0 around its equilibrium point at the origin.
V ( x1 , x 2 ) = x12 + x22
V = V3 V3 > V2 > V1
Fig. 3.10 Illustrating Definition 3.8 for n=2 using contour its derivative V& along any system trajectory is
curves
V& = 2( x12 + x 22 )( x12 + x 22 − 2)
3.4.2 Equilibrium point theorems
Lyapunov’s theorem for local stability Thus, is locally negative definite in the 2-dimensional ball B 2 ,
Theorem 3.2 (Local stability) If, in a ball B R0 , there exists a i.e., in the region defined by ( x12 + x22 ) < 2 . Therefore, the
scalar function V (x) with continuous first partial derivatives above theorem indicates that the origin is asymptotically
such that stable.
__________________________________________________________________________________________
A simple pendulum with viscous damping is described as Consider the nonlinear system
x& + c( x) = 0
θ&& + θ& + sin θ = 0
Consider the following scalar function where c is any continuous function of the same sign as its
scalar argument x , i.e., such as x c( x) > 0 ∀x ≠ 0 . Intuitively,
1 this condition indicates that − c(x ) ’pushes’ the system back
V (x) = (1 − cosθ ) + θ& 2
2 towards its rest position x = 0 , but is otherwise arbitrary.
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 11
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Since c is continuous, it also implies that c(0) = 0 (Fig. 3.13). Local invariant set theorem
Consider as the Lyapunov function candidate the square of The invariant set theorem reflect the intuition that the decrease
distance to the origin V = x 2 . The function V is radially of a Lyapunov function V has to graduate vanish (i.e., ) V&
has to converge to zero) because V is lower bounded. A
unbounded, since it tends to infinity as x → ∞ . Its derivative
precise statement of this result is as follows.
is V& = 2 x x& = −2 x c ( x) . Thus V& < 0 as long as x ≠ 0 , so
that x = 0 is a globally asymptotically stable equilibrium point. Theorem 3.4 (Local Invariant Set Theorem) Consider an
autonomous system of the form (3.2), with f continuous, and
let V (x) be a scalar function with continuous first partial
c (x )
derivatives. Assume that
• for some l > 0 , the region Ω l defined by V (x) < l is
bounded
0 x • V& (x) ≤ 0 for all x in Ω l
Let R be the set of all points within Ω where V& (x) = 0 , and l
M be the largest invariant set in R. Then, every solution x(t )
originating in Ω l tends to M as t → ∞ .
Fig. 3.13 The function c(x )
⊗ Note that:
For instance, the system x& = sin 2 x − x is globally convergent - M is the union of all invariant sets (e.g., equilibrium
to x = 0 , since for x ≠ 0 , sin 2 x ≤ sin x ≤ x . Similarly, the points or limit cycles) within R
- In particular, if the set R is itself invariant (i.e., if once
system x& = − x 3 is globally asymptotically convergent to x = 0 .
V& = 0 , then ≡ 0 for all future time), then M=R
Notice that while this system’s linear approximation ( x& ≈ 0) is
inconclusive, even about local stability, the actual nonlinear The geometrical meaning of the theorem is illustrated in Fig.
system enjoys a strong stability property (global asymptotic 3.14, where a trajectory starting from within the bounded
stability). region Ω l is seen to converge to the largest invariant set M.
__________________________________________________________________________________________
Note that the set R is not necessarily connected, nor is the set
Example 3 .10______________________________________ M.
The asymptotic stability result in the local Lyapunov theorem
Consider the nonlinear system can be viewed a special case of the above invariant set
theorem, where the set M consists only of the origin.
x&1 = x2 − x1 ( x12 + x22 )
x& 2 = − x1 − x 2 ( x12 + x22 ) V =l
V
The origin of the state-space is an equilibrium point for this
Ωl
system. Let V be the positive definite function V = x12 + x22 .
Its derivative along any system trajectory is V& = −2( x12 + x 22 ) 2 R
which is negative definite. Therefore, the origin is a globally
asymptotically stable equilibrium point. Note that the
globalness of this stability result also implies that the origin is M
the only equilibrium point of the system. x0 x2
__________________________________________________________________________________________
x1
⊗ Note that:
Fig. 3.14 Convergence to the largest invariant set M
- Many Lyapunov function may exist for the same system.
- For a given system, specific choices of Lyapunov
Let us illustrate applications of the invariant set theorem using
functions may yield more precise results than others.
some examples.
- Along the same line, the theorems in Lyapunov analysis
are all sufficiency theorems. If for a particular choice of
Example 3 .11______________________________________
Lyapunov function candidate V , the condition on V& are
not met, we cannot draw any conclusions on the stability Asymptotic stability of the mass-damper-spring system
or instability of the system – the only conclusion we
For the system (3.13), we can only draw conclusion of
should draw is that a different Lyapunov function
marginal stability using the energy function (3.14) in the local
candidate should be tried.
equilibrium point theorem, because V& is only negative semi-
3.4.3 Invariant set theorem definite according to (3.15). Using the invariant set theorem,
Definition 3.9 A set G is an invariant set for a dynamic system however, we can show that the system is actually
if every system trajectory which starts from a point in G asymptotically stable. TO do this, we only have to show that
remains in G for all future time. the set M contains only one point.
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 12
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
x&1 = x2 − x17 ( x14 + 2 x 22 − 10) Corollary: Consider the autonomous system (3.2), with f
continuous, and let V (x) be a scalar function with continuous
x& 2 = − x13 − 3 x 25 ( x14 + 2 x 22 − 10)
partial derivatives. Assume that in a certain neighborhood Ω
of the origin
Note that the set defined by x14 + 2 x22 = 10 is invariant, since • is locally positive definite
• V& (x) is negative semi-definite
d 4
( x1 + 2 x 22 − 10) = −(4 x110 + 12 x 26 )( x14 + 2 x22 − 10) • the set R defined by V& (x) = 0 contains no trajectories of
dt (3.2) other than the trivial trajectory x ≡ 0
Then, the equilibrium point 0 is asymptotically stable.
which is zero on the set. The motion on this invariant set is Furthermore, the largest connected region of the form
described (equivalently) by either of the equations (defined by V (x) < l ) within Ω is a domain of attraction of the
equilibrium point.
x&1 = x 2
x& 2 = − x13 Indeed, the largest invariant set M in R then contains only the
equilibrium point 0.
Therefore, we see that the invariant set actually represents a
⊗ Note that:
limit circle, along which the state vector moves clockwise. Is
- The above corollary replaces the negative definiteness
this limit circle actually attractive ? Let us define a Luapunov
condition on V& in Lyapunov’s local asymptotic stability
function candidate V = ( x14 + 2 x22 − 10) 2 which represents a
theorem by a negative semi-definiteness condition on V& ,
measure of the “distance” to the limit circle. For any arbitrary
combined with a third condition on the trajectories within
positive number l , the region Ω l , which surrounds the limit
R.
circle, is bounded. Its derivative - The largest connected region of the form Ω l within Ω is a
domain of attraction of the equilibrium point, but not
V& = −8( x110 + 3x 26 )( x14 + 2 x22 − 10) 2 necessarily the whole domain of attraction, because the
function V is not unique.
Thus V& is strictly negative, except if x14 + 2 x 22 = 10 or - The set Ω itself is not necessarily a domain of attraction.
Actually, the above theorem does not guarantee that Ω is
x110 + 3 x 26 = 0 , in which cases V& = 0 . The first equation is invariant: some trajectories starting in Ω but outside of
simply that defining the limit cycle, while the second equation the largest Ω l may actually end up outside Ω .
is verified only at the origin. Since both the limit circle and the
origin are invariant sets, the set M simply consists of their Global invariant set theorem
union. The above invariant set theorem and its corollary can be
simply extended to a global result, by enlarging the involved
Thus, all system trajectories starting in Ω l converge either to region to be the whole space and requiring the radial
the limit cycle or the origin (Fig. 3.15) unboundedness of the scalar function V .
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 13
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Theorem 3.5 (Global Invariant Set Theorem) Consider an as long as x ≠ 0 . Thus the system cannot get “stuck” at an
autonomous system of the form (3.2), with f continuous, and equilibrium value other than x = 0 ; in other words, with R
let V (x) be a scalar function with continuous first partial being the set defined by x& = 0 , the largest invariant set M in R
derivatives. Assume that contains only one point, namely [ x = 0, x& = 0] . Use of the
• V& (x) ≤ 0 over the whole state space local invariant set theorem indicates that the origin is a locally
asymptotically stable point.
• V (x) → ∞ as x → ∞
Let R be the set of all points where V& (x) = 0 , and M be the b(x& ) c (x )
largest invariant set in R. Then all solutions globally
asymptotically converge to M as t → ∞
For instance, the above theorem shows that the limit cycle
0 x& 0 x
convergence in Example 3.13 is actually global: all system
trajectories converge to the limit cycle (unless they start
exactly at the origin, which is an unstable equilibrium point).
Fig. 3.17 The functions b(x& ) and c(x )
Because of the importance of this theorem, let us present an
additional (and very useful) example.
x
where b and c are continuous functions verifying the sign Example 3 .15 Multimodal Lyapunov Function___________
conditions x& b( x& ) > 0 for x& ≠ 0 and x c( x& ) > 0 for x ≠ 0 . The Consider the system
dynamics of a mass-damper-spring system with nonlinear
damper and spring can be described by the equation of this πx
&x& + x 2 − 1 x& 3 + x = sin
form, with the above sign conditions simply indicating that the 2
otherwise arbitrary function b and c actually present π y 1 2 x
“damping” and “spring” effects. A nonlinear R-L-C (resistor-
inductor-capacitor) electrical circuit can also be represented by
Chose the Lyapunov function V =
2
dy .
2
x& +
∫ y − sin
0
the above dynamic equation (Fig. 3.16) This function has two minima, at x = ±1, x& = 0 , and a local
maximum in x (a saddle point in the state-space) at
vC = c (x ) v L = &x& v R = b(x& ) x = 0, x& = 0 . Its derivative V& = − x 2 − 1 x& 4 , i.e., the virtual
which can be thought of as representing the power dissipated 3.5 System Analysis Based on Lyapunov’s Direct Method
in the system. Furthermore, by hypothesis, x& b( x& ) = 0 only if
How to find a Lyapunov function for a specific problem ?
x& = 0 . Now x& = 0 implies that &x& = −c (x) , which is non-zero
There is no general way of finding Lyapunov function for
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 14
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
nonlinear system. Faced with specific systems, we have to use Lyapunov functions for linear time-invariant systems
experience, intuition, and physical insights to search for an Given a linear system of the form x& = A x , let us consider a
appropriate Lyapunov function.
quadratic Lyapunov function candidate V& = xT P x , where P
In this section, we discuss a number of techniques which can is a given symmetric positive definite matrix. Its derivative
facilitate the otherwise blind of Lyapunov functions. yields
3.5.1 Lyapunov analysis of linear time-invariant systems V& = x& T P x + x T P x& = -xT Q x (3.18)
Symmetric, skew-symmetric, and positive definite matrices where
Definition 3.10 A square matrix M is symmetric if M=MT (in
other words, if ∀i, j M ij = M ji ). A square matrix M is skew- A T P + P A = -Q (3.19)
symmetric if M = −M T (i.e., ∀i, j M ij = − M ji ). (3.19) is so-called Lyapunov equation. Note that Q may be not
p.d. even for stable systems.
⊗ Note that:
Example 3 .17 ______________________________________
- Any square n × n matrix can be represented as the sum of
a symmetric and a skew-symmetric matrix. This can be
Consider the second order linear system with A = 0 4 .
shown in the following decomposition − 8 − 12
M + MT M - MT
M= + If we take P = I , then - Q = P A + A T P = 0 −4 . The
142 243 142 243 − 4 − 24
symmetric skew− symmetric matrix Q is not p.d.. Therefore, no conclusion can be draw
- The quadratic function associated with a skew-symmetric from the Lyapunov function on whether the system is stable or
matrix is always zero. Let M be a n × n skew-symmetric not.
__________________________________________________________________________________________
matrix and x is an arbitrary n × 1 vector. The definition of
skew-symmetric matrix implies that xT M x = − xT M T x . A more useful way of studying a given linear system using
T T T quadratic functions is, instead, to derive a p.d. matrix P from a
Since x M x is a scalar, x M x = − x M x which yields given p.d. matrix Q, i.e.,
∀x, x T M x = 0 (3.16)
• choose a positive definite matrix Q
In the designing some tracking control systems for robot, • solve for P from the Lyapunov equation
this fact is very useful because it can simplify the control • check whether P id p.d.
law.
- (3.16) is a necessary and sufficient condition for a matrix
M to be skew-symmetric. If P is p.d., then (1 / 2)x T P x is a Lyapunov function for the
linear system. And the global asymptotical stability is
Definition 3.11 A square matrix M is positive definite (p.d.) if guaranteed.
x ≠ 0 ⇒ xT M x > 0 .
Theorem 3.6 A necessary and sufficient condition for a LTI
⊗ Note that: system x& = A x to be strictly stable is that, for any symmetric
- A necessary condition for a square matrix M to be p.d. is p.d. matrix Q, the unique matrix P solution of the Lyapunov
that its diagonal elements be strictly positive. equation (3.19) be symmetric positive definite.
- A necessary and sufficient condition for a symmetric
matrix M to be p.d. is that all its eigenvalues be strictly Example 3 .18 ______________________________________
positive. Consider again the second order linear system in Example
- A p.d. matrix is invertible.
p p
- A .d. matrix M can always be decomposed as 3.18. Let us take Q = I and denote P by P = 11 12 ,
p 21 p 22
M = U T ΛU (3.37)
where due to the symmetry of P, p 21 = p12 . Then the
where U T U = I , Λ is a diagonal matrix containing the
Lyapunov equation is
eigenvalues of M
- There are some following facts
p11 p12 0 4 0 − 8 p11 p12 − 1 0
p 21 p 22 − 8 − 12 + 4 − 12 p 21 p 22 = 0 − 1
2 2
• λmin (M ) x ≤ xT Mx ≤ λmax (M ) x
• x T Mx = xT U T ΛUx = z T Λz where Ux = z
• λmin (M ) I ≤ Λ ≤ λmax (M ) I whose solution is p11 = 5 , p12 = p 22 = 1 . The corresponding
• zT z = x
2
matrix P = 5 1 is p.d., and therefore the linear system is
1 1
globally asymptotically stable.
The concepts of positive semi-definite, negative definite, and __________________________________________________________________________________________
negative semi-definite can be defined similarly. For instance, a
square n × n matrix M is said to be positive semi-definite 3.5.2 Krasovskii’s method
(p.s.d.) if ∀x, x T M x ≥ 0 . A time-varying matrix M(t) is Krasovskii’s method suggests a simplest form of Lyapunov
uniformly positive definite if ∃α > 0, ∀t ≥ 0, M (t ) ≥ α I . function candidate for autonomous nonlinear systems of the
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 15
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
form (3.2), namely, V = f T f . The basic idea of the method is function for this system. If the region Ω is the whole state
simply to check whether this particular choice indeed leads to space, and if in addition, V (x) → ∞ as x → ∞ , then the
a Lyapunov function. system is globally asymptotically stable.
Theorem 3.7 (Krasovkii) Consider the autonomous system 3.5.3 The Variable Gradient Method
defined by (3.2), with the equilibrium point of interest being The variable gradient method is a formal approach to
the origin. Let A(x) denote the Jacobian matrix of the system, constructing Lyapunov functions.
i.e.,
∂f To start with, let us note that a scalar function V (x) is related
A(x ) =
∂x to its gradient ∇V by the integral relation
If the matrix F = A + A T is negative definite in a
x
neighborhood Ω , then the equilibrium point at the origin is
asymptotically stable. A Lyapunov function for this system is
V ( x ) = ∇V dx
∫ 0
∂∇V1 ∂∇V2 ∂a12 ∂a The convergence rate estimate is largest for Q = I . Indeed, let
= ⇒ a12 + x 2 = a 21 + x1 21
∂x2 ∂x1 ∂x 2 ∂x1 P0 be the solution of the Lyapunov equation corresponding to
Q = I is
If the coefficients are chosen to be a11 = a 22 = 1, a12 = a 21 = 0
A T P0 + P0 A = −I
which leads to ∇V = x , ∇V = x then V& can be computed
1 1 2 2 and let P the solution corresponding to some other choice of
as Q
x1 x2 x12 + x 22 A T P + PA = −Q1
V ( x) =
∫ 0
x1 dx1 +
∫ 0
x2 dx 2 =
2
(3.22)
Without loss of generality, we can assume that λmin (Q1 ) = 1
since rescaling Q1 will rescale P by the same factor, and
This is indeed p.d., and therefore, the asymptotic stability is
guaranteed. therefore will not affect the value of the corresponding γ .
Subtract the above two equations yields
If the coefficients are chosen to be a11 = 1, a12 = x 22 , A T (P - P0 ) + (P - P0 ) A = −(Q1 - I )
a 21 = 3 x22 , a 22 = 3 , we obtain the p.d. function Now since λmin (Q1 ) = 1 = λmax (I ) , the matrix (Q1 - I) is
positive semi-definite, and hence the above equation implies
x12 3 2 that (P - P0 ) is positive semi-definite. Therefore
V ( x) = + x2 + x1 x 23 (3.23)
2 2
λmax (P ) ≥ λmax (P0 )
Since λmin (Q1 ) = 1 = λmin (I ) , the convergence rate estimate
whose derivative is V& = −2 x12 − 6 x 22 − 2 x 22 ( x1 x 2 − 3 x12 x22 ) .
γ = λmin (Q) / λmax (P )
corresponding to Q = I the larger than (or equal to) that
We can verify that V& is a locally negative definite function
(noting that the quadratic terms are dominant near the origin), corresponding to Q = Q1 .
and therefore, (3.23) represents another Lyapunov function for
the system. Estimating convergence rates for nonlinear systems
__________________________________________________________________________________________
The estimation convergence rate for nonlinear systems also
3.5.4 Physically motivated Lyapunov functions involves manipulating the expression of V& so as to obtain an
explicit estimate of V . The difference lies in that, for
3.5.5 Performance analysis nonlinear systems, V and V& are not necessarily quadratic
function of the states.
Lyapunov analysis can be used to determine the convergence
rates of linear and nonlinear systems. Example 3 .22 ______________________________________
where α is a real number. Then W (t ) ≤ W (0) e −α t Choose the Lyapunov function candidate V = x
2
, its
dV
The above Lemma implies that, if W is a non-negative derivative is V& = 2V (V − 1) . That is = −2dt . The
function, the satisfaction of (3.26) guarantees the exponential V (1 − V )
convergence of W to zero. solution of this equation is easily found to be
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 17
Applied Nonlinear Control Nguyen Tan Tien - 2002.3
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
α e −2dt V ( 0)
V (x) = , where α = .
1 + α e −2dt 1 − V (0)
2
If x(0) = V (0) < 1 , i.e., if the trajectory starts inside the
unit circle, then α > 0 , and V (t ) < α e −2t . This implies that
the norm x(t ) of the state vector converges to zero
exponentially, with a rate of 1.
However, if the trajectory starts outside the unit circle, i.e., if
V (0) > 1 , then α < 0 , so that V (t ) and therefore x tend to
infinity in a finite time (the system is said to exhibit finite
escape time, or “explosion”).
__________________________________________________________________________________________
___________________________________________________________________________________________________________
Chapter 3 Fundamentals of Lyapunov Theory 18
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
x& = f (x, t ) (4.1) The asymptotic stability requires that there exists an attractive
region for every initial time t 0 .
equilibrium points x* are defined by
Definition 4.3 The equilibrium point 0 is exponentially stable
f (x* , t ) ≡ 0 ∀t ≥ t 0 (4.2) if there exist two positive numbers, α and λ , such that for
sufficiently small x(t 0 ) ,
Note that this equation must be satisfied ∀t ≥ t 0 , implying that x(t ) ≤ α x 0 e −λ (t −t0 ) ∀t ≥ t 0
the system should be able to stay at the point x* all the time.
For instance, we can easily see that the linear time-varying
Definition 4.4 The equilibrium point 0 is global stable ∀x(t 0 ) ,
system
x(t ) → 0 as t → ∞
x& = A(t ) x (4.3)
Example 4.2 A first-order linear time-varying system_______
has a unique equilibrium point at the origin 0 unless A(t) is
always singular. Consider the first-order system x& (t ) = −a(t ) x(t ) . Its solution is
t
Definition 4.1 The equilibrium point 0 is stable at t 0 if for convergent with rate 1 / 2 .
__________________________________________________________________________________________
any R > 0 , there exists a positive scalar r ( R, t 0 ) such that
Uniformity in stability concepts
x (t 0 ) < r ⇒ x(t ) < R ∀t ≥ t 0 (4.6) The previous concepts of Lyapunov stability and asymptotic
stability for non-autonomous systems both indicate the
importance effect of initial time. In practice, it is usually
Otherwise, the equilibrium point 0 is unstable. desirable for the systems to have a certain uniformity in its
behavior regardless of when the operation starts. This
The definition means that we can keep the state in ball of motivates us to consider the definitions of uniform stability
arbitrarily small radius R by starting the state trajectory in a and uniform asymptotic stability. Non-autonomous systems
ball of sufficiently small radius r . with uniform properties have some desirable ability to
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 19
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
withstand disturbances. The behavior of autonomous systems Definition 4.7 A scalar time-varying function V (x, t ) is locally
is dependent of the initial time, all the stability properties of an positive definite if V (0, t ) = 0 and there exits a time-variant
autonomous system are uniform.
positive definite function V0 (x) such that
Definition 4.5 The equilibrium point 0 is locally uniformly
stable if the scalar r in Definition 4.1 can be chosen ∀t ≥ t 0 , V (x, t ) ≥ V0 (x) (4.7)
independent of t 0 , i.e., if r = r (R) .
Thus, a time-variant function is locally positive definite if it
Definition 4.6 The equilibrium point at the origin is locally dominates a time-variant locally positive definite function.
uniformly asymptotically stable if Globally positive definite functions can be defined similarly.
• it is uniformly stable
• there exits a ball of attraction B R0 , whose radius is Definition 4.8 A scalar time-varying function V (x, t ) is said to
be decrescent if V (0, t ) = 0 , and if there exits a time-variant
independent of t 0 , such that any trajectory with initial
positive definite function V1 (x) such that
states in B R0 converges to 0 uniformly in t 0 .
∀t ≥ 0, V (x, t ) ≥ V1 (x) (4.7)
By uniform convergence in terms of t 0 , we mean that for
all R1 and R2 satisfying 0 < R2 < R1 ≤ R0 , ∃T ( R1 , R2 ) > 0 such In other word, a scalar function V (x, t ) is decrescent if it is
dominated by a time-invariant p.d. function.
that, ∀t ≥ t 0
Example 4.4________________________________________
x(t 0 ) < R1 ⇒ x(t ) < R2 ∀t ≥ t 0 + T ( R1 , R2 )
Consider time-varying positive definite functions as follows
i.e., the trajectory, starting from within a ball B R1 , will
V (x, t ) = (1 + sin 2 t )( x12 + x22 )
converges into a smaller ball B R2 after a time period T which
V0 (x) = x12 + x22
is independent of t 0 .
V1 (x) = 2( x12 + x22 )
By definition, uniform asymptotic stability always implies
asymptotic stability. The converse (ñaû o ñeà ) is generally not ⇒ V (x, t ) dominates V0 (x) and is dominated by V1 (x)
true, as illustrated by the following example. because V0 (x) ≤ V (x, t ) ≤ V1 (x) .
__________________________________________________________________________________________
Example 4.3________________________________________
x Given a time-varying scalar function V (x, t ), its derivative
Consider the first-order system x& = − . This system has
1+ t along a system trajectory is
1 + t0
general solution x(t ) = x(t 0 ). The solution asymptotically d V ∂V ∂V ∂V ∂V
1+ t = + x& = + f ( x, t ) (4.8)
converges to zero. But the convergence is not uniform. dt ∂t ∂x ∂t ∂x
Intuitively, this is because a larger t 0 requires a longer time to
Lyapunov theorem for non-autonomous system stability
get close to the origin.
__________________________________________________________________________________________ The main Lyapunov stability results for non-autonomous
systems can be summarized by the following theorem.
The concept of globally uniformly asymptotic stability can be
defined be replacing the ball of attraction B R0 by the whole Theorem 4.1 (Lyapunov theorem for non-autonomous
systems)
state space.
Stability: If, in a ball B R0 around the equilibrium point 0,
4.2 Lyapunov Analysis of Non-Autonomous Systems there exits a scalar function V (x, t ) with continuous partial
derivatives such that
In this section, we extend the Lyapunov analysis results of
chapter 3 to the stability of non-autonomous systems. 1. V is positive definite
2. V& is negative semi-definite
4.2.1 Lyapunov’sdirect method for non-autonomous systems
then the equilibrium point 0 is stable in the sense of
The basic idea of the direct method, i.e., concluding the
Lyapunov.
stability of nonlinear systems using scalar Lyapunov functions,
can be similarly applied to non-autonomous systems. Besides
Uniform stability and uniform asymptotic stability: If,
more mathematical complexity, a major difference in non-
furthermore
autonomous systems is that the powerful La Salle’s theorems
do not apply. This drawback will partially be compensates by 3. V is decrescent
a simple result in section 4.5 called Barbalat’s lemma.
then the origin is uniformly stable. If the condition 2 is
Time-varying positive definite functions and decrescent strengthened by requiring that V& be negative definite, then
functions the equilibrium point is asymptotically stable.
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 20
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Global uniform asymptotic stability: If, the ball B R0 is A simple result, however, is that the time-varying system
(4.17) is asymptotically stable if the eigenvalues of the
replaced by the whole state space, and condition 1, the
strengthened condition 2, condition 3, and the condition symmetric matrix A(t ) + A T (t ) (all of which are real) remain
strictly in the left-half complex plane
4. V (x, t ) is radially unbounded
∃λ > 0, ∀i, ∀t ≥ 0, λi ( A(t ) + A T (t )) ≤ −λ (4.19)
are all satisfied, then the equilibrium point at 0 is globally
uniformly asymptotically stable. This can be readily shown using the Lyapunov function
Similarly to the case of autonomous systems, if in a certain V = x T x , since
neighborhood of the equilibrium point, V is positive definite
and V& , its derivative along the system trajectories, is negative V& = x T x& + x& T x = x T ( A(t ) + A T (t )) ≤ −λ xT x& = −λ V
semi-definite, then V is called Lyapunov function for the non-
autonomous system. so that ∀t ≥ 0, 0 ≤ xT x = V (t ) ≤ V (0) e − λt and therefore x
tends to zero exponentially.
Example 4.5 Global asymptotic stability_________________ It is important to notice that the result provides a sufficient
Consider the system defined by
condition for any asymptotic stability.
x& = [ A1 + A 2 (t )] x (4.20)
Chose the Lyapunov function candidate
Since LTI systems are asymptotically stable if their Sufficient smoothness conditions on the A(t ) matrix
eigenvalues all have negative real parts ⇒ Will the system
(4.17) be stable if any time t ≥ 0 , the eigenvalues of A(t ) all Consider the linear system (4.17), and assume that at any time
have negative parts ? t ≥ 0 , the eigenvalues of A(t ) all have negative real parts
x&1 − 1 e 2t x1
x& 2 = 0 − 1 x 2 (4.18) If, in addition, the matrix A(t ) remains bounded, and
∞
Both eigenvalues of A(t ) equal to -1 at all times. The solution
−t t
∫ 0
A T (t ) A(t ) dt < ∞ (i.e., the integral exists and is finite)
of (4.18) can be rewritten as x2 = x 2 (0) e , x&1 + x1 = x 2 (0) e .
Hence, the system is unstable. then the system is globally exponentially stable.
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 21
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
4.2.3 The linearization method for non-autonomous systems Therem 4.3 If the Jacobian matrix A(t ) is constant,
Lyapunov’s linearization method can also be developed for A(t ) = A 0 , and if (4.23) is satisfied, then the instability of the
non-autonomous systems. Let a non-autonomous system be
described by (4.1) and 0 be an equilibrium point. Assume that linearized system implies that of the original non-autonomous
f is continuously differentiable with respect to x. Let us denote nonlinear system, i.e., (4.1) is unstable if one or more of the
eigenvalues of A 0 has a positive real part .
∂f
A(t ) = (4.22) 4.3 Instability Theorems
∂ x x =0
4.4 Existence of Lyapunov Functions
The for any fixed time t (i.e., regarding t as a parameter), a
Taylor expansion of f leads to 4.5 Lyapunov-Like Analysis Using Barbalat’s Lemma
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 22
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
- A simple sufficient condition for a differentiable function This implies that V (t ) ≤ V (0) , and therefore, that e and θ are
to be uniformly continuous is that its derivative be bound. bounded. But the invariant set cannot be used to conclude the
This can be seen from the finite different theorem: ∀t , t1 , convergence of e , because the dynamics is non-autonomous.
∃t 2 (t ≤ t 2 ≤ t1 ) such that g (t ) − g (t1 ) = g& (t 2 )(t − t1 ) . To use Barbalat’s lemma, let us check the uniform continuity
And therefore, if R1 > 0 is an upper bound on the of V& . The derivative of V& is V&& = −4e (−e + θ w) . This shows
function g& , we can always use η = R / R1 independently that V&& is bounded, since w is bounded by hypothesis, and e
of t1 to verify the definition of uniform continuity. and θ were shown above to be bounded. Hence, V& is
uniformly continuous. Application of Babarlat’s lemma then
Example 4.12_______________________________________ indicates that e → 0 as t → ∞ .
Note that, although e converges to zero, the system is not
Consider a strictly stable linear system whose input is bounded. asymptotically stable, because θ is only guaranteed to be
Then the system output is uniformly continuous. bounded.
__________________________________________________________________________________________
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 23
Applied Nonlinear Control Nguyen Tan Tien - 2002.4
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Example 4.14 A strictly positive real function_____________ which shows that h4 is SPR (since it is also strictly stable). Of
course, condition (4.34) can also be checked directly on a
1 computer.
Consider the rational function h( p ) = , which is the
p+λ __________________________________________________________________________________________
transfer function of a first-order system, with λ > 0 . ⊗ The basic difference between PR and SPR transfer
Corresponding to the complex variable p = σ + jω , functions is that PR transfer functions may tolerate poles on
the jω axis, while SPR functions cannot.
1 σ + λ − jω
h( p ) = =
(σ + jω ) + λ (σ + λ ) 2 + ω 2 Example 4.16_______________________________________
1
Consider the transfer function of an integrator h( p ) = . Its
Obviously, Re[ h( p )] ≥ 0 if σ ≥ 0 . Thus, h( p ) is a positive real p
function. In fact, one can easily see that h( p ) is strictly σ − jω
value corresponding to p = σ + jω is h( p ) = . We
positive real, for example by choosing ε = λ / 2 in Definition σ 2 +ω 2
4.9. can easily see from Definition 4.9 that h( p ) is PR but not SPR.
__________________________________________________________________________________________
__________________________________________________________________________________________
A T P + P A = -q q T − ε L
Pb = c + γ q
___________________________________________________________________________________________________________
Chapter 4 Advanced Stability Theory 25
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
6. Feedback Linearization
u = (v − f ) / b (6.7)
A(h) h& = u − a 2 gh (6.2)
we can cancel the nonlinearities and obtain the simple input-
If u(t) is chosen as output relation (multiple-integrator form) x ( n) = v . Thus, the
control law v = − k 0 x − k1 x& − K − k n −1 x ( n −1) with the ki chosen
u (t ) = a 2 gh + A(h)v (6.3)
so that the polynomial p n + k n −1 p n −1 + K + k 0 has its roots
with v being an “equivalent input” to be specified, the strictly in the left-half complex plane, lead to exponentially
resulting dynamics is linear h& = v stable dynamics x ( n ) + k n −1 x ( n −1) + K + k 0 x = 0 which
Choosing v as implies that x(t ) → 0 . For tasks involving the tracking of the
~ desired output xd (t), the control law
v = −α h (6.4)
~
with h = h(t ) − hd is the level error, α is a strictly positive v = x d ( n) − k 0 e − k1e& − K − k n −1e ( n −1) (6.8)
constant. Now, the close loop dynamics is
(where e(t ) = x(t ) − x d (t ) is the tracking error) leads to
~
h& + α h = 0 (6.4) exponentially convergent tracking.
___________________________________________________________________________________________________________
Chapter 6 Feedback linearization 26
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Example 6.2: Feedback linearization of a two-link robot first put the dynamics into the controllability canonical form
Consider the two-link robot as in the Fig. 6.2 before using the above feedback linearization design.
The dynamics of a two-link robot Example: Consider a simple second order system
Now, consider the new dynamics (6.15). It is linear and &y& = ( x 2 + 1) u + f1 ( x ) (6.21)
controllable. Using the well known linear state feedback
control law v = − k1 z1 − k 2 z 2 , one could chose k1 = 2, k 2 = 0 f1 ( x ) = ( x15 + x 3 )( x 3 + cos x 2 ) + ( x 2 + 1) x12 (6.22)
or
Clearly, (6.21) represents an explicit relationship between y
v = −2 z 2 (6.16) and u . If we choose the control input to be in the form
1
resulting in the stable closed-loop dynamics z&1 = −2 z1 + z 2 u= (v − f 1 ) (6.23)
x2 +1
and z& 2 = −2 z 2 . In term of the original state, this control law
corresponds to the original input where v is the new input to be determined, the nonlinearity in
(6.21) is canceled, and we apply a simple linear double-
1 integrator relationship between the output and the new input v,
u= (−2 a x 2 − 2 sin x1 − cos x1 sin x1 + 2 x1 cos x1 )
a cos(2 x1 ) &y& = v . The design of tracking controller for this double-
(6.17) integrator relation is simple using linear technique. For
The original state x is given from z by instance, letting e = y (t ) − y d (t ) be the tracking error, and
choosing the new input v such as
x1 = z1 (6.18a)
x 2 = ( z 2 − sin z1 ) / a (6.18b) v = &y&d − k1e − k 2 e& (6.24)
The closed-loop system under the above control law is where k1 , k 2 are positive constant. The tracking error of the
represented in the block diagram in Fig. 6.3. closed-loop system is given by
To generate a direct relationship between the output and input, x& 2 + x 23 = y& d − e (6.30)
let us differentiate the output y& = x&1 = sin x 2 + ( x 2 + 1) x 3 .
Since y& is still not directly relate to the input u , let us which is non-autonomous and nonlinear. However, in view of
differentiate again. We now obtain the facts that e is guaranteed to be bound by (6.29) and y& d is
___________________________________________________________________________________________________________
Chapter 6 Feedback linearization 28
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
x&1 x 23 + u
& = (6.27a)
x 2 u
y = x1 (6.27b)
x& 2 + x 23 = 0 (6.45)
___________________________________________________________________________________________________________
Chapter 6 Feedback linearization 29
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
7. Sliding Control
f ( x ) : unknown, bounded nonlinear function = (Φ / λ )(1 − e − λt ) ≤ Φ / λ . Apply the same procedure, we get
b ( x ) : control gain, unknown bounded but known sign ~
x ≤ Φ/ λ−λt = ε .
7.1.1 A Notation Simplification From previous results, one has z1 ≤ φ / λn −1−i , where z1 is
- Tracking error in the variable x
~ the output of the ( n − i − 1) th filter. Furthermore, noting that
x ≡ x − xd
i
- Tracking error vector p λ ~ Φ λ
= 1− ⇒ x (i ) ≤ 1 + = ( 2 λ ) i ε is
x~ ≡ x − x = ~ x ~ x& L ~ x ( n −1)
d [ ]T p+λ p+λ λn −1−i λ
bounded. In the case that x~(0) ≠ 0 , bounds (7.4) are obtained
- Time-varying surface S (t ) in the state-space R (n) by the asymptotically, i.e., within a short time-constant ( n − 1) /λ .
scalar equation s ( x; t ) = 0 , where The simplified, 1st-order problem of keeping the scalar s at
n −1
zero can now be achieved by choosing the control law u of
d ~ (7.1) such that outside of S (t )
s ( x; t ) = + λ x , λ is positive constant (7.3)
dt
1 d 2
For example, n = 2 → s = ~
x& + λ ~
x , n = 3 → s = &~
x& + 2 λ ~
x& + λ2 ~
x. 2 dt
s ≤ −η s (7.5)
⊗ Given initial condition (7.2), the problem of tracking where η is a strictly positive constant. Essentially, (7.5) states
x ≡ x d is equivalent to that of remaining on the surfaces
that the squared “distance” to the surface, as measured by s 2 ,
S (t ) for all t > 0 ; indeed s ≡ 0 represents a linear differential decrease along system trajectories. Thus it constrains
equation whose unit solution is ~ x ≡ 0 , given initial condition trajectories to point towards the surface S (t ) , as illustrated in
(7.2). ⇒ The problem of tracking the n-dimensional vector Fig. 7.2.
x d can be reduced to that of keeping the scalar quantity s at
zero.
S (t )
Bounds on s can be directly translated into bounds on the
tracking error vector ~x , and therefore the scalar s represents
a true measure of tracking performance. Assume that
x~ ( 0) = 0 , we have
∀t ≥ 0, s (t ) ≤ Φ ⇒ ∀t ≥ 0, ~
x (i ) t ≤ (2 λ ) i ε, i = 0, L , n − 1
(7.4)
n −1
where ε = Φ/ λ . Fig. 7.2 The sliding condition
___________________________________________________________________________________________________________
Chapter 7 Sliding Control 30
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Condition (7.5) is called sliding condition. S (t ) verifying (7.5) Geometrically, the equivalent control can be constructed as
is referred to as sliding surface. The system‘s behavior once
on the surface is called sliding regime or sliding mode. The u eq = α u + + (1 − α )u −
other interesting aspect of the invariant set S (t ) is that once on
it, the system trajectories are defined by the equation of the set i.e., as a convex combination of the values of u on both side of
itself, namely (d / dt + λ )n −1 ~
x = 0 . Satisfying (7.5) guarantees the surface S (t ) . The value of α can be obtained formally
that if condition (7.2) is not exactly verified, from (7.6), which corresponds to requiring that the system
i.e., x (0) ≠ x d (0) , the surface S (t ) will be reach in a finite trajectories be tangent to the surface. This intuitive
construction is summarized in Fig. 7.5
time smaller that s (t = 0) / η .
s<0
s=0
The typical system behavior implied by satisfying sliding f+
condition (7.5) is illustrated in Fig. 7.3 for n = 2 . feq
x&
f- s>0
sliding mode
exponential convergence
finite -time xd (t)
reaching phase Fig. 7.5 Filippov’s construction of the equivalent
x
slope - λ dynamics in sliding mode
s=0
7.1.3 Perfect Performance – At a Price
Fig. 7.3 Graphical interpretation of Eqs. (7.3) and (7.5),n=2
A Basic Example:
When the switching control is imperfect, there is chattering as Consider the second-order system
shown in Fig. 7.4
&x& = −a(t ) x& 2 cos 3x + u (7.10)
x&
where,
sliding mode u : control input
exponential convergence y=x : scalar output of interest
finite -time xd (t)
reaching phase f = − a (t ) x& 2 cos 3 x : unknown bounded nonlinear function
x
slope - λ with 1 ≤ a ≤ 2 . Let fˆ be an estimation value of f , assume
s=0 that the estimation error is bounded by some known function
F = F ( x, x& ) as follows
Fig. 7.4 Chattering as a result of imperfect control switchings.
fˆ − f ≤ F (7.9)
7.1.2 Filippov’s Construction of the Equivalent Dynamics
The dynamics while in sliding mode can be written as assume that fˆ = −1.5 x& 2 cos 3x ⇒ F = 0.5 x& 2 cos 3 x . In
order to have the system track x(t ) = x d (t ) , we define a
s& = 0 (7.6) sliding surface s = 0 according to (7.3), namely
By solving (7.6), we obtain an expression for u called the
d
equivalent control, u eq , which can be interpreted as the s = + λ ~
x=~
x& + λ ~
x (7.11)
dt
continuous control law that would maintain s& = 0 if the
dynamics were exactly known. Fro instance, for a system of We then have
the form &x& = f + u , we have
s& = ~
&x& + λ ~
x& = ( &x& − &x&d ) + λ ~
x& = f + u − &x&d + λ ~
x& (7.12)
u eq → u = − f + &x& = − f + ( &x&d + ~
&x&)
To achieve s& = 0 , we choose control law as u = − f + &x&d − λ ~
x& .
From (7.6)
Because f is unknown and replaced by its estimation fˆ , the
control is chosen as
s& = ~ x = 0 ⇒ &~
x& + λ ~ x& = −λ ~
x&
u → uˆ = − fˆ + &x&d − λ ~
x& (7.13)
Hence,
û can be seen as our best estimate of the equivalent control. In
u eq = − f + &x&d − λ ~
x& (7.7) order to stratify the sliding condition (7.5), despite the
uncertainty on the dynamics f , we add to û a term
And the system dynamics while in sliding mode is, of course, discontinuous across the surface s = 0
&x& = f + u eq = &x&d − λ ~
x& (7.8) u = uˆ − k sgn( s) (7.14)
___________________________________________________________________________________________________________
Chapter 7 Sliding Control 31
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
sgn = +1 if s>0 where, β = bmax / bmin . Since the control law will be
where “sgn” is the sign function:
sgn = −1 if s<0 designed to be robust to the bounded multiplicative
uncertainty (7.18), we shall call β the gain margin of our
By choosing k = k ( x, x& ) in (7.14) to be large enough, we can design. With s and û defined as before, one can then easily
now guarantee that (7.5) is verified. Indeed, from (7.12) and show that the control law
1 d 2
(7.14), s = s& s = [ f − fˆ − k sgn( s )] s = ( f − fˆ ) s − k s .
2 dt u = bˆ −1 [uˆ − k sgn( s )] (7.19)
So that, letting
with
k = F +η (7.15)
k ≥ β ( F + η ) + ( β − 1) uˆ (7.20)
1 d 2
we get from (7.9): s ≤ −η s as desired.
2 dt satisfies the sliding condition. Indeed, using (7.19) in the
⊗ Note that: expression of s& leads to
- from (7.15), the control discontinuity k across the
s& = ( f − b bˆ −1 fˆ ) + (1 − b bˆ −1 )(− &x&d + λ ~
x& ) − b bˆ −1k sgn( s )
surface s = 0 increases with the extent of parametric
uncertainty.
Condition (7.5) can be rewritten as s& s ≤ −η s = −η s sgn(s ) .
- fˆ and F need not depend only on x or x& . They may
Hence we have
more generally be functions of any measured variables
external to system (7.8), and may also depend explicitly
on time. (( f − bbˆ −1
fˆ ) + (1 − b bˆ −1 )(− &x&d + λ ~
x& ) − b bˆ −1k sgn( s ) s )
- To the first order system, the sliding mode can be ≤ −η s sgn( s )
interpreted that “if the error is negative, push hard or
enough in the positive direction, and conversely”. It does
not for higher-order systems. (
b bˆ −1k s sgn( s) ≥ ( f − b bˆ −1 fˆ ) + (1 − b bˆ −1 )(− &x&d + λ ~
x& ) s )
+ η s sgn( s)
Integral Control
A similar result would be obtained by using integral control, (
⇒ k ≥ (bˆb −1 f − fˆ ) + (bˆb −1 − 1)(− &x&d + λ ~
x& ) sgn( s) + bˆb −1η )
t
i.e., formally letting
∫ ~x (r )dr be the variable of interest. The
0
so that k must verify
&x& = f + b u where
(7.16)
x : position of vehicle
where the (possibly time-varying or state-dependent) control u : control input (force provided by a propeller)
gain b is unknown but of known bounds m : mass of the vehicle
c : drag coefficient
0 < bmin ≤ b ≤ bmax (7.16) In practice, m and c are not known accurately, because they
only describe loosely the complex hydrodynamic effects that
Since the control input enters multiplicatively in the dynamics, govern the vehicle’s motion. From (7.3),
it is natural to choose our estimate b̂ of gain b as the
s=~
x& + λ ~
x
geometric mean of the above bounds bˆ = bmin bmax . Bound
⇒ s& = x + λ ~
~
&& x& = ( &x& − &x&d ) + λ ~
x&
(7.17) can then be written in the form
⇒ m s& = m &x& − m &x&d + m λ ~ x& = −c x& x& + u − m &x&d + m λ ~
x&
b̂
β −1 ≤ ≤β (7.18)
b The estimated controller is chosen as
___________________________________________________________________________________________________________
Chapter 7 Sliding Control 32
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
uˆ = mˆ ( &x&d − λ ~
x& ) + cˆ x& x& u
boundary
and a control law satisfying the sliding condition can be layer
derived as
û
u = uˆ − k sgn( s ) = mˆ ( &x&d − λ ~
x& ) + cˆ x& x& − k sgn( s ) (7.22)
−φ φ s
where k is calculated from (7.20)
k ≥ β ( F + η ) + ( β − 1) uˆ
≥ β ( F + η ) + ( β − 1) mˆ ( &x&d − λ ~
x& ) + cˆ x& x&
Fig. 7.6.b Control interpolation in the boundary layer
Hence k can be chosen as follows
Given the results of section 7.1.1, this leads to tracking to
k = ( F + β η ) + mˆ ( β − 1) ( &x&d − λ ~
x& ) (7.23) within a guaranteed precision ε , and more generally
guarantees that for all trajectories starting inside B(t = 0)
Note that the expression (7.23) is “tighter” than the general
form (7.20), reflecting the simpler structure of parametric ∀t ≥ 0, ~
x i (t ) ≤ (2 λ ) i ε i = 0, K , n − 1
uncertainty: intuitively, u can compensate for c x& x& directly,
regardless of uncertainty on m . In general, for a given Example 7.2________________________________________
problem, it is a good idea to quickly rederive a control law
Consider again the system (7.10): &x& = − a (t ) x& 2 cos 3 x + u ,
satisfying the sliding condition, rather than apply some pre-
packed formula. and assume that the desired trajectory is x d = sin(π t / 2) . The
__________________________________________________________________________________________
constants are chosen as λ = 20, η = 0.1 , sampling time
7.1.4 Direct Implementations of Switching Control Laws dt = 0.001 sec.
Switching control law:
The main direct applications of the above switching controller
include the control of electric motors, and the use of artificial u = uˆ − k sgn( s)
dither to reduce stiction effects.
= 1.5 x& 2 cos 3 x + &x&d − 20 ~
x&
- Switching control in place of pulse-width modulation
- Switching control with linear observer − (0.5 x& 2 cos 3 x + 0.1) sgn( ~
x& + 20 ~
x)
- Switching control in place of dither Smooth control law with a thin boundary layer φ = 0.1 :
0.5
Control input
the case n = 2 .
-0.5
-1
-2
-1.0
-3
-4 -1.5
bo ε 6 5
un 5 4
da
ry
3
Tracking Error (x10-3)
lay 3 2
Control Input
er x 2
1
1
0
0 -1
-1 -2
ε -2
-3
-3
-4
-4 -5
0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4 0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4
Fig. 7.6.a The boundary layer Fig. 7.8 Smooth control input and tracking performance
__________________________________________________________________________________________
b = bˆ = 1 . In order to maintain attractiveness of the boundary Conceptually, the perturbations are filtered according to (7.29)
layer now that φ is allowed to vary with time, we must to give s , which in turn provides tracking error ~ x by further
actually modify condition (7.5). Indeed, we now need to low-pass filtering, according to definition (7.3)
guarantee that the distance to the boundary layer always
decreases. ~
− ∆ f (x d ) + O (ε ) 1storder filter s 1 x
d
s≥φ ⇒ ( s − φ) ≤ −η (7.29) ( p + λ ) n −1
dt
d choice of φ definition of s
s ≤ −φ ⇒ ( s − φ) ≥ η
dt
Fig. 7.9 Structure of the closed-loop error dynamics
Thus, instead of simply required that (7.5) be satisfy outside
the boundary layer, we now required that
Control action is a function of x and x d . Since λ is break-
1 d 2 frequency of filter (7.3), it must be chosen to be “small” with
s ≥φ ⇒ s ≤ (φ& - η ) s (7.26) respect to high-frequency un-modeled dynamics (such as un-
2 dt modeled structural modes or neglected time-delays).
Furthermore, we can now turn the boundary layer thickness φ
The additional term φ& s in (7.26) reflects the fact that the so that (7.29) also presents a first-order filter of bandwidth λ .
boundary layer attraction condition is more stringent during It suffices to let
boundary layer contraction ( φ& < 0 ) and less stringent during
k (x d )
boundary layer expansion ( φ& > 0 ).In order to satisfy (7.26), =λ (7.30)
φ
the quantity −φ& is added to control discontinuity gain k (x) ,
i.e., in our smooth implementation the term which can be written from (7.27) as
k (x) sgn( s ) obtained from switched control law u is actually
φ& + λ φ = k (x d ) (7.31)
replaced by k (x) sat( s / φ) , where
sat( y ) = y if y ≤ 1 k ( x) = k ( x) − k ( x d ) + λ φ (7.32)
and sat is the saturation function
sat( y ) = sgn( y ) otherwise
and can be seen graphically as in the following figure ⊗ Note that:
- The s-trajectory is a compact descriptor of the closed-
loop behavior: control activity directly depends on s ,
sat( y )
while tracking error ~x is merely a filtered version of s
- The s-trajectory represents a time-varying measure of the
y validity of the assumptions on model uncertainty.
−1 1
- The boundary layer thickness φ describes the evolution
of dynamics model uncertainty with time. It is thus
particularly informative to plot s (t ) , φ(t ) , and −φ(t ) on
a single diagram as illustrated in Fig. 7.11b.
___________________________________________________________________________________________________________
Chapter 7 Sliding Control 34
Applied Nonlinear Control Nguyen Tan Tien - 2002.5
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
2 1
1 0
d
0 -1
6
φ
S-trajectories (x10-2)
2
k ( x) = max cˆ − c x& 2 + max mˆ − m &x& − λ ~
x& + max(m)η
0 s
d
-2
φ& = k ( x ) − λ φ
-4
−φ d
-6
-8 k = k ( x) − φ&
0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4
Time (s)
2
20
Control Input
1
of the boundary layer allow us to make better use of the 0
15
10
available bandwidth.
-1
acceleration (m/s2 )
5
-2
velocity (m/s)
__________________________________________________________________________________________ -3
distance (m)
0
-4 -5
In the case that β ≠ 1 , one can easily show that (7.31) and
-5 -10
0 0.5 1.0 1.5 2.0 2.5 4 0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4
Time (s) Time (s)
λφ 10
φ
k (x d ) ≥ ⇒ φ& + λ φ = β d k (x d )
1.0
(7.33)
Tracking Error (x10-2)
S-trajectories (x10-2)
βd 0
0.5
s
-5 0
λφ λφ k (x d ) -10
k (x d ) ≤ ⇒ φ& + =
-0.5
(7.34) -15
βd β d2 βd -20
-1.0
−φ
-25 -1.5
0 1 2 3 4 5 6 0 1 2 3 4 5 6
constant and large, so that the term k sat( s / φ) simply where T A is the largest un-modeled time-delay (for
equals λ s / β in the boundary layer. instance in the actuators).
- A well-designed controller should be capable of
iii. sampling rate: with a full-period processing delay, one
gracefully handling exceptional disturbances, i.e.,
gets a condition of the form
disturbances of intensity higher than the predicted
bounds which are used in the derivation of the control 1
λ ≤ λS ≈ ν sampling (7.43)
law. 5
- In the case that λ is time-varying, the term
where, ν sampling is the sampling rate.
u ′ = −λ& ~
x should be added to the corresponding û , while
the augmenting gain k (x) according by the quantity The desired control bandwidth λ is the minimum of three
u ′ ( β − 1) . It will be discussed in next section. bounds (7.41-43). Ideally, the most effective design
corresponds to matching these limitations, i.e., having
7.3 The Modeling/Performance Trade-Offs
λR ≈ λ A ≈ λS ≈ λ (7.44)
The balance conditions (7.33)-(7.36) have practical
implications in term of design/modeling/performance trade- 7.4 Multi-Input System
offs. Neglecting time-constants of order 1 / λ , condition (7.33)
and (7.34) can be written Consider a nonlinear multi-input system of the form
m
λn ε ≈ β d k d (7.39) x i ( ni ) = f i (x) + ∑bj =1
ij ( x) u j , i = 1, L , m , j = 1, L , m
∆ε ≈ ∆ ( β d k d / λ n ) (7.40)
1
λ ≤ λA ≈ (7.42)
3TA
___________________________________________________________________________________________________________
Chapter 7 Sliding Control 36
Applied Nonlinear Control Nguyen Tan Tien - 2002.6
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
8. Adaptive Control
8.1.2 What is Adaptive Control ? Example 8.1 MRAC control of unknown mass____________
An adaptive controller differs from an ordinary controller in
Consider the control of a mass on a frictionless surface by a
that the controller parameters are variable, and there is a
mechanism for adjusting these parameters on-line based on motor force u , with the plant dynamics being
signals in the system. There are two main approaches for
m &x& = u (8.1)
constructing adaptive controllers: so-called model-reference
adaptive control method and so-called self-tuning method.
Choose the following model reference
Model-Reference Adaptive Control (MRAC) &x&m + λ1 x& m + λ2 x m = λ2 r (t ) (8.2)
ym
reference model where,
λ1 , λ1 : positive constants chosen to reflect the performance
r
u y e specifications
controller plant xm : the reference model output (ideal out put of the
controlled system)
r (t ) : reference position
â
adaptation law
* m is known exactly, we can choose the following control
Fig. 8.3 A model-reference adaptive control system law to achieve perfect tracking &~
x& + 2λ ~
x& + λ2 ~
x = 0 , with
~
x = x − xm representing the tracking error and λ is a strictly
A MRAC can be schematically represented by Fig. 8.3. It is
composed of four parts: a plant containing unknown positive number. This control law leads to the exponentially
parameters, a reference model for compactly specifying the convergent tracking error dynamics: u = mˆ ( &x&m − 2λ ~
x& − λ2 ~
x).
desired output of the control system, a feedback control law
containing adjustable parameters, and an adaptation * m is not known exactly, we may use the control law
mechanism for updating the adjustable parameters.
u = mˆ ( &x&m − 2λ ~
x& − λ2 ~
x) (8.3)
The plant is assumed to have a known structure, although the
parameters are unknown. which contains the adjustable parameter m̂ . Substitution this
- For linear plants, the numbers of poles and zeros are control law into the plant dynamics, yields
assumed to be known, but their locations are not.
- For nonlinear plants, this implies that the structure of the m &x& = mˆ ( &x&m − 2λ ~x& − λ2 ~
x)
dynamic equations is known, but that some parameters are = (m + m ) ( &x&m − 2λ x& − λ2 ~
~ ~ x ), with m~ ≡ mˆ − m
not.
⇒ m &~ x& + 2m λ ~ x& + m λ2 ~x =m~ ( &x& − 2λ ~
x& − λ2 ~
x) m
A reference model is used to specify the ideal response of the
adaptive control system to external command. The choice of ⇒ m( ~
&x& + λ ~
x& ) + λ m ( ~
x& + λ ~ ~ ( &x& − 2λ ~
x)=m m x& − λ2 ~
x)
the reference model has to satisfy two requirements:
- It should reflect the performance specification in the control Let the combined tracking error measure be
tasks such as rise time, settling time, overshoot or frequency
s =~
x& + λ ~
x (8.5)
domain characteristics.
- This ideal behavior should be achievable for the adaptive
control system, i.e., there are some inherence constrains on and the signal quantity v is defined as v = &x&m − 2λ ~
x& − λ2 ~
x .
the structure of reference model given the assumed structure The closed-loop zero dynamics
of the plant model.
~v
m s& + λ m s = m (8.4)
The controller is usually parameterized by a number of
adjustable parameters. The controller should have perfect Consider Lyapunov function
___________________________________________________________________________________________________________
Chapter 8 Adaptive Control 37
Applied Nonlinear Control Nguyen Tan Tien - 2002.6
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
0.5
2
2.0
Tracking Performance
Parameter Estimation
0.4 0
1.5
0.3
0.2
1.0 with w = &x& . If actually, the unknown parameter m is slowly
time-varying, the above estimate has to be recalculated at
0.1
0.5
0.0
-0.1
0.0 0.5 1.0 1.5 2.0 2.5 3.0
0.0
1.0 1.5 2.0 3.0
every new time instant. To increase computational efficiency,
0.0 0.5 2.5
Time (s) Time (s) it is desirable to adopt a recursive formulation instead of
Fig. 8.4 Tracking performance and parameter estimation for repeatedly using (8.11). To do this, we define
an unknown mass with reference path r (t ) = 0
0.8 2.5 1
0.6 P (t ) ≡ t
(8.12)
∫w
2.0
Tracking Performance
Parameter Estimation
0.4
2
0.2 1.5
dr
0.0 0
-0.2 1.0
-0.8
0.0 0.5 1.0 1.5 2.0 2.5 3.0
0.0
0.0 0.5 1.0 1.5 2.0 2.5 3.0 be directly obtained by using
Time (s) Time (s)
( )
Fig. 8.5 Tracking performance and parameter estimation for d −1
an unknown mass with reference path r (t ) = sin( 4t ) P = w2 (8.13)
__________________________________________________________________________________________
dt
mˆ& = − P (t ) w e (8.14)
Lemma 8.1: Consider two signals e and φ related by the Now we choose the adaptation laws for â r and â y . Let the
following dynamic equation tracking error be e = y − y m and the error of parameter
estimation be
e(t ) = H ( p ) [ k φT (t ) v (t )] (8.15)
a~r = aˆ r − a r a~ y = aˆ y − a y (8.25)
where e(t ) is a scalar output signal, H ( p) is strictly positive
real transfer function, k is an unknown constant with know The dynamics of tracking error can be found by subtracting
sign, φ(t ) is m × 1 vector function of time, and v (t ) is a (8.23) and (8.21)
measurable m × 1 vector. If the vector φ(t ) varies according to
e& = − a m ( y − y m ) + (a m − a p + b p aˆ y ) y + (b p aˆ r − bm )r
φ& (t ) = − sgn(k ) γ e v (t ) (8.16) = −a m e + b p (a~r r − a~ y y ) (8.26)
with γ being a positive constant, then e(t ) and φ(t ) are The Lemma 8.1 suggests the following adaptation laws
globally bounded. Furthermore, if v (t ) is bounded, then
aˆ& r = − sgn(b p ) γ e r (8.27)
e(t ) → 0 as t → ∞ .
a&ˆ y = − sgn(b p ) γ e y (8.28)
8.2 Adaptive Control of First-Order Systems
Let us discuss the adaptive control of first-order plants using with γ being a positive constant representing the adaptation
MRAC method. Consider the first-order differential equation gain. The sgn(b p ) in (8.27-28) determines the direction of the
y& = − a p y + b p u (8.20) search for the proper controller parameters.
___________________________________________________________________________________________________________
Chapter 8 Adaptive Control 39
Applied Nonlinear Control Nguyen Tan Tien - 2002.6
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Tracking Performance
Parameter Estimation
Parameter Estimation
1.0 1.0
3.5
4
0.5 âr 3.0 0.5 â y
3 0.0 2.5 0.0
Fig. 8.9 Tracking performance and parameter estimation with Fig. 8.11 Tracking performance and parameter estimation
reference path r (t ) = 4 with reference path r (t ) = 4
5 2.0 5 2.0
4 1.5 4 1.5 âr
3 3
Tracking Performance
Tracking Performance
Parameter Estimation
Parameter Estimation
1.0 1.0
2 2
0.5 âr 0.5
1 1
â f
0 0.0 0 0.0
-1 -0.5 -1 -0.5
-2
-1.0
â y -2
-1.0
-3 -3 â y
-1.5 -1.5
-4 -4
-5 -2.0 -5 -2.0
0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10 0 1 2 3 4 5 6 7 8 9 10
Time (s) Time (s) Time (s) Time (s)
Fig. 8.10 Tracking performance and parameter estimation Fig. 8.12 Tracking performance and parameter estimation
with reference path r (t ) = 4 sin(3t ) with reference path r (t ) = 4 sin(3t )
__________________________________________________________________________________________ __________________________________________________________________________________________
Parameter convergence analysis ⇒ refer text book 8.3 Adaptive Control of Linear Systems with Full States
Feedback
Extension to nonlinear plant
The same method of adaptive control design can be used for Consider the nth-order linear system in the canonical form
the non-linear first-order plant describe by the differential
equation a n y ( n ) + a n−1 y ( n−1) + K + a0 y = u (8.36)
y& = − a p y − c p f ( y ) + b p u (8.32) where the state components y, y& ,K, y ( n−1) are measurable,
coefficient vector a = [a n L a1 a0 ]T is unknown, but
where f is any known nonlinear function. The nonlinear in
their signs are known. The objective of the control system is to
these dynamics is characterized by its linear parametrization in make y closely track the response of a stable reference model
terms of the unknown constant c . Instead of using (8.22), now
we use the control law α n y m ( n) + α n−1 y m ( n−1) + K + α 0 y m = r (t ) (8.37)
where the second term in (8.33) is introduced with the Choice of control law
intention of adaptively canceling the nonlinear term. Using the Define a signal z (t ) as follows
same procedure for the linear plant,
(n)
z (t ) = y m − β n−1e ( n−1) − K − β 0 e (8.38)
y& = − a p y − c p f ( y ) + b p [aˆ y y + aˆ f f ( y ) + aˆ r r ]
= −(a p − b p aˆ y ) y − (c p − b p aˆ f ) f ( y ) + b p aˆ r r with β1 ,K, β n being positive constants chosen such that
p n + β n−1 p n−1 + K + β 0 is a stable (Hurwitz) polynomial.
Comparing to (8.21) and define a f ≡ c p / bp and Adding both side of (8.36) and rearranging, we can rewrite the
a~ f ≡ aˆ f − a f . The adaptation laws are plant dynamics as
a n [ y ( n) − z ] = u − a n z − a n−1 y ( n−1) − K − a0 y
a&ˆ y = − sgn(b p ) γ e y (8.34a)
Let us choose the control law to be
a&ˆ f = − sgn(b p ) γ e f (8.34b)
Choice of adaptation law theory that the relative degree of the reference model has to be
Rewrite the closed-loop system (3.40) in state space form larger or equal to that the plant in order to allow the possibility
of perfect tracking. Therefore, in our treatment, we will
x& = A x + b [(1 / a n ) v T ~
a] (3.41a) assume that nm − mm ≥ n − m .
e = cx (3.41b)
where The objective of the design is to determine a control law, and
0 1 0 L 0 0 1 an associated adaptation law, so that the plant
output y asymptotically approaches y m . We assume as follows
0 0 1 L 0 0 0 - the plant order n is known
A= M M M O M , b = M , cT = M
- the relative degree n − m is known
0 0 0 L 1 0 0 - the sign of k p is known
− β − β n−1 1 0
0 − β1 − β 2 L - the plant is minimum phase
Consider Lyapunov function candidate 8.4.1 Linear systems with relative degree one
Choice of the control law
V ( x, ~
a ) = xT P x + ~
a T Γ −1~
a To determine the appropriate control law for the adaptive
controller, we must first know what control law can achieve
perfect tracking when the plant parameters are perfect known.
where both Γ and P are symmetric positive constant matrix, Many controller structures can be used for this purpose. The
and P satisfies following one is particularly convenient for later adaptation
design.
PA + A T P = −Q Q = QT > 0
Example 8.5 A controller for perfect tracking_____________
for a chosen Q . The derivative V& can be computed easily as Consider the plant described by
V& = − xT Q x + 2~
a T v bT P x + 2~
a Γ −1~
a& k p ( p + bp )
y= 2
u (8.45)
p + a p1 p + a p2
Therefore, the adaptation law
and the reference model
aˆ& = − Γ v bT P x (8.42)
k m ( p + bm )
ym = r (8.46)
leads to V& = − x T Q x ≤ 0 . p 2 + a m1 p + a m2
ym (t )
8.4 Adaptive Control of Linear Systems with Output r (t ) Wm ( p )
Feedback u0 u1 e
u
Consider the linear time-invariant system presented bu the k W p ( p)
transfer function α1
p + bm
Z p ( p) b0 + b1 p + K + bm−1 p m−1 + p m
W ( p) = k p = kp (8.43) β1 p + β 2
R p ( p) a0 + a1 p + K + a n−1 p n−1 + p n p + bm
where k p is called the high-frequency gain. The reason for Fig. 8.13 Model-reference control system for relative degree 1
this term is that the plant frequency response at high frequency
Let the controller be chosen as shown in Fig. 8.13, with the
kp
verifies W ( jω ) = n−m , i.e., the high frequency response is control law being
ω
essentially determined by k p . The relative degree r of this β1 p + β 2
u = α1 z + y+kr (8.47)
system is r = n − m . In our adaptive control problem, the p + bm
coefficients ai , b j (i = 0,1,K, n − 1; j = 0,1,K, m − 1) and the
high frequency gain k p are all assumed to be unknown. where z = u /( p + bm ) , i.e., z is the output of a first-order
filter with input u , and α1 , β1 , β 2 , k are controller parameters.
The desired performance is assumed to be described by a If we take these parameters to be α1 = b p − bm ,
reference model with transfer function
a m1 − a p1 a m2 − a p2 km
β1 = , β2 = , k= , the transfer
Zm kp kp kp
Wm ( p ) = k m (8.44)
Rm function from the reference input r to the plant output is
Therefore, perfect tracking is achieved with this control law, - The vector θ1* contains (n − 1) parameters which intend to
i.e., y (t ) = y m (t ), ∀t ≥ 0 .
cancel the zeros of plant.
Why the closed-loop transfer function can become exactly the - The vector θ 2* contains (n − 1) parameters which, together
same as that of the reference model ? To know this, note that with the scalar gain θ 0* can move the poles of the closed-loop
the control input in (8.47) is composed of three parts:
- The first part in effect replaces the plant zero by the control system to the locations of the reference model poles.
reference model zero, since the transfer function from u1 to
As before, the control input in this system is a linear
y is combination of:
- the reference signal r (t )
p + bm k p ( p + b p ) k p ( p + bm )
Wu1, y = = 2 - the vector signal ω1 obtained by filtering the control input u
p + bp p2 + a p p + a p p + a p1 p + a p2
1 2 - the vector signal ω 2 obtained by filtering the plant output y
- The second part places the closed-loop poles at locations of and the output itself.
those of reference model. This is seen by noting that the The control input can be rewritten in terms of the adjustable
transfer function from u 0 to y is parameters and the various signals, as
u0 u1
r (t ) u y Let the controller be chosen as shown in Fig. 8.16. Noting that
* W p ( p)
k
bm in the filter in Fig. 8.13 has been replaced by a positive
ω1
θ1* Λ, h number λ . The closed-loop transfer function from the
T
φ ω ω2 reference signal r to the plant output y is
k* θ 2* Λ, h
p + λ0 kp
θ 0* p + λ0 + α1 p 2 + a p p + a p
Fig. 8.15 An equivalent control system for time-varying gains W ry = k 1 2
p + λ0 β1 p + β 2 kp
1+
Since y m (t ) = Wm ( p ) r , the tracking error is seen to be related p + λ0 + α 1 p + λ0 p 2 + a p p + a p
1 2
km k k p Z p λ1 ( p ) Z m ( p )
ym = r Wry = (8.58)
2
p + a m1 p + a m2 R p ( p )[λ ( p ) + C ( p )] + k p Z p D( p )
ym (t )
Wm ( p ) The question now is whether in this general case, there exists
r (t )
u0 u1 choice of values for k , θ1 , θ 2 and θ 0 such that the above
e
u transfer function becomes exactly the same as Wm ( p ) , or
k W p ( p)
equivalently
α1
p + λ0 R p (λ ( p ) + C ( p )) + k p Z p D( p ) = λ1Z p Rm ( p ) (8.59)
β1 p + β 2
p +λ 0 The answer to this question can be obtained from the
Fig. 8.16 Model-reference control system for relative degree 2 following lemma
___________________________________________________________________________________________________________
Chapter 8 Adaptive Control 43
Applied Nonlinear Control Nguyen Tan Tien - 2002.6
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
u = m &x&r − α s + c f1 + k f 2 ( → 8.71)
where k is constant of the same sign as h , and y r(n) is the
which yields m &x& + c f1 + k f 2 = m &x&r − α s + c f1 + k f 2 or
derivative of y r( n−1) , i.e., y r( n) = y d( n) − λn−2 e ( n−1) − K − λ0 e& .
m ( &x& − &x&r ) + α s = 0 . Because the unknown parameters, the
Noting that y r(n) , the so-called “reference” value of y (n) , is controller is
obtained by modifying y d(n ) according to the tracking errors. If
u → uˆ = mˆ &x&r − α s + cˆ f1 + kˆ f 2 ( → 8.72)
the parameters are all known, this choice leads to the tracking
error dynamics h s& + k s = 0 and therefore gives exponential which leads to the tracking error
convergence of s , which in turn, guarantees the convergence
of e . m &x& + c f1 + k f 2 = mˆ &x&r − α s + cˆ f1 + kˆ f 2
Choice of adaptation law (m &x& − m &x&r ) − (mˆ &x&r − m &x&r ) − (cˆ − c) f1 − (kˆ − k ) f 2 + α s = 0
For our adaptive control, the control law (8.71) is replaced by ~ &x& − c~ f − k~ f + α s = 0
m ( &x& − &x& ) − mr r 1 2
n
&x&r
u = hˆ y r( n) − k s + ∑ aˆ f (x, t )
i =1
i i (8.72) ~ c~ k~ f
m s& + α s = m 1 [ ] ( → 8.74)
where h, ai have been replaced by their estimated values. The f 2
tracking error yields &
Adaptation laws: m&ˆ = − γ s &x&r , c&ˆ = − γ s f1 , kˆ = − γ s f 2
n
∑ a~ f (x, t )
~
h s& + k s = h y r( n) + i i (8.73) ~ 2 + c~ 2 + k~ 2 ) and its
Lyapunov function: V = m s 2 + γ −1 ( m
i =1
~ derivative with the above adaptation laws yields
where, hi = hˆi − hi , a~i = aˆ i − ai . (8.73) can be written in the
form
V& = 2m s s& + 2γ −1 ( m~ m&ˆ + c~ c&ˆ + k~ k&ˆ)
= −2α s 2 + 2γ −1 m
2
~ (mˆ& + γ s &x& ) + c~ (cˆ& + γ s f ) + k~ (kˆ& + γ s f )
i =1 r 1 2
y r( n) (8.74)
= −2α s 2 ≤ 0
=
1/ h ~ ~
h a1 L a n
p + ( k / h)
[
~ f1 (x, t )
M
] __________________________________________________________________________________________
qualitative answers can improve our understanding of adaptive of non-parametric uncertainties present in the above example,
control system behavior in practical applications. the observed instability can seem quite surprising.
_________________________________________________________________________________________
may lead to slower estimation, we may multiply both side of
(8.79) by a constant number, i.e., λ f .
In view of the global tracking convergence proven in the
absence of non-parametric uncertainties and the small amount
___________________________________________________________________________________________________________
Chapter 8 Adaptive Control 46
Applied Nonlinear Control Nguyen Tan Tien - 2002.6
_________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________________
Generally, for a linear SISO system, its dynamics can be H 12 = H 21 = m 2 l1c 2 cos q 2 + m 2 l c22 + I 2
described by
H 22 = m 2 l c22 + I 2
A( p ) y = B ( p ) u (8.80) h = m 2 l1l c 2 sin q 2
where g1 = m1l c1 g cos q1 + m 2 g[l c 2 cos(q1 + q 2 ) + l1 cos q1 ]
A( p ) = a0 + a1 p + K + a n−1 p n−1 + p n g 2 = m 2 l c 2 g cos(q1 + q 2 )
B ( p ) = b0 + b1 p + K + a n−1 p n−1
Let us define a1 = m2 , a 2 = m2lc 2 , a3 = I1 + m1lc21 , and
Divide both sides of (8.80) by a known monic polynomial of
order n, leading to a4 = I 2 + m2 lc22 .
Then each term on the left-hand side of (6.9)
is linear terms of the equivalent inertia parameters
y=
A0 ( p ) − A( p) B( p )
u (8.81) a = [a1 a 2 a3 a 4 ]T . Specially,
A0 ( p ) A0 ( p )
where H11 = a3 + a 4 + a1l12 + 2a 2 l1 cos q 2
n −1 n
A0 = α 0 + α1 p + K + α n−1 p +p H 22 = a 4
H12 = H 21 = a 2 l1 cos q 2 + a 4
has known coefficients. In view of the fact that
Thus we can write
A0 ( p ) − A( p ) = (α 0 − a0 ) + (α1 − a1 ) p + K + (α n−1 − a n−1 ) p n−1
τ = Y1 (q, q& , q
&&) a (8.83)
(8.81) can be rewritten in the form
This linear parametrization property actually applies to any
y = θT w (t ) (8.82) mechanical system, including multiple-link robots.
t t d
∫ ∫ dr [wH]q& dr
I2, m2 t
w(t − r )Hq
&& dr = w(t − r )Hq& −
lc2 0 0 0
l2
l1 q 2, τ 2 = w(0)H (q)q& − w(0)H[q (0)] q& (0) −
t
Fig. 6.2 A two-link robot where y is the filtered torque and W is the filtered version of
Y1 . Thus the matrix W can be computed from available
Consider the nonlinear dynamics of a two-link robot measurements of q and q& . The filtered torque y can also be
computed because the torque signals issued by the computer
H 11 H 12 q&&1 − h q& 2 − h q&1 − h q& 2 q&1 g1 τ 1 are known.
= & + = _________________________________________________________________________________________
H 21 H 22 q&&2 h q&1 0 q 2 g 2 τ 2
(6.9) 8.7.2 Prediction-error-based estimation model
where,
8.7.3 The gradient estimator
q = [q1 q 2 ]T : joint angles
τ = [τ 1 τ 2 ]T : joint inputs (torques) 8.7.4 The standard least-squares estimator
___________________________________________________________________________________________________________
Chapter 8 Adaptive Control 48