Professional Documents
Culture Documents
net/publication/335219538
CITATIONS READS
2 151
3 authors, including:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Ahmed Kamel on 17 August 2019.
of time-dependency of the Hamilton- Jacobi-Bellman (HJB) III. N ONLINEAR F INITE -H ORIZON R EGULATOR V IA
differential equation. In finite-horizon optimal nonlinear con- SDRE
trol problem, the DRE can not be solved in real time by Finite-horizon optimal control of nonlinear systems is a
backward integration from t f to t0 because we do not know the challenging problem in the control field due to the complexity
value of the states ahead of present time step. To overcome the of time-dependency of the Hamilton- Jacobi-Bellman (HJB)
problem, an approximate analytical approach is used to convert differential equation. In finite-horizon optimal nonlinear con-
the original nonlinear Ricatti equation to a linear differential trol problem, the DRE can not be solved in real time by
Lyapunov equation that can be solved in closed format each backward integration from t f to t0 because we do not know the
time step. value of the states ahead of present time step. To overcome the
In this section, the relation between the proposed method problem, an approximate analytical approach is used [9], [10],
and the exact optimal solution to the problem will be dis- [11], [12] to convert the original nonlinear Ricatti equation to
cussed. a linear differential Lyapunov equation that can be solved in
Given the nonlinear system in the form: closed format each time step.
ẋ(t) = f(x) + g(x)u(t). (1)
A. Problem Formulation
The nonlinear system can be expressed in a state-dependent The nonlinear system considered in this paper is assumed
like linear form, as: to be in the form:
ẋ(t) = A(x)x(t) + B(x)u(t), (2) ẋ(t) = f(x) + g(x)u(t), (10)
where f(x) = A(x)x(t), B(x) = g(x). y(t) = h(x). (11)
The exact solution to the optimal control of nonlinear system That nonlinear system can be expressed in a state-dependent
(2) subject to the cost function like linear form, as:
1 ẋ(t) = A(x)x(t) + B(x)u(t), (12)
J(x, u) = x′ (t f )Fx(t f ) +
2
1 tf ′
Z y(t) = C(x)x(t), (13)
x (t)Q(x)x(t) + u′(x)R(x)u(x) dt, (3)
2 t0 where f(x) = A(x)x(t), B(x) = g(x), h(x) = C(x)x(t).
The goal is to find a state feedback control law of the form
where Q(x) and F is a symmetric positive semi-definite matrix,
u(x) = −kx(t), that minimizes a cost function given by [13]:
and R(x) is a symmetric positive definite matrix, is given by
1
u(x,t) = −R−1 (x)B′ (x)[P(x,t)x(t) + Π], (4) J(x, u) = x′ (t f )Fx(t f )
2
1 tf ′
Z
where
+ x (t)Q(x)x(t) + u′(x)R(x)u(x) dt, (14)
2 t0
1 ∂ P(x,t)
Π = [x′ Px1 x . . . x′ Pxn x]′ , Pxi = . (5) where Q(x) and F is a symmetric positive semi-definite matrix,
2 ∂ xi
and R(x) is a symmetric positive definite matrix. Moreover,
P(x,t) is a symmetric positive definite solution to the equation
x′ Q(x)x is a measure of control accuracy and u′ (x)R(x)u(x)
−Ṗ(x,t) = P(x,t)A(x) + A′(x)P(x,t) is a measure of control effort.
−P(x,t)B(x)R−1 (x)B′ (x)P(x,t) + Q(x) + Ω, (6)
B. Solution for Finite-Horizon SDRE Regulator
with the final condition P(x,t f ) = F To minimize the above cost function (14), a state feedback
where control law can be given as
1 n n
Ω= ∑ ∑ Pxi x[B(x)R−1(x)B′ (x)]i j x′Px j , (7) u(x) = −kx(t) = −R−1 (x)B′ (x)P(x)x(t), (15)
4 i=1 j=1
where P(x,t) is a symmetric, positive-definite solution of the
and [ ]i j is the ith element of the jth row of that matrix. State-Dependent Differential Riccati Equation (SDDRE) of the
Performing some approximations by neglecting terms Ω in form
(6) and Π in (4), which leads to the optimal control given by
−Ṗ(x) = P(x)A(x) + A′(x)P(x)
−1 ′
u(x,t) = −R (x)B (x)P(x,t)x(t), (8) −P(x)B(x)R−1 (x)B′ (x)P(x) + Q(x), (16)
resulted from solving DRE with the final condition
As the SDRE is a function of (x,t), we do not know the value B. Solution for Finite-Horizon SDDRE Tracking
of the states ahead of present time step. Consequently, the
To minimize the above cost function (25), a feedback
state dependent coefficients cannot be calculated to solve (16)
control law can be given as
with the final condition (17) by backward integration from
t f to t0 . To overcome the problem, an approximate analytical
u(x) = −R−1 B′ (x)[P(x)x − g(x)], (26)
approach is used [9]. Which converts the original nonlinear
Ricatti equation to a differential Lyapunov equation. At each
where P(x) is a symmetric, positive-definite solution of the
time step, the Lyapunov equation can be solved in closed form.
SDDRE of the form
In order to solve the DRE (16), one can follow the following
steps at each time step: −Ṗ(x) = P(x)A(x) + A′(x)P(x)
1) Solve Algebraic Riccati Equation to calculate the steady −P(x)B(x)R−1 B′ (x)P(x) + C′(x)Q(x)C(x), (27)
state value Pss (x)
with the final condition
Pss (x)A(x) + A′ (x)Pss (x)
−Pss (x)B(x)R−1 (x)B′ (x)Pss (x) + Q(x) = 0. (19) P(x,t f ) = C′ (t f )FC(t f ). (28)
2) Use changing of variables technique and assume that The resulting SDRE-controlled trajectory becomes the solution
K(x, t) = [P(x, t) − Pss(x)]−1 . of the state-dependent closed-loop dynamics
3) Calculate the value of Acl (x) as
Acl (x) = A(x) − B(x)R−1B′ (x)Pss (x). ẋ(t) = [A(x) − B(x)R−1(x)B′ (x)P(x)]x(t)
4) Solve the algebraic Lyapunov equation [14] +B(x)R−1 (x)B′ (x)g(x), (29)
Acl D + DA′cl − BR−1 B′ = 0. (20) where g(x) is a solution of the state-dependent non-
homogeneous vector differential equation
5) Solve the differential Lyapunov equation
ġ(x) = −[A(x) − B(x)R−1(x)B′ (x)P(x)]′ g(x)
K̇(x, t) = K(x, t)A′cl (x) + Acl (x)K(x, t)
−C′ (x)Q(x)z(x), (30)
−B(x)R−1 B′ (x). (21)
with the final condition
The solution of (21), as shown by [15], is given by
′ g(x,t f ) = C′ (t f )Fz(t f ). (31)
K(t) = eAcl (t−tf ) (K(x, tf ) − D)eAcl (t−tf ) + D. (22)
Similar to Section III, an approximate analytical approach is
6) Calculate the value of P(x, t) from the equation
used and the DRE can be solved in the following steps at each
time step:
P(x, t) = K−1 (x, t) + Pss(t). (23)
1) Solve for P(x, t) similar to the SDDRE regulator prob-
7) Finally, calculating the value of the optimal control lem in Section III
u(x,t) as 2) Calculate the steady state value gss (x) from the equation
u(x,t) = −R−1 B′ (x)P(x, t)x. (24) gss (x) = [A(x) − B(x)R−1 (x)B′ (x)Pss (x)]′−1
C′ (x)Q(x)z(x). (32)
IV. N ONLINEAR F INITE -H ORIZON T RACKING U SING 3) Use changing of variables technique and assume that
SDRE Kg (x, t) = [g(x, t) − gss(x)].
4) Solve the differential equation
A. Problem Formulation
−1
B′ P)′ (t−tf )
Consider nonlinear system given in (10) and (11),which can Kg (x,t) = e−(A−BR [g(x,t f ) − gss (x)]. (33)
be redescibed in the form (12) and (13), Let z(t) be the desired
output. 5) Calculate the value of g(x,t) from the equation
The goal is to find a state feedback control law that
minimizes a cost function given by : g(x, t) = Kg (x, t) + gss(x). (34)
1.8 2
Actual Angle
Desired Angle
1.6 1.5
1.4
1
1.2
1
0
0.8
−0.5
0.6
−1
0.4
Missile Trajectory
Target Trajectory −1.5
0.2
0 −2
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5 0 1 2 3 4 5 6 7 8
Range [Km] Time [sec]
Fig. 5. Missile-Target Engagement Scenario (Case 1) Fig. 6. Angle Trajectories for Gimbaled System (Case 1)
1
The system nonlinear state equations can be written in the 0.8
form: 0.6
0.4
x˙1 = x2 , (38)
where: θ = x1 , θ̇ = x2 ,i = x3 , V = u. −1
0 1 2 3 4 5 6 7 8
Time [sec]
Or alternatively in state dependent form:
0 1 0
x˙1 x1 0 Fig. 7. Optimal Control for Gimbaled System (Case 1)
x˙2 = (g/l)sin(x1 ) km x2 + 0 u
x1 0 ml 2 (42)
x˙3 x3 1
0 − kLb R
L L
The weight matrices are chosen to be and the engagement scenario is shown in Fig. 9. The resulting
trajectories for the demanded and achieved seeker angles are
Q = diag(3000, 0, 0), R = 30, F = diag(1, 1, 1). (43) illustrated in Fig. 10, the optimal control is shown in Fig. 11,
and the optimal error is shown in Fig. 12.
A. Case 1: Fixed Target In Fig. 10, the solid line denotes the actual (achieved) angle
Consider a fixed target, in this case the desired seeker angle trajectory of the finite-horizon tracking controller, the dashed
will be z(t) = 0o , i.e the problem is now a regulator problem. line denotes the desired seeker angle.
The simulations were performed for final time of 8 seconds, Fig. 9 show that a successful hit is observed with accept-
and the engagement scenario is shown in Fig. 5. The resulting able miss-distance. Comparing these trajectories in Fig. 10,
trajectories for the demanded and achieved seeker angles are it’s clear that the developed finite-horizon SDDRE nonlinear
presented in Fig. 6, the optimal control is shown in Fig. 7,
and the optimal error is shown in Fig. 8.
In Fig. 6, the solid line denotes the actual (achieved) angle
0.25
trajectory of the finite-horizon tracking controller, the dashed 0.2
line denotes the desired seeker angle. 0.15
As shown in Fig. 5, a successful hit is observed. Fig. 0.1
Optimal Error [deg]
−0.05
is able to solve the SDRE finite-horizon nonlinear regulator −0.1
problem with a zero average optimal error and 0.003o standard −0.15
deviation. −0.2
−0.25
0 1 2 3 4 5 6 7 8
B. Case 2: Non-Maneuvering Target Time [sec]
9
6
8
5 7
6
4
Altitude [Km]
Altitude [Km]
3 4
3
2
2
1 1 Missile Trajectory
Missile Trajectory Target Trajectory
Target Trajectory 0
0 1 2 3 4 5 6 7 8 9
0 Range [Km]
0 0.5 1 1.5 2 2.5 3 3.5 4 4.5 5
Range [Km]
10
10
8
5
6
0
Actual Angle
Desired Angle 4
−5
Optimal Angles [deg]
−25 −2
−30 −4
−35
−6
−40
0 1 2 3 4 5 6 7 8
Time [sec] −8 Actual Angle
Desired Angle
−10
0 1 2 3 4 5 6 7 8 9 10
Fig. 10. Angle Trajectories for Gimbaled System (Case 2) Time [sec]
−5
ation error of 0.075o.
−10
−15
C. Case 3: Maneuvering Target
−20
Consider a highly maneuvering target. The simulations were
−25
performed for final time of 10 seconds, and the engagement
−30
0 1 2 3 4 5 6 7 8
scenario is shown in Fig. 13. The resulting trajectories for the
Time [sec]
demanded and achieved seeker angles are illustrated in Fig.
14, the optimal control is shown in Fig. 15, and the optimal
Fig. 11. Optimal Control for Gimbaled System (Case 2)
error is shown in Fig. 16.
In Fig. 14, the solid line denotes the actual (achieved) angle
2
trajectory of the finite-horizon tracking controller, the dashed
line denotes the desired seeker angle.
1.5 Fig. 13 show that a successful hit is observed with accept-
able miss-distance. Comparing these trajectories in Fig. 14,
1
Optimal Error [deg]
10
R EFERENCES
[1] C. A. Woolsey, “Review of marine control systems: Guidance, navi-
5
gation, and control of ships, rigs and underwater vehicles,” Journal of
Guidance, Control, and Dynamics, vol. 28, no. 3, pp. 574–575, 2005.
0
[2] G. M. Siouris, Missile Guidance and Control Systems. Springer, 2004.
[3] T. Çimen, “Recent advances in nonlinear optimal feedback control
Optimal Control [V]
−5
design,” Proceedings of the 9th WSEAS International Conference on
Applied Mathematics, Istanbul, Turkey, pp. 460–465, May 2006.
−10
[4] J. R. Cloutier, “State-dependent Riccati equation techniques: An
overview,” Proc. American Control Conference, vol. 2, pp. 932–936,
−15
1997.
[5] T. Çimen, “Development and validation of a mathematical model for
−20
control of constrained nonlinear oil tanker motion,” Mathematical and
Computer Modeling of Dynamical Systems, vol. 15, no. 1, p. 1749, 2009.
−25
[6] A. Ratnoo and D. Ghose, “State-dependent Riccati-equation-based guid-
−30
ance law for impact-angle-constrained trajectories,” Journal of Guid-
0 2 4 6 8 10 ance, Control, and Dynamics, vol. 32, no. 1, pp. 320–326, 2009.
Time [sec]
[7] S. Vaddi, P. K. Menon, and E. J. Ohlmeyer, “Numerical state-dependent
Riccati equation approach for missile integrated guidance control,”
Fig. 15. Optimal Control for Gimbaled System (Case 3) Journal of guidance, control, and dynamics, vol. 32, no. 2, pp. 699–
703, 2009.
[8] T. Çimen, “A generic approach to missile autopilot design using state-
2
dependent nonlinear control,” in 18th IFAC World Congress, 2011, pp.
9587–9600.
1.5 [9] A. Heydari and S. N. Balakrishnan, “Path planning using a novel
finite-horizon suboptimal controller,” Journal of Guidance, Control, and
1
Dynamics, pp. 1–5, 2013.
[10] T. Nguyen and Z. Gajic, “Solving the matrix differential Riccati equa-
Optimal Error [deg]
0.5
tion: a Lyapunov equation approach,” IEEE Trans. Automatic Control,
0 vol. 55, no. 1, pp. 191–194, 2010.
[11] J. Nazarzadeh, M. Razzaghi, and K. Nikravesh, “Solution of the matrix
−0.5
Riccati equation for the linear quadratic control problems,” Mathemati-
−1 cal and Computer Modelling, vol. 27, no. 7, pp. 51–55, 1998.
[12] A. Khamis and D. Naidu, “Nonlinear optimal tracking using finite-
−1.5 horizon state dependent Riccati equation ( SDRE),” Proceedings of
the 4th International Conference on Circuits, Systems, Control, Signals
−2
0 1 2 3 4 5
Time [sec]
6 7 8 9 10 (WSEAS), pp. 37–42, August 2013, valencia, Spain.
[13] D. S. Naidu, Optimal Control Systems. CRC Press, 2003.
[14] Z. Gajic and M. Qureshi, “The Lyapunov matrix equation in system
Fig. 16. Optimal Error for Gimbaled System (Case 3) stability and control,” New York: Dover Publications, 2008.
[15] A. Barraud, “A new numerical solution of xdot=a1*x+x*a2+d, x(0)=c,”
IEEE Transaction on Automatic Control, vol. 22, no. 6, pp. 976–977,
Dec. 1977.
variables that converts the differential Riccati equation to a [16] A. Khamis, “Design and realization of hardware in loop simulation for a
linear Lyapunov equation.The Lyapunov equation is solved homing guided missile,” Ph.D. dissertation, Military Technical Collage,
Cairo, 2007.
in a closed form at the given time step. Simulation results [17] B. Etkin and L. D. Reid, Dynamics of Flight: Stability and Control.
for gimbaled system in missile seeker are included. Three Wiley New York, 1982.
engagement scenarios including fixed target, non-maneuvering [18] A. Calise, S. Lee, and M. Sharma, “Direct adaptive reconfigurable
control of a tailless fighter aircraft,” AIAA Guidance, Navigation, and
target, and maneuvering target are considered to demonstrate Control Conference, pp. 88–97, 1998.
the effectiveness of the developed technique. [19] Y. Ulybyshev, “Terminal guidance law based on proportional naviga-
tion,” Journal of Guidance, Control, and Dynamics, vol. 28, no. 4, pp.
821–824, 2005.