Professional Documents
Culture Documents
L1 Intro
L1 Intro
1. Introduction
Adaptive Control 2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
Karl Johan Åström and Anders Rantzer 4. Estimation and Excitation
5. Minimum Variance Control
Department of Automatic Control, LTH
Lund University 6. Self-Tuning Regulators
7. Learning and Dual Control
January 20, 2021
8. Applications
9. Related Fields
10. Summary
Most of you know that with the advent a few years ago of hypersonic and
Pubications per year
4
10
supersonic aircraft, the Air Force was faced with a control problem. This
problem was two-fold; one, it was taking a great deal of time to develop a
flight control system; and two, the system in existence were not capable of
2
10
fulfilling future Air Force requirements. These systems lacked the ability to
10
0
control the aircraft satisfactorily under all operating conditions.
1920 1930 1940 1950 1960 1970 1980 1990 2000 2010 2020
Year
Contracts on two important test flights
Blue control red adaptive control ◮ Honeywell self oscillating adaptive system X-15
◮ Lear will test MIT model reference adaptive system on F-101A
1
Some Landmarks Yakov Z Tsypkin 1919 - 1997
◮ Early flight control systems 1955 ◮ BS, Moscow Electrical Engineering Institute 1941
◮ Dynamic programming Bellman 1957 ◮ BS, Moscow State University 1943
◮ Dual control Feldbaum 1960 ◮ MS Engineering, Moscow State University 1945
◮ System identification 1965 ◮ PhD, Moscow State University 1948
◮ Learning control Tsypkin 1971 ◮ Engineer, senior engineer, Chief of Department,
Research Institute Aircraft Equipment 1941-1949
◮ Algorithms MRAS STR 1970
◮ Senior researcher, Institute Control Sciences,
◮ Stability analysis 1980
Moscow 1950-1957
Lyapunov - Tsypkin
◮ Head of laboratory, Institute Control Sciences,
Passivity - Popov Landau
Moscow, since 1957
Augmented Error - Monopoli
◮ Yakov Z. Tsypkin and C. Constanda Relay Control ◮ Lenin Prize 1960
◮ PID auto-tuning 1980
Systems.
◮ ◮ Quazza Medal IFAC
Industrial products 1980 ◮ Sampling Systems Theory and Its Application 1984
◮ Robustness 1985 Volume 1 and 2 (NY 1964 Macmillan. Translated
◮ Hartley Medal IMC 1985
◮ from Russian by A. Allen an...)
Automatic tuning 1985
◮ Rufus Oldenburger
◮ Foundations of the Theory of Learning Systems by
◮ Autonomous Control 1995 Medal ASME 1989
Tsypkin, Ya. Z. (1973) Paperback
◮ Adaptation and Learning - a renaissance
The Self-Oscillating Adaptive System H. Schuck Honeywell 1959 The Self-Oscillating Adaptive System
It’s is rather hard to tell when we at Honeywell first became interested in adaptive control. Gain
... In retrospect, it seems that we first conciously articulated the need in connection with our Filter
changer
early work in automatic approach and landing. ...
Let us look at the adaptive flight control system that was proven in the flight tests on the
F-94C. Conceptually it is simple, deceptively so. The input is applied to a model whose Filter Process
ym y
dynamic performance is what we whish the dynamic performance of the aircraft to be. The
actual response is compared with the reponse of the model and the difference is used as
Model Σ G f (s) Σ G p (s)
e
an input to the servo. If the gain of the servo is sufficiently high, the response of the aircraft
will be identical to that of the model, no matter what the elevator effectiveness, so long as it
is finite and has the right direction. Dither
Design of the model is fairly simple. ... The big problem comes inconnection with the need
to make the gain of the servo sufficiently high. An ordinary linear servo loop will not do. It
simply cannot be given a sufficiently high gain and still be stable. So we go in for −1
non-linearity, the most extreme form of non-linearity, in fact - the bang-bang type. Full
available power is applied one way, or the other, depending on the the direction of the
◮ Shape behavior by the block called Model
swithching order.
Now it is well known that a simple bang-bang sytem is oscillatorye. And we don’t want an ◮ Make the inner loop as fast as possible
oscillatory aircraft. .. So we look for ways to tame it down, keeping its high-gain ◮ Relay feedback automatically adjusts to gain margin for low
characteristic while reducing it oscillatory activity. ... frequency signals to gm = 2! Dual input describing functions!!!
◮ Relay amplitude adjusted by logic
k
SOAS P (s) = SOAS Simulation - Adding Lead Network
s(s + 1)2
1 1
y y
ym ym
−1 −1
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
1 u 1 u
−1 −1
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
0.5
0.5 e e
−0.5 −0.5
0 10 20 30 40 50 0 10 20 30 40 50
Time Time
2
SOAS Simulation - Adding Gain Changer The X-15 with MH-96 Autopilot Crash Nov 11 1967
1
y
ym
−1
0 10 20 30 40 50
Time
1 u
−1
0 10 20 30 40 50
Time
0.5
e
−0.5
Logic for gain change did not increase gain fast enough
0 10 20 30 40 50
Time
Dydek, Zachary, Anuradha Annaswamy, and Eugene Lavretsky. “Adaptive Control and the
Gain increases by a factor of 5 at time t = 25 NASA X-15-3 Flight Revisited.” IEEE Control Systems Magazine 30.3 (2010): 32–48.
Model Reference Adaptive Control – P. Whitaker MIT 1959 Model Reference Adaptive Control – MRAS
ym
We have further suggested the name model-reference adaptive system for Model
the type of system under consideration. A model-reference system is
Controller parameters
characterized by the fact that the dynamic specification for a desired Adjustment
system output are embodied in a unit which is called the model-reference mechanism
for the system, and which forms part of the equipment installation. The uc
commandsignal input to the control system is also fed to the model. The u
Plant
y
Controller
difference between the output signal of the model and the corresponding
output quantity of the system is then the response error. The design
objectie of the adaptive portion of this type of system is to minimze this
Linear feedback from e = y − ym is not adequate for parameter
response error under all operational conditions of the system. Specifically
adjustment!
the adjustment is done by the MIT Rule.
The MIT rule
dθ e
= −γ e
dt θ
Many other versions
u ( t ) = θ 1 uc ( t ) − θ 2 y ( t ) e b2θ 1 b
=− uc = − y
θ 2 (p + a + bθ 2 )2 p + a + bθ 2
Ideal controller parameters
Approximate
bm p + a + bθ 2 ( p + am
θ1 = θ 10 =
b
am − a Hence
θ 2 = θ 20 = dθ 1 am
3 4
b = −γ uc e
dt p + am
Find a feedback that changes the controller parameters so that the closed dθ 2
3
am
4
loop response is equal to the desired model =γ y e
dt p + am
3
Block Diagram Simulation a = 1, b = 0.5, am = bm = 2.
ym
G m (s) 1
−
y
e
Σ
−1
uc + u y +
Π Σ G(s)
0 20 40 60 80 100
− Time
θ1
Π 5 u
θ2
0
γ γ
−
s s −5
0 20 40 60 80 100
am am Time
Π Π
s + am s + am
θ1 γ =5
4
γ =1
2
dθ 1 am γ = 0.2
3 4
= −γ uc e
0
0 20 40 60 80 100
dt p + am
θ2
Time
γ =5
dθ 2 am
3 4
2
=γ y e γ =1
dt p + am 0
γ = 0.2
Example a = 1, b = 0.5, am = bm = 2. 0 20 40 60 80 100
Time
1 1 1
3 4
V ( e, θ 1 , θ 2 ) = e2 + (bθ 2 + a − am )2 + (bθ 1 − bm )2
2 bγ bγ
Lyapunov MIT
Derivative of error and Lyapunov function
de G m (s)
= −am e − (bθ 2 + a − am )y + (bθ 1 − bm ) uc −
dt
e G m (s)
Σ −
e
dV de 1 dθ 2 1 dθ 1 uc
Π
+
Σ
u y + Σ
=e (bθ 2 + a − am ) (bθ 1 − bm )
G(s) +
+ + uc u y +
γ γ Π Σ
−
dt dt dt dt
G(s)
θ1 −
Π θ1
Π
1 dθ 2 θ2
3 4
= − am e 2 + (bθ 2 + a − am ) − γ ye
θ2
γ γ
−
γ
γ γ
dt s s −
s s
1 dθ 1
3 4 am am
Π Π Π Π
+ (bθ 1 − bm ) + γ uc e s + am s + am
γ dt
Adaptation law
dθ 1 dθ 2 de
= −γ uc e, = γ ye [ = − e2
dt dt dt
Error will always go to zero, what about parameters, Barbara’s lemma!
4
Indirect MRAS - Estimate Process Model Indirect MRAS
Process and estimator
dx d x̂
= ax + bu, = âx̂ + b̂u
dt dt
Nominal controller gains: kx = kx0 = (a − am )/b, kr = kr0 = bm /b. Process and estimator
Estimation error e = x̂ − x has the derivative dx d x̂
= ax + bu, = âx̂ + b̂u
de dt dt
= âx + b̂u − ax − bu = ae + (â − a)x̂ + (b̂ − b)u = ae + ãx̂ + b̃u,
dt Control law
â − am bm
where ã = â − a and b̃ = b̂ − a. Lyapunov function u=− x+ r
b̂ b̂
11 2 Very bad to divide by b̂!
2V = e2 + ã + b̃2 .
2
γ
dV de 1 1 d â d b̂ 2 1 d ã 2 1 d b̃ 2
= ae2 + ex̂ +
1 1
=e + ã + b̃ ã + eu + b̃
dt dt γ dt dt γ dt γ dt
â − am bm
u=− x+ r 1. Introduction
b̂ b̂
b̂u + (â − am )x − bm r = 0 2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
with the differential equation
4. Estimation and Excitation
du 5. Minimum Variance Control
= K bm r − (â − am )x − b̂u
! "
dt 6. Self-Tuning Regulators
Avoid division by b̂, can be interpreted as sending the signal 7. Learning and Dual Control
b̂m r + (am − â)x through a first order filter with the pole s = −K b̂. 8. Applications
9. Related Fields
L1 has been tested on several airplanes
10. Summary
I. M. Gregory, C. Cao, E. Xargay, N. Hovakimyan and X. Zou L1 Adaptive Control Design
for NASA AirSTAR Flight Test Vehicle. AIAA Guidance, Navigation, and Control
Conference Portland Oregon August 2011
5
Persistent Excitation - Examples Lack of Identifiability due to Feedback
a aˆ
Parameters
−1
a = −0.9, b = 0.5, σ = 0.5, θˆ(0) = 0, P (0) = 100I
P (s)C(s) P (s)
Y (s) = R(s) + V (s) Excitation
1 + P (s)C(s) 1 + P (s)C(s)
◮ Unit pulse at t = 50
C(s) C(s)P (s)
U (s) = R(s) − V (s) ◮ Square wave of unit amplitude and period 100
1 + P (s)C(s) 1 + P (s)C(s)
◮ Two cases
1
Y (s) = P (s)U (s) if v = 0, and Y (s) = − U (s) if r = 0. a) u(t ) = −0.2y (t )
C(s) b) u(t ) = −0.32y (t − 1)
Identification will then give the negative inverse of controller transfer
function! Any signal entering between u and v will influence closed loop
identification severely A good model can only be obtained if v = 0, or if v
is much smaller than r!
6
The Drying Section Steady State Regulation of Basis Weight and Moisture Content
◮ Servomechanism theory 1945 Set up as an independent non-profit research organization (Think Tank) for
the US Airforce by Douglas Aircraft Corporation in 1945.
◮ IFAC 1956 (50 year jubilee in 2006)
Saab R System
◮ Widespread education and industrial use of
control
◮ The First IFAC World Congress Moscow 1960 ◮ Richard Bellman
◮ Exciting new ideas
◮ Georg Danzig LP
Dynamic Programming Bellman 1957
Maximum Principle Pontryagin 1961
◮ Henry Kissinger
Kalman Filtering ASME 1960 ◮ John von Neumann
◮ Exciting new development ◮ Condolezza Rice
The space race (Sputnik 1957)
◮ Donald Rumsfeld
Computer Control Port Arthur 1959
◮ IBM and IBM Nordic Laboratory 1960 ◮ Paul Samuelson
Computerized Process Control
7
Modeling from Data (Identification) Practical Issues
The Likelihood function (Bayes rule)
N
1 X ε 2 (t ) N
p(Yt , θ ) = p(y (t )pYt −1 , θ ) = · · · = − − log 2πσ 2
2 σ2 2
1 ◮ Sampling period
θ = (a1 , . . . , an , b1 , . . . , bn , c1 , . . . , cn , ε (1), .., ) ◮ To perturb or not to perturb
Ay (t ) = Bu(t ) + Ce(t ) Cε (t ) = Ay (t ) − Bu(t ) ◮ Open or closed loop
ε = one step ahead prediction error experiments
Efficient computations ◮ Model validation
N ◮ 20 min for two-pass compilation
J X ε (t ) ε (t )
= ε (t ) C = qk y ( t ) of Fortran program!
ak ak ak
1 ◮ Control design
◮ Good match identification and control. Prediction error is minimized ◮ Skills and experiences
in both cases!
KJÅ and T. Bohlin, Numerical Identification of Linear Dynamic Systems from Normal
Operating Records. In Hammond, Theory of Self-Adaptive Control Systems, Plenum Press,
January 1966.
KJÅ, P Hagander, J Sternby Zeros of sampled systems. Automatica 20 (1), 31-38, 1984
◮ Robustness and tuning
Td Td + Ts Tp
Ts t
8
Back to Billerud - Performance of Minimum Variance Control IBM Scientific Symposium Control Theory and Applications 1964
9
The Self-Tuning Regulator 2 The Self-Tuning Regulator 3
Process model, estimation model and control law
yt + a1 yt −1 + · · · + an yt −n = b0 ut −k + · · · + bm ut −m
Estimates of parameters in
+ et + c1 et −1 + · · · + cn et −n
yt +a1 yt −1 +· · ·+an yt −n = b0 ut −k +· · ·+bm ut −m +et +c1 et −1 +· · ·+cn et −n yt +k = s0 yt + s1 yt −1 + · · · + sm yt −m + r0 (ut + r1 ut −1 + · · · rn ut −{ )
ut + r̂1 ut −1 + · · ·r̂n ut −{ = −(ŝ0 yt + ŝ1 yt −1 + · · · + ŝm yt −m )/r0
gives unbiased estimates if ci = 0.
Surprisingly the self-tuner works with ci ,= 0. Discovered empirically in If estimate converge and 0.5 < r0 /b0 < ∞
simulations by Wieslander, Wittenmark and independently by Peterka. ry (τ ) = 0, τ = k , k + 1, · · · k + m + 1
Johan Wieslander and Björn Wittenmark An approach to adaptive control ryu (τ ) = 0, τ = k , k + 1, · · · k + {
using real time identification. Proc 2nd IFAC Symposium on Identification
If degrees sufficiently large ry (τ ) = 0, ∀τ ≥ k
and Process Parameter Estimation, Prague 1970
◮ Essentially Kalman’s algorithm
Vaclav Peterka Adaptive digital regulation of noisy systems. Proc 2nd IFAC ◮ Converges to minimum variance control even if ci ,= 0, Surprising!
Symposium on Identification and Process Parameter Estimation, Prague ◮ Automates identification and minimum variance control in about 30 lines of code.
1970 ◮ The controller that drives covariances to zero
yt + a1 yt −1 + · · · + an yt −n = b0 ut −k + · · · bm ut −n
+ et + c1 et −1 + · · · + cn et −n
Estimation model
yt +k = s0 yt + s1 yt −1 + · · · + sm yt −m + r0 (ut + r1 ut −1 + · · · rn ut −{ )
p fp 1 2 2
dx = f (x )dt + g(x )dw =− + g f
t x 2 x2 1. Introduction
dθ
θ t + 1 = θ t + γt φ e = f (θ ) = Eφ e 2. Self-oscillating Adaptive Control
dτ
3. Model Reference Adaptive Control
Method for convergence of recursive algorithms. Global stability of STR
4. Estimation and Excitation
(Ay = Bu + Ce) if G(z ) = 1/C(z ) − 0.5 is SPR
5. Minimum Variance Control
L. Ljung, Analysis of Recursive Stochastic Algorithms IEEE Trans AC-22 (1967) 551–575.
6. Self-Tuning Regulators
Converges locally if ℜC(zk ) > 0 for all zk such that B(zk ) = 0 7. Learning and Dual Control
Jan Holst, Local Convergence of Some Recursive Stochastic Algorithms. 5th IFAC 8. Applications
Symposium on Identification and System Parameter Estimation, 1979 9. Related Fields
Lei Gui and Han-Fu Chen, The Åström-Wittenmark Self-tuning Regulator Revisited and
ELS-Based Adaptive Trackers. IEEE Trans AC36:7 802–812.
10
The Problem - Integrator with unknown gain The Hamilton-Jakobi-Bellman Equation
Consider the system
yt +1 = yt + but + et +1 The solution to the problem is given by the Bellman equation
where et is a sequence of independent normal (0, σ 2 ) random variables
Vt (Xt ) = EXt min E yt2+1 + Vt +1 (Xt +1 )Xt
1 2
and b a constant but unknown parameter with a normal b̂, P (0) prior or a ut
random wai.
Find a control llaw such that ut based on the information available at time t The state is Xt = yt , yt −1 , yt −2 , . . . , y0 , ut −1 , ut −2 , . . . , u0 . The derivation
is general applies also to
Xt = yt , yt −1 , . . . , y0 , ut −1 , ut −2 , . . . , u0 ,
xt +1 = f (xt , ut , et )
that minimizes the cost function yt = g(xt , ut , vt )
X
T
X min E q(x1 , ut )
2
V =E y (k ).
k =1 How to solve the optimization problem?
The curse of dimensionality: Xt has high dimension
KJÅ and A. Helmersson. Dual Control of an Integrator with Unkown Gain, Computers and
Mathematics with Applications 12:6A, pp 653–662, 1986.
It can be shown that a sufficient statistic for estimating future outputs is yt Vt (Xt ) = EXt min E yt2+1 + Vt +1 (Xt +1 )Xt
1 2
and the conditional distribution of b given Xt . In our setting the conditional u t
Short Time Horizon - One Step Ahead Controller Gain - Cautious Control
Consider situation at time t and look one step ahead
b̂ y b̂
T
X u=− y = Ky , η= . β = √ ,
VT −1 (y , b̂, P ) = min E yk2 = min yT2 b̂2 + P σ P
u u
k =T
yT = yT −1 + buT −1 + eT
Py 2
= min y + 2y b̂u + u2 (b̂2 + P ) + σ 2 = σ 2 + 2
2
1 2
u b̂ + P
where minimum occurs for
b̂ 1
u=− y [ u=− y as P → 0
b̂2 + P b̂
These control laws are called cautious control and certainty equivalence
control (Herbert Simon).
The Solution and Scaling Understanding Probing - Policy function has multiple minima
Py 2
VT (y , b̂, P ) = σ 2 +
+P b̂2
Iterate backward in time. Scale variable
√
y b̂ u P
η= . β = √ , µ=
σ P σ
The scaled Bellman equation
1 2
Wt (η, β ) = min Ut (η, β , µ), φ (x ) = √ e−x /2
µ 2π
Ut (η, β , µ) = (η + β µ)2 + 1 + µ2
Z∞1 p p 2
+ Wt + 1 ( η + β µ + ε 1 + µ2 , β 1 + µ2 + µε φ (ε )dε
−∞
Solving minimization gives control law µ = Π(η, β ), where η is scaled
control error and β is the relative parameter uncertainty Two functions: the
value function W (η, β ) and the policy function U (η, β , µ) Notice jump!!
11
Controller Gain - 3 Steps Controller gain for 30 Steps
Comparison Summary
◮ Use dynamic programming (move optimization inside the integral)
T T
X 1X
2
Cautious Control Dual Control min E yt2 = min EXt E yk2 Xt
Ut Ut
k =t +1 k =t +1
T
1X 2 T
1X 2
= EXt min E yk2 Xt = EXt min min E yk2 Xt
Ut ut Ut +1
k =t +1 k =t +1
T 2
1 X
= EXt min E yt2+1 + min yk2 Xt
ut Ut +1
k =t +2
√
◮ State reduction y , b̂, P → η = y /σ , β = b̂/ P
◮ Certainty equivalence, cautious and dual
◮ K. J. Åström Control of Markov Chains with Incomplete State Information
JMAA, 10, pp. 174–205, 1965.
◮ K. J. Åström and A. Helmersson. Dual Control of an Integrator with Unkown
Gain, Computers and Mathematics with Applications 12:6A, pp 653–662,
1986.
w xt +1 = Ai xt + Bi ut + wt t≥0
(x , u ) ✛
✛ x + = Ax + Bu + w ✛ ut = µt (x0 , . . . , xt , u0 , . . . , ut −1 ).
u
◮ If N = 1, then this gives standard H∞ control.
✲ ◮ γ quantifies robustness to unmodeled dynamics.
Dual controller
◮ In general, nonlinear feedback with memory is needed.
◮ Early work by [Didinsky/Basar, CDC 1994]
◮ Extension to output feedback difficult
12
Example: Double Integrator with Unknown Sign Adaptive Control
1. Introduction
2 −1 1 0
Results Steermaster
Kockums
◮ Ship dynamics
◮ SSPA Kockums
◮ Full scale tests on ships
Significant improvement of production 15 %! in operation
13
First Controll Relay Auto-Tuner Tore
◮ One-button tuning
◮ Automatic generation of
gain schedules
◮ Adaptation of feedback and
◮ Gunnar Bengtsson
feedforward gains
◮ Founder of ABB Novatune ◮ Many versions
◮ Rolling mills Single loop controllers
◮ Continuous casting DCS systems
Addaline - Bernard Widrow & Ted Hoff (Intel 4004) BOXES 1960 - Donald Michie
◮ Cryptography Bletchley Park
◮ Director Department of Machine
Intelligence and Perception, University
of Edinburgh (previously the
Experimental Programming Unit 1965)
◮ Quantize state in boxes
◮ Each box stores information about
control actions taken and the
performance
◮ Adaline - ADAptive LInear Neuron - Single layer neuron network
Pm ◮ Local demon and global demons
y = sign j =1 wij uj , least Mean Square (LMS)
◮ Used for playing games like
◮ B. Widrow and M. E. Hoff, “Adaptive Switching Circuits,” 1960 IRE
Tick-Tack-Toe and broom balancing
WESCON Convention Record, 1960, pp. 96-104.
◮ Analog implementation potentiometers D. Michie and R. A. Chambers, “BOXES: An experiment in adaptive
control,” in Machine Intelligence 2, E. Dale and D. Michie, Eds. Edinburgh:
◮ Separating hyperplane
Oliver and Boyd, (1968), pp. 137–152.
◮ Madaline a multilayer version
14
Development of Image Recognition Adaptive Control
1. Introduction
2. Self-oscillating Adaptive Control
3. Model Reference Adaptive Control
4. Estimation and Excitation
5. Self-Tuning Regulators
6. Learning and Dual Control
7. Applications
8. Related Fields
9. Summary
Very rapid development because of good test batch and competing algorithms
Summary
15