PCACCSMDec 2021
PCACCSMDec 2021
Adaptive Control
A NUMERICAL INVESTIGATION OF PERSISTENCY,
CONSISTENCY, AND EXIGENCY
ZARMEEN AZMAT
A
mong the multitude of modern control methods, (LQR) and linear quadratic Gaussian (LQG), and nonlinear
model predictive control (MPC) is one of the most control methods, such as feedback linearization and sliding
successful [1]–[4]. As noted in “Summary,” this mode control. Although MPC is computationally intensive,
success is largely due to the ability of MPC to it is more broadly applicable than Hamilton–Jacobi–Bellman-
respect constraints on controls and enforce con- based control and more suitable for feedback control than the
straints on outputs, both of which are difficult to handle with minimum principle. In many cases, the constrained optimi-
linear control methods, such as linear quadratic regulator zation problem for receding-horizon optimization is convex,
which facilitates computational efficiency [5].
Digital Object Identifier 10.1109/MCS.2021.3107647 MPC uses a model of the plant with constrained reced-
Date of current version: 12 November 2021 ing-horizon optimization to compute a sequence of
M
discarded. At a first glance, this strategy seems ad hoc and odel predictive control, which is based on receding-
wasteful. However, receding-horizon optimization is horizon optimization, is a widely used modern control
known to be asymptotically stabilizing for a sufficiently technique with numerous successful application in diverse
long prediction horizon or with terminal constraints, areas. Much of this success is due to the ability of reced-
and, if the initial state is feasible and the model matches ing-horizon optimization to enforce state and control con-
the plant, it is recursively feasible with respect to con- straints, which are crucial in many applications of control.
straints [6]–[9]. To avoid the need for an observer, predictive cost adaptive
MPC handles convex control and output constraints ef- control (PCAC) uses the block observable canonical form,
ficiently compared to conventional feedback control meth- whose state consists of past values of the control inputs and
ods that treat control constraints as input nonlinearities. In measured outputs. PCAC also uses recursive least squares
particular, linear dynamics with control magnitude and (RLS) with variable-rate forgetting for online identification.
move-size (rate) saturation comprise a Hammerstein system The article describes the algorithmic details of PCAC
[10]–[12], whose importance is reflected by the vast litera- and numerically investigates its performance through a col-
ture on bounded control and antiwindup strategies [13]– lection of numerical examples that highlight various control
[15]. Within the context of MPC, however, magnitude and challenges, such as model-order uncertainty, sensor noise,
move-size control constraints are treated as optimization prediction horizon, stabilization, magnitude and move-size
constraints rather than explicit nonlinearities, which simpli- saturation, and stabilization. The numerical examples are
fies the treatment of this problem. used to probe the performance of PCAC in terms of persis-
From the perspective of classical control and modern tency, consistency, and exigency. Since, unlike dual con-
optimal state-space control, MPC is a paradigm shift. For trol, PCAC does not employ a separate control perturbation
example, conventional feedback methods use integral ac- to enhance persistency, the focus is on self-generated per-
tion to asymptotically follow step commands and reject sistency during transient operation. For closed-loop iden-
step disturbances, and it is common practice to embed inte- tification using RLS, sensor noise gives rise to bias in the
grators within LQR and LQG controllers. MPC, however, identified model, and the goal is to determine the effect of
performs numerical optimization at each step, and thus no the lack of consistency. Finally, the numerical examples are
integrator per se appears [8]. Without integral action, MPC used to assess exigency, which is the extent to which the
lacks one of the key strengths of classical control, namely, online identification emphasizes model characteristics that
the ability to follow setpoint commands and reject step dis- are most relevant to meeting performance objectives.
turbances without knowledge of the dc gain of the plant.
On the other hand, by eschewing integral action, MPC is
not degraded by integrator windup. This dependence is mitigated by robust MPC techniques [31],
Since MPC uses model-based prediction to determine [32], minimax techniques for minimizing the loss of a worst-case
future control inputs, it follows that these predictions require scenario, and robust tube-based MPC [33]–[35].
knowledge of future commands and disturbances. Future com- As an alternative to robust MPC, extensions of MPC to in-
mands are sometimes known; this is the aim of preview control clude online identification and learning are considered in [36]–
and trajectory-tracking methods [16], [17]. On the other hand, [39]. These techniques can potentially overcome the worst-case
except for disturbance-feedforward control architectures [18], considerations of robust MPC by allowing MPC to learn the
disturbances are rarely known, especially in the future. The true plant dynamics and disturbances. For system identifica-
lack of future knowledge of commands and disturbances thus tion, the commands, disturbances, and control inputs must
represents a potential obstacle for MPC. provide sufficient persistency to facilitate identification. Concur-
Although MPC is often based on linear models, this tech- rent learning with MPC can be viewed as a form of indirect
nique is also applicable to nonlinear systems [19]. An addition- adaptive control [40]. Within this context, the role of persistency
al challenge is the case of output feedback, where not all of the is a longstanding issue [41]. More generally, indirect adaptive
plant states are measured. In this case, state estimation can be control can be viewed as a specialized version of data-driven
used to provide estimates of unmeasured states, where the control [42]–[48].
state estimates serve as ersatz states for the receding-horizon Beyond persistency, since online identification and learning
optimization [20]–[24]. Transient errors in the state estimates, occur during closed-loop operation, the control input is corre-
however, can impede the ability to enforce constraints on un- lated with the measurements due to disturbances and sensor
measured states. Output-feedback MPC that does not rely on a noise. When recursive least squares (RLS) is used for closed-
state estimator is developed in [25]–[30] and [8, Ch. 5]. loop identification, this correlation may obstruct consistency and,
In addition to the lack of knowledge about future commands thus, lead to asymptotic bias in the parameter estimates [49]–
and disturbances, a potential weakness of MPC is the need for a [51]. Alternative identification methods, such as the prediction
sufficiently accurate model for effective predictive optimization. error method (PEM) [52] and instrumental variables (IV) [53],
where x (t) ! R lx is the state, w (t) ! R lw is the disturbance, where k $ 0 is the time step, nt $ 1 is the data window for esti-
y (t) ! R p is the output of G, A ! R lx # lx, B ! R lx # m, C ! R p # lx, mation, u k ! R m is the control, yt k ! R p is the model output,
D ! R p # m, and D 1 ! R lx # lw. The disturbance w(t) is matched if t 0, f, G
and Ft 1, f, Ft nt ! R p # p and G t nt ! R p # m are the coefficient
there exists Ur ! R m # m such that D 1 = BUr ; otherwise, the dis- matrices to be estimated. Based on the structure of (8), the
turbance is unmatched. Note that, if B = D 1, then the distur- 1-step predicted output is given by
bance is matched, and Ur = I. The disturbance w(t) is assumed nt nt
to be piecewise constant within each interval kTs to (k + 1) Ts, y 1|k = - / Ft i, k + 1 y k + 1 - i + G
t 0, k + 1 u 1|k + / G
t i, k + 1 u k + 1 - i, (9)
i=1 i =1
at the value w k, where Ts 2 0. In the case where w(t) is stochas-
tic, the standard deviation of w k is specified. The output y(t) of where u 1|k ! R m is the 1-step computed input, and Ft 1, k + 1, f,
G is corrupted by sensor noise v(t). The sample operation yields t 0, k + 1, f, G
Ft nt , k + 1 ! R p # p and G t nt , k + 1 ! R p # m are the estimated
9 9
y k = y (kTs) + v k, where v k = v (kTs) ! R p is the sampled sensor coefficient matrices at step k. As explained below, the coefficient
noise, and Ts is the sampling time. For each example, the statis- t 0, k + 1, f, G
matrices Ft 1, k + 1, f, Ft nt , k + 1, G t nt , k + 1 are available at step
tics of the sampled sensor noise v k are specified. k. Similarly, using the 1-step predicted output y 1|k and the
The tracking output y t,k ! R p t is defined by 2-step computed control u 2|k, the 2-step predicted output is
9
given by
y t,k = C t y k, (3)
nt
where C t ! R p t # p. The performance objective is to have y t,k fol- y 2|k = - Ft 1, k + 1 y 1|k - / Ft i, k + 1 y k + 2 - i + G
t 0, k + 1 u 2|k
i=2
low a commanded trajectory rk ! R p t, whose future values may nt
or may not be known, and, thus, the command preview may or t 1, k + 1 u 1|k + / G
+G t i, k + 1 u k + 2 - i . (10)
i=2
may not be available.
In addition to the performance objective, the constrained Hence, for all j $ 1, the j-step predicted output is given by
output y c,k ! R p c is defined by j -1 nt j -1
it = vec 6F t nt @ . (14)
9 t0 f G
t 1 f Ft nt G
k
z i (i) z i (it) + 1 (it - i 0)T P 0-1 (it - i 0), (12)
ti T t
J k (it) = / t
i=0
tk tk Defining the regressor matrix z k ! R p # [np (m + p) + mp] by
T his sidebar compares the accuracy of the prediction error shown. IV is implemented using the Matlab function iv4 with a
method (PEM) and instrumental variables (IV) with recur- second-order identification model.
sive least squares (RLS) in open- and closed-loop frameworks.
All of the examples in this sidebar consider the asymptotically Example S1: Open-Loop Identification Using Recursive
stable nonminimum-phase (NMP) plant Least Squares and the Prediction Error Method
Let u k, w k, and v k be zero-mean, Gaussian white noise
4 (z - 1.2) sequences with a standard deviation of one. Figure S1
G (z) = , (S1)
z 2 - z + 0.8 shows the error of the estimate of each coefficient of (S1)
versus the time step. Note that all of the PEM coefficient
with matched disturbance w k . The output y k of (S1) is cor- errors approach zero, which indicates consistency. How-
rupted by sensor noise v k, and the noisy output is used for ever, the RLS estimates are biased, which shows that RLS
identification. The input u k of (S1) is assumed to be known and lacks consistency.
is also used for identification.
For RLS, let nt = 2 and P0 = 1000. PEM is implemented Example S2: Open-Loop Identification Using Recursive
using the Matlab function recursiveBJ with initial parameter Least Squares and Instrumental Variables
covariance 1000, a second-order identification model, and a Let u k, w k, and v k be as in Example S1. Figure S2 shows the
zeroth-order noise model. Numerical tests with higher-order error of the estimate of each coefficient of (S1) versus the time
noise models yielded less accurate results and, thus, are not step. Note that all of the IV coefficient errors approach zero,
|θ2,k – θ2,k|
|θ1,k – θ1,k|
|θ2,k – θ2,k|
10–2 10–2
10–4 10–4
" " " "
|θ4,k – θ4,k|
|θ3,k – θ3,k|
|θ4,k – θ4,k|
10–2 10–2
" " " 10–2 "
10–4 10–4 10–2
10–6 10–6 10–4
RLS RLS
10–8 10–8 PEM 10–6 10–4 IV
10–10 0 10–10 0
10 102 104 106 10 102 104 106 100 101 102 103104 105 100 101 102 103104 105
k (Step) k (Step) k (Step) k (Step)
FIGURE S1 Example S1: Open-loop identification accuracy of FIGURE S2 Example S2: Open-loop identification accuracy of
the prediction error method (PEM) and recursive least squares recursive least squares (RLS) and instrumental variables (IV).
(RLS). Note that PEM coefficient errors approach zero. Note that the IV coefficient errors approach zero.
100
100 104 104
|θ1,k – θ1,k|
|θ2,k – θ2,k|
|θ2,k – θ2,k|
"
10–1
" 102
10–1 " "
100 100
10–2 10–2 10–2
10–2 10–4
10–4
10–6
104 104
|θ3,k – θ3,k|
|θ4,k – θ4,k|
|θ3,k – θ3,k|
|θ4,k – θ4,k|
" " 100 " 102 " 102
100
100 100
10–1 10–2 10–2
RLS IV
10–1 PEM 10–4 10–4 RLS
10–2
100 102 104 106 100 102 104 106 100 101 102 103 104 105 100 101 102 103 104 105
k (Step) k (Step) k (Step) k (Step)
which indicates consistency. However, the RLS estimates are Example S4: Closed-Loop Identification Using Recursive
biased, which shows that RLS lacks consistency. Least Squares and Instrumental Variables
Let u k, w k, and v k be as in Example S3. Figure S4 shows the
Example S3: Closed-Loop Identification error of the estimate of each coefficient of (S1) versus the time
Using Recursive Least Squares and the step. Note that the RLS estimates are biased, and IV produces
Prediction Error Method large estimation errors.
Let w k and v k be zero-mean, Gaussian white noise sequenc- These numerical examples, although limited, demonstrate
es with a standard deviation of 0.1. For closed-loop identifi- that PEM is more accurate than RLS for both open-loop and
cation examples, the input u k is given by a linear quadratic closed-loop identification. Furthermore, although IV showed
Gaussian controller designed using the Matlab command lqg clear advantages over RLS for open-loop identification, its per-
with weights Q xu = Q wv = I 3. Figure S3 shows the error of the formance for closed-loop identification is poor, at least for the
estimate of each coefficient of (S1) versus the time step. Note instruments implemented using the Matlab command iv4. Note
that, after convergence, the PEM coefficient errors are uni- that these closed-loop identification examples are based on a
formly smaller than the RLS coefficient errors; the improve- linear time-invariant controller. The performance of PEM within
ment ranges from 0.71 to 9.05. However, all of the PEM and the context of adaptive control, where the controller is linear time
RLS estimates are biased, which shows that neither method varying, is considered in “Closed-Loop Identification Under Pre-
is consistent for this example. dictive Cost Adaptive Control With the Prediction Error Method.”
T his sidebar investigates the performance of predictive cost output y k is corrupted by sensor noise v k, and the noisy output
adaptive control (PCAC) when the prediction error method is used for identification. The control input u k, which is given by
(PEM) is used instead of recursive least squares (RLS). Nu- receding-horizon optimization, is also used for identification.
merical examples are presented to compare the performance For PCAC/PEM, PEM-based identification is combined
of PCAC/PEM and PCAC/RLS. Note that, unlike the numerical with receding-horizon optimization. PEM is implemented using
examples considered in “Open- and Closed-Loop Identification the Matlab function recursiveBJ with a zeroth-order noise
Using the Prediction Error Method and Instrumental Variables,” model, model order nt , and initial parameter covariance P0.
where the controller is linear time invariant, the PCAC control- Furthermore, the identification model in PEM is initialized at
ler is adaptive. The examples in this sidebar involve discrete- i 0. Both PCAC/PEM and PCAC/RLS use u min = - 1, u max = 1,
time plants with matched disturbances w k . For each plant, the r = 50I , - 1, Pr = 50, R = 10I ,,
Tu min = - 1, Tu max = 1, , = 50, Q
5
100
|θ1,k – θ1,k|
|θ2,k – θ2,k|
10–1
10–4
–5 PCAC/PEM 10–2
1 PCAC/RLS
100
0.5
|θ3,k – θ3,k|
|θ4,k – θ4,k|
–0.5 10–1
PCAC/PEM
10–4 PCAC/RLS
–1
10–2
0 2,000 4,000 6,000 8,000 10,000 100 102 104 100 102 104
k (Step) k (Step) k (Step)
100
2000 100
10–1
|θ1,k – θ1,k|
|θ2,k – θ2,k|
" "
10–2
0 10–1
10–3
10–2 10–4 PCAC/PEM
–2000 PCAC/PEM 102 PCAC/RLS
PCAC/RLS 102
1
101
|θ3,k – θ3,k|
|θ4,k – θ4,k|
100
10–1
–1 10–1 10–2
0 20 40 60 80 100 100 101 102 100 101 102
k (Step) k (Step) k (Step)
4 (z - 1.2)
Example S5: An Asymptotically Stable Plant G (z) = , (S2)
z 2 - z + 1.2
Let G(z) be given by (S1), and let w k and v k be as in Example
S3. Figure S5 shows y k and u k for PCAC/PEM and PCAC/RLS. which is unstable, and let w k and v k be as in Example S3.
Figure S6 shows the error of each coefficient of (S1) versus the Figure S7 shows y k and u k for PCAC/PEM and PCAC/RLS.
time step obtained with PCAC/PEM and PCAC/RLS. Note that Figure S8 shows the error of each coefficient of (S2) versus the
all of the PEM and RLS estimates are biased, which shows that, time step obtained with PEM and RLS operating within PCAC/
for this example, neither method is consistent. However, RLS is PEM and RLS in PCAC. Note that PCAC/PEM fails to stabilize
uniformly faster in reducing the identification errors. the plant.
t k, i ! R p # m is defined by
where, for all i = 1, f, , - 1, H
y ,|k u ,|k
3
1.5 rk rk
2
1 1
yk
yk
0.5 0
0 –1
6
2 4
2
uk
uk
1
0
0 –2
0
log ∆IR log yk – rk
0
–2
–2
–4
–4
0
0
–1
–0.5
–2
–1 –3
–1 0
log ∆DG
log ∆DG
–2 –2
–3
–4 –4
–5 –6
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 2 Example 1: Command following for the discrete-time asymptotically stable single-input, single-output plant (46) using nt = n = 2
t 0,k = 0. (a) The single-step command is rk / 1. The output y k approaches the step command
with a strictly proper model, that is, with G
with a decreasing command-following error, which is 5.6e-6 at k = 60. Note that, although the impulse-response error TIR stops
decreasing after the initial transient, the dc gain estimation error approaches zero, which indicates exigency within the closed-loop
identification; the identification of the dc gain is prioritized by recursive least squares to decrease the command-following error. The
bottom-most plot compares the poles and zero of the identified model at k = 60 to the poles and zero of the plant. (b) The multistep
command rk is given by (47). For each step command, y k approaches the command with a decreasing command-following error, which
is 1.6e-4 at k = 19, 1.2e-4 at k = 39, and 1.6e-4 at k = 60. Note that, although TIR stops decreasing after each transient, the dc gain
estimation error approaches zero. The bottom-most plot compares the poles and zero of the identified model at k = 60 to the poles and
zero of the plant. DDG: the absolute value of the difference between the dc gain of the plant and model; TIR: the L 2-norm difference
between the 30-step impulse response of the plant and model.
3
1.5 rk rk
2
1 1
yk
yk
0.5 0
0 –1
6
2 4
2
uk
uk
1
0
0 –2
0
log ∆IR log yk – rk
0
–2
–4 –2
–6 –4
0
0
–1
–0.5
–2
–1 –3
0 0
log ∆DG
log ∆DG
–2 –2
–4 –4
–6 –6
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 3 Example 1: Command following for the discrete-time asymptotically stable single-input, single-output plant (46) using
nt = 3 2 n = 2 with a strictly proper model. (a) The single-step command is rk / 1. The output y k approaches the step command with a
decreasing command-following error, which is 4.7e-7 at k = 60. (b) The multistep command rk is given by (47). For each step command,
y k approaches the command with a decreasing command-following error, which is 2e-4 at k = 19, 1.3e-4 at k = 39, and 1.7e-4 at k = 60.
Note that, in (b), the estimates of the poles and zeros, including the pole-zero cancellation, are more accurate than in (a) due to the
persistency arising from the changing step command. TDG: the absolute value of the difference between the dc gain of the plant and
model; TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
3
1.5 rk rk
2
1 1
yk
yk
0.5 0
0 –1
6
2 4
2
uk
uk
1
0
0 –2
0
log ∆IR log yk – rk
0
–1
–1
–2
–2
–3
0.2 3
0 2
–0.2 1
–0.4 0
0 0
log ∆DG
log ∆DG
–1 –1
–2 –2
–3
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 4 Example 1: Command following for the discrete-time asymptotically stable single-input, single-output plant (46) using
nt = 1 1 n = 2 with a strictly proper model. (a) The single-step command is rk / 1. The output y k approaches the step command with a
decreasing command-following error, which is 7.8e-4 at k = 60. (b) The multistep command rk is given by (47). For each step command,
y k approaches the command with a decreasing command-following error, which is 5.3e-3 at k = 19, 5.7e-3 at k = 39, and 1.8e-2 at
k = 60. Note that the accuracies of the impulse response and the poles and zeros of the identified model in both (a) and (b) are poor,
due to the fact that nt 1 n. Nevertheless, in both (a) and (b), the estimate of the dc gain is sufficiently accurate to allow predictive cost
adaptive control to approach the step command. TDG: the absolute value of the difference between the dc gain of the plant and model;
TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
2 rk rk
2
1
yk
yk
1
0 0
4 dk dk
4
"
"
2 dk dk
uk
uk
2
0 0
0
log dk – dk log ∆IR log yk – rk
0
–1 –1
–2 –2
–3 –3
0.2 0.2
0 0
–0.2 –0.2
–0.4 –0.4
–0.6 –0.6
" 0 " 0
–1 –1
–2 –2
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 5 Example 1: Command following and disturbance rejection for the discrete-time asymptotically stable single-input, single-output
plant (46) using rk / 1 and nt = n = 2 with a strictly proper model. (a) The single-step disturbance is d k / 0.8. The output y k approaches
the step command with a decreasing command-following error, which is 3e-3 at k = 60. Note that, although TIR stops decreasing after
the initial transient, recursive least squares (RLS) continues to refine the estimate dtk of the disturbance for the remainder of the simula-
tion. This refinement indicates exigency within the closed-loop identification, where the estimation of dtk is prioritized to decrease the
command-following error. The bottom-most plot shows a pole-zero cancellation in the identified model at k = 60. (b) The multistep distur-
bance d k is given by (52). For each step disturbance, y k approaches the command with a decreasing command-following error, which is
1.3e-2 at k = 19, 2.4e-2 at k = 39, and 8.8e-2 at k = 60. Note that, although TIR stops decreasing within a few steps after each change
in the disturbance, RLS continues to refine the estimate dtk of the disturbance while dk is constant. Note that, in both (a) and (b), the step
disturbances induce self-generated persistency, which facilitates the ability of RLS to effectively construct an estimate dtk of the distur-
bance d k, which, in turn, facilitates disturbance rejection. The bottom-most plot shows a pole-zero cancellation in the identified model at
k = 60, due to the step disturbance. TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
2 rk 2 rk
yk
yk
1 1
0 0
4 dk 4 dk
"
"
2 dk 2 dk
uk
uk
0 0
0 0
log dk – dk log ∆IR log yk – rk
–1 –1
–2
–2
–3
0 0
–0.5 –0.5
–1
–1
0
" " 0
–1 –1
–2 –2
–3 –3
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 6 Example 1: Command following and disturbance rejection for the discrete-time asymptotically stable single-input, single-out-
put plant (46) using rk / 1 and nt = 3 2 n = 2 with a strictly proper model. (a) The single-step disturbance is d k / 0.8. The output y k
approaches the step command with a decreasing command-following error, which is 1.3e-3 at k = 60. (b) The multistep disturbance d k
is given by (52). For each step disturbance, y k approaches the command with a decreasing command-following error, which is 7.4e-3
at k = 19, 2e-2 at k = 39, and 7.1e-2 at k = 60. Note that, although the accuracies of the poles and zeros of the identified model in both
(a) and (b) are poor, the estimate of the disturbance is sufficiently accurate to allow predictive cost adaptive control to approach the
command and reject the disturbance. TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
3, k $ 40. and for a given model order nt , the output of the model identi-
fied by RLS at step k is given by
Figures 2–4 show that, for asymptotically stable minimum- nt nt
phase plants, PCAC follows step commands and is robust to y k = - / Ft i, k y k - i + / G
t i, k u k - i . (50)
i =1 i =1
the choice of model order nt .
Next, a matched disturbance d k ! R is applied to the plant Using (50) in (49), the disturbance dt k estimated by RLS is
(46), for which the plant output is given by given by
1.5 rk 2 rk
1
1
yk
yk
0.5
0 0
4 dk 4 dk
"
"
2 dk 2 dk
uk
uk
0 0
log dk – dk log ∆IR log yk – rk
0 0
–0.5 –0.5
–1 –1
–1.5
–1.5
0.4 0.4
0.2 0.2
0 0
" 0 " 0
–1 –1
–2 –2
–3 –3
0 20 40 60 80 100 0 20 40 60 80 100
k (Step) k (Step)
Im (z)
FIGURE 7 Example 1: Command following and disturbance rejection for the discrete-time asymptotically stable single-input, single-out-
put plant (46) using rk / 1 and nt = 1 1 n = 2 with a strictly proper model. (a) The single-step disturbance is d k / 0.8. The output y k
approaches the step command with a decreasing command-following error, which is 2.9e-2 at k = 100. (b) The multistep disturbance
d k is given by (52). For each step disturbance, y k approaches the command with a decreasing command-following error, which is 1.2e-1
at k = 19, 2.1e-1 at k = 39, and 1.7e-1 at k = 100. Note that, in both (a) and (b), y k approaches the command more slowly than in the
cases nt = 2 and nt = 3 considered in Figures 5 and 6, respectively. TIR: the L 2-norm difference between the 30-step impulse response
of the plant and model.
i=1 i =1 i =1 i =1 i =1
minimum-phase plant (46), PCAC rejects step disturbances
(51) with nt $ n. Note that self-generated persistency is evident in
the control input arising from the step-command changes and
Figures 5–7 show the response of PCAC for nt = 2, nt = 3, step disturbances. Furthermore, it is seen that, although the
and nt = 1, respectively, with constant command rk / 1, impulse-response error TIR stops decreasing after each tran-
where the error of the disturbance estimate dt k is comput- sient, RLS keeps refining the dc gain of the model for the re-
ed. In Figures 5(a), 6(a), and 7(a), the constant disturbance mainder of the simulation. This indicates exigency within
d k / 0.8 is applied, and in Figures 5(b), 6(b), and 7(b), the identification, as RLS prioritizes the identification of the dc gain
three-step disturbance over the impulse response to achieve command following. G
Example 1 showed that PCAC can follow step commands
0.8, 0 # k 1 20, and reject step disturbances. This ability is further investigated
d k = * -0.4, 20 # k 1 40, (52) within the context of fixed-gain integral control in “Following
1.2, k $ 40 Step Commands Without an Integrator.”
S ince receding-horizon optimization determines control in- pared with linear time-invariant integral control as well as with
puts without using an explicit feedback controller, it is de- PCAC using a fixed model with erroneous dc gain. To do this,
sirable to compare the performance of predictive cost adaptive consider the discrete-time plants
control (PCAC) with the performance of linear time-invariant z - 0.5
G (z) = , (S3)
controllers. In particular, within the context of linear time-invari- (z - 0.2) (z - 0.8)
ant control, asymptotic step-command following is achievable 0.1 (z - 0.5)
G (z) = , (S4)
if and only if the controller or plant includes an integrator. To (z - 0.2) (z - 0.8)
5 (z - 0.5)
investigate asymptotic step-command following under PCAC, G (z) = , (S5)
(z - 0.2) (z - 0.8)
the performance of PCAC using online identification is com-
0.1953 (z - 0.2)
G (z) = , (S6)
(z - 0.5) (z - 0.9)
rk uk yk
Gc (z) G (z) which are asymptotically stable and have dc gains 3.125, 0.3125,
− 15.625, and 3.125, respectively. Note that (S3)–(S5) have the
same poles and zeros and that (S3) and (S6) have the same dc
gains but different poles and zeros. Disturbance and sensor noise
FIGURE S9 Basic servo loop with step command rk and integral are not considered in these examples. A strictly proper model is
controller G c. t 0,k = 0 is enforced.
used for identification in PCAC; that is, G
0 tegral controller
–2 G c (z) = 0.1 . (S7)
3 z-1
2 Figure S10 shows the closed-loop response with the controller
uk
0
S3 thus, the command-following error converges to zero. On the
–2 S4 other hand, the closed-loop dynamics are unstable for G(z)
–4 S5
S6 given by (S5), and, thus, y k diverges.
0 10 20 30 40 50 60 70 80 90 100
k (Step) Example S8: Predictive Cost Adaptive
Control With Fixed Models
FIGURE S10 Example S7: Linear time-invariant integral control for Let the true plant G(z) be given by (S3). PCAC is applied with the fixed
G(z) given by (S4)–(S6) with the controller (S7). For plants (S3),
models (S3)–(S6) with u min = - 1, u max = 1, Tu min = - 1, Tu max = 1,
(S4), and (S6), the closed-loop system is asymptotically stable,
, = 50, Q r = I , - 1, Pr = 1, and R = I ,. Figure S11 shows that the
and, thus, asymptotic step-command following is achieved. For
the plant (S5), the closed-loop system is unstable. command-following error approaches zero in the cases where
3
2 rk 1
yk
1 0.5
yk
0 rk
0
1 1
0.5
uk
0.5 0
uk
S4
S5
Log |∆DG| Log |yk – rk|
0 0 S6
–1 10–2.12n × 1
"
0
Log |yk – rk|
–5 S3 –2
S4 –3
–10 S5
S6 1
–15 0
0 5 10 15 20 25 30 35 40 45 50
–1
k (Step) –2
FIGURE S11 Example S8: Step-command following using pre- 0 5 10 15 20 25 30
dictive cost adaptive control with fixed models (S3)–(S6). k (Step)
Asymptotic command following is achieved in the cases (S3)
and (S6), where the dc gain of the model is the same as the dc
gain of the plant. Note that asymptotic command following is FIGURE S12 Example S9: Step-command following using pre-
achieved in the case of (S6) despite the fact that all of the poles dictive cost adaptive control with the initial models (S4)–(S6)
and zeros of the fixed model are incorrect relative to the true as well as the initial model given by i 0 = 10 -2 1 2nt # 1. For all four
plant (S3). For (S4) and (S5), where the dc gain of the model is initial models, the command-following error approaches zero.
not equal to the dc gain of the plant, the command-following TDG: the absolute value of the difference between the dc gain
error remains bounded but does not approach zero. of the plant and model.
PCAC uses the fixed models (S3) and (S6), both of which have Pr = 1, R = I ,, m = 1, nt = 2, and P0 = 10 3 I 2nt for the initial models
the correct dc gain. However, in the case where PCAC uses (S4)–(S6) as well as the initial model given by i 0 = 10 -2 1 2nt # 1.
the fixed models (S4) and (S5), both of which have an incor- Figure S12 shows that, for all four initial models, the
rect dc gain, the command-following error remains bounded command-following error approaches zero. Furthermore,
but does not approach zero. the dc gain of the identified model approaches the dc
gain of (S3).
Example S9: Predictive Cost Adaptive Control With Various These examples demonstrate that using (TU 1; k, ,) T RTU 1; k, ,
Initial Models That Are Different From the True Plant in (41) does not yield integral action, and, in addition, that esti-
Let the true plant G(z) be given by (S3). PCAC is applied mation of the dc gain of the plant is necessary for asymptotic
r = I , - 1,
with u min = - 1, u max = 1, Tu min = - 1, Tu max = 1, , = 50, Q step-command following.
20 rk 10 rk
0 0
α = –10 α = –10
yk
yk
–20 –10
α = –1 α = –1
–40 α = –0.1 α = –0.1
–20
α = 0.1 α = 0.1
20 α=1 10 α=1
α = 10 α = 10
0
uk
uk
0
–20 –10
log yk – rk
0 1
0
–2
–1
–4
15 15
log ∆IR
log ∆IR
10 10
5 5
0 0
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
Im (z)
FIGURE 8 Example 2: Command following for the discrete-time unstable nonminimum-phase single-input, single-output plant (53) using
rk / 1 and nt = n = 2 with a strictly proper finite-impulse response model initialized with i 0 = a [0 0 0 1] T, a ! {- 10, - 1, - 0.1, 0.1, 1, 10}.
(a) The single-step command is rk / 1. The output y k approaches the step command with a decreasing command-following error, which
is in the interval [2.1e-3, 1.0e-2] at k = 60. Note that, for all a, TIR approaches asymptotic values for k $ 25. The bottom-most plot com-
pares the poles and zero of the identified model at k = 60 to the poles and zero of the plant. (b) The multistep command rk is given by
(47). For a ! - 10, y k approaches the three-step command with a decreasing command-following error, which is in the interval [4.3e-1, 1.1]
at k = 19, [3.3e-2, 6.7e-2] at k = 39 , and [8e-2, 1.2e-1] at k = 60. For a = - 10, y k approaches the last command with a decreasing com-
mand-following error, which is 1.3e-1 at k = 60. Note that, for all a, TIR approaches asymptotic values for k $ 25. The bottom-most plot
compares the poles and zero of the identified model at k = 60 to the poles and zero of the plant. TIR: the L 2-norm difference between
the 30-step impulse response of the plant and model.
0 20
0
yk
yk
–5 –20
rk rk
–10 –40
2
20
0
–2 0
uk
uk
–4
–20
0
log yk – rk
0
–2 –2
–4 –4
–6 –6
10 20
log ∆IR
log ∆IR
5 10
0 0
0 10 20 30 40 50 60 0 10 20 30 40 50 60
k (Step) k (Step)
0.1 PlantPole
Plant Pole 0.1 PlantPole
Plant Pole
PlantZero
Plant Zero PlantZero
Plant Zero
Im (z)
Im (z)
0 ModelPole
Pole 0 ModelPole
Pole
Model Model
ModelZero
Model Zero ModelZero
Model Zero
–0.1 –0.1
0 0.5 1 1.5 2 2.5 0 0.5 1 1.5 2 2.5
Re (z) Re (z)
(a) (b)
FIGURE 9 Example 2: Command following for the discrete-time unstable nonminimum-phase single-input, single-output plant (53) using
rk / 1 and nt = n = 2 with a strictly proper model, which is initialized with 100 Gaussian-distributed samples of i 0 with a standard devia-
tion of v. (a) For v = 1, the output y k approaches the step command with a decreasing command-following error, which is in the inter-
val [9.5e-4, 5.5e-3] at k = 60. (b) For v = 2, the output y k approaches the step command with a decreasing command-following error,
which is in the interval [6.7e-3, 2.1e-2] at k = 60. Note that, in both (a) and (b), although TIR approaches an asymptotic value for k $ 25,
indicating the presence of bias and, thus, lack of consistency in the estimate of the model, the poles and zero of the model are close to
the poles and zero of the plant at k = 60. TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
20
5 rk rk
0
yn,k
yn,k
0
" " "
n=2
–2 uk –5
–10
–4
log ∆IR log yn,k – rk
1
0 0
–2 –1
–2
–4 –3
2 3
1 2
0 1
0
–1
0 20 40 60 80 100 120 0 20 40 60 80 100 120
k (Step) k (Step)
Re (z) Re (z)
–1 –0.5 0 0.5 1 1.5 2 2.5 –1 –0.5 0 0.5 1 1.5 2 2.5
0.5 0.5
Im (z)
Im (z)
0 0
–0.5 –0.5
0.01 0.01
Im (z)
Im (z)
0 0
–0.01 –0.01
1 1.1 1.2 1.3 1.4 1.5 1 1.1 1.2 1.3 1.4 1.5
Re (z) Re (z)
"
"
"
"
"
"
"
"
"
(a) (b)
FIGURE 10 Example 2: Command following for the discrete-time unstable nonminimum-phase single-input, single-output plant (53) using
rk / 1; noisy measurement with a noise standard deviation of v; and finite-impulse response strictly proper initial model i 0 = [0 T(2nt - 1) # 1 1] T,
where nt = 2, nt = 3, and nt = 4. (a) For v = 0.05, in all cases, y k approaches the step command with a decreasing command-following
error. (b) For v = 0.15, y k approaches the step command with a decreasing command-following error for nt = 3 and nt = 4. For nt = 2, y k
does not approach the step command, and the command-following error is 7.9 at k = 120. Note that, in both (a) and (b), larger values of nt
yield smaller values of TIR. The bottom-most plots show that larger values of nt facilitate identification, and the poles and zero of the model
are closer to the poles and zero of the plant. TIR: the L 2-norm difference between the 30-step impulse response of the plant and model.
10 30
n=2 n=2
" " " " "
25
8 n=3 n=3
n=4 20 n=4
n=5 n=5
6
log (∆IR)
n=6 15 n=6
∆IR
4 10
5
2
0
0 –5
0 0.02 0.04 0.06 0.08 0.1 0 0.02 0.04 0.06 0.08 0.1
σ σ
(a) (b)
FIGURE 11 Example 2: Accuracy of the identified model versus the standard deviation v of the sensor noise for the discrete-time
unstable nonminimum-phase single-input, single-output plant (53) with rk / 1 and with the controllers given by predictive cost adaptive
control (PCAC) and linear quadratic Gaussian (LQG) with an integrator. The model accuracy is determined by TIR at k = 140 averaged
over 100 simulations. PCAC uses a strictly proper finite-impulse response initial model, and LQG with an integrator uses (53) to compute
the control gain with output and control weights Q = 2 and R = 1, respectively. (a) PCAC. For each v, larger values of nt yield smaller
values of TIR. Therefore, for each sensor-noise level, the identified model is more accurate for larger values of nt . (b) LQG with an inte-
grator. For each nt , larger values of v yield larger values of TIR. Note that, unlike (a), for all v $ 0.05, an increase in the model order nt
results in a less accurate model. In addition, for each value of nt , the models in (b) are less accurate than those in (a). TIR: the L 2-norm
difference between the 30-step impulse response of the plant and model.
4 rk 4 rk
2 2
0
yk
yk
0
–2
–4 –2
20 40
20
0
uk
uk
0
–20 –20
1
1
log yk – rk
0 0
–1 –1
–2 –2
–3
–3
2
2
log ∆IR
log ∆IR
1 1
0 0
–1 –1
0 50 100 150 200 250 300 0 50 100 150 200 250 300
k (Step) t (s)
Im (z)
FIGURE 12 Example 3: Command following for the unstably stabilizable single-input, single-output plants (54) and (56) using nt = n = 3
and the three-step command (55). (a) For the discrete-time plant (54), the output y k approaches each step command with a decreasing
command-following error. The bottom-most plot compares the poles and zeros of the identified model at k = 300 to the poles and zeros
of the plant. It was observed (not shown) that the response time was larger than 100 steps for 32 # , # 40, and PCAC was unable to
follow the step commands for , # 31. For , 2 85, it was observed (not shown) that the output diverged. (b) For the continuous-time plant
(56), the output y(t) approaches each step command with a decreasing command-following error. The bottom-most plot compares the
poles and zeros of the identified model at t = 300 s to the poles and zeros of the exact discretization of the continuous-time plant (56).
Note that the discretized plant is unstably stabilizable. TIR: the L 2-norm difference between the 30-step impulse response of the plant
and model.
2 2
0 0
Open Loop Open Loop
y (t )
y (t )
uk
–2 –2
0 100 200 300 400 500 0 100 200 300 400 500
t (s) t (s)
0 0
Gain (dB)
Gain (dB)
–20 –20
–40 –40
Plant Plant
0 n=4 0
n=5 n=5
Phase (°)
Phase (°)
0 π /4 π /2 3π /4 π 0 π /4 π /2 3π /4 π
θ (rad/Sample) θ (rad/Sample)
1 1
Im (z)
Im (z)
0 0
–1 –1
–1 –0.5 0 0.5 1 1.5 2 2.5 –1 –0.5 0 0.5 1 1.5 2 2.5
Re (z ) Re (z)
"
"
"
"
"
"
"
"
"
(a) (b)
FIGURE 13 Example 4: Matched disturbance rejection for the continuous-time asymptotically stable single-input, single-output plant (59)
and harmonic disturbance (60) with nt = n = 4, nt = 5, and nt = 6. (a) Zero initial conditions. Note that larger values of nt yield a better
performance and that u k approximately cancels d(t) for nt = 6. Furthermore, larger values of nt yield a more accurate frequency response
of the identified models at t = 500 s. For nt = 6, the identified model has an approximate pole-zero cancellation at the disturbance fre-
quency i d = r/15 rad/sample, which is denoted by the vertical dash–dotted line; the spurious pole in the identified model allows
receding-horizon optimization to predict the future harmonic response of the plant, which facilitates the optimization of the future control
inputs. This pole-zero cancellation serves as an implicit internal model of the unknown disturbance. The bottom-most plot compares the
poles and zeros of the identified model at t = 500 s to the poles and zeros of the exact discretization of the plant. For nt = 6, the approx-
imate pole-zero cancellations on the unit circle occur at the frequency i d of the harmonic disturbance. (b) Nonzero initial conditions
x (0) = [10 - 5 - 8 300] T for the state-space realization (61), (62). Compared to the case of zero initial conditions, the transient
response is more pronounced, and the control saturation occurs over a longer interval of time. Moreover, since nonzero initial conditions
result in more persistency during the transient, the identified models in (b) are more accurate than those in (a).
1 200
y (t )
0
y (t )
0
Open Loop Open Loop
–1 –200
" "
n=4
" " "
n=4 –400
n=5 n=5
2 dk 2 dk
"
n=6 n=6
0 0
uk
uk
–2 –2
0 1000 2000 3000 4000 5000 0 1000 2000 3000 4000 5000
t (s) t (s)
0 0
Gain (dB)
Gain (dB)
n=4
" " "
n=4 n=5
360 0
n=5 n=6
180
Phase (°)
Phase (°)
n=6 –180
0
–180 –360
–360
0 0
log (PSD)
log (PSD)
–2 –2
–4 –4
–6 –6
0 π /4 π /2 3π /4 π 0 π /4 π /2 3π /4 π
θd (rad/Sample) θd (rad/Sample)
(a) (b)
FIGURE 14 Example 4: Matched disturbance rejection for the continuous-time asymptotically stable single-input, single-output plant (59)
and zero-mean, Gaussian white noise sequence d k with a standard deviation of 0.5 for nt = n = 4, nt = 5, and nt = 6. (a) Zero initial con-
ditions. Note that larger values of nt yield more accurate magnitudes of the frequency response of the identified models at t = 5000 s.
The bottom-most plot shows that, for all nt , predictive cost adaptive control (PCAC) reduces the two peaks of the open-loop power
spectral density (PSD) of y k, which indicates that PCAC rejects the broadband disturbance d k . (b) Nonzero initial conditions
x (0) = [10 - 5 - 8 300] T for the state-space realization (61) and (62). As in (a), the bottom-most plot shows that, for all nt , PCAC
reduces the two peaks of the open-loop PSD of y k . Note that, for all nt , the frequency responses of the identified models at t = 5000 s
are more accurate in (b) than in (a) due to the nonzero initial conditions.
200 600
rk rk
400
200
y (t )
y (t )
0
0
No Slack –200 S = 0.1
–200 S = 0.1 S = 10
20 S = 10 S = 1000
S = 1000 10 S = 5000
10
0
uk
uk
0
–10
–10
40 8000
6000
εk∞
εk∞
20 4000
0 2000
0
log ∆IR
log ∆IR
2 2
0 0
–2
–2
0 20 40 60 80 100 120 0 50 100 150 200
t (s) t (s)
Im (z )
FIGURE 15 Example 5: Command following with the output constraint - 20 # y k # 20 and continuous-time unstable single-input, single-
output plant (63) for various values of S. (a) Predictive cost adaptive control (PCAC) is applied using (37)–(40) and (41)–(45) with
nt = n = 4, g = 0.7, x (0) = [1 - 3 2 0.5] T, and the three-step command (67). At t = 9 s, (37)–(40) become infeasible, and the three-step
command cannot be followed. In contrast, with (41)–(45), y(t) approaches the first two step commands, and the slack is activated at
t = 8 s, t = 50 s, and t = 80 s, as can be seen in the third plot. Since the last step command is not achievable, y(t) moves further away
from the command as S increases, where the command-following error at t = 120 s is approximately 0.02 for S = 0.1, 1.00 for S = 10,
and 4.81 for S = 1000. Furthermore, note that the overshoot at t = 19 s violates the output constraint due to the poor initial model and
nonzero initial condition. The distance between y(t) and the constraint at t = 19 s for S = 0.1 is 41.47, which is reduced by 47% for
S = 10 and 89% for S = 1000. (b) PCAC is applied using (41)–(45) with nt = 6 2 n = 4, g = 0.01, x (0) = [10 - 5 8 3] T, and the three-
step command (68). Compared to (a), note that, for all S, the output-constraint violation is more severe during the first transient. How-
ever, as the model becomes more accurate, y(t) approaches the first and second step commands, where the constraint is enforced after
t = 106 s. For the third step command, as in (a), y(t) moves away from the step command as S increases. TIR: the L 2-norm difference
between the 30-step impulse response of the plant and model.
0 -20, k $ 80.
Figure 18 shows that, for the unstable MIMO plant (71), (72),
–0.5 and (74) with NMP TZ and NMP CZ, PCAC stabilizes the plant
–1 0 1 2 3
Re (s) while asymptotically satisfying the output constraints, and the
output approaches the step commands. G
TZ From u to y~
TZ From u to z~
LINEAR TIME-VARYING EXAMPLES
(a) This section considers linear time-varying SISO plants, where
2 y t, k = y c, k = y k .
0
The goal of this example is to investigate the applicability of
–1 PCAC with VRF on time-varying plants. Consider a discrete-
time asymptotically stable SISO plant (46) that abruptly changes
–2 at step k c to a discrete-time unstable NMP SISO plant (53). Note
–10 –5 0 5
Re (s) that the orders of both plants are n = 2 with relative degree one.
Let u min = -50, u max = 50, Tu min = -25, Tu max = 25, , = 20,
CZ From u1 to y~1 CZ From u2 to y~1 Qr = 2I , - 1, Pr = 5, R = I ,, nt = n = 2, and P0 = 10 3 I 2nt . No output
CZ From u to y~
1 2 CZ From u to y~
2 2
constraint is considered in this example. The plant is initialized
CZ From u1 to z~ CZ From u2 to z~ with y -1 = -0.2, y -2 = 0.4, and u -1 = u -2 = 0. PCAC uses the
strictly proper FIR initial model i 0 = 60 1 # 3 1@ T and uses (37)–
(b)
(40) with the measurement y k . Let v k = 0.
Figure 20 shows the response of PCAC for rk / 1 using CRF
FIGURE 16 Example 6: Transmission zeros (TZs) and channel with m = 1 and VRF (21) with h = 0.9, x n = 5, and x d = 10 for
zeros (CZs) of the continuous-time asymptotically stable multiple- k c = 30, k c = 200, k c = 400, and k c = 600. Figure 20(a) shows
input, multiple-output plant (71)–(73). (a) The transfer function
from u to [yu T zu T] T has two nonminimum-phase (NMP) TZs at that, in the absence of forgetting, that is, m = 1, reidentification
approximately z t = 0.94. (b) Each transfer function has one NMP of the plant after the abrupt change is poor, and the output does
CZ at approximately 0.94. not approach the step command. In contrast, Figure 20(b)
z (t )
0
~
Open Loop
–10
"
n=3
"
n=4
20
"
n=6
y (t )
0
~
–20
0
uk
–2
0 500 1000 1500 2000 2500 3000 3500 4000 4500 5000
t (s)
2
log (PSD(z ))
~
0
–2
–4
0 π /4 π /2 3π /4 π
θ d (rad/Sample)
40
Gain (dB)
Gain (dB)
40
u2 to y~1
20 Plant
u1 to y~1
Plant
0
20
–20
360 180
Phase (°)
Phase (°)
u2 to y~1
u1 to y~1
180
0
0
–180 –180
20
Gain (dB)
20
Gain (dB)
0
u2 to y~2
u1 to y~2
–20 0
–40 –20
–60 –40
–80
0 720
Phase (°)
Phase (°)
u2 to y~2
u1 to y~2
360
–360
0
–720 –360
Gain (dB)
20
Gain (dB)
20
u2 to z~
u1 to z~
0
10 –20
–40
0 –60
0 360
Phase (°)
Phase (°)
u2 to z~
180
u1 to z~
–180 0
–180
–360 –360
0 π /4 π /2 3π /4 π 0 π /4 π /2 3π /4 π
θ d (rad/Sample) θ d (rad/Sample)
FIGURE 17 Example 6: Unmatched broadband disturbance rejection for the continuous-time asymptotically stable multiple-input, multiple-
output plant (71)–(73) with a nonminimum-phase (NMP) transmission zero and an NMP channel zero for nt = 3, nt = 4, and nt = 6 using the
discrete-time broadband disturbance w k = [w 1, k w 2, k] T ! R 2, where w 1,k and w 2,k are zero-mean, Gaussian white noise sequences with 1.7
and 2.5, respectively. Note that, for all nt , the magnitude of the open-loop power spectral density (PSD) of zu in the range r/16 # i d # r/4
is reduced, which indicates suppression of the broadband disturbance w k . For all six channels, namely, from u 1 and u 2 to yu 1, yu 2, and zu ,
the 12 bottom-most plots compare the frequency response of the identified model at t = 5000 s to the frequency response of the plant. As
in Example 4, this example shows that a larger model order nt yields more accurate estimates of the plant dynamics.
0 TZ From u to z~
Example 9: Abrupt Change From an Asymptotically
Stable Plant to an Unstable Nonminimum-Phase
–0.5 Plant With Sensor Noise
–1 0 1 2 3
Re (s) The goal of this example is to investigate the effect of zero-
(a) mean, Gaussian white sensor noise on VRF. This example uses
the same setup as Example 8 except that PCAC now uses noisy
2
measurements with the command profile
1 Z
] 20, 0 # k 1 100,
] -20, 100 # k 1 200,
Im (s)
0 rk = [ (76)
]] 3, 200 # k 1 400,
–1 -50, k $ 400.
\
–2 Figure 21 shows the response of PCAC for two different levels
–10 –5 0 5
of sensor noise using k c = 300 and VRF (21) with h = 0.5,
Re (s)
x d = 5x n, and various values of x n . As shown in Figure 21(a)
and (b), as x n increases, the duration during which forgetting is
CZ From u1 to y~1 CZ From u2 to y~1
active increases. Hence, as x n increases, reidentification be-
CZ From u to y~
1 2 CZ From u to y~
2 2
comes slower and, for large values of x n, the output does not
CZ From u1 to z~1 CZ From u1 to z~2
approach the command. Note that, as x n increases, the VRF m k
CZ From u2 to z~1 CZ From u2 to z~2
becomes less sensitive to sensor noise.
(b) Note that the speed of convergence of the coefficients after
the abrupt change determines whether or not the output ap-
FIGURE 18 Example 7: The transmission zero (TZ) and channel proaches the command. In particular, with a large choice of x n,
zero (CZ) of the continuous-time unstable multiple-input, multiple- the reconvergence of the model coefficients is slow, which
output plant (74). (a) The transfer function from u to [yu zu T] T has yields a divergence of the output. Furthermore, in the presence
four nonminimum-phase (NMP) TZs at approximately 0.00, 0.94,
and 3.33. (b) The eight transfer functions have 21 NMP CZs at of zero-mean, Gaussian white sensor noise, where x d = 5x n,
approximately 0.00, 0.13 ! 1.42., 0.14 ! 1.18., 0.15 ! 0.99., 0.19, increasing the value of x n reduces the activation of forgetting
0.95, 2.10, and 3.33. due to sensor noise. G
–20
–40 S = 0.1 I2
40 S = 10 I2
20 S = 1000 I2
z2(t )
0
~
–20
100
0
y (t )
~
–100
10
0
uk
–10
6000 40
20
εk∞
4000
0
2000 10 15 20 25
0
0 10 20 30 40 50 60
t (s)
0.5 0.5
u1 to y~1
u2 to y~1
Im (z)
Im (z)
0 0
–0.5 –0.5
0.5 0.5
u1 to y~2
u2 to y~2
Im (z)
Im( z)
0 0
–0.5 –0.5
0.5 0.5
u1 to z~1
u2 to z~1
Im (z)
Im (z)
0 0
–0.5 –0.5
0.5 0.5
u1 to z~2
u2 to z~2
Im (z)
Im (z)
0 0
–0.5 –0.5
–2 –1 0 1 2 –2 –1 0 1 2
Re (z) Re (z)
FIGURE 19 Example 7: Command following for the continuous-time unstable multiple-input, multiple-output plant (74) with the nonmini-
mum-phase (NMP) transmission zero and NMP channel zero for nt = 2 and various values of S using the three-step command (75). For
each S, the tracking output approaches the three-step command. Note that, as S increases, the output-constraint violation becomes
less severe. Due to the poor initial model, the output constraint is relaxed by the slack f k until t = 20 s. For all t $ 20 s, note that, for
each S, as the model becomes more accurate and zu 1 (t) approaches zero, the output constraint is enforced. The eight bottom-most plots
compare the poles and zeros of the identified model at t = 60 s to the poles and zeros of the plant.
50 rk
0
50
yk
rk –50 kc = 30
0 –100 kc = 200
kc = 400
yk
–50 50
–100 kc = 30 kc = 600
kc = 200 uk 0
50 kc = 400
kc = 600 –50
0
log ∆IR log yk – rk
uk
2
0
–50 –2
4 –4
log ∆IR log yk – rk
2 –6
0 10
–2
–4 5
–6 0
10
5 1
Jk (Θk+1)
0 0.5
0
20 1
Jk (Θk+1)
10
λk
0.8
0
0 100 200 300 400 500 600 0 100 200 300 400 500 600 700
k (Step) k (Step)
k = kc – 1 k = 650 k = kc – 1 k = 700
0.5 0.5
0.5 0.5
Im (z )
Im (z )
Im (z )
Im (z )
0 0 0 0
–0.5 –0.5 –0.5 –0.5
–0.5 0 0.5 0.2 0.4 0.6 0.8 1 1.2 1.4 –0.5 0 0.5 0.2 0.4 0.6 0.8 1 1.2 1.4
Re (z) Re (z ) Re (z ) Re (z )
Plant Pole Plant Zero Model Pole Model Zero Unit Circle
(a) (b)
FIGURE 20 Example 8: Command following with dynamics that abruptly change at step k c represented by the dashed vertical line for
rk / 1 and various values of k c. At step k c, the discrete-time asymptotically stable minimum-phase single-input, single-output plant (46)
changes to the discrete-time unstable nonminimum-phase plant (53). (a) Constant-rate forgetting with m = 1. Note that, as for larger
values of k c, the command-following error after the change is larger. In particular, for k c = 600, the slow reidentification after the abrupt
change causes the control input to saturate and the output error to diverge. Note that the maximum value of the recursive least squares
(RLS) cost J k (i k + 1) (12) increases as k c increases. This is due to the fact that m = 1, and, thus, RLS does not discount data from before
the abrupt change. The two bottom-most plots compare the poles and zeros of the identified model with the poles and zeros of the plant
at k = k c - 1 and k = 650. (b) Variable-rate forgetting (VRF) (21) with h = 0.9, x n = 5, and x d = 10. In contrast with (a), for each k c, y k
approaches the command after the abrupt change. The sixth plot from the top shows that the VRF factor m k is activated near k c. TIR:
the L 2-norm difference between the 30-step impulse response of the plant and model.
50 50
rk rk
0 0
–50 –50
yk
yk
tn = 1 tn = 1
–100 tn = 5 –100 tn = 5
tn = 20 tn = 15
50 tn = 25 50 tn = 20
0 0
uk
uk
–50 –50
log ∆IR log yk – rk
6 6
4 4
2 2
0 0
–2 –2
–4 –4
10 10
5 5
0 0
15 15
Jk (Θk+1)
10 10
5 5
0 0
1 1
0.8 0.8
λk
λk
0.6 0.6
0 100 200 300 400 500 600 700 0 100 200 300 400 500 600 700
k (Step) k (Step)
k = kc – 1 k = 700 k = kc – 1 k = 700
0.5 0.5 0.5 0.5
Im (z )
Im (z )
Im (z )
Im (z )
0 0 0 0
–0.5 –0.5 –0.5 –0.5
–0.5 0 0.5 0.2 0.4 0.6 0.8 1 1.2 1.4 –0.5 0 0.5 0.2 0.4 0.6 0.8 1 1.2 1.4
Re (z ) Re (z ) Re (z ) Re (z )
Plant Pole Plant Zero Model Pole Model Zero Unit Circle
(a) (b)
FIGURE 21 Example 9: Command following with dynamics that abruptly change from (46) to (53) at step k c = 300 with sensor noise.
Predictive cost adaptive control uses the command profile (76) and variable-rate forgetting (VRF) (21) with h = 0.5, x d = 5x n, and various
values of x n. (a) The sensor noise v k is a zero-mean, Gaussian white noise sequence with a standard deviation of 0.1. Note that, before
the abrupt change and for all values of x n, y k approaches each step command. After the abrupt change, however, y k approaches the
command for x n = 1, x n = 5, and x n = 20, and it diverges for x n = 25. Furthermore, for x n = 25, the impulse-response error and recur-
sive least squares (RLS) cost diverge approximately at k = 430. Note that, as x n increases, the VRF factor m k becomes less sensitive
to sensor noise and is activated only at the abrupt change and at each step-command change. (b) The sensor noise v k is a zero-mean,
Gaussian white noise sequence with a standard deviation of 0.3. Note that, compared to (a), since the sensor-noise level is larger, the
value of x n for which the output, impulse-response error, and RLS cost diverge is lower; namely, x n = 20. TIR: the L 2-norm difference
between the 30-step impulse response of the plant and model.
400
200 rk 1000 rk
0
y(t)
y(t)
–200 0
"
n=2
"
n=2
" "
n=4
" "
n=4
–400 –1000
"
n=3
"
n=5 n=3 n=5
2 2
0 0
uk
uk
–2 –2
Jk (Θk+1) log ∆IR log yk – rk
λk
0.6 0.6
0 100 200 300 400 500 600 0 100 200 300 400 500 600
t (s) t (s)
t = (tc – 1) s t = 600 s t = (tc – 1) s t = 600 s
1 1 1 1
Im (z)
Im (z)
Im (z)
Im (z)
0 0 0 0
–1 –1 –1 –1
–1 0 1 –2 –1 0 1 –1 0 1 –2 –1 0 1
Re (z) Re (z) Re (z) Re (z)
Plant Pole Plant Zero Model Pole Model Zero Unit Circle
(a) (b)
FIGURE 22 Example 10: Command following with dynamics that abruptly change from (77) to (63) at 300 s with sensor noise and an
output constraint for various values of nt . Predictive cost adaptive control uses the command profile (76) and variable-rate forgetting (21)
with h = 0.5, x n = 25, and x d = 125. (a) The sensor noise v k is a zero-mean, Gaussian white noise sequence with a standard deviation
of 0.01. Note that, before the abrupt change occurs and for each value of nt , y (t) approaches the step commands, and the output con-
straint is satisfied at all times. In the underparameterized cases nt 1 n = 4, after the abrupt change occurs, the double integrator and
lightly damped poles are poorly identified by recursive least squares (RLS), causing y(t) to oscillate and violate the constraint until the
end of the simulation. In contrast, in the exact and overparameterized cases nt $ n = 4 , after the abrupt change occurs, y(t) only tempo-
rarily violates the constraint during reidentification Additionally, after t = 338 s, y(t) is more damped than the underparameterized cases,
and the output constraint is enforced at all times. Note that the last segment of y(t) is infeasible. (b) The sensor noise v k is a zero-mean,
Gaussian white noise sequence with a standard deviation of 0.1. Note that, due to the higher sensor-noise level, the oscillation amplitude
and output-constraint violation for nt 1 n = 4 are more pronounced in (b) than in (a). Furthermore, note that, in both (a) and (b), for nt = 2,
RLS is able to capture the two integrators of the plant, whereas, for nt = 3, the three poles of the model are unable to capture the two
integrators and two undamped poles of the plant, causing the identification and transient response to be poorer. TIR: the L 2-norm dif-
ference between the 30-step impulse response of the plant and model.