Chapter 8: Learning-based Adaptive Control and Contraction Theory for
8.1 Learning-based Adaptive Control
In this section, we consider the following smooth nonlinear system with an uncertain parameterπ βRπ:
Β€
π₯ = π(π₯ , π) +π΅(π₯ , π)π’, π₯Β€π = π(π₯π, π) +π΅(π₯π, π)π’π (8.1) where π₯, π’, π₯π, and π’π are as defined in (6.1) and (6.3), and π : RπΓ Rπ β¦β Rπ andπ΅:RπΓRπ β¦βRπΓπare known smooth functions with the uncertain parameter π βRπ. Due to Theorems 2.4 and 5.2, we can see that the robust control techniques presented earlier in Theorems 4.2, 4.6, 6.4, and 6.5 are still useful in this case if the modeling errors β₯π(π₯ , π) β π(π₯ , ππ) β₯andβ₯π΅(π₯ , π) βπ΅(π₯ , ππ) β₯are bounded, where ππis a nominal guess of the true parameter π. However, there are situations where such assumptions are not necessarily true.
We present a method of deep learning-based adaptive control for nonlinear systems with parametric uncertainty, thereby further improving the real-time performance of robust control in Theorems 5.2 and 5.3 for model-based systems, and Theorems 8.3 and 8.4 for model-free systems. Although we consider continuous-time dynamical systems in this chapter, discrete-time changes could be incorporated in this frame- work using [1], [2]. Also, note that the techniques in this chapter [3] can be used with differential state feedback frameworks [4]β[9] as described in Theorem 4.6, trading off added computational cost for generality (see Table 3.1 and [10], [11]).
8.1.I Adaptive Control with CV-STEM and NCM Let us start with the following simple case.
Assumption 8.1. The matrix π΅ in(8.1) does not depend onπ, andβΞ (π₯) β RπΓπ s.t. Ξ (π₯)β€π = π(π₯ , ππ) β π(π₯ , π), where π = ππβπ and ππ is a nominal guess of the uncertain parameterπ.
Under Assumption 8.1, we can write (8.1) as
Β€
π₯ = π(π₯ , ππ) +π΅(π₯)π’βΞ (π₯)β€π
leading to the following theorem [3] for the NCM-based adaptive control. Note that we could also use the SDC formulation with respect to a fixed point as delineated in Theorem 3.2 [12]β[14].
Theorem 8.1. Suppose that Assumption 8.1 holds and let M defines the NCM of Definition 6.2, which modelsπof the CV-STEM contraction metric in Theorem 4.2 for the nominal system (8.1) with π = ππ, constructed with an additional convex constraint given as ππ
π(π₯)πΒ― + ππ
π(π₯π)πΒ― = 0, where ππ
π(π)πΒ― = Γ
π(ππΒ―/π ππ)ππ(π) for π΅ = [π1,Β· Β· Β· , ππ] (see [4], [10]). Suppose also that the matched uncertainty condition [10] holds, i.e.,(Ξ (π₯) βΞ (π₯π))β€πβspan(π΅(π₯))forΞ (π₯), and that(8.1) is controlled by the following adaptive control law:
π’=π’πΏ+π(π₯ , π₯π)β€πΛ (8.2)
Β€Λ
π=βΞ(π(π₯ , π₯π)π΅(π₯)β€M (π₯ , π₯π, π’π) (π₯βπ₯π) +ππΛ) (8.3) where π’πΏ is given by (6.7) of Theorem 6.3, i.e., π’πΏ = π’π β π β1π΅β€M (π₯ β π₯π), Ξ β RπΓπ is a diagonal matrix with positive elements that governs the rate of adaptation,π βRβ₯0, and(Ξ (π₯) βΞ (π₯π))β€π=π΅(π₯)π(π₯ , π₯π)β€π.
If βπΎ , πΎ ,π,Β― π,Β― π,Β― πΒ― β R>0 s.t. πΎI βͺ― Ξ βͺ― πΎI, β₯π΅(π₯) β₯ β€ πΒ―, β₯π β1(π₯ , π₯π, π’π) β₯ β€ πΒ―,
β₯π(π₯ , π₯π) β₯ β€ πΒ―, and β₯πβ₯ β€ πΒ―, and if Ξ and π of (8.3) are selected to satisfy the following relation for the learning errorβ₯M βπβ₯ β€ πβ in some compact setS as in(6.18)of Theorem 6.3:
"
β2πΌβπ πΒ―ππΒ― β
Β―
πππΒ― β β2π
#
βͺ― β2πΌπ
"
π 0
0 1/πΎ
#
(8.4) for βπΌπ β R>0, πΌβ given in Theorem 6.3, andπ andπ given inπI βͺ― π βͺ― πIof (2.26), then the system(8.1)is robust against bounded deterministic and stochastic disturbances withπ β 0, and we have the following bound in the compact setS:
β₯e(π‘) β₯ β€ (πβ(0)πβπΌππ‘+πΌβ1
π π
βοΈ
πΎπΒ―(1βπβπΌππ‘))/β
π (8.5)
where e = π₯βπ₯π, andπβ = β«π1
π0 β₯ΞπΏπβ₯ is defined in Theorem 2.3 with π = Ξβ€Ξ replaced by diag(π ,Ξβ1) for π0 = [π₯β€
π, πβ€]β€ and π1 = [π₯β€,πΛβ€]β€. Furthermore, if the learning error πβ = 0 (CV-STEM control), (8.3) with π = 0 guarantees asymptotic stability ofπ₯toπ₯π in(8.1).
Proof. The proof can be found in [3], but here we emphasize the use of contraction theory. Forπ’given by (8.2), the virtual system of a smooth pathπ(π, π‘) = [πβ€
π₯, πβ€
π]β€ parameterized by π β [0,1], which hasπ(π =0, π‘) = [π₯β€
π
, πβ€]β€ andπ(π=1, π‘) = [π₯β€,πΛβ€]β€ as its particular solutions, is given as follows:
Β€ π =
"
π(ππ₯, π₯ , π₯π, π’π) +π΅ πβ€ππβΞ (π₯)β€π
βΞ(π π΅β€M (ππ₯ βπ₯π) +π(ππβπ)) +ππ π(π, π)
#
where ππ π(π, π) = βππ π. Note that π is as given in (3.13), i.e., π = (π΄ β π΅ π β1π΅β€M) (ππ₯βπ₯π) + Β€π₯π, where the SDC matrix π΄is defined as (see Lemma 3.1) π΄(π₯βπ₯π) = π(π₯ , ππ) +π΅(π₯)π’πβ π(π₯π, ππ) βπ΅(π₯π)π’π.
The arguments of π΄(π, π₯ , π₯π, π’π), π΅(π₯),M (π₯ , π₯π, π’π), andπ(π₯ , π₯π) are omitted for notational simplicity. Since πΒ€π₯ = π(ππ₯, π₯ , π₯π, π’π, π‘) is contracting due to Theo- rems 4.2 and 6.3 with a contraction rate given byπΌβ in Theorem 6.3, we have for a Lyapunov functionπ =πΏπβ€
π₯π πΏππ₯+πΏπβ€
πΞβ1πΏππthat πΒ€/2 β€ βπΌβπΏπβ€
π₯π πΏππ₯+πΏπβ€
π₯(πβ M)π΅ πβ€πΏππβπβ₯πΏππβ₯2+πΏπβ€
ππΏ ππ π. (8.6) Applying (8.4) withβ₯M βπβ₯ β€πβ of (6.18), we get
πΒ€/2βπΏπβ€
ππΏ ππ π β€ β(πΌβπ) β₯πΏππ₯β₯2+πΒ―ππΒ― ββ₯πΏππ₯β₯ β₯πΏππβ₯ βπβ₯πΏππβ₯2
β€ βπΌπ(πβ₯πΏππ₯β₯2+ β₯πΏππβ₯2/πΎ) β€ βπΌππ .
Since we have β₯π ππ π/π πβ₯ = ππΒ―, this impliesπΒ€β β€ βπΌππβ +π
βοΈ
πΎπΒ―, yielding the bound (8.5) due to Lemma 2.1 [3]. Robustness against deterministic and stochastic disturbances follows from Theorem 3.1 ifπ β 0. Ifπβ =0 andπ =0, the relation (8.6) reduces toπΒ€/2 β€ βπΌπΏπβ€
π₯π πΏππ₯, which results in asymptotic stability of π₯ to π₯π in (8.1) due to Barbalatβs lemma [15, p. 323] as in the proof of Theorem 2 in [10].
Remark 8.1. Although Theorem 8.1 is for the case where π(π₯) is affine in its parameter, it is also useful for the following types of systems with an uncertain parameterπ βRπand a control inputπ’(see [3]):
π»(π₯)π(π) +β(π₯) +Ξ (π₯)π =π’
where π β Rπ, π’ β Rπ, β : Rπ β¦β Rπ, π» : Rπ β¦β RπΓπ, Ξ : Rπ β¦β RπΓπ, π₯ = [(π(πβ2))β€,Β· Β· Β· ,(π)β€]β€, andπ(π)denotes theπth time derivative ofπ. In particular, adaptive sliding control [16] designsπ’to render the system of the composite variable π given asπ = π(πβ1)βπ(πβ1)
π to be contracting, whereπ(πβ1)
π = π(πβ1)
π βΓπβ2 π=0 ππe(π), e = π β ππ, ππ is a target trajectory, and π πβ1+ππβ2π πβ2+ Β· Β· Β· +π0 is a stable (Hurwitz) polynomial in the Laplace variableπ (see Example 2.6). Since we have e(πβ1) = π βΓπβ2
π=0 ππe(π) and the system for [e(0),Β· Β· Β· ,e(πβ2)] is also contracting if π =0due to the Hurwitz property, the hierarchical combination property [17], [18]
of contraction in Theorem 2.7 guaranteeslimπ‘βββ₯π β ππβ₯ = 0[19, p. 352] (see Example 2.7).
Example 8.1. Using the technique of Remark 8.1, we can construct adaptive control for the Lagrangian system in Example 2.6 as follows:
H (q) Β₯q+ C (q,q) €€ q+ G (q) =π’(q,qΒ€, π‘) (8.7) π’(q,qΒ€, π‘) =βK (π‘) ( Β€qβ Β€qπ) +H (q) Β₯Λ qπ +C (qΛ ,q) €€ qπ +G (q)Λ (8.8) where H,Λ C, andΛ GΛ are the estimates of H, C, and G, respectively, and the other variables are as defined in Example 2.6. Suppose that the terms H, C, and G depend linearly on the unknown parameter vectorπas follows [19, p. 405]:
H (q) Β₯qπ + C (q,q) €€ qπ + G (q)=π(q,qΒ€,qΒ€π,qΒ₯π)π .
Updating the parameter estimateπΛusing the adaptation law,πΒ€Λ=βΞπ(q,qΒ€,qΒ€π,qΒ₯π)β€( Β€qβ
qΒ€π), as in(8.3)where Ξ β» 0, we can define the following virtual system which has π =π0= [ Β€qβ€π, πβ€]β€andπ =π1= [ Β€qβ€,πΛβ€]β€as its particular solutions:
"
H 0
0 Ξβ1
#
( Β€πβ Β€π0) +
"
C + K βπ
πβ€ 0
#
(πβπ0) =0
where the relationπ’ =βK (π‘) ( Β€qβ Β€qπ) +π(q,qΒ€,qΒ€π,qΒ₯π)πΛis used. Thus, for a Lyapunov functionπ =πΏπβ€ H 0
0 Ξβ1
πΏπ, we have that πΒ€ =πΏπβ€
"
K π
βπβ€ 0
#
πΏπ =πΏπβ€
"
K 0
0 0
# πΏπ
which results in asymptotic stability of πΏπ (i.e., semi-contraction [20], see also Barbalatβs lemma [19, p. 405-406]).
8.1.II Parameter-Dependent Contraction Metric (aNCM)
Although Theorem 8.1 utilizes the NCM designed for the nominal system (8.1) with π = ππ, we could improve its representational power by explicitly taking the parameter estimate Λπ as one of the NCM arguments [3], leading to the concept of an adaptive NCM (aNCM).
In this section, we consider multiplicatively-separable nonlinear systems given in Assumption 8.2, which holds for many types of systems including robotics sys- tems [19], spacecraft high-fidelity dynamics [21], [22], and systems modeled by basis function approximation and DNNs [23], [24].
Assumption 8.2. The dynamical system(8.1)is multiplicatively separable in terms ofπ₯ andπ, i.e.,βππ :Rπβ¦β RπΓππ§,ππ
π :Rπβ¦β RπΓππ§, andπ :Rπ β¦βRππ§ s.t.
ππ(π₯)π(π) = π(π₯ , π), ππ
π(π₯)π(π) =ππ(π₯ , π)
whereπ΅(π₯ , π) = [π1(π₯ , π),Β· Β· Β· , ππ(π₯ , π)]. We could defineπas[πβ€, π(π)β€]β€so we haveππ(π)π = π(π, π)andππ
π(π)π =ππ(π;π). Such an over-parameterized system could be regularized using the Bregman divergence as in [25] (see Example 8.4), and thus we denote[πβ€, π(π)β€]β€asπin the subsequent discussion.
Under Assumption 8.2, the dynamics fore = π₯β π₯π of (8.1) can be expressed as follows:
Β€
e= π΄(π, π₯ , π₯π, π’π,πΛ)e+π΅(π₯ ,πΛ) (π’βπ’π) βπΛ(πΛβπ) (8.9) where π΄ is the SDC matrix of Lemma 3.1, Λπ is the current estimate of π, and Λπ is defined as
Λ
π =πβππ =(ππ(π₯) +ππ(π₯ , π’)) β (ππ(π₯π) +ππ(π₯π, π’π)) (8.10) whereππ(π₯ , π’)=Γπ
π=1ππ
π(π)π’π.
Definition 8.1. The adaptive Neural Contraction Metric (aNCM) in Fig. 8.1 is a DNN model for the optimal parameter-dependent contraction metric, given by solving the adaptive CV-STEM, i.e., (4.2) of Theorem 4.2 (or Theorem 4.6 for differential feedback) with its contraction constraint replaced by the following convex constraint:
βΞ+2 sym π΄πΒ―
β2π π΅ π β1π΅β€ βͺ― β2πΌπΒ― (8.11)
for deterministic systems, and
"
βΞ+2 sym(π΄πΒ―) β2π π΅ π β1π΅β€+2πΌπΒ― πΒ―
Β―
π β π
πΌπ I
#
βͺ― 0 (8.12)
for stochastic systems, whereπ = π(π₯ , π₯π, π’π,πΛ)β1 β» 0(orπ = π(π₯ ,πΛ)β1 β» 0, see Theorem 3.2),πΒ― = ππ, π = π, π = π (π₯ , π₯π, π’π) β» 0is a weight matrix on π’, Ξ = (π/π π‘) |πΛπΒ― is the time derivative ofπΒ― computed along(8.1)withπ =πΛ, π΄and π΅are given in(8.9),πΌ,π,π, andπΌπ are as given in Theorem 4.2, and the arguments (π₯ , π₯π, π’π,πΛ) are omitted for notational simplicity.
The aNCM given in Definition 8.1 has the following stability property along with its optimality due to the CV-STEM of Theorem 4.2 [3].
Theorem 8.2. Suppose that Assumption 8.2 holds and let M define the aNCM of Definition 8.1. Suppose also that the true dynamics (8.1) is controlled by the following adaptive control law:
π’ =π’πβπ (π₯ , π₯π, π’π)β1π΅(π₯ ,πΛ)β€M (π₯ , π₯π, π’π,πΛ)e (8.13)
Β€Λ
π = Ξ( (πβ€πMπ₯+πβ€
π πMπ₯π+πΛβ€M)eβππΛ) (8.14)
where πMπ = [(πM/π π1)e,Β· Β· Β· ,(πM/π ππ)e]β€/2, e = π₯ β π₯π, Ξ β» 0, π β Rβ₯0, and π,ππ, πΛ are given in (8.10). Suppose further that the learning error in
β₯M βπβ₯ β€ πβ of Theorem 6.3 additionally satisfies β₯πMπ₯π β πππ₯
πβ₯ β€ πβ and
β₯πMπ₯βπππ₯β₯ β€ πβ in some compact setS.
If βπ,Β― π,Β― π¦Β― β R>0 s.t. β₯π΅(π₯ ,πΛ) β₯ β€ πΒ―, β₯π β1(π₯ , π₯π, π’π) β₯ β€ πΒ―, β₯πβ₯ β€ π¦Β―, β₯ππβ₯ β€ π¦Β―, and β₯πΛβ₯ β€ π¦Β― in(8.13) and(8.14), and if Ξ andπ of (8.14)are selected to satisfy the following as in Theorem 8.1:
"
β2πΌβπ π¦πΒ― β
Β―
π¦πβ β2π
#
βͺ― β2πΌπ
"
π 0
0 1/πΎ
#
(8.15)
π π β ππ
πππ π
π½ΰ·‘
Neural Network Adaptive NCM
True Dynamical System
aNCM Adaptation Reference Model
Target Trajectory Generation π΄
π΄
π π½ΰ·‘
π, π
π, π
π, π
ππ , ππ
ππ , ππ
Neural net model of adaptive CV-STEM
ππ
Figure 8.1: Illustration of aNCM (π: positive definite matrix that defines aNCM; Λπ: estimated parameter;π: error signal, see (8.10);π₯: system state;π’: system control input; and(π₯π, π’π): target state and control input trajectory).
for βπΌπ β R>0, πΌβ given in Theorem 6.3, and π and π given in πI βͺ― π βͺ― πI of (2.26), then the system (8.1) is robust against bounded deterministic and stochastic disturbances, and we have the exponential bound (8.5) in the compact set S. Furthermore, if πβ = 0 (adaptive CV-STEM control), (8.14) with π = 0 guarantees asymptotic stability ofπ₯ toπ₯πin(8.1).
Proof. Replacing the contraction constraints of the CV-STEM in Theorem 4.2 by (8.11) and (8.12), the bound (8.5) and the asymptotic stability result can be derived as in the proof of Theorem 8.1 (see Theorem 4 and Corollary 2 of [3] for details).
Theorems 2.4 and 2.5 guarantee robustness of (8.1) against bounded deterministic and stochastic disturbances forπβ 0.
Let us again emphasize that, by using Theorem 4.6, the results of Theorems 8.1 and 8.2 can be extended to adaptive control with CCM-based differential feed- back [10], [11] (see Table 3.1 for the trade-offs).
Since the adaptation laws (8.3) and (8.14) in Theorems 8.1 and 8.2 yield an explicit bound on the steady-state error as in (8.5), it could be used as the objective function of the CV-STEM in Theorem 4.2, regardingΞandπas extra decision variables to get π optimal in a sense different from Theorem 4.2. Smallerπβ would lead to a weaker condition on them in (8.4) and (8.15). Also, the size of β₯πβ₯ β€ πΒ― in (8.5) can be adjusted simply by rescaling it (e.g.,πβ π/πΒ―). However, such a robustness guarantee comes with a drawback of having limπ‘βββ₯πΛ(π‘) β₯ =0 forπ β 0 in (8.3), leading to the trade-offs in different types of adaptation laws, some of which are given in Examples 8.2 β 8.4 (see also Remark 8.2).
Example 8.2. Let us briefly describe the following projection operator-based adap- tation law, again for the Lagrangian system(8.7)of Example 8.1 with the unknown parameter vectorπ:
Β€Λ
π =Proj(π ,Λ βΞπ(q,qΒ€,qΒ€π,qΒ₯π)β€( Β€qβ Β€qπ), π) (8.16) where Proj is the projection operator and π is a convex boundary function (e.g., π(πΛ) =(πΛβ€πΛβπ2
max)/(πππ2
max)for given positive constantsπmaxandππ). Ifπ(πΛ) > 0 andβπ(πΛ)β€π > 0,
Proj(π , π , πΛ ) =πβ (βπ(πΛ)βπ(πΛ)β€/β₯βπ(πΛ) β₯2)π π(πΛ),
otherwise, Proj(π , π , πΛ ) = π. The projection operator has the following useful property which allows bounding the parameter estimateπΛ.
Lemma 8.1. If πΛ is given by (8.16) with πΛ(0) β Ξ©π = {π β Rπ|π(π) β€ 1} for a convex function π(π), then πΛ(π‘) β Ξ©π, βπ‘ β₯ 0 (e.g., β₯πΛβ₯ β€ πmax
β1+ππ if π(πΛ) =(πΛβ€πΛβπ2
max)/(πππ2
max)).
Proof. See [26].
Since Lemma 8.1 guarantees the boundedness ofβ₯πΛβ₯, the adaptive controller(8.8), π’ =βK ( Β€qβ Β€qπ) +π π+ππΛ, can be viewed as the exponentially stabilizing controller
βK ( Β€qβ Β€qπ) +π π(see Example 2.6) plus a bounded external disturbanceππΛ, implying robustness due to Theorem 2.4 [11], [26]β[28].
Let us also remark that, as in Example 8.1, the projection operator-based adaptation law (8.16) still achieves asymptotic stabilization. Applying the control law (8.8) with the adaptation (8.16) yields the following virtual system which has π = π0 = [ Β€qβ€π, πβ€]β€ andπ =π1= [ Β€qβ€,πΛβ€]β€as its particular solutions:
"
H 0
0 I
#
( Β€πβ Β€π0) +
"
(C + K) (ππ β Β€qπ) βπ(ππβπ) Proj(π ,Λ Ξπβ€(ππ β Β€qπ), π)
#
=0 whereπ= [πβ€
π , πβ€
π]β€. Thus, for a Lyapunov functionππ β = 1
2
β«π1
π0
πΏπβ€ hH (q)
0 0 Ξβ1
i πΏπ, we have that
πΒ€π β =
β« π1
π0
βπΏπβ€
π KπΏππ +πΏπβ€
π(Proj(π , πΛ β€πΏππ , π) βπβ€πΏππ ).
Using the convex property of the projection operator, i.e.,πΛβ€(Proj(π , π , πΛ ) βπ) β€ 0[26], [27], this givesπΒ€π β β€ ββ«π1
π0
πΏπβ€
π KπΏππ , which results in asymptotic stability ofπΏπdue to Barbalatβs lemma [19, p. 405-406] as in Example 8.1.
Example 8.3. The dependence on π’andπΒ€Λin (π/π π‘) |πΛπ can be removed by using
ππ
π(π₯)π + ππ
π(π₯π)π = 0 as in Theorem 8.1, and by using adaptation rate scaling introduced in [11]. In essence, the latter multiplies the adaptation law (8.14) by any strictly-increasing and strictly-positive scalar functionπ£(2π), and updateπ as
Β€ π = 1
2 π£(2π) π£π(2π)
π
βοΈ
π=1
1 πe+π
ππe
ππΛπ
πΒ€π (8.17)
where π£π = π π£/π π, π β R>0, andπe = eβ€π(π₯ , π₯π, π’π,πΛ)e for e = π₯βπ₯π and π given in Definition 8.1, so the additional term due to (8.17) cancels out the term involving πΒ€Λ in (π/π π‘) |Λ
ππΒ― of (8.11) (see [11] for details). Its robustness property follows from Theorem 8.2 also in this case.
Example 8.4. Using the Bregman divergence-based adaptation in [25], we could implicitly regularize the parameter estimateπΛas follows:
lim
π‘ββ
Λ
π =arg min
πβπ΄
ππ(πβ₯πβ) =arg min
πβπ΄
ππ(πβ₯πΛ(0))
where ππ is the Bregman divergence defined as ππ(π₯β₯π¦) = π(π₯) β π(π¦) β (π₯ β π¦)β€βπ(π¦) for a convex functionπ, andA is a set containing only parameters that interpolate the dynamics along the entire trajectory. IfπΛ(0) =arg minπβRππ(π), we havelimπ‘ββπΛ=arg minπβπ΄π(π), which regularizesπΛto converge to a parameter that minimizesπ. For example, using 1-norm forπ would impose sparsity on the steady-state parameter estimateπΛ[25].
These extensions of adaptive control techniques described in Examples 8.2 β 8.4 could be utilized with contraction theory and learning-based control as in Theo- rems 8.1 and 8.2.
Remark 8.2. Note that the results presented earlier in this chapter do not necessarily mean parameter convergence,limπ‘βββ₯πΛβ₯ =0, as the adaptation objective is to drive the tracking errorβ₯π₯βπ₯πβ₯to zero [19, p. 331], not to find out the true parameterπ out of the many that achieve perfect tracking.
Asymptotic parameter convergence, limπ‘βββ₯πΛβ₯ = 0 for π of (8.1) under Assump- tion 8.2, could be guaranteed if there is no disturbance and learning error with π = 0, and we have βπ , πΌπ πΈ β R>0s.t. β«π‘+π
π‘
Λ
πβ€π π πΛ βͺ° πΌπ πΈI, βπ‘ forπΛ given in (8.10) (the persistent excitation condition [19, p. 366]). We could also utilize the Bregman divergence-based adaptation to regularize the behavior oflimπ‘βββ₯πΛβ₯ as in Example 8.4.