Chapter 3: Robust Nonlinear Control and Estimation via Contraction Theory . 50
3.4 KYP Lemma in Contraction Theory
Analogously to Theorem 3.3, the LMI (3.18) of Theorem 3.1 can be understood using the Kalman-Yakubovich-Popov (KYP) lemma [39, p. 218]. Consider the quadratic Lyapunov function๐ = ๐ โคQ๐ withQ โป 0, satisfying the following output strict passivity (dissipativity) condition:
๐ยค โ2๐โค๐+ (2/๐พ)๐โค๐ โค0. (3.31)
This which can be expanded by completing the square to have ๐ยค โค โโฅ๐พ(๐ โ๐/๐พ) โฅ2+๐พโฅ๐โฅ2โ (1/๐พ) โฅ๐โฅ2
โค ๐พโฅ๐โฅ2โ (1/๐พ) โฅ๐โฅ2. (3.32)
This implies that we have โฅ (๐)๐โฅL2 โค ๐พโฅ (๐)๐โฅL2 +โ๏ธ
๐พ๐(0) by the comparison lemma [25, pp. 102-103, pp. 350-353], leading to L2 gain stability with its L2 gain less than or equal to ๐พ [39, p. 218]. The condition (3.31) can be expressed equivalently as an LMI form as follows:
๏ฃฎ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฐ
Q +ยค 2sym(QA) + 2C๐พโคC QB + Cโค 2D
๐พ โI
BโคQ +
2Dโค ๐พ โI
C โ(Dโค+๐ท) + 2DโคD
๐พ
๏ฃน
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃป
โชฏ 0 (3.33)
whereA,B, C, andD are as defined in (3.29).
Theorem 3.4. If(3.33)holds, the LMI(3.30)for the bounded real lemma holds with P =๐พQ, i.e., the system(3.29)isL2gain stable with itsL2gain less than or equal to๐พ. Thus, for systems withA,B, C, andD defined in Theorem 3.3, the condition (3.33)guarantees the contraction condition(3.18)of Theorem 3.1.
Proof. Writing the inequality in (3.32) in a matrix form, we have that
๏ฃฎ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฐ
Q +ยค 2sym(QA) + 2CโคC
๐พ QB + Cโค
2D ๐พ โI
BโคQ +
2Dโค ๐พ โI
C โ(Dโค+๐ท) + 2DโคD
๐พ
๏ฃน
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃป
โชฐ
"Q +ยค 2sym(QA) QB
BโคQ 0
# +
"CโคC ๐พ
CโคD ๐พ DโคC
๐พ
DโคD ๐พ
#
โ
"
0 0 0 ๐พI
#
=
"Q +ยค 2sym(QA) + Cโค๐พC QB + CโคD
๐พ
BโคQ + DโคC
๐พ
DโคD ๐พ โ๐พI
# .
Therefore, a necessary condition of (3.33) reduces to (3.30) ifQ = P/๐พ. The rest follows from Theorem 3.3.
Example 3.5. Theorems 3.3 and 3.4 imply the following statements.
โข If(3.30)holds, the system is finite-gainL2stable with๐พ as itsL2gain.
โข If(3.33)holds, the system is finite-gainL2stable with๐พ as itsL2gain.
โข (3.31) is equivalent to(3.33), and to the output strict passivity condition with dissipation1/๐พ [25, p. 231].
โข If(3.33)holds (KYP lemma),(3.30)holds (bounded real lemma). The converse is not necessarily true.
References
[1] T. Basar and P. Bernhard,HโOptimal Control and Related Minimax Design Problems: A Dynamic Game Approach. Birkhauser, 1995.
[2] W. Zhang and B. Chen, โState feedbackHโcontrol for a class of nonlinear stochastic systems,โSIAM J. Control Optim., vol. 44, no. 6, pp. 1973โ1991, 2006.
[3] I. R. Manchester and J.-J. E. Slotine, โRobust control contraction metrics:
A convex approach to nonlinear state-feedbackHโcontrol,โIEEE Control Syst. Lett., vol. 2, no. 3, pp. 333โ338, 2018.
[4] A. J. Van der Schaft, โL2-gain analysis of nonlinear systems and nonlinear state-feedback Hโ control,โ IEEE Trans. Autom. Control, vol. 37, no. 6, pp. 770โ784, Jun. 1992.
[5] A. Isidori and A. Astolfi, โDisturbance attenuation and Hโ-control via measurement feedback in nonlinear systems,โIEEE Trans. Autom. Control, vol. 37, no. 9, pp. 1283โ1293, Sep. 1992.
[6] J. C. Doyle, K. Glover, P. P. Khargonekar, and B. A. Francis, โState-space solutions to standard H2 and Hโ control problems,โ IEEE Trans. Autom.
Control, vol. 34, no. 8, pp. 831โ847, Aug. 1989.
[7] P. P. Khargonekar and M. A. Rotea, โMixed H2/Hโ control: A convex optimization approach,โIEEE Trans. Autom. Control, vol. 36, no. 7, pp. 824โ
837, Jul. 1991.
[8] L. Xie and E. de Souza Carlos, โRobust Hโ control for linear systems with norm-bounded time-varying uncertainty,โIEEE Trans. Autom. Control, vol. 37, no. 8, pp. 1188โ1191, Aug. 1992.
[9] C. Scherer, P. Gahinet, and M. Chilali, โMultiobjective output-feedback control via LMI optimization,โIEEE Trans. Autom. Control, vol. 42, no. 7, pp. 896โ911, Jul. 1997.
[10] B.-S. Chen and W. Zhang, โStochasticH2/Hโcontrol with state-dependent noise,โ IEEE Trans. Autom. Control, vol. 49, no. 1, pp. 45โ57, Jan. 2004, issn: 0018-9286.
[11] D. Hinrichsen and A. Pritchard, โStochasticHโ,โSIAM J. Control Optim., vol. 36, no. 5, pp. 1504โ1538, 1998.
[12] H. Tsukamoto and S.-J. Chung, โRobust controller design for stochastic nonlinear systems via convex optimization,โ IEEE Trans. Autom. Control, vol. 66, no. 10, pp. 4731โ4746, 2021.
[13] H. Tsukamoto and S.-J. Chung, โConvex optimization-based controller de- sign for stochastic nonlinear systems using contraction analysis,โ in IEEE Conf. Decis. Control, Dec. 2019, pp. 8196โ8203.
[14] R. Wang, R. Tรณth, and I. R. Manchester,Virtual CCMs: Convex nonlinear feedback design via behavioral embedding, arXiv:2003.08513, Mar. 2020.
[15] R. Wang, R. Tรณth, and I. R. Manchester, โA comparison of LPV gain scheduling and control contraction metrics for nonlinear control,โ in 3rd IFAC Workshop on LPVS, vol. 52, 2019, pp. 44โ49.
[16] H. Tsukamoto and S.-J. Chung, โNeural contraction metrics for robust esti- mation and control: A convex optimization approach,โ IEEE Control Syst.
Lett., vol. 5, no. 1, pp. 211โ216, 2021.
[17] H. Tsukamoto, S.-J. Chung, and J.-J. E. Slotine, โNeural stochastic contrac- tion metrics for learning-based control and estimation,โIEEE Control Syst.
Lett., vol. 5, no. 5, pp. 1825โ1830, 2021.
[18] A. P. Dani, S.-J. Chung, and S. Hutchinson, โObserver design for stochastic nonlinear systems via contraction-based incremental stability,โIEEE Trans.
Autom. Control, vol. 60, no. 3, pp. 700โ714, Mar. 2015.
[19] I. R. Manchester and J.-J. E. Slotine, โControl contraction metrics: Convex and intrinsic criteria for nonlinear feedback design,โ IEEE Trans. Autom.
Control, vol. 62, no. 6, pp. 3046โ3053, Jun. 2017.
[20] S. Singh, A. Majumdar, J.-J. E. Slotine, and M. Pavone, โRobust online motion planning via contraction theory and convex optimization,โ inIEEE Int. Conf. Robot. Automat., May 2017, pp. 5883โ5890.
[21] J.-J. E. Slotine and W. Li,Applied Nonlinear Control. Upper Saddle River, NJ: Pearson, 1991.
[22] F. Tan, W. Lohmiller, and J.-J. E. Slotine, โAnalytical SLAM without lin- earization,โInt. J. Robot. Res., vol. 36, no. 13-14, pp. 1554โ1578, 2017.
[23] E. Sontag, โA Lyapunov-like characterization of asymptotic controllability,โ
SIAM J. Control Optim., vol. 21, no. 3, pp. 462โ471, 1983.
[24] E. D. Sontag, โA โuniversalโ construction of Artsteinโs theorem on nonlinear stabilization,โSyst. Control Lett., vol. 13, no. 2, pp. 117โ123, 1989.
[25] H. K. Khalil,Nonlinear Systems, 3rd. Upper Saddle River, NJ: Prentice-Hall, 2002.
[26] W. Li and E. Todorov, โIterative linear quadratic regulator design for non- linear biological movement systems,โ inInt. Conf. Inform. Control Automat.
Robot., 2004, pp. 222โ229.
[27] W. Li and E. Todorov, โAn iterative optimal control and estimation design for nonlinear stochastic system,โ inIEEE Conf. Decis. Control, 2006, pp. 3242โ
3247.
[28] S. Bonnabel and J.-J. E. Slotine, โA contraction theory-based analysis of the stability of the deterministic extended Kalman filter,โ IEEE Trans. Autom.
Control, vol. 60, no. 2, pp. 565โ569, Feb. 2015.
[29] J. R. Cloutier, โState-dependent Riccati equation techniques: An overview,โ
inProc. Amer. Control Conf., vol. 2, 1997, pp. 932โ936.
[30] H. T. Banks, B. M. Lewis, and H. T. Tran, โNonlinear feedback controllers and compensators: A state-dependent Riccati equation approach,โComput.
Optim. Appl., vol. 37, no. 2, pp. 177โ218, Jun. 2007, issn: 1573-2894.
[31] I. Chang and S.-J. Chung, โExponential stability region estimates for the state-dependent Riccati Equation controllers,โ inIEEE Conf. Decis. Control, 2009, pp. 1974โ1979.
[32] T. รimen, โSurvey of state-dependent Riccati equation in nonlinear optimal feedback control synthesis,โ AIAA J. Guid. Control Dyn., vol. 35, no. 4, pp. 1025โ1047, Jul. 2012.
[33] S. Singh, S. M. Richards, V. Sindhwani, J.-J. E. Slotine, and M. Pavone,
โLearning stabilizable nonlinear dynamics with contraction-based regular- ization,โInt. J. Robot. Res., Aug. 2020.
[34] Q. C. Pham and J.-J. E. Slotine,Stochastic contraction in Riemannian met- rics, arXiv:1304.0340, Apr. 2013.
[35] I. R. Manchester, โContracting nonlinear observers: Convex optimization and learning from data,โ in Proc. Amer. Control Conf., 2018, pp. 1873โ
1880.
[36] H. Tsukamoto and S.-J. Chung, โLearning-based robust motion planning with guaranteed stability: A contraction theory approach,โ IEEE Robot.
Automat. Lett., vol. 6, no. 4, pp. 6164โ6171, 2021.
[37] S. Boyd, L. El Ghaoui, E. Feron, and V. Balakrishnan, Linear Matrix In- equalities in System and Control Theory(Studies in Applied Mathematics).
Philadelphia, PA: SIAM, Jun. 1994, vol. 15, isbn: 0-89871-334-X.
[38] T. Baลar and P. Bernhard,HโOptimal Control and Related Minimax Design Problems, A Dynamic Game Approach(Modern Birkhรคuser Classics), 2nd.
Birkhรคuser Basel, 2008, isbn: 978-0-387-98945-7.
[39] G. E. Dullerud and F. Paganini, A Course in Robust Control Theory, A Convex Approach (Texts in Applied Mathematics), 1st. Springer, 2000, isbn: 978-0-387-98945-7.
C h a p t e r 4
CONVEX OPTIMALITY IN ROBUST NONLINEAR CONTROL AND ESTIMATION VIA CONTRACTION THEORY
[1] H. Tsukamoto and S.-J. Chung, โRobust controller design for stochastic nonlinear systems via convex optimization,โ IEEE Trans. Autom. Control, vol. 66, no. 10, pp. 4731โ4746, 2021.
[2] H. Tsukamoto and S.-J. Chung, โConvex optimization-based controller de- sign for stochastic nonlinear systems using contraction analysis,โ in IEEE Conf. Decis. Control, Dec. 2019, pp. 8196โ8203.
Theorem 3.1 indicates that the problem of finding contraction metrics for general nonlinear systems could be formulated as a convex feasibility problem. This chap- ter thus delineates one approach, called the method of ConVex optimization-based Steady-state Tracking Error Minimization (CV-STEM) [1]โ[4], to optimally design ๐ of Theorem 3.1 that defines a contraction metric and minimizes an upper bound of the steady-state tracking error in Theorem 2.4 or in Theorem 2.5 via convex optimization. In particular, we present an overview of the SDC- and CCM-based CV-STEM frameworks for provably stable and optimal feedback control and state estimation of nonlinear systems, perturbed by deterministic and stochastic distur- bances. It is worth noting that the steady-state bound is expressed as a function of the condition number ๐ =๐/๐as to be seen in (4.1) and (4.17), which renders the CV-STEM applicable and effective even to learning-based and data-driven control frameworks as shall be seen in Chapter 5.
4.1 CV-STEM Control
As a result of Theorems 2.4, 2.5, and 3.1, the control law๐ข = ๐ข๐ โ๐พ(๐ฅ โ๐ฅ๐) of (3.10) gives a convex steady-state upper bound of the Euclidean distance between the system trajectories, which can be used as an objective function for the CV-STEM control framework in Theorem 4.2 [1]โ[4].
Theorem 4.1. If one of the matrix inequalities of Theorem 3.1 holds, we have the following bound for ๐ =๐/๐:
lim
๐กโโ
โ๏ธ
E
โฅ๐ฅโ๐ฅ๐โฅ2
โค ๐0(๐ผ, ๐ผ๐บ)โ
๐ โค ๐0(๐ผ, ๐ผ๐บ)๐ (4.1)
where ๐0(๐ผ, ๐ผ๐บ) = ๐ยฏ๐
โ๏ธ
(2๐ผโ1
๐บ +1)/(2๐ผ) for stochastic systems and ๐0(๐ผ, ๐ผ๐บ) = ๐ยฏ๐/๐ผfor deterministic systems, with the variables given in Theorem 3.1.
Proof. Taking lim๐กโโin the exponential tracking error bounds (3.20) and (3.21) of Theorem 3.1 gives the first inequality of (4.1). The second inequality follows from the relation 1โค โ๏ธ
๐/๐ โค ๐/๐= ๐due to๐ โค ๐.
The convex optimization problem of minimizing (4.1), subject to the contraction constraint of Theorem 3.1, is given in Theorem 4.2, thereby introducing the CV- STEM control [1]โ[4] for designing an optimal contraction metric and contracting control policy as depicted in Fig. 4.1.
Theorem 4.2. Suppose that๐ผ,๐ผ๐บ,๐ยฏ๐, and๐ยฏ๐in(4.1)and the Lipschitz constant๐ฟ๐ of๐ ๐/๐ ๐ฅ๐in Theorem 3.1 are given. If the pair(๐ด, ๐ต)is uniformly controllable, the non-convex optimization problem of minimizing the upper bound(4.1)is equivalent to the following convex optimization problem, with the convex contraction constraint (3.17)or(3.18)of Theorem 3.1 andI โชฏ๐ยฏ โชฏ ๐Iof(3.19):
๐ฝโ
๐ถ๐ = min
๐โR>0, ๐โR,๐ยฏโป0
๐0๐+๐1๐+๐2๐(๐, ๐,๐ยฏ) (4.2)
s.t.(3.17) and (3.19) for deterministic systems s.t.(3.18) and (3.19) for stochastic systems
where๐ = ๐(๐ฅ , ๐ฅ๐, ๐ข๐, ๐ก)โ1 for Theorem 3.1 or๐ = ๐(๐ฅ , ๐ก)โ1for Theorem 3.2,
ยฏ
๐ =๐๐, ๐ =๐, ๐ =๐/๐, ๐0 is as defined in(4.1)of Theorem 4.1, ๐1, ๐2 โ Rโฅ0, and๐is a performance-based convex cost function.
The weight ๐1 for ๐ can be viewed as a penalty on the 2-norm of the feedback control gain ๐พ of ๐ข = ๐ข๐ โ๐พ(๐ฅ โ๐ฅ๐) in (3.10). Using non-zero ๐1 and ๐2 thus enables finding contraction metrics optimal in a different sense. Furthermore, the coefficients of the SDC parameterizations ๐in Lemma 3.1 (i.e.,๐ด =ร
๐๐๐ด๐in(3.6)) can also be treated as decision variables by convex relaxation, thereby adding a design flexibility to mitigate the effects of external disturbances while verifying the system controllability.
Proof. As proven in Theorems 3.1 and 4.1,๐Iโชฏ ๐ โชฏ ๐I of (2.26), the contraction constraint (3.15), and the objective (4.1) reduce to (4.2) with ๐1 = 0 and ๐2 = 0.
Since the resultant constraints are convex and the objective is affine in terms of the decision variables๐, ๐, and ยฏ๐, the problem (4.2) is indeed convex. Since we have
๐ข๐
CV-STEMdesigns contraction metric
& robust control CV-STEM
Robust Control
Contracting Policy ๐ขโ Contraction Metric w.r.t.๐
๐ฅ๐ฅ๐ ๐ก ๐
CV-STEM Robust Control
๐ขโ State ๐ฅ
Time ๐ก
Target Trajectory
(๐ฅ๐, ๐ข๐)
Convex Program (CV-STEM)
Figure 4.1: Block diagram of CV-STEM.
sup๐ฅ ,๐กโฅ๐โฅ โค ๐ = ๐, ๐1 can be used as a penalty to optimally adjust the induced 2-norm of the control gain (see Example 4.1). The SDC coefficients ๐can also be utilized as decision variables for controllability due to Proposition 1 of [1].
Example 4.1. The weights ๐0 and ๐1 of the CV-STEM control of Theorem 4.2 establish an analogous trade-off to the case of the LQR with the cost weight matrices of๐for state and ๐ for control, since the upper bound of the steady-state tracking error(4.1)is proportional to ๐, and an upper bound ofโฅ๐พโฅfor the control gain ๐พ in(3.10),๐ข=๐ข๐โ๐พ(๐ฅโ๐ฅ๐), is proportional to๐. In particular [2], [3],
โข if ๐1 is much greater than ๐0 in (4.2) of Theorem 4.2, we get smaller control effort but with a large steady-state tracking error, and
โข if๐1is much smaller than๐0in(4.2)of Theorem 4.2, we get a smaller steady-state state tracking error but with larger control effort.
This is also because the solution ๐ โป 0 of the LQR Riccati equation [5, p. 89],
โ ยค๐= ๐ ๐ด+๐ดโค๐โ๐ ๐ต ๐ โ1๐ตโค๐+๐, can be viewed as a positive definite matrix๐ that defines a contraction metric as discussed in Example 2.3.
4.2 CV-STEM Estimation
We could also design an optimal state estimator analogously to the CV-STEM control of Theorem 4.2, due to the differential nature of contraction theory that enables LTV systems-type approaches to stability analysis. In particular, we exploit the estimation and control duality in differential dynamics similar to that of the Kalman filter and LQR in LTV systems.
Let us consider the following smooth nonlinear systems with a measurement๐ฆ(๐ก), perturbed by deterministic disturbances๐๐0(๐ฅ , ๐ก)and๐๐1(๐ฅ , ๐ก)with sup๐ฅ ,๐กโฅ๐๐0(๐ฅ , ๐ก) โฅ=
ยฏ
๐๐0 โ Rโฅ0and sup๐ฅ ,๐กโฅ๐๐1(๐ฅ , ๐ก) โฅ = ๐ยฏ๐1 โ Rโฅ0, or by Gaussian white noise, driven
by Wiener processes ๐ฒ0(๐ก) and ๐ฒ1(๐ก) with sup๐ฅ ,๐กโฅ๐บ๐0(๐ฅ , ๐ก) โฅ๐น = ๐ยฏ๐0 โ Rโฅ0 and sup๐ฅ ,๐กโฅ๐บ๐1(๐ฅ , ๐ก) โฅ๐น =๐ยฏ๐1 โRโฅ0:
ยค
๐ฅ = ๐(๐ฅ , ๐ก) +๐๐0(๐ฅ , ๐ก), ๐ฆ =โ(๐ฅ , ๐ก) +๐๐1(๐ฅ , ๐ก) (4.3) ๐๐ฅ = ๐(๐ฅ , ๐ก)๐ ๐ก+๐บ๐0๐๐ฒ0, ๐ฆ ๐ ๐ก =โ(๐ฅ , ๐ก)๐ ๐ก+๐บ๐1๐๐ฒ1 (4.4) where๐ก โRโฅ0is time,๐ฅ :Rโฅ0โฆโ R๐is the system state,๐ฆ :Rโฅ0โฆโR๐is the system measurement, ๐ : R๐รRโฅ0 โฆโ R๐ and โ : R๐ รRโฅ0 โฆโ R๐ are known smooth functions,๐๐0:R๐รRโฅ0โฆโ R๐,๐๐1 :R๐รRโฅ0 โฆโR0,๐บ๐0:R๐รRโฅ0โฆโR๐ร๐ค0, and ๐บ๐1:R๐รRโฅ0โฆโ R๐ร๐ค1 are unknown bounded functions for external disturbances, ๐ฒ0 :Rโฅ0โฆโR๐ค0 and๐ฒ1:Rโฅ0โฆโ R๐ค1 are two independent Wiener processes, and the arguments of๐บ๐0(๐ฅ , ๐ก)and๐บ๐1(๐ฅ , ๐ก)are suppressed for notational convenience.
Let ๐ด(๐๐, ๐ฅ ,๐ฅ , ๐กห )and๐ถ(๐๐, ๐ฅ ,๐ฅ , ๐กห ) be the SDC matrices given by Lemma 3.1 with (๐ , ๐ ,๐ ,ยฏ ๐ขยฏ) replaced by(๐ ,๐ฅ , ๐ฅ ,ห 0)and (โ,๐ฅ , ๐ฅ ,ห 0), respectively, i.e.
๐ด(๐๐, ๐ฅ ,๐ฅ , ๐กห ) (๐ฅหโ๐ฅ) = ๐(๐ฅ , ๐กห ) โ ๐(๐ฅ , ๐ก) (4.5) ๐ถ(๐๐, ๐ฅ ,๐ฅ , ๐กห ) (๐ฅหโ๐ฅ) =โ(๐ฅ , ๐กห ) โโ(๐ฅ , ๐ก). (4.6) We design a nonlinear state estimation law parameterized by a matrix-valued func- tion๐(๐ฅ , ๐กห )as follows:
ยคห
๐ฅ = ๐(๐ฅ , ๐กห ) +๐ฟ(๐ฅ , ๐กห ) (๐ฆโโ(๐ฅ , ๐กห )) (4.7)
= ๐(๐ฅ , ๐กห ) +๐(๐ฅ , ๐กห )๐ถยฏ(๐๐,๐ฅ , ๐กห )โค๐ (๐ฅ , ๐กห )โ1(๐ฆโโ(๐ฅ , ๐กห ))
where ยฏ๐ถ(๐๐,๐ฅ , ๐กห ) = ๐ถ(๐๐,๐ฅ ,ห ๐ฅ , ๐กยฏ ) for a fixed trajectory ยฏ๐ฅ (e.g., ยฏ๐ฅ = 0, see Theo- rem 3.2), ๐ (๐ฅ , ๐กห ) โป 0 is a weight matrix on the measurement ๐ฆ, and ๐(๐ฅ , ๐กห ) โป 0 is a positive definite matrix (which satisfies the matrix inequality constraint for a contraction metric, to be given in (4.12) of Theorem 4.5). Note that we could use other forms of estimation laws such as the EKF [2], [6], [7], analytical SLAM [8], or SDC with respect to a fixed point [1], [4], [9], depending on the application of interest, which result in a similar stability analysis as in Theorem 3.2.
4.2.I Nonlinear Stability Analysis of SDC-based State Estimation using Con- traction Theory
Substituting (4.7) into (4.3) and (4.4) yields the following virtual system of a smooth path๐(๐, ๐ก), parameterized by๐โ [0,1]to have๐(๐=0, ๐ก) =๐ฅand๐(๐=1, ๐ก) =๐ฅห:
ยค
๐(๐, ๐ก) =๐(๐(๐, ๐ก), ๐ฅ ,๐ฅ , ๐กห ) +๐(๐, ๐ฅ ,๐ฅ , ๐กห ) (4.8) ๐๐(๐, ๐ก) =๐(๐(๐, ๐ก), ๐ฅ ,๐ฅ , ๐กห )๐ ๐ก+๐บ(๐, ๐ฅ ,๐ฅ , ๐กห )๐๐ฒ(๐ก) (4.9)
where ๐(๐, ๐ฅ ,๐ฅ , ๐กห ) = (1 โ ๐)๐๐0(๐ฅ , ๐ก) + ๐ ๐ฟ(๐ฅ , ๐กห )๐๐1(๐ฅ , ๐ก), ๐บ(๐, ๐ฅ ,๐ฅ , ๐กห ) = [(1 โ ๐)๐บ๐0(๐ฅ , ๐ก), ๐ ๐ฟ(๐ฅ , ๐กห )๐บ๐1(๐ฅ , ๐ก)],๐ฒ = [๐ฒโค
0 ,๐ฒ1โค]โค, and ๐(๐, ๐ฅ , ๐ฅ๐, ๐ข๐, ๐ก)is defined as
๐(๐, ๐ฅ ,๐ฅ , ๐กห ) =(๐ด(๐๐, ๐ฅ ,๐ฅ , ๐กห ) โ๐ฟ(๐ฅ , ๐กห )๐ถ(๐๐, ๐ฅ ,๐ฅ , ๐กห )) (๐โ๐ฅ) + ๐(๐ฅ , ๐ก). (4.10) Note that (4.10) is constructed to contain๐ =๐ฅหand๐ =๐ฅas its particular solutions of (4.8) and (4.9). If๐ =0 and๐ฒ =0, the differential dynamics of (4.8) and (4.9) for๐๐๐=๐ ๐/๐ ๐is given as
๐๐๐ยค =(๐ด(๐๐, ๐ฅ ,๐ฅ , ๐กห ) โ๐ฟ(๐ฅ , ๐กห )๐ถ(๐๐, ๐ฅ ,๐ฅ , ๐กห ))๐๐๐ . (4.11) The similarity between (3.14) (๐๐๐ยค= (๐ดโ๐ต๐พ)๐๐๐) and (4.11) leads to the following theorem [1]โ[4]. Again, note that we could also use the SDC formulation with respect to a fixed point as delineated in Theorem 3.2 and as demonstrated in [1], [4], [9].
Theorem 4.3. Suppose โ๐,ยฏ ๐ยฏ โ Rโฅ0 s.t. โฅ๐ โ1(๐ฅ , ๐กห ) โฅ โค ๐ยฏ, โฅ๐ถ(๐๐, ๐ฅ ,๐ฅ , ๐กห ) โฅ โค
ยฏ
๐, โ๐ฅ ,๐ฅ , ๐กห . Suppose also that ๐I โชฏ ๐ โชฏ ๐ ๐ผ of (2.26) holds, or equivalently, I โชฏ ๐ยฏ โชฏ ๐ ๐ผ of(3.19)holds with๐ = ๐(๐ฅ , ๐กห )โ1,๐ยฏ = ๐๐, ๐ =๐, and ๐ = ๐/๐. As in Theorem 3.1, let๐ฝbe defined as๐ฝ =0for deterministic systems(4.3)and ๐ฝ=๐ผ๐ =๐ผ๐0+๐2๐ผ๐1 =๐ฟ๐๐ยฏ2
๐0(๐ผ๐บ +1/2)/2+๐2๐ฟ๐๐ยฏ2๐ยฏ2๐ยฏ2
๐1(๐ผ๐บ +1/2)/2 for stochastic systems(4.4), where2๐ผ๐0= ๐ฟ๐๐ยฏ2
๐0(๐ผ๐บ+1/2),2๐ผ๐1= ๐ฟ๐๐ยฏ2๐ยฏ2๐ยฏ2
๐1(๐ผ๐บ+ 1/2), ๐ฟ๐ is the Lipschitz constant of ๐๐/๐ ๐ฅ๐, ๐ยฏ๐0 and ๐ยฏ๐1 are given in(4.4), and
โ๐ผ๐บ โR>0is an arbitrary constant as in Theorem 2.5.
If ๐(๐ฅ , ๐กห ) in (4.7) is constructed to satisfy the following convex constraint for
โ๐ผโR>0:
ยคยฏ
๐+2 sym(๐ ๐ดยฏ โ๐๐ถยฏโค๐ โ1๐ถ) โชฏ โ2๐ผ๐ยฏ โ๐ ๐ฝI (4.12) then Theorems 2.4 and 2.5 hold for the virtual systems(4.8)and(4.9), respectively, i.e., we have the following bounds fore =๐ฅหโ๐ฅwith๐ =๐and ๐=๐/๐:
โฅe(๐ก) โฅ โค
โ
๐๐โ(0)๐โ๐ผ๐ก+
ยฏ ๐๐0
โ
๐+๐ยฏ๐ยฏ๐ยฏ๐1๐ ๐ผ
(1โ๐โ๐ผ๐ก) (4.13)
E
โฅe(๐ก) โฅ2
โค ๐E[๐๐ โ(0)]๐โ2๐ผ๐ก+ ๐ถ๐0๐+๐ถ๐1๐ ๐2 2๐ผ
(4.14) where ๐๐ โ = โซ๐ฅห
๐ฅ
๐ฟ๐โค๐ ๐ฟ๐ and ๐โ = โซ๐ฅห
๐ฅ โฅฮ๐ฟ๐โฅ are given in Theorem 2.3 with ๐ = ๐โ1 = ฮโคฮ defining a contraction metric, the disturbance bounds ๐ยฏ๐0, ๐ยฏ๐1,
ยฏ
๐๐0, and ๐ยฏ๐1 are given in (4.3)and (4.4), respectively,๐ถ๐0 = ๐ยฏ2
๐0(2๐ผ๐บโ1+1), and ๐ถ๐1 = ๐ยฏ2๐ยฏ2๐ยฏ2
๐1(2๐ผ๐บโ1+1). Note that for stochastic systems, the probability that
โฅeโฅ is greater than or equal to๐โR>0is given as P[โฅe(๐ก) โฅ โฅ ๐] โค 1
๐2
๐E[๐๐ โ(0)]๐โ2๐ผ๐ก+ ๐ถ๐ธ 2๐ผ
(4.15) where๐ถ๐ธ =๐ถ๐0๐+๐ถ๐1๐ ๐2.
Proof. Theorem 3.1 indicates that (4.12) is equivalent to
๐ยค +2 sym(๐ ๐ดโ๐ถยฏโค๐ โ1๐ถ) โชฏ โ2๐ผ๐โ๐ฝI. (4.16) Computing the time derivative of a Lyapunov function๐ =๐๐๐โค๐ ๐๐๐with๐๐๐=
๐ ๐/๐ ๐for the unperturbed virtual dynamics (4.11), we have using (4.16) that ๐ยค =๐๐๐โค๐ ๐๐๐ =๐๐๐โค( ยค๐ +2๐ ๐ดโ2 ยฏ๐ถโค๐ โ1๐ถ)๐๐๐ โค โ2๐ผ๐ โ๐ฝโฅ๐๐๐โฅ2
which implies that๐ = ๐โ1defines a contraction metric. Since we have ๐โ1I โชฏ ๐ โชฏ ๐โ1I,๐ โฅ ๐โ1โฅ๐๐๐โฅ2, and
โฅฮ(๐ฅ , ๐กห )๐๐๐โฅ โค ๐ยฏ๐0/โ
๐+๐ยฏ๐1๐ยฏ๐ยฏ
โ ๐
โฅ๐๐๐บโฅ2๐น โค ๐ยฏ2
๐0+๐ยฏ2๐ยฏ2๐ยฏ2
๐1๐2
for ๐ in (4.8) and๐บ in (4.9), the bounds (4.13) โ (4.15) follow from the proofs of Theorems 2.4 and 2.5 [2], [3].
Remark 4.1. Although (4.12) is not an LMI due to the nonlinear term โ๐ ๐ฝI on its right-hand side for stochastic systems(4.4), it is a convex constraint asโ๐ ๐ฝ =
โ๐๐ผ๐ =โ๐๐ผ๐0โ๐3๐ผ๐1is a concave function for๐ โR>0[3], [10].
4.2.II CV-STEM Formulation for State Estimation
The estimator (4.7) gives a convex steady-state upper bound of the Euclidean distance between๐ฅand ห๐ฅ as in Theorem 4.1 [1]โ[4].
Theorem 4.4. If(4.12)of Theorem 4.3 holds, then we have the following bound:
๐กโโlim
โ๏ธ
E
โฅ๐ฅหโ๐ฅโฅ2
โค ๐0(๐ผ, ๐ผ๐บ)๐+๐1(๐ผ, ๐ผ๐บ)๐๐ (4.17) where ๐0 = ๐ยฏ๐0/๐ผ, ๐1 = ๐ยฏ๐ยฏ๐ยฏ๐1/๐ผ, ๐ = 1 for deterministic systems (4.8), and ๐0=โ๏ธ
๐ถ๐0/(2๐ผ),๐1=๐ถ๐1/(2โ
2๐ผ๐ถ๐0), and๐ =2for stochastic systems(4.9), with ๐ถ๐0and๐ถ๐0given as๐ถ๐0=๐ยฏ2
๐0(2๐ผโ1
๐บ +1)and๐ถ๐1= ๐ยฏ2๐ยฏ2๐ยฏ2
๐1(2๐ผโ1
๐บ +1).
Proof. The upper bound (4.17) for deterministic systems (4.8) follows from (4.13) with the relation 1 โค โ
๐ โค ๐due to๐ โค ๐. For stochastic systems, we have using (4.14) that
๐ถ๐0๐+๐ถ๐1๐2๐ โค ๐ถ๐0(๐+ (๐ถ๐1/(2๐ถ๐0))๐2)2
due to 1โค ๐ โค ๐2and๐ โR>0. This gives (4.17) for stochastic systems (4.9).
Finally, the CV-STEM estimation framework is summarized in Theorem 4.5 [1]โ[4].
Theorem 4.5. Suppose that๐ผ, ๐ผ๐บ, ๐ยฏ๐0, ๐ยฏ๐1, ๐ยฏ๐0, ๐ยฏ๐1, and๐ฟ๐ in(4.12) and(4.17) are given. If the pair (๐ด, ๐ถ) is uniformly observable, the non-convex optimization problem of minimizing the upper bound(4.17)is equivalent to the following convex optimization problem with the contraction constraint (4.12) and I โชฏ ๐ยฏ โชฏ ๐I of (3.19):
๐ฝโ
๐ถ๐ = min
๐โR>0, ๐โR,๐ยฏโป0
๐0๐+๐1๐๐ +๐2๐(๐, ๐,๐ยฏ) (4.18) s.t.(4.12) and (3.19)
where๐0,๐1, and๐ are as defined in(4.17)of Theorem 4.4,๐2 โRโฅ0, and๐is some performance-based cost function as in Theorem 4.2.
The weight ๐1 for ๐๐ indicates how much we trust the measurement ๐ฆ(๐ก). Using non-zero๐2enables finding contraction metrics optimal in a different sense in terms of ๐. Furthermore, the coefficients of the SDC parameterizations ๐๐ and ๐๐ in Lemma 3.1 (i.e., ๐ด = ร
๐๐,๐๐ด๐ and๐ถ = ร
๐๐,๐๐ถ๐ in (4.5) and (4.6)) can also be treated as decision variables by convex relaxation [9], thereby adding a design flexibility to mitigate the effects of external disturbances while verifying the system observability.
Proof. The proposed optimization problem is convex as its objective and constraints are convex in terms of decision variables ๐, ๐, and ยฏ๐ (see Remark 4.1). Also, larger ยฏ๐๐1and ยฏ๐๐1in (4.3) and (4.4) imply larger measurement uncertainty. Thus by definition of๐1in Theorem 4.4, the larger the weight of๐, the less confident we are in ๐ฆ(๐ก)(see Example 4.2). The last statement on the SDC coefficients for guaranteeing observability follows from Proposition 1 of [9] and Proposition 1 of [1].
Example 4.2. The weights ๐0 and๐1 of the CV-STEM estimation of Theorem 4.5 has an analogous trade-off to the case of the Kalman filter with the process and
sensor noise covariance matrices,๐and๐, respectively, since the term๐0๐in upper bound of the steady-state tracking error in(4.17)becomes dominant if measurement noise is much smaller than process noise (๐ยฏ๐0 โซ ๐ยฏ๐1 or๐ยฏ๐0 โซ ๐ยฏ๐1), and the term ๐1๐๐ becomes dominant if measurement noise is much greater than process noise (๐ยฏ๐0 โช ๐ยฏ๐1or๐ยฏ๐0โช ๐ยฏ๐1). In particular [2], [3],
โข if๐1is much greater than๐0, large measurement noise leads to state estimation that responds slowly to unexpected changes in the measurement ๐ฆ (i.e. small estimation gain due to๐ =๐ โฅ โฅ๐โฅ), and
โข if ๐1is much smaller than๐0, large process noise leads to state estimation that responds fast to changes in the measurement (i.e. large๐ =๐ โฅ โฅ๐โฅ).
This is also because the solution ๐ = ๐โ1 โป 0of the Kalman filter Riccati equa- tion [7, p. 375], ๐ยค = ๐ด๐+๐ ๐ดโค โ๐๐ถโค๐ โ1๐ถ ๐+๐, can be viewed as a positive definite matrix that defines a contraction metric as discussed in Example 2.3.
4.3 Control Contraction Metrics (CCMs)
As briefly discussed in Remark 3.4, the concept of a CCM [11]โ[16] is introduced to extend contraction theory to design differential feedback control laws for control- affine deterministic nonlinear systems (3.1), ๐ฅยค = ๐(๐ฅ , ๐ก) + ๐ต(๐ฅ , ๐ก)๐ข. Let us show that the CCM formulation could also be considered in the CV-STEM control of Theorem 4.2, where similar ideas have been investigated in [12], [17], [18].
Theorem 4.6. Suppose the CV-STEM control of Theorem 4.2 is designed with its contraction condition replaced by the following set of convex constraints along with Iโชฏ ๐ยฏ โชฏ ๐Iof(3.19):
๐ตโคโฅ
โ๐๐ยฏ
๐ ๐ก
โ๐๐๐ยฏ +2 sym ๐ ๐
๐ ๐ฅ
ยฏ ๐
+2๐ผ๐ยฏ
๐ตโฅ โบ0 (4.19)
๐ตโคโฅ
๐๐
๐๐ยฏ โ2 sym ๐ ๐๐
๐ ๐ฅ
ยฏ
๐ ๐ตโฅ =0,โ๐ฅ , ๐ก , ๐ (4.20)
where ๐ตโฅ(๐ฅ , ๐ก) is a matrix whose columns span the cokernel of ๐ต(๐ฅ , ๐ก) defined as coker(๐ต) = {๐ โ R๐|๐ตโค๐ = 0} satisfying ๐ตโค๐ตโฅ = 0, ๐๐(๐ฅ , ๐ก) is the ๐th column of ๐ต(๐ฅ , ๐ก), ๐ยฏ (๐ฅ , ๐ก) = ๐๐(๐ฅ , ๐ก), ๐(๐ฅ , ๐ก) = ๐(๐ฅ , ๐ก)โ1 for ๐(๐ฅ , ๐ก) that defines a contraction metric,๐ =๐, ๐=๐/๐, and๐๐๐น =ร๐
๐=1(๐ ๐น/๐ ๐ฅ๐)๐๐for ๐(๐ฅ , ๐ก) โR๐ and ๐น(๐ฅ , ๐ก) โ R๐ร๐. Then the controlled system (3.1) with ๐๐ = 0, i.e., ๐ฅยค =
๐(๐ฅ , ๐ก) +๐ต(๐ฅ , ๐ก)๐ข, is
1) universally exponentially open-loop controllable
2) universally exponentially stabilizable via sampled-data feedback with arbitrary sample times
3) universally exponentially stabilizable using continuous feedback defined almost everywhere, and everywhere in a neighborhood of the target trajectory
all with rate๐ผ and overshoot ๐ = โ๏ธ
๐/๐ for ๐ and๐ given in ๐I โชฏ ๐ โชฏ ๐Iof (2.26). Such a positive definite matrix๐(๐ฅ , ๐ก)defines a CCM.
Given a CCM, there exists a differential feedback controller๐ฟ๐ข= ๐(๐ฅ , ๐ฟ๐ฅ , ๐ข, ๐ก)that stabilizes the following differential dynamics of(3.1)with๐๐ =0along all solutions (i.e., the closed-loop dynamics is contracting as in Definition 2.3):
๐ฟ๐ฅยค = ๐ด(๐ฅ , ๐ข, ๐ก)๐ฟ๐ฅ+๐ต(๐ฅ , ๐ก)๐ฟ๐ข (4.21)
where ๐ด=๐ ๐/๐ ๐ฅ+ร๐
๐=1(๐ ๐๐/๐ ๐ฅ)๐ข๐and๐ฟ๐ขis a tangent vector to a smooth path of controls at๐ข. Furthermore, it can be computed as follows [11], [12], [16]:
๐ข(๐ฅ(๐ก), ๐ก) =๐ข๐+
โซ 1 0
๐(๐พ(๐, ๐ก), ๐๐๐พ(๐, ๐ก), ๐ข(๐พ(๐, ๐ก), ๐ก), ๐ก)๐๐ (4.22) where๐๐๐พ =๐ ๐พ/๐ ๐,๐พ is the minimizing geodesic with๐พ(0, ๐ก)=๐ฅ๐and๐พ(1, ๐ก) =๐ฅ for ๐ โ [0,1], (๐ฅ๐, ๐ข๐) is a given target trajectory in ๐ฅยค๐ = ๐(๐ฅ๐, ๐ก) + ๐ต(๐ฅ๐, ๐ก)๐ข๐ of (3.3), and the computation of ๐ is given in [11], [12] (see Remark 4.2). Further- more, Theorem 2.4 for deterministic disturbance rejection still holds and the CCM controller (4.22) thus possesses the same sense of oplimality as for the CV-STEM control in Theorem 4.2.
Proof. Since the columns of๐ตโฅ(๐ฅ , ๐ก) span the cokernel of ๐ต(๐ฅ , ๐ก), the constraints (4.19) and (4.20) can be equivalently written as๐ตโฅ(๐ฅ , ๐ก)replaced by๐in coker(๐ต)= {๐ โR๐|๐ตโค๐=0}. Let๐ = ๐ ๐ฟ๐ฅ =๐โ1๐ฟ๐ฅ. Then multiplying (4.19) and (4.20) by ๐โ1> 0 and rewriting them using๐yields
๐ฟ๐ฅโค๐ ๐ต=0 โ ๐ฟ๐ฅโค( ยค๐+2 sym(๐ ๐ด) +2๐ผ ๐)๐ฟ๐ฅ <0 (4.23) where ๐ด = ๐ ๐/๐ ๐ฅ + ร๐
๐=1(๐ ๐๐/๐ ๐ฅ)๐ข๐ for the differential dynamics (4.21). The relation (4.23) states that ๐ฟ๐ฅ orthogonal to the span of actuated directions ๐๐ is naturally contracting, thereby implying the stabilizability of the system (3.1) with ๐๐ =0, i.e.,๐ฅยค = ๐(๐ฅ , ๐ก) +๐ต(๐ฅ , ๐ก)๐ข. See [11], [12], [16] (and [17] on the CV-STEM formulation) for the rest of the proof.
Example 4.3. Let us consider a case where๐of the differential feedback controller does not depend on๐ข, and is defined explicitly by๐(๐ฅ , ๐ก) โป0as follows [17]:
๐ข =๐ข๐โ
โซ ๐ฅ ๐ฅ๐
๐ (๐, ๐ก)โ1๐ต(๐, ๐ก)โค๐(๐, ๐ก)๐ฟ๐ (4.24) where ๐ (๐ฅ , ๐ก) โป 0 is a given weight matrix and ๐ is a smooth path that connects ๐ฅ to ๐ฅ๐. Since (4.24) yields ๐ฟ๐ข = โ๐ (๐ฅ , ๐ก)โ1๐ต(๐ฅ , ๐ก)โค๐(๐ฅ , ๐ก)๐ฟ๐ฅ, the contraction conditions(4.19)and(4.20)could be simplified as
โ ๐๐ยฏ
๐ ๐ก
โ๐๐๐ยฏ +2 sym ๐ ๐
๐ ๐ฅ
ยฏ ๐
โ๐ ๐ต ๐ โ1๐ตโค โชฏ โ2๐ผ๐ยฏ
โ๐๐
๐
ยฏ
๐ +2 sym ๐ ๐๐
๐ ๐ฅ
ยฏ ๐
=0 (4.25)
yielding the convex optimization-based control synthesis algorithm of Theorem 4.6 independently of(๐ฅ๐, ๐ข๐) similar to that of Theorem 4.2 (see [17] for details).
Example 4.4. If๐ตis of the form [0,I]โค for the zero matrix0 โR๐1ร๐ and identity matrix I โ R๐2ร๐ with ๐ = ๐1+๐2, the condition (4.25) says that ๐ should not depend on the last๐2state variables [11].
Remark 4.2. We could consider stochastic perturbation in Theorem 4.6 using Theorem 2.5, even with the differential control law of the form (4.22) or (4.24) as demonstrated in [17]. Also, although the relation(4.20)or(4.25)is not included as a constraint in Theorem 4.2 for simplicity of discussion, the dependence of๐ยคยฏ on ๐ขin Theorem 4.2 can be removed by using it in a similar way to [17].
As stated in(4.22), the computation of the differential feedback gain ๐(๐ฅ , ๐ฟ๐ฅ , ๐ข, ๐ก) and minimizing geodesics ๐พ is elaborated in [11], [12]. For example, if ๐ is state-independent, then geodesics are just straight lines.
Let us again emphasize that, as delineated in Sec. 3.1, the differences between the SDC- and CCM-based CV-STEM frameworks in Theorems 4.2 and 4.6 arise only from their different form of controllers in ๐ข = ๐ข๐ โ๐พ(๐ฅ โ๐ฅ๐) of (3.10) and ๐ข =๐ข๐+โซ1
0 ๐ ๐๐of (4.22), leading to the trade-offs outlined in Table 3.1.
4.4 Remarks in CV-STEM Implementation
We propose some useful techniques for the practical application of the CV-STEM in Theorems 4.2, 4.5, and 4.6. Note that the CV-STEM requires solving a convex optimization problem at each time instant, but its solution can be approximated with formal stability guarantees to enable faster computation using machine learning techniques as shall be seen in Chapter 6.