Chapter V: Data-driven Approach in the Noiseless Case
5.5 Distributed AND Localized algorithm for Data-driven MPC
amount of data needed to parametrize the behavior of the system does not scale with the size of the network but rather withπ, the size of the localized region, which is usually much smaller than π. To the best of our knowledge, this is the first such result. As we show next, this will prove key in extending data-driven SLS to the distributed setting.
Definition 12. Given the partition{π―1, ...,π―π}, a functional/set isrow-wise separable if:
β’ for a functional, π(π½) = Γπ
π=1ππ π½(π―π,:)
for some functionals ππ for π = 1, ..., π.
β’ for a set,π½β Pif and only ifπ½(π―π,:) β Pπβπfor some setsPπforπ=1, ..., π. An analogous definition exists forcolumn-wise separablefunctionals and sets, where the partition{π 1, ...,π π}entails the columns ofπ½, i.e.,π½(:,π π).
By Assumption 1 the objective function and the safety/saturation constraints in equation (5.11) are row-separable in terms ofπ½. At the same time, the achievability and locality constraints (5.9), (5.10) are column-separable in terms of G. Hence, the data-driven MPC subroutine becomes:
minimize
π½,πΏ,{Gπ}π
π=1
π
βοΈ
π=1
ππ( [π½]π[π₯0]inπ(π))
s.t.
π₯0=π₯(π‘), Gπsatisfies (5.10),
[π½π₯]π[π₯0]inπ(π) β Xπ, [π½π’]π[π₯0]inπ(π) β Uπ, [πΏπ]in
π(π) =π»πΏ( [xΛ]in
π(π),[uΛ]in
π(π+1))Gπ
βπ =1, . . . , π , π½=πΏ.
(5.12)
Notice that the objective function and the constraints decompose across the π- localized neighborhoods of each subsystemπ. Given this structure, we can make use of ADMM to decompose this problem into row-wise local subproblems in terms of [π½]π and column-wise local subproblems in terms of Gπ andπΏπ, both of which can also be parallelized across the subsystems. The ADMM subroutine iteratively updates the variables as
[π½]{π+1}
π =





ο£²





ο£³
argmin
[π½]π
ππ( [π½]π[π₯0]inπ(π))+
π 2
ππ(π½,πΏ{π},π²{π})
2
πΉ
π .π‘ . [π½π₯]π[π₯0]in
π(π) β Xπ,
[π½π’]π[π₯0]inπ(π) β Uπ, π₯0=π₯(π‘).





ο£½





ο£Ύ
(5.13a)
[πΏπ]{π+1}
inπ(π) =




ο£²



ο£³
argmin
[πΏπ]inπ(π),Gπ
β₯πinπ(π) [π½π]{π+1},[πΏπ],[π²π]{π}
β₯2πΉ
π .π‘ . [πΏπ]inπ(π) = [π»πΏ(xΛ,uΛ)]inπ(π)Gπ, Gπ satisfies (5.10).




ο£½



ο£Ύ
(5.13b)
[π²]{π+1}
π =ππ(π½{π+1},πΏ{π+1},π²{π}) (5.13c) where we define
πβ(π½,πΏ,π²) := [π½]ββ [πΏ]β+ [π²]β withβdenoting a subsystem or collection of subsystems, and
[π»πΏ(xΛ,uΛ)]inπ(π) :=π»πΏ( [Λx]inπ(π),[uΛ]inπ(π+1)).
We note that to solve this subroutine, and in particular optimization (5.13b), each subsystem only needs to collect trajectory of states and control actions of subsystems that are at mostπ+2 hops away. This subroutine thus constitutes a distributed and localized solution. We also emphasize that the trajectory length only needs to scale with theπ-localized system size instead of the global system size. The full algorithm is given in Algorithm 5.
Theoretical guarantees
It is worth noting that this reformulation is equivalent to the closed-loop DLMPC also introduced in Chapter 2, with the achievability constraint ππ΄ π΅π½= πΌ replaced by the data-driven parametrization in terms of Gand the Hankel matrix π»πΏ. For this reason, guarantees derived for the DLMPC formulation (3.1) directly apply to problem (5.11) when the constraint setsX andU are polytopes.
Convergence
Algorithm 5 relies on ADMM to separate both row, and column-wise computations, in terms of π½ and G, respectively. Each of these is then distributed into the subsystems in the network, and a communication protocol is established to ensure the ADMM steps are properly followed. Hence, one can guarantee the convergence
Algorithm 5Subsystemβsπimplementation of D3LMPC
Input: tolerance parameters ππ, ππ > 0, Hankel matrices [π»πΏ(xΛ,uΛ)]inπ(π+1) constructed from arbitrarily generated PE trajectories Λxinπ(π+1),uΛinπ(π+1).
1: Measure local state [π₯0]π, π β 0.
2: Share measurement [π₯0]πwithoutπ(π) and receive[π₯0]π from π βinπ(π).
3: Solve optimization problem (5.13a).
4: Share [π½]{π+1}
π withoutπ(π). Receive the corresponding [π½]{π+1}
π frominπ(π) and construct [π½π]{π+1}
inπ(π).
5: Perform update (5.13b).
6: Share [πΏπ]{inπ+1}
π(π) withoutπ(π). Receive the corresponding [πΏπ]{π+1}in
π(π) from π β
inπ(π)and construct [πΏ]{π+1}
π .
7: Perform update (5.13c).
8: if
[πΏ]{π+1}
π β [π½]{π+1}
π
πΉ
β€ ππ and
[ [π½]{π+1}
π β [π½]{π}
π
πΉ
β€ ππ: Apply computed control action:
[π’0]π =HπΏ( [u]Λ inπ(π))Gπ[π₯0]inπ(π), and return to step 1.
else:
Setπ β +1 and return to step 3.
of the data-driven version of DLMPC in the same way that convergence of model- based DLMPC is shown: by leveraging the convergence result of ADMM in [19].
For additional details see Lemma 6.
Recursive feasibility
Recursive feasibility for formulation (2.12) is guaranteed by means of a localized maximally invariant terminal set Xπ. This set can be computed in a distributed manner and with local information only as described in Chapter 3. In particular, a closed-loop mapπ½for the unconstrained localized closed-loop system has to be computed. In the model-based SLS problem with quadratic cost, a solution exists for the infinite-horizon case [71], which can be done in a distributed manner and with local information only. When no model is available, the same problem can be solved using the localized data-driven SLS approach introduced in Β§IV with a finite-horizon approximation, which also allows for a distributed synthesis with only local data. The length of the time horizon chosen to solve the localized data-driven SLS problem might slightly impact the conservativeness of the terminal set, but since the conservatism in the FIR approach decays exponentially with the horizon length, this harm in performance is not expected to be substantial for usual values of the horizon. Onceπ½for the unconstrained localized closed-loop system has been
computed, Algorithm 2 can be used to synthesize this terminal set in a distributed and localized manner. Therefore, a terminal set that guarantees recursive feasibility for D3LMPC can be computed in a distributed manner offline using only local information and without the need for a system model.
Asymptotic stability
Similar to recursive feasibility, asymptotic stability for the D3LMPC problem (5.11) is directly inherited from the asymptotic stability guarantee for model-based DLMPC. In particular, adding a terminal cost based on the terminal set previously described is a sufficient condition to guarantee asymptotic stability of the DLMPC problem (3.1). Moreover, such cost can be incorporated in the D3LMPC formula- tion in the same way as in the model-based DLMPC problem, and the structure of the resulting problem is analogous. This terminal cost introduces coupling among subsystems, but the coupling can be dealt with by solving step 3 of Algorithm 5 via local consensus. Notice that since step 3 of Algorithm 5 is written in terms of π½, Algorithm 3 can be directly used to handle this coupling.