LEARNING AND CONTROL IN LINEAR TIME-VARYING SYSTEMS
4.3 Stability and Identification of Random Asynchronous LTI Systems In this section, we introduce the random asynchronous LTI systems, a new form ofIn this section, we introduce the random asynchronous LTI systems, a new form of
4.3.4 System Identification for Randomized LTI Systems
mean-square stability of the randomized LTI system and the stability of Ado not imply each other, i.e.,π(S)<1 andπ(A)<1 are not equivalent in general. Note that Lemma 4.4 provides the precise characterization of the rightmost axes of the plots in Figure 4.4.
In order to visualize the convergence behavior of the randomized updates, we consider a numerical test example of size ππ₯ =2 with a constant input (i.e., fixed- point iteration), and initializeπ₯0with independent Gaussian random variables (left- most block in Figure 4.5). Then, the distribution of the state vector π₯π‘ (at time π‘) follows a Gaussian mixture model (GMM) due to the randomized nature of the updates (See Figure 4.5). Furthermore, the stability of the matrix S ensures that the mean ofπ₯π‘ converges to the fixed-point of the system while the variance of π₯π‘ converges to zero.
The key insight to the convergence behavior in Figure 4.5 is as follows: When rep- resented as a switching system, the randomized LTI model (4.29) switches between 2ππ₯ systems randomly, and it can be shown that all these 2ππ₯ systems (including the original system) have the same fixed-point. It should also be noted that not all 2ππ₯ systems are stable by themselves, and an arbitrary switching does not necessarily ensure the convergence. Nevertheless, with a careful selection of the probability, the randomized model can obtain convergence even when the synchronous system is unstable.
When the system is mean-square stable, the steady-state covariance matrix, πͺ, is given as
vec(πͺ) =(IβS)β1 π2I+ (πβ π2)J
vec(B U Bβ€) +π2
π€vec(I)
. (4.33) Whenπ =1, we haveπ(π₯) =Aπ₯Aβ€, which implies thatπͺ=AπͺAβ€+B U Bβ€+π2
π€I and π(S) =π2(A). So, we have π(S) < 1 if and only ifπ(A) < 1 for synchronous LTI systems, which recovers the well-known stability result in the classical systems theory.
Figure 4.5: Evolution of the state vector for a mean-square stable (but synchronously unstable) 2-dimensional randomized LTI system with a fixed input and Gaussian initialization
First, recall the Markov chain central limit theorem (MC-CLT). Assume that we have a Markov chain at its stationary distribution. MC-CLT states that, the sample average of any measurable, finite-variance and real-valued function of a sequence of πvariables from this Markov chain converge to a Gaussian distribution asπβ β, where mean is the expected value of this function at the stationary distribution and the variance linearly decays inπ[129].
Notice that the randomized updates of (4.29) form an ergodic Markov chain (due to independent selection in every iteration) and the stability of the system guar- antees the stationary distribution. We also know that the stable systems converge exponentially fast to their steady state, i.e., Markov chain formed by (4.29) quickly approaches to its stationary distribution. In light of these observations, we can deduce that, as the number of collected input-output samplesπ increases, the sam- ple state correlation and input-output cross correlation matrices converge to their expected values with the rate of 1/β
π. In particular, given a sequence of inputs and outputs{π₯0, π’0, π₯1, . . . , π’πβ1, π₯π}, let
C0= 1 π
βοΈπβ1 π‘=0
"
π₯π‘ π’π‘
# "
π₯π‘ π’π‘
#β€
, C1= 1 π
βοΈπ
π‘=1
π₯π‘
"
π₯π‘β1 π’π‘β1
#β€
. (4.34)
According to MC-CLT, as π β β, C0 and C1 converge to E[C0] and E[C1] respectively, where
E[C0] =
"
πͺ 0 0 U
#
, E[C1] = h
Aπͺ BU i
. (4.35)
Therefore, usingC1Cβ01converges to the average state transition and input matrices h
A B i
. Notice that C1Cβ01 is in fact the solution of the following least squares problem:
arg min
Ξ
βοΈπ
π‘=1tr π₯π‘βΞ
"
π₯π‘β1 π’π‘β1
#
π₯π‘ βΞ
"
π₯π‘β1 π’π‘β1
# β€
. (4.36)
Thus, we are guaranteed to recover the average system consistently via (4.36). This result could be extended to recover first πΎ Markov parameters of the randomized partially observable LTI systems. Define E = [Iππ₯ 0] β Rππ₯Γ(ππ₯+ππ’). Then, the extended Lyapunov equation can be written as
E
"
πͺ 0 0 U
# Eβ€=
h A B
i
"
πͺ0 0 U
# h
A B iβ€
+
1/πβ1 h
AβI B i
"
πͺ0 0 U
# h
AβI B iβ€
βI+π2
π€I, (4.37) We know that covariance matrices of the state variables πͺ and inputs U must satisfy (4.37) for a stable randomized LTI system. The central idea for our system identification method is to exploit this fact and recover the randomization probability π, the noise covarianceπ2
π€ and the system parametersA,Bof a stable randomized LTI system. Therefore, we can write extended Lyapunov equation (4.37) in terms ofC0andC1and due to (4.35) expect to havely(C0,C1) =0, where
ly(C0,C1)BEC0Eβ€βC1Cβ10 Cβ€1β1 π
β1 (C1Cβ10 βE)C0(C1Cβ10 βE)β€
βIβπ2
π€I. Thus, to identify the underlying system dynamics, we propose to solve the following:
π,b bπ2
π€ =arg min
π,ππ€2
β₯ly(C0,C1) β₯2F. (4.38) This problem can be further simplified to
bπ, bπ2
π€ =arg min
π,ππ€2
π΄1β (1/π) π΄2βπ2
π€I
2
F, (4.39)
whereπ΄2=
(C1Cβ10 βE)C0(C1Cβ10 βE)β€
βIandπ΄1=E C0Eβ€βC1Cβ10 Cβ€1+π΄2. Notice thatπandπ2
π€appear decoupled in (4.39). Therefore, we can first solve (4.39) forbπ2
π€ for a fixed value of π to get an optimal solution. Then, substitutingbπ2
π€ into the problem and solving for bπwe obtain the optimal estimate for π. The described procedure yields the following optimal estimates:
πb=
ππ₯ tr(π΄β€2 π΄2) βtr2(π΄2) ππ₯tr(π΄β€1 π΄2) βtr(π΄1)tr(π΄2),
bπ2
π€ = tr(π΄1) β (1/πb) tr(π΄2) ππ₯
. (4.40) Using the estimate of randomization probability bπ and C1Cβ10 = [b
A bB], i.e., the estimate of average system transition parameters, the underlying system parameters could be recovered as Ab=(1/πb)b
A+ (1β1/bπ)Iππ₯ and bB=(1/bπ)b
B. To study the performance of the proposed system identification method, we consider a random- ized LTI system with state transition matrix of A1 and a random B with π = 0.5
100 101 102 103 104 105 106
T(number of samples)
10-3 10-2 10-1 100 101
AverageEstimationError
|pβpb|
|ΟwβbΟw| 1/β
T
100 101 102 103 104 105 106
T(number of samples)
10-3 10-2 10-1 100 101
AverageEstimationError
kAβAkbF/kAkF
kBβBkbF/kBkF
1/β T
Figure 4.6: Average estimation error for the unknown system parameters of the stable randomized LTI system with state transition matrix ofA1and randomBfor 100 independent single trajectories
which guarantees the stability (verified by Lemma 4.4) and set ππ€ = 1. We run 100 independent single trajectories and present the average rate of decay for the estimation errors of π, π2
π€ in the first plot andAandBin the second plot of Figure 4.6. Notice that the estimation errors behave irregularly at the beginning where there are few samples, corresponding to burn-in period to converge to steady-state.
On the other hand, Figure 4.6 show that, as predicted by MC-CLT, the estimation errors decay with 1/β
π rate as we get more samples. This estimation error rate is the optimal behavior in linear regression problems with independent noise and covariates [106]. This depicts the consistency and efficiency of the proposed system identification method for randomized LTI systems.