• Tidak ada hasil yang ditemukan

System Identification for Randomized LTI Systems

Dalam dokumen Learning and Control of Dynamical Systems (Halaman 139-142)

LEARNING AND CONTROL IN LINEAR TIME-VARYING SYSTEMS

4.3 Stability and Identification of Random Asynchronous LTI Systems In this section, we introduce the random asynchronous LTI systems, a new form ofIn this section, we introduce the random asynchronous LTI systems, a new form of

4.3.4 System Identification for Randomized LTI Systems

mean-square stability of the randomized LTI system and the stability of Ado not imply each other, i.e.,𝜌(S)<1 and𝜌(A)<1 are not equivalent in general. Note that Lemma 4.4 provides the precise characterization of the rightmost axes of the plots in Figure 4.4.

In order to visualize the convergence behavior of the randomized updates, we consider a numerical test example of size 𝑑π‘₯ =2 with a constant input (i.e., fixed- point iteration), and initializeπ‘₯0with independent Gaussian random variables (left- most block in Figure 4.5). Then, the distribution of the state vector π‘₯𝑑 (at time 𝑑) follows a Gaussian mixture model (GMM) due to the randomized nature of the updates (See Figure 4.5). Furthermore, the stability of the matrix S ensures that the mean ofπ‘₯𝑑 converges to the fixed-point of the system while the variance of π‘₯𝑑 converges to zero.

The key insight to the convergence behavior in Figure 4.5 is as follows: When rep- resented as a switching system, the randomized LTI model (4.29) switches between 2𝑑π‘₯ systems randomly, and it can be shown that all these 2𝑑π‘₯ systems (including the original system) have the same fixed-point. It should also be noted that not all 2𝑑π‘₯ systems are stable by themselves, and an arbitrary switching does not necessarily ensure the convergence. Nevertheless, with a careful selection of the probability, the randomized model can obtain convergence even when the synchronous system is unstable.

When the system is mean-square stable, the steady-state covariance matrix, πšͺ, is given as

vec(πšͺ) =(Iβˆ’S)βˆ’1 𝑝2I+ (π‘βˆ’ 𝑝2)J

vec(B U B⊀) +𝜎2

𝑀vec(I)

. (4.33) When𝑝 =1, we haveπœ™(π‘₯) =Aπ‘₯A⊀, which implies thatπšͺ=AπšͺA⊀+B U B⊀+𝜎2

𝑀I and 𝜌(S) =𝜌2(A). So, we have 𝜌(S) < 1 if and only if𝜌(A) < 1 for synchronous LTI systems, which recovers the well-known stability result in the classical systems theory.

Figure 4.5: Evolution of the state vector for a mean-square stable (but synchronously unstable) 2-dimensional randomized LTI system with a fixed input and Gaussian initialization

First, recall the Markov chain central limit theorem (MC-CLT). Assume that we have a Markov chain at its stationary distribution. MC-CLT states that, the sample average of any measurable, finite-variance and real-valued function of a sequence of 𝑛variables from this Markov chain converge to a Gaussian distribution as𝑛→ ∞, where mean is the expected value of this function at the stationary distribution and the variance linearly decays in𝑛[129].

Notice that the randomized updates of (4.29) form an ergodic Markov chain (due to independent selection in every iteration) and the stability of the system guar- antees the stationary distribution. We also know that the stable systems converge exponentially fast to their steady state, i.e., Markov chain formed by (4.29) quickly approaches to its stationary distribution. In light of these observations, we can deduce that, as the number of collected input-output samples𝑇 increases, the sam- ple state correlation and input-output cross correlation matrices converge to their expected values with the rate of 1/√

𝑇. In particular, given a sequence of inputs and outputs{π‘₯0, 𝑒0, π‘₯1, . . . , π‘’π‘‡βˆ’1, π‘₯𝑇}, let

C0= 1 𝑇

βˆ‘οΈπ‘‡βˆ’1 𝑑=0

"

π‘₯𝑑 𝑒𝑑

# "

π‘₯𝑑 𝑒𝑑

#⊀

, C1= 1 𝑇

βˆ‘οΈπ‘‡

𝑑=1

π‘₯𝑑

"

π‘₯π‘‘βˆ’1 π‘’π‘‘βˆ’1

#⊀

. (4.34)

According to MC-CLT, as 𝑇 β†’ ∞, C0 and C1 converge to E[C0] and E[C1] respectively, where

E[C0] =

"

πšͺ 0 0 U

#

, E[C1] = h

Aπšͺ BU i

. (4.35)

Therefore, usingC1Cβˆ’01converges to the average state transition and input matrices h

A B i

. Notice that C1Cβˆ’01 is in fact the solution of the following least squares problem:

arg min

Θ

βˆ‘οΈπ‘‡

𝑑=1tr π‘₯π‘‘βˆ’Ξ˜

"

π‘₯π‘‘βˆ’1 π‘’π‘‘βˆ’1

#

π‘₯𝑑 βˆ’Ξ˜

"

π‘₯π‘‘βˆ’1 π‘’π‘‘βˆ’1

# ⊀

. (4.36)

Thus, we are guaranteed to recover the average system consistently via (4.36). This result could be extended to recover first 𝐾 Markov parameters of the randomized partially observable LTI systems. Define E = [I𝑑π‘₯ 0] ∈ R𝑑π‘₯Γ—(𝑑π‘₯+𝑑𝑒). Then, the extended Lyapunov equation can be written as

E

"

πšͺ 0 0 U

# E⊀=

h A B

i

"

πšͺ0 0 U

# h

A B i⊀

+

1/π‘βˆ’1 h

Aβˆ’I B i

"

πšͺ0 0 U

# h

Aβˆ’I B i⊀

βŠ™I+𝜎2

𝑀I, (4.37) We know that covariance matrices of the state variables πšͺ and inputs U must satisfy (4.37) for a stable randomized LTI system. The central idea for our system identification method is to exploit this fact and recover the randomization probability 𝑝, the noise covariance𝜎2

𝑀 and the system parametersA,Bof a stable randomized LTI system. Therefore, we can write extended Lyapunov equation (4.37) in terms ofC0andC1and due to (4.35) expect to havely(C0,C1) =0, where

ly(C0,C1)BEC0EβŠ€βˆ’C1Cβˆ’10 C⊀1βˆ’1 𝑝

βˆ’1 (C1Cβˆ’10 βˆ’E)C0(C1Cβˆ’10 βˆ’E)⊀

βŠ™Iβˆ’πœŽ2

𝑀I. Thus, to identify the underlying system dynamics, we propose to solve the following:

𝑝,b b𝜎2

𝑀 =arg min

𝑝,πœŽπ‘€2

βˆ₯ly(C0,C1) βˆ₯2F. (4.38) This problem can be further simplified to

b𝑝, b𝜎2

𝑀 =arg min

𝑝,πœŽπ‘€2

𝑴1βˆ’ (1/𝑝) 𝑴2βˆ’πœŽ2

𝑀I

2

F, (4.39)

where𝑴2=

(C1Cβˆ’10 βˆ’E)C0(C1Cβˆ’10 βˆ’E)⊀

βŠ™Iand𝑴1=E C0EβŠ€βˆ’C1Cβˆ’10 C⊀1+𝑴2. Notice that𝑝and𝜎2

𝑀appear decoupled in (4.39). Therefore, we can first solve (4.39) forb𝜎2

𝑀 for a fixed value of 𝑝 to get an optimal solution. Then, substitutingb𝜎2

𝑀 into the problem and solving for b𝑝we obtain the optimal estimate for 𝑝. The described procedure yields the following optimal estimates:

𝑝b=

𝑑π‘₯ tr(π‘΄βŠ€2 𝑴2) βˆ’tr2(𝑴2) 𝑑π‘₯tr(π‘΄βŠ€1 𝑴2) βˆ’tr(𝑴1)tr(𝑴2),

b𝜎2

𝑀 = tr(𝑴1) βˆ’ (1/𝑝b) tr(𝑴2) 𝑑π‘₯

. (4.40) Using the estimate of randomization probability b𝑝 and C1Cβˆ’10 = [b

A bB], i.e., the estimate of average system transition parameters, the underlying system parameters could be recovered as Ab=(1/𝑝b)b

A+ (1βˆ’1/b𝑝)I𝑑π‘₯ and bB=(1/b𝑝)b

B. To study the performance of the proposed system identification method, we consider a random- ized LTI system with state transition matrix of A1 and a random B with 𝑝 = 0.5

100 101 102 103 104 105 106

T(number of samples)

10-3 10-2 10-1 100 101

AverageEstimationError

|pβˆ’pb|

|Οƒwβˆ’bΟƒw| 1/√

T

100 101 102 103 104 105 106

T(number of samples)

10-3 10-2 10-1 100 101

AverageEstimationError

kAβˆ’AkbF/kAkF

kBβˆ’BkbF/kBkF

1/√ T

Figure 4.6: Average estimation error for the unknown system parameters of the stable randomized LTI system with state transition matrix ofA1and randomBfor 100 independent single trajectories

which guarantees the stability (verified by Lemma 4.4) and set πœŽπ‘€ = 1. We run 100 independent single trajectories and present the average rate of decay for the estimation errors of 𝑝, 𝜎2

𝑀 in the first plot andAandBin the second plot of Figure 4.6. Notice that the estimation errors behave irregularly at the beginning where there are few samples, corresponding to burn-in period to converge to steady-state.

On the other hand, Figure 4.6 show that, as predicted by MC-CLT, the estimation errors decay with 1/√

𝑇 rate as we get more samples. This estimation error rate is the optimal behavior in linear regression problems with independent noise and covariates [106]. This depicts the consistency and efficiency of the proposed system identification method for randomized LTI systems.

Dalam dokumen Learning and Control of Dynamical Systems (Halaman 139-142)