• Tidak ada hasil yang ditemukan

Fundamental Trade-offs

Dalam dokumen Learning-Augmented Control and Decision-Making (Halaman 191-198)

LEARNING POWER SYSTEM PARAMETERS FROM LINEAR MEASUREMENTS

6.3 Fundamental Trade-offs

informationI(𝐺;A|B) from above byH(A) βˆ’H(Z). The general converse stated in Theorem 6.3.1 is used in asserting the results on worst-case sample complexity in Theorem 6.4.1. Next, we analyze the sufficient condition for recovering a graph matrixY(𝐺). Before proceeding to the results, we introduce a novel characterization of the distributionG𝑛, from which a graph𝐺 is sampled. In particular, the graph𝐺 is allowed to have high-degree nodes.

Characterization of Graph Distributions

Let 𝑑𝑗(𝐺) denote the degree of node 𝑗 ∈ V in𝐺. Denote byVLarge(πœ‡) := {𝑗 ∈ V

𝑑𝑗(𝐺) > πœ‡}the set of nodes having degrees greater than thethreshold parameter 0 ≀ πœ‡ ≀ π‘›βˆ’2 and VSmall(πœ‡) := V\VLarge(πœ‡) the set of nodes for all πœ‡-sparse column vectors ofY. With acounting parameter0 ≀ 𝐾 ≀ 𝑛, we define a set of graphs wherein each graph consists of no more than𝐾 nodes with degree larger than πœ‡, denoted byC(𝑛, πœ‡, 𝐾) :={𝐺 ∈C(𝑛) |

VLarge(πœ‡)

≀ 𝐾}.The following definition characterizes graph distributions.

Definition 6.3.1 ((πœ‡, 𝐾 , 𝜌)-sparse distribution). A graph distribution G𝑛 is said to be (πœ‡, 𝐾 , 𝜌)-sparseif assuming that𝐺 is distributed according to G𝑛, then the probability that𝐺belongs toC(𝑛, πœ‡, 𝐾) is larger than 1βˆ’πœŒ,i.e.,

PG𝑛(𝐺 βˆ‰C(𝑛, πœ‡, 𝐾)) ≀ 𝜌 . (6.10) 1) Uniform Sampling of Trees:

Based on the definition above, for particular graph distributions, we can find the associated parameters. We exemplify by considering two graph distributions introduced in Example 7. Denote byUT(𝑛) the uniform distribution on the setT(𝑛) of all trees with𝑛nodes.

Lemma 22. For anyπœ‡β‰₯ 1and𝐾 >0, the distributionUT(𝑛) is(πœ‡, 𝐾 ,1/𝐾)-sparse.

2) ErdΕ‘s-RΓ©nyi(𝑛, 𝑝)model:

Denote by GER(𝑛, 𝑝) the graph distribution for the ErdΕ‘s-RΓ©nyi (𝑛, 𝑝) model.

Similarly, the lemma below classifiesGER(𝑛, 𝑝) into a(πœ‡, 𝐾 , 𝜌)-sparse distribution with appropriate parameters.

Lemma 23. For any πœ‡(𝑛, 𝑝)that satisfies πœ‡(𝑛, 𝑝) β‰₯ 2𝑛 β„Ž(𝑝)/(ln 1/𝑝)and 𝐾 >0, the distributionGER(𝑛, 𝑝)is(πœ‡, 𝐾 , 𝑛exp(βˆ’π‘› β„Ž(𝑝))/𝐾)-sparse.

The proofs of Lemmas 22 and 23 are in Appendix 6.D.

Remark 11. It is worth noting that the(πœ‡, 𝐾 , 𝜌)-sparsityis capable of characterizing any arbitrarily chosen distribution. The interesting part is that for some of the well-known distributions, such as GER(𝑛, 𝑝), this sparsity characterization offers a method that can be used in the analysis and moreover, it leads to an exact characterizationof sample complexity for the noiseless case. Therefore, for the particular examples presented in Lemma 22 and Lemma 23, the selected threshold and counting parameters for both of them are β€œtight” (up to multiplicative factors), in the sense that the corresponding sample complexity matches (up to multiplicative factors) the lower bounds derived from Theorem 6.3.1. This can be seen in Corollary 6.4.1 and 6.4.2.

Algorithm 9: A Three-stage Recovery Scheme. The first stage focuses on solving each column of the matrixYindependently usingβ„“1-minimization. In the second stage, the recovery correctness of the first stage is further verified via consistency-checking, which utilizes the fact that the matrix to be recoveredYis symmetric. The parameter𝛾 is set to zero for the analysis of noiseless parameter reconstruction.

Data: Matrices of measurementsAandB Result: Estimated graph matrixX

Step (a): Recovering columns independently:

for 𝑗 ∈ V do

Solve the followingβ„“1-minimization and obtain an optimalX:

minimize 𝑋𝑗

1

subject to||B𝑋𝑗 βˆ’ 𝐴𝑗||2 ≀ 𝛾 , 𝑋𝑗 ∈F𝑛.

end

Step (b): Consistency-checking:

forS βŠ† V with|S |=π‘›βˆ’πΎ do

if |𝑋𝑖, 𝑗 βˆ’π‘‹π‘— ,𝑖|> 2𝛾 for some𝑖, 𝑗 ∈ Sthen continue;

end else

for 𝑗 ∈ S do

Step (c): Resolving unknown entries:

Update 𝑋S

𝑗 by solving the linear system:

BS𝑋S

𝑗 = π΄π‘—βˆ’BS𝑋S

𝑗 . end

end break;

end

return X=(𝑋1, . . . , 𝑋𝑛);

Sufficient Conditions

In this subsection, we consider the sufficient conditions (achievability) for parameter reconstruction. The proofs rely on constructing a three-stage recovery scheme (Algorithm 9), which contains three steps –column-retrieving,consistency-checking andsolving unknown entries. The worst-case running time of this scheme depends on the underlying distributionG𝑛4. The scheme is presented as follows.

Figure 6.3.1: The recovery of a graph matrixY using the three-stage scheme in Algorithm 10. Theπ‘›βˆ’πΎ columns ofYcolored by gray are first recovered via the β„“1-minimization (6.11a)-(6.11c) in step (a), after they are accepted by passing the consistency check in step (b). Then, symmetry is used for recovering the entries in the matrix marked by green. Leveraging the linear measurements again, in step (c), the remaining𝐾2entries in the white symmetric sub-matrix are solved using Equation (6.12).

1) Three-stage Recovery Scheme:

Step (a): Retrieving columns. In the first stage, using β„“1-norm minimization, we recover each column ofYbased on (6.1):

minimize 𝑋𝑗

1 (6.11a)

subject to ||Bπ‘‹π‘—βˆ’ 𝐴𝑗||2 ≀ 𝛾 , (6.11b)

𝑋𝑗 ∈F𝑛. (6.11c)

Let 𝑋S

𝑗 := (𝑋𝑖, 𝑗)π‘–βˆˆS be a length-|S |column vector consisting of |S|coordinates in 𝑋𝑗, the 𝑗-th retrieved column. We do not restrict the methods for solving theβ„“1-norm

4Although for certain distributions, the computational complexity is not polynomial in𝑛, the scheme still provides insights on the fundamental trade-offs between the number of samples and the probability of error for recovering graph matrices. Furthermore, motivated by the scheme, a polynomial-time heuristic algorithm is provided in Section 6.5 and experimental results are reported in Section 6.6.

minimization in (6.11a)-(6.11c), as long as there is a unique solution for sparse columns with fewer than πœ‡non-zeros (provided enough number of measurements and the parameter πœ‡ >0 is defined in Definition 6.3.1).

Step (b): Checking consistency.

In the second stage, we check for error in the decoded columns𝑋1, . . . , 𝑋𝑛using the symmetry property (perturbed by noise) of the graph matrixY. Specifically, we fix a subsetS βŠ† V with a given size|S|=π‘›βˆ’πΎfor some integer50≀ 𝐾 ≀ 𝑛. Then we check if|𝑋𝑖, 𝑗 βˆ’ 𝑋𝑗 ,𝑖| ≀ 2𝛾 for all𝑖, 𝑗 ∈ S. If not, we choose a different setSof the same size. This procedure stops until either we find such a subsetSof columns, or we go through all possible subsets without finding one. In the latter case, an error is declared and the recovery is unsuccessful. It remains to recover the vectors 𝑋𝑗 for

𝑗 ∈ S.

Step (c): Resolving unknown entries. In the former case, for each vector𝑋𝑗, 𝑗 ∈ S, we accept its entries 𝑋𝑖, 𝑗, 𝑖 ∈ S, as correct and therefore, according to the symmetry assumption, we know the entries 𝑋𝑖, 𝑗, 𝑖 ∈ S, 𝑗 ∈ S (equivalently {𝑋S

𝑗 : 𝑗 ∈ S}), which are used together with the sub-matrices BS and BS to compute the other entries 𝑋𝑖, 𝑗, 𝑖 ∈ S, of𝑋𝑗 using (6.11b):

BS𝑋S

𝑗 = 𝐴𝑗 βˆ’BS𝑋S

𝑗 , 𝑗 ∈ S. (6.12)

Note that to avoid being over-determined, in practice, we solve BK

S

𝑋S

𝑗 = 𝐴K

𝑗 βˆ’BKS𝑋S

𝑗 , 𝑗 ∈ S

whereBK

S is a𝐾 ×𝐾 matrix whose rows are selected fromBS corresponding to K βŠ† V with|K | =𝐾 andBKS selects the rows ofBS in the same way. We combine 𝑋S

𝑗 and𝑋S

𝑗 to obtain a new estimate 𝑋𝑗 for each 𝑗 ∈ S. Together with the columns 𝑋𝑗, 𝑗 ∈ S, that we have accepted, they form the estimated graph matrixX. We illustrate the three-stage scheme in Figure 6.3.1. In the sequel, we analyze the sample complexity of the three-stage scheme based on the (πœ‡, 𝐾 , 𝜌)-sparse distributions defined in Definition 6.3.1.

2) Analysis of the Scheme:

5The choice of𝐾depends on the structure of the graph to be recovered and more specifically,𝐾 is the counting parameter in Definition 6.3.1. In Theorem 6.3.2 and Corollary 6.3.1, we analyze the sample complexity of this three-stage recovery scheme by characterizing an arbitrary graph into the classes specified by Definition 6.3.1 with a fixed𝐾.

LetF≑Rfor the simplicity of representation and analysis. We now present another of our main theorems. Consider the models defined in Section 6.2 and 6.2. The Ξ“-probability of erroris defined to be the maximal probability that theβ„“2-norm of the difference between the estimated vector 𝑋 ∈R𝑛and the original vectorπ‘Œ ∈R𝑛 (satisfying 𝐴=Bπ‘Œ+𝑍 and both 𝐴andBare known to the estimator) is larger than Ξ“ >0:

πœ€P(Ξ“) := sup

π‘ŒβˆˆY(πœ‡)

P(||π‘‹βˆ’π‘Œ||2 > Ξ“)

where Y(πœ‡) is the set of all πœ‡-sparse vectors in R𝑛 and the probability is taken over the randomness in the generator matrixBand the additive noise 𝑍. Given a generator matrixB, the correspondingrestricted isometry constantdenoted byπ›Ώπœ‡is the smallest positive number with

1βˆ’π›Ώπœ‡

||x||22≀ ||BSx||22≀ 1+π›Ώπœ‡

||x||22 (6.13) for all subsetsS βŠ† V of size |S| ≀ πœ‡and allx∈R|S|. Below we state a sufficient condition6 derived form the three-stage scheme for parameter reconstruction.

Theorem 6.3.2(Achievability). Suppose the generator matrix satisfies thatBK

S ∈ R𝐾×𝐾 is invertible for all S βŠ† V and K βŠ† V with |S | = |K | = 𝐾. Let the distribution G𝑛 be (πœ‡, 𝐾 , 𝜌)-sparse. If the three-stage scheme in Algorithm 10 is used for recovering a graph matrixY(𝐺𝑛)of𝐺𝑛that is sampled according toG𝑛, then the probability of error satisfiesπœ€P(πœ‚) ≀ 𝜌+ (π‘›βˆ’πΎ)πœ€P(Ξ“)withπœ‚greater or equal to

2

𝑛Γ+ Ξ“||B||2+𝛾 1βˆ’π›Ώ2𝐾

2(π‘›βˆ’πΎ) +𝐾 πœ‰(B)

where𝛿2𝐾is the corresponding restricted isometry constant ofBwithπœ‡=2𝐾defined in(6.13)and

πœ‰(B):= max

S,K βŠ†V,|S |=|K |=𝐾

BS

2

BK

S

βˆ’1 2.

The proof is in Appendix 6.B. The theory of classical compressed sensing (see [182, 183, 193]) implies that for noiseless parameter reconstruction, if the generator matrix B has restricted isometry constants 𝛿2πœ‡ and 𝛿3πœ‡ satisfying 𝛿2πœ‡ + 𝛿3πœ‡ < 1, then

6Note that𝛾 cannot be chosen arbitrarily andΞ“depends on 𝛾; otherwise the probability of errorπœ€P(Ξ“)will blow up. Theorem 6.4.2 indicates that for Gaussian ensembles settingΞ“ =𝑂(𝛾)= 𝑂(√

π‘›πœŽN)is a valid choice where𝜎Nis the standard deviation of each independent𝑍𝑖 , 𝑗inZ.

all columnsπ‘Œπ‘— with 𝑗 ∈ VSmall are correctly recovered using the minimization in (6.11a)-(6.11c). Denote by spark(B) the smallest number of columns in the matrix B that are linearly dependent (see [194] for the requirements on the spark of the generator matrix to guarantee desired recovery criteria). The following corollary is an improvement of Theorem 6.3.2 for the noiseless case. The proof is postponed to Appendix 6.C.

Corollary6.3.1. LetZ=0and suppose the generator matrixBhas restricted isometry constants𝛿2πœ‡ and𝛿3πœ‡ satisfying𝛿2πœ‡+𝛿3πœ‡ <1and furthermore,spark(B) > 2𝐾. If the distributionG𝑛is(πœ‡, 𝐾 , 𝜌)-sparse, then the probability of error for the three-stage scheme to recover the parameters of a graph matrixY(𝐺𝑛)of𝐺𝑛that is sampled according toG𝑛satisfiesπœ€P ≀ 𝜌.

Dalam dokumen Learning-Augmented Control and Decision-Making (Halaman 191-198)