• Tidak ada hasil yang ditemukan

RANDOM NODE-ASYNCHRONOUS UPDATES ON GRAPHS

2.6 Asynchronous Polynomial Filters on Graphs

Some updates might be adversarial that increase the variation of the signal over the graph. See Figure 1 of [184] for an illustrative example. Nevertheless, some other updates cancel them out in the long run as proven by Theorem 2.2.

Then, random updates on โ„Ž(A) as in(2.70) converge to an eigenvector of Awith eigenvalue๐œ†๐‘— for any amount of asynchronicity0โ‰ค ๐›ฟT โ‰ค 1.

Proof. Sinceโ„Ž(ยท)is assumed to satisfy (2.72), โ„Ž(A)has a unit eigenvalue, and the non-unit eigenvalues are strictly less than one in magnitude. Hence, Corollary 2.2 ensures that the updates converge to a point in the eigenspace of โ„Ž(A) with the eigenvalue 1, which is equivalent to the eigenspace ofAwith the eigenvalue๐œ†๐‘— due

to the property in (2.72).

Polynomial filters play an important role in the area of graph signal processing.

Starting from the early works [153, 151, 160], polynomial filters are designed to achieve a desired frequency response on the graph. On the contrary, the condition (2.72) disregards the overall response of the filter since its objective is to isolate a single eigenvector. Thus, the design procedure and the properties of the poly- nomials (to be presented in the subsequent sections) differ from the polynomial approximation ideas considered in [153, 151, 160].

Theorem 2.4 tells that an arbitrary eigenvector of the graph can be computed in a decentralized manner if a low order polynomial satisfying (2.72) is constructed.

In this regard the updates on polynomial filters resemble the beam steering in antenna arrays [202]: assume that the order of the polynomial, ๐ฟ, is fixed. Then, the communication pattern between the nodes is completely determined by the operator A and the order ๐ฟ (See Algorithm 1). Once the nodes start to update their values randomly and asynchronously, the behavior of the signal is controlled by the polynomial coefficients. By changing the coefficients in the light of (2.72), one cansteer the signal over the graph to desired โ€œdirections,โ€ which happen to be the eigenvectors of the operator. Here, ๐ฟ corresponds to the number of elements (sensors) in the array, and the filter coefficients serve the purpose of steering in both cases.

Note that the condition in (2.72) is not necessary in general for the convergence of random updates with a fixed amount of asynchronicity. (See Section 2.4.) However, (2.72) is necessary to guarantee the convergence for all levels of asynchronicity including the most restrictive case of power iteration.

Notice that if the operator A is the graph Laplacian, and the target eigenvalue is ๐œ†๐‘— =0, then the corresponding eigenvector is the constant vector. In this case the problem reduces to the consensus problem, and the condition in (2.72) becomes a relaxed version of the polynomial considered in [150].

In the following sections we will assume that eigenvalues ofAare real valued, which is the case for undirected graphs, and assume that โ„Ž๐‘› โˆˆR. Although the complex case can also be considered in this framework, as we shall see, some of the results do not extend to the complex case.

2.6.1 The Optimal Polynomial

In this section we consider the construction of the polynomial that has the largest gap between the unit eigenvalue and the rest. In order to represent the condition (2.72) in the matrix-vector form we will use a vector of length๐ฟ+1 to denote the polynomial in (2.69), that is, h=[โ„Ž

0 ยท ยท ยท โ„Ž๐ฟ]T. In addition, let๐šฝbe a Vandermonde matrix constructed with the eigenvalues ofAin the following form:

๐šฝ=

๏ฃฎ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฏ

๏ฃฐ 1 ๐œ†

1 ๐œ†2

1 ยท ยท ยท ๐œ†๐ฟ

1

1 ๐œ†

2 ๐œ†2

2 ยท ยท ยท ๐œ†๐ฟ

2

.. .

.. .

..

. ยท ยท ยท . .. 1 ๐œ†

๐‘ ๐œ†2

๐‘ ยท ยท ยท ๐œ†๐ฟ

๐‘

๏ฃน

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃบ

๏ฃป

โˆˆR๐‘ร—(๐ฟ+1). (2.73)

In the case of repeated eigenvalues, we assume that the repeated rows of the matrix ๐šฝare removed. Letฯ†(๐‘—)denote the row of๐šฝcorresponding to the target eigenvalue ๐œ†๐‘—, and let ยฏ๐šฝ๐‘— denote the remaining rows of๐šฝ.

In order to find the optimal๐ฟ๐‘ก โ„Ž order polynomial satisfying (2.72), we consider the following optimization problem:

max ๐‘, h

๐‘ s.t.

ฯ†(๐‘—) h=1,

๐šฝยฏ ๐‘— h

โ‰ค (1โˆ’๐‘) 1. (2.74) First of all notice that the constraints in (2.74) are linear due to the fact that๐šฝand hare real valued. The objective function is linear as well. Hence, (2.74) is a linear programming that can be solved efficiently given the eigenvalue matrix๐šฝ.

The constraints of (2.74) enforce the polynomial to satisfy the desired condition in (2.72) while the objective function maximizes the distance between the unit and non-unit eigenvalues ofโ„Ž(A). Therefore, the formulation in (2.74) searches for the polynomial that yields the fastest rate of convergence onโ„Ž(A)among all polynomials of order ๐ฟ satisfying (2.72). Hence, we will refer to the solution of (2.74) asthe optimal polynomial.

It should be noted that the solution of (2.74) is optimal with respect to the worst case scenario of the synchronous updates. In general, the polynomial generated

via (2.74) may not give the fastest rate of convergence for an arbitrary amount of asynchronicity.

2.6.2 Sufficiency of Second Order Polynomials

In order to make use of the construction in (2.74), the order of the polynomial, ๐ฟ, should be selected appropriately so that the problem is feasible and admits a solution. One way to guarantee the feasibility is to select๐ฟ = ๐‘โˆ’1, in which case a solution always exists due to the invertibility of the Vandermonde matrix๐šฝ (by disregarding the multiplicity of eigenvalues). In this case, however, updates are no longer localized, which prevents the asynchronous model of (2.5) from being useful.

At the other extreme, the case of๐ฟ =1 is insufficient to ensure the condition (2.72) in general. Therefore, nonlocal updates are required for the sake of flexibility in the eigenvectors. Nevertheless, the locality of the updates needs only to be compromised marginally, as the following theorem shows that๐ฟ =2 is in factsufficientto satisfy (2.72).

Theorem 2.5. Assume that the operatorAhas real eigenvalues๐œ†๐‘– โˆˆR. For a given target eigenvalue ๐œ†๐‘—, the condition in (2.72) is satisfied by the following second order polynomial:

โ„Ž(๐œ†) =1โˆ’2๐œ– (๐œ†โˆ’๐œ†๐‘—)2/๐‘ 2

๐‘—, (2.75)

for any๐œ– in0< ๐œ– < 1and๐‘ ๐‘— satisfying the following:

๐‘ ๐‘— โ‰ฅ max

1โ‰ค๐‘–โ‰ค๐‘

|๐œ†๐‘–โˆ’๐œ†๐‘—|. (2.76)

Proof. It is clear thatโ„Ž(๐œ†๐‘—) =1. In the following we will show thatโˆ’1< โ„Ž(๐œ†๐‘–) < 1 for all๐œ†๐‘– โ‰ ๐œ†๐‘—. For the upper bound note that(๐œ†๐‘–โˆ’๐œ†๐‘—)2> 0 for all๐œ†๐‘– โ‰ ๐œ†๐‘—. There- fore,

1โˆ’โ„Ž(๐œ†๐‘–)=2๐œ– (๐œ†๐‘–โˆ’๐œ†๐‘—)2/๐‘ 2

๐‘— > 0, (2.77)

which proves thatโ„Ž(๐œ†๐‘–) < 1 for all๐œ†๐‘– โ‰ ๐œ†๐‘—. For the lower bound notice that we have ๐‘ 2

๐‘— โ‰ฅ (๐œ†๐‘–โˆ’๐œ†๐‘—)2for all๐œ†๐‘–by the condition in (2.76). Therefore we have โ„Ž(๐œ†๐‘–) =1โˆ’2๐œ– (๐œ†๐‘–โˆ’๐œ†๐‘—)2/๐‘ 2

๐‘— โ‰ฅ1โˆ’2๐œ– > โˆ’1, (2.78)

for all๐œ†๐‘–.

Notice that ๐œ– in (2.75) is a free parameter which can be tuned to increase the gap between the eigenvalues. Thus, the polynomial given in (2.75) isnotguaranteed to

be optimal in general. It merely shows that a second order polynomial satisfying (2.72) always exists, which also implies the feasibility of (2.74) in the case of๐ฟ =2, or larger.

An important remark is as follows: the sufficiency of second order polynomials does not extend to the complex case in general. To see this consider the following set of ๐‘ complex numbers: ๐œ†๐‘›=๐‘’๐‘—2๐œ‹๐‘›/(๐‘-1) for 1โ‰ค ๐‘› โ‰ค ๐‘-1 and๐œ†๐‘ =0. As shown in the supplementary document, no polynomial of order ๐ฟ โ‰ค ๐‘-2 (possibly with complex coefficients) can satisfy|โ„Ž(๐œ†๐‘›) | < 1 for 1โ‰ค ๐‘› โ‰ค ๐‘-1 andโ„Ž(๐œ†๐‘) =1. This adversarial example shows not only that second order polynomials are insufficient, but also that a polynomial of order ๐‘-1 is in fact necessary in the complex case in general. Although no guarantee can be provided, low order polynomials might still exist in the complex case depending on the values of the eigenvalues of a given operatorA.

2.6.3 Spectrum-Blind Construction of Suboptimal Polynomials

Although the solution of (2.74) provides the optimal polynomial, it requires the knowledge of all the eigenvalues of A. Such information is not available and difficult to obtain in general. By compromising the optimality, we will discuss a way of constructing second order polynomials satisfying (2.72) withoutusing the knowledge of all eigenvalues ofA, except the target eigenvalue๐œ†๐‘—.

First of all notice that a value for the coefficient ๐‘ ๐‘— used in (2.75) can be found using only the minimum and the maximum eigenvalues of the operator. That is, the following selection

๐‘ ๐‘— =max{๐œ†

maxโˆ’๐œ†๐‘—, ๐œ†๐‘— โˆ’๐œ†

min}, (2.79)

satisfies (2.76). Therefore, the minimum, the maximum and the target eigenvalues suffice to construct the polynomial (2.75).

In fact (2.76) can be satisfied by using an appropriate upper bound for ๐œ†

max and lower bound for๐œ†

min. For example if Ais the adjacency matrix or Laplacian, we can use the largest degree๐‘‘

maxof the graph to select๐‘ ๐‘— as follows:

The Laplacian

In this case the eigenvalues are bounded as 0 โ‰ค๐œ†๐‘– โ‰ค 2๐‘‘

max. Hence, ๐‘ ๐‘— =๐‘‘

max+ |๐œ†๐‘—โˆ’๐‘‘

max|. (2.80)

The Adjacency

In this case the eigenvalues are bounded asโˆ’๐‘‘

maxโ‰ค ๐œ†๐‘– โ‰ค ๐‘‘

max. Hence, ๐‘ ๐‘— = ๐‘‘

max+ |๐œ†๐‘—|. (2.81)

The Normalized Laplacian

In this case the eigenvalues are bounded as 0 โ‰ค๐œ†๐‘– โ‰ค 2. Hence,

๐‘ ๐‘— =1+ |๐œ†๐‘—โˆ’1|. (2.82)

Since the selections in (2.80), (2.81), and (2.82) do not use the eigenvalues of the (corresponding) operatorA, the polynomial in (2.75) can be constructed using only the target eigenvalue๐œ†๐‘—.

Aforementioned constructions also provide a trade-off between the available infor- mation and the rate of convergence. This point will be elaborated in Section 2.7.

2.6.4 Inexact Spectral Information and The Use of Nonlinearity

In this subsection, we will focus on the construction of a second order polynomial when the target eigenvalue๐œ†๐‘— is not known exactly. In this regard, we first assume that we are given an interval [๐‘Ž ๐‘] to whichonlythe eigenvalue๐œ†๐‘— belongs. More precisely, we assume the following:

๐œ†๐‘—

-1 < ๐‘Ž โ‰ค๐œ†๐‘— โ‰ค ๐‘ < ๐œ†๐‘—

+1, (2.83)

where we consider only the distinct eigenvalues indexed in ascending order, and assume that eigenvalues are real. Then, we consider the following polynomial:

โ„Ž(๐œ†)=1โˆ’2๐œ–

(๐œ†โˆ’๐‘Ž) (๐œ†โˆ’๐‘) max{๐œ†

uppโˆ’๐‘, ๐‘Žโˆ’๐œ†

low} + (๐‘โˆ’๐‘Ž)/22 (2.84) for some ๐œ– in 0< ๐œ– < 1. It should be noted that when the target eigenvalue ๐œ†๐‘— is known exactly, we can take ๐‘Ž =๐‘=๐œ†๐‘—, in which case the polynomial in (2.84) reduces to the one in (2.75) with๐‘ ๐‘— selected as in (2.79). In the case of ๐‘Žโ‰  ๐‘, one can observe that the polynomial in (2.84) maps the eigenvalues of the operatorAas follows:

โ„Ž(๐œ†๐‘—) >1 and |โ„Ž(๐œ†๐‘–) | < 1 โˆ€๐œ†๐‘– โ‰ ๐œ†๐‘—, (2.85) which shows that โ„Ž(A) does not have a unit eigenvalue, thus the asynchronous updates running on โ„Ž(A) do not converge. In fact, the signal would diverge due

to the dominant eigenvalue, โ„Ž(๐œ†๐‘—), being strictly larger than 1. (See Figure 2.2).

In order to prevent the signal from diverging, we consider the followingsaturated update model:

x๐‘˜

๐‘– =

๏ฃฑ๏ฃด

๏ฃด

๏ฃฒ

๏ฃด๏ฃด

๏ฃณ ๐‘“

โ„Ž(A)x๐‘˜-1

๐‘–

, ๐‘– โˆˆ T๐‘˜, x๐‘˜-1

๐‘–, ๐‘–โˆ‰T๐‘˜,

(2.86) where ๐‘“(ยท)is the โ€œsaturation nonlinearityโ€ defined as follows:

๐‘“(๐‘ฅ) =sign(๐‘ฅ) ยทmin{|๐‘ฅ|, 1}, (2.87) which is visualized in Figure 2.6.

-5 -4 -3 -2 -1 0 1 2 3 4 5

x -1

0 1

f(x)

Figure 2.6: Visualization of the saturation nonlinearity defined in (2.87).

Notice that the boundedness of the function ๐‘“(ยท)ensures that the signalx๐‘˜ in (2.86) does not diverge. In fact, it is numerically observed that randomized asynchronous updates in (2.86) indeed converge to the fixed point of the model. That is, x๐‘˜

converges tobx, wherebxsatisfies the following equation:

๐‘“ โ„Ž(A)bx

=bx, (2.88)

where ๐‘“(ยท)is assumed to operate element-wise on a vector.

The key observation regarding the solution of (2.88) is the following: when the interval [๐‘Ž ๐‘] in (2.83) is small, then bx is a good approximation of the target eigenvectorv๐‘—. That is,

bxโ‰ˆ ๐›พ v๐‘— (2.89)

for some scale factor๐›พ โˆˆR. In fact, when the target eigenvalue is known exactly, i.e., ๐‘Ž=๐‘ =๐œ†๐‘—, then the approximation in (2.89) becomes equality. Therefore, we conclude that the asynchronous saturated update scheme in (2.86) allows us to find an arbitrary eigenvector approximately when the corresponding eigenvalue is not known exactly. Furthermore, as we have a better approximation of the eigenvalue, we get a better approximation of the corresponding eigenvector. Section 2.7 will make use of this observation to obtain an autonomous clustering of a network.

2.6.5 Implementation of Second Order Polynomials

In Section 2.6.2 graph signals are shown to converge to an arbitrary eigenvector of the underlying graph operatorAthrough random asynchronous updates running on an appropriate second order polynomial filter. In this section, we will show that asynchronous updates on a second order polynomial can be implemented as a node-asynchronous communication protocol in which nodes follow a collect- compute-broadcast scheme independently from each other. For this purpose, we first write a second polynomial ofAexplicitly as follows:

โ„Ž(A)= โ„Ž

0I+โ„Ž

1A+โ„Ž

2A2, (2.90)

where the filter coefficients โ„Ž

0, โ„Ž

1, โ„Ž

2 are assumed to be pre-determined such that (2.72) is satisfied for the eigenvalue of the target eigenvector. Then, we define an auxiliary variableyas:

y=A x, (2.91)

where xdenotes the signal on the graph, and yis the โ€œgraph shiftedโ€ signal. We will assume that the๐‘–๐‘ก โ„Ž node stores๐‘ฅ๐‘– and ๐‘ฆ๐‘– simultaneously. Thus, (๐‘ฅ๐‘–, ๐‘ฆ๐‘–) can be considered as the state of the๐‘–๐‘ก โ„Ž node. Then, we can write the following:

(โ„Ž(A)x)๐‘– = โ„Ž

0๐‘ฅ๐‘–+โ„Ž

1 ๐‘ฆ๐‘–+โ„Ž

2 (A y)๐‘–. (2.92) Using (2.92), asynchronous updates with ๐›ฟT =1 (only one node is updated per iteration) running onโ„Ž(A)can be equivalently written in the following three steps:

๐‘ข โ† (โ„Ž

0โˆ’1)๐‘ฅ๐‘–+โ„Ž

1๐‘ฆ๐‘–+โ„Ž

2A[๐‘–,:]y, (2.93)

๐‘ฅ๐‘– โ†๐‘ฅ๐‘–+๐‘ข, (2.94)

yโ†y+A[:,๐‘–]๐‘ข, (2.95)

whereA[๐‘–,:] andA[:,๐‘–] denote the๐‘–๐‘ก โ„Ž row and column ofA, respectively.

It is important to note that equations in (2.93)-(2.95) are in the form of a collect- compute-broadcast scheme. In (2.93), the termA[๐‘–,:]yrequires the๐‘–๐‘ก โ„Žnode to collect ๐‘ฆ๐‘—โ€™s from all๐‘— โˆˆ Nin(๐‘–). In (2.94), the node simply updates its own signal. In (2.95), the termA[:,๐‘–]๐‘ข requires the๐‘–๐‘ก โ„Ž node to broadcast๐‘ขto all ๐‘— โˆˆ Nout(๐‘–). These three steps can be converted into a node-asynchronous communication protocol as in Algorithm 1.

Algorithm 1 consists of three procedures: initialization, active stage, and passive stage. In the initialization, the ๐‘–๐‘ก โ„Ž node assigns a random value to its own signal

Algorithm 1Node-Asynchronous Communication Protocol procedureInitialization(๐‘–)

Initialize๐‘ฅ๐‘– randomly.

Collect๐‘ฅ๐‘— from nodes ๐‘— โˆˆ Nin(๐‘–). ๐‘ฆ๐‘– โ†ร

๐‘—โˆˆNin(๐‘–) ๐ด๐‘–, ๐‘— ๐‘ฅ๐‘—. procedurePassive Stage(๐‘–)

ifa broadcast๐‘ขis received from the node ๐‘— then ๐‘ฆ๐‘– โ† ๐‘ฆ๐‘–+ ๐ด๐‘–, ๐‘— ๐‘ข.

ifthe node ๐‘— sends a requestthen Send๐‘ฆ๐‘– to node ๐‘—.

procedureActive Stage(๐‘–)

Collect๐‘ฆ๐‘— from nodes ๐‘— โˆˆ Nin(๐‘–). ๐‘ขโ† (โ„Ž

0โˆ’1)๐‘ฅ๐‘–+โ„Ž

1๐‘ฆ๐‘–+โ„Ž

2

ร

๐‘—โˆˆNin(๐‘–) ๐ด๐‘–, ๐‘— ๐‘ฆ๐‘—. ๐‘ฅ๐‘– โ†๐‘ฅ๐‘–+๐‘ข.

Broadcast๐‘ขto all nodes ๐‘— โˆˆ Nout(๐‘–).

๐‘ฅ๐‘–, then it constructs the auxiliary variable ๐‘ฆ๐‘– by collecting๐‘ฅ๐‘— from its neighbors.

Once the initialization is completed, the๐‘–๐‘ก โ„Žnode waits in the passive stage, in which the graph signal๐‘ฅ๐‘– is not updated. However, its neighbors can request ๐‘ฆ๐‘–, or send a broadcast. When a broadcast is received, the๐‘–๐‘ก โ„Ž node updates only its auxiliary variable ๐‘ฆ๐‘–. When the๐‘–๐‘ก โ„Ž node wakes up randomly, it gets into the active stage, in which it collects the auxiliary variable ๐‘ฆ๐‘— from its neighbors, updates its signal๐‘ฅ๐‘–, and then broadcasts the amount of update to its neighbors. Then, the node goes back to the passive stage.

Five comments are in order: 1) The random update model in (2.9) implies that all nodes have the same probability of going into the active stage. 2) As the signal x converges to the target eigenvector ofA, the ratio ๐‘ฆ๐‘–/๐‘ฅ๐‘– converges to the corresponding eigenvalue ofA(assuming๐‘ฅ๐‘– is non-zero), thus โ„Ž(๐‘ฆ๐‘–/๐‘ฅ๐‘–) converges to 1 due to (2.72). 3) In the active stage, the broadcast (the step in (2.95)) is essential to ensure thatxandysatisfy (2.91). 4) The amount of update for๐‘ฅ๐‘–is computed by the๐‘–๐‘ก โ„Ž node itself. The amount of update for๐‘ฆ๐‘– is dictated by the neighbors of the ๐‘–๐‘ก โ„Ž node. Thus,๐‘ฆ๐‘–can also be considered as a buffer. 5) Since edges are allowed to be directed, a node may collect data from the ๐‘—๐‘ก โ„Ž node in the active stage, but may not send data back to the ๐‘—๐‘ก โ„Ž node.

As a final remark we note that Algorithm 1 assumes reliable communication between the nodes, i.e., no link failures. In this case Algorithm 1 is exactly equivalent to the model in (2.5) running onโ„Ž(A). As long as the polynomial coefficients are selected

properly (see Theorem 2.4), the signalxin Algorithm 1 is guaranteed to converge to the eigenvector targeted by the polynomial. In the case of link failures, Algorithm 1 deviates from the model in (2.5), thus the convergence guarantees presented here are not applicable. Nevertheless, we have numerically observed that Algorithm 1 converges even in the case of random link failures. This case will be studied in future.