RANDOM NODE-ASYNCHRONOUS UPDATES ON GRAPHS
2.6 Asynchronous Polynomial Filters on Graphs
Some updates might be adversarial that increase the variation of the signal over the graph. See Figure 1 of [184] for an illustrative example. Nevertheless, some other updates cancel them out in the long run as proven by Theorem 2.2.
Then, random updates on โ(A) as in(2.70) converge to an eigenvector of Awith eigenvalue๐๐ for any amount of asynchronicity0โค ๐ฟT โค 1.
Proof. Sinceโ(ยท)is assumed to satisfy (2.72), โ(A)has a unit eigenvalue, and the non-unit eigenvalues are strictly less than one in magnitude. Hence, Corollary 2.2 ensures that the updates converge to a point in the eigenspace of โ(A) with the eigenvalue 1, which is equivalent to the eigenspace ofAwith the eigenvalue๐๐ due
to the property in (2.72).
Polynomial filters play an important role in the area of graph signal processing.
Starting from the early works [153, 151, 160], polynomial filters are designed to achieve a desired frequency response on the graph. On the contrary, the condition (2.72) disregards the overall response of the filter since its objective is to isolate a single eigenvector. Thus, the design procedure and the properties of the poly- nomials (to be presented in the subsequent sections) differ from the polynomial approximation ideas considered in [153, 151, 160].
Theorem 2.4 tells that an arbitrary eigenvector of the graph can be computed in a decentralized manner if a low order polynomial satisfying (2.72) is constructed.
In this regard the updates on polynomial filters resemble the beam steering in antenna arrays [202]: assume that the order of the polynomial, ๐ฟ, is fixed. Then, the communication pattern between the nodes is completely determined by the operator A and the order ๐ฟ (See Algorithm 1). Once the nodes start to update their values randomly and asynchronously, the behavior of the signal is controlled by the polynomial coefficients. By changing the coefficients in the light of (2.72), one cansteer the signal over the graph to desired โdirections,โ which happen to be the eigenvectors of the operator. Here, ๐ฟ corresponds to the number of elements (sensors) in the array, and the filter coefficients serve the purpose of steering in both cases.
Note that the condition in (2.72) is not necessary in general for the convergence of random updates with a fixed amount of asynchronicity. (See Section 2.4.) However, (2.72) is necessary to guarantee the convergence for all levels of asynchronicity including the most restrictive case of power iteration.
Notice that if the operator A is the graph Laplacian, and the target eigenvalue is ๐๐ =0, then the corresponding eigenvector is the constant vector. In this case the problem reduces to the consensus problem, and the condition in (2.72) becomes a relaxed version of the polynomial considered in [150].
In the following sections we will assume that eigenvalues ofAare real valued, which is the case for undirected graphs, and assume that โ๐ โR. Although the complex case can also be considered in this framework, as we shall see, some of the results do not extend to the complex case.
2.6.1 The Optimal Polynomial
In this section we consider the construction of the polynomial that has the largest gap between the unit eigenvalue and the rest. In order to represent the condition (2.72) in the matrix-vector form we will use a vector of length๐ฟ+1 to denote the polynomial in (2.69), that is, h=[โ
0 ยท ยท ยท โ๐ฟ]T. In addition, let๐ฝbe a Vandermonde matrix constructed with the eigenvalues ofAin the following form:
๐ฝ=
๏ฃฎ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฏ
๏ฃฐ 1 ๐
1 ๐2
1 ยท ยท ยท ๐๐ฟ
1
1 ๐
2 ๐2
2 ยท ยท ยท ๐๐ฟ
2
.. .
.. .
..
. ยท ยท ยท . .. 1 ๐
๐ ๐2
๐ ยท ยท ยท ๐๐ฟ
๐
๏ฃน
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃบ
๏ฃป
โR๐ร(๐ฟ+1). (2.73)
In the case of repeated eigenvalues, we assume that the repeated rows of the matrix ๐ฝare removed. Letฯ(๐)denote the row of๐ฝcorresponding to the target eigenvalue ๐๐, and let ยฏ๐ฝ๐ denote the remaining rows of๐ฝ.
In order to find the optimal๐ฟ๐ก โ order polynomial satisfying (2.72), we consider the following optimization problem:
max ๐, h
๐ s.t.
ฯ(๐) h=1,
๐ฝยฏ ๐ h
โค (1โ๐) 1. (2.74) First of all notice that the constraints in (2.74) are linear due to the fact that๐ฝand hare real valued. The objective function is linear as well. Hence, (2.74) is a linear programming that can be solved efficiently given the eigenvalue matrix๐ฝ.
The constraints of (2.74) enforce the polynomial to satisfy the desired condition in (2.72) while the objective function maximizes the distance between the unit and non-unit eigenvalues ofโ(A). Therefore, the formulation in (2.74) searches for the polynomial that yields the fastest rate of convergence onโ(A)among all polynomials of order ๐ฟ satisfying (2.72). Hence, we will refer to the solution of (2.74) asthe optimal polynomial.
It should be noted that the solution of (2.74) is optimal with respect to the worst case scenario of the synchronous updates. In general, the polynomial generated
via (2.74) may not give the fastest rate of convergence for an arbitrary amount of asynchronicity.
2.6.2 Sufficiency of Second Order Polynomials
In order to make use of the construction in (2.74), the order of the polynomial, ๐ฟ, should be selected appropriately so that the problem is feasible and admits a solution. One way to guarantee the feasibility is to select๐ฟ = ๐โ1, in which case a solution always exists due to the invertibility of the Vandermonde matrix๐ฝ (by disregarding the multiplicity of eigenvalues). In this case, however, updates are no longer localized, which prevents the asynchronous model of (2.5) from being useful.
At the other extreme, the case of๐ฟ =1 is insufficient to ensure the condition (2.72) in general. Therefore, nonlocal updates are required for the sake of flexibility in the eigenvectors. Nevertheless, the locality of the updates needs only to be compromised marginally, as the following theorem shows that๐ฟ =2 is in factsufficientto satisfy (2.72).
Theorem 2.5. Assume that the operatorAhas real eigenvalues๐๐ โR. For a given target eigenvalue ๐๐, the condition in (2.72) is satisfied by the following second order polynomial:
โ(๐) =1โ2๐ (๐โ๐๐)2/๐ 2
๐, (2.75)
for any๐ in0< ๐ < 1and๐ ๐ satisfying the following:
๐ ๐ โฅ max
1โค๐โค๐
|๐๐โ๐๐|. (2.76)
Proof. It is clear thatโ(๐๐) =1. In the following we will show thatโ1< โ(๐๐) < 1 for all๐๐ โ ๐๐. For the upper bound note that(๐๐โ๐๐)2> 0 for all๐๐ โ ๐๐. There- fore,
1โโ(๐๐)=2๐ (๐๐โ๐๐)2/๐ 2
๐ > 0, (2.77)
which proves thatโ(๐๐) < 1 for all๐๐ โ ๐๐. For the lower bound notice that we have ๐ 2
๐ โฅ (๐๐โ๐๐)2for all๐๐by the condition in (2.76). Therefore we have โ(๐๐) =1โ2๐ (๐๐โ๐๐)2/๐ 2
๐ โฅ1โ2๐ > โ1, (2.78)
for all๐๐.
Notice that ๐ in (2.75) is a free parameter which can be tuned to increase the gap between the eigenvalues. Thus, the polynomial given in (2.75) isnotguaranteed to
be optimal in general. It merely shows that a second order polynomial satisfying (2.72) always exists, which also implies the feasibility of (2.74) in the case of๐ฟ =2, or larger.
An important remark is as follows: the sufficiency of second order polynomials does not extend to the complex case in general. To see this consider the following set of ๐ complex numbers: ๐๐=๐๐2๐๐/(๐-1) for 1โค ๐ โค ๐-1 and๐๐ =0. As shown in the supplementary document, no polynomial of order ๐ฟ โค ๐-2 (possibly with complex coefficients) can satisfy|โ(๐๐) | < 1 for 1โค ๐ โค ๐-1 andโ(๐๐) =1. This adversarial example shows not only that second order polynomials are insufficient, but also that a polynomial of order ๐-1 is in fact necessary in the complex case in general. Although no guarantee can be provided, low order polynomials might still exist in the complex case depending on the values of the eigenvalues of a given operatorA.
2.6.3 Spectrum-Blind Construction of Suboptimal Polynomials
Although the solution of (2.74) provides the optimal polynomial, it requires the knowledge of all the eigenvalues of A. Such information is not available and difficult to obtain in general. By compromising the optimality, we will discuss a way of constructing second order polynomials satisfying (2.72) withoutusing the knowledge of all eigenvalues ofA, except the target eigenvalue๐๐.
First of all notice that a value for the coefficient ๐ ๐ used in (2.75) can be found using only the minimum and the maximum eigenvalues of the operator. That is, the following selection
๐ ๐ =max{๐
maxโ๐๐, ๐๐ โ๐
min}, (2.79)
satisfies (2.76). Therefore, the minimum, the maximum and the target eigenvalues suffice to construct the polynomial (2.75).
In fact (2.76) can be satisfied by using an appropriate upper bound for ๐
max and lower bound for๐
min. For example if Ais the adjacency matrix or Laplacian, we can use the largest degree๐
maxof the graph to select๐ ๐ as follows:
The Laplacian
In this case the eigenvalues are bounded as 0 โค๐๐ โค 2๐
max. Hence, ๐ ๐ =๐
max+ |๐๐โ๐
max|. (2.80)
The Adjacency
In this case the eigenvalues are bounded asโ๐
maxโค ๐๐ โค ๐
max. Hence, ๐ ๐ = ๐
max+ |๐๐|. (2.81)
The Normalized Laplacian
In this case the eigenvalues are bounded as 0 โค๐๐ โค 2. Hence,
๐ ๐ =1+ |๐๐โ1|. (2.82)
Since the selections in (2.80), (2.81), and (2.82) do not use the eigenvalues of the (corresponding) operatorA, the polynomial in (2.75) can be constructed using only the target eigenvalue๐๐.
Aforementioned constructions also provide a trade-off between the available infor- mation and the rate of convergence. This point will be elaborated in Section 2.7.
2.6.4 Inexact Spectral Information and The Use of Nonlinearity
In this subsection, we will focus on the construction of a second order polynomial when the target eigenvalue๐๐ is not known exactly. In this regard, we first assume that we are given an interval [๐ ๐] to whichonlythe eigenvalue๐๐ belongs. More precisely, we assume the following:
๐๐
-1 < ๐ โค๐๐ โค ๐ < ๐๐
+1, (2.83)
where we consider only the distinct eigenvalues indexed in ascending order, and assume that eigenvalues are real. Then, we consider the following polynomial:
โ(๐)=1โ2๐
(๐โ๐) (๐โ๐) max{๐
uppโ๐, ๐โ๐
low} + (๐โ๐)/22 (2.84) for some ๐ in 0< ๐ < 1. It should be noted that when the target eigenvalue ๐๐ is known exactly, we can take ๐ =๐=๐๐, in which case the polynomial in (2.84) reduces to the one in (2.75) with๐ ๐ selected as in (2.79). In the case of ๐โ ๐, one can observe that the polynomial in (2.84) maps the eigenvalues of the operatorAas follows:
โ(๐๐) >1 and |โ(๐๐) | < 1 โ๐๐ โ ๐๐, (2.85) which shows that โ(A) does not have a unit eigenvalue, thus the asynchronous updates running on โ(A) do not converge. In fact, the signal would diverge due
to the dominant eigenvalue, โ(๐๐), being strictly larger than 1. (See Figure 2.2).
In order to prevent the signal from diverging, we consider the followingsaturated update model:
x๐
๐ =
๏ฃฑ๏ฃด
๏ฃด
๏ฃฒ
๏ฃด๏ฃด
๏ฃณ ๐
โ(A)x๐-1
๐
, ๐ โ T๐, x๐-1
๐, ๐โT๐,
(2.86) where ๐(ยท)is the โsaturation nonlinearityโ defined as follows:
๐(๐ฅ) =sign(๐ฅ) ยทmin{|๐ฅ|, 1}, (2.87) which is visualized in Figure 2.6.
-5 -4 -3 -2 -1 0 1 2 3 4 5
x -1
0 1
f(x)
Figure 2.6: Visualization of the saturation nonlinearity defined in (2.87).
Notice that the boundedness of the function ๐(ยท)ensures that the signalx๐ in (2.86) does not diverge. In fact, it is numerically observed that randomized asynchronous updates in (2.86) indeed converge to the fixed point of the model. That is, x๐
converges tobx, wherebxsatisfies the following equation:
๐ โ(A)bx
=bx, (2.88)
where ๐(ยท)is assumed to operate element-wise on a vector.
The key observation regarding the solution of (2.88) is the following: when the interval [๐ ๐] in (2.83) is small, then bx is a good approximation of the target eigenvectorv๐. That is,
bxโ ๐พ v๐ (2.89)
for some scale factor๐พ โR. In fact, when the target eigenvalue is known exactly, i.e., ๐=๐ =๐๐, then the approximation in (2.89) becomes equality. Therefore, we conclude that the asynchronous saturated update scheme in (2.86) allows us to find an arbitrary eigenvector approximately when the corresponding eigenvalue is not known exactly. Furthermore, as we have a better approximation of the eigenvalue, we get a better approximation of the corresponding eigenvector. Section 2.7 will make use of this observation to obtain an autonomous clustering of a network.
2.6.5 Implementation of Second Order Polynomials
In Section 2.6.2 graph signals are shown to converge to an arbitrary eigenvector of the underlying graph operatorAthrough random asynchronous updates running on an appropriate second order polynomial filter. In this section, we will show that asynchronous updates on a second order polynomial can be implemented as a node-asynchronous communication protocol in which nodes follow a collect- compute-broadcast scheme independently from each other. For this purpose, we first write a second polynomial ofAexplicitly as follows:
โ(A)= โ
0I+โ
1A+โ
2A2, (2.90)
where the filter coefficients โ
0, โ
1, โ
2 are assumed to be pre-determined such that (2.72) is satisfied for the eigenvalue of the target eigenvector. Then, we define an auxiliary variableyas:
y=A x, (2.91)
where xdenotes the signal on the graph, and yis the โgraph shiftedโ signal. We will assume that the๐๐ก โ node stores๐ฅ๐ and ๐ฆ๐ simultaneously. Thus, (๐ฅ๐, ๐ฆ๐) can be considered as the state of the๐๐ก โ node. Then, we can write the following:
(โ(A)x)๐ = โ
0๐ฅ๐+โ
1 ๐ฆ๐+โ
2 (A y)๐. (2.92) Using (2.92), asynchronous updates with ๐ฟT =1 (only one node is updated per iteration) running onโ(A)can be equivalently written in the following three steps:
๐ข โ (โ
0โ1)๐ฅ๐+โ
1๐ฆ๐+โ
2A[๐,:]y, (2.93)
๐ฅ๐ โ๐ฅ๐+๐ข, (2.94)
yโy+A[:,๐]๐ข, (2.95)
whereA[๐,:] andA[:,๐] denote the๐๐ก โ row and column ofA, respectively.
It is important to note that equations in (2.93)-(2.95) are in the form of a collect- compute-broadcast scheme. In (2.93), the termA[๐,:]yrequires the๐๐ก โnode to collect ๐ฆ๐โs from all๐ โ Nin(๐). In (2.94), the node simply updates its own signal. In (2.95), the termA[:,๐]๐ข requires the๐๐ก โ node to broadcast๐ขto all ๐ โ Nout(๐). These three steps can be converted into a node-asynchronous communication protocol as in Algorithm 1.
Algorithm 1 consists of three procedures: initialization, active stage, and passive stage. In the initialization, the ๐๐ก โ node assigns a random value to its own signal
Algorithm 1Node-Asynchronous Communication Protocol procedureInitialization(๐)
Initialize๐ฅ๐ randomly.
Collect๐ฅ๐ from nodes ๐ โ Nin(๐). ๐ฆ๐ โร
๐โNin(๐) ๐ด๐, ๐ ๐ฅ๐. procedurePassive Stage(๐)
ifa broadcast๐ขis received from the node ๐ then ๐ฆ๐ โ ๐ฆ๐+ ๐ด๐, ๐ ๐ข.
ifthe node ๐ sends a requestthen Send๐ฆ๐ to node ๐.
procedureActive Stage(๐)
Collect๐ฆ๐ from nodes ๐ โ Nin(๐). ๐ขโ (โ
0โ1)๐ฅ๐+โ
1๐ฆ๐+โ
2
ร
๐โNin(๐) ๐ด๐, ๐ ๐ฆ๐. ๐ฅ๐ โ๐ฅ๐+๐ข.
Broadcast๐ขto all nodes ๐ โ Nout(๐).
๐ฅ๐, then it constructs the auxiliary variable ๐ฆ๐ by collecting๐ฅ๐ from its neighbors.
Once the initialization is completed, the๐๐ก โnode waits in the passive stage, in which the graph signal๐ฅ๐ is not updated. However, its neighbors can request ๐ฆ๐, or send a broadcast. When a broadcast is received, the๐๐ก โ node updates only its auxiliary variable ๐ฆ๐. When the๐๐ก โ node wakes up randomly, it gets into the active stage, in which it collects the auxiliary variable ๐ฆ๐ from its neighbors, updates its signal๐ฅ๐, and then broadcasts the amount of update to its neighbors. Then, the node goes back to the passive stage.
Five comments are in order: 1) The random update model in (2.9) implies that all nodes have the same probability of going into the active stage. 2) As the signal x converges to the target eigenvector ofA, the ratio ๐ฆ๐/๐ฅ๐ converges to the corresponding eigenvalue ofA(assuming๐ฅ๐ is non-zero), thus โ(๐ฆ๐/๐ฅ๐) converges to 1 due to (2.72). 3) In the active stage, the broadcast (the step in (2.95)) is essential to ensure thatxandysatisfy (2.91). 4) The amount of update for๐ฅ๐is computed by the๐๐ก โ node itself. The amount of update for๐ฆ๐ is dictated by the neighbors of the ๐๐ก โ node. Thus,๐ฆ๐can also be considered as a buffer. 5) Since edges are allowed to be directed, a node may collect data from the ๐๐ก โ node in the active stage, but may not send data back to the ๐๐ก โ node.
As a final remark we note that Algorithm 1 assumes reliable communication between the nodes, i.e., no link failures. In this case Algorithm 1 is exactly equivalent to the model in (2.5) running onโ(A). As long as the polynomial coefficients are selected
properly (see Theorem 2.4), the signalxin Algorithm 1 is guaranteed to converge to the eigenvector targeted by the polynomial. In the case of link failures, Algorithm 1 deviates from the model in (2.5), thus the convergence guarantees presented here are not applicable. Nevertheless, we have numerically observed that Algorithm 1 converges even in the case of random link failures. This case will be studied in future.