• Tidak ada hasil yang ditemukan

getdocf93c. 288KB Jun 04 2011 12:05:13 AM

N/A
N/A
Protected

Academic year: 2017

Membagikan "getdocf93c. 288KB Jun 04 2011 12:05:13 AM"

Copied!
29
0
0

Teks penuh

(1)

El e c t ro n ic

Jo ur

n a l o

f P

r o

b a b il i t y

Vol. 14 (2009), Paper no. 86, pages 2463–2491. Journal URL

http://www.math.washington.edu/~ejpecp/

Limiting spectral distribution of circulant type matrices

with dependent inputs

Arup Bose

, Rajat Subhra Hazra

and Koushik Saha

Statistics and Mathematics Unit

Indian Statistical Institute

203 B. T. Road, Kolkata 700108

I

NDIA

Abstract

Limiting spectral distribution (LSD) of scaled eigenvalues of circulant, symmetric circulant and a class ofk-circulant matrices are known when the input sequence is independent and identically distributed with finite moments of suitable order. We derive the LSD of these matrices when the input sequence is a stationary, two sided moving average process of infinite order. The limits are suitable mixtures of normal, symmetric square root of the chisquare, and other mixture distribu-tions, with the spectral density of the process involved in the mixtures.

Key words: Large dimensional random matrix, eigenvalues, circulant matrix, symmetric circu-lant matrix, reverse circucircu-lant matrix,kcirculant matrix, empirical spectral distribution, limiting spectral distribution, moving average process, spectral density, normal approximation.

AMS 2000 Subject Classification:Primary 15A52, 60F99, 62E20, 60G57. Submitted to EJP on April 6, 2009, final version accepted November 5, 2009.

email: [email protected] Research supported by J.C.Bose Fellowship, Govt. of India.

email: [email protected]

(2)

1

Introduction

Suppose λ1,λ2, ...,λk are all the eigenvalues of a square matrixAof order k. Then the empirical

spectral distribution (ESD)ofAis defined as

FA(x,y) =k−1 k

X

i=1

I{Ri)x, Ii) y},

where for anyz∈C,R(z), I(z)denote the real and imaginary part ofz respectively. Let{An}n=1

be a sequence of square matrices (with growing dimension) with the corresponding ESD{FA

n}

n=1.

Thelimiting spectral distribution(or measure) (LSD) of the sequence is defined as the weak limit of

the sequence{FAn}∞n=1, if it exists.

Suppose elements of{An}are defined on some probability space(Ω,F,P), that is{An}are random. Then{FA

n(·)} are random and are functions ofω∈Ω but we suppress this dependence. Let F be

a nonrandom distribution function. We say the ESD ofAn converges to F in L2 if at all continuity points(x,y)ofF, Z

FAn(x,y)−F(x,y)

2

dP(ω)0 asn→ ∞. (1.1)

If the eigenvalues are real then it is understood that{FAn} and F are functions of one variable. It

may be noted that (1.1) holds ifE[FAn(t)] F(t)andV[FAn(t)]0 at all continuity points t of

F. We often write Fn forFA

nwhen the matrix under consideration is clear from the context.

For detailed information on existence and identification of LSD in different contexts, see Bai[1999]

and also Bose and Sen[2008]. There are many universality results available in this context. How-ever, most of the existing work on LSD assumes the inputsequence{xi}to be independent. With the current methods used to establish LSD, such as the moment method or the Stietljes transform method, it does not appear to be easy to extend the known results on LSD to general dependent situations for all types of matrices. There are very few works dealing with dependent inputs. For instance, Bose and Sen[2008]establishes LSD for some specific type of dependent entries for a few matrices and Bai and Zhou [2008]establishes LSD of large sample covariance matrices with AR(1) entries.

We focus on the LSD of circulant type matrices— the circulant, the symmetric circulant, the reverse circulant and the so calledk-circulant for suitable values of k. We assume that{xi}is a stationary linear process. Stationary linear process is an important class of dependent sequence. For instance the widely used stationary time series models such as AR, MA, ARMA are all linear processes. Under very modest conditions on the process, we are able to establish LSD for these matrices. These LSD are functions of the spectral density of the process and hence are universal. Consider the following condition.

Assumption A{xi}are independent,E(xi) =0, V(xi) =1 and sup

iE|xi|3<∞.

We first describe the different matrices that we deal with. These may be divided into four classes:

(3)

Cn =p1n

Sen [2006] shows that if Assumption A is satisfied then the ESD of Cn converges in L2 to the two-dimensional normal distribution given by N(0,D) where D is a 2×2 diagonal matrix with diagonal entries 1/2. Meckes[2009]shows similar type of result for independent complex entries. In particular, ifE(xj) =0,E|xj|2=1 and

for everyε >0, then the ESD converges inL2to the standard complex normal distribution.

(ii) Thesymmetric circulant matrix is defined as

SCn =

Bose and Mitra [2002] show that if {xi} satisfies Assumption A, then the ESD of SCn converges weakly inL2 to the standard normal distribution.

Thepalindromic Toeplitz matrixis the palindromic version of the usual symmetric Toeplitz matrix. It

is defined as (see Massey et.al.[2007]),

Its behavior is closely related to the symmetric circulant matrix. It may be noted that then×n princi-pal minor ofP Tn+1 isSCn. Massey et.al.[2007]establish the Gaussian limit forFP Tn. Bose and Sen

[2008]show that if the input sequence{xi}is independent with mean zero and variance 1 and are either(i)uniformly bounded or(ii)identically distributed, then the LSD of P Tn is standard Gaus-sian. They also observe that if the LSD of any one of P Tn andSCn exist, then the other also exists and they are equal.

(4)

RCn = p1n

       

x0 x1 x2 . . . xn−2 xn−1

x1 x2 x3 . . . xn1 x0 x2 x3 x4 . . . x0 x1

.. .

xn1 x0 x1 . . . xn3 xn2        

n×n .

Bose and Mitra [2002] show that if {xi} satisfies Assumption A then the ESD of RCn converges weakly inL2toF, which is the symmetric square root of the chisquare with two degrees of freedom, having density

f(x) =|x|exp(x2), −∞< x<∞. (1.2)

(iv) For positive integerskandn, then×n k-circulantmatrix is defined as

Ak,n =

     

x0 x1 x2 . . . xn−2 xn−1

xnk xnk+1 xnk+2 . . . xnk−2 xnk−1

xn2k xn2k+1 xn2k+2 . . . xn2k2 xn2k1

.. .

     

n×n .

We emphasize that all subscripts appearing above are calculated modulo n. For 1≤ j< n−1, its

(j+1)-th row is obtained by giving its j-th row a right circular shift by k positions (equivalently,

k mod n positions). We have dropped the n−1/2 factor from the definition of the matrix so that subsequent formulae for eigenvalues remain simple. Establishing the LSD for general k-circulant matrices appears to be a difficult problem. Bose, Mitra and Sen [2008]show that if{xi} are i.i.d

N(0, 1), k = no(1) ( 2) and g cd(k,n) = 1 then the LSD of Fn−1/2A

k,n is degenerate at zero, in

probability. They also derive the following LSD. Suppose {xi} are i.i.d. satisfying Assumption C given below. Let{Ei}be i.i.d. E x p(1), U1 be uniformly distributed over (2g)-th roots of unity, U2

be uniformly distributed over the unit circle where{Ui}, {Ei}are mutually independent. Then as

n→ ∞, Fn−1/2A

k,n converges weakly in probability to

(i) U1(

Qg

i=1Ei)1/2g ifkg=−1+sn, g≥1, s=o(n1/3),

(ii) U2(Qig=1Ei)1/2g ifkg=1+sn, g2 and

s=

(

o(n) if gis even

o(n

g+1

g−1) if gis odd.

We investigate the existence of LSD of the above matrices under the following situation.

Assumption B {xn;n0}is a two sided moving average process

xn=

X

i=−∞

aiεni, where an∈R and

X

n∈Z

|an|<∞.

Assumption C {εi},iZare i.i.d. random variables with mean zero, variance one andE|εi|2+δ<

(5)

We show that the LSD of the above matrices continue to exist in this dependent situation under appropriate conditions on the spectral density of the process. The LSD turn out to be appropriate mixtures of, the normal distribution, the “symmetric” square root of the chisquare distribution, and, some other related distributions. Quite expectedly, the spectral density of the process is involved in these mixtures. Our results also reduce to the results quoted above for the i.i.d. situation.

In Section 2 we describe the nature of the eigenvalues of the above matrices, describe the spectral density and set up notation. In Section 3 we state the main results and report some simulation which demonstrate our theoretical results. The main proofs are given in Section 4 and the proofs of some auxiliary Lemma are given in the Appendix.

Some of the results reported in this article have been reported in the (not to be published) technical reports Bose and Saha[2008a], Bose and Saha[2008b], Bose and Saha[2009].

2

Preliminaries

2.1

Spectral density and related facts

Under AssumptionsBandC, γh =C ov xt+h,xt

is finite andPjZ|γj|<∞. Thespectral density

function f of{xn}exists, is continuous, and is given by

f(ω) = 1

2π

X

k∈Z

γkexp(ikω) = 1 2π

γ0+2

X

k≥1

γkcos(kω)

for ω∈[0, 2π].

Let

Ink) = 1

n

n−1

X

t=0

xtei tωk2, k=0, 1, . . . ,n1, (2.1)

denote the periodogram of{xi}whereωk=2πk/nare the Fourier frequencies. Let

C0={t[0, 1]: f(2πt) =0}andC0′={t [0, 1/2]: f(2πt) =0}. (2.2) Define

ψ(eiω) =

X

j=−∞

ajei jω, ψ1(eiω) =R[ψ(eiω)], ψ2(eiω) =I[ψ(eiω)], (2.3)

whereai’s are the moving average coefficients in the definition of xn. It is easy to see that

|ψ(eiω)|2= [ψ1(eiω)]2+ [ψ2(eiω)]2=2πf(ω).

Let

B(ω) =

‚

ψ1(eiω) −ψ2(eiω) ψ2(eiω) ψ1(eiω)

Œ

and for g2,

B(ω1,ω2, ..,ωg) =

          

ψ1(eiω1) −ψ2(eiω1) 0 0 · · · 0 ψ2(eiω1) ψ1(eiω1) 0 0 · · · 0

0 0 ψ1(eiω2) −ψ2(eiω2) · · · 0

0 0 ψ2(eiω2) ψ1(eiω2) · · · 0

0 0 0 0 ... 0

0 0 0 · · · ψ1(eiωg) −ψ2(eiωg)

0 0 0 · · · ψ2(eiωg) ψ1(eiωg)

          

(6)

The above functions will play a crucial role in the statements and proofs of the main results later.

2.2

Description of eigenvalues

We now describe the eigenvalues of the four classes of matrices. Let[x]be the largest integer less than or equal tox.

(i) Circulant matrix. Its eigenvalues{λi}are (see for example Brockwell and Davis[2002]),

λk=

(ii)Symmetric circulant matrix. The eigenvalues{λi}ofSCnare given by:

(a) fornodd:

(iii)Palindromic Toeplitz matrix. As far as we know, there is no formula solution for the eigenval-ues of the palindromic Toeplitz matrix. As pointed out already, since the n×n principal minor of

P Tn+1 isSCn, by interlacing inequality P TnandSCn have identical LSD.

(iv)Reverse circulant matrix. The eigenvalues are given in Bose and Mitra[2002]:

(7)

developed in Section 2 of Bose, Mitra and Sen[2008]. Let

ν=νn:=cos(2π/n) +isin(2π/n)and λk= n−1

X

l=0

xlνkl, 0 j<n. (2.5)

For any positive integersk,n, letp1<p2<. . .<pc be all their common prime factors so that,

n=n

c

Y

q=1

pqβq and k=k

c

Y

q=1

pαqq.

Here αq, βq ≥ 1 and n′, k′, pq are pairwise relatively prime. For any positive integer s, let Zs = {0, 1, 2, . . . ,s−1}. Define the following sets

S(x) ={x kbmodn′:b0}, 0x<n′.

For any setA, let|A|denote its cardinality. Let gx =|S(x)|and

υk,n′=

{x∈Zn′:gx <g1}

. (2.6)

We observe the following about the setsS(x).

1. S(x) ={x kbmodn′: 0b<|S(x)|}.

2. For x 6=u, eitherS(x) =S(u)orS(x)S(u) =φ. As a consequence, the distinct sets from the collection{S(x): 0x <n}forms a partition ofZn.

We shall call{S(x)}theeigenvalue partitionof{0, 1, 2, . . . ,n1}and we will denote the partitioning sets and their sizes by

{P0,P1, . . . ,Pl−1}, and ni=|Pi|, 0≤i<l. (2.7) Define

yj:= Y

t∈Pj

λt y, j=0, 1, . . . ,l−1 where y=n/n′.

Then the characteristic polynomial ofAk,n (whence its eigenvalues follow) is given by

χ€Ak,nŠ=λnn

−1

Y

j=0

€

λnj y

j

Š

. (2.8)

3

Main results

(8)

3.1

Circulant matrix

Define for(x,y)R2andω[0, 2π],

HC(ω,x,y) =

¨

P B(ω)(N1,N2)p2(x,y)if f(ω)6=0, I(x 0,y0) if f(ω) =0,

whereN1 andN2are i.i.d. standard normal distributions.

Lemma 1. (i) For fixed x,y, HC is a bounded continuous function inω.

(ii) FC defined as follows is a proper distribution function.

FC(x,y) =

Z 1

0

HC(2πs,x,y)ds. (3.1)

(iii) IfLeb(C0) =0then FC is continuous everywhere and can be expressed as

FC(x,y) =

ZZ I{(v1,v

2)≤(x,y)}

h Z 1

0

1

2π2f(2πs)e

v

2 1+v22 2πf(2πs)ds

i

d v1d v2. (3.2)

Further, FC is bivariate normal if and only if f is constant almost everywhere (Lebesgue).

(iv) IfLeb(C0)6=0then FC is discontinuous only on D1={(x,y):x y=0}.

The proof of the Lemma is easy and we omit it. The normality claim in (iii) follows by applying Cauchy Schwartz inequality to compare fourth moment and square of the variance and using the fact that for the normal distribution their ratio equals 3. We omit the details.

Theorem 1. Suppose AssumptionsBandChold. Then the ESD of Cn converges in L2to FC(·)given in (3.1)–(3.2).

Remark 1. If {xi} are i.i.d with finite (2+δ) moment, then f(ω)≡ 1/2π, and FC reduces to the

bivariate normal distribution whose covariance matrix is diagonal with entries1/2each. This agrees

with Theorem15, page57of Sen [2006]who proved the result under AssumptionA.

3.2

Symmetric Circulant Matrix

Forx ∈Randω[0,π]define,

HS(ω,x) =

¨

P p2πf(ω)N(0, 1)x) if f(ω)6=0,

I(x0) if f(ω) =0. (3.3)

As before, we now have the following Lemma. We omit the proof.

Lemma 2. (i) For fixed x, HS is a bounded continuous function inωand HS(ω,x) +HS(ω,x) =1.

(ii) FS defined below is a proper distribution function and FS(x) +FS(x) =1.

FS(x) =2

Z 1/2

0

(9)

(iii) IfLeb(C0′) =0then FS is continuous everywhere and may be expressed as

FS(x) =

Z x

−∞

h Z 1/2

0

1

πpf(2πs)

e

t2

4πf(2πs)ds i

d t. (3.5)

Further, FS is normal if and only if f is constant almost everywhere (Lebesgue).

(iv) IfLeb(C0′)6=0then FS is discontinuous only at x=0.

Theorem 2. Suppose AssumptionsBandChold and

lim n→∞

1

n2

[npX/2]

k=1

f(2πk

n )

−3/2

−→0 for all 0<p<1. (3.6)

Then the ESD of SCn converges in L2 to FS given in (3.4)–(3.5). The same limit continues to hold for

P Tn.

Remark 2. (i) (3.6) is satisfied ifinfωf(ω)>0.

(ii) It is easy to check that the variance,µ2 and the fourth momentµ4 of FS equal

R1/2

0 4πf(2πs)ds

andR01/224π2f2(2πs)ds respectively. By Cauchy-Schwartz inequality it follows thatµ4

µ2 2 ≥

3and equal

to3iff f 21π. In the latter case, FS is standard normal. This agrees with Remark2of Bose and Mitra

[2002](under AssumptionA).

3.3

Reverse circulant matrix

DefineHR(ω,x)on[0, 2π]×Ras

HR(ω,x) =

( G

x2

2πf(ω)

if f(ω)6=0

1 if f(ω) =0,

whereG(x) =1ex for x>0, is the standard exponential distribution function. Lemma 3. (i) For fixed x, HR(ω,x)is bounded continuous on[0, 2π].

(ii) FRdefined below is a valid symmetric distribution function.

FR(x) =

  

1 2+

R1/2

0 HR(2πt,x)d t if x>0 1

2−

R1/2

0 HR(2πt,x)d t if x≤0.

(3.7)

(iii) IfLeb(C0′) =0then FRis continuous everywhere and can be expressed as

FR(x) =

  

1R1/2

0 e

2πfx(22πt)d t if x

>0

R1/2

0 e

2πfx(22πt)d t if x ≤0.

(3.8)

Further, FRis the distribution of the symmetric version of the square root of chisquare variable with two

degrees of freedom if and only if f is constant almost everywhere (Lebesgue).

(10)

The proof of the above lemma is omitted.

Theorem 3. Suppose Assumptions BandChold. Then the ESD of RCn converges in L2 to FR given in

(3.7)–(3.8).

Remark 3. If{xi}are i.i.d, with finite(2+δ)moment, then f(ω) =1/2πfor allω∈[0, 2π]and the

LSD FR(·)agrees with (1.2) given earlier.

3.4

k-Circulant matrix

As mentioned before, it appears difficult to prove general results for all possible pairs(k, n). We investigate two subclasses of thek-circulant.

3.4.1 n=kg+1for some fixed g≥2

For anyd1, let

Gd(x) =P

d

Y

i=1

Ei x,

where {Ei} are i.i.d. E x p(1). Note that Gd is continuous. For any integer d ≥ 1, define

Hd(ω1, . . . ,ωd,x)on[0, 2π]d×R≥0as

Hd(ω1, . . . ,ωd,x) =

  

Gd x2d

(2π)dQd i=1f(ωi)

if Qdi=1fi)6=0

1 if Qdi=1fi) =0.

Lemma 4. (i) For fixed x, Hd(ω1, . . . ,ωd,x)is bounded continuous on[0, 2π]d.

(ii) Fd defined below is a valid continuous distribution function.

Fd(x) =

Z 1

0

· · ·

Z 1

0

Hd(2πt1, . . . , 2πtd,x) d

Y

i=1

d ti for x ≥0. (3.9)

The proof of lemma is omitted.

Theorem 4. Suppose AssumptionsBandChold. Suppose n=kg+1for some fixed g 2. Then as

n→ ∞, Fn−1/2A

k,n converges in L2 to the LSD U1(

Qg

i=1Ei)1/2g where{Ei}are i.i.d. with distribution

function Fg given in (3.9) and U1 is uniformly distributed over the(2g)th roots of unity, independent

of the{Ei}.

Remark 4. If {xi}are i.i.d, then f(ω) =1/2π for allω∈[0, 2π]and the LSD is U1(

Qg

i=1Ei)1/2g

where {Ei}are i.i.d. E x p(1), U1 is as in Theorem 4 and independent of{Ei}. This limit agrees with

Theorem3of Bose, Mitra and Sen[2008].

Remark 5. Using the expression (2.8) for the characteristic polynomial, it is then not difficult to

(11)

suppose the sequences k and n satisfy kg=1+sn where g 1is fixed and s= o(n1/3). Fix primes p1,p2, . . . ,pt and positive integersβ1,β2, . . . ,βt. Define

e

n=1

1 p

β2 2 . . .p

βt

t n.

Suppose k = p1p2. . .ptm → ∞. Then the ESD of en−1/2Ak,ne converges weakly in probability to the

LSD which has 1−Πt

s=1p

βs

s

1

mass at zero, and rest of the probability mass is distributed as

U1(Qig=1Ei)1/2g where U1 and{Ei}are as in Theorem 4.

3.4.2 n=kg1for some g2

Forzi,wiR,i=1, 2, ..,g, and with{Ni}i.i.d.N(0, 1), define

Hg(ωi,zi,wi,i=1, . . . ,g) =P B(ω1,ω2, ..,ωg)(N1, ...,N2g)′≤(zi,wi,i=1, 2, ..,g)′

.

Lemma 5. (i)Hg is a bounded continuous in(ω1, . . . ,ωg)for fixed{zi,wi,i=1, . . . ,g}.

(ii)Fgdefined below is a proper distribution function.

Fg(zi,wi,i=1, . . . ,g) =

Z 1

0

· · ·

Z 1

0

Hg(2πti,zi,wi,i=1, . . . ,g)Yd ti. (3.10)

(iii) IfLeb(C0) =0thenFg is continuous everywhere and may be expressed as

Fg(zi,wi,i=1, ..,g)

=

Z

· · ·

Z I

{t(zk,wk,k=1,..,g)}

h Z 1

0

· · ·

Z 1

0

I

{Qf(2πui)6=0}

(2π)gQg

i=1[πf(2πui)] g

Y

i=1

e

1 2

t2

2i−1+t22i

πf(2πui) Ydu

i

i dt.

wheret= (t1,t2, . . . ,t2g−1,t2g)and dt=

Q

d ti. FurtherFg is multivariate (with independent

com-ponents) if and only if f is constant almost everywhere (Lebesgue).

(iv) IfLeb(C0)6=0thenFg is discontinuous only on Dg={(zi,wi,i=1, . . . ,g):

Qg

i=1ziwi =0}.

The proof of lemma is omitted.

Theorem 5. Suppose Assumptions B and C hold. Suppose n = kg1 for some g 2. Then as

n→ ∞, Fn−1/2A

k,n converges in L2 to the LSD (

Qg

i=1Gi)1/g where (R(Gi),I(Gi); i= 1, 2, . . .g)has

the distributionFggiven in (3.10).

Remark 6. If {xi} are i.i.d, with finite (2+δ) moment, then f(ω) ≡ 1/2π and the LSD

simpli-fies to U2(Qig=1Ei)1/2g where {Ei} are i.i.d. E x p(1), U2 is uniformly distributed over the unit circle

independent of{Ei}. This agrees with Theorem4of Bose, Mitra and Sen[2008].

3.5

Simulations

(12)

0 2 4 6 8 10 0

0.2 0.4 0.6 0.8 1 1.2 1.4

−− dependent entries

−6 −4 −2 0 2 4 6

0 0.1 0.2 0.3 0.4 0.5 0.6 0.7

symmetric circulant

Figure 1: (i) (left) dashed line represents the density ofF2when f(ω) = 1

2π(1.25+cosx)and the continuous line represents the same with f 1

2π. (ii) (right) dashed line represents the LSD of symmetric circulant matrix with entries xt=0.3εt+εt+1+0.5εt+2where{εi}i.i.d.N(0, 1)and the continuous line represents the kernel density estimate of the

ESD of the same matrix of order 5000×5000 and same{xt}.

−5 0 5

0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45

reverse circulant with N(0,1) entries

−5 0 5

0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45

reverse circulant with Binomail(1,0.5)

Figure 2: (i) (left) dashed line represents the LSD of the reverse circulant matrix with entriesxt=0.3εtt+1+0.5εt+2

where{εi}i.i.d. N(0, 1). The continuous line represents the kernel density estimate of ESD of the same matrix of order

(13)

4

Proofs of main results

ThroughoutcandC will denote generic constants depending only ond. We use the notationanbn ifanbn →0 andanbn if an

bn →1. As pointed out earlier, to prove that Fnconverges to F (say)

in L2, it is enough to show that

E[Fn(t)]F(t) and V[Fn(t)]0 (4.1)

at all continuity points t ofF. This is what we shall show in every case.

If the eigenvalues have the decompositionλk=ηk+ yk for 1≤kn, where yk→0 in probability then{λk}and{ηk}have similar behavior. We make this precise in the following lemma.

Lemma 6. Suppose {λn,k}1≤kn is a triangular sequence of Rd-valued random variables such that

λn,k=ηn,k+yn,kfor1≤kn. Assume the following holds:

(i) limn→∞1nPnk=1Pn,k˜x) =F(˜x), for ˜x Rd, (ii) limn→∞n12

Pn

k,l=1P(ηn,k≤˜x,ηn,ly˜) =Fx)F(y˜),for˜x, ˜y ∈Rd

(iii) For anyε >0,max1knP(|yn,k|> ε)0as n→ ∞.

Then,

1. limn→∞1n

Pn

k=1P(λn,k≤˜x) =Fx).

2. limn→∞n12

Pn

k,l=1P(λn,k≤˜x,λn,l≤ ˜y) =Fx)Fy).

Proof. We define new random variablesΛn with P(Λn=λn,k) =1/nfork=1, . . . ,n. Then

P(Λnx˜) = 1

n

n

X

k=1

Pn,kx˜).

Similarly define En with P(En = ηn,k) = 1/nfor 1 ≤ kn and Yn with P(Yn = yn,k) = 1/n for 1kn. Now observe thatΛn=En+Yn and for anyε >0,

P(|Yn|> ε) = 1

n

n

X

k=1

P(|yn,k|> ε)0, asn→ ∞

by assumption (iii). ThereforeΛnandEn have the same limiting distribution. Now asn→ ∞,

P(En˜x) = 1

n

n

X

k=1

Pn,kx˜)F(˜x).(by assumption(i))

Therefore asn→ ∞,

1

n

n

X

k=1

(14)

and this is conclusion (i). To prove (ii) we use similar type of argument. Here we define new random variables ˜Λn with P(Λ˜n = (λn,k,λn,l)) =1/n2 for 1≤ k,ln. Similarly define ˜En and ˜Yn. Again ˜

Λn=E˜n+Y˜nand

P(kYnk> ε) = 1

n2

n

X

k,l=1

P(k(yn,k,yn,l)k> ε)→0, asn→ ∞.

So ˜Λnand ˜En will have same limiting distribution and hence conclusion (ii) holds.

We use normal approximation heavily in our proofs. Lemma 7 is a fairly standard consequence of normal approximation and follows easily from Bhattacharya and Ranga Rao[1976] (Corollary 18.1, page 181 and Corollary 18.3, page 184). We omit its proof. Part (i) will be used in Section 4.1– 4.4 and Part (ii) will be used in Section 4.4.

Lemma 7. Let X1, . . . ,Xkbe independent random vectors with values inRd, having zero means and an

average positive-definite covariance matrix Vk =k−1

Pk

j=1C ov(Xj). Let Gk denote the distribution of

k−1/2Tk(X1+. . .+Xk), where Tkis the symmetric, positive-definite matrix satisfying Tk2=Vk−1, n1.

If for someδ >0,EkXjk(2+δ)<, then there exists C>0(depending only on d), such that

(i)

sup B∈C|

Gk(B)Φd(B)| ≤ C kδ/2[λmin(Vk)]−(2+δ)ρ2+δ

(ii) for any Borel set A,

|Gk(A)Φd(A)| ≤ C kδ/2[λmin(Vk)]−(2+δ)ρ2+δ+2 sup

y∈Rd

Φd((∂A)η y)

whereΦd is the standard d dimensional normal distribution function,C is the class of all Borel

mea-surable convex subsets ofRd,ρ2+δ=k−1Pk

j=1EkXjk

(2+δ)andη=Cρ

2+δnδ/2.

4.1

Proof of Theorem 1

The proof for circulant matrix mainly depends on Lemma 7 which helps to use normal approximation and, Lemma 8 given below which allows us to approximate the eigenvalues by appropriate partial sums of independent random variables. The latter follows easily from Fan and Yao[2003](Theorem 2.14(ii), page 63). We have provided a proof in Appendix for completeness. For k = 1, 2,· · ·,n, define

ξ2k−1=

1 pn

n−1

X

t=0

εtcos(ωkt), ξ2k= 1 pn

n−1

X

t=0

εtsin(ωkt).

Lemma 8. Suppose AssumptionBholds and{εt}are i.i.d random variables with mean0, variance1.

For k=1, 2,· · ·,n, write

λn,k= 1 p

n

n−1

X

l=0

xleiωkl =ψ(eiωk)[ξ2k−1+2k] +Ynk),

(15)

Proof of Theorem 1: We first assume Leb(C0) =0. Note that we may ignore the eigenvalue λn and alsoλn/2 whenevernis even since they contribute atmost 2/nto the ESDFn(x,y). So forx,y∈R,

E[Fn(x,y)] s n−1

n−1

X

k=1,k6=n/2

P(bkx,cky).

Define fork=1, 2,· · ·,n,

ηk= (ξ2k−1,ξ2k)′, Y1nk) =R[Ynk)], Y2nk) =I[Ynk)],

where Ynk) are same as defined in Lemma 8. Then (bk,ck)′ = Bkk+ (Y1nk),Y2nk))′. Now in view of Lemma 6 and Lemma 8, to showE[Fn(x,y)]FC(x,y)it is sufficient to show that

1

n

n−1

X

k=1,k6=n/2

P(Bkk≤(x,y)′)→FC(x,y).

To show this, define for 1≤kn−1, (except fork=n/2) and 0≤ln−1,

Xl,k= p

2εlcos(ωkl), p

2εlsin(ωkl)

.

Note that

E(Xl,k) =0 (4.2)

n−1

n−1

X

l=0

C ov(Xl,k) =I (4.3)

sup n

sup

1≤kn

[n−1

n−1

X

l=0

EkXl kk(2+δ)]C <. (4.4)

Fork6=n/2

Bkk≤(x,y)′ =

Bk)(n−1/2 n−1

X

l=0

Xl,k)≤( p

2x,p2y)′ .

Since {(r,s) : Bk)(r,s)′ ≤ ( p

2x,p2y)′} is a convex set in R2 and {Xl,k, l = 0, 1, . . .(n1)}

satisfies (4.2)–(4.4), we can apply Part (i) of Lemma 7 fork6=n/2 to get

P Bk)(n−1/2

n−1

X

l=0

Xl,k)≤( p

2x,p2y)′P Bk)(N1, N2)′≤(

p

2x,p2y)′

C nδ/2[n−1

n−1

X

l=0

EkXl kk(2+δ)]C nδ/20, asn→ ∞.

Therefore

lim n→∞

1

n

n−1

X

k=1,k6=n/2

P Bkk(x,y)= lim

n→∞ 1

n

n−1

X

k=1,k6=n/2

HC(2πk

n ,x,y) =

Z 1

0

(16)

Hence

Along the lines of the proof used to show (4.5) one may now extend the vectors of two coordinates defined above to ones with four coordinates and proceed exactly as above to verify this. We omit the routine details. This completes the proof for the case Leb(C0) =0.

When Leb(C0)6=0, we have to show (4.1) only onD1c(of Lemma 1). All the above steps in the proof will go through for all(x,y)inD1c. Hence if Leb(C0)6=0, we have our required LSD. This completes the proof of Theorem 1.

4.2

Proof of Theorem 2

For convenience, we prove the result for symmetric circulant matrix only for oddn=2m+1. The even case follows by appropriate easy changes in the proof. The partial sum approximation now takes the following form. For the interested reader, we provide a proof in the Appendix.

Lemma 9. Suppose AssumptionBholds and{εt}are i.i.d random variables with mean0, variance1.

For n=2m+1and k=1, 2,· · ·,m, write

Proof of Theorem 2: As before, we provide the detailed proof only when Leb(C0′) =0. Note that we

(17)

Following the argument given in circulant case and using Lemma 6 and Lemma 9 it is sufficient to

whereΦ0,Vk is a bivariate normal distribution with mean zero and covariance matrix Vk. Note that for largem,σ2n≈2 andδ2n≈2. HenceCk′ =(u,v):ψ1(eiωk)u+ψ2(eiωk)v

p

x serves as a good approximation toCkand we get

(18)

Let 0<p<1. Now asn→ ∞, using Assumption (3.6),

Therefore, by first lettingn→ ∞and then letting p1, (4.8) holds. Hence

lim

Rest of the argument in the proof for symmetric circulant is same as in the proof of Theorem 1.

To prove the result forP Tn, we use Cauchy’s interlacing inequality (see Bhatia, 1997, page 59):

Interlacing inequality:SupposeAis ann×nsymmetric real matrix with eigenvaluesλnλn−1≥

. . .λ1. Let Bbe the(n−1)×(n−1)principal submatrix ofAwith eigenvalues µn−1 ≥. . .≥µ1.

4.3

Proof of Theorem 3

For reverse circulant we need the following Lemma. Its proof is given in Fan and Yao[2003] (The-orem 2.14(ii), page 63).

(19)

Proof of Theorem 3: As earlier, we give the proof only for the case Leb(C0′) =0. From the structure of the eigenvalues, the LSD, if it exists, is going to be that of a symmetric distribution. So, it is enough to concentrate on the casex >0. As before we may ignore the two eigenvaluesλ0andλn/2.

Hence forx >0,

E[Fn(x)]1/2+n−1

[n−1

2 ]

X

k=1

P(Ink) x2). (4.9)

Along the same lines as in the proof of Theorem 1, using Lemma 6 and Lemma 10 it is sufficient to show

1

n

[n−1 2 ]

X

k=1

P Lnk)x2HR(2πt,x),

where Lnk)is same as in Lemma 10. Define fork=1, 2,· · ·,[n21]andl =0, 1, 2,· · ·,n−1,

Xl,k= p2εlcos(lωk), p

2εlsin(lωk)

, A

kn=

(r1,r2):πfk)(r12+r 2 2)≤x

2 .

Note that{Xl,k}satisfies (4.2)– (4.4) andLnk)≤ x2 =

n−1/2Pnl=01Xl,kAkn . SinceAknis a convex set inR2, we can apply Part (i) of Lemma 7 to get, asn→ ∞

1

n

[n21]

X

k=1

|P(Lnk) x2)Φ0,I(Akn)| ≤C nδ/20.

But

1

n

[n−1

2 ]

X

k=1

Φ0,I(Akn) = 1

n

[n−1

2 ]

X

k=1

HR(2πk

n ,x)→

Z 1/2

0

HR(2πt,x)d t.

Hence forx ≥0,

E[Fn(x)]→ 1 2+

Z 1/2

0

HR(2πt,x)d t=FR(x).

Now rest of the argument in the proof is same as in the proof of Theorem 1.

4.4

Proofs for Theorems 4 and 5

The proofs of the two theorems for k circulant matrices draw substantially from the method of Bose and Mitra[2002]and uses the eigenvalue description given earlier.

4.4.1 n=kg+1for some fixed g≥2

Proof of Theorem 4: For simplicity we first prove the result wheng=2. Note thatg cd(k,n) =1 and

hence in this casen′=nin (2.8). Thus the index of each eigenvalue belongs toexactly oneof the sets Pl in the eigenvalue partition of{0, 1, 2, ...,n−1}. Recall thatvk,nis the total number of eigenvalues

(20)

havevk,n/n2/n0 and hence these eigenvalues do not contribute to the LSD. Hence it remains to consider only the eigenvalues corresponding to the setsPl which have sizeexactly equaltog1.

Note thatS(1) ={1,k,n−1,nk}and henceg1=4. Recall the quantitiesnj=|Pj|, yj=

Q

t∈Plλt,

whereλj =

Pn−1

l=0 xlνjl, 0≤ j<ngiven in (2.5). Also, for every integer t ≥0, t k2 =−t modn, so that,λt andλnt belong to the same partition blockS(t) =S(nt). Thus each yt is real. Let us define

In={l:|Pl|=4}.

It is clear thatn/|In| →4. Without any loss, let In={1, 2, ...,|In|}.

Let 1,ω,ω2,ω3 be all the fourth roots of unity. Note that for every j, the eigenvalues of A

k,n corresponding to the set Pj are: y1j/4,y1j/4ω,y1j/4ω2,y1j/4ω3. Hence it suffices to consider only

the modulus of eigenvalues y1j/4 as jvaries: if these have an LSDF, say, then the LSD of the whole sequence will be(r,θ)in polar coordinates wherer is distributed according toFandθis distributed uniformly across all the fourth roots of unity and r andθ are independent. With this in mind and remembering the scalingpn, we consider forx >0,

Fn(x) =|In|−1 |In|

X

i=1

I

‚•y

j

n2 ˜1

4

x Œ

.

Since the set of λ values corresponding to any Pj is closed under conjugation, there exists a set Ai⊂ Pi of size 2 such that

Pi={x :x∈ Ai ornx ∈ Ai}.

Combining eachλj with its conjugate, we may write yjin the form,

yj=

Y

t∈Aj

(nb2t +nct2)

where{bt}and{ct}are given in (2.4). Note that for x>0,

E[Fn(x)] =|In|−1

|In|

X

j=1

P yj

n2 ≤x

4.

Now our aim is to show

|In|−1 |In|

X

j=1

P yj

n2 ≤ x

4

F2(x).

We can write yj

n2 =Ln,j+Rn,j for 1≤ j≤ |In|, where

Ln,j = 4π2fjyj

n2, yj=

Y

t∈Aj

(22t1+22t), fj= Y

t∈Aj

ft), 1≤ j≤ |In|,

Rn,j = Lnj1)Rnj2) +Lnj2)Rnj1) +Rnj1)Rnj2),

Lnjk) = 2πfjk)(ξ

2

2jk−1+ξ

2

(21)

Now using Lemma 10 it is easy to see that for anyε >0, max1j≤|I

To do this we will use normal approximation. Define

Xl,j=21/2 (ii) of Lemma 7 and Lemma 4 of Bose, Mitra and Sen[2008], arguing as in the previous proofs,

(22)

= lim

4 coordinates defined above to ones with 8 coordinates and proceed exactly as above to verify this. We omit the routine details. This proves the Theorem when g = 2. The above argument can be extended to cover the general(g >2)case. We highlight only a few of the technicalities and omit the other details. We now need the following lemma.

Lemma 11. Given anyε,η >0there exist an N ∈Nsuch that

and iterating this argument,

(23)

≤ P Rnj

Combining all the above we get

P

First term in the right side can be made smaller than η/2 by choosing M large enough and since max1≤knE|Rnk)| →0 as n→ ∞, we can choose N ∈N such that the second term is less than

η/2 for allnN, proving the lemma.

For general g≥2, as before, n′=n, the index of each eigenvalue belongs to one of the setsPl in the eigenvalue partition of{0, 1, 2, ...,n1}and vk,n/n0. Hence it remains to consider only the eigenvalues corresponding to the setsPl which have size exactly equal to g1 and it follows from the argument in the proof of Lemma 3(i)of Bose, Mitra and Sen[2008]that g1=2g. We can now

Lemma 6, to show (4.12) it is sufficient to show that

1

For this we use normal approximation as we did ing=2 case and define

(24)

Now note that forn=kg+1 we can write{1, 2, . . . ,n1}as{b1kg−1+b2kg−2+· · ·+bg1k+bg; 0

bik−1, for 1 ≤ ik−1; 1 ≤ bgk}. So we can write the sets Aj explicitly using this decomposition of {1, 2, . . . ,n−1}as done in g =2 case, that is, n= k2+1 case. For example if

g=3,Aj ={b1k2+b2k+b3, b2k2+b3kb1, b3k2b1kb2}for j=b1k2+b2k+b3(except for finitely many j, bounded by vk,n and they do not contribute to this limit). Using this fact and proceeding as before we conclude that the LSD is nowFg(·), proving Theorem 4 completely.

4.4.2 n=kg−1for some fixed g≥2

Proof of Theorem 5. First we assume Leb(C0) = 0. Note that gcd(k,n) = 1. Since kg = 1+n =

1 modn, we haveg1|g. If g1<g, theng1g/αwhereα=2 ifg is even andα=3 ifg is odd. In either case, it is easy to check that

kg1kg/α(1+n)1=o(n).

Hence, g = g1. By Lemma 3(ii)of Bose, Mitra and Sen[2008]the total number of eigenvaluesγj ofAk,n such that j∈ Al and|Al|<gis asymptotically negligible.

Unlike the previous theorem, here the partition setsAl are not necessarily self-conjugate. However, the number of indiceslsuch thatAl is self-conjugate is asymptotically negligible compared ton. To show this, we need to bound the cardinality of the following set for 1l< g:

Dl={t∈ {1, 2, . . . ,n}:t kl =t modn}={t∈ {1, 2, . . . ,n}:n|t(kl+1)}.

Note thatt0=n/gcd(n,kl+1)is the minimum element ofDl and every other element is a multiple oft0. Thus

|Dl| ≤

n

t0 ≤gcd(n,k

l+1).

Let us now estimate gcd(n,kl+1). Forl>[g/2],

gcd(n,kl+1)gcd(kg−1,kl+1) =gcd kgl(kl+1)(kgl−1),kl+1≤kgl, which implies gcd(n,kl+1)k[g/2]for all 1l< g. Therefore,

gcd(n,kl+1)

n =

k[g/2]

(kg1) ≤ 2

k[(g+1)/2] ≤

2

((n)1/g)[(g+1)/2] =o(1).

So, we can ignore the partition setsPj which are self-conjugate. For otherPj,

yj= Y

t∈Pj

(pnbt+ipnct)

will be complex.

Now for simplicity we will provide the detailed argument assuming thatg=2. Then,n=k2−1 and we can write{0, 1, 2, . . . ,n}as{ak+b; 0ak1, 0 bk1}and using the construction ofS(x)we havePj={ak+b,bk+a}and|Pj|=2 for j=ak+b; 0ak−1, 0 bk−1 (except for finitely many jand hence such indices do not contribute to the LSD). Let us define

(25)

It is clear that n/|In| → 2. Without any loss, let In = {1, 2, . . . ,|In|}. Suppose Pj = {j1,j2}. We

For this we use normal approximation and define

(26)

= lim

Hence the empirical distribution of yj for those j for which|Pj|=2 converges to the distribution ofQ2i=1Gi such that (R(Gi),I(Gi); i= 1, 2) has distributionF2. Hence the LSD ofn−1/2Ak,n is

This proves the Theorem when g=2.

For generalg>2, note that we can write{0, 1, 2, . . . ,n}as{b1kg−1+b2kg−2+· · ·+bg−1k+bg; 0≤

bi k1, for 1 i k}. So we can write the sets Aj explicitly using this decomposition of {0, 1, 2, . . . ,n} as done in n=k21 case. For example if g = 3,Aj = {b1k2+b2k+b3, b2k2+

b3k+b1, b3k2+b1k+b2}for j=b1k2+b2k+b3(except for finitely many j, bounded byvk,nand they do not contribute to this limit). Using this fact and proceeding as before we will have the LSD as Qgi=1Gi

1/g

such that(R(Gi),I(Gi); i=1, 2, . . .g)has distributionFg.

Acknowledgement. We thank Professor G.J. Babu, Pennysylvania State University for his help. Professor Z.D. Bai, National University of Singapore helped us to access the Chinese article. We also thank two anonymous referees for their insightful comments and valuable suggestions. These have lead to significant improvements in the proofs.

5

Appendix

Proof of Lemma 8.

(27)

= p1

The right side of the above expression is independent of k and as n→ ∞, it can be made smaller than any given positive constant by choosingl large enough. Hence, max1knE|Ynk)| →0. Proof of Lemma 9.

1

for any fixed positive integerl andm>l,

(28)

≤ 2 p

2 p

n

X

j=−∞

|aj|{min(|j|,m)}1/2

≤ 2p2p1

n X

|j|≤l

|aj||j|1/2+X

|j|>l |aj|.

The right side of the above expression is independent of k and as n→ ∞, it can be made smaller than any given positive constant by choosingl large enough. Hence, max1≤kmE(Yn,k)→0.

References

Z. D. Bai. Methodologies in spectral analysis of large-dimensional random matrices, a review. With comments by G. J. Rodgers and Jack W. Silverstein; and a rejoinder by the author. Statist. Sinica, 9(3):611–677, 1999. MR1711663

Z. D. Bai and Wang Zhou. Large sample covariance matrices without independent structure in columns.

Statist. Sinica, 18:425–442, 2008. MR2411613

R. Bhatia. Matrix Analysis. Springer, New York, 1997. MR1477662

R. N. Bhattacharya and R. Ranga Rao.Normal approximation and asymptotic expansions. John Wiley & Sons, New York-London-Sydney, 1976. MR0436272

Arup Bose and Joydip Mitra. Limiting spectral distribution of a special circulant. Statist. Probab. Lett., 60 (1):111–120, 2002. MR1945684

Arup Bose, Joydip Mitra and Arnab Sen. Large dimensional random k-circulants. Technical Report No.R10/2008, Stat-Math Unit, Indian Statistical Institute, Kolkata. Submitted for publication.

Arup Bose and Arnab Sen. Another look at the moment method for large dimensional random matrices.

Electron. J. Probab., 13:no. 21, 588–628, 2008. MR2399292

Arup Bose and Koushik Saha. Limiting spectral distribution of reverse circulant matrix with dependent entries.Technical Report No.R9/2008, Stat-Math Unit, Indian Statistical Institute, Kolkata.

Arup Bose and Koushik Saha. Limiting spectral distribution of circulant matrix with dependent entries.

Technical Report No.R11/2008, Stat-Math Unit, Indian Statistical Institute, Kolkata.

Arup Bose and Koushik Saha. Limiting spectral distribution ofk circulant matrix with dependent entries.

Technical Report No.R2/2009, Stat-Math Unit, Indian Statistical Institute, Kolkata.

Włodzimierz Bryc, Amir Dembo and , Tiefeng Jiang. Spectral measure of large random Hankel, Markov and Toeplitz matrices.The Annals of Probability, 34(1):1–38, 2006.

Peter J. Brockwell and Richard A. Davis. Introduction to time series and forecasting. Springer Texts in Statistics. Springer-Verlag, New York, second edition, 2002. MR1894099

Sourav Chatterjee. A generalization of the Lindeberg principle.The Annals of Probability, 34(6):2061–2076, 2006. MR2294976

(29)

Massey, A., Miller, S.J. and Sinsheimer, J. Distribution of eigenvalues of real symmetric palindromic Toeplitz matrices and circulant matrices. Journal of Theoretical Probability.20, no. 3, 637-662, 2007. MR2337145 Arnab Sen. Large dimensonal random matrices. M. Stat Project Report, Indian Statistical Institute, Kolkata, 2006.

Gambar

Figure 1:(i) (left) dashed line represents the density ofESD of the same matrix of order 5000 F2 when f (ω) =12π(1.25 + cos x) and the continuous linerepresents the same with f ≡12π

Referensi

Dokumen terkait

Minimum rank, Symmetric matrix, Finite field, Projective geometry, Polarity graph, Bilinear symmetric form. AMS

Circulant matrix, A -Factor circulant matrices, Block Vandermonde and Fourier matrices, Rotation and hyperbolic matrices, Generalized Euler formula matrices, Periodic solutions..

Young subgroups, Spherical functions, Finite symmetric spaces, Ramanujan graphs, Symmetric groups, Representations, Characters, Spectral graph theory, Gelfand pair.. AMS

In this paper, the problem of constructing a real symmetric doubly arrow matrix A from two special kinds of spectral information is considered.. The first kind is the minimal

The spectral norm is also of the same order for other related random matrix ensembles, including random Hankel matrices.. A family of random variables is uniformly subgaussian if

In this paper we will study the spectral properties of Wigner real symmetric and Hermitian random matrices in the case when the matrix entries have heavy tails of distribution..

Keywords: Random matrix, concentration of measure, empirical distribution,

For a coordinate symmetric random vector (Y 1 ,.. and is said to be coordinate symmetric. In this note we show that when a random vector is coordinate sym- metric, even though