El e c t ro n ic
Jo ur
n a l o
f P
r o
b a b il i t y
Vol. 14 (2009), Paper no. 35, pages 978–1011. Journal URL
http://www.math.washington.edu/~ejpecp/
Rates of convergence for minimal distances in the central
limit theorem under projective criteria
Jérôme Dedecker
∗Florence Merlevède
†Emmanuel Rio
‡Abstract
In this paper, we give estimates of ideal or minimal distances between the distribution of the normalized partial sum and the limiting Gaussian distribution for stationary martingale differ-ence sequdiffer-ences or stationary sequdiffer-ences satisfying projective criteria. Applications to functions of linear processes and to functions of expanding maps of the interval are given.
Key words:Minimal and ideal distances, rates of convergence, Martingale difference sequences, stationary sequences, projective criteria, weak dependence, uniform mixing.
AMS 2000 Subject Classification:Primary 60F05.
Submitted to EJP on November 5, 2007, final version accepted June 6, 2008.
∗Université Paris 6, LSTA, 175 rue du Chevaleret, 75013 Paris, FRANCE. E-mail: jerome.dedecker@upmc.fr
†Université Paris Est, Laboratoire de mathématiques, UMR 8050 CNRS, Bâtiment Copernic, 5 Boulevard Descartes,
77435 Champs-Sur-Marne, FRANCE. E-mail: florence.merlevede@univ-mlv.fr
‡Université de Versailles, Laboratoire de mathématiques, UMR 8100 CNRS, Bâtiment Fermat, 45 Avenue des
1
Introduction and Notations
LetX1,X2, . . . be a strictly stationary sequence of real-valued random variables (r.v.) with mean zero and finite variance. Set Sn = X1+X2+· · ·+Xn. By Pn−1/2S
n we denote the law of n −1/2S
n and by Gσ2 the normal distribution N(0,σ2). In this paper, we shall give quantitative estimates of the
approximation ofPn−1/2S
n byGσ2 in terms of minimal or ideal metrics.
LetL(µ,ν) be the set of the probability laws onR2 with marginalsµ andν. Let us consider the
following minimal distances (sometimes called Wasserstein distances of orderr)
Wr(µ,ν) =
infn
Z
|x−y|rP(d x,d y):P∈ L(µ,ν)o if 0<r<1
infn
Z
|x−y|rP(d x,d y)1/r:P∈ L(µ,ν)o ifr≥1 .
It is well known that for two probability measuresµandν onRwith respective distributions
func-tions (d.f.) F andG,
Wr(µ,ν) =
Z 1
0
|F−1(u)−G−1(u)|rdu 1/r
for anyr≥1. (1.1)
We consider also the following ideal distances of order r (Zolotarev distances of order r). For two probability measuresµandν, andr a positive real, let
ζr(µ,ν) =sup
nZ
f dµ− Z
f dν: f ∈Λro,
whereΛr is defined as follows: denoting byl the natural integer such thatl < r ≤l+1,Λr is the class of real functions f which arel-times continuously differentiable and such that
|f(l)(x)−f(l)(y)| ≤ |x−y|r−l for any(x,y)∈R×R. (1.2)
It follows from the Kantorovich-Rubinstein theorem (1958) that for any 0<r≤1,
Wr(µ,ν) =ζr(µ,ν). (1.3)
For probability laws on the real line, Rio (1998) proved that for anyr>1,
Wr(µ,ν)≤cr ζr(µ,ν)
1/r
, (1.4)
wherecris a constant depending only onr.
For independent random variables, Ibragimov (1966) established that if X1 ∈ Lp for p ∈]2, 3],
then W1(Pn−1/2S
n,Gσ2) = O(n
1−p/2) (see his Theorem 4.3). Still in the case of independent r.v.’s,
Zolotarev (1976) obtained the following upper bound for the ideal distance: ifX1∈Lpforp∈]2, 3],
thenζp(Pn−1/2S
n,Gσ2) =O(n
1−p/2). From (1.4), the result of Zolotarev entails that, for p
∈]2, 3],
Wp(Pn−1/2S
n,Gσ2) =O(n
1/p−1/2)(which was obtained by Sakhanenko (1985) for any p>2). From
(1.1) and Hölder’s inequality, we easily get that for independent random variables inLp withp ∈
]2, 3],
Wr(Pn−1/2S
n,Gσ2) =O(n
−(p−2)/2r) for any 1
In this paper, we are interested in extensions of (1.5) to sequences of dependent random variables. More precisely, for X1 ∈ Lp and p in ]2, 3] we shall give Lp-projective criteria under which: for r∈[p−2,p]and(r,p)6= (1, 3),
Wr(Pn−1/2S
n,Gσ2) =O(n
−(p−2)/2 max(1,r)). (1.6)
As we shall see in Remark 2.4, (1.6) applied tor =p−2 provides the rate of convergenceO(n−
p−2 2(p−1))
in the Berry-Esseen theorem.
When (r,p) = (1, 3), Dedecker and Rio (2008) obtained that W1(Pn−1/2S
n,Gσ2) = O(n
−1/2) for
stationary sequences of random variables in L3 satisfying L1 projective criteria or weak
depen-dence assumptions (a similar result was obtained by Pène (2005) in the case where the vari-ables are bounded). In this particular case our approach provides a new criterion under which
W1(Pn−1/2S
n,Gσ2) =O(n
−1/2logn).
Our paper is organized as follows. In Section 2, we give projective conditions for stationary martin-gales differences sequences to satisfy (1.6) in the case(r,p)6= (1, 3). To be more precise, let(Xi)i∈Z
be a stationary sequence of martingale differences with respect to some σ-algebras (Fi)i∈Z (see
Section 1.1 below for the definition of(Fi)i∈Z). As a consequence of our Theorem 2.1, we obtain
that if(Xi)i∈Zis inLpwithp∈]2, 3]and satisfies
∞
X
n=1
1
n2−p/2
E
S2n
n
F0
−σ2
p/2<∞, (1.7)
then the upper bound (1.6) holds provided that(r,p)6= (1, 3). In the case r = 1 and p = 3, we obtain the upper boundW1(Pn−1/2S
n,Gσ2) =O(n
−1/2logn).
In Section 3, starting from the coboundary decomposition going back to Gordin (1969), and using the results of Section 2, we obtain Lp-projective criteria ensuring (1.6) (if (r,p) 6= (1, 3)). For
instance, if(Xi)i∈Zis a stationary sequence ofLp random variables adapted to(Fi)i∈Z, we obtain
(1.6) for any p ∈]2, 3[ and any r ∈[p−2,p]provided that (1.7) holds and the series E(Sn|F0)
converge inLp. In the case wherep=3, this last condition has to be strengthened. Our approach
makes also possible to treat the case of non-adapted sequences.
Section 4 is devoted to applications. In particular, we give sufficient conditions for some functions of Harris recurrent Markov chains and for functions of linear processes to satisfy the bound (1.6) in the case (r,p) 6= (1, 3) and the rate O(n−1/2logn) when r = 1 and p = 3. Since projective criteria are verified under weak dependence assumptions, we give an application to functions of
φ-dependent sequences in the sense of Dedecker and Prieur (2007). These conditions apply to unbounded functions of uniformly expanding maps.
1.1
Preliminary notations
Throughout the paper, Y is aN(0, 1)-distributed random variable. We shall also use the following notations. Let(Ω,A,P)be a probability space, andT :Ω7→Ωbe a bijective bimeasurable
transfor-mation preserving the probabilityP. For aσ-algebraF0 satisfyingF0 ⊆ T−1(F0), we define the nondecreasing filtration(Fi)i∈Z byFi = T−i(F0). LetF−∞ =
T
k∈ZFk andF∞=
W
k∈ZFk. We shall denote sometimes byEi the conditional expectation with respect toFi. LetX0be a zero mean
2
Stationary sequences of martingale differences.
In this section we give bounds for the ideal distance of order r in the central limit theorem for stationary martingale differences sequences(Xi)i∈Zunder projective conditions.
Notation 2.1. For anyp>2, define the envelope normk.k1,Φ,p by
kXk1,Φ,p=
Z 1
0
(1∨Φ−1(1−u/2))p−2QX(u)du
whereΦdenotes the d.f. of theN(0, 1)law, andQX denotes the quantile function of|X|, that is the cadlag inverse of the tail function x →P(|X|>x).
Theorem 2.1. Let(Xi)i∈Zbe a stationary martingale differences sequence with respect to(Fi)i∈Z. Let
σdenote the standard deviation of X0. Let p∈]2, 3]. Assume thatE|X0|p<∞and that
∞
X
n=1
1
n2−p/2
E
Sn2
n
F0
−σ2
1,Φ,p<∞, (2.1)
and
∞
X
n=1
1
n2/p
E
S2n
n
F0
−σ2
p/2<∞. (2.2)
Then, for any r ∈ [p−2,p] with (r,p) 6= (1, 3), ζr(Pn−1/2S
n,Gσ2) = O(n
1−p/2), and for p = 3,
ζ1(Pn−1/2S
n,Gσ2) =O(n
−1/2logn).
Remark 2.1. Leta>1 and p>2. Applying Hölder’s inequality, we see that there exists a positive constant C(p,a) such thatkXk1,Φ,p ≤ C(p,a)kXka. Consequently, if p ∈]2, 3], the two conditions (2.1) and (2.2) are implied by the condition (1.7) given in the introduction.
Remark 2.2. Under the assumptions of Theorem 2.1, ζr(Pn−1/2S
n,Gσ2) = O(n
−r/2) if r < p−2.
Indeed, let p′ =r+2. Since p′< p, if the conditions (2.1) and (2.2) are satisfied for p, they also hold forp′. Hence Theorem 2.1 applies withp′.
From (1.3) and (1.4), the following result holds for the Wasserstein distances of orderr.
Corollary 2.1. Under the conditions of Theorem 2.1, Wr(Pn−1/2S
n,Gσ2) =O(n
−(p−2)/2 max(1,r))for any
r in[p−2,p], provided that(r,p)6= (1, 3).
Remark 2.3. Forpin]2, 3],Wp(Pn−1/2S
n,Gσ2) =O(n
−(p−2)/2p). This bound was obtained by
Sakha-nenko (1985) in the independent case. For p < 3, we have W1(Pn−1/2S
n,Gσ2) = O(n
1−p/2). This
bound was obtained by Ibragimov (1966) in the independent case.
Remark 2.4. Recall that for two real valued random variables X,Y, the Ky Fan metricα(X,Y) is defined byα(X,Y) =inf{ǫ >0 :P(|X−Y|> ǫ)≤ǫ}. LetΠ(µ,ν)be the Prokhorov distance between
µandν. By Theorem 11.3.5 in Dudley (1989) and Markov inequality, one has, for anyr>0,
Taking the minimum over the random couples (X,Y) with law L(µ,ν), we obtain that, for any 0< r ≤1, Π(µ,ν)≤(Wr(µ,ν))1/(r+1). Hence, if Πn is the Prokhorov distance between the law of
n−1/2Sn and the normal distributionN(0,σ2),
Πn≤(Wr(Pn−1/2S
n,Gσ2))
1/(r+1)for any 0<r ≤1 .
Taking r=p−2, it follows that under the assumptions of Theorem 2.1,
Πn=O(n−
p−2
2(p−1)) if p<3 andΠ
n=O(n−1/4
p
logn) if p=3. (2.3)
For p in]2, 4], under (2.2), we have thatkPni=1E(X2i −σ2|Fi−1)kp/2 =O(n2/p)(apply Theorem
2 in Wu and Zhao (2006)). Applying then the result in Heyde and Brown (1970), we get that if
(Xi)i∈Zis a stationary martingale difference sequence inLpsuch that (2.2) is satisfied then
kFn−Φσk∞=O n−
p−2 2(p+1).
whereFnis the distribution function ofn−1/2SnandΦσis the d.f. of Gσ2. Now
kFn−Φσk∞≤ 1+σ−1(2π)−1/2
Πn.
Consequently the bounds obtained in (2.3) improve the one given in Heyde and Brown (1970), provided that (2.1) holds.
Remark 2.5. If (Xi)i∈Zis a stationary martingale difference sequence inL3 such thatE(X02) =σ2
and
X
k>0
k−1/2kE(Xk2|F0)−σ2k3/2<∞, (2.4)
then, according to Remark 2.1, the conditions (2.1) and (2.2) hold forp=3. Consequently, if (2.4) holds, then Remark 2.4 giveskFn−Φσk∞=O n−1/4plogn
. This result has to be compared with Theorem 6 in Jan (2001), which states thatkFn−Φσk∞=O(n−1/4) if
P
k>0kE(X 2
k|F0)−σ
2
k3/2<
∞.
Remark 2.6. Notice that if(Xi)i∈Zis a stationary martingale differences sequence, then the
condi-tions (2.1) and (2.2) are respectively equivalent to
X
j≥0
2j(p/2−1)k2−jE(S2
2j|F0)−σ2k1,Φ,p<∞, and
X
j≥0
2j(1−2/p)k2−jE(S2
2j|F0)−σ2kp/2<∞.
To see this, letAn=kE(S2
n|F0
−E(S2
n)k1,Φ,pandBn=kE(Sn2|F0
−E(S2
n)kp/2. We first show that AnandBnare subadditive sequences. Indeed, by the martingale property and the stationarity of the sequence, for all positiveiand j
Ai+j = kE(S2
i + (Si+j−Si)2|F0
−E(S2
i + (Si+j−Si)2)k1,Φ,p
≤ Ai+kE (Si+j−Si)2−E(S2j)|F0k1,Φ,p.
Proceeding as in the proof of (4.6), p. 65 in Rio (2000), one can prove that, for anyσ-fieldA and any integrable random variableX,kE(X|A)k1,Φ,p≤ kXk1,Φ,p. Hence
kE (Si+j−Si)2−E(S2
j)|F0
k1,Φ,p≤ kE((Si+j−Si)2−E(S2j)|Fi
k1,Φ,p.
3
Rates of convergence for stationary sequences
In this section, we give estimates for the ideal distances of order r for stationary sequences which are not necessarily adapted toFi.
Theorem 3.1. Let(Xi)i∈Zbe a stationary sequence of centered random variables inLp with p∈]2, 3[,
and letσ2n=n−1E(S2
n). Assume that
X
n>0
E(Xn|F0)and
X
n>0
(X−n−E(X−n|F0))converge inLp, (3.1)
and
X
n≥1
n−2+p/2kn−1E(S2
n|F0)−σ2nkp/2<∞. (3.2)
Then the seriesPk∈ZCov(X0,Xk)converges to some nonnegativeσ2, and
1. ζr(Pn−1/2S
n,Gσ2) =O(n
1−p/2)for r
∈[p−2, 2],
2. ζr(Pn−1/2S
n,Gσ2n) =O(n
1−p/2)for r
∈]2,p].
Remark 3.1. According to the bound (5.40), we infer that, under the assumptions of Theorem 3.1, the condition (3.2) is equivalent to
X
n≥1
n−2+p/2kn−1E(Sn2|F0)−σ2kp/2<∞. (3.3)
The same remark applies to the next theorem withp=3.
Remark 3.2. The result of item 1 is valid withσn instead ofσ. On the contrary, the result of item 2 is no longer true ifσn is replaced byσ, because for r ∈]2, 3], a necessary condition forζr(µ,ν)
to be finite is that the two first moments ofν and µ are equal. Note that under the assumptions of Theorem 3.1, both Wr(Pn−1/2S
n,Gσ2) and Wr(Pn−1/2Sn,Gσ2n) are of the order of n
−(p−2)/2 max(1,r).
Indeed, in the case wherer∈]2,p], one has that
Wr(Pn−1/2S
n,Gσ2)≤Wr(Pn−1/2Sn,Gσ2n) +Wr(Gσ2n,Gσ2), and the second term is of order|σ−σn|=O(n−1/2).
In the case wherep=3, the condition (3.1) has to be strengthened.
Theorem 3.2. Let(Xi)i∈Z be a stationary sequence of centered random variables inL3, and letσ2n =
n−1E(S2n). Assume that (3.1) holds for p=3and that
X
n≥1
1
n
X
k≥n
E(Xk|F0)
3<∞ and
X
n≥1
1
n
X
k≥n
(X−k−E(X−k|F0))
3<∞. (3.4)
Assume in addition that
X
n≥1
n−1/2kn−1E(S2
n|F0)−σ2nk3/2<∞. (3.5)
Then the seriesPk∈ZCov(X0,Xk)converges to some nonnegativeσ
1. ζ1(Pn−1/2S
n,Gσ2) =O(n
−1/2logn),
2. ζr(Pn−1/2S
n,Gσ2) =O(n
−1/2)for r
∈]1, 2],
3. ζr(Pn−1/2S
n,Gσ2n) =O(n
−1/2)for r
∈]2, 3].
4
Applications
4.1
Martingale differences sequences and functions of Markov chains
Recall that the strong mixing coefficient of Rosenblatt (1956) between twoσ-algebras A and B is defined by α(A,B) =sup{|P(A∩B)−P(A)P(B)| : (A,B)∈ A × B }. For a strictly stationary
sequence(Xi)i∈Z, letFi=σ(Xk,k≤i). Define the mixing coefficientsα1(n)of the sequence(Xi)i∈Z
by
α1(n) =α(F0,σ(Xn)).
For the sake of brevity, letQ=QX0 (see Notation 2.1 for the definition). According to the results of
Section 2, the following proposition holds.
Proposition 4.1. Let (Xi)i∈Z be a stationary martingale difference sequence in Lp with p ∈]2, 3].
Assume moreover that the series
X
k≥1
1
k2−p/2
Z α1(k)
0
(1∨log(1/u))(p−2)/2Q2(u)du and X
k≥1
1
k2/p
Z α1(k)
0
Qp(u)du2/p (4.1)
are convergent.Then the conclusions of Theorem 2.1 hold.
Remark 4.1. From Theorem 2.1(b) in Dedecker and Rio (2008), a sufficient condition to get
W1(Pn−1/2S
n,Gσ2) =O(n
−1/2logn)is
X
k≥0
Z α1(n)
0
Q3(u)du<∞.
This condition is always strictly stronger than the condition (4.1) whenp=3.
We now give an example. Consider the homogeneous Markov chain(Yi)i∈Zwith state spaceZ
de-scribed at page 320 in Davydov (1973). The transition probabilities are given bypn,n+1=p−n,−n−1= an forn≥0, pn,0=p−n,0=1−an forn>0, p0,0=0,a0=1/2 and 1/2≤an<1 for n≥1. This chain is irreducible and aperiodic. It is Harris positively recurrent as soon asPn≥2Πnk−=11ak<∞. In that case the stationary chain is strongly mixing in the sense of Rosenblatt (1956).
Denote by K the Markov kernel of the chain (Yi)i∈Z. The functions f such that K(f) =0 almost
everywhere are obtained by linear combinations of the two functions f1 and f2 given by f1(1) =1,
f1(−1) =−1 and f1(n) = f1(−n) =0 ifn6=1, and f2(0) =1, f2(1) = f2(−1) =0 and f2(n+1) = f2(−n−1) =1−an−1ifn>0. Hence the functions f such thatK(f) =0 are bounded.
If(Xi)i∈Zis defined byXi= f(Yi), with K(f) =0, then Proposition 4.1 applies if
which holds as soon as P0(τ = n) = O(n−1−p/2(logn)−p/2−ε), where P0 is the probability of the chain starting from 0, and τ= inf{n> 0,Xn = 0}. Now P0(τ = n) = (1−an)Πin=−11ai for n≥ 2. Consequently, if
ai=1− p 2i
1+ 1+ε
logi
forilarge enough ,
the condition (4.2) is satisfied and the conclusion of Theorem 2.1 holds.
Remark 4.2. If f is bounded and K(f) 6= 0, the central limit theorem may fail to hold forSn =
Pn
i=1(f(Yi)−E(f(Yi))). We refer to the Example 2, page 321, given by Davydov (1973), whereSn properly normalized converges to a stable law with exponent strictly less than 2.
Proof of Proposition 4.1. Let Bp(F0) be the set of F0-measurable random variables such that
kZkp≤1. We first notice that
kE(X2k|F0)−σ2kp/2= sup
Z∈Bp/(p−2)(F 0)
Cov(Z,Xk2).
Applying Rio’s covariance inequality (1993), we get that
kE(X2
k|F0)−σ2kp/2≤2
Z α1(k)
0
Qp(u)du2/p,
which shows that the convergence of the second series in (4.1) implies (2.2). Now, from Fréchet (1957), we have that
kE(X2
k|F0)−σ2k1,Φ,p=sup
E
((1∨ |Z|p−2)|E(X2
k|F0)−σ2|),Z F0-measurable, Z∼ N(0, 1) .
Hence, settingǫk=sign(E(Xk2|F0)−σ2),
kE(X2
k|F0)−σ2k1,Φ,p=sup
Cov(ǫk(1∨ |Z|p−2),Xk2),ZF0-measurable,Z∼ N(0, 1) .
Applying again Rio’s covariance inequality (1993), we get that
kE(X2
k|F0)−σ2k1,Φ,p≤C
Z α1(k)
0
(1∨log(u−1))(p−2)/2Q2(u)du,
which shows that the convergence of the first series in (4.1) implies (2.1).
4.2
Linear processes and functions of linear processes
In what follows we say that the series Pi∈Zai converges if the two series
P
i≥0ai and
P
i<0ai converge.
Theorem 4.1. Let(ai)i∈Z be a sequence of real numbers in ℓ2 such that P
i∈Zai converges to some
real A. Let (ǫi)i∈Z be a stationary sequence of martingale differences in Lp for p ∈]2, 3]. Let Xk =
P
j∈Zajǫk−j, andσ
2
n=n−
1E(S2
n). Let b0=a0−A and bj=aj for j6=0. Let An=
P
j∈Z( Pn
k=1bk−j)2.
If An=o(n), thenσ2
n converges toσ2=A2E(ǫ02). If moreover
∞
X
n=1
1
n2−p/2
E
1
n
Xn
j=1
ǫj
2 F0
−E(ǫ20)
p/2<∞, (4.3)
1. If An=O(1), thenζ1(Pn−1/2S
Remark 4.3. If the condition given by Heyde (1975) holds, that is
∞
Proof of Theorem 4.1.We start with the following decomposition:
Sn=A Burkholder’s inequality, there exists a constantC such that
kRnkp≤C
According to Remark 2.1, since (4.3) holds, the two conditions (2.1) and (2.2) of Theorem 2.1 are satisfied by the martingale Mn = A
Pn
k=1ǫk. To conclude the proof, we use Lemma 5.2 given in Section 5.2, with the upper bound (4.7).
Proof of Remarks 4.3 and 4.4. To prove Remark 4.3, note first that
LetTi=
In the next result, we shall focus on functions of real-valued linear processes
Xk=h X
Proof of Theorem 4.2.Theorem 4.2 is a consequence of the following proposition:
Proposition 4.2. Let(ai)i∈Z,(ǫi)i∈Zand(Xi)i∈Zbe as in Theorem 4.2. Let(ǫi′)i∈Zbe an independent
To prove Theorem 4.2, it remains to check (4.10). We only check the first condition. Since
From Burkholder’s inequality, for anyβ >0,
(4.9). The second part can be handled in the same way.
Proof of Proposition 4.2. Let Fi = σ(ǫk,k ≤ i). We shall first prove that the condition (3.2) of Theorem 3.1 holds. We write
kE(S2n|F0) − E(S2n)kp/2≤2 conditional expectation with respect toǫ. Define
Yi=
Applying first the triangle inequality, and next Hölder’s inequality, we get that
kE(XiXk+i|F0)−E(XiXk+i)kp/2 ≤ kh0(Yk′+i+Zk+i)kpkh0(Yi′+Zi)−h0(Yi′+Zi′)kp
Letm1,i =|Yi′+Zi| ∨ |Yi′+Zi′|. Sincewh
By subadditivity, we obtain that
distributed, it follows that
In the same way
provided that the first condition in (4.10) holds.
We turn now to the control ofPni=1Pnk=ikE(XiXk+i|F0)kp/2. We first write that
Using the same arguments as before, we get that
kE(Xk+i|Fi+[k/2])kp=kE(Xb(k)|F0)kp≤2
In the same way,
Let
m2,i,k=
X
j∈Z
aiǫi−j
∨
X
j<−[k/2]
ajǫi′−j+
X
j≥−[k/2]
ajǫi−j
.
Using again the subbadditivity of t → wh(t,M), and the fact that (Pj<−[k/2]ajǫi−j,m2,i,k),
(Pj<−[k/2]ajǫi′−j,m2,i,k)and(
P
j<−[k/2]ajǫ−j,M2,−[k/2])are identically distributed, we obtain that
Xi−E(Xi|Fi+[k/2])
p=
X−[k/2]−E(X−[k/2]|F0)
p≤2
wh
X
j<−[k/2]
ajǫ−j
,M2,−[k/2]
p.
(4.12) Consequently
X
n≥1
1
n3−p/2
n
X
i=1
n
X
k=i
kE(XiXk+i|F0)kp/2<∞
provided that (4.10) holds. This completes the proof of (3.2).
Using the bounds (4.11) and (4.12) (taking b(k) = nin (4.11) and[k/2] = nin (4.12)), we see that the condition (3.1) of Theorem 3.1 (and also the condition (3.4) of Theorem 3.2 in the case
p=3) holds under (4.10).
4.3
Functions of
φ-dependent sequences
In order to include examples of dynamical systems satisfying some correlations inequalities, we introduce a weak version of the uniform mixing coefficients (see Dedecker and Prieur (2007)).
Definition 4.1. For any random variable Y = (Y1,· · ·,Yk) with values in Rk define the function
gx,j(t) =1It≤x−P(Yj≤ x). For anyσ-algebraF, let
φ(F,Y) = sup
(x1,...,xk)∈Rk
E
Yk
j=1
gxj,j(Yj)
F−E
Yk
j=1
gxj,j(Yj)
∞.
For a sequenceY= (Yi)i∈Z, whereYi =Y0◦Ti andY0 is aF0-measurable and real-valued r.v., let
φk,Y(n) = max
1≤l≤k i sup l>...>i1≥n
φ(F0,(Yi1, . . . ,Yil)).
Definition 4.2. For anyp≥1, letC(p,M,PX)be the closed convex envelop of the set of functions f
which are monotonic on some open interval ofRand null elsewhere, and such thatE(|f(X)|p)<M.
Proposition 4.3. Let p∈]2, 3]and s≥p. Let Xi= f(Yi)−E(f(Yi)), where Yi=Y0◦Ti and f belongs
toC(s,M,PY0). Assume that
X
i≥1
i(p−4)/2+(s−2)/(s−1)φ2,Y(i)(s−2)/s<∞. (4.13)
Then the conclusions of Theorem 4.2 hold.
Remark 4.5. Notice that ifs= p= 3, the condition (4.13) becomesPi≥1φ2,Y(i)1/3 < ∞, and if
Proof of Proposition 4.3. Let Bp(F0) be the set of F0-measurable random variables such that
kZkp≤1. We first notice that
kE(Xk|F0)kp≤ kE(Xk|F0)ks= sup
Z∈Bs/(s−1)(F 0)
Cov(Z,f(Yk)).
Applying Corollary 6.2 with k= 2 to the covariance on right hand (take f1 = Id and f2 = f), we
obtain that
kE(Xk|F0)ks ≤ sup
Z∈Bs/(s−1)(F 0)
8(φ(σ(Z),Yk))(s−1)/skZks/(s−1)(φ(σ(Yk),Z))1/sM1/s
≤ 8(φ1,Y(k))(s−1)/sM1/s, (4.14)
the last inequality being true becauseφ(σ(Z),Yk)≤φ1,Y(k) andφ(σ(Yk),Z)≤ 1. It follows that the conditions (3.1) (for p∈]2, 3]) and (3.4) (forp=3) are satisfied under (4.13). The condition (3.2) follows from the following lemma by takingb= (4−p)/2.
Lemma 4.1. Let Xi be as in Proposition 4.3, and let b∈]0, 1[.
If X
i≥1
i−b+(s−2)/(s−1)φ2,Y(i)(s−2)/s<∞, then
X
n>1
1
n1+bk
E(S2
n|F0)−E(Sn2)kp/2<∞.
Proof of Lemma 4.1. Since,
kE(S2
n|F0)−E(S2n)kp/2≤2
n
X
i=1
n−i
X
k=0
kE(XiXk+i|F0)−E(XiXk+i)kp/2,
we infer that there existsC>0 such that
X
n>1
1
n1+bk
E(S2
n|F0)−E(Sn2)kp/2≤C
X
i>0
X
k≥0
1
(i+k)bkE(XiXk+i|F0)−E(XiXk+i)kp/2. (4.15)
We shall bound upkE(XiXk+i|F0)−E(XiXk+i)kp/2in two ways. First, using the stationarity and the
upper bound (4.14), we have that
kE(XiXk+i|F0)−E(XiXk+i)kp/2≤2kX0E(Xk|F0)kp/2≤16kX0kpM1/s(φ1,Y(k))(s−1)/s. (4.16)
Next, note that
kE(XiXk+i|F0)−E(XiXk+i)kp/2 = sup
Z∈Bs/(s−2)(F 0)
Cov(Z,XiXk+i)
= sup
Z∈Bs/(s−2)(F 0)
E((Z−E(Z))XiXk+i).
Applying Corollary 6.2 withk=3 to the termE((Z−E(Z))XiXk+i)(take f1=Id, f2= f3= f), we
obtain thatkE(XiXk+i|F0)−E(XiXk+i)kp/2is smaller than
sup Z∈Bs/(s−2)(F
0)
Sinceφ(σ(Z),Yi,Yk+i)≤φ2,Y(i)andφ(σ(Yi),Z,Yk+i)≤1,φ(σ(Yk+i),Z,Yi)≤1, we infer that
kE(XiXk+i|F0)−E(XiXk+i)kp/2≤32(φ2,Y(i))(s−2)/sM2/s. (4.17)
From (4.15), (4.16) and (4.17), we infer that the conclusion of Lemma 4.1 holds provided that
X
i>0
[i
(s−2)/(s−1)]
X
k=1
1
(i+k)b
(φ2,Y(i))(s−2)/s+
X
k≥0
[k
(s−1)/(s−2)]
X
i=1
1
(i+k)b
(φ1,Y(k))(s−1)/s<∞.
Here, note that
[i(s−2)/(s−1)]
X
k=1
1
(i+k)b ≤i −b+ss−2
−1 and
[k(s−1)/(s−2)]
X
i=1
1
(i+k)b ≤
[2k(s−1)/(s−2)]
X
m=1
1
mb ≤Dk
(1−b)(s−1)
(s−2),
for someD>0. Sinceφ1,Y(k)≤φ2,Y(k), the conclusion of Lemma 4.1 holds provided
X
i≥1
i−b+ss−2−1φ 2,Y(i)
s−2
s <∞ and
X
k≥1 k(1−b)
(s−1) (s−2)φ
2,Y(k)
s−1
s <∞.
To complete the proof, it remains to prove that the second series converges provided the first one does. If the first series converges, then
lim n→∞
2n
X
i=n+1
i−b+ss−−21φ 2,Y(i)
s−2
s =0 . (4.18)
Since φ2,Y(i) is non increasing, we infer from (4.18) that φ2,Y(i)1/s = o(i−1/(s−1)−(1−b)/(s−2)). It follows that φ2,Y(k)(s−1)/s ≤ Cφ2,Y(k)(s−2)/sk−1/(s−1)−(1−b)/(s−2) for some positive constant C, and the second series converges.
4.3.1 Application to Expanding maps
LetBV be the class of bounded variation functions from[0, 1]toR. For anyh∈BV, denote bykdhk
the variation norm of the measuredh.
LetT be a map from[0, 1]to[0, 1]preserving a probabilityµon[0, 1], and let
Sn(f) =
n
X
k=1
(f ◦Tk−µ(f)).
Define the Perron-Frobenius operatorK fromL2([0, 1],µ)toL2([0, 1],µ)viathe equality
Z 1
0
(Kh)(x)f(x)µ(d x) =
Z 1
0
h(x)(f ◦T)(x)µ(d x). (4.19)
A Markov KernelK is said to beBV-contracting if there existC >0 andρ∈[0, 1[such that
The mapT is said to beBV-contracting if its Perron-Frobenius operator isBV-contracting.
Let us present a large class ofBV-contracting maps. We shall say thatT is uniformly expanding if it belongs to the classC defined in Broise (1996), Section 2.1 page 11. Recall that if T is uniformly expanding, then there exists a probability measureµon[0, 1], whose densityfµ with respect to the
Lebesgue measure is a bounded variation function, and such thatµis invariant byT. Consider now the more restrictive conditions:
(a) T is uniformly expanding.
(b) The invariant measureµis unique and(T,µ)is mixing in the ergodic-theoretic sense.
(c) 1
fµ
1fµ>0 is a bounded variation function.
Starting from Proposition 4.11 in Broise (1996), one can prove that if T satisfies the assumptions (a), (b) and (c) above, then it is BV contracting (see for instance Dedecker and Prieur (2007), Section 6.3). Some well known examples of maps satisfying the conditions (a), (b) and (c) are:
1. T(x) =βx−[βx]forβ >1. These maps are calledβ-transformations.
2. Iis the finite union of disjoint intervals(Ik)1≤k≤n, andT(x) =akx+bkon Ik, with|ak|>1.
3. T(x) =a(x−1−1)−[a(x−1−1)]for somea>0. Fora=1, this transformation is known as the Gauss map.
Proposition 4.4. Letσ2n=n−1E(S2
n(f)). If T is BV -contracting, and if f belongs toC(p,M,µ)with
p∈]2, 3], then the seriesµ((f−µ(f))2)+2Pn>0µ(f◦Tn·(f−µ(f)))converges to some nonnegative
σ2, and
1. ζ1(Pn−1/2S
n(f),Gσ2) =O(n
−1/2logn), for p=3,
2. ζr(Pn−1/2S
n(f),Gσ2) =O(n
1−p/2)for r
∈[p−2, 2]and(r,p)6= (1, 3),
3. ζr(Pn−1/2S
n(f),Gσ2n) =O(n
1−p/2)for r
∈]2,p].
Proof of Proposition 4.4. Let (Yi)i≥1 be the Markov chain with transition KernelK and invariant measureµ. Using the equation (4.19) it is easy to see that(Y0, . . . ,Yn)is distributed as(Tn+1, . . . ,T). Consequently, to prove Proposition 4.4, it suffices to prove that the sequence Xi = f(Yi)−µ(f)
satisfies the condition (4.13) of Proposition 4.3.
According to Lemma 1 in Dedecker and Prieur (2007), the coefficientsφ2,Y(i)of the chain(Yi)i≥0
with respect toFi =σ(Yj,j≤i)satisfyφ2,Y(i)≤Cρifor someρ∈]0, 1[and some positive constant
C. It follows that (4.13) is satisfied fors=p.
5
Proofs of the main results
From now on, we denote byC a numerical constant which may vary from line to line.
Notation 5.1. Forlinteger,qin]l,l+1]and f l-times continuously differentiable, we set
|f|Λq =sup{|x−y| l−q
5.1
Proof of Theorem 2.1
We prove Theorem 2.1 in the caseσ=1. The general case follows by dividing the random variables by σ. Since ζr(PaX,PaY) = |a|rζ
r(PX,PY), it is enough to bound up ζr(PSn,Gn). We first give an upper bound forζp,N:=ζp(PS2N,G2N).
Proposition 5.1. Let(Xi)i∈Zbe a stationary martingale differences sequence inLp for p in]2, 3]. Let
Mp=E(|X0|p). Then for any natural integer N ,
2−2N/pζ2p/,Np≤Mp+ 1
2p2 N
X
K=0
2K(p/2−2)kZKk1,Φ,p2/p+2
p∆N, (5.1)
where ZK =E(S22K|F0)−E(S22K)and∆N =
PN−1
K=02−2K
/pkZ Kkp/2.
Proof of Proposition 5.1. The proof is done by induction on N. Let (Yi)i∈N be a sequence of
N(0, 1)-distributed independent random variables, independent of the sequence(Xi)i∈Z. Form>0,
letTm=Y1+Y2+· · ·+Ym. SetS0=T0=0. For any numerical function f andm≤n, set
fn−m(x) =E(f(x+Tn−Tm)).
Then, from the independence of the above sequences,
E(f(Sn)−f(Tn)) =
n
X
m=1
Dm with Dm=E fn
−m(Sm−1+Xm)− fn−m(Sm−1+Ym)
. (5.2)
For any two-times differentiable functiong, the Taylor integral formula at order two writes
g(x+h)−g(x) =g′(x)h+1
2h
2g′′(x) +h2
Z 1
0
(1−t)(g′′(x+th)−g′′(x))d t. (5.3)
Hence, for anyqin]2, 3],
|g(x+h)−g(x)−g′(x)h−1
2h
2g′′(x)| ≤h2
Z 1
0
(1−t)|th|q−2|g|Λqd t≤ 1
q(q−1)|h|
q
|g|Λq. (5.4)
Let
D′m=E(f′′
n−m(Sm−1)(Xm2 −1)) =E(fn−m′′ (Sm−1)(Xm2 −Y
2
m))
From (5.4) applied twice with g = fn−m, x = Sm−1 and h = Xm or h = Ym together with the martingale property,
Dm−
1 2D
′ m
≤
1
p(p−1)|fn−m|ΛpE(|Xm| p+
|Ym|p).
NowE(|Ym|p)≤p−1≤(p−1)Mp. Hence
|Dm−(D′m/2)| ≤Mp|fn−m|Λ
p (5.5)
Assume now that f belongs toΛp. Then the smoothed function fn−m belongs to Λp also, so that
|fn−m|Λp ≤1. Hence, summing onm, we get that
E(f(Sn)− f(Tn))≤nMp+ (D′/2) where D′=D′
1+D′2+· · ·+D′n. (5.6)
Notation 5.2. Set m0 = m−1 and write m0 in basis 2: m0 = PiN=0bi2i with bi = 0 or bi = 1 (note that bN =0). SetmL=PiN=Lbi2i, so thatmN =0. LetIL,k=]k2L,(k+1)2L]∩N(note that
IN,1=]2N, 2N+1]),U(Lk)=Pi∈I
L,kXi and ˜U
(k)
L =
P
i∈IL,kYi. For the sake of brevity, letU
(0)
L =ULand ˜
U(L0)=U˜L.
SincemN=0, the following elementary identity is valid
Dm′ =
N−1
X
L=0
E
(fn′′−1−m
L(SmL)− f ′′
n−1−mL+1(SmL+1))(X 2
m−1)
.
NowmL6=mL+1 only if bL=1, then in this casemL=k2L withkodd. It follows that
D′=
N−1
X
L=0
X
k∈IN−L,0
kodd
E
(fn−′′1−k2L(Sk2L)−f′′
n−1−(k−1)2L(S(k−1)2L))
X
{m:mL=k2L}
(Xm2 −σ2). (5.7)
Note that{m:mL=k2L}=IL,k. Now by the martingale property,
Ek2L
X
i∈IL,k
(X2i −σ2)=Ek2L((U(k) L )
2)
−E((UL(k))2):=Z(Lk).
Consequently
D′ =
N−1
X
L=0
X
k∈IN−L,0
kodd
E
fn′′−1−k2L(Sk2L)−f′′
n−1−(k−1)2L(S(k−1)2L)ZL(k)
=
N−1
X
L=0
X
k∈IN−L,0
kodd
E
fn′′
−1−k2L(Sk2L)−f′′
n−1−k2L(S(k−1)2L+Tk2L−T(k−1)2L)Z(k) L
, (5.8)
since(Xi)i∈N and(Yi)i∈Nare independent. By using (1.2), we get that
D′≤
N−1
X
L=0
X
k∈IN−L,0
kodd
E(|UL(k−1)−U˜(k−1)
L | p−2
|Z(Lk)|).
From the stationarity of(Xi)i∈Nand the above inequality,
D′≤ 1
2 N−1
X
K=0
2N−KE(|UK−U˜K|p−2|Z(1)
K |). (5.9)
Now letVK be the N(0, 2K)-distributed random variable defined fromUK via the quantile transfor-mation, that is
VK=2K/2Φ−1(FK(UK−0) +δK(FK(UK)−FK(UK−0)))
whereFKdenotes the d.f. ofUK, and(δK)is a sequence of independent r.v.’s uniformly distributed on
[0, 1], independent of the underlying random variables. Now, from the subadditivity of x →xp−2,
|UK−U˜K|p−2≤ |UK−VK|p−2+|VK−U˜K|p−2. Hence
E(|UK−U˜K|p−2|Z
(1)
K |)≤ kUK−VKkpp−2kZ
(1)
K kp/2+E(|VK−U˜K|p−2|Z
(1)
By definition of VK, the real number kUK −VKkp is the so-called Wasserstein distance of order p
between the law of UK(0)and the N(0, 2K) normal law. Therefrom, by Theorem 3.1 of Rio (2007) (which improves the constants given in Theorem 1 of Rio (1998)), we get that, forp∈]2, 3],
kUK−VKkp≤2(2(p−1)ζp,K)1/p≤2(4ζp,K)1/p. (5.11)
Now, sinceVK and ˜UK are independent, their difference has theN(0, 2K+1)distribution. Note that if Y is a N(0, 1)-distributed random variable,Q|Y|p−2(u) = (Φ−1(1−u/2))p−2. Hence, by Fréchet’s
inequality (1957) (see also Inequality (1.11b) page 9 in Rio (2000)), and by definition of the norm
k.k1,Φ,p,
E(|VK−U˜K|p−2|Z(1)
K |)≤2
(K+1)(p/2−1)
kZKk1,Φ,p. (5.12)
From (5.10), (5.11) and (5.12), we get that
E(|UK−U˜K|p−2|Z(1)
K |)≤2p−
4/pζ(p−2)/p
p,K kZKkp/2+2(K+1)(p/2−1)kZKk1,Φ,p. (5.13)
Then, from (5.6), (5.9) and (5.13), we get
2−Nζp,N ≤Mp+2p/2−3∆′N+2 p−2−4/p
N−1
X
K=0
2−Kζp(p−,K2)/pkZKkp/2,
where∆′N =PNK=−012K(p/2−2)kZKk1,Φ,p. Consequently we get the induction inequality
2−Nζp,N≤Mp+ 1
2p2∆ ′ N+
N−1
X
K=0
2−Kζ(pp,K−2)/pkZKkp/2. (5.14)
We now prove (5.1) by induction onN. First by (5.6) applied withn=1, one hasζp,0≤Mp, since
D1′ = f′′(0)E(X2
1 −1) =0. Assume now thatζp,L satisfies (5.1) for any L in [0,N−1]. Starting from (5.14), using the induction hypothesis and the fact that∆′K ≤∆′N, we get that
2−Nζp,N ≤Mp+ 1
2p2∆ ′ N+
N−1
X
K=0
2−2K/pkZKkp/2
Mp+ 1
2p2∆ ′ N
2/p
+ 2 p∆K
p/2−1
.
Now 2−2K/pkZKkp/2= ∆K+1−∆K. Consequently
2−Nζp,N ≤Mp+ 1
2p2∆ ′ N+
Z ∆N
0
Mp+ 1
2p2∆ ′ N
2/p
+2 px
p/2−1
d x,
which implies (5.1) forζp,N.
In order to prove Theorem 2.1, we will also need a smoothing argument. This is the purpose of the lemma below.
Lemma 5.1. Let S and T be two centered and square integrable random variables with the same variance. For any r in]0,p],ζr(PS,PT)≤2ζr(PS∗G1,PT ∗G1) +4
p
Proof of Lemma 5.1. Throughout the sequel, letY be aN(0, 1)-distributed random variable, inde-pendent of theσ-field generated by(S,T).
Forr≤2, sinceζris an ideal metric with respect to the convolution,
ζr(PS,PT)≤ζr(PS∗G1,PT∗G1) +2ζr(δ0,G1)≤ζr(PS∗G1,PT∗G1) +2E|Y|r
which implies Lemma 5.1 forr≤2. Forr>2, from (5.4), for any f inΛr,
f(S)− f(S+Y) + f′(S)Y−1
2f
′′(S)Y2
≤ 1
r(r−1)|Y|
r.
Taking the expectation and noting thatE|Y|r≤r−1 forr in]2, 3], we infer that
E f(S)−f(S+Y)−1
2f ′′(S)
≤1
r.
Obviously this inequality still holds for T instead of S and−f instead of f, so that adding the so obtained inequality,
E(f(S)− f(T))≤E(f(S+Y)−f(T+Y)) +1
2E(f
′′(S)−f′′(T)) +1.
Since f′′belongs toΛr−2, it follows that
ζr(PS,PT)≤ζr(PS∗G1,PT∗G1) +
1
2ζr−2(PS,PT) +1. Nowr−2≤1. Hence
ζr−2(PS,PT) =Wr−2(PS,PT)≤(Wr(PS,PT))r−2.
Next, by Theorem 3.1 in Rio (2007),Wr(PS,PT)≤(32ζr(PS,PT))1/r. Furthermore
(32ζr(PS,PT))1−2/r≤ζr(PS,PT)
as soon asζr(PS,PT)≥2(5r/2)−5. This condition holds for anyr in]2, 3]ifζr(PS,PT)≥4
p
2. Then, from the above inequalities
ζr(PS,PT)≤ζr(PS∗G1,PT ∗G1) +
1
2ζr(PS,PT) +1,
which implies Lemma 5.1.
We go back to the proof of Theorem 2.1. Letn∈]2N, 2N+1]andℓ=n−2N. The main step is then to prove the inequalities below: for r ≥p−2 and (r,p)6= (1, 3), for someε(N)tending to zero as
N tends to infinity,
ζr(PSn,Gn)≤cr,p2
N(r−p)/2ζ
p(PSℓ,Gℓ) +C(2N(r+2−p)/2+2N((r−p)/2+2/p)ε(N)(ζp(PSℓ,Gℓ))(p−2)/p)
(5.15) and forr=1 andp=3,
ζ1(PSn,Gn)≤C(N+2 −Nζ