Upload
alexandre-ripamonti
View
217
Download
0
Embed Size (px)
Citation preview
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
1/33
Econometric Theory,30, 2014, 252284.doi:10.1017/S0266466613000182
ASYMPTOTIC NORMALITY FOR
WEIGHTED SUMS OF LINEAR
PROCESSES
KARIM M. ABADIR AND WALTER DISTASO
Imperial College London
LIUDAS GIRAITIS
Queen Mary, University of London
HIRA
L. KOUL
Michigan State University
We establish asymptotic normality of weighted sums of linear processes with gen-
eral triangular array weights and when the innovations in the linear process are
martingale differences. The results are obtained under minimal conditions on the
weights and innovations. We also obtain weak convergence of weighted partial sum
processes. The results are applicable to linear processes that have short or long mem-
ory or exhibit seasonal long memory behavior. In particular, they are applicable to
GARCH and ARCH() models and to their squares. They are also useful in de-riving asymptotic normality of kernel-type estimators of a nonparametric regression
function with short or long memory moving average errors.
1. INTRODUCTION
Numerous inference procedures in statistics and econometrics are based on the
sums
Sn=n
j=1
Xj , Wn=n
j=1
zn jXj
of a linear process{Xj }, where{zn j , 1 j n} is an array of known real num-bers. In this paper we focus on deriving asymptotic distributions ofSn and Wn .
In addition, the weak convergence property of the corresponding partial sum
The authors would like to thank Donatas Surgailis for providing part (iii) of Theorem 2.2, Lemma 3.1, and other
useful comments, and to the Editor and the three reviewers for their constructive suggestions. Research is in part
supported by ESRC grant RES062230790 and USA NSF DMS Grant 0704130. Address correspondence to Liudas
Giraitis, School of Ecnomics and Finance, Queen Mary, University of London, Mile End Rd., London E14NS, United
Kingdom; e-mail: [email protected]
252 c Cambridge University Press 2013
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
2/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 253
processes is also discussed. The linear process is assumed to be a moving average
with martingale differences innovations and may exhibit short- or long-range de-
pendence. It will be shown that{Var(Tn)}1/2(Tn ETn ), with Tn= Sn orWn ,converges weakly to a normal distribution under easily verifiable minimal as-sumptions on weights and innovations. In particular, these assumptions are valid
for the nonlinear squared autoregressive conditional heteroskedasticity (ARCH)
process where innovations are conditionally heteroskedasticity martingale differ-
ences. The proofs use the central limit theorem (CLT) for martingale differences.
Numerous testing procedures, e.g., testing for unit root, Cumulative Sum
(CUSUM) change-point detection, and Kwiatkowski, Phillips, Schmidt and Shin
(1992) (KPSS) test for stationarity, are based on the weak convergence of the par-
tial sum process S[nt],0 t 1 to a Gaussian process, and, in particular, requireverification of the CLT for Sn . The CLT for weighted sums Wn , in turn, is used inkernel estimation and in spectral analysis, e.g., obtaining asymptotic normality of
the discrete Fourier transforms of{Xj }. It is thus of interest to provide easy-to-useCLTs and invariance principles for the above statistics.
The work of Ibragimov and Linnik (1971) contains a number of useful results
on classical asymptotic theory of weakly dependent random variables. Davydov
(1970) obtained a weak convergence result for the partial sum process of lin-
ear processes with independent and identically distributed (i.i.d.) innovations,
whereas Phillips and Solo (1992) developed CLT and invariance principles for
sums of linear processes based on Beveridge-Nelson decomposition. Peligrad andUtev (2006) extended Ibragimov and Linnik (Thm. 18.6.5) for linear processes
with innovations following a more general dependence framework. Gordin (1969)
introduced a general method for proving CLTs for stationary processes using a
martingale approximation. In the case of short memory, his method gives the same
result as the Beveridge-Nelson decomposition. Wu and Woodroofe (2004) ob-
tained a CLT for the sums of stationary and ergodic sequences using a martingale
approximation method. Merlevede, Peligrad, and Utev (2006) provide a further
survey of some recent results on CLT and its weak invariance principle for sta-
tionary processes.Section 2 deals with asymptotic normality ofSn and Wn , whereas in Section 3
we discuss weak convergence of the corresponding partial sum processes.
Section 4 contains examples and applications, while Section 5 includes simula-
tions. In the sequel, all limits are taken as n , unless specified otherwise; pand D , respectively, denote the convergence in probability and in distribution,Z = {0,1,2, . . .}; and Nk(,) denotes the k-dimensional normal distribu-tion with mean vectorand covariance matrix,k 1. We write N for N1. Forany two sequences of real numbers,an, bn ,an bn meansan /bn 1.
2. CLT FOR WEIGHTED SUMS
In this section we consider asymptotic normality of the sums Sn and Wn when
{Xj } is a moving average process,
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
3/33
254 KARIM M. ABADIR ET AL.
Xj=
k=0akjk=
j
k=
ajkk, j Z,
k=0a2k< . (2.1)
Here,(j
,Fj
), j
Z is a martingale difference sequence (m.d.s.) with constant
variance, where Fj := -field{i , i j}, j Z, i.e., E(j |Fj1) = 0, E2j= 2 K) 0, as K .First, we discuss the asymptotic normality ofSn . Theorem 18.6.5 of Ibragimov
and Linnik (1971) gives a CLT forSnin the case of i.i.d. innovations under general
conditions. Theorem 2.1 below extends it to m.d. innovations and shows that any
rate of divergence Var(Sn) of the variance guarantees the CLT. Peligradand Utev (2006, Prop. 4) proved this result when
j
is a stationary and ergodic
m.d.s. For clarity, we provide a brief proof based on the ideas of these works.
Note that if{Xj } is a zero-mean Gaussian process, then (Var(Sn))1/2 Sn=DN(0, 1) for all n 1, and the question about the asymptotic distribution of Snreduces to finding the asymptotics of the Var(Sn).
THEOREM 2.1. Suppose {Xj } is a linear process (2.1) where {j } is either astationary and ergodic m.d.s., or satisfies Assumption 2.1. Then
2n := Var(Sn) implies
1n SnD N(0, 1). (2.2)Proof. For simplicity of notation, set ak= 0, k= 1,2, . . . in (2.1), and let
cn j= 1
n
n
k=max(j,1) akj= 1
n
n
k=1 akj , j Z. Without loss of generality,assume 2= 1. Then,
1n Sn =n
j=
cn j j , 2n Var(Sn ) =
n
j=
c2n j= 1, n 1. (2.3)
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
4/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 255
Next we show
n
j=cn j cn,j1
2
=o(1), cn :
=max
jn |cn j
| =o(1), (2.4)
which together with Lemma 2.1 below implies (2.2).
To prove (2.4), note that for any s n,t= 1, 2, . . . ,s
j=st
c2n j=s
j=st
cn,j1 +
cn j cn,j1
2,
c2n,s = c2n,st1 + 2s
j=
s
t
cn,j1
cn j cn,j1
+
s
j=
s
tcn j cn,j1
2
.
Because jn c2n,j= 1, then for every s Z and n 1, limt c2n,st1= 0.Sincetis arbitrary, take the limit t and use the Cauchy-Schwarz inequalityto obtain
c2n,s = 2s
j=
cn,j1
cn j cn,j1+ s
j=
cn j cn,j1
2
2
n
j=
c2n j1/2
Bn
+B2n
2Bn
+B2n , (2.5)
where Bn :=
nj=(cn j cn,j1)2
1/2does not depend ons . Since by defini-
tioncn j cn,j1 = 1n (a1j anj+1),
B2n 4 2n
j=0a2j= o(1),
because 2n
and
j
=0 a
2j 0. (2.12)
Letqn := nj=Md2n j Vj denote the left-hand side (l.h.s.) of (2.11). We shall showthat
Eqn 1, Var(qn) 0,
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
6/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 257
which yields E(qn 1)2 0 and, together with Chebyshevs inequality, will im-ply (2.11). The first claim follows from EVj=E2j= 1, (2.9) and nj= d2n j= 1.To prove the second claim, let 0 :
=maxj
ZEV
2j K
[. . .] +n
j,k=M: |jk|K
[. . .]
K
n
j=
d2n j
2+ d2n
n
j=M
d2n j k: |jk|K
|V(j,j )V(k, k)|1/2 K + 0(2K+ 1)d2n 0,
by Assumptions 2.1(a,b) and (2.6). This completes the proof of (2.11).
To prove (2.12), note that the expected value of the l.h.s. of (2.14) is bounded
above by
n
j=M
d2n jE
2j I|j | d1n max
jE
2j I|j | d1n 0,
by Assumptions 2.1(c) and (2.6), which together with the Markov inequality
implies (2.12), and also completes the proof of part (i) of the lemma.
Case (ii). The proof of (2.10) combines arguments used in the proofs of
Peligrad and Utev (1997, Thm. 2.1; 2006, Prop. 4). Let Xn j := dn j j ,Mj n. By Hall and Heyde (1980, Thm. 3.2) to prove (2.10), it suffices to verify(a) maxMjn |Xn j | p 0, (b) nj=MX2n jp1, and (c) EmaxMjnX2n j=O(1).
First, we show that (2.8) implies (2.6). To see this, use the bound analogous
to (2.5), which does not depend on any particular form ofcn j , to obtain d2n,s
2Bn +B2n ,where now Bn :=
nj=
dn j dn,j1
21/2. Hence, (2.8) implies
dn 0.Now, claims (a) and (c) follow because for any >0,
E n
j=M
dn j j
2I(|dn j j | )
E21I(|1| d1n ) nj=
d2n j 0.
To show (b), we need to verify
n
j=M
d2n j 2
jp1. (2.13)
We do this by truncation. For ak 1, writen
j=M
d2n j 2j=n
j=M
d2n j +n
j=M
d2n j2j k1 kl=1
2j+l+
n
j=M
d2n j
k1
k
l=1
2j+l 1
=:qn1 + qn2 + qn3.
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
7/33
258 KARIM M. ABADIR ET AL.
In view of (2.9),qn1 = nj= d2n j + O(1/ log n) 1. Since E2j= 1,
E|qn2| k1E
n
j=Mkd2
n j
2
jk
l=1
n+l
j=M+l d2
n,jl 2
j k1
n
j=
|kd2n j d2n,j1 . . . d2n,jk|E
2j
+ dnE
2n+1 + . . . + 2n+k+ 2M
kn
j= |d2n j d
2n,j1|+ (k+ 1)d
2n 0, k 1,
because nj= |d2n j d2n,j1|
nj=
dn j dn,j1
21/22
nj= d
2n j
1/2 0, by (2.8).
Finally, since {j } is stationary and ergodic, by the ergodic theorem, see, e.g.,Stout (1974, Cor. 3.5.2), E
k1 kl=1 2j+l E21 = Ek1 kl=1 2l1 0,k, and thus E|qn3| o(1)nj= d2n j = o(1), which implies (2.13) and completesthe proof. n
Remark 2.2. Note that the above lemma clearly includes the case of{j } IID(0, 1). Furthermore, it gives a generalization of Theorem V.1.2.1 of Hajek and
Sidak (1967) wheredn j= 0, j 0 and {j } IID(0, 1). Ifj IID(0, 1), assump-tion (2.6) implies the weak Lindeberg condition, see Hall and Heyde (1980, p. 53),
and is minimal. Indeed, let Sn= nj=1 dn j j , dn j=
1 + cnj n , c> 0. If >0, then max1jn |dn j | = o(1) and the CLT holds, as also shown in Phillipsand Magdalinos (2007). However, if = 0, then max1jn |dn j | =(1 + c)1,(2.6) does not hold and SnD j=1(1 + c)j j , which is Gaussian only if{j }is Gaussian, as pointed out in Anderson (1959). Peligrad and Utev (1997, Thm.2.1) derived Lemma 2.1, requiring instead of Assumption 2.1 that the m.d.s. {j }be a pairwise mixing sequence. In their Example 2.1 they showed that (2.6) and
stationarity and ergodicity of m.d.s. alone are not sufficient for the CLT.
The next corollary provides a straightforward generalization of Lemma 2.1 to a
two-sided moving average of m.d.s., while the result for weighted sums of a sta-
tionary ergodic process stated in Proposition 2.1 is useful in various applications.
COROLLARY 2.1. Suppose Sn= j= dn j j, n 1, where{j } is anm.d.s., and{dn j } are such thatjZ d2n j= 1.
(i) If{j } satisfies Assumption 2.1 and maxjZ |dn j | = o(1), then SnDN(0, 2).
(ii) In addition, if{j } is stationary ergodic andjZ(dn j dn,j1)2 = o(1),
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
8/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 259
then SnD N(0, 2), and
jZ
d2n j 2
j= E21+op(1). (2.14)
PROPOSITION 2.1. Suppose {j } is a stationary ergodic sequence, E|1| 0, |u| u0, for some c > 0, and u 0> 0, and (2.20) holds, thenthe conditions of Theorem 2.2(iii)are satisfied, and CLT (2.18) holds.(c) Suppose Assumption 2.1 is satisfied, there exists c> 0such that f(u) c>0, u , and
max1jn
|zn j | = o n
j=1
z2n j
1/2. (2.22)
Then the conditions of Theorem 2.2(i)are satisfied, and CLT (2.18) holds.
Proof.(a) Let
G(
u):=
n
j=1e
ijuzn j ,
u . SinceX(
k) = eiku f(u)d u,
2n=n
j,k=1
zn j X(j k)znk=
f(u)|G(u)|2du
= 2f(0)n
j=1
z2n j + in, in := 2n2f(0)n
j=1
z2n j . (2.23)
It remains to show that
in= 2n2f(0) n
j=1z2n j
= o nj=1
z2n j
, (2.24)
which proves (2.21) and together with (2.20) verifies condition (iii) of Theorem
2.2 and thus (2.18).
Verification of (2.24) is based on two facts, which will be proved later:
|G(u)|2du = 2n
j=1z2n j , sup|u| |uG(u)| = o
n
j=1z2n j. (2.25)
We proceed as follows. Let >0. Choose > 0, such that sup0u|f(u) f(0)| . Then by (2.25),
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
12/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 263
in |u|
|f(u) f(0)||G(u)|2du
|u||G(u)|2du +
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
13/33
264 KARIM M. ABADIR ET AL.
Remark 2.5. Proposition 2.2 verifies the CLT for smooth weights zn j and
stationary and ergodic m.d.s.{j }; e.g., in Corollary 4.4 below, it is shown that(2.20) holds for kernel weights zn j= K((nx j )/(nb)) in nonparametric re-gression setups. As long as fis continuous at zero, (2.20) yields the asymptoticbehavior of the variance (2.21) and the asymptotic normality ofWn .
If (2.20) does not hold, Theorem 2.2(i) or Theorem 2.2(ii) may be applied.
For instance, alternating weights zn j= (1)j = eij do not satisfy (2.20), butTheorem 2.2(i) is applicable. Indeed, maxj |zn j | = 1 and nj=1z2n j = n. In addition,if f(u) f() >0, u , then, letting Dn(u):= nl=1 eilu ,
Var(Wn) =n
j,k
=1
zn j X(j k)znk=
|Dn(u + )|2 f(u)du
=
0f(u )|Dn (u)|2du +
0
f(u + )|Dn(u)|2du
= f()
|Dn(u)|2du + o(n) = 2f( )n + o(n),
using f( ) = f(), as in the proof of (2.21). This shows the applicability ofTheorem 2.2(i) and hence the CLT for Wn . Note that here (2.21) for Var(Wn)does
not hold.
The following proposition, which is valid for any short memory covariance-stationary process{Xj }, provides an upper bound for Var(Wn) and analyzes itsasymptotic behavior.
PROPOSITION 2.3. Let{Xj } be a covariance-stationary process with zeromean, finite variance, and covariance function such that
kZ
| (k)| < . (2.27)
Let Wn be as in (2.17) with zn j satisfy conditions (2.20) and (2.22). Then, n
j=1z2n j
1Var(Wn)
kZ| (k)|,
n
j=1
z2n j
1Var(Wn) 2 :=
kZ (k). (2.28)
Proof.Under (2.27),
f(u)= (2 )1 kZ
eiku (k) (2 )1 kZ
| (k)| < , u ,
f(u) f(0) = (2)1 kZ
(k) = (2 )1 2, u 0.
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
14/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 265
Thus, by (2.23) and (2.25),
Var(Wn) =
f(u)|G(u)|2du supu
f(u)
|G(u)|2du
kZ
| (k)|n
j=1
z2n j ,
which proves the first bound of (2.28). The proof of the second bound is the same
as that of (2.21). This completes the proof of the proposition. n
The following result is a multivariate generalization of Theorem 2.2.
THEOREM 2.3. Let z(i )n,j, i
=1, . . . , k be k arrays of real weights, and
{Xj
}is
a linear process (2.1) with m.d.s. {j }. Assume that sums W(i)n :=nj=1z(i )n,jXj and(
(i )n )
2 := Var(W(i)n ), i= 1, . . . , k satisfy one of the conditions (a) or (b).(a){j } is stationary and ergodic, and each sum W(i)n , i= 1, . . . , k satisfies
condition (iii) of Theorem 2.2.
(b){j } satisfies Assumption 2.1, and each sum W(i )n ,i= 1, . . . , k, satisfies oneof the conditions of (i)(iii) of Theorem 2.2.
Let for some (positive definite) matrix,CovW(i )n / (i )n , W(j )n / (j)n i,j=1,...,k
. (2.29)
Then,W(1)n /
(1)n , . . . , W
(k)n /
(k)n
D Nk(0,). (2.30)
Proof.Similarly, as in (2.19), write
S
(i )
n := W(i )
n /(i )
n =n
j=d(i )
n jj , i= 1, . . . , k.To prove (2.30), in view of the Cramer-Wold device, it suffices to show that for
everya = (a1, . . . , ak) Rk andk 1,Sn := a1 S(1)n + + akS(k)n D N(0, aaT). (2.31)Write Sn= nj= dn j j where dn j= a1d(1)n j + + akd(k)n j . Condition (2.29)implies that
Var(Sn) = E21n
j=
d2n j aaT.
Assume that m.d.s. satisfies Assumption 2.1. As seen in the proof of Theorem
2.2, any one of the conditions (i), (ii), or (iii) assures that maxjn |d(i )n j | 0,
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
15/33
266 KARIM M. ABADIR ET AL.
i= 1, . . . , k, which yields maxjn |dn j | 0.Hence, coefficientsdn j ofSn satisfyassumptions of Lemma 2.1, which implies (2.31).
Assume that m.d.s. is stationary ergodic. In the proof of Theorem 2.2, it was
shown that condition (iii) yields jn d(i )n j d(i)n,j12 0,i= 1, . . . , k. Conse-quently, jn(dn j dn,j1)2 0, and (2.31) follows by Lemma 2.1. n
3. WEAK CONVERGENCE OF PARTIAL SUM PROCESSES
A number of econometric applications require the weak convergence of a suitably
standardized partial sums process Sn ( ) = [n ]j=1Xj , >0 to some limit processS(),0 1. Observe that for each n , Sn(),0 1,is a step function in , belonging to the Skorokhod functional space D[0, 1].
From Billingsley (1968), we recall that a sequence of stochastic processes
{Zn()}, n 1 in D[0, 1] is said to converge weakly to a stochastic processZ() C[0, 1], and we write Zn Z, if every finite dimensional distribution of{Zn()} converges to that of Z() and if{Zn ()} is tight with respect to the uni-form metric; see also Pollard (1984). The uniform topology is stronger than the
Skorokhod J1-topology, and the verification of tightness in the uniform metric is
relatively easier.
Using the arguments in Section 12 and Theorem 15.5, p. 127 of Billingsley
(1968), see also Pollard (1984, Ch. V.1, Thm. 3), one can show that a sufficient
condition for tightness of{Zn} is the following: There exists a sequence of nonde-creasing right continuous functions Fn on [0, 1] that are uniformly bounded and
converge uniformly to a continuous function Fsuch that for some >1, >0,
and 0 s< t 1,E|Zn(t) Zn (s)| C[Fn (t) Fn(s)], n 1, (3.1)
whereCmay depend on , but not ons,t, andn.
Now, let Sn= Sn(1)and 2n := Var(Sn). Our goal here is to establish the weakconvergence of
{
1
n S
n()}
. For this purpose, we need to establish that its finite di-
mensional distributions converge to those of the limit process, denoted by f d d,and that the process is tight in uniform metric. We first focus on the finite dimen-
sional convergence.
According to the Lamperti theorem (1962), if 1n Sn( )
f d d S( ), (3.2)then for some H (0, 1)and a positive slowly varying function L, 2n
=Var(Sn)
=n2HL(n). (3.3)
In most applications
2n= Var(Sn) s2n2H , for some 0< H
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
16/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 267
In the case of a linear process{Xj } of (2.1), with m.d. stationary and ergodicinnovations {j }, (3.4) is also sufficient for (3.2).
The limits in this section will be described by the fractional Brownian motion
(fBm),BH(),0 1, with parameter 0 0
f d d
BH( )
>0, (3.6)
where BHis the fBm with parameter H.
Proof. LetTn ( ) := 1n Sn( ). Assumption (3.4) implies that for any 0 < < 1,
Var(
Tn ()) =
Var(Sn ())
Var(Sn) =(n )2H(1
+o(1))
n2H(1 + o(1)) 2H
,and hence,
Cov(Tn(t), Tn (s)) = (1/2){Var(Tn (t)) + Var(Tn(s)) Var
Tn(t) Tn(s)}
rH(t, s), 00, and k 1, Sn= a1 Sn ([nt1]) + + akSn ([ntk])satisfy
1
n SnD S:= a1BH(t1) + + akBH(tk).By (3.7), Var(Sn) , and the rest of the proof repeats the lines of proof ofTheorem 2.1. n
The following proposition is a simplified version of the result obtained by
Taqqu (1975). It shows that long memory of the summands{Xj } a priori guar-antees the tightness of the normalized partial sum process. It does not require
{Xj } to be a linear process.PROPOSITION 3.2. Let
{Xj
}be a second-order stationary process satisfying
(3.4), with1/2< H
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
17/33
268 KARIM M. ABADIR ET AL.
Proof. To check tightness, we shall verify (3.1) for the process Tn(t) := 1n Sn (t). Let Fn (t):= [nt]/n, F(t):= t,0 t 1. Observe that supt |Fn(t) F(t)| 0 and Fis continuous on [0, 1]. By covariance stationarity of the incre-ments,
ETn (t) Tn(s)2 =E 1n [nt][ns ]
j=1Xj
2
=
[nt] [ns ]n
2H(1 + o(1))(1 + o(1)) C
[nt] [ns]
n
2H= C
Fn(t) Fn(s)
2H
, (3.9)
for some >0. Since := 2H >1, this verifies (3.1) for the Tn process with= 2 and completes the proof. nThe following lemma, whereSn= Sn(1), gives a useful bound for moments of
the sums of a linear process with martingal difference (m.d.) innovations and is
useful in proving the tightness of the process {Tn(t)} in the cases of short and neg-ative memory processes{Xj }. It extends the well-known Burkholder-Rosenthalinequality for martingales and some other inequalities involving m.d.s. of Dhar-
madhikari, Fabian, and Jogdeo (1968, Theorem in Section 1) and Borovskikh and
Korolyuk (1997, Chap. 3).
LEMMA 3.1. Let{Xj } be a linear process (2.1) with m.d. innovations{j },such thatE2j= 2 for all j, and:= maxj E|j |p < , for some p 2. Then
E|Sn|p cES2n
p/2, n 1, (3.10)
where c>0 depends only on p and.
Proof. Setting dn j= nk=max(j,1) akj , write Sn= nj= dn j j . BecauseE2
j= 2
, ES2
n= 2
n
j=d2
n j
2, (3.11)with a constant Cp>0 depending only on p.
The inequality (3.11) remains valid also for n = .
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
18/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 269
Proof. Let n< . For p 2, by the Burkholder-Rosenthals inequality (seeHall and Heyde, 1980, p. 24),
E nj=1
Yj p Cp nj=1
E|Yj |p +E nj=1
E[Y2j|Fj1]p/2. (3.12)Recall the fact that for any real numbers a,b, and for 0 < 1,|a + b| |a| +|b| . Apply this fact with = 2/p 1 to obtain
n
j=1
E|Yj |p2/p
n
j=1
E|Yj |p
2/p. (3.13)
To bound the second term on the r.h.s. of (3.12), first use the CauchySchwarz
inequality for the conditional expectation with p/2 1 to obtain E[Y2j|Fj1] (E[|Yj |p|Fj1])2/p, 1 j n. Next, by Minkowski inequality for any r.v.s XandY,
(E|X+ Y|r)1/r (E|X|r)1/r + (E|Y|r)1/r, r 1.These facts, in turn, imply
E
nj=1E
Y2j|Fj1
p/2 Enj=1(E|Yj |p|Fj1])2/pp/2 nj=1 E|Yj |p2/pp/2,
which together with (3.13) and (3.12) proves (3.11).
Consider now the case of an infinite sum. Let p 2. We claim
E
j=1
Yj
p Cpj=1
(E|Yj |p)2/pp/2
. (3.14)
This inequality is trivially true if the r.h.s. of (3.14) is infinite. Assume it is finite.
By (3.11), for any 1
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
19/33
270 KARIM M. ABADIR ET AL.
Claim (3.14) now is proved upon taking limit as n in this bound. nWe are now ready to state and prove the following weak convergence result.
THEOREM 3.1. Assume that the m.d.s. {j } in the linear process {Xj } is eitherstationary and ergodic or satisfies Assumption 2.1. Suppose that Sn satisfies (3.4)
with0< H1/H.Then,
1n Sn () BH(), in D[0, 1]and uniform metric, (3.15)where BHis a fBm.
Proof. Proposition 3.1 implies the finite dimensional convergence. To prove
tightness, we shall use (3.10). For H
1/2, by (3.10) and (3.9),
E|Tn (t) Tn(s)|p cE(Tn(t) Tn(s))2p/2 cFn(t) Fn(s)H p.This verifies (3.1) for the Tn process with = p, = H p>1. For H >1/2,(3.15) follows from Proposition 3.2. This completes the proof. n
Remark 3.1. A functional CLT for the partial sums process of a short memory
linear process {Xj } of (2.1) with m.d. innovations andaj s such that j=0 j |aj | 0, bk 0, k 1,
k=1bk
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
20/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 271
4.1.2. Squared ARCH Process. A centered squared ARCH process
Xj= r2jEr2j,
withrj as in (4.1) and satisfying E1/2
[4
0 ]k=1 bk
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
21/33
272 KARIM M. ABADIR ET AL.
Use this bound, C-S, and BE1/240m, Cov( 2mt, 2m j ) = 0. We will show that for |t j | >m,max
jE( 2j 2m j )2 m 0, m . (4.7)
The latter with the bound |Cov(X, Y)| (EX2EY2)1/2 and (4.5) implies
|Cov( 2t,
2j)
| = Cov( 2mt, 2m j ) + Cov( 2t 2mt, 2m j )+ Cov( 2mt, 2j 2m j ) + Cov( 2t 2mt, 2j 2m j )
C(1/2m + m ) 0, m , (4.8)
which verifies Assumption 2.1(b) with K= m. To verify (4.7), notice thatbj 0implies hk,j hm,k,j . Hence, with Bm := ms=1 bs andbs= bsI(1 s m), asin (4.4),
E(hk,j
hm,k,j )
2
=
s1,...,sk=1(bs1 . . . bsk
bs1 . . . bsk)
2j
s1
. . . 2j
s1
sk
2(4.9)
s1,...,sk=1
{bs1 . . . bsk bs1 . . . bsk}2
(E40)k
= (BkBkm )2(E40)k.Next, since 0< Bm B, by the mean value theorem, Bk Bkm k Bk1(B Bm ) k Bkm ,wherem := (B Bm )/B 0, asm . Hence, as in (4.5),
E( 2j
2m j )
2
=b20E
k=1{hk,j
hm,k,j
}2
b20
k=1kBE1/2[40 ]
k2 2m= C2m 0, m , (4.10)which proves (4.7).
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
22/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 273
Assumption 2.1(c) follows by Markovs inequality, noting that Er4j =E41 E
41 < .
(iii) Here, Vj
= 4j Var(
20)
=E[2j
|Fj
1]. Thus, verifying Assumption 2.1(a)
is equivalent to proving maxj E 8j < . Toward this goal, first note that by theHolders inequality and by arguing as for (4.4), Eh4k,j B4k(E80)k. Together withassumption BE1/4[80 ]m ,
maxj
E
4j 4m j2 m 0, m , (4.12)
which together with (4.11), as in (4.8), implies that
|Cov 4t, 4j | = Cov 4mt, 4m j+ Cov 4t 4mt, 4m j
+Cov
4mt,
4j
4m j+ Cov 4t 4mt, 4j 4m j C1/2m + m 0, m .
Finally, to prove (4.12), use Holder inequality, as in (4.9) and (4.10), to obtain
E
hk,j hm,k,j4
s1,...,sk=1bs1 . . . bsk
m
s1,...,sk=1
bs1 . . . bsk
4E80
k
BkBkm
4
E80
k
E80
k
k4B4k4m ,
E 2j 2m j4 b40Ek=1
hk,j hm,k,j4
b40
k=1k(BE1/4[80 ])
k
4 4m= C4m 0, m ,
because BE1/4[80 ]
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
23/33
274 KARIM M. ABADIR ET AL.
Assumption 2.1(c) follows by Markovs inequality, noting that E4j CE81 E
81 < , which completes the proof of the proposition. n
4.1.3. Stochastic Volatility m.d. Processes. By a stochastic volatility model
one usually understands a stationary processj ,j Z of the formj= j j , j IID(0, 1), jZ,where the (volatility) process j >0 is a function of the past information up to
time j 1. Let Fj1 be the -field generated by past innovations s , s j 1,and E 2j < . Then E[j |Fj1] = 0, 2j= Var(j |Fj1), and {j } is a white noiseprocess, which by Theorem 3.5.8 of Stout (1974) is also stationary and ergodic.
It is often assumed that the volatility process j= h(j ), j Z is a nonlin-ear function of a stationary Gaussian or linear process{j }, see, e.g., Robinson(2001). The choice of h(j )= exp(j ) includes the exponential generalizedARCH (EGARCH) model, proposed by Nelson (1991). A related class of stochas-
tic volatility models with long memory in{j } was introduced and studied inBreidt, Crato, and de Lima (1998), Harvey (1998), and Surgailis and Viano
(2002). As a rule, the volatility process Vj 2j in these models is a stationaryprocess with autocovariances decaying to zero and thus satisfies Assumption 2.1.
4.1.4. Sample ACF. Let {j } be a stationary and ergodic m.d.s. such that 2=E20 < . Consider the sample autocorrelation k= k/0, k 1, wherek=n1 nj=k+1(j )(jk ) and= n1 nj=1 j . It is well known that in thecase of i.i.d. random variables,
nkD N(0, 1) and (
n1, . . . ,
nk) D
N(0,I), where the limit is a vector ofk independent standard normal variables.
These results are widely used for testing for the absence of correlation, see, e.g.,
Brockwell and Davis (1991). The next proposition shows that, in general, this
asymptotic normality result is true only for i.i.d. sequences.
Letbe m
mmatrix with(j, k)th elementj,k=
E[21
1j
1k
]/ 4
.
PROPOSITION 4.2. Let{j } be a stationary and ergodic m.d.s. with 2 :=E20 < such thatE
21
21k
< , for k= 1, . . . , m, where m is a given positive
integer. Then,n1, . . . ,
nm
D Nm (0,). (4.13)In particular, (4.13) holds in case of ARCH processj= rj of (4.1), such thatE
1/2(40)k=1 bk
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
24/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 275
ergodicity of 2j , 0p E20 . Hence,
nk= ( 2)1n1/2 nj=k+1 j jk+O(n1/2) N(0, k,k).
To prove (4.13), let (k)
j
=j j
k , k
=1, . . . , m. Then for any real numbers
c1, . . . , cm ,j := c1(1)j + . . . + cm (m)j is a stationary ergodic m.d.s., with E2j= 2 := c211,1 + . . . + c2m m,m . Let Sn := nj=m+1 j . Then ES2n= 2 n , andby Theorem 2.1, n1/2 SnD N(0, 2 ), which by the Cramer-Wold device im-plies the claim of (4.13).
In case of ARCH processrj , (4.1) and (4.3) implyk,k 1 and show thatisdiagonal if, in addition, E3j= 0. n
Remark 4.1. Proposition 4.2 shows that for the ARCH m.d.s. (4.1), the limit
variance ofnk, as the rule, is greater than one, and so the 95% confidence bandfor k= 0 is wider than in the i.i.d. case. Moreover, the limit matrix may benondiagonal, unless the distribution ofj in (4.1) is symmetric.
4.2. Regression Models
In this section we discuss application of the above results for obtaining asymptotic
normality of least squares (LS) estimators in parametric regression models and
kernel-type estimators in nonparametric regression models.
But first we verify the above conditions for asymptotic normality Sn and weakconvergence of the partial sum process {Sn(), 0 1} of a linear process {Xj }of (2.1) in some typical cases.
Let (j ), j= 0, 1, 2, . . . denote the autocovariance function of{Xj } and fits spectral density. Note that (j ) := Cov(Xj ,X0)= 2k=0 akak+j , j=0, 1, 2,. . . .Consider the assumption in terms of f.
f(u) c|u|2d, u 0, |d| 0. (4.14)In terms of (j ), consider the condition
jZ
| (j )| < , 2 := jZ
(j ) >0, ford = 0,
(j ) c |j |1+2d, 0
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
25/33
276 KARIM M. ABADIR ET AL.
where
s2 = 2f(0), d= 0;
s2 = cf R
sin
2
(u/2)(u/2)2
|u|2ddu, 0< |d|
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
26/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 277
Now, note that= nj=1zn jXj /nj=1z2n j=:Wn/nj=1z2n j . Define
v2d:= 10 g2(u)dukZ (k), d= 0,
:= c 1
0
10
g(u)g(v)|u v|1+2ddudv, 0K
| (s)| 0, K ,
whereas for any |i | K,n1 nk=1zn,k+iznk (i ) (i )1
0 g2(u)du.Whence,
limK
limn
n
j,k
=1:
|j
k|
K
zn jznk (j k) =
1
0g2(u)du lim
K
|i|
K
(i ) = v20 ,
which proves 2n v20.Next, consider the case 0
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
27/33
278 KARIM M. ABADIR ET AL.
n12d 2n= n12dn
k,j=1
zn jzn,kX(j k)
= n12dcn
k,j=1:k=jg jng kn|j k|1+2d+ o(1)
c 1
0
10
g(u)g(v)|u v|1+2ddudv= v2d.
This completes the proof of (4.22) and the corollary. n
4.2.2. A Non-Stationary Process. As another application of Theorem 2.2,
consider the weighted sum Vn := n
j=1zn j Yj of a nonstationary unit root processYj := js=1Xs , j= 1, 2, . . .. Set (u) =
1u g(v)dv,0 u 1. Also, define
v2d:= 1
02(u)du
kZ (k)
, d= 0,
:= c 1
0
10
(u)(v)|u v|1+2ddudv, 0
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
28/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 279
Then,n (x) (x) = Dn (x) + n (x) (x). Typically the bias term n(x) (x) is negligible compared to Dn(x), and asymptotic distribution ofn (x) (x)is determined by that ofDn (x). Fix anx (0, 1)and let
zn j := K1nx Kx n j
nb
. (4.23)
Then, clearly Dn(x)is like aWn . Define
2d,K := K22 kZ
(k), d= 0,
:= c 1
0
10
K(u)K(v)|u v|1+2ddudv, 0
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
29/33
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
30/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 281
FIGURE3. Kernel density ofVar( n (x))1/2( n (x)(x))(solid line),N(0, 1)density(dashed line).5.0.5. ARFIMA-ARCH
We start from the case where Xj in (2.1) is ARFIMA(1, d, 0) model with the
AR parameterr= 0.8 and m.d. innovationsj generated as ARCH(1)process:j= j j , j N(0, 1),
2j= 0 + 12j1, j= 1, . . . , n. (5.1)
We take 0= 0.2, 1= 0.8, so that the E2
j = 1. We simulated the cases d=0.1, 0.2, 0.3, 0.4, which gave qualitatively similar results. For the sake of brevity,we report the results ford= 0.3 only. We analyze the asymptotic normality of thesuitably standardized sums
Wn=n
j=1
zn jXj , Vn=n
j=1
zn j Yj ,
whereYj= js=1Xs , with the weights zn j= (j/n)2 + cos(j/n), j= 1, 2, . . . , n.Var(Wn) and Var(Vn) are estimated by the Monte Carlo variances. We then plota kernel estimate of the densities ofVar(Wn )1/2Wn andVar(Vn)1/2 Vn usinga Gaussian kernel with bandwidth b chosen according Silvermans (1986) rule,
bn = (4/3)1/5n1/5, and superimpose the standard normal density. The results forARFIMA(1, 0.3, 0),r= 0.8 process {Xj } with ARCH(1)innovations in Figure 1
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
31/33
282 KARIM M. ABADIR ET AL.
show a very close resemblance between the solid line representing the kernel
density estimate and the dashed line representing N(0, 1)density.
AR-ARCH
Next we analyze asymptotic normality of Wn and Vn in the case of AR(1)
processXj= Xj1 +j , j Z,with = 0.6 and ARCH(1)errorsj as in (5.1).The results in Figure 2 show that the kernel density estimate ofVar(Wn)1/2Wnseems to have less probability mass close to the mean than the N(0, 1) densityand is slightly positively skewed.
Nonparametric regression
To illustrate the fit of normal approximation in nonparametric estimation, we
use the regression model Yj
=(j/n)
+Xj , j
=1, . . . , n,where Xj s follow an
ARFIMA(1, 0.3, 0)-ARCH(1) process with AR parameter r= 0.8 generated asin (5.1). We set(j/n) = (j/n)2 + cos(j/n), and(x) is estimated at x= 1/4by a kernel-type estimator
n (x) =
nj=1K
nxj
nb
Yj
nj=1K
nxj
nb
,using a Gaussian kernel and setting b = n1/4(4/3)1/5 S D(J), where S D(J) isthe standard deviation of the regressor J= j/n, j= 1, 2, . . . , n. Var(n(x)) isestimated by the Monte Carlo variance. Figure 3 shows close fit of the estimateddensity ofVar(n(x))1/2(n(x) (x)), and the standard normal density.REFERENCES
Abadir, K.M. (1993) The limiting distribution of the autocorrelation coefficient under a unit root.
Annals of Statistics21, 10581070.
Abadir, K.M. (1995) The limiting distribution of the tratio under a unit root.Econometric Theory11,
775793.
Anderson, T.W. (1959) On asymptotic distributions of estimates of parameters of stochastic difference
equations.Annals of Mathematical Statistics 30, 676687.
Billingsley, P. (1968)Convergence of Probability Measures. Wiley.
Bollerslev, T. (1988) On the correlation structure for the generalized autoregressive conditional het-
eroskedastic process.Journal of Time Series Analysis., 9, 121131.
Borovskikh, Y.V. & V.S. Korolyuk (1997)Martingale Approximation. VSP.
Breidt, F.J., Crato, N. and de Lima, P. (1998) On the detection and estimation of long memory in
stochastic volatility.Journal of Econometrics83, 325348.
Brockwell, P.J. & R.A. Davis (1991) Time Series: Theory and Methods, 2nd ed. Springer Series in
Statistics. Springer-Verlag.
Davydov, Y.A. (1970) The invariance principle for stationary processes.Theory of Probability and its
Applications15, 487498.
Dharmadhikari, S.W., V. Fabian & K. Jogdeo (1968) Bounds on the moments of martingales. Annalsof Mathematical Statistics39, 17191723.
Dickey, D.A., & W.A. Fuller (1979) Distribution of estimators of autoregressive time series with a unit
root.Journal of the American Statistical Association 74, 427431.
Dickey, D.A. & W.A. Fuller (1981) Likelihood ratio statistics for autoregressive time series with a unit
root.Econometrica49, 10571072.
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
32/33
CLT FOR WEIGHTED SUMS OF LINEAR PROCESS 283
Dolado, J.J., J. Gonzalo, & L. Mayoral (2002) A fractional Dickey-Fuller test for unit roots. Econo-
metrica70, 19632006.
Engle, R.F. (1982) Autoregressive conditional heteroscedasticity with estimates of the variance of
United Kingdom inflation.Econometrica50, 9871008.
Giraitis, L., P. Kokoszka, R. Leipus (2000) Stationary ARCH models: Dependence structure and cen-
tral limit theorems.Econometric Theory16, 322.
Giraitis, L.R. Leipus & D. Surgailis (2007) Recent advances in ARCH modelling. In G. Teyssiere &
A.P. Kirman (eds.),Long Memory in Economics, pp. 338 Springer.
Gordin, M.I. (1969) The central limit theorem for stationary processes. Soviet Mathematics Doklady
10, 11741176.
Granger, C.W.J. & R. Joyeux (1980) An introduction to long-memory time series models and frac-
tional differencing.Journal of Time Series Analysis 1, 1529.
Gray, H.L., N.-F. Zhang & W.A. Woodward (1989) On generalized fractional processes. Journal of
Time Series Analysis10, 233257.
Hajek, J. & Z. Sidak (1967)Theory of Rank Tests. Academic Press.
Hall, P. & C.C. Heyde (1980) Martingale Limit Theory and Applications.Academic Press.
Harvey, A.C. (1998) Long memory in stochastic volatility. In J. Knight, & S. Satchell (eds.), Forecast-
ing Volatility in the Financial Markets, pp. 307320, Butterworth & Heineman.
Hosking, J.R.M. (1981) Fractional differencing. Biometrika68, 165176
Ibragimov, I.A. & Y.V. Linnik (1971) Independent and Stationary Sequences of Random Variables.
Wolters-Noordhoff.
Kwiatkowski, D., Phillips, P.C.B., Schmidt, P. and Shin, Y. (1992). Testing the null hypothesis of
stationary against the alternative of a unit root: how sure are we that economic time series have a
unit root?Journal of Econometrics54, 159178.
Lamperti, J.W. (1962) Semi-stable stochastic processes. Transactions of the American Mathematical
Society104, 6278.Marinucci, D. & P.M. Robinson (1999) Alternative forms of fractional Brownian motion. Journal of
Statistical Planning and Inference80, 111122.
Merlevede, F., M. Peligrad, & S. Utev (2006) Recent advances in invariance principles for stationary
sequences.Probability Surveys3, 136.
Nelson, D.B. (1991) Conditional heteroskedasticity in asset returns: A new approach. Econometrica
59, 347370.
Peligrad, M. & S. Utev (1997) Central limit theorem for stationary linear processes.Annals of Proba-
bility25, 443456.
Peligrad, M. & S. Utev (2005) A new maximal inequality and invariance principle for stationary
sequences.Annals of Probability33, 798815.
Peligrad, M. & S. Utev (2006) Central limit theorem for stationary linear processes.Annals of Proba-bility34, 16081622.
Phillips, P.C.B. (1987) Time series regression with a unit root. Econometrica55, 277301.
Phillips, P.C.B. & T. Magdalinos (2007) Limit theory for moderate deviations from a unit root.Journal
of Econometrics136, 115130.
Phillips, P.C.B. & V. Solo (1992) Asymptotics for linear processes. Annals of Statistics 20,
9711001.
Pollard, D. (1984)Convergence of Stochastic Processes. Springer Series in Statistics. Springer-Verlag.
Robinson, P.M. (1991) Testing for strong serial correlation and dynamic conditional heteroskedasticity
in multiple regression.Journal of Econometrics47, 6784.
Robinson, P.M. (1997) Large-sample inference for non-parametric regression with dependent errors.
Annals of Statistics25, 20542083.Robinson, P.M. (2001) The memory of stochastic volatility models. Journal of Econometrics 101,
195218.
Silverman, B.W. (1986)Density Estimation. Chapman & Hall.
Stout, W. (1974)Almost Sure Convergence. Academic Press.
8/13/2019 2014 Abadir Distaso Giraitis Koul Asymptotic Normality for Weighted Sums of Linear Processes
33/33
284 KARIM M. ABADIR ET AL.
Surgailis, D. & M.-C. Viano (2002) Long memory properties and covariance structure of the EGARCH
model.ESAIM: Probability & Statistics6, 311329.
Taqqu, M.S. (1975) Weak convergence to fractional Brownian motion and to the Rosenblatt process.
Zeitschritt f ur Wahrscheinlichkeitstheorie und verwandte Gebiete31, 287302.
Wu, W.B. & M. Woodroofe (2004) Martingale approximations for sums of stationary processes. An-
nals of Probability32, 16741690.