www.elsevier.com / locate / econbase
The bias of the 2SLS variance estimator
a b ,
*
Jan F. Kiviet , Garry D.A. Phillips
a
Tinbergen Institute and Faculty of Economics and Econometrics, University of Amsterdam, Amsterdam, The Netherlands
b
School of Business and Economics, University of Exeter, Streatham Court, Exeter EX4 4PU, UK
Accepted 8 April 1999
Abstract
In simultaneous equation models the two stage least squares (2SLS) estimator of the coefficients, though consistent, is biased in general and the nature of this bias has given rise to a good deal of research. However, little if any attention has been given to the bias that arises when an estimate of the asymptotic variance is used to approximate the small sample variance. In this paper we use asymptotic expansions to show that, in general, the asymptotic variance estimator has an upwards bias. 2000 Elsevier Science S.A. All rights reserved.
Keywords: 2SLS estimation; Nagar expansions; Asymptotic variance; Variance estimation bias
JEL classification: C30
1. Introduction
The seminal paper of Nagar (1959) presented approximations for the bias of the 2SLS estimator to
21 22
order T and for the mean squared error to the order of T , where T is the sample size. By subtracting the square of the bias approximation from the mean squared error approximation we
22
obtain an estimator for the variance to the order of T . Little use seems to have been made of this particular approximation; indeed, the problem of bias in the estimation of the coefficient estimator variance seems generally to have been neglected. However, this approximation can be used to explore the bias of the estimated asymptotic variance as an estimator of the small sample variance. By finding
22 an approximation for the expectation of the asymptotic variance estimator to the order of T , and comparing it with the variance approximation of the same order, we may deduce immediately the
22
approximate bias in the asymptotic variance estimator. This bias, which is of order T , is found to be non-negative for all coefficients in the 2SLS estimator showing that, in general, the traditional
*Corresponding author. Tel.: 144-1392-263-241; fax: 144-1392-263-242. E-mail address: [email protected] (G.D.A. Phillips)
estimator is upwards biased. This is the main theoretical result in the paper. Given that an explicit expression for the bias approximation is obtained, a bias correction can routinely be applied.
2. Model and notation
Consider a general static simultaneous equation model containing G equations which may be written as
A9yt1B9xt5et, t51, . . . ,T, (1)
where y is a Gt 31 vector of endogenous variables, x is a Kt 31 vector of strongly exogenous variables which we shall treat as non-stochastic, and et is a G31 vector of structural disturbances.
A9and B9 are respectively, G3G and G3K matrices of structural coefficients. With T observations
on the above system, we may write
YA1XB5E (2)
where Y is a T3G matrix of observations on the endogenous variables, X is a T3K matrix of
observations on the exogenous variables, and E is a T3G matrix of structural disturbances. We shall
be particularly concerned with that part of the system (2) which relates to the first equation. The reduced form of the system includes
Y15XP11V1 (3)
where Y 5( y :Y ), X5(X :X ), P 5(p :P ) and V 5(v :V ).P is a K3( g11) matrix of reduced
1 1 2 1 2 1 1 2 1 1 2 1
form parameters and V is a T1 3( g11)matrix of reduced form disturbances. In addition, the following assumptions are made:
• The rows of V are independently and normally distributed with mean vector 01 9 and non-singular covariance matrix V5
h j
vij.• The T3K matrix X is of rank K (,T ), and the elements of the K3K matrix X9X are O(T ).
• The first equation of system (1) is overidentified with the order of overidentification, L, being at least 2. This ensures that the first two moments of 2SLS exist; see Kinal (1980).
3. Asymptotic approximations
The first equation of (2) may be written as
y15Y2b1X1g1e1 (4)
where y and Y are, respectively, a T1 2 31 vector and a T3g matrix of observations on g endogenous
variables and X is a T1 3k matrix of observations on k non-stochastic exogenous variables. The
2 of independently and identically distributed normal random variables with mean zero and variances . The 2SLS estimators of the unknown parameters of (4) are given by
21
on X. From (5) we may write the estimation error as
21
In what follows it will be convenient to re-write (4) in the form
y15Z1a1e1 (7)
Before stating the approximations that are the focus of interest, we shall define the following:
¯ ¯ ¯
With the above definitions we may state the following:
21
2 21
E(a*2a)5s (L21)Qc1o(T ) (9)
22
• 2SLS mean squared error to order T : Nagar (1959, p. 579).
2 2
E (ha*2a)(a*2a)9 5j s Q1s
f
tr(CQ )22(L21)tr(C Q ) Q1g
2 2 22
1s
f
(L 23L14)QC Q1 2(L22)QCQg
1o(T ) (10)21
• Bias of the residual variance estimator to order T : Nagar (1961, p. 240).
2 2 2 21
E(s*2s )5 2s
f
2(L21)tr(QC )1 2tr(QC )g
1o(T ) (11)e9*e*
2 ]]]]
where s*5T2( g1k) and e*5y12Z1a* is a T31 vector of 2SLS residuals.
These are slight adaptations of the published results which we shall use later in the paper. In fact Nagar (1961) deflates the sum of squared residuals by T and, as a result, the estimator is biased to
21
order T . We prefer to use the less biased version: see also Kiviet and Phillips (1998).
4. The bias of the asymptotic variance estimator
Subtracting the outer product of (9) from (10), we may deduce an approximation to the variance of 2SLS as follows: used to estimate the variance in finite samples and it is the bias of this estimator which is the main focus of interest in this paper. However, we shall first consider the bias of a non-operational estimator
2 21 2
˜ ˆ ˆ
of the variance given by Var(a*)5s (Z9Z ) where s is known. Since none of the resulting bias 2
can be attributed to the estimator of s , a consideration of this case will be helpful in analysing the source of the bias in the estimated asymptotic variance. In Appendix 1 the following result is proved:
2 21
˜ ˆ ˆ
Lemma 1. The expected value of the non-operational variance estimator Var(a*)5s (Z9Z ) can
be approximated as
2 2 22
˜
E Var(
f
a*)g
5s Q1s ftr(CQ )Q2(L22)QCQg1o(T ). (13)The result of this lemma, combined with (12), leads to the following.
22
˜
Theorem 1. The bias of the non-operational variance estimator Var(a*), to order T , is given by
2 22
˜
Notice that tr(C Q )Q and QC Q are both positive semi-definite matrices where tr(C Q )Q1 1 1 $QC Q;1 see Kadane (1971, p. 728). Hence the bias matrix above is positive semi-definite for L$2. However,
L$2 is a requirement for the variances to exist so that, in general, the non-operational variance
22
estimator is biased upwards to order T . Next we examine the bias of the asymptotic variance estimator. In Appendix 2 we show the following result.
2 21
ˆ ˆ ˆ
9
Lemma 2. The expected value of the asymptotic variance estimator Var(a*)5s (Z Z ) to the * 1 1
An approximation for the bias can now be readily obtained. Combining the result in Lemma 2 with the approximation in (12) gives the next theorem.
2 21
Noting that both tr(QC )Q and QC Q are positive semi-definite, it is clear that the estimated1 asymptotic variance is, in general, biased upwards to the order of the approximation. Comparing 2 Theorems 1 and 2, it is seen that the direction of the bias is unaltered by the need to estimate s although the bias expression itself changes. It does not appear possible to make general statements
ˆ
about the relative magnitudes of the two biases. This is partly because the bias of Var(a*) depends on
˜
the matrix C whereas the bias of Var(2 a*) does not.
ˆ
Given that an explicit expression for the bias of the asymptotic variance estimator Var(a*) has been found, a bias corrected estimator can be obtained straightforwardly. Estimates are available for the relevant terms in the bias approximation so that an estimate of the bias can be obtained which is then subtracted from the original estimator. We shall not pursue the matter further in this paper however.
5. Conclusion
This paper shows that the traditional variance estimator for the 2SLS coefficient estimator is, in
22
Appendix A
2 21 22
˜ ˆ ˆ
9
Here we derive the expected value of Var(a*)5s (Z Z )1 1 to the order T . We need only
2 ˆ ˆ
consider the inverse matrix since s is constant. First we examine the matrix Z15(Y :X ) where2 1
21 21
ˆ
Y25X(X9X ) X9Y25XP21X(X9X ) X9V . Noting that this last term is stochastic, it is seen that we2
ˆ
may write the matrix Z as the sum of two matrices, one stochastic and one non-stochastic, as follows:1
21 21
expanded in terms of decreasing orders of smallness. To obtain E (Z Z )
f
1 1g
we shall take expectations of each of the relevant terms in (A1) as follows:¯ ¯ ¯ we may use the results in Mikhail (1972) or Kiviet and Phillips (1996, p. 166) to show directly that:
¯ ¯ ¯ ¯
˘
9
˘9
˘9
9
˘E Q(V Z )Q(V Z )Q
f
2 1 2 1g
5QCQ, E Q(V Z )Q(Z V )Qf
2 1 1 2g
5( g1k)QCQ¯
9
˘ ˘9
¯ ¯9
˘ ¯9
˘E Q(Z V )Q(V Z )Q
f
1 2 2 1g
5tr(QC )Q, E Q(Z V )Q(Z V )Qf
1 2 1 2g
5QCQ.Adding terms we find that in (iii)
¯ ¯ ¯ ¯
˘
9
9
˘ ˘9
9 9
˘E Q(V Z
f
2 11Z V )Q(V Z1 2 2 11Z V )Q1 2g
5( g1k12)QCQ1tr(QC )Q. (A.2)2
Finally, using (i)–(iii) above, we have found that the expected value ofs times (A1) is given by:
2 ˆ ˆ
9
2 22Appendix B
2 21
ˆ ˆ ˆ
9
In this Appendix we prove Lemma 2. Thus we derive the expected value of Var(a*)5s*(Z Z )1 1
22 2
to o(T ) where s* is given in (11). To find the required expansion we first note that the 2SLS residual vector is given by
*
e1 5y12Z1a*5e12Z (1 a*2a) and the sum of squared residuals is then
9
*
9
9
9
The inverse matrix (Z Z )1 1 is similarly expanded in terms of decreasing stochastic order of magnitude in (A.1) as
To find the appropriate expansion for s*(Z Z )1 1 we combine (B.1) and (B.2) to yield
9
We shall take expectations of each of these terms. However the analysis of the first is simplified by noting that
9
where terms involving products of an odd number of zero mean normal random variables have been ignored. The first two expected values are;9
see Kiviet and Phillips (1996, p. 166).
The third term can be readily evaluated as in (A.2) so that
2 ¯
9
˘ ˘9
¯ ¯9
˘ ˘9
¯ 2 2E
h
sf
Q(Z V1 21V Z )Q(Z V2 1 1 21V Z )Q2 2g
j
5s ( g1k12)QCQ1s tr(QC )Q (B.5) Gathering terms from (B.4) and (B.5), we have found the expectation of the first term of (B.3). Next we consider the second term in (B.3). To analyse this term we first note the expansion21
The first of these terms involves a product of an odd number of normal random variables with zero mean and so has expected value zero. Taking expected values for the remaining terms we have;
¯ ¯
Here the results for evaluating the second, third and fourth terms are given in Nagar (1961, p. 242).
˘
An evaluation of the last term proceeds from putting V25W1e t1 9 where W and e1 are independent. The required analysis is straightforward but lengthy and so is not included here. The authors will provide details on request. Collecting the terms in (B.8) and multiplying by 22 yields the expectation of the second term of (B.3). To complete the analysis we need the expected value of the last term in (B.3). From Nagar (1961, p. 243) we find that
2
Collecting the various terms we have the result given in Lemma 2.
References
Kadane, J.B., 1971. Comparison of k-class estimators when the disturbances are small. Econometrica 39, 723–737. Kinal, T.W., 1980. The existence of moments of k-class estimators. Econometrica 48, 241–249.
Kiviet, J.F., Phillips, G.D.A., 1996. The bias of the ordinary least squares estimator in simultaneous equation models. Economics Letters 53, 161–167.
Kiviet, J.F., Phillips, G.D.A., 1998. Degrees of freedom adjustment for disturbance variance estimators in dynamic regression models. Econometrics Journal 1, 44–70.
Mikhail, W.M., 1972. The bias of the two stage least squares estimator. Journal of the American Statistical Association 67, 625–627.
Nagar, A.L., 1959. The bias and moment matrix of the general k-class estimators of the parameters in simultaneous equations. Econometrica 27, 575–595.