1
Quiz 2- Linear Regression Analysis (Based on Lectures 15-31) Time: 1 Hour
1. The random errors ε in multiple linear regression model y=Xβ ε+ are assumed to be identically and independently distributed following the normal distribution with zero mean and constant variance. Here y is a n×1 vector of observations on response variable, X is a n K× matrix of n observations on each of the K explanatory variables,
β is a K×1 vector of regression coefficients and ε is a n×1 vector of random errors.
The residuals ˆε = −y yˆ based on the ordinary least squares estimator of β have, in general,
(A) zero mean, constant variance and are independent (B) zero mean, constant variance and are not independent (C) zero mean, non constant variance and are not independent (D) non zero mean, non constant variance and are not independent
Answer: (C)
Solution:The residual is
1 1
2
ˆ ˆ
where ( ' ) '
( ) where ( ' ) '
( )
( )ˆ 0
( )ˆ ( )
y y
y Xb b X X X y
I H y H X X X X
I H E
V I H
ε
ε ε
ε σ
−
−
= −
= − =
= − =
= −
=
= −
• Since E( )εˆ =0, so ˆεi's have zero mean.
• Since I−H is not generally a diagonal matrix, so ˆ 'εi sdo not have necessarily the same variances.
• The off-diagonal elements in (I−H) are not zero, in general. So ˆ 'εi s are not independent.
2
2. Consider the multiple linear regression model y= Xβ ε+ , ( )E ε =0,
2 2 2
1 2
( ) ( , ,..., n)
V ε =diag σ σ σ where y is a n×1 vector of observations on response variable, X is a n K× matrix of n observations on each of the K explanatory variables,
β is a K×1 vector of regression coefficients and ε is a n×1 vector of random errors.
Let hii be the ith diagonal element of matrix H =X X X( ' )−1. The variance of the ith PRESS residual is
(A) σi2(1−hii) (B) σi2(1−hii)2 (C)
2
1
i
hii
σ
− (D)
2
(1 )2 i
hii
σ
−
Answer: (C)
Solution:
Let ei = −yi yˆibe the ith ordinary residual based on the ordinary least squares estimator of β. The ith PRESS residual e(i) is( ) 1
i i
ii
e e
= h
− .
The variance of e( )i is obtained as follows:
( ) 2
( ) ( )
(1 )
i i
ii
Var e Var e
= h
− and
2
( ) ( ) ( )
( )i i (1 ii).
V e I H V
Var e h
ε σ
= −
⇒ = −
Thus
2
( ( )) . 1
i i
ii
Var e
h
= σ
−
3
3. Consider the linear model y=β1X1+β2X2+ε, ( )E ε =0, ( )V ε =Iwhere the study variable y and the explanatory variables X1 and X2are scaled to length unity and the correlation coefficient between X1 and X2is 0.5. Let b1 and b2be the ordinary least squares estimators of β1 and β2respectively. The covariance between b1 and b2 is
(A) 2/3 (B) -2/3 (C) -0.5 (D) 1/3
Answer: (B)
Solution:
If ris the correlation coefficient betweenX1 and X2, then the ordinary least squares estimators b1 and b2 of β1 and β2 are the solutions of1 1
2
( ' ) ' where ' 1 . 1
b r
b X X X y X X
b r
−
= = =
Then
1
2
1
2
1 2 2
1 1 ( ' )
1 1
1 1 ( ) ( ' )
1 1
( , ) .
1 X X r
r r V b X X r
r r Cov b b r
r
−
−
−
= − −
−
= = − −
= − −
1 2
When 0.5, cov( , ) 2. r= b b = −3
4
4. Under the multicolinearity problem in the data in the modely=β1X1+β2X2+ε, where the study variable y and the explanatory variables X1 and X2are scaled to length unity and the random error ε is normally distributed withE
( )
ε =0,V( )
ε =σ2I whereσ2is unknown. What do you conclude about the null hypothesis H01:β1=0 and H02:β2 =0 out of the following when sample size is small.(A) Both H01and H02 are more often accepted.
(B) Both H01and H02 are more often rejected.
(C) H01is accepted more often and H02 is rejected more often.
(D) H01is rejected more often and H02 is accepted more often.
Answer: (A)
Solution:
If ris the correlation coefficient betweenX1 and X2, then the ordinary least squares estimators b1 and b2are the solutions of1 1
( ' ) ' where ' . 1
b X X X y X X r
r
−
= =
Then
2
2 1
2
( )
( ) ( ' ) 1 .
1 1
E b
V b X X r
r r β
σ − σ
=
−
= = − −
As rbecomes higher, the variance of b1 and b2become larger and so the test statistic
, 1, 2 var( )
i i
i
t b i
b
= =
for testing H0i:β =i 0 becomes very small and so H0iis more often accepted.
5
5. Consider the setup of multiple linear regression model y=Xβ ε+ where y is a n×1 vector of observations on response variable, X is a n K× matrix of n observations on each of the K explanatory variables, β is a K×1 vector of regression coefficients and ε is a n×1 vector of random errors following N(0,σ2I)with all usual assumptions. Let
( ' ) and 1 ii
H =X X X − h be the ith diagonal element of H. A data point is a leverage point if hiiis
(A) greater than the value of average size of hat diagonal.
(B) less than the value of average size of hat diagonal.
(C) greater or less than the value of average size of hat diagonal.
(D) none of the above.
Answer: (A)
Solution:
The ithdiagonal element of His hii = x (1i X X1 )−1xi where xi is the ith row of X matrix. Thehiiis a standardized measure of distance of ith observation from the center (or centroid) of the x-space.( )
1 ( )Average size of hat diagonal .
n ii i
h rank H trH K
h n n n n
=
∑
= = = =Ifhii >2hthen the point is remote enough to be considered as a leverage point.