
CHAPTER 4
✦
The Least Squares Estimator
105
is included in the regression than when it is not. Does the same hold for the sample
estimate of this covariance matrix? Why or why not? Assume that X and z are
nonstochastic and that the coefficient on z is nonzero.
9. For the classical normal regression model y = Xβ + ε with no constant term and
K regressors, assuming that the true value of β is zero, what is the exact expected
value of F[K, n − K] = (R
2
/K)/[(1 − R
2
)/(n − K)]?
10. Prove that E [b
b] = β
β + σ
2
K
k=1
(1/λ
k
) where b is the ordinary least squares
estimator and λ
k
is a characteristic root of X
X.
11. For the classical normal regression model y = Xβ + ε with no constant term and
K regressors, what is plim F[K, n − K] = plim
R
2
/K
(1−R
2
)/(n−K)
, assuming that the true
value of β is zero?
12. Let e
i
be the ith residual in the ordinary least squares regression of y on X in the
classical regression model, and let ε
i
be the corresponding true disturbance. Prove
that plim(e
i
− ε
i
) = 0.
13. For the simple regression model y
i
= μ + ε
i
,ε
i
∼ N[0,σ
2
], prove that the sam-
ple mean is consistent and asymptotically normally distributed. Now consider the
alternative estimator ˆμ =
i
w
i
y
i
, w
i
=
i
(n(n+1)/2)
=
i
i
i
. Note that
i
w
i
= 1.
Prove that this is a consistent estimator of μ and obtain its asymptotic variance.
[Hint:
i
i
2
= n(n + 1)(2n + 1)/6.]
14. Consider a data set consisting of n observations, n
c
complete and n
m
incomplete,
for which the dependent variable, y
i
, is missing. Data on the independent variables,
x
i
, are complete for all n observations, X
c
and X
m
. We wish to use the data to
estimate the parameters of the linear regression model y = Xβ + ε. Consider the
following the imputation strategy: Step 1: Linearly regress y
c
on X
c
and compute
b
c
. Step 2: Use X
m
to predict the missing y
m
with X
m
b
c
. Then regress the full sample
of observations, (y
c
, X
m
b
c
), on the full sample of regressors, (X
c
, X
m
).
a. Show that the first and second step least squares coefficient vectors are identical.
b. Is the second step coefficient estimator unbiased?
c. Show that the sum of squared residuals is the same at both steps.
d. Show that the second step estimator of σ
2
is biased downward.
15. In (4-13), we find that when superfluous variables X
2
are added to the regression of
y on X
1
the least squares coefficient estimator is an unbiased estimator of the true
parameter vector, β = (β
1
, 0
)
. Show that in this long regression, e
e/(n −K
1
−K
2
)
is also unbiased as estimator of σ
2
.
16. In Section 4.7.3, we consider regressing y on a set of principal components, rather
than the original data. For simplicity, assume that X does not contain a constant
term, and that the K variables are measured in deviations from the means and
are “standardized” by dividing by the respective standard deviations. We consider
regression of y on L principal components, Z = XC
L
, where L < K. Let d denote
the coefficient vector. The regression model is y = Xβ + ε. In the discussion, it is
claimed that E[d] = C
L
β. Prove the claim.
17. Example 4.10 presents a regression model that is used to predict the auction prices
of Monet paintings. The most expensive painting in the sample sold for $33.0135M
(log = 17.3124). The height and width of this painting were 35” and 39.4”, respec-
tively. Use these data and the model to form prediction intervals for the log of the
price and then the price for this painting.