linear dependencies among the regressors. The variance proportions associated
with eigenvalue number 10 indicate a correlation between male age and male age
at union. Those associated with eigenvalue number 11 suggest that female age,
female age at union, and union duration are somewhat linearly dependent. These
are all precisely the variables that would be exactly collinear were it not for
missing data.
Alternatives to OLS When Regressors Are Collinear
Several simple remedies for collinearity problems were discussed in Chapter 3,
including dropping redundant variables, incorporating variables into a scale, employ-
ing nonlinear transformations, and centering (for collinearity arising from cross-prod-
uct terms). However, there are times when none of these solutions are satisfactory. For
example, in the body fat data, I may want to know the effect of, say, triceps skinfold
thickness on body fat, net of (i.e., controlling for) the effects of thigh circumference
and midarm circumference. Or, in the NSFH example, I may want to tease out the sep-
arate effects of male- and female age, male- and female age at the beginning of the
union,and union duration, on couple disagreement. None of the simple remedies are
useful in these situations. With this in mind, I will discuss two alternatives to OLS:
ridge regression and principal components regression. These techniques are some-
what controversial (see, e.g., Draper and Smith, 1998; Hadi and Ling, 1998).
Nevertheless, they may offer an improvement in the estimates of regressor effects
when collinearity is severe.
First we need to consider a key tool in the evaluation of parameter estimators: the
mean squared error of the estimator, denoted MSQE (to avoid confusion with the MSE
in regression). Let θ be any parameter and θ
ˆ
its sample estimator. Then MSQE(θ
ˆ
)
E
θ
ˆ
(θ
ˆ
θ)
2
. That is, MSQE(θ
ˆ
) is the average, over the sampling distribution of θ
ˆ
, of the
squared distance of θ
ˆ
from θ. All else equal, estimators with a small MSQE are pre-
ferred, since they are by definition closer, on average, to the true value of the parame-
ter, compared to other estimators. It can be shown that MSQE(θ
ˆ
) V(θ
ˆ
) [B(θ
ˆ
)]
2
,
where B(θ
ˆ
) is the bias of θ
ˆ
(defined in Chapter 1). Both ridge and principal components
regression employ biased estimators. However, both techniques offer a trade-off of a
small amount of bias in the estimator for a large reduction in its sampling variance.
Ideally, this means that these techniques bring about a substantial reduction in the
MSQE of the regression coefficients compared to OLS.
At the same time, both techniques have a major drawback, particularly in the
social sciences, where hypothesis testing is so important: The extent of bias in the
regression coefficients is unknown. Therefore, significance tests are not possible. To
understand why, consider the test statistic for the null hypothesis that β
k
0. For sim-
plicity, suppose that the true variance of b
k
is known and that n is large, so that t tests
and z tests are equivalent. The test relies on the fact that b
k
is unbiased for β
k
. Now
the test statistic is
z
b
k
σ
b
β
k
k,0
,
REGRESSION DIAGNOSTICS II: MULTICOLLINEARITY 231