where, for emphasis, we let denote an F
random variable with (q,n k 1) degrees
of freedom, and F is the actual value of the
test statistic. The p-value still has the same
interpretation as it did for t statistics: it is the
probability of observing a value of F at least
as large as we did, given that the null hypoth-
esis is true. A small p-value is evidence
against H
0
. For example, p-value .016
means that the chance of observing a value
of F as large as we did when the null hypoth-
esis was true is only 1.6%; we usually reject
H
0
in such cases. If the p-value .314, then
the chance of observing a value of the F sta-
tistic as large as we did under the null
hypothesis is 31.4%. Most would find this to
be pretty weak evidence against H
0
.
As with t testing, once the p-value has been computed, the F test can be carried out at
any significance level. For example, if the p-value .024, we reject H
0
at the 5% signif-
icance level but not at the 1% level.
The p-value for the F test in Example 4.9 is .238, and so the null hypothesis that
motheduc
and
fatheduc
are both zero is not rejected at even the 20% significance level.
Many econometrics packages have a built-in feature for testing multiple exclusion
restrictions. These packages have several advantages over calculating the statistics by
hand: we will less likely make a mistake, p-values are computed automatically, and the
problem of missing data, as in Example 4.9, is handled without any additional work on
our part.
The F Statistic for Overall Significance of a Regression
A special set of exclusion restrictions is routinely tested by most regression packages.
These restrictions have the same interpretation, regardless of the model. In the model with
k independent variables, we can write the null hypothesis as
H
0
: x
1
, x
2
,…,x
k
do not help to explain y.
This null hypothesis is, in a way, very pessimistic. It states that none of the explanatory
variables has an effect on y. Stated in terms of the parameters, the null is that all slope
parameters are zero:
H
0
:
1
2
…
k
0,
(4.44)
and the alternative is that at least one of the
j
is different from zero. Another useful way
of stating the null is that H
0
:E(yx
1
, x
2
,…,x
k
) E(y), so that knowing the values of x
1
,
x
2
,…,x
k
does not affect the expected value of y.
The data in ATTEND.RAW were used to estimate the two equations
atndrte (47.13) (13.37) priGPA
atn
ˆ
drte (2.87) (1.09) priGPA
n 680, R
2
.183,
and
atndrte (75.70) (17.26) priGPA 1.72 ACT
atn
ˆ
drte (3.88) (1.08) priGPA 1(?) ACT,
n 680, R
2
.291,
where, as always, standard errors are in parentheses; the standard
error for ACT is missing in the second equation. What is the t sta-
tistic for the coefficient on ACT? (Hint: First compute the F statis-
tic for significance of ACT.)
160 Part 1 Regression Analysis with Cross-Sectional Data
QUESTION 4.5