MODELS WITH BOTH CATEGORICAL AND CONTINUOUS PREDICTORS 141
and marketability. The same comment can be made regarding the other coefficient
reductions.
Model 3 in Table 4.5 is the same as model 2 except that effect coding is used to
represent colleges. In comparison to model 2 in Table 4.3, we see that the departures
of each college’s mean salary from the grand mean are smaller than was the case
without controlling for covariates. Moreover, whereas three of these departures were
significant before, only two are significant now. To further foreground the closing of
the salary gaps across colleges, the last column of Table 4.5 shows the adjusted mean
salaries for each college after accounting for the four key covariates. Because the
covariates are centered and their means are therefore all zero, calculating adjusted
means is quite straightforward. One just ignores the coefficients for the covariates in
model 2 in Table 4.5 and uses the intercept and dummy coefficients to calculate the
means. For example, the adjusted mean for “arts and sciences” is just the intercept
in model 2—47950. The adjusted mean for “firelands” is 47950 6098.382
41851.618, and so on. It is evident that compared to the unadjusted means for each
college in the last column of Table 4.1, the adjusted means exhibit less variability.
Mean Contrasts with an Adjusted Alpha Level. Recall that with five categories of
the variable college, there are 10 possible mean salary contrasts between pairs of col-
leges that can be tested. Up until now, I have been conducting these tests without
controlling for the increased risk of type I error—or capitalization on chance—that
accrues to making multiple tests. There are several procedures that accomplish this
control; here I discuss one, the Bonferroni comparison procedure. The Bonferroni
technique is advantageous because of its great generality. It is not only limited to
tests of mean contrasts. It can be used to adjust for capitalization on chance when-
ever multiple tests of hypothesis are conducted, regardless of whether or not they are
the same type of test. The rationale for the procedure is quite simple. Suppose that I
were making 10 tests and I wanted my overall chance of making at least one type I
error to be .05 for the collection of tests. That is, I want the probability of rejecting
at least one null hypothesis that is, in fact, true, to be no more than .05 over all tests.
If I make each test at an α level of .05, the probability of making a type I error on
each test is .05. This means that the probability of not making a type I error on any
given test is .95, and the probability of not making any type I errors across all 10
tests is therefore (.95)
10
.599. This implies that the probability of making at least
one type I error in all these tests is 1 .599 .401. In other words, we have about a
40% chance of declaring one H
0
to be false when it is not. The Bonferroni solution
in this case is to conduct each test at an α level of .05/10 .005. This way, the prob-
ability of not making a type I error on any given test is .995, and the probability of
making at least one type I error across all 10 tests is 1 (.995)
10
.049. In general,
if one is making K tests and one wants the probability of making at least one type I
error to be held at α across all tests, the Bonferroni procedure calls for each test to
be made at an α level of α α/K.
Although the Bonferroni procedure has the advantages of simplicity and flexi-
bility, it tends to be somewhat low in power. Holland and Copenhaver (1988) discuss
several modifications of the Bonferroni procedure that result in enhanced power to