
Machine Learning
202
(15)
We use the AUC as the primary measure, as it allows us to gauge the trade off between the
FP and TP rates at different cut-off points. Although the error rate W
Err
(or accuracy) has
been widely used in comparing classifiers’ performance, it has been criticized as it highly
depends on the probability of the threshold chosen to approximate the positive classes. Here
we note that we assign new classes to the positive class if the probability of the class is
greater than or equal to 0.5 (threshold=0.5). In addition, in (Huang & Ling, 2005) the authors
prove theoretically and empirically that AUC is more accurate than accuracy to evaluate
classi_ers' performance. Moreover, although classifiers might have different error rates,
these rates may not be statistically significantly different. Therefore, we use the Wilcoxon
signed-ranks test (Wilcoxon, 1945) to compare the error rates of classifiers and find whether
the di_erences among these accuracies is significant or not (Demšar, 2006).
4.3 Experimental studies
We optimize the classifiers’ performance by testing them using different input parameters.
In order to find the maximum AUC, we test the classifiers using the complete dataset
applying different input parameters. Also, we apply 10-fold-cross-validation and average the
estimates of all 10 folds (sub-samples) to evaluate the average error rate for each of the
classifiers, using the 70 features and 6561 emails. We do not perform any preliminary
variable selection since most classifiers discussed here can perform automatic variable
selection. To be fair, we use L1-SVM and penalized LR, where variable selection is
performed automatically.
We test NNet using different numbers of units in the hidden layer (i.e. different sizes (s))
ranging from 5 to 35. Further, we apply different weight decays (w) on the interconnections,
ranging from 0.1 to 2.5. We find that a NNet with s = 35 and w = 0.7 achieves the maximum
AUC of 98.80%.
RF is optimized by choosing the number of trees used. Specifically, the number of trees we
consider in this experiment is between 30 and 500. When using 50 trees on our dataset, RF
achieves the maximum AUC of 95.48%.
We use the L1-SVM C-Classification machine with radial basis function (RBF) kernels. L1-
SVM can automatically select input variables by suppressing parameters of irrelevant
variables to zero. To achieve the maximum AUC over different parameter values, we
consider cost of constraints violation values (i.e. the “c” constant of the regularization term
in the Lagrange formulation) between 1 and 16, and values of the γ parameter in the kernels
between 1 ×10
-8
and 2. We find that γ= 0.1 and c = 12 achieve the maximum AUC of 97.18%.
In LR we use penalized LR and apply different values of the lambda regularization
parameter under the L2 norm, ranging from 1 × 10
-8
to 0.01. In our dataset
λ
= 1 × 10
-4
achieves the maximum AUC of 54.45%.
We use two BART models; the first is the original model and as usual, we refer to this as
“BART”. The second model is the one we modify so as to be applicable to classification,
referred to as “CBART”. We test both models using different numbers of trees ranging from
30 to 300. Also, we apply different power parameters for the tree prior, to specify the depth
of the tree, ranging from 0.1 to 2.5. We find that BART with 300 trees and power = 2.5
achieves the maximum AUC of 97.31%. However, CBART achieves the maximum AUC of
99.19% when using 100 trees and power = 1.