
missing. Drawing with replacement from the data is equivalent to Monte Carlo
realizations from the empirical CDF. The statistic of interest is computed on all of
the replicate bootstrap data sets. The distribution of the bootstrap replicates of the
statistic is a measure of uncertainty of the statistic.
Drawing bootstrap replicates from the empirical CDF in this way is sometimes
termed nonparametric bootstrap.Inparametric bootstrap the data are first mod-
eled by a parametric CDF (e.g., a multivariate Gaussian), and then bootstrap data
replicates are drawn from the modeled CDF. Both simple bootstrap techniques
described above assume the data are independent and identically distributed. More
sophisticated bootstrap techniques exist that can account for data dependence.
Statistical classification
The goal in statistical classification problems is to predict the class of an unknown
sample based on observed attributes or features of the sample. For example, the
observed attributes could be P and S impedances, and the classes could be lithofacies,
such as sand and shale. The classes are sometimes also called states, outcomes, or
responses, while the observed features are called the predictors. Discussions concerning
many modern classification methods may be found in Fukunaga (1990), Duda et al.
(2000), Hastie et al. (2001), and Bishop (2006).
There are two general types of statistical classification: supervised classification,
which uses a training data set of samples for which both the attributes and classes
have been observed; and unsupervised learning, for which only the observed
attributes are included in the data. Supervised classification uses the training data
to devise a classification rule, which is then used to predict the classes for new data,
where the attributes are observed but the outcomes are unknown. Unsupervised
learning tries to cluster the data into groups that are statistically different from each
other based on the observed attributes.
A fundamental approach to the supervised classification problem is provided by
Bayesian decision theory. Let x denote the univariate or multivariate input attributes,
and let c
j
, j ¼1,..., N denote the N different states or classes. The Bayes formula
expresses the probability of a particular class given an observed x as
Pðc
j
jxÞ¼
Pðx; c
j
Þ
PðxÞ
¼
Pðx jc
j
ÞPðc
j
Þ
PðxÞ
where P(x, c
j
) denotes the joint probability of x and c
j
; P(x jc
j
) denotes the conditional
probability of x given c
j
; and P(c
j
) is the prior probability of a particular class. Finally,
P(x) is the marginal or unconditional pdf of the attribute values across all N states.
It can be written as
PðXÞ¼
X
N
j¼1
PðX jc
j
ÞPðc
j
Þ
17 1.3 Statistics and probability