12.4 Knowledge-Based Image Analysis 343
of a single source of multispectral data, as was the case in Chapter 3, consider now
that the data to be analysed consists of three parts: a Landsat multispectral image, a
radar image of the same region and a soil map of that region. From what has been
said above, standard methods of quantitative analysis cannot cope with trying to
draw inferences about the cover types in the region since they will not function well
on two numerical sources of quite different characteristics (multispectral and radar
data) and also since they cannot handle non-numerical data at all.
In contrast, consider how a skilled photointerpreter might approach the problem
of analysing this multiple source of spatial data. Certainly he or she would not wish
to work at the individual pixel level, as discussed in Sect. 3.1, but would more likely
concentrate on regions. Suppose a particular region was observed to have a predomi-
nantly pink tone on a standard false colour composite print of the multispectral data,
leading the photointerpreter to infer initially that the region is vegetated; whether it is
a grassland, crop or forest region may not yet be clear. However the photointerpreter
could then refer to the radar imagery. If its tone is dark, then the region would be
thought to be almost smooth at the radar wavelength being used. Combining this
evidence with that from the multispectral source, the photointerpreter is then led to
consider the region as being either grassland or a small crop. He or she might then
resolve this conflict by referring to the soil map of the region. Noting that the soil
type is not that normally associated with agriculture, the photointerpreter would then
conclude that the region is same form of natural grassland.
In practice the process of course may not be so straightforward, and the photoin-
terpreter may need to refer backwards and forwards over the data sets in order to
finalise an interpretation, especially if the multispectral and radar tones were not uni-
form for the region. For example, some spots on the radar imagery may be bright. The
photointerpreter would probably regard these as indicating shrubs or trees, consistent
with the overall region being labelled as natural grassland. The photointerpreter will
also account for differences in data quality, placing most reliance on data that is seen
to be most accurate or most relevant to a particular exercise, and weighting down
unreliable or marginally relevant data.
The question we need to ask at this stage is how the photointerpreter is able to
make these inferences so easily. Even apart from spatial processing, as discussed
in Table 3.1 (where the photointerpreter would also use spatial clues such as shape
and texture), the key to the photointerpreter’s success lies in his or her knowledge
– knowledge about spectral reflectance characteristics, knowledge of radar response
and also of how to combine the information from two or more sources (for example,
pink multispectral appearance and dark radar tone indicates a low level vegetation
type). We are led therefore to consider whether the knowledge possessed by an expert
such as a skilled photointerpreter can be given to and used by a machine and so devise
a method for analysis that is able to handle the varieties of spatial data type available in
GIS-like systems. In other words can we emulate the photointerpreter’s approach? If
we can then we will have available an analytical procedure capable of handling mixed
data types, and also able to work repetitively, at the pixel level if necessary. With
respect to the latter point, it is important to recognise that photointerpreters generally