
9.3
The Nonlinear Inverse Problem with Gaussian Data
147
9.2
Linearizing Parameterizations
One of the reasons for changing parameterizations is that it is
sometimes possible to transform an inverse problem into a form that
can be solved by a known method. The problems that most commonly
benefit from such transformations involve fitting exponential and
power functions to data. Consider a set of
(z,
d)
data pairs that are
thought to obey the model
di
=
m,
exp(m2zi). By making the transfor-
mation
m;
=
log(m,),
m;
=
m2,
and
dj=
log(di), we can write the
model as the linear equation
di
=
m{
+
m;z2,
which can be solved by
simple least squares techniques. To justify rigorously the application
of least squares techniques to this problem, we must assume that the
dj
are independent random variables with a Gaussian distribution of
uniform variance. The distribution of the data in their original pa-
rameterization must therefore be non-Gaussian.
For example,
if
the exponential decays with increasing
z
for all
m2
<
0,
then the process of taking a logarithm amplifies the scattering
of the near-zero points that occurs at large
z.
The assumption that the
dj
have uniform variance, therefore, implies that the data
d
were
measured with an accuracy that increases with
z
(Fig.
9.2).
This
assumption may well be inconsistent with the facts of the experiment.
Linearizing transformations must be used with some caution.
9.3
The Nonlinear Inverse Problem with
Gaussian Data
Linearizing transformations cannot be found for most inverse prob-
lems. We must consider other methods for directly solving the nonlin-
ear problem. We shall begin by considering the very general implicit
equation
f(d, m)
=
0
(where
f
is of length
p
5
M
+
N).
We simplify by
assuming that the data
d
and a priori model parameters
(m)
have
Gaussian distributions with covariance [cov
d]
and [cov
m],
respec-
tively. If we let
x
=
[d, mIT,
we can think of the a priori distribution of
the data and model as a cloud in the space
S(x)
centered about the
observed data and mean a priori model parameters, with a shape
determined by the covariance matrix [cov
x]
(Fig.
9.3).
This matrix
[cov
x]
contains [cov
d]
and [cov
m]
on diagonal blocks. In principle
the off-diagonal blocks could be made nonzero, indicating some cor-