04.08.2013 Views

P. Schmoldt, PhD - MTNet - DIAS

P. Schmoldt, PhD - MTNet - DIAS

P. Schmoldt, PhD - MTNet - DIAS

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

6.3. Deriving subsurface structure using magnetotelluric data<br />

[Siripunvaraporn and Egbert, 2000; Siripunvaraporn, 2010]. The advantages of the code<br />

are that it converges in a small number of iterations, and that it searches for a minimum<br />

structure model [Constable et al., 1987; Siripunvaraporn and Egbert, 2000]. The latter<br />

means that all structures in a model are more reliable as they are required by the data and<br />

not unconstrained artefacts of the inversion [Siripunvaraporn, 2010]. The disadvantage<br />

of the code is its large computational cost, similar to the GN methods, i.e. inverting the<br />

M × M matrix τR −1<br />

mm + J T<br />

k R−1 ddJk <br />

and computing the N × M Jacobian.<br />

The issue of large computational cost can be mitigated by transforming the computational<br />

space from the model space into the data space, rearranging Equation 6.38 as<br />

m − m0 = RmmJ Tβ, (6.39)<br />

where β is a unknown coefficient vector of length N [Parker, 1994; Siripunvaraporn and<br />

Egbert, 2000; Siripunvaraporn et al., 2005a]. Inserting Equation 6.39 into Equation 6.32<br />

and differentiating the resulting function with respect to β yields the iterative sequence of<br />

approximate solutions<br />

mk+1 − m0 = RmmJ T<br />

k R−1/2<br />

dd<br />

<br />

−1/2 T<br />

τI + Rdd JkRmmJk R−1/2<br />

−1 dd R −1/2<br />

dd d ′ k, (6.40)<br />

where I is the identity matrix. Therefore, only a N × N matrix τI + R −1/2 T<br />

dd<br />

JkRmmJk R−1/2<br />

<br />

dd<br />

(instead of the M × M matrix in the model space) needs to be inverted, which significantly<br />

reduces CPU time and memory usage [Siripunvaraporn and Egbert, 2000; Siripunvaraporn<br />

et al., 2005a]. However, the algorithm still requires the costly computation of the<br />

Jacobian.<br />

In the data space Occam’s inversion and in the GN approach, inversion of the matrices<br />

(Eqs. 6.38 and 6.37, respectively) is usually computational the most expensive part of the<br />

process. In the quasi-Newton (QN) method the inverse matrix is therefore approximated<br />

through a recursive process, in which H −1 is replaced by a successively adapted matrix<br />

[Fletcher and Powell, 1963; Shanno, 1970]. As a result the main computation for each<br />

QN iteration is reduced to computing ∇T m and performing the line search, meaning that<br />

the memory requirement of the QN method is insignificant in comparison with the GN<br />

method [Siripunvaraporn, 2010]. Because of the QN method’s slow convergence rate<br />

[Haber, 2005] modifications are made to permit large datasets; e.g. approximating only<br />

the part of H that is related to the data misfit, rather than the full H [Haber, 2005]; or<br />

introducing additional regularisations [Avdeev and Avdeeva, 2009].<br />

Like the QN method, the Gauss-Newton with Conjugate Gradient (GN-CG) method<br />

is set up to reduce the computational load of inverting the large matrices in Equations<br />

6.37, 6.38, and 6.40. The method is referred to as model space GN-CG or data space<br />

GN-CG, respectively, dependent on whether it is set up to solve Equation 6.38 or 6.40.<br />

The advantage of the method is that instead of J only the product of J and J T with an<br />

arbitrary vector V is required, i.e. J Va and J T Vb. Vb and Vb can be computed by solving<br />

one forward problem [Mackie and Madden, 1993; Newman and Alumbaugh, 2000; Rodi<br />

125

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!