It is important to extend the normal mode procedure to
variable sound speed profiles. The
strategy here is a new one, based on representing the solution in terms of
spanning functions from an iso-velocity solution that is in some average sense close
to the variable profile and making use of elements of Sturm-Louiville
theory. The following method was
developed after an unsuccessful attempt to include a variable sound speed
profile as a perturbation of the iso-velocity case via the usual perturbation
theory from elementary quantum mechanics.
The first order perturbation is doomed to failure because the lowest
modes are relatively small even compared to the smallest perturbation, which
makes it impossible to reconstruct the eigenfunctions. It is a well-known fact that the standard
perturbation method did not treat perturbed wave functions correctly in terms
of the expansion coefficients even though the theory often gave reasonable
results for perturbed eigenvalues. Any
attempt to add second order terms in a poorly converging first order theory
will never work, since, with the exception of forbidden first order processes
such as occur in quantum mechanics and result in Fermi’s “first golden rule”,
second order corrections are always refinements to the first order calculations
at best. The new method excludes any
truncation in space or any limiting of parameters in some asymptotic sense;
rather it retains the complete space throughout the calculation. Thus, in a
sense it is not a perturbation method, but a transformation method, which may
be viewed as a new type of matrix - a warping
matrix - transforming an iso-velocity space to a variable velocity space.
Superficially, the method resembles a Galerkin formulation which has its origin
in a completely different philosophy, namely in demanding that any subset of a
complete set used to expand an exact solution must be orthogonal to those
states not included in the expansion - the residuals. The Galerkin method is
one of the weighted residual methods and had its origin in polynomial fitting
methods. The Galerkin method is easily
subject to error due to the arbitrary nature of basis states and the difficulty
of dealing with nested boundary conditions.
The Generalized
Eigenvalue Formulation
At this point we examine two ocean
models with the same bottom properties. One model has an iso-velocity water
column while the other has a variable velocity profile. The bottom properties
are the same and are as described above. The solution for the depth-separated wave
equation for the iso-velocity case in the water column has been developed above
and is expressed now as:
. (1)
The equation that
governs the solution for the mode Um
in the water column for arbitrary sound speed profile is:
. (2)
The terms in Eq. (2)
can be rearranged in such a way that the left-hand side of the equation
contains the operator for the iso-velocity solution:
, (3)
where , and
Recall that for Sturm-Louiville problems
we have the orthogonality condition:
(4)
for any two function and . We can assume
completeness of the space, namely that the eigenfunctions of the iso-velocity space
span the variable velocity space. Then, in particular, within the water column
we can decompose the solution for the variable velocity space as:
. (5)
It is to be understood that for solutions
other than in the water column, the functional form is the same, so that for
the layers only the normalization and the values of the eigenvalues need to be
adjusted for the ’s defined in terms of the ’s. Substituting (5) in the differential Eq. (3), we obtain:
Then using Eq. (1), we arrive at:
Multiplying both sides of Eq. (6) by and integrating over
only the water column, we obtain:
and, which can be calculated as soon as we have solved the iso-velocity
problem, so at this stage they are known.
The set of equations (7) can be written in matrix form as follows:
Keep
in mind that matrices Q and B are both symmetric. Here
, 0, and K
are diagonal matrices of the respective eigenvalues, where the diagonal term
for row m is the eigenvalue etc.
Matrix Eq. (8) relates the variable velocity
eigenvalues with the iso-velocity
eigenvalue and the vertical
iso-velocity eigenvalues . Here k0 is the wave number for the
iso-velocity (spanning) space, and and are the iso-velocity modal eigenangles and the variable
velocity modal eigenangles, respectively. Although the modal angle for the
iso-velocity case corresponds to the angle of the mode phase intensity, the
angle for the variable velocity is not
physical in the same sense, due to refraction - the paths of intensity bend
contrary to what a fixed angle would imply.
Further, it is somewhat arbitrary since it depends on the iso-velocity
wave number. However, the overall
eigenvalue should be and is an
invariant with respect to the iso-velocity wave number proved the number is within reasonable bounds. We discuss those bounds below but it is
suitable to choose the reference wave number corresponding to a sound speed
bracketed by the values in the water column.
However, the angles do have meaning in the time domain in connection
with energy fronts since they are averaged over a large number of frequencies.
Eq.
(8) is actually a generalized eigenvalue problem which may be reduced to the
following form by using the fact that :
. (9)
Here
I
is the identity (unit) matrix and k0
is a scalar. Eq. (9) may be solved by
post multiplying both sides by the inverse of B to arrive at a matrix which is not generally symmetric. We then may make use of the QR-algorithm for
general non-symmetric or non-Hermitian matrices, which gives us both the
eigenvalues and the eigenfunctions at somewhat more computational cost than if
we had a symmetric matrix. Because we
are essentially dealing with a Sturm-Louiville problem (or a pseudo- Sturm-Louiville problem if we require the
continuum), we know that we can transform the problem to one of symmetric form
which is computationally desirable and mathematically more aesthetic.
Numerical Algorithm for the Solution of
the Generalized Eigenvalue Problem
The first observation which we can make on the
approach to solving Eq. (9) by an optimal numerical procedure is that the
matrix B is close to the unit matrix
since its elements are the inner products of the elements of the major part of
the normalized vertical wave functions. Only that part of the wave function
that penetrates the sediment is left out, and that component is usually
small. Indeed, for a strongly ducted
problem with the duct away from the bottom of the water column there exists a
subset of B that is essentially a
unit matrix. However, B is not only symmetric but positive definite since any part of the norm
will be positive. Thus we may employ the most efficient method for
LU-decomposition due to Cholesky to decompose B into lower and upper triangular matrices, namely B=L,
where is the transpose of
the lower triangular matrix L and
is, thus, upper triangular. The diagonal elements under the Cholesky
scheme are the square roots of the
elements of D, the pivot matrix, in
standard LDU-decomposition. The scheme
requires that both L and its
transpose have the same diagonal elements. The next discussion illustrates how simply and efficiently
Cholesky decomposition can work for our eigenvalue problem. Here it should be
remembered that B is both symmetric
and positive definite. We have that:
.
This decomposition involves about N3/6
floating point operations (FPO), where N is the size of the matrix, and is probably the fastest known
factorization method for a dense matrix. One FPO consists of one floating point
multiplication and one addition (operational count).
Following
decomposition of B, we post multiply both
sides of Eq. (9) by and use
the fact that to obtain:
, (10)
which now involves a symmetric matrix U= and a
new matrix X=AL that is related to
the variable velocity case modal functions of Eq. (5). We may efficiently
obtain U by solving for the two
expressions:
Q=LY (11)
and
. (12)
Since both Q and U are symmetric,
we need only find the lower triangular elements of Y from Eq. (11) and the lower triangular elements of U from Eq. (12) to obtain all of the
elements of U. Each of these involve
N(N+1)/2 FPO’s and are extremely simple to program and very stable.
It
remains to determine . To determine this quantity, we first find:
. (13)
We note that C is lower triangular since is lower triangular,
which means that . Then we determine:
(14)
This result follows from the fact that and C are lower triangular, which means that and ; thus k=i=j. Now we need only to determine the diagonal elements of the inverse of , which may be carried out in a manner similar to the above proof and leads to:
.
Thus we have arrived at:
. (15)
This leads to the following equation
which replaces Eq. (10):
. (16)
Eq.
(16) can be viewed as a generalized eigenvalue problem. In the formulation of
our variable velocity profile problem, we may refer to the matrix W as a warping matrix that transforms the iso-velocity eigenfunctions and
eigenvalues into the variable velocity space. We may now take advantage of the
fact that W is symmetric and
transform W to a tridiagonal
matrix. This may be performed by either
a Givens transformation or a sequence of Householder transformations. A Givens transformation is related to a
rotation in two dimensions while the more efficient Householder transformation
is related to “inflection” in two dimensions.
Both have the effect of leading to a tridiagonal form but the
Householder method is about twice as efficient for dense matrices. There are many other advantages to the
Householder transformation. First, the Householder matrix is symmetric,
unitary, Hermitian and “idempotent” (M*M=M). The Householder transformation for a general symmetric matrix
requires only 2N3/3 operations for large values of N. It performs
the same procedure as the Gram-Schmidt orthogonalization method but is also
much more stable than that procedure, and even more stable than the modified
Gram-Schmidt method which sometimes fails to maintain true orthogonality. The
product of Householder transformations that leads to a tridiagonal form is a
similarity transformation, and thus eigenvalues of the tridiagonal matrix are the
same as the original matrix. We then
employ the QL-algorithm, which for a tridiagonal symmetric matrix requires on
the order of or less than 2N iterations, to obtain the eigenvalues for the
variable velocity space. Typically the average number of iterations is
fewer than 2N. In the process if we save the sequence of
unitary transformations that transforms the matrix W to tridiagonal form, which is easy to do using the Householder
method, it is possible to construct the original eigenvectors in a straightforward
manner, namely X=AL, which only requires an additional
2N3/3 FPO’s. The procedure
is also very stable. Since W may be viewed as being obtained from a similarity transformation of
, it has the same eigenvalues as the
original problem, namely those associated with A.
Once we obtain X, we may easily obtain A from
X=AL with N(N+1) FPO’s. The
matrix A has rows that correspond to
the expansion coefficients we require for expression (3.31). The complete process is on the order of (3N3/2)+N(1.5N+2) FPO’s.
The method presented here should not be
confused with the Galerkin method for finding eigenvalues and eigenfunctions
for variable sound speed profiles
published earlier in the literature (for example, see [14]). This formulation
is based on a completely different theoretical strategy. Appendix 3B contains some comments on the
comparison of the two methods, because this question was raised by the audience
several times after we had given talks on our technique.
Building
the matrix elements of Q may
be the time consuming operation, so
special attention must be paid to performing this operation in an efficient manner. We avoid numerical quadrature as that would
be too costly, particularly as one increases the frequency. Therefore an effort has been made to obtain
a closed form representation of the matrix elements. We may do this in one of two ways. Recall that
. (17)
We may approximate over any suitably small segment, so we may always express Eq. (17) in that form. This is the so-called pseudo-linear sound speed variation. This allows us to obtain the elements of the matrix Q in closed form.
A more elegant and precise way that generally leads to a faster calculation is the following. Any analytical function can be decomposed in a truncated series of Legendre polynomials:
, (18)
where the ’s are Legendre polynomials of order j. Gauss-Legendre quadrature requires about twice the number of terms as the highest order of the Legendre polynomial (almost never more than about 10) and are thus quite simple to do. We will not need many expansion terms since q(z) does not have a complicated structure. Typically we may need from 2 to 5 terms. This expansion also smoothes the profile, which is desirable. Regardless of what one measures, nature is never so abrupt.
We now make the transformation of the remainder of the integrand for involving the eigenfunctions of the iso-velocity case:
. (19)
We may rewrite the cosine functions
taking into account a change of variables for :
.
We note that :
and (20)
where is a spherical Bessel
function of order 2k. Thus, we can express the elements of matrix , in a convenient analytical form:
(21)
where , . Here the number of terms reduces by a factor of 2 in
comparison with the total number of Legendre expansion terms (almost never more
than 5) because only even functions give non-zero integrals. The diagonal terms
are a special case and much easier to construct. It is noted that Q is symmetric in the indices so we
only need to construct half of the
off-diagonal elements. We can also save much computer time by noting that [51]:
, (22)
where and are 3-j coefficients from quantum mechanics.
We need only make 2L rather than L2 sets of spherical Bessel
functions, where L is the number of perfectly trapped modes for the iso-velocity
case. Whenever possible, we may also make a similar transformation for the sine
and cosine functions and store them for multiple use. Thus, rather than L2 , we need to construct only 2L
sine and cosine functions. Since these
functions are fast and easy to make, we have a fast closed form solution of the
integrals that define the matrix Q.
For high frequency the ’s fall off as they move off the matrix diagonal at a rate
such that we do not need to calculate all of the values, and we may set values
to zero and invoke bandedness at no loss of accuracy in suitable cases. Clearly,
the sine of a large argument varies rapidly, so when indices of q differ much, the integrals approach small numbers relative to the diagonal
terms. Indeed, bandedness is more a characteristic of higher frequencies than of lower.
These
comments end the part describing the solution of the depth-separated
wave-equation for an arbitrary sound speed profile. It is important to note that one of the advantages of the normal mode approach is that vertical
modal functions do not depend on the receiver-source geometry. Once calculated, they can be stored and used
many times if the ocean environment investigated is the same.