A parallel solution with the accumulated matrices
and
is only feasible if they posses the block structure
required in Sec. 4.3.1
(again we have to fulfill the mesh conditions).
Then the submatrices
from
are tridiagonal matrices for edges with linear FE-test functions.
There local inverting can be done easily by a Thompson/Progonka-algorithm.
Matrix
is usually a diagonal matrix.
Between the blocks ''
'' and inside the subdomains
(block ''
'') a Gauß-Seidel iteration will be performed.
Due to the partial use of direct solvers, the convergence rate is
slightly better as in a Gauß-Seidel iteration.
The accumulation of cross points (''
'') and interface data (''
'')
is usually performed separately so that we need
exactly the same amount of communication as in the
-Jacobi iteration (Alg. 5.6).
On the other hand, the accumulation of matrix
(tridiagonal) in the setup of the iteration is three times more expensive
then the normal accumulation of the main diagonal.
The accumulated matrices
und
have to be stored
additionally because the distributed matrices
and
are still needed in the matrix-times-vector
operation.
In case of using the iteration as a smoother with a fixed
number of iterations the calculation of the inner product is
no longer necessary. This saves the ALL/SMALL>_REDUCE operation
in the parallel code
and vectors
and
can be stored in one place.
Exercise 13:
Find a modification of Alg. 5.11 without the
double storing of
and
!
Next: 5.5 Incomplete factorizations
Up: 5.4.4 Parallel algorithms
Previous: 5.4.4.2 Variant 2 : Gauß-Seidel-Jacobi
  Contents
Gundolf Haase
2000-03-20