next up previous
Next: Bibliography

Nikolaos, N M Missirlis
Preconditioning augmented linear systems

Department of Informatics and Telecommunications
University of Athens
Panepistimiopolis
15784
Ilisia
Athens
Greece
nmis@di.uoa.gr
Maria Louka

Let $ A \in \mathbb{R}^{m\times m}$ be a symmetric positive definite matrix and $ B\in\mathbb{R}^{m\times n}$ be a matrix of full column rank, where $ m\geq n$ . Then, the augmented linear system is of the form

$\displaystyle \ca Au=b$ (1)

where

$\displaystyle \ca=\left( \begin{array}{cc} A & B\\ -B^{T} & 0 \end{array} \righ...
...{array} \right),\;\; b= \left( \begin{array}{cc} b_1\\ -b_2 \end{array} \right)$ (2)

with $ B^T$ be the transpose of the matrix B. When A and B are large and sparse, iterative methods for solving ([*]) are effective and more attractive than direct methods, because of storage requirements and preservation of sparsity.
The difficulty in applying iterative methods such as the successive overrelaxation (SOR) method [10] to system ([*]) is the singularity of the block diagonal part of the coefficient matrix. Some methods have been developed to overcome this difficulty such as the Uzawa and the Preconditioned Uzawa method [1]. In 2001, Golub et al. [4] generalized the Uzawa and the Preconditioned Uzawa method for the augmented system ([*]) by introducing an additional acceleration parameter and produced the SOR-like method. Other known methods, competitive and specially attractive for solving ([*]) are the preconditioned MINRES and GMRES methods [3]. In case where the matrix A is symmetric and positive definite, the Preconditioned Conjugate Gradient (PCG) method [5] is the most suitable. In 2003, Li, Evans and Zhang [6] proved that the PCG method is at least as accurate as the SOR-like method. In 2005, Bai et al. [2] proposed the Generalized SOR (GSOR) method and proved that it has the same rate of convergence but less complexity than the PCG method. Furthermore, the Generalized Modified Extrapolated SOR (GMESOR) method, a generalization of the GSOR method, was also proposed for further study. The latter is a generalization of the GSOR method as it uses one additional parameter.
In this paper, we study the Generalized Modified Extrapolated SOR (GMESOR) method and the Generalized Modified Preconditioned Simultaneous Displacement (GMPSD) method. The latter is derived by using as preconditioning matrix the product of a lower and upper triangular part of $ \ca$ [7], [8]. In an attempt to generalize our theory, we introduce an additional parameter $ a$ in the structure of the matrix $ \ca$ , with the hope of increasing the rate of convergence for the aforementioned methods. Our starting point, for studying these methods, is the derivation of functional relations which relate the eigenvalues of their iteration matrices with those of the key matrix $ J=Q^{-1}B^TA^{-1}B$ . Assuming that the matrix Q is symmetric positive or negative definite matrix, the eigenvalues of the matrix $ J$ are real and either positive if the matrix Q is positive definite or negative if the matrix Q is negative definite. Under these assumptions we find sufficient conditions for the convergence of the aforementioned methods and determine the optimum values of their parameters. For the determination of the optimum values of the parameters in GMESOR and GMPSD methods we followed a similar approach to Varga [9] p. 110, for the optimum $ \omega$ of SOR.




next up previous
Next: Bibliography
root 2012-02-20