next up previous
Next: About this document ...

Victor Y. Pan
Novel Techniques for Linear Systems and Eigen-solving

Department of Mathematics and Computer Science
Lehman College
The City University of New York
Bronx NY 10468 USA
victor.pan@lehman.cuny.edu

Summary

Additive preconditioning facilitates the computation of a vector in the null space of a matrix as well as a basis for this space. This amounts to new effective algorithms for solving linear systems of equations. We further incorporate our technique into the inverse iteration for computing the eigenvectors and eigenspaces of a matrix, which are the null vectors and null spaces of the same matrix shifted by its eigenvalues. This facilitates every iteration step but does not slow down convergence, according to our analysis and extensive experiments. We elaborate upon this approach for simple and multiple eigenvalues as well as for clusters of eigenvalues.

1. Solving Linear Systems of Equations with Additive Preconditioning

Given an $ n \times n$ matrix $ A$ of a rank $ \rho<n$, suppose we seek its null vector or null basis, that is, a vector in the (right) null space $ RN(A)$ or a basis for this space. This is equivalent to solving the homogeneous linear system of equations $ AY=0$ and is readily extended to the solution of a nonhomogeneous linear system $ M{\bf y}={\bf b}$. Indeed we can just write $ A=(-{\bf b},M)$ and observe that $ {\bf y}$ is a null vector for the matrix $ A$ that has its first coordinate equal to one.

We can obtain the solution to these problems via computing the SVD, QRP or PLU factorizations, or the inverse of a nonsingular $ \rho \times \rho$ submatrix of matrices $ A$ or $ MAN$ for some nonsingular multipliers $ M$ and $ N$.

Our alternative approach employs addititive preprocessing of the input matrix $ A$. See detailed descriptions of this approach in the author's technical reports in the Computer Science Program of the Graduate Center of the City University of New York, 2005-2008 and in his paper in Computers and Mathematics with Applications (in press). Hereafter ``A-" and ``APP" abbreviate ``additive" and ``additive preprocessor", respectively.

Define two generators $ U$ and $ V$ of size $ n \times r$, suppose an APP $ UV^H$ has rank $ r=n-\rho$ equal to the nullity of the matrix $ A$, and let the A-modification $ C=A+UV^H$ have full rank. Then the columns of the null aggregate $ C^{-1}U$ form a null basis for the matrix $ A$, and so we call the matrix $ C^{-1}U$ a null matrix basis for the matrix $ A$.

According to our analysis and experiments, A-preprocessing of an $ n \times n$ ill conditioned input matrix $ A$ with a random well conditioned and properly scaled APPs $ UV^H$ of a rank $ r$ (normalized so that the ratio $ \vert\vert UV^H\vert\vert _2/\vert\vert A\vert\vert _2$ is neither large nor small) is expected to yield an A-modification $ C=A+UV^H$ with the condition number of the order of $ \sigma_{1}(A)/\sigma_{n-r}(A)$ where $ \sigma_j(A)$ denotes the $ j$th largest singular value of the matrix $ A$. If $ \sigma_{n-r}(A)\gg
\sigma_{n}(A)$, then our A-preprocessing is expected to be A-preconditioning, that is, expected to decrease the condition number substantially.

Furthermore, even very weak randomization is actually sufficient, allowing us to choose structured or sparse APPs. Since our techniques preserve matrix structure and improve conditioning, they enable effective application of the Conjugate Gradient algorithms.

2. Extension to Eigen-solving

The eigenspace of a matrix $ M$ associated with its eigenvalue $ \lambda$ is just the null space of the matrix $ A=A(\lambda)=\lambda I_n-M$, and so the above approach can be incorporated into the known eigen-solvers. We elaborate upon its incorporation into the inverse iteration. Our study can be readily extended to the shift-and-invert enhancements of the Lanczos, Arnoldi, Jacobi-Davidson, and other effective eigen-solvers.

Our analysis and experiments show that our modification does not affect the convergence rate, even though it improves conditioning of every iteration step. We elaborate upon this approach for simple and multiple eigenvalues as well as for clusters of eigenvalues and point out its various natural extensions. In spite of apparent similarity of the classical and our algorithms, our analysis and in particular our treatment of multiple and clustered eigenvalues and our proof of local quadratic convergence show the distinct nature of the power of the two approaches. This suggests concurrent application of both iterations (also performed concurrently for a number of distinct initial approximations to the eigenvalues) to improve global convergence.

We apply A-preconditioning to the inverse iteration in its both linear and multilinear settings. The latter variant is natural for approximating multiple and clustered eigenvalues as well as complex conjugate pairs of eigenvalues of a real matrix.




next up previous
Next: About this document ...
Marian 2008-02-26