Structured preconditioners for nonsingular matrices of block two-by-two structures

By Zhong-Zhi Bai

Abstract

For the large sparse block two-by-two real nonsingular matrices, we establish a general framework of practical and efficient structured preconditioners through matrix transformation and matrix approximations. For the specific versions such as modified block Jacobi-type, modified block Gauss-Seidel-type, and modified block unsymmetric (symmetric) Gauss-Seidel-type preconditioners, we precisely describe their concrete expressions and deliberately analyze eigenvalue distributions and positive definiteness of the preconditioned matrices. Also, we show that when these structured preconditioners are employed to precondition the Krylov subspace methods such as GMRES and restarted GMRES, fast and effective iteration solvers can be obtained for the large sparse systems of linear equations with block two-by-two coefficient matrices. In particular, these structured preconditioners can lead to efficient and high-quality preconditioning matrices for some typical matrices from the real-world applications.

1. Introduction

Let represent the real -dimensional vector space, and the real matrix space. Consider an iterative solution of the large sparse system of linear equations

In this paper, we will study algorithmic constructions and theoretical properties of practical and efficient structured preconditioners to the matrix which is of the block two-by-two structure

where nonsingular, , and , with , such that is nonsingular. Evidently, when the matrix block is nonsingular, the matrix is nonsingular if and only if its Schur complement is nonsingular.

Linear systems of the form Equation 1.1Equation 1.2 arise in a variety of scientific and engineering applications, including computational fluid dynamics Reference 21Reference 23Reference 26, mixed finite element approximation of elliptic partial differential equations Reference 16Reference 38, optimization Reference 25Reference 30Reference 34, optimal control Reference 13, weighted and equality constrained least squares estimation Reference 14, stationary semiconductor device Reference 36Reference 42Reference 43, structural analysis Reference 44, electrical networks Reference 44, inversion of geophysical data Reference 31, and so on.

As we have known, preconditioned Krylov subspace methods Reference 40 are efficient iterative solvers for the system of linear equations Equation 1.1Equation 1.2, and effective and high-quality preconditioners play a crucial role to guarantee their fast convergence and economical costs. A number of structured preconditioners have been studied in the literature for some special cases of the block two-by-two matrix in Equation 1.2. Besides specialized incomplete factorization preconditioners Reference 17Reference 18 we mention, among others, algebraic multilevel iteration preconditioners Reference 2Reference 3Reference 4Reference 5Reference 12, block and approximate Schur complement preconditioners Reference 21Reference 23, splitting iteration preconditioners Reference 15Reference 19Reference 22Reference 28Reference 29Reference 39Reference 45, block definite and indefinite preconditioners Reference 24Reference 34Reference 38Reference 10, and block triangular preconditioners Reference 35Reference 37Reference 10. Theoretical analyses and experimental results have shown that these preconditioners may lead to nicely clustered eigenvalue distributions of the preconditioned matrices and, hence, result in fast convergence of the preconditioned Krylov subspace iteration methods for solving the large sparse system of linear equations Equation 1.1Equation 1.2. However, exact inversions of the matrix block or , as well as the Schur complement , are demanded for most of these preconditioners, which makes them less practical and effective in actual applications.

In this paper, by sufficiently utilizing the matrix structure and property, we first establish a general framework of a class of practical and efficient structured preconditioners to the matrix in Equation 1.2 through matrix transformation and several steps of matrix approximations; these preconditioners can avoid the exact inversions of the matrix blocks and , as well as the Schur complement , and cover the known preconditioners mentioned previously as special cases. Then, with this framework we further present a family of practical and efficient preconditioners by technically combining it with the modified block relaxation iterations Reference 6Reference 7, which includes the modified block Jacobi-type, the modified block Gauss-Seidel-type and the modified block unsymmetric (symmetric) Gauss-Seidel-type preconditioners as typical examples. Moreover, we particularly discuss the eigenvalue distributions and the positive definiteness of the preconditioned matrices with respect to the modified block Jacobi-type, the modified block Gauss-Seidel-type, and the modified block unsymmetric (symmetric) Gauss-Seidel-type preconditioners to the block two-by-two matrix , and deliberately address the applications of these preconditioners to three classes of real-world matrices, i.e., the symmetric positive definite matrix, the saddle point matrix and the Hamiltonian matrix. Besides, we show that when these structured preconditioners are employed to precondition the Krylov subspace methods such as GMRES or restarted GMRES, fast and effective iteration solvers can be obtained for the large sparse system of linear equations Equation 1.1Equation 1.2.

The organization of this paper is as follows. After establishing the general framework of the structured preconditioners in Section 2, we present the modified block splitting iteration preconditioners and study the eigenvalue distributions and the positive definiteness of the corresponding preconditioned matrices in Section 3; connections of these preconditioners to Krylov subspace iteration methods are also briefly discussed in this section. Specifications of these preconditioners to three classes of real-world matrices are investigated in Section 4. Finally, in Section 5, we use a brief conclusion and several remarks to end the paper.

2. General framework of the structured preconditioners

The construction of our structured preconditioners basically includes the following three steps: Firstly, seek two nonsingular block two-by-two matrices such that and are easily invertible and holds for a block two-by-two matrix of certain good properties; secondly, approximate the matrix by another block two-by-two matrix by dropping some higher-order small block quantities; thirdly, approximate the matrix further by another block two-by-two matrix that is also easily invertible. Then, the resulting preconditioners are of the form . See Reference 9Reference 11.

Let and be nonsingular matrices such that

or equivalently,

where is a matrix approximating the identity matrix , and is a matrix approximating the identity matrix when it is positive definite and approximating when it is negative definite. For simplicity, in the sequel we will abbreviate the identity matrices and as , with their dimensions being inferred from the context.

Evidently, , and , can be considered as split preconditioners to the matrix blocks and , respectively, whose preconditioning properties can be measured by the approximation degrees of the matrices and to the identity matrix . There are many possible choices of the matrices , and , . For example, they may be the incomplete lower-upper triangular factors Reference 2Reference 40, the incomplete orthogonal triangular factors Reference 8, the approximate inverse preconditioners Reference 40, the splitting iteration matrices Reference 2Reference 6Reference 7Reference 27, the multigrid or the algebraic multilevel approximations Reference 2Reference 3Reference 4Reference 5Reference 12, or even technical combinations of the above-mentioned matrices, to the matrix blocks and , respectively.

In particular, when is singular, besides the possible choices mentioned above, we may choose and according to the following cases:

(i)

If is a symmetric positive semidefinite matrix, we may let . Hence, is also symmetric positive semidefinite.

(ii)

If is a symmetric negative semidefinite matrix, we may let and (or and ). Hence, is symmetric positive semidefinite. Or we may let . Hence, is also symmetric negative semidefinite.

(iii)

If is a general singular matrix, we may let . Hence, is also singular.

To construct a high-quality structured preconditioner to the block two-by-two matrix , we introduce matrices

and

where denotes the zero matrix. Then from Equation 2.2 we have

where

Furthermore, we can find a unit lower triangular matrix and a unit upper triangular matrix of block two-by-two structures such that is block-diagonally dominant as far as possible and may also possess some other desired good properties.

In fact, if we let

then by concrete computations we obtain

with

and

with

and

We can now choose the matrices and such that either of the following two principles is satisfied as far as possible:

()

the matrix is block-diagonally dominant and symmetric;

()

the matrix is block-diagonally dominant and skew-symmetric.

This is because if the matrix satisfies either of the principles () and (), we can easily construct a good approximation to it, and hence, obtain a high-quality preconditioner to the original matrix .

According to both () and (), we can take and such that

Recalling that , we can let

Thus, for both cases, it follows from Equation 2.4 and Equation 2.5 that the matrices and have the following expressions:

Therefore, for these choices of the matrices and , we have

Because the nonsingularity of the matrix implies that the matrix and its Schur complement are nonsingular, and

and the Schur complement of is

we immediately know that when

both matrices and are nonsingular.

Now, if we let be a nonsingular “replacement” of the matrix , or in other words, a “replacement” to the matrix , then the matrix

is a natural preconditioner to the original matrix , and under the condition Equation 2.9 this preconditioner is well defined.

Note that here we use the term “replacement” rather than “approximation”. This is because sometimes we may choose the matrix being not an approximation to in the usual sense so that the obtained preconditioner and the preconditioned matrix can possess some desired properties such as positive definiteness and, hence, a specified Krylov subspace iteration method may exploit its efficiency sufficiently.

If is used as a left preconditioner to , then

with

Therefore, the preconditioning property of to is determined by the properties of the matrices and . If is used as a right preconditioner to , then

with

Therefore, the preconditioning property of to is determined by the properties of the matrices and . In general, if the matrix admits a split form

then Equation 2.10 straightforwardly leads to a split preconditioner

to the original matrix . Because

we see that the preconditioning property of to is determined by the property of the matrix .

Evidently, the matrices , and are similar, and hence, they have exactly the same spectrum. However, the eigenvectors of these kinds of preconditioned matrices are usually quite different, which may lead to different performance results of the corresponding preconditioned Krylov subspace iteration methods.

In actual applications, when the matrix defined in Equation 2.10 is employed as a preconditioner to some Krylov subspace iteration method for solving the block two-by-two system of linear equations Equation 1.1, we need to solve a generalized residual equation of the form

at each iteration step, where is the current residual vector. By making use of the two-by-two block structure of , we can obtain the following practical procedure for computing the generalized residual vector .

Procedure for computing the generalized residual vector.

Let and , with and .

1.

Solve and to get and , and let .

2.

Solve to get , with .

3.

Solve and to get and .

When the approximation matrix to the matrix is specified, a concrete procedure for computing the generalized residual vector defined by Equation 2.18 can be straightforwardly obtained from this procedure.

Usually, the matrix may involve information about the matrices , , and . Therefore, to solve the linear system we may need to compute the vectors

These vectors can be economically computed by the following formulas:

1.

Solve .

2.

Solve , .

3.

Solve .

4.

Solve , .

3. Several practical structured preconditioners

In this section, we will construct three classes of structured approximations to the block two-by-two matrix , or in other words, to the block two-by-two matrix in Equation 2.7, by making use of the modified block Jacobi, the modified block Gauss-Seidel and the modified block unsymmetric Gauss-Seidel splittings of . See Reference 6Reference 7 for details. Therefore, three types of structured preconditioners to the original block two-by-two matrix , called the modified block Jacobi-type (MBJ-type) preconditioner, the modified block Gauss-Seidel-type (MBGS-type) preconditioner and the modified block unsymmetric Gauss-Seidel-type (MBUGS-type) preconditioner, can be obtained, correspondingly.

To analyze the spectral property of the preconditioned matrices with respect to the above-mentioned preconditioners, we need the following two basic facts.

Lemma 3.1.

Let and be unit lower and upper triangular matrices of the block two-by-two forms

where and . Let

be a monotone increasing function with respect to in the interval . Then it follows that

Proof.

By direct computations we have

Without loss of generality, we assume . From Theorem 2.5.2 in Reference 27, page 70 we know that the matrix admits a singular value decomposition (SVD), i.e., there exist two orthogonal matrices and and a matrix , with being a nonnegative diagonal matrix having the maximum diagonal entry , such that holds. Define

Then is an orthogonal matrix, too. It follows from concrete computations that

Therefore, detailed analysis shows that the eigenvalues of the matrix are with multiplicity and

It then follows straightforwardly that the spectral radius of the matrix , say , is given by

and therefore,

The proof of the second equality can be demonstrated in a similar fashion.

We remark that for the real one-variable function defined by Equation 3.1, the estimate holds for all because of and .

Lemma 3.2.

Let be a diagonal matrix, and a given matrix, where represents the complex matrix space. If there exists a positive constant such that , then all eigenvalues of the matrix are located within , where denotes the circle having center and radius on the complex plane.

Proof.

Let be an eigenvalue of the matrix and be the corresponding normalized eigenvector. Then we have . Hence,

It then follows that . Therefore, it follows that (), or equivalently, .

For the simplicity of our statements, in the sequel we always use to represent the function defined by Equation 3.1. For the matrices in Equation 2.1 and , in Equation 2.3, we write

and denote the -block entry of the matrix in Equation 2.7 by , i.e.,

Assume and are nonsingular, let be a nonsingular matrix that is a replacement to (e.g., or , etc.), and define the quantities

In addition, in the case that is an approximation to , we define the quantities

and in the case that is an approximation to , instead of and we use the quantities

For two positive constants and , to be specified later, we use to denote the circle having center and radius , and use to denote the union of the two circles having centers and and radius , on the complex plane, respectively.

By making use of the above notation, the nonsingularity of the matrices and can be precisely described by the following lemma.

Lemma 3.3.

The matrices and are nonsingular, provided either of the following conditions holds:

(1)

is an approximation to , and

(a)

, or

(b)

;

(2)

is an approximation to , and

(a)

, or

(b)

.

Proof.

We only prove (1a), as the other conclusions can be demonstrated analogously.

Because , it follows that

From Equation 2.8 we have

Hence,

It then follows that

Now, we easily see that Equation 2.9 holds when

or equivalently, . Therefore, when

the matrices and are nonsingular.

We first consider the case that . The case that will be discussed in Section 3.4.

3.1. The MBJ-type preconditioners

If the matrix in Equation 2.10 is taken to be the modified block Jacobi splitting matrix Reference 6Reference 7 of the matrix in Equation 2.7, i.e.,

then we obtain the modified block Jacobi-type (MBJ-type) preconditioner to the original matrix . Note that when is symmetric positive definite, is a symmetric positive definite matrix, and when is symmetric negative definite, is a symmetric indefinite matrix.

The following theorem describes the eigenvalue distribution of the preconditioned matrix with respect to the MBJ-type preconditioner.

Theorem 3.1.

Let be the MBJ-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.3. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Proof.

We only prove (i), as (ii) can be verified analogously.

From Equation 2.7 and Equation 3.3 we have

Hence,

By making use of Lemma 3.1 we can immediately obtain

Furthermore, when the matrix is positive definite, we can demonstrate the positive definiteness of the matrices and .

Theorem 3.2.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Proof.

We only prove the validity of (i), as (ii) can be demonstrated similarly.

Some straightforward computations immediately show that . Let

Then from the proof of Theorem 3.1 we easily obtain

Because is positive definite, we know that its symmetric part is symmetric positive definite. Therefore, the matrix is symmetric positive definite if and only if so is its Schur complement .

Since

we have

By direct computations we immediately obtain

and

It then follows that

Noticing that

holds if and only if

or equivalently,

we therefore know that holds true when . Hence, is a symmetric positive definite matrix, and is a positive definite matrix.

3.2. The MBGS-type preconditioners

If the matrix in Equation 2.10 is taken to be the modified block Gauss-Seidel splitting matrix Reference 6Reference 7 of the matrix in Equation 2.7, i.e.,

then we obtain the modified block Gauss-Seidel-type (MBGS-type) preconditioner to the original matrix .

The following theorem describes the eigenvalue distribution of the preconditioned matrix with respect to the MBGS-type preconditioner.

Theorem 3.3.

Let be the MBGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.4. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Proof.

We only prove (i), as (ii) can be verified analogously.

From Equation 2.7 and Equation 3.4 we have

Hence,

By making use of Lemma 3.1 we can immediately obtain

Furthermore, when the matrix is positive definite, we can demonstrate the positive definiteness of the matrices and .

Theorem 3.4.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Proof.

We only prove the validity of (i), as (ii) can be demonstrated similarly.

Some straightforward computations immediately show that . Let

Then from the proof of Theorem 3.3 we can easily obtain

Because is positive definite, we know that its symmetric part is symmetric positive definite. Therefore, the matrix is symmetric positive definite if and only if so is its Schur complement .

Since

we have

By direct computations we immediately get

and

It then follows that

Notice that

holds if and only if

and this inequality holds when

or equivalently,

Therefore, we know that holds true when . Hence, is a symmetric positive definite matrix, and is a positive definite matrix.

Alternatively, if the matrix in Equation 2.10 is taken to be the modified block Gauss-Seidel splitting matrix Reference 6Reference 7 of the matrix in Equation 2.7, i.e.,

then we obtain another modified block Gauss-Seidel-type (MBGS-type) preconditioner to the original matrix . Exactly following the demonstrations of Theorems 3.3 and 3.4, we can obtain the following results for the eigenvalue distribution and the positive definiteness of the preconditioned matrix with respect to the MBGS-type preconditioner Equation 3.5.

Theorem 3.5.

Let be the MBGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.5. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Theorem 3.6.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

3.3. The MBUGS-type preconditioners

If the matrix in Equation 2.10 is taken to be the modified block unsymmetric Gauss-Seidel splitting matrix Reference 6Reference 7 of the matrix in Equation 2.7, i.e.,

then we obtain the modified block unsymmetric Gauss-Seidel-type (MBUGS-type) preconditioner to the original matrix .

The following theorem describes the eigenvalue distribution of the preconditioned matrix with respect to the MBUGS-type preconditioner.

Theorem 3.7.

Let be the MBUGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.6. Let and . Then it follows that

(i)

, with

(ii)

, with

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Proof.

It is analogous to the proofs of Theorems 3.1 and 3.3 and hence is omitted.

Furthermore, when the matrix is positive definite, we can demonstrate the positive definiteness of the matrices and .

Theorem 3.8.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Proof.

It is analogous to the proofs of Theorems 3.2 and 3.4 and hence is omitted.

Alternatively, if the matrix defined by Equation 3.6 is considered to possess the split form , with

or

then we can obtain other modified block unsymmetric Gauss-Seidel-type preconditioners to the original matrix , where

and and are given by Equation 2.6. Exactly following the demonstrations of Theorems 3.7 and 3.8, we can obtain the results about the eigenvalue distributions and the positive definiteness of the preconditioned matrices with respect to the MBUGS-type preconditioners Equation 3.7Equation 3.8.

We remark that when , the above-discussed modified block unsymmetric Gauss-Seidel-type preconditioners naturally reduce to the modified block symmetric Gauss-Seidel-type (MBSGS-type) preconditioners to the matrix in Equation 1.2, correspondingly.

3.4. The case

In the case that is negative definite, we may let be an approximation to in order to obtain a preconditioner of positive definiteness in nature. Hence, some specified preconditioned Krylov subspace iteration method can exploit its efficiency sufficiently.

When , for the MBJ-, the MBGS-, and the MBUGS-type preconditioners discussed above, we can demonstrate that the eigenvalues of the preconditioned matrices are, correspondingly, located within two circles having center and in the complex plane. These results are precisely summarized in the following theorem. Since their proofs are essentially the same as those of Theorems 3.1, 3.3, 3.5 and 3.7 with only the identity matrix being replaced by the matrix

we only state the theorem but omit its proof.

Theorem 3.9.

Let in Equation 2.10 be the preconditioner to the block two-by-two matrix in Equation 1.2, with and being given by Equation 2.6 and being given by Equation 2.7. Let and

(i)

If is defined by Equation 3.3, then

where and are the same as in Theorem 3.1.

(ii)

If is defined by Equation 3.4, then

where and are the same as in Theorem 3.3.

(iii)

If is defined by Equation 3.5, then

where and are the same as in Theorem 3.5.

(iv)

If is defined by Equation 3.6, then

where and are the same as in Theorem 3.7.

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the preconditioned matrix are located within the union of two circles having centers and and radius , and those of the preconditioned matrix are located within the union of two circles having centers and and radius , respectively. Therefore, they are all within . Here, , and .

We observe from the demonstrations of Theorems 3.1-3.9 that when or , the results in these theorems can be considerably improved and made more accurate.

3.5. Connections to Krylov subspace methods

The preconditioning matrix defined in Equation 2.10 can be used to accelerate the Krylov subspace methods such as GMRES or its restarted variant GMRES() Reference 41Reference 40 for solving the large sparse system of linear equations Equation 1.1Equation 1.2. This preconditioning matrix can be used as a left (see Equation 2.11Equation 2.12), a right (see Equation 2.13Equation 2.14), or a split (see Equation 2.15Equation 2.17) preconditioner to the system of linear equations Equation 1.1. The obtained equivalent linear systems can be solved by GMRES or GMRES().

Assume that the coefficient matrices of the above preconditioned linear systems are diagonalizable, i.e., there exist a nonsingular matrix and a diagonal matrix such that . Then it is well known from Reference 41, Theorem 4 that the residual norm at the -th step of the preconditioned GMRES is bounded by , where is the Euclidean condition number of and . Here, denotes the set of all polynomials of degree not greater than such that , and denotes the spectrum of the matrix .

Consider defined by Equation 3.2; see also Equation 2.1 and Equation 2.3. When the matrix is an approximation to the matrix , from Theorems 3.1, 3.3, 3.5 and 3.7 we know that all eigenvalues of the matrix are contained in either of the circles , , and . Therefore, when , a special case of Theorem 5 in Reference 41 implies that , , and .

Alternatively, the preconditioning matrix can also be used as a left, a right, or a split preconditioner to the system of linear equations Equation 1.1 to obtain a preconditioned linear system of coefficient matrix , , or , respectively. Because Theorems 3.2, 3.4, 3.6 and 3.8 guarantee the positive definiteness of the preconditioned matrix , it is known from Reference 20 and Reference 41, p. 866 that the following error bound for the correspondingly preconditioned GMRES holds:

where denotes the symmetric part of the matrix , and and denote, respectively, the smallest and the largest eigenvalues of the corresponding matrix. This gives a guarantee for the convergence of the restarted preconditioned GMRES iteration, say PGMRES(), for all , when the coefficient matrix is positive definite.

When the matrix is an approximation to the matrix , because the preconditioned matrix or may be usually not positive definite, instead of GMRES and GMRES() we may use other Krylov subspace methods such as BiCGSTAB, QMR and TFQMR to solve the preconditioned linear systems. In particular, when the original coefficient matrix is symmetric indefinite, MINRES is a possible candidate if a symmetric positive definite or indefinite preconditioner is obtainable. See Reference 2Reference 27Reference 40.

4. Applications to three typical matrices

In this section, we will investigate the concretizations of the structured preconditioners established in Sections 2 and 3 to three special classes of matrices arising from real-world applications.

4.1. The symmetric positive definite matrix

When the matrix blocks and are symmetric positive definite, and the Schur complement is symmetric positive definite, the matrix reduces to the block two-by-two symmetric positive definite matrix

These kinds of matrices may arise in the red/black ordering of a symmetric positive definite linear system, or in discretization incorporated with a domain decomposition technique of a boundary value problem of a self-adjoint elliptic partial differential equation, etc. See Reference 2Reference 3Reference 6Reference 7Reference 27Reference 40.

Let and be nonsingular matrices such that either Equation 2.1 or Equation 2.2 holds with and . Then from Equation 2.10 and Equation 2.6 we know that is the structured preconditioner to the matrix , where

and is an approximation to the matrix

defined by Equation 2.7, with and .

Note that and are symmetric positive definite. Let be an approximation to the matrix . To guarantee the symmetric positive definiteness of the preconditioning matrix , we can choose to be the modified block Jacobi splitting matrix in Equation 3.3 or the modified block symmetric Gauss-Seidel splitting matrix in Equation 3.6, obtaining the modified block Jacobi-type preconditioner or the modified block symmetric Gauss-Seidel-type preconditioner to the matrix , respectively.

4.2. The saddle point matrix

When the matrix block is symmetric positive definite, and is of full row rank, the matrix reduces to the saddle point matrices

These kinds of matrices may arise in constrained optimization as well as least-squares, saddle-point and Stokes problems, without a regularizing/stabilizing term, etc. See Reference 14Reference 16Reference 24Reference 25Reference 28Reference 37Reference 44.

Let be a nonsingular matrix such that either Equation 2.1 or Equation 2.2 holds with and . Then from Equation 2.10 and Equation 2.6 we know that are the preconditioners to the matrices , respectively, where

and are approximations to the matrices

defined by Equation 2.7, with and .

Let be approximations to the matrices . By choosing the matrices to be the modified block Jacobi splitting matrices in Equation 3.3, the modified block Gauss-Seidel splitting matrices in Equation 3.4 or Equation 3.5, or the modified block unsymmetric Gauss-Seidel splitting matrices in Equation 3.6, we can obtain the modified block Jacobi-type preconditioners, the modified block Gauss-Seidel-type preconditioners, or the modified block unsymmetric Gauss-Seidel-type preconditioners to the matrices , respectively.

4.3. The Hamiltonian matrix

When the matrix block is symmetric positive definite and is symmetric positive/negative definite (denoted by , respectively), and , the matrix reduces to the Hamiltonian matrices

These kinds of matrices may arise in stationary semiconductor devices Reference 36Reference 43Reference 42, in constrained optimization as well as least-squares, saddle-point and Stokes problems, with a regularizing/stabilizing term Reference 28.

Let and be nonsingular matrices such that either Equation 2.1 or Equation 2.2 holds with and . Then from Equation 2.10 and Equation 2.6 we know that are the preconditioners to the matrices , where

and are approximations to the matrices

defined by Equation 2.7, with and .

Let be approximations to the matrices . By choosing the matrices to be the modified block Jacobi splitting matrices in Equation 3.3, the modified block Gauss-Seidel splitting matrices in Equation 3.4 or Equation 3.5, or the modified block unsymmetric Gauss-Seidel splitting matrices in Equation 3.6, we can obtain the modified block Jacobi-type preconditioners, the modified block Gauss-Seidel-type preconditioners, or the modified block unsymmetric Gauss-Seidel-type preconditioners to the matrices , respectively.

4.4. An illustrative example

Let us consider the electromagnetic scattering problem from a large rectangular cavity on the -plane in which the medium is -directional inhomogeneous. In the transverse magnetic polarization case, when the model Helmholtz equation with positive wave number is discretized by the five-point finite difference scheme with uniform stepsize , we obtain a block two-by-two system of linear equations Equation 1.1Equation 1.2, in which

and , where , , is a real constant, is the -th unit vector in , is the -by- identity matrix, is a tridiagonal matrix, is a nonnegative diagonal matrix, , and denotes the Kronecker product. See Reference 33Reference 1.

Concretely, in our computations we take , , and .

Let be an incomplete triangular factorization of the matrix block , and . Then we have

Now, by choosing with being the band matrix of half-band width truncated from the matrix , after straightforward computations we can obtain the results listed in Tables 1, 2, 3, and 4 for the discretization stepsizes , , and , or equivalently, for the problem sizes , , and , respectively.

In Table 1 we list the half-band width , the quantities

with respect to the matrix norms, and

with respect to the matrix approximation accuracies. For , and , in Tables 2, 3, and 4 we list the radii and of the circles centered at where all eigenvalues of the matrices and are located within and the radii of the smallest circles that include all eigenvalues of the corresponding preconditioned matrices (see Theorems 3.1, 3.3 and 3.7), and the quantities and that guarantee the positive definiteness of the preconditioned matrices and whenever and (see Theorems 3.2, 3.4 and 3.8), respectively.

The results in Tables 2, 3, and 4 clearly show that

(i)

for , and , and . It follows that . As and are quite small, the eigenvalues of the preconditioned matrices, with respect to the MBJ-, the MBGS- and the MBUGS-type preconditioners, are tightly clustered around the point ; see Theorems 3.1, 3.3 and 3.7. Hence, a Krylov subspace method such as GMRES, when applied to the preconditoned systems of linear equations, will achieve fast convergence; see Section 3.5.

(ii)

for , and , and . It follows that the preconditioned matrices, with respect to the MBJ-, the MBGS- and the MBUGS-type preconditioners, are positive definite, and the convergence of the restarted GMRES methods preconditioned by these preconditioners are guaranteed; see Theorems 3.2, 3.4 and 3.8 as well as Section 3.5.

(iii)

for , and , . This shows that the eigenvalues of the preconditioned matrices, with respect to the MBJ-, the MBGS- and the MBUGS-type preconditioners, are really located within the theoretically estimated circles centered at with radii given in Theorems 3.1, 3.3 and 3.7, respectively.

In summary, this example shows that the conditions of our theorems are reasonable and the conclusions of them are correct.

5. Conclusion and remarks

We have established a general framework of practical and efficient structured preconditioners to the large sparse block two-by-two nonsingular matrices. For several special cases associated with the modified block relaxation iteration methods, we have studied the eigenvalue distributions and the positive definiteness of the preconditioned matrices. Theoretical analyses have shown that this preconditioning technique can afford effective and high-quality preconditioners to the Krylov subspace iteration methods for solving large sparse systems of linear equations with block two-by-two coefficient matrices.

We remark that our preconditioning technique and the corresponding theory can be straightforwardly developed to the following cases.

(a)

The approximation matrix in Equation 2.10 that is generated by a multi-step variant of the modified block Jacobi, the modified block Gauss-Seidel or the modified block unsymmetric Gauss-Seidel splitting matrix of the matrix in Equation 2.7 Reference 6Reference 7.

(b)

Alternatively, the approximation matrix in Equation 2.10 that is generated by a single- or multiple-step variant of the modified block successive overrelaxation (SOR), the modified block unsymmetric SOR, the modified block accelerated overrelaxation (AOR) or the modified block unsymmetric AOR splitting matrix of the matrix in Equation 2.7 Reference 32Reference 6Reference 7.

(c)

More generally, the approximation matrix in Equation 2.10 that is generated by any suitable direct or iterative method induced by the matrix in Equation 2.7.

(d)

The matrix that is of a general -by- block structure. More concretely, , where , , and are positive integers satisfying .

For the structured preconditioners based on the relaxation iteration methods involving parameters, we can further optimize them through choices of the optimal parameters. In addition, we should point out that, although all results in this paper are demonstrated in the -norm, they trivially hold for other consistent matrix norms such as the -norm and the -norm.

Acknowledgments

The author is very much indebted to the referees for their constructive and valuable comments and suggestions which greatly improved the original version of this paper.

Figures

Table 1.

Quantities with respect to the preconditioned matrices

24630
11.970415.843225.367928.8844
6.053397.380688.9545314.7829
5.771086.5752311.077539.2410
2.823233.289383.9356019.5354
Table 2.

Bounds with respect to the MBJ-type preconditioner with being defined by Equation 3.3

0.1115668.62e-026.78e-023.26e-02
0.1525070.1084977.98e-022.70e-02
0.1115668.62e-026.78e-022.70e-02
Table 3.

Bounds with respect to the MBGS-type preconditioner with being defined by Equation 3.4

Table 4.

Bounds with respect to the MBUGS-type preconditioner with being defined by Equation 3.6

0.1025180.166135
0.1484790.1063650.216704
0.1025180.166135

Mathematical Fragments

Equation (1.1)
Equation (1.2)
Equation (2.1)
Equation (2.2)
Equation (2.3)
Equation (2.4)
Equation (2.5)
Equation (2.6)
Equation (2.7)
Equation (2.8)
Equation (2.9)
Equation (2.10)
Equation (2.11)
Equation (2.12)
Equation (2.13)
Equation (2.14)
Equation (2.15)
Equation (2.17)
Equation (2.18)
Lemma 3.1.

Let and be unit lower and upper triangular matrices of the block two-by-two forms

where and . Let

be a monotone increasing function with respect to in the interval . Then it follows that

Lemma 3.2.

Let be a diagonal matrix, and a given matrix, where represents the complex matrix space. If there exists a positive constant such that , then all eigenvalues of the matrix are located within , where denotes the circle having center and radius on the complex plane.

Equation (3.2)
Equation (3.3)
Theorem 3.1.

Let be the MBJ-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.3. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Theorem 3.2.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Equation (3.4)
Theorem 3.3.

Let be the MBGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.4. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Theorem 3.4.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Equation (3.5)
Theorem 3.5.

Let be the MBGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.5. Let and . Then it follows that

(i)

, with ; and

(ii)

, with .

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Theorem 3.6.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Equation (3.6)
Theorem 3.7.

Let be the MBUGS-type preconditioner to the block two-by-two matrix in Equation 1.2, where and are given by Equation 2.6, is given by Equation 2.7, and is defined by Equation 3.6. Let and . Then it follows that

(i)

, with

(ii)

, with

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the matrices and are located within a circle having center and radii and , respectively, and therefore, they are all within the circle .

Theorem 3.8.

Let the matrix be positive definite. Then

(i)

the matrix is positive definite, provided , where

(ii)

the matrix is positive definite, provided , where

Equation (3.7)
Equation (3.8)
Theorem 3.9.

Let in Equation 2.10 be the preconditioner to the block two-by-two matrix in Equation 1.2, with and being given by Equation 2.6 and being given by Equation 2.7. Let and

(i)

If is defined by Equation 3.3, then

where and are the same as in Theorem 3.1.

(ii)

If is defined by Equation 3.4, then

where and are the same as in Theorem 3.3.

(iii)

If is defined by Equation 3.5, then

where and are the same as in Theorem 3.5.

(iv)

If is defined by Equation 3.6, then

where and are the same as in Theorem 3.7.

It follows from Lemma 3.2 as well as Equation 2.11 and Equation 2.13 that the eigenvalues of the preconditioned matrix are located within the union of two circles having centers and and radius , and those of the preconditioned matrix are located within the union of two circles having centers and and radius , respectively. Therefore, they are all within . Here, , and .

References

Reference [1]
H. Ammari, G. Bao and A.W. Wood, An integral equation method for the electromagnetic scattering from cavities, Math. Methods Appl. Sci., 23(2000), 1057-1072. MR 1773922 (2001g:78013)
Reference [2]
O. Axelsson, Iterative Solution Methods, Cambridge University Press, Cambridge, 1994. MR 1276069 (95f:65005)
Reference [3]
Z.-Z. Bai, Parallel Iterative Methods for Large-Scale Systems of Algebraic Equations, Ph.D. Thesis of Shanghai University of Science and Technology, Shanghai, June 1993. (In Chinese)
Reference [4]
Z.-Z. Bai, A class of hybrid algebraic multilevel preconditioning methods, Appl. Numer. Math., 19(1996), 389-399. MR 1377785 (96j:65116)
Reference [5]
Z.-Z. Bai, Parallel hybrid algebraic multilevel iterative methods, Linear Algebra Appl., 267(1997), 281-315.MR 1479124 (99c:65081)
Reference [6]
Z.-Z. Bai, A class of modified block SSOR preconditioners for symmetric positive definite systems of linear equations, Adv. Comput. Math., 10(1999), 169-186.MR 1680610 (2000c:65020)
Reference [7]
Z.-Z. Bai, Modified block SSOR preconditioners for symmetric positive definite linear systems, Ann. Operations Research, 103(2001), 263-282.MR 1868455 (2002k:65046)
Reference [8]
Z.-Z. Bai, I.S. Duff and A.J. Wathen, A class of incomplete orthogonal factorization methods. I: methods and theories, BIT, 41(2001), 53-70. MR 1829661 (2002b:65040)
Reference [9]
Z.-Z. Bai and G.-Q. Li, Restrictively preconditioned conjugate gradient methods for systems of linear equations, IMA J. Numer. Anal., 23(2003), 561-580. MR 2011340 (2004i:65025)
Reference [10]
Z.-Z. Bai and M.K. Ng, On inexact preconditioners for nonsymmetric matrices, SIAM J. Sci. Comput., 26(2005), 1710-1724.
Reference [11]
Z.-Z. Bai and Z.-Q. Wang, Restrictive preconditioners for conjugate gradient methods for symmetric positive definite linear systems, J. Comput. Appl. Math., 187(2006), 202-226.
Reference [12]
Z.-Z. Bai and D.-R. Wang, A class of new hybrid algebraic multilevel preconditioning methods, Linear Algebra Appl., 260(1997), 223-255. MR 1448358 (98i:65035)
Reference [13]
J.T. Betts, Practical Methods for Optimal Control Using Nonlinear Programming, SIAM, Philadelphia, PA, 2001. MR 1826768 (2002e:49001)
Reference [14]
A. Björck, Numerical Methods for Least Squares Problems, SIAM, Philadelphia, PA, 1996. MR 1386889 (97g:65004)
Reference [15]
J.H. Bramble, J.E. Pasciak and A.T. Vassilev, Analysis of the inexact Uzawa algorithm for saddle point problems, SIAM J. Numer. Anal., 34(1997), 1072-1092. MR 1451114 (98c:65182)
Reference [16]
F. Brezzi and M. Fortin, Mixed and Hybrid Finite Element Methods, Springer-Verlag, New York, 1991.MR 1115205 (92d:65187)
Reference [17]
I.S. Duff, N.I.M. Gould, J.K. Reid, J.A. Scott and K. Turner, The factorization of sparse symmetric indefinite matrices, IMA J. Numer. Anal., 11(1991), 181-204. MR 1105227 (92a:65143)
Reference [18]
I.S. Duff and J.K. Reid, Exploiting zeros on the diagonal in the direct solution of indefinite sparse symmetric linear systems, ACM Trans. Math. Software, 22(1996), 227-257. MR 1408491 (97c:65085)
Reference [19]
N. Dyn and W.E. Ferguson, Jr., The numerical solution of equality constrained quadratic programming problems, Math. Comput., 41(1983), 165-170. MR 0701631 (85b:90051)
Reference [20]
S.C. Eisenstat, H.C. Elman and M.H. Schultz, Variational iterative methods for nonsymmetric systems of linear equations, SIAM J. Numer. Anal., 20(1983), 345-357.MR 0694523 (84h:65030)
Reference [21]
H.C. Elman, Preconditioners for saddle point problems arising in computational fluid dynamics, Appl. Numer. Math., 43(2002), 75-89. MR 1936103
Reference [22]
H.C. Elman and G.H. Golub, Inexact and preconditioned Uzawa algorithms for saddle point problems, SIAM J. Numer. Anal., 31(1994), 1645-1661. MR 1302679 (95f:65065)
Reference [23]
H.C. Elman, D.J. Silvester and A.J. Wathen, Performance and analysis of saddle point preconditioners for the discrete steady-state Navier-Stokes equations, Numer. Math., 90(2002), 665-688.MR 1888834 (2002m:76071)
Reference [24]
B. Fischer, R. Ramage, D.J. Silvester and A.J. Wathen, Minimum residual methods for augmented systems, BIT, 38(1998), 527-543.MR 1652781 (99i:65031)
Reference [25]
P.E. Gill, W. Murray and M.H. Wright, Practical Optimization, Academic Press, New York, NY, 1981.MR 0634376 (83d:65195)
Reference [26]
R. Glowinski, Numerical Methods for Nonlinear Variational Problems, Springer-Verlag, New York, 1984.MR 0737005 (86c:65004)
Reference [27]
G.H. Golub and C.F. Van Loan, Matrix Computations, 3rd Edition, The Johns Hopkins University Press, Baltimore and London, 1996.MR 1417720 (97g:65006)
Reference [28]
G.H. Golub and A.J. Wathen, An iteration for indefinite systems and its application to the Navier-Stokes equations, SIAM J. Sci. Comput., 19(1998), 530-539.MR 1618828 (99d:65107)
Reference [29]
G.H. Golub, X. Wu and J.-Y. Yuan, SOR-like methods for augmented systems, BIT, 41(2001), 71-85.MR 1829662 (2002a:65055)
Reference [30]
N.I.M. Gould, M.E. Hribar and J. Nocedal, On the solution of equality constrained quadratic programming problems arising in optimization, SIAM J. Sci. Comput., 23(2001), 1375-1394.MR 1885606 (2002k:90072)
Reference [31]
E. Haber, U.M. Ascher and D. Oldenburg, On optimization techniques for solving nonlinear inverse problems, Inverse Problems, 16(2000), 1263-1280.MR 1798355 (2001j:78025)
Reference [32]
A. Hadjidimos, Accelerated overrelaxation method, Math. Comput., 32(1978), 149-157.MR 0483340 (58:3353)
Reference [33]
J. Jin, The Finite Element Method in Electromagnetics, John Wiley & Sons, New York, NY, 1993.MR 1903357 (2004b:78019)
Reference [34]
C. Keller, N.I.M. Gould and A.J. Wathen, Constrained preconditioning for indefinite linear systems, SIAM J. Matrix Anal. Appl., 21(2000), 1300-1317.MR 1780274 (2002b:65050)
Reference [35]
A. Klawonn, Block-triangular preconditioners for saddle point problems with a penalty term, SIAM J. Sci. Comput., 19(1998), 172-184.MR 1616885 (99f:65051)
Reference [36]
P.A. Markowich, The Stationary Semiconductor Device Equations, Springer-Verlag, New York, 1986.MR 0821965 (87b:78042)
Reference [37]
M.F. Murphy, G.H. Golub and A.J. Wathen, A note on preconditioning for indefinite linear systems, SIAM J. Sci. Comput., 21(2000), 1969-1972.MR 1762024 (2001a:65055)
Reference [38]
I. Perugia and V. Simoncini, Block-diagonal and indefinite symmetric preconditioners for mixed finite element formulations, Numer. Linear Algebra Appl., 7(2000), 585-616.MR 1802361 (2001j:65187)
Reference [39]
R.J. Plemmons, A parallel block iterative method applied to computations in structural analysis, SIAM J. Alg. Disc. Meth., 7(1986), 337-347.MR 0844035 (88h:49058)
Reference [40]
Y. Saad, Iterative Methods for Sparse Linear Systems, 2nd Edition, SIAM, Philadelphia, PA, 2003. MR 1990645 (2004h:65002)
Reference [41]
Y. Saad and M.H. Schultz, GMRES: a generalized minimal residual algorithm for solving nonsymmetric linear systems, SIAM J. Sci. Stat. Comput., 7(1986), 856-869.MR 0848568 (87g:65064)
Reference [42]
G.E. Sartoris, A 3D rectangular mixed finite element method to solve the stationary semiconductor equations, SIAM J. Sci. Stat. Comput., 19(1998), 387-403.MR 1618875 (99b:65147)
Reference [43]
S. Selberherr, Analysis and Simulation of Semiconductor Devices, Springer-Verlag, New York, 1984.
Reference [44]
G. Strang, Introduction to Applied Mathematics, Wellesley-Cambridge Press, MA, 1986.MR 0870634 (88a:00006)
Reference [45]
W. Zulenher, Analysis of iterative methods for saddle point problems: a unified approach, Math. Comput., 71(2001), 479-505.MR 1885611 (2003f:65183)

Article Information

MSC 2000
Primary: 65F10 (Iterative methods for linear systems), 65F50 (Sparse matrices)
Keywords
  • Block two-by-two matrix
  • preconditioner
  • modified block relaxation iteration
  • eigenvalue distribution
  • positive definiteness
Author Information
Zhong-Zhi Bai
State Key Laboratory of Scientific/Engineering Computing, Institute of Computational Mathematics and Scientific/Engineering Computing, Academy of Mathematics and System Sciences, Chinese Academy of Sciences, P.O. Box 2719, Beijing 100080, People’s Republic of China
bzz@lsec.cc.ac.cn
Additional Notes

Subsidized by The Special Funds For Major State Basic Research Projects (No. G1999032803), The National Basic Research Program (No. 2005CB321702), and The National Natural Science Foundation (No. 10471146), P.R. China.

Journal Information
Mathematics of Computation, Volume 75, Issue 254, ISSN 1088-6842, published by the American Mathematical Society, Providence, Rhode Island.
Publication History
This article was received on , revised on , and published on .
Copyright Information
Copyright 2005 American Mathematical Society
Article References
  • Permalink
  • Permalink (PDF)
  • DOI 10.1090/S0025-5718-05-01801-6
  • MathSciNet Review: 2196992
  • Show rawAMSref \bib{2196992}{article}{ author={Bai, Zhong-Zhi}, title={Structured preconditioners for nonsingular matrices of block two-by-two structures}, journal={Math. Comp.}, volume={75}, number={254}, date={2006-04}, pages={791-815}, issn={0025-5718}, review={2196992}, doi={10.1090/S0025-5718-05-01801-6}, }

Settings

Change font size
Resize article panel
Enable equation enrichment

Note. To explore an equation, focus it (e.g., by clicking on it) and use the arrow keys to navigate its structure. Screenreader users should be advised that enabling speech synthesis will lead to duplicate aural rendering.

For more information please visit the AMS MathViewer documentation.