DOI QR코드

DOI QR Code

FAST ONE-PARAMETER RELAXATION METHOD WITH A SCALED PRECONDITIONER FOR SADDLE POINT PROBLEMS

  • OH, SEYOUNG (Department of Mathematics, Chungnam National University) ;
  • YUN, JAE HEON (Department of Mathematics, College of Natural Sciences, Chungbuk National University)
  • Received : 2015.11.02
  • Accepted : 2015.12.10
  • Published : 2016.01.30

Abstract

In this paper, we first propose a fast one-parameter relaxation (FOPR) method with a scaled preconditioner for solving the saddle point problems, and then we present a formula for finding its optimal parameter. To evaluate the effectiveness of the proposed FOPR method with a scaled preconditioner, numerical experiments are provided by comparing its performance with the existing one or two parameter relaxation methods with optimal parameters such as the SOR-like, the GSOR and the GSSOR methods.

Keywords

1. Introduction

We consider a fast one-parameter relaxation method with a scaled preconditioner for solving the saddle point problem

where A ∈ ℝm×m is a symmetric positive definite matrix, and B ∈ ℝm×n is a matrix of full column rank. This problem (1) appears in many different scientific applications, such as constrained optimization [9], the finite element approximation for solving the Navier-Stokes equation [5,6], and the constrained least squares problems and generalized least squares problems [1,3,12]. So many authors have proposed one or two parameter relaxation iterative methods for solving the saddle point problem (1). Golub et al. [7] proposed the one-parameter SOR-like method and presented an incomplete formula for finding one optimal parameter, Bai et al. [3] proposed the two-parameter GSOR (Generalized SOR) method and presented a formula for finding two optimal parameters for the GSOR and a complete formula for finding one optimal parameter for SOR-like method, Zhang and Lu [14] studied the two-parameter GSSOR (Generalized symmetric SOR) method and Chao et al [4] presented a formula for finding two optimal parameters for the GSSOR, and so on [10,13].

This paper is organized as follows. In Section 2, we propose a fast oneparameter relaxation (FOPR) method with a scaled preconditioner, and then we present a formula for finding its optimal parameter. In Section 3, numerical experiments are provided to examine the effectiveness of the proposed FOPR method with a scaled preconditioner by comparing its performance with the existing one or two parameter relaxation methods with optimal parameters such as the SOR-like, the GSOR and the GSSOR methods. Lastly, some conclusions are drawn.

 

2. Convergence of a fast one-parameter relaxation (FOPR) method

For the coefficient matrix of the augmented linear system (1), we consider the following splitting

where

and Q ∈ ℝn×n is a symmetric positive definite matrix which approximates BTA−1B. Let

where ω > 0 is a relaxation parameter, Im ∈ ℝm×m and In ∈ ℝn×n denote the identity matrices of order m and n, respectively. Then a fast one-parameter relaxation (FOPR) method for solving the saddle point problem (1) is defined by

where Tω = (D − ΩL)−1((I − Ω)D + ΩU) is an iteration matrix for the FOPR method, gω = Ωc, and I is an identity matrix of order m + n. That is, the FOPR method is defined by

Note that the FOPR method can be viewed as a special case of the GSOR method [3] with

Lemma 2.1 ([11]). Consider the quadratic equation x2 − bx + c = 0, where B and c are real numbers. Both roots of the equation are less than one in modulus if and only if |c| < 1 and |b| < 1 + c.

Let λ be an eigenvalue of Tω and be the corresponding eigenvector. Then we have

or equivalently,

From now on, let ρ(H) denote the spectral radius of a square matrix H. The following theorem provides the convergence result for the FOPR method.

Theorem 2.2. Let μmax be the spectral radius of Q−1BTA−1B. If μmax < 4, then the FOPR method converges for all

Proof. Let μ be an eigenvalue of Q−1BTA−1B and λ be an eigenvalue of Tω. Then μ > 0. From equation (5), one can obtain the following quadratic equation for λ

Applying Lemma 2.1 to (6), one easily obtains , then ρ(Tω) < 1, which completes the proof. □

Notice that if μmax ≥ 4 in Theorem 2.2, then the convergence region for which the FOPR method converges may be an empty set. Next theorem provides an optimal parameter ω for which the FOPR method performs best.

Theorem 2.3. Let μmin and μmax be the minimum and maximun eigenvalues of Q−1BTA−1B, respectively. Assume that μmax < 4. Then the optimal parameter ω for the FOPR method is given by ω = ωo, where

Moreover . That is,

Proof. Let μ be an eigenvalue of Q−1BTA−1B and λ be an eigenvalue of Tω. From the quadratic equation (6) for λ, one obtains two roots

Let f(ω) = 2 − ω − μ and g(ω) = (ω + μ)2 − 4μ. The necessary and sufficient condition for the roots λ to be real is g(ω) ≥ 0, which is equivalent to . Since Hence one obtains

Notice that for μ ∈ (0, 4) and it has the maximum value 1 at μ = 1. Since is an increasing function for is a decreasing function for . Thus, (8) implies that given μ, |λ| takes the minimum . If S is a set containing all eigenvalues of Q−1BTA−1B, then

where . Hence the theorem follows. □

As can be shown in Theorems 2.2 and 2.3, a big disadvantage of the FOPR method is that it requires a rather strong condition μmax < 4 which may not be true for some types of preconditioners Q. To remedy this problem, we need to scale the preconditioner Q so that 0 < μmin, μmax < 4. From Theorem 2.3, it can be also seen that in order to minimize ρ(Tωo), Q needs to be scaled so that . Next theorem shows how to scale the preconditioner Q such that 0 < μmin, μmax < 4 and ρ(Tωo) can be minimized. Next theorem also shows an optimal convergence factor of the FOPR method with a scaled preconditioner.

Theorem 2.4. Let μmin and μmax be the minimum and maximun eigenvalues of Q−1BTA−1B, respectively. Let Qs = s Q be a scaled preconditioner, where s > 0 is a scaling factor, and let νmin and νmax be the minimum and maximun eigenvalues of , respectively. If , then 0 < νmin, νmax < 4 and and

where refer to the optimal parameter and the iteration matrix for the FOPR method with the scaled preconditioner Qs, respectively.

Proof. Since

Since , one obtains

Using (10), it can be easily shown that

The remaining part of this theorem can be proved by simple calculation. □

From Theorem 2.4, it can be seen that optimal convergence factor of the FOPR method with the scaled preconditioner Qs is the same as that of the GSOR method [3] with the preconditioner Q. Notice that the scaling factor s in Theorem 2.4 can be easily computed using MATLAB by computing only the largest and smallest eigenvalues of Q−1BTA−1B.

We next summarize the formulas for finding optimal parameters of the SOR-like, the GSOR and the GSSOR methods which are used for numerical experiments in Section 3.

Remark 2.1 ([7,3,4]). Let μmin and μmax be the minimum and maximun eigenvalues of Q−1BTA−1B, respectively. Then

 

3. Numerical results

To evaluate the effectiveness of the FOPR method with a scaled preconditioner, we provide numerical experiments by comparing its performance with the SOR-like, the GSOR and the GSSOR methods. For performance comparison, both the FOPR method with preconditioner Q and the FOPR method with scaled preconditioners Qs = s Q and Qs+ϵ = (s + ϵ)Q are provided, where s is the scaling factor defined in Theorem 2.4 and ϵ is a positive number which is chosen appropriately small as compared with s. In Tables 2 to 5, Iter denotes the number of iteration steps and CPU denotes the elapsed CPU time in seconds. In all experiments, the right hand side vector (bT ,−qT )T ∈ ℝm+n was chosen such that the exact solution of the saddle point problem (1) is = (1, 1,...,1)T ∈ ℝm+n, and the initial vector was set to the zero vector. From now on, let ║·║ denote the L2-norm.

Example 3.1 ([2]). We consider the saddle point problem (1), in which

and

with ⊗ denoting the Kronecker product and the discretization mesh size. For this example, m = 2p2 and n = p2. Thus the total number of variables is 3p2. We choose the matrix Q as an approximation to the matrix BTA−1B, according to four cases listed in Table 1. The iterations for the relaxation iterative methods are terminated if the current iteration satisfies ERR < 10−9, where ERR is defined by

Numerical results for this example are listed in Tables 2 to 5. In Tables 4 and 5, numerical results for the FOPR method are not listed since it does not work because of μmax > 4, and thus only those for the FOPR method with scaled preconditioners Qs and Qs+ϵ are listed.

Example 3.2 ([3]). We consider the same problem as Example 3.1 except that F is defined by

Note that the matrix F is a highly ill-conditioned Toeplitz matrix. So we choose only Cases III and IV of Q in Table 1 as an approximation to the matrix BTA−1B, and all iterations are terminated if the current iteration satisfies RES < 10−9, where RES is defined by

Since μmax > 4 for these choices of Q, numerical results for the FOPR method with scaled preconditioners Qs and Qs+ϵ are listed in Tables 4 to 5.

Table 1.Choices of the matrix Q.

Table 2.Numerical results for Example 3.1 with Case I of Q.

Table 3.Numerical results for Example 3.1 with Case II of Q.

Table 4.Numerical results for Case III of Q.

Table 5.Numerical results for Case IV of Q.

All numerical tests are carried out on a PC equipped with Intel Core i5-4570 3.2GHz CPU and 8GB RAM using MATLAB R2014a. For numerical experiments of all relaxation iterative methods used in this paper, the optimal parameters described in Remark 2.1 are used. For test runs of the FOPR method with the scaled preconditioner Qs+ϵ , we have tried the values of ϵ ∈ [0.0001, 0.0005] in Tables 2 and 3, and the values of ϵ ∈ [0.01, 0.05] in Tables 4 and 5. For all of these values of ϵ, the FOPR method with Qs+ϵ performs better than the GSOR method, and the values of ϵ for which it performs best are reported in Tables 2 to 5.

As can be expected from Theorem 2.4, the FOPR method with the scaled preconditioner Qs performs as well as the GSOR method. The FOPR method with the scaled preconditioner Qs+ϵ performs best of all methods considered in this paper, and specifically it performs much better than other methods for Cases III and IV of Q for which μmax > 4 (see Tables 2 to 5). On the other hand, the GSSOR method performs worse than the FOPR and the GSOR methods since its computational cost for each iteration is higher than others.

 

4. conclusions

In this paper, we proposed a fast one-parameter relaxation (FOPR) method with a scaled preconditioner for solving the saddle point problems, and then we presented a formula for finding its optimal parameter. Both theoretical and computational results showed that the FOPR methods with the scaled preconditioner Qs performs as well as the GSOR method. In addition, the FOPR method with the scaled preconditioner Qs+ϵ performs better than other one or two parameter relaxation methods with optimal parameters, and specifically it performs much better than other methods for Cases III and IV of Q for which μmax > 4 (see Tables 2 to 5). Hence, it may be concluded that the FOPR method with the scaled preconditioner Qs+ϵ is strongly recommended for solving the saddle point problems when μmax = ρ(Q−1BTA−1B) > 4.

References

  1. M. Arioli, I.S. Duff, P.P.M. de Rijk, On the augmented system approach to sparse least squares problems, Numer. Math. 55 (1989), 667-684. https://doi.org/10.1007/BF01389335
  2. Z.Z. Bai, G.H. Golub, J.Y. Pan, Preconditioned Hermitian and skew-Hermitian splitting methods for non-Hermitian positive semidefinite linear systems, Numer. Math. 98 (2004), 1-32. https://doi.org/10.1007/s00211-004-0521-1
  3. Z.Z. Bai, B.N. Parlett, Z.Q. Wang, On generalized successive overrelaxation methods for augmented linear systems, Numer. Math. 102 (2005), 1-38. https://doi.org/10.1007/s00211-005-0643-0
  4. Z. Chao, N. Zhang, Y.Z. Lu, Optimal parameters of the generalized symmetric SOR method for augmented systems, J. Comput. Appl. Math. 266 (2014), 52-60. https://doi.org/10.1016/j.cam.2014.01.023
  5. H. Elman, D.J. Silvester, Fast nonsymmetric iteration and preconditioning for Navier-Stokes equations, SIAM J. Sci. Comput. 17 (1996), 33-46. https://doi.org/10.1137/0917004
  6. B. Fischer, A. Ramage, D.J. Silvester, A.J. Wathen, Minimum residual methods for augmented systems, BIT 38 (1998), 527-543. https://doi.org/10.1007/BF02510258
  7. G.H. Golub, X. Wu, J.Y. Yuan, SOR-like methods for augmented systems, BIT 41 (2001), 71-85. https://doi.org/10.1023/A:1021965717530
  8. G.H. Santos, B.P.B. Silva, J.Y. Yuan, Block SOR methods for rank deficient least squares problems, J. Comput. Appl. Math. 100 (1998), 1-9. https://doi.org/10.1016/S0377-0427(98)00114-9
  9. S. Wright, Stability of augmented system factorization in interior point methods, SIAM J. Matrix Anal. Appl. 18 (1997), 191-222. https://doi.org/10.1137/S0895479894271093
  10. S.L. Wu, T.Z. Huang, X.L. Zhao, A modified SSOR iterative method for augmented systems, J. Comput. Appl. Math. 228 (2009), 424-433. https://doi.org/10.1016/j.cam.2008.10.006
  11. D.M. Young, Iterative solution of large linear systems, Academic Press, New York, 1971.
  12. J.Y. Yuan, A.N. Iusem, Preconditioned conjugate gradient methods for generalized least squares problem, J. Comput. Appl. Math. 71 (1996), 287-297. https://doi.org/10.1016/0377-0427(95)00239-1
  13. SeYoung Oh, J.H. Yun, K.S. Kim, ESOR method with diagonal preconditioners for SPD linear systems, J. Appl. Math. & Informatics 33 (2015), 111-118. https://doi.org/10.14317/jami.2015.111
  14. G.F. Zhang, Q.H. Lu, On generalized symmetric SOR method for augmented systems, J. Comput. Appl. Math. 219 (2008), 51-58. https://doi.org/10.1016/j.cam.2007.07.001