An incomplete generalized minimum backward perturbation algorithm for large nonsymmetric linear systems

Lei SUN

Front. Math. China ›› 2023, Vol. 18 ›› Issue (3) : 203 -222.

PDF (1307KB)
Front. Math. China ›› 2023, Vol. 18 ›› Issue (3) : 203 -222. DOI: 10.3868/s140-DDD-023-0014-x
RESEARCH ARTICLE
RESEARCH ARTICLE

An incomplete generalized minimum backward perturbation algorithm for large nonsymmetric linear systems

Author information +
History +
PDF (1307KB)

Abstract

This paper gives the truncated version of the generalized minimum backward error algorithm (GMBACK)—the incomplete generalized minimum backward perturbation algorithm (IGMBACK) for large nonsymmetric linear systems. It is based on an incomplete orthogonalization of the Krylov vectors in question, and gives an approximate or quasi-minimum backward perturbation solution over the Krylov subspace. Theoretical properties of IGMBACK including finite termination, existence and uniqueness are discussed in details, and practical implementation issues associated with the IGMBACK algorithm are considered. Numerical experiments show that, the IGMBACK method is usually more efficient than GMBACK and GMRES, and IMBACK, GMBACK often have better convergence performance than GMRES. Specially, for sensitive matrices and right-hand sides being parallel to the left singular vectors corresponding to the smallest singular values of the coefficient matrices, GMRES does not necessarily converge, and IGMBACK, GMBACK usually converge and outperform GMRES.

Graphical abstract

Keywords

Nonsymmetric linear systems / Krylov subspace methods / minimum backward perturbation / incomplete orthogonalization process / GMBACK / GMRES

Cite this article

Download citation ▾
Lei SUN. An incomplete generalized minimum backward perturbation algorithm for large nonsymmetric linear systems. Front. Math. China, 2023, 18(3): 203-222 DOI:10.3868/s140-DDD-023-0014-x

登录浏览全文

4963

注册一个新账户 忘记密码

1 Introduction

In many applied sciences and engineering calculations, it is often necessary to solve large nonsymmetric sparse linear systems:

Ax=b,

where ARn×n is nonsingular, x,bRn. The Krylov subspace method is a very effective method for solving (1). Krylov subspace methods [20] such as the generalized minimum residual method (CMRES algorithm [4, 22]) often use the residual norm as a condition for determining the termination of the algorithm. If the approximate solution is accurate, then the residual norm is small. Conversely, a small residual norm does not mean that the approximate solution is accurate, especially when A is a ill-posed matrix. In order to overcome the shortcomings of residual norm as a termination condition, the idea of using backward perturbation norm [10] as the termination condition is proposed in reference [1]. Kasenally minimized the backward perturbation norm ΔAF on A by finding an approximate solution satisfying the perturbation equation when solving nonsymmetric linear systems (AΔA)xm=b, that is, minxmx0+Km(A,r0)ΔAF makes

(AΔA)xm=b.

The generalized minimum backward perturbation method (CMBACK algorithm [13]) for solving large nonsymmetric linear systems is presented. In this notation, xm represents the approximate solution of (1) with the form: xm=x0+tm, where tmKm(A,r0), x0 represents the original estimate of the approximate solution, denoted r0:=bAx0. ΔA and Δb represent perturbations to matrix A and vector b, respectively, forming a joint backward perturbation matrix [ΔA,Δb]. In the late 1990s, Kasenally, Simoncini and Zhihao Cao popularized the GMBACK algorithm by perturbing 6 on top of 4, that is, minxmx0+Km(A,r0)ΔA,ΔbF makes

(AΔA)xm=(b+Δb).

The minimum joint backward perturbation method for solving large nonsymmetric linear systems is obtained (Minpert algorithm [8, 14]). The GMRES algorithm can be regarded as minimizing the perturbation norm Δb2 on b by finding an approximate solution satisfying the equation Axm=bΔb, that is, minxmx0+Km(A,r0)Δb2 makes

Axm=bΔb.

Therefore, GMBACK and GMRES algorithms can be regarded as minimizing the perturbation norm on the coefficient matrix A and vector b, respectively. While the Minpert algorithm is minimizing the norm of joint backward perturbation matrix [ΔA,Δb] on matrix A and vector b.

The GMBACK, GMRES, and Minpert algorithms are all a set of bases v1,v2,,vm that utilize the Arnoldi process to generate Km(A,r0). This means that these algorithms use long recursive formulas, which leads to a dramatic increase in computation and storage capacity with the number of steps, and the algorithms become unusable when the number of steps increases to a certain point. Therefore, these algorithms usually need to be restarted, but for difficult problems, even when combined with preprocessing techniques, the number of steps taken in a restart is still quite large to ensure convergence of the algorithm. In order to overcome the shortcomings of long recursive formulas, a popular technique is to resort to truncation strategies. The idea is to use only a few rather than all the previously calculated vectors to construct a new short recursive formula to calculate the following vectors, which greatly reduces the computation and storage capacity. The truncation form of CMRES algorithm is given in references [5, 6]: quasi-incomplete orthogonalization method (QGMRES) or incomplete generalized minimum Residue Method (ICMRES). Sun Lei provided the truncated form of the Minpert algorithm in reference [25]: incomplete minimum joint backward perturbation method (IMinpert algorithm). However, the truncated form of the CMIBACK algorithm has not yet been given. In this paper, by means of incomplete orthogonalization process [12, 20, 21], a group of bases v1,v2,,vm of Km(A,r0) will be generated. The truncation form of CMBACK algorithm is given: incomplete generalized minimum backward perturbation method (IGMBACK). As with GMBACK, the approximate solution of IGMBACK is generated by solving the minimum problem (2).

The structure of this paper is as follows: Section 2 gives the theoretical derivation process of the ICMBACK algorithm, followed by the IGMBACK algorithm, and some theoretical studies are made on the algorithm, including the finite termination of the algorithm, existence and uniqueness of the solution. Section 3 gives the execution of IGMBACK algorithm. In Section 4, we will show with numerical examples that IGMBACK is generally more efficient than GMBACK and GMRES, and IGMBACK and GMBACK generally converge better than GMRES. In particular, the restarted GMRES algorithm does not necessarily converge if the coefficient matrix is a sensitive matrix and the vector b on the right-hand side of equations is parallel to the left singular vector corresponding to the minimum singular value of A. However, IGMBACK and GMBACK algorithms generally converge and converge better than GMRES. Section 5 is the summary of the full text and the prospect of the work.

2 IGMBACK algorithm

2.1 Analysis of the backward perturbation matrix

The following lemma gives the general formula for the backward perturbed matrix ΔA.

Lemma 2.1  Suppose that an incomplete orthogonalization process of m steps has been performed and a set of basis vectors v1,v2,,vm and vm+1 of Km(A,r0) is obtained, denoted Vm=[v1,v2,,vm], Vm+1=[v1,v2,,vm,vm+1]. At the same time, an upper Hessenberg matrix HmR(m+1)×m is obtained satisfying:

AVm=Vm+1Hm=VmH¯m+hm+1,mvm+1emT,

where the last line of Hm is removed to get H¯m. The approximate solution of the equation (1) can be written as follows:

xm=x0+Vmym,ymRm.

Denoted β=r02. Thus the set S={ΔA} of backward perturbation matrices satisfying (AΔA)xm=b can be expressed as:

S={Vm+1(Hmymβe1)xm22xmT+R(IwmwmT):RRn×n},

where wm=xmxm21.

Proof Substituting AVm=Vm+1Hm and (3) into (AΔA)xm=b, we get

ΔAwm=Vm+1(Hmymβe1)xm21.

Thus for any RRn×n, we immediately obtain (4) (see [3]).

By Lemma 2.1, we can further obtain F- in S the backward perturbation matrix Δmin with the minimum norm and its norm ΔminF.

Theorem 2.1  The backward perturbed matrix Δmin with the minimum norm of F in S can be expressed as

Δmin:=Vm+1(Hmymβe1)xm22xmT=rmwmTxm2,

where rm=bAxm. Thus the backward perturbation norm ΔminF=rm2xm2.

Proof Require F- in S the backward perturbation matrix with the minimum norm. According to Lemma 2.1, let

g(ym,R)=Vm+1(Hmymβe1)xm22xmT+R(IwmwmT)F2.

We can write g(ym,R) as

g(ym,R)=vec{Vm+1(Hmymβe1)xm22xmT}+{(IwmwmT)I}vec(R)22,

where vec() indicates that the columns of the matrix are straightened, denotes tensor product. From the above equation, the minimum value of g(ym,R) is required. By the knowledge of optimization theory, we have vec(R)g(ym,R)=0. Thus we have R(IwmwmT)=0, and the conclusion of the theorem holds. □

2.2 Generation of IGMBACK algorithm

This section will derive the IGMBACK algorithm. Lemma 2.1 gives the general formula (4) for the backward perturbation matrix ΔA. In the proof of Theorem 2.1, we establish the function g(ym,R) according to the general formula (4). It can be seen that the only free variables in g(ym,R) are R and ym, so that the minimum value problem (2) can be transformed into minymRm,RRn×ng(ym,R), i.e.,

minymRm,RRn×nVm+1(Hmymβe1)xm22xmT+R(IwmwmT)F2.

According to Theorem 2.1, problem (5) can be further transformed into

minymRmVm+1(Hmymβe1)xm22xmTF2.

Since Vm+1=[v1,v2,,vm,vm+1] is not a column orthogonal norm matrix, this can be computationally intensive, and since

Vm+1(Hmymβe1)xm22xmTFVm+1FHmymβe12xm2=m+1Hmymβe12xm2,

we can choose ym such that Hmymβe12xm2 is minimal. That is, problem (6) is approximated as

minymRmHmymβe122xm22.

For convenience, we define two matrices LmR(m+1)×(m+1) and GmRn×(m+1), where

Lm=[Hmβe1],Gm=[Vmx0].

The following theorem will give the solution to the minimum value problem (8).

Theorem 2.2  Assuming that the m-step incomplete orthogonalization process has been performed, the form of xm in problem (8) is as in (3). Let {λi,μi}i=1,2,,m+1 be the combination of all generalized eigenvalues and corresponding eigenvectors of (LmTLm,GmTGm), where λiλi+1,i=1,2,,m. If the last element of the vector μm+1 is not 0, i.e., μm+1,m+10, then the solution ym of the minimum value problem (8) satisfies

[ym1]=1μm+1,m+1μm+1

and

ΔminFm+1λm+112.

Proof The preceding reasoning process eventually transforms problem (2) into the minimum value problem (8). By

minymRmHmymβe122xm22=minymRm[Hmβe1][ymT1]T22[Vmx0][ymT1]T22=minμRm+1Lmμ22Gmμ22,

and the Courant−Fischer theorem [24], we have,

minμRm+1Lmμ22Gmμ22=λm+1,

where λm+1 is the minimum generalized eigenvalue of (LmTLm,GmTGm). Since μm+1,m+10, the solution ym of (8) can be calculated by using equation (10). From (7), we soon obtain (11).

By Theorem 2.2, we derive the IGMBACK algorithm. In order to reduce the storage and computation capacity of the algorithm, the algorithm in this paper adopts the m step loop format. The main steps for IGMBACK(m) are given as follows.

Algorithm 1 Restart IGMBACK(m):

① Choose the initial estimate x0 of (1) and the parameter q, where q satisfies 2qm, compute r0:=bAx0 and v1:=r0β, where β:=r02.

② Incomplete orthogonalization process (q) : Perform m steps of the incomplete orthogonalization process to compute Vm+1 and Hm, for j= 1,2,,m.

i. Calculate v^j+1:=Avj.

ii. For i=max{1,jq+1},,j1,j, compute hij=viTv^j+1,v^j+1:=v^j+1hijvi.

iii. Calculate hj+1,j=v^j+12. If hj+1,j0, compute vj+1:=v^j+1hj+1,j, otherwise stop.

③ Solving generalized eigenvalue problems.

LmTLmu=λGmTGmu,

where Lm and Gm are defined in (9).

④ The approximate solution xmIGB:=x0+Vmym is calculated from Equation (10) in Theorem 2.2.

⑤ Start over: Define (Hmymβe1)xm22xmTF=λm+112. If the termination condition is met, stop, otherwise redefine x0:=xmIGB, compute r0:=bAx0,v1:=r0β, where β:=r02, and return to step (2).

Denote

P=LmTLm=[HmTHmβHmTe1βe1THmβ2],Q=GmTGm=[VmTVmVmTx0x0TVmx0Tx0].

Therefore, the generalized eigenvalue problem (13) can then be written as Pu=λQu.

The minimum value problem (12) can be rewritten as

λm+112=minμRm+1(Gmμ2μ2)1Lmμ2μ2.

Using equation (15), we can give an estimate of the range of (Hmymβe1)xm22xmTF=λm+112 in the termination condition of the algorithm. This is shown by the following theorem. Here xm=x0+Vmym is the approximate solution of the IGMBACK algorithm.

Theorem 2.3  Assuming that m steps of incomplete orthogonalization have been performed, σ1(Gm) denotes the maximum singular value of Gm. σm+1(Lm),σm+1(Gm) are the minimum singular values of Lm,Gm respectively. Let σm+1(Gm)>0, we have

σm+1(Lm)σ11(Gm)λm+112σm+1(Lm)σm+11(Gm).

Proof By employing equation (15), we have

λm+112minz1Rm+1(Gmz12z12)1minz2Rm+1Lmz22z22=σ11(Gm)σm+1(Lm)

and

λm+112maxz1Rm+1(Gmz12z12)1minz2Rm+1Lmz22z22=σm+11(Gm)σm+1(Lm).

Thus, the conclusion of the theorem holds.□

2.3 The theoretical research of IGMBACK algorithm

This section presents a theoretical study of the IGMBACK algorithm in terms of finite termination, existence and uniqueness of the solution.

2.3.1 Finite termination of the algorithm

Suppose that after m steps of incomplete orthogonalization, we get hm+1,m=0, i.e., P is singular, then the vector vm+1 cannot be formed. When this happens, we call it a lucky break [22].

Let the column vectors of Hm1 be linearly independent and the F- backward perturbation matrix with minimum norm Δi generated in the IGMBACK algorithm satisfies Δi0 when i<m. According to Theorem 2.1, a sufficient condition for Δmin=Δm=0 is Hmymβe1=0, then we have

Δmin=0hm+1,m=0.

In fact, assume hm+1,m=0, ym=H¯m1βe1, then Δmin=0, thus xm=x0+VmH¯m1βe1 is an exact solution to the system of equations (1). Conversely, if Δmin=0, then Hmymβe1=0. Let ym=[(ym(1))T,ym(2)]T, where ym(2)R. From the last line of the equation Hmymβe1=0, we get hm+1,mym(2)=0. If ym(2)=0, then Hm1ym(1)βe1=0, and thus Δm1=0, which contradicts the previous assumption and gives hm+1,m=0.

We know that the algorithm can execute at most m=n steps before a lucky break occurs, so the following corollary holds.

Corollary 2.1  For any ARn×n and bRn, the IGMBACK algorithm can converge by up to n steps.

According to the above discussion, hi+1,i0,im is assumed in the rest of this paper. That is, P is assumed to be non-singular.

2.3.2 Existence and uniqueness of solutions

There are two cases where the solution of the algorithm does not exist: one is when the generalized eigenvalue problem Pu=λQu has degenerate generalized eigenvalues, the other is when μm+1,m+1=0 in (10).

Lemma 2.2  Let the incomplete orthogonalization process in m steps have been performed, then, the necessary and sufficient condition for Q=GmTGm to be a singular matrix is x0=0 or x0Km(A,r0).

Proof According to the basic knowledge of algebra, Q=GmTGm is a singular matrix if and only if the rank r(Gm)< m+1 of matrix Gm, where Gm is shown in (9). On the one hand, if Q=GmTGm is a singular matrix, then r(Gm)<m+1, i.e., the set of column vectors of Gm=[Vmx0] is linearly related. The set of column vectors of Vm is linearly independent, so x0 can be uniquely linearly tabulated by the set of column vectors of Vm, i.e., x0Km(A,r0). On the other hand, if x0=0 or x0Km(A,r0), i.e., x0 can be linearly tabulated by the set of column vectors of Vm, then the set of column vectors of Gm is linearly related. Thus r(Gm)<m+1, i.e., Q is a singular matrix.

By Lemma 2.2, Q is potentially singular. We choose x0 in the actual implementation of the IGMBACK algorithm such that x0Km(A,r0), which ensures that Q is a non-singular matrix. Thus, the generalized eigenvalue problem Pu=λQu belongs to the non-degenerate generalized eigenvalue problem [7], without degenerate generalized eigenvalues and eigenvectors. In general, the dimension m of Km(A,r0) chosen in the IGMBACK algorithm is much smaller than the order n of A. Therefore, there is plenty of scope for choosing x0 in the n-dimensional vector space Rn such that x0Km(A,r0). Throughout the rest of the paper, we assume that x0Km(A,r0), i.e., that Q is a non-singular matrix.

The following theorem gives a sufficient condition for the IGMBACK algorithm to fail to form an approximate solution when the multiplicity of λm+1 is 1.

Theorem 2.4  Suppose {λm+1,μm+1} is the combination of the minimum generalized eigenvalue and the corresponding eigenvector of (P,Q), Q is a non-singular matrix, and the multiplicity of λm+1 is 1. Let μm+1=[llμ^Tμm+1,m+1]T. Then

μm+1,m+1=0HmTHmμ^=λm+1VmTVmμ^,μ^(λm+1VmTx0+HmTβe1).

Proof  Let μm+1,m+1=0. Substituting the expressions for P,Q in (14) into Pu=λQu, we have

[HmTHmβHmTe1βe1THmβ2][μ^0]=λm+1[VmTVmVmTx0x0TVmx0Tx0][μ^0].

From the first line of the matrix equation above, we get HmTHmμ^=λm+1VmTVmμ^. From the second line we get βe1THmμ^= λm+1x0TVmμ^, i.e., (λm+1VmTx0+HmTβe1)Tμ^=0, that is,

μ^(λm+1VmTx0+HmTβe1).

Assume HmTHmμ^=λm+1VmTVmμ^, and (λm+1VmTx0+HmTβe1)Tμ^=0. Then [μ^T,0]T is the eigenvector of Pu=λQu, and the corresponding generalized eigenvalue is λm+1. Since Q is a non-singular matrix, Pu=λQu does not have degenerate generalized eigenvalues and eigenvectors, and since the multiplicity of λm+1 is 1, k[μ^T,0]T(kR,k0) is the eigenvector of all Pu=λQu corresponding to λm+1. Therefore, there must be μm+1,m+1=0.

The existence and uniqueness of the solution are discussed further below. For convenience, according to (14) and (8), denote

Pm=P=LmTLm,Qm=Q=GmTGm=[VmTVmVmTx0x0TVmx0Tx0],Δ~m=minymRm[ymT1]Pm[ymT1]T[ymT1]Qm[ymT1]T.

Let’s start with a lemma.

Lemma 2.3  Suppose that an incomplete orthogonalization process of m+1 steps has been performed. {λi}i=1,2,,m+1 and {λ^i}i=1,2,,m+2 represent all generalized eigenvalues of (Pm,Qm) and (Pm+1,Qm+1) in non-increasing order, respectively, then

λ^iλiλ^i+1,i=1,2,,m+1.

Proof Let {λi,μi}i=1,2,,m+1 be all generalized characteristic pairs of (Pm,Qm). According to Subsection 2.3.1, if hm+2,m+10, then Lm+1,Lm are non-singular matrices. Therefore, we can write Pu=λQu as (Lm1)TQmLm1 zi=ξizi, where zi=Lmμi,ξi is the reciprocal of λi. The size order of {λi}i=1,2,,m+1 determines 0<ξ1ξ2 ξm+1. By

K=[Hm+1βe1]=[Lmh^m+10hm+2,m+1]Pm+2(m+1,m+2),

where h^m+1=[Im+10]Hm+1em+1,Pm+2(m+1,m+2) represents the primary matrix after interchanging the m+2 and m+1 columns of the unit matrix Im+2, one gets that

Lm+11=Pm+2(m+1,m+2)[Lm1Lm1h^m+1hm+2,m+110hm+2,m+11].

Substituting the above equation into (Lm+11)TQm+1Lm+11, we have

(Lm+11)TQm+1Lm+11=[(Lm1)TQmLm1ffTg],

where

g=hm+2,m+12+2hm+2,m+11dm+1hm+1+dm+1Qmdm+1T,f=hm+2,m+11(Lm1)T(hm+1QmLm1h^m+1),hm+1=[VmTvm+1x0Tvm+1](m+1)×1,dm+1=hm+2,m+11h^m+1T(Lm1)T.

Thus, according to Theorem IV.4.2 in reference [24], one obtains that

ξ^m+2ξm+1ξ^m+1ξmξ^2ξ1ξ^1,

where {ξ^i}i=1,2,,m+2 is all the eigenvalues of (Lm+11)TQm+1Lm+11. Since ξi,ξ^i are the reciprocals of λi and λ^i, respectively, the conclusion of the lemma holds.

According to Lemma 2.3, we obtain the following conclusions about the convergence, existence and uniqueness of the solution.

Corollary 2.2  Assuming that the incomplete orthogonalization process has been performed in m+1 steps, the following conclusion holds:

(a) Δ~m+1Δ~m.

(b) If λ^m+2<λ^m+1 and μm+2,m+20, then there is a unique approximate solution to the IGMBACK algorithm.

(c) If λ^m+2=λ^m+1, and the approximate solution of the IGMBACK algorithm exists, then the approximate solution of the IGMBACK algorithm is not unique, and Δ~m+1=Δ~m, i.e., the IGMBACK algorithm stalls from step m to step m+1.

Let the minimum generalized eigenvalue λm+1 of (Pm,Qm) be of multiplicity not 1. Assume λmk+1>λmk+2==λm+1, i.e., let λm+1 be of k multiplicity, and the corresponding standard orthogonal eigenvectors are μmk+2,μmk+3,,μm+1, respectively. Let

Uk=[μmk+2μmk+3μm+1]R(m+1)×k.

Then for any z=[z1Tz2]Tspan{Uk},z2R, if z20, the approximate solution of the IGMBACK algorithm takes the form

xm=x0+Vmz1z2.

In the practical implementation of the algorithm, we often choose z1z2 such that z122 is minimal to make the resulting approximate solution unique. Maximize span{Uk} in all unit vectors containing span {z:z=[z1Tz2]Tspan{Uk},z2=1} maximizes |z2| such that z1z22 is minimized.

3 Execution of the IGMBACK algorithm

This section considers how to solve the generalized eigenvalue problem (13), i.e., Pu=λQu.

In the implementation of Algorithm 1, we focus on solving the generalized eigenvalue problem (13). Based on Theorem 2.2, it is sufficient to compute the minimum generalized eigenvalue of (LmTLm,GmTGm) and the corresponding combination of eigenvectors {λm+1,μm+1}. Since LmTLmλGmTGm is symmetric, we can compute {λm+1,μm+1} with the help of the inverse iteration method. However, when the condition numbers of LmTLm and GmTGm are relatively large, it becomes difficult to compute {λm+1,μm+1} by inverse iteration. Therefore, in the GMBACK algorithm [13], the authors solve the corresponding generalized eigenvalue problem by solving the associated singular value decomposition problem, which overcomes the difficulties associated with the increasing number of conditions for LmTLm and GmTGm. Can Algorithm 1 also calculate {λm+1,μm+1} by solving the associated singular value decomposition problem?

Noting that the matrix Q contains VmTVm, make the Cholesky decomposition of VmTVm: VmTVm=XXT, where X is the lower triangular matrix, thus

Q=[X0x0TVm(X1)Tx022X1VmTx022][XTX1VmTx00x022X1VmTx022]=ZZT,

where

Z=[X0x0TVm(X1)Tx022X1VmTx022].

By substituting Q=ZZT into Pu=λQu, the original generalized eigenvalue problem Pu=λQu is transformed into an ordinary eigenvalue problem for symmetric matrices:

Z1P(Z1)Tv=λv,

where v=ZTu. When λ is small, based on P=[Hmβe1]T[Hmβe1], to solve exactly for the matrix Z1P(Z1)T with the minimum eigenvalue λmin and the corresponding eigenvector v is not an easy task. Therefore, we transform the original problem (16) into a solution for the minimum singular value of another matrix

=[Hm(X1)THm([VmTVm]1)TVmTx0xlβxle1],

and the corresponding right singular vector, which solves the original problem very well. In equation (17), xl=1x022X1VmTx022.

So we get the specific execution process of IGMBACK algorithm.

Algorithm 2 Restart IGMBACK algorithm, IGMBACK(m):

① Select initial estimated value x0 and parameter q of (1), where q satisfies 2qm. Calculate r0:=bAx0 and v1:=r0β, where β:=r02.

② Incomplete orthogonalisation process (q): Performing an incomplete orthogonalisation process in m steps, Vm+1 and Hm were calculated, for j=1,2,,m.

i. Calculate v^j+1:=Avj.

ii. For i=max{1,jq+1},,j, compute hij=viTv^j+1, v^j+1:=v^j+1hijvi.

iii. Calculate hj+1,j=v^j+12, if hj+1,j0, calculate vj+1=v^j+1hj+1,j, otherwise stop.

③ Compute the minimum singular value σ of the matrix [Hmβe1](Z1)T and the corresponding right singular vector v. Calculate u=(Z1)Tv=[y~mT,η]T. Standardize the vector u to derive ym:ym=y~mη. Forming an approximate solution xmIGB:=x0+Vmym.

④ Starting over: Define (Hmymβe1)xm22xmTF2=σ. If the termination condition is met, stop. Otherwise, redefine x0:=xmIGB and calculate r0:=bAx0, v1:=r0β where β:=r02, and return to step ②.

4 Numerical experiments

It is well known that when the coefficient matrix A is a ill-posed matrix, a small perturbation of A or b causes a large change in the approximate solution of (1). This is further exacerbated [9] when b is parallel to the left singular vector corresponding to the minimum singular value of A. In this section, we illustrate the effectiveness of the ICMBACK(m) algorithm in solving the preceding problem by means of several numerical examples, and the IGMBACK(m) algorithm is often more efficient than the GMBACK(m) and CMRES(m) algorithms in solving many problems. The programming software used for the numerical experiments in this paper is MATLAB 7.0.

We mainly compare the following three algorithms:

1) IGMBACK(m,q)(q is the parameter in the incomplete orthogonalisation process).

2) GMBACK(m).

3) CMRES(m). The different convergence rates of the three algorithms during execution are reflected by comparing the change in the backward perturbation norm ΔminF of the approximate solutions of the three algorithms with the increase in the number of selected generations (restarts). In the GMRES(m) algorithm, let the backward perturbation norm ΔminF=rm2xm2. The stop criteria is ΔminF107.

4.1 Numerical experiments with strong informal matrices

This part of the numerical experiment will consider two matrices that are sensitive to small perturbations (called sensitivity matrices).

We notate the singular value decomposition of the matrix A with the following symbols: A=USVT, where

U=[u1,u2,,un],V=[v1,v2,,vn],S=(s1,s2,,sn).

For i=1,2,,n1, sisi+1 in S, then u1 and v1 are the left singular vectors and the right singular vector corresponding to the maximum singular value sn of matrix A, respectively.

Example 4.1 Consider a Toeplitz matrix of order 100×100: A=Toeplitz([1_,1,1]). Calculate the maximum singular value s1=2.9990, and minimum singular value sn=2.6784102 of the matrix A. For the first experiment Test 1, set b=u1, x0=v1. Thus, we obtain r0=u1Av1, that is to say, r0 is parallel to u1 (noted as r0u1). For the second experiment Test 2, assume b=un, x0=vn. Thus, we obtain r0=unAvn, that is to say, r0un. In these two experiments, let m=20, q=15. The experimental results are shown in Fig.1 (a) and Fig.1 (b), respectively. 10 in the vertical coordinate represents 1010, and so on.

Fig.1 (a) shows that the convergence rates of the three algorithms in the first experiment are essentially the same. From Fig.1 (b), it can be seen that the convergence rates of IGMBACK(m) and GMBACK(m) in the second experiment are equivalent (IGMBACK (m) converges slightly faster), and both converge significantly faster than GMRES(m). The coefficient matrices are the same in both experiments (both sensitive matrices), what makes IGMBACK(m) and GMBACK(m) significantly better than GMRES(m) in the second experiment? This is mainly because the second experiment was set up with bun and x0=vn making r0un.

Example 4.2 Consider another matrix of order 100×100: A=Toeplitz([1,1_,1,1,1]). Matrix A has sensitive eigenvalues. Calculate the maximum singular value s1=4.2394, and minimum singular value sn=9.0205101 of the matrix A. The first experiment Test 1, solves the equation Ax=b, with a setting of b=u0, x0=vn. Let m=20, q=15. The result is shown in Fig.2 (a).

From Fig.2 (a), it can be seen that the convergence rates of IGMBAC-K(m), GMBACK(m), and GMRES(m) are completely consistent, that is, the setting of vector b and the selection of initial estimates for approximate solutions cannot cause differences in the convergence rates of the three algorithms, as shown in Example 4.1 (even if set at bun). This is mainly owing to the fact that the coefficient matrix A is well conditioned. However, by reducing the minimum singular value of A, the convergence rates of the above three algorithms will become different. Use E=0.89unvnT to perturb the matrix A. The first n1 singular values of the perturbed coefficient matrix AE are the same as A, while the minimum singular value Sn decreases to 1.2048102. The left and right singular vectors also remain unchanged. The second experiment Test 2 considers solving the system of equations (AE)x=b using the three algorithms described above, still setting b=un, x0=vn. Let m=20, q=15. The result is shown in Fig.2 (b).

As depicted in Fig.2 (b), stagnation occurs in GMRES, while IGMBACK(m) and GMBACK(m) are convergent, and IGMBACK(m) converges at a rate comparable to GMBACK(m). Obviously, the latter two algorithms are clearly superior to the restarted GMRES(m), which is largely attributable to the the foundation of setting bun. In the second experiment, when the minimum singular value of A is reduced, the condition of the coefficient matrix becomes worse and the coefficient matrix becomes sensitive matrix.

From the previous two examples, we can conclude that the GMRES(m) algorithm, which uses the least-squares problem to derive approximate solutions, is not sufficient to guarantee convergence if the coefficient matrix is a sensitive matrix and the vector b on the right-hand side of the system of equations is parallel to the left singular vector corresponding to the least singular value of the coefficient matrix. However, GMBACK(m) and IGMBACK(m) algorithms can get better convergence effect by minimizing or approximately minimizing the backward perturbation norm. Interestingly, when comparing the TLS (Total Least Sguares) algorithm with the classical least square method, we can come to a similar conclusion [26].

In order to further illustrate that the algorithms IGMBACK(m) can be compared with GMBACK(m), we compare the CPU time taken by each of the two algorithms to run in Example 4.1 and Example 4.2. The comparison results are shown in Tab.1. In the table, CPU and CPUa represent the entire CPU time used and the average CPU time used every reboot, respectively, in seconds. iter represents the number of restarts; ratio is the ratio of CPU time used by algorithm IGMBACK(m) and GMBACK(m); BPN represents the norm of backward perturbation. Note that IGMBACK(m) becomes GMBACK(m) when q=m.

From Tab.1, we can see that the algorithm IGMBACK(m) is more efficient than GMBACK(m).

4.2 Numerical experiments related to partial differential equations

In practical applications, a large number of large-scale nonsymmetric linear systems arise from the discretization of partial differential equations.

Example 4.3 Consider the convective diffusion equation defined over the region (0,1) × (0,1)

2ux22uy2+γ(xux+yuy)+βu=f,

where γ=1000, β=10. The boundary condition is u(x,y)=0. The above convective diffusion equation is discretized by the central difference method with lattice length h=132 to obtain an nonsymmetric matrix A of order 312. Take b such that x=[111]T is an exact solution to the system of equations Ax=b. Suppose x0 is a random matrix of order 961×1, with the values of the elements in the interval (0.0,1.0), and let m=15, q=10. The experimental result is shown in Fig.3.

In this example, the GMRES algorithm does not converge in 400 steps when m is set at 15, while Fig.3 shows that the ICMBACK and GMBACK can drop to around 108 in 40 steps by the backward perturbation norm. This indicates that the latter two algorithms are significantly better than GMRES algorithm. Tab.2 shows the CPU time spent by each of the three algorithms when m is 15 and 25 respectively in Example 4.3.

Tab.2 illustrates that IGMBACK is more efficient than GMBACK and that both algorithms are significantly better than GMRES algorithm.

Example 4.4 Consider the second order elliptic partial differential equation defined on the region [0,1]×[0,1]:

ey2ux2ex2uy2+(x+y)ux+(xy)uy+u=(y2ey+x2ex)exy+(y2+x2+1)exy,

the boundary condition is u(0,y)=1, u(1,y)=ey, u(x,0)=1, u(x,1)=ex. Differentiating the above second-order elliptic partial differential equations by the central difference method with lattice lengths A and h duals, respectively, yields a nonsymmetric difference linear systems Ax=b of coefficient matrix orders 392 and 692. Let x0=111T. The experimental results are shown in Tab.3 and Tab.4, respectively. The symbols in Tab.3 and Tab.4 are used as in Tab.1, where ratio represents the ratio of CPU time used by the algorithm IGMBACK(m) or GMRES(m) and GMBACK(m) for the same value of m.

As seen in Tab.3 and Tab.4: overall IGMBACK(m) and GMBACK(m) converge faster than CMRES(m). GMBACK(m) is more effective than GMRES(m), which in turn is more effective than IGMBACK(m). GMBACK(m) is more effective than GMRES(m), and IGMBACK(m) is more effective than GMBACK(m). We have additionally done a number of numerical experiments and found that IGMBACK(m) and GMBACK(m) often converge better than GMRES(m), and that IGMBACK(m) is generally more efficient than GMBACK(m) and GMRES(m).

5 Summary and expectation of the work

In today's increasingly popular Krylov subspace methods [15], this paper presents another method for solving large sparse linear systems of equations Krylov subspace method: The truncated form of GMBACK algorithm, that is, IGMBACK algorithm. The IGMBACK algorithm uses an incomplete orthogonalization process to generate a set of bases for Km(A,r0), which overcomes the disadvantage of using Arnold process in CMBACK algorithm with large amount of computation and storage capacity. The detailed theoretical derivation and some theoretical analysis of the new algorithm are given. Numerical experiments have shown that IGMBACK is generally more effective than CMBACK and CMRES.

It should be noted that the GMRES algorithm has been greatly improved in recent years by the efforts of many authors, and has become a major iterative method for solving large nonsymmetric linear systems. Reference [18] lists the GMRES algorithm including some of the major deformations [11, 17, 19, 23] and theoretical developments [2, 16] in recent years. At present, the international exploration of GMRES algorithm is deepening, and GMRES algorithm itself has become a relatively independent research field. At the same time, this algorithm is also widely used in many fields such as science and engineering calculation. In this case, the advantages of IGMBACK and GMBACK over CMRES in solving many problems illustrate the value of IGMBACK and GMBACK.

The convergence problem of the GMBACK algorithm has not yet been solved. In the truncated version of the GMBACK algorithm, the cost is the loss of some important properties of the original method, such as orthogonality or A-orthogonality, minimization of the backward perturbation parametrization, etc., which are the main basis for the analysis of the convergence of the method, thus making the convergence of the IGMBACK algorithm more complicated. Therefore, one of our major research directions in the future is how to establish the convergence theory of these two algorithms and compare them with GMRES algorithm theoretically, so as to further determine theoretically what kind of equations the GMBACK series algorithm is better than the restarted GMRES for solving. In addition, in many cases and applications, direct use of the iterative method does not converge or converges very slowly. The current general solution is to combine iterative methods with preprocessing techniques. Therefore, another major research direction in the future is how to combine IGMBACK and GMBACK algorithm with preprocessing technology, so as to obtain a new algorithm with better convergence effect and higher accuracy.

References

[1]

Arioli M, Duff I, Ruiz D. Stopping criteria for iterative solvers. SIAM J Matrix Anal Appl 1992; 13(1): 138–144

[2]

Baker A H, Jessup E R, Kolev T V. A simple strategy for varying the restart parameter in GMRES(m). J Comput Appl Math 2009; 230(2): 751–761

[3]

Ben-IsraelAGrevilleT N E. Generalized Inverses: Theory and Applications. New York: Wiley Interscience, 1974

[4]

Brown P N. A theoretical comparison of the Arnoldi and GMRES algorithms. SIAM J Sci Stat Comput 1991; 12(1): 58–78

[5]

Brown P N, Hindmarsh A C. Reduced storage matrix methods in stiff ODE systems. Appl Math Comput 1989; 31: 40–91

[6]

Brown P N, Saad Y. Hybrid Krylov methods for nonlinear systems of equations. SIAM J Sci Stat Comput 1990; 11(3): 450–481

[7]

Cao Z H. On a deflation method for the symmetric generalized eigenvalue problem. Linear Algebra Appl 1987; 92: 187–196

[8]

Cao Z H. Total generalized minimum backward error algorithm for solving nonsymmetric linear systems. J Comput Math 1998; 16(6): 539–550

[9]

GolubG HVan LoanC F. Matrix Computations, 2nd ed. Baltimore, MD: John Hopkins Univ Press, 1990

[10]

Higham D J, Higham N J. Backward error and condition of structured linear systems. SIAM J Matrix Anal Appl 1992; 13(1): 162–175

[11]

Huhtanen M, Permki A. Orthogonal polynomials of the R-linear generalized minimal residual method. J Approx Theory 2013; 167(3): 220–239

[12]

Jia Z X. On IOM(q): the incomplete orthogonalization method for large unsymmetric linear systems. Numer Linear Algebra Appl 1996; 3(6): 491–512

[13]

Kasenally E M. GMBACK: a generalized minimum backward error algorithm for nonsymmetric linear systems. SIAM J Sci Comput 1995; 16(3): 698–719

[14]

Kasenally E M, Simoncini V. Analysis of a minimum perturbation algorithm for nonsymmetric linear systems. SIAM J Numer Anal 1997; 34(1): 48–66

[15]

Li X A, Chen Y H, Zhang Y, Wang X P. Development of the Krylov subspace method for solving large sparse linear systems. science & Technology Review 2013; 11: 68–73

[16]

LiesenJTichyP. The field of values bound on ideal GMRES, 2012, arXiv: 1211.5969v1

[17]

Liu Y Q, Yin K X, WU E H. Fast GMRES-GPU algorithm for solving large scale sparse linear systems. Journal of Computer-Aided Design & Computer Graphics 2011; 23(4): 553–560

[18]

Ma X F. An overview of recent developments and applications of the GMRES method. Pure Math 2013; 3: 181–187

[19]

Najafi H S, Zareamonghaddam H. A new computational GMRES method. Appl Math Comput 2008; 199(2): 527–534

[20]

Saad Y. Krylov subspace methods for solving large unsymmetric linear systems. Math Comp 1981; 37(155): 105–126

[21]

Saad Y. Practical use of some Krylov subspace methods for solving indefinite and nonsymmetric linear systems. SIAM J Sci Stat Comput 1984; 5(1): 203–228

[22]

Saad Y, Schultz M H. GMRES: a generalized minimal residual algorithm for solving nonsymmetric linear systems. SIAM J Sci Stat Comput 1986; 7(3): 856–869

[23]

Sterck H D. Steepest descent preconditioning for nonlinear GMRES optimization. Numer Linear Algebra Appl 2013; 20(3): 453–471

[24]

StewartG WSunJ-G. Matrix Perturbation Theory. New York: Academic Press, 1990

[25]

Sun L, Wang X H, Guan Y. IMinpert: an incomplete minimum perturbation algorithm for large unsymmetric linear systems. Numer Math J Chinese Univ 2007; 16(4): 300–312

[26]

Van HuffelSVandewalleJ. The Total Least Squares Problem: Computational Aspects and Analysis. Frontiers in Applied Mathematics, Vol 9. Philadelphia, PA: SIAM, 1991

RIGHTS & PERMISSIONS

Higher Education Press 2023

AI Summary AI Mindmap
PDF (1307KB)

696

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/