Three-term derivative-free projection method for solving nonlinear monotone equations

Jinkui LIU , Xianglin DU

Front. Math. China ›› 2023, Vol. 18 ›› Issue (4) : 287 -299.

PDF (410KB)
Front. Math. China ›› 2023, Vol. 18 ›› Issue (4) : 287 -299. DOI: 10.3868/s140-DDD-023-0018-x
RESEARCH ARTICLE
RESEARCH ARTICLE

Three-term derivative-free projection method for solving nonlinear monotone equations

Author information +
History +
PDF (410KB)

Abstract

In this paper, a three-term derivative-free projection method is proposed for solving nonlinear monotone equations. Under some appropriate conditions, the global convergence and R-linear convergence rate of the proposed method are analyzed and proved. With no need of any derivative information, the proposed method is able to solve large-scale nonlinear monotone equations. Numerical comparisons show that the proposed method is effective.

Keywords

Nonlinear monotone equations / conjugate gradient method / derivative-free projection method / global convergence / R-linear convergence rate

Cite this article

Download citation ▾
Jinkui LIU, Xianglin DU. Three-term derivative-free projection method for solving nonlinear monotone equations. Front. Math. China, 2023, 18(4): 287-299 DOI:10.3868/s140-DDD-023-0018-x

登录浏览全文

4963

注册一个新账户 忘记密码

1 Introduction

In this paper, we consider solving the nonlinear monotonic system of equations:

F(x)=0,

where F:RnRn is continuous nonlinear mapping. The monotonicity of Fmeans:

F(x)F(y),xy0,x,yRn.

Nonlinear monotone equations are originated from many scientific and practical problems, such as first-order necessity conditions for unconstrained convex optimization problems and subproblems of generalized proximity point algorithms [9], some monotone variational inequality problems [20], chemical equilibrium problems [14], economic equilibrium problems [6], and power equations [4]. Therefore, many algorithms have been proposed to solve nonlinear monotone systems of equations, among which Newton’s Method and Quasi-Newton Methods [4, 5, 11, 17, 21-23] are well known by researchers due to their excellent convergence speed. However, these algorithms are not suitable for solving large-scale non-smooth systems of equations, because they need to solve a linear system of equations using the Jacobian matrix of the objective function or an approximate Jacobian matrix at each iteration to determine the next search direction.

In recent years, in order to efficiently solve large-scale systems of equations, many researchers have promoted the nonlinear conjugate gradient method under the hyperplane projection technique [17]. For example, Cheng [3] combined the well-known PRP method [15, 16] and hyperplane projection technique to study the derivative-free PRP-type projection method, and proved the global convergence for nonlinear monotone systems of equations without requiring the objective function to be differentiable. Li et al. [12] and Ahookhosh et al. [1] further studied three derivative-free PRP-type projection algorithms. Wu et al. [18] established a derivative-free three-term HS projection algorithm based on the well-known HS conjugate gradient method [8] and hyperplane projection technique. The algorithm inherits the advantages of the HS conjugate gradient method with small storage and does not need to calculate any derivatives, thus it can solve large-scale non-smooth nonlinear monotonic systems of equations. Xiao and Zhu [19], Liu and Li [13] proposed a derivative-free CG_DESCENT type projection method based on the CG_DESCENT method [7] and hyperplane projection technique, and achieved good numerical results and proved the global convergence of the method under appropriate conditions.

Recently, Andrei [2] proposed a three-term nonlinear conjugate gradient method (denoted as TTCG method) based on the well-known CG_DESCENT method. This method is a modified memoryless BFGS algorithm whose search direction satisfies the descent and D-L conjugacy conditions. Numerous numerical experiments have shown that the TTCG method is more efficient than the well-known CG_DESCENT method. To solve the large-scale system of equations problem more effectively, this paper extends the TTCG method based on the hyperplane projection technique and establishes a three-term derivative-free projection algorithm. Under appropriate assumptions, we prove the global convergence and R-linear convergence rate of the algorithm.

2 Derivative-free projection method

The iterative method is one of the most used methods for solving nonlinear systems of equations problems. It generates an iterative column of points {xk}, which is usually based on the following unconstrained optimization problem, namely

minxRnf(x),

where f(x)=12F(x)2. Various iterative methods based on solving Problem (2.1) have been widely studied and applied in practice. However, the stable point of Problem (2.1) is not necessarily the solution of Problem (1.1). Moreover, in many practical problems, the mapping F is usually non-smooth.

To solve Problem (1.1) directly, Solodov and Svaiter [17] discussed the hyperplane projection method. Let xk and dk be the current iteration point and the search direction, respectively. A certain line search was used to obtain the step αk so that the point zk=xk+αkdk is obtained such that:

FT(zk)(xkzk)>0.

If F(x~)=0 for some point x~, then it follows from the monotonicity of F:

FT(zk)(x~zk)0.

Thus, the hyperplane Hk={FT(zk)(xzk)=0} strictly separates the current iteration point xk from the solution of the nonlinear monotonic system of equations. Using the hyperplane, the next iteration point xk+1 is obtained:

xk+1=xkFT(zk)(xkzk)F(zk)2F(zk).

The following is a three-term derivative-free projection algorithm for solving a nonlinear monotonic system of equations based on the TTCG method and the hyperplane projection method.

Algorithm 2.1

Step 0: Given the initial point x0Rn and the constant σ(0,1),t0,β>0. If F0=0, stop.

Step 1: Set d0F0, put k:=0.

Step 2: Calculate the step αk=βmk such that mk satisfies the following equation for the smallest non-negative integer m:

FT(xk+αkdk)dkσαkF(xk+αkdk)dk2.

Step 3: Let zk=xk+αkdk. If F(zk)=0, stop. Otherwise, use (2.2) to solve for xk+1.

Step 4: If Fk+1=0, stop. Otherwise, calculate the search direction:

dk+1=Fk+1+βk+1dk+θk+1(dk+yk),

where βk+1=1dkTωk(yktdkyk2dkTωk)TFk+1, θk+1=Fk+1TdkdkTωk, yk=Fk+1Fk, ωk=yk+tkdk, tk = 1+max{0,dkTykdkTdk}.

Step 5: Set k:=k+1, turn to Step 2.

The following theorem shows that Algorithm 2.1 is a descending iterative algorithm when F is the gradient of the real-valued function f:RnR.

Theorem 2.1  Let the sequence {Fk} and {dk} be generated by Algorithm 2.1, then

FkTdkFk2,k0.

Proof According to the definition of parameterωk, it can be seen that

dkTωk=dkTyk+tkdk2=dkTyk+dk2+max{0,dkTyk}dk2.

Multiply both ends of (2.4) by Fk+1T and use (2.6) to obtain

Fk+1Tdk1=Fk+12+βk+1Fk+1Tdk+θk+1(Fk+1Tdk+Fk+1Tyk)=Fk+12(1+tyk2dkTωk)(Fk+1Tdk)2dkTωkFk+12.

Known that d0=F0, then FkTd0=F02. The conclusion is proved.

Note 2.1 According (2.5) and (2.6), we know that

dkTωkFk2.

Since that, before Algorithm 2.1 stops, the parameter ωk and θk is always meaningful.

3 Global convergences

To prove the global convergence of Algorithm 2.1, assume that the mapping F satisfies:

(H) Let the mapping F be Lipschitz continuous, i.e., there exists a constant L>0 such that

F(x)F(y)Lxy,x,yRn.

Lemma 3.1  Let the mapping F satisfy (H). The sequence {Fk} and {dk} are generated by Algorithm 2.1, then the step factor αk satisfies

αkβL+σF(xk+β1αkdk)Fk2dk2.

Proof If αk1, it follows from the definition of αk that β1αk does not satisfy the line search condition (2.3). Thus,

(F(xk+β1αkdk),dk)<σβ1αkF(xk+β1αkdk)dk2.

According to (2.5) and (3.1), it follows that

Fk2FkTdk=[F(xk+β1αkdk)F(xk)]TdkF(xk+β1αkdk)TdkLβ1αkdk2+σβ1αkF(xk+β1αkdk)dk2.

So, it can be concluded that

αkβL+σF(xk+β1αkdk)Fk2dk2.

The lemma is proven.

The following lemma shows that sequence {xkx~} is descending and convergent, where sequence {xk} is generated by Algorithm 2.1 and point x~ is an arbitrary solution to Problem (1.1). The proof process of this lemma can be found in Reference [17].

Lemma 3.2  Assuming that mapping F satisfies (H) and sequence {xk} and {zk} are generated by Algorithm 2.1, then for any solution x~ to Problem (1.1), there is

xk+1x~2xkx~2xk+1xk2.

Especially, the sequence {xk} is bounded and

limkxk+1xk=0.

Note 3.1 According to (2.2) and (2.3), we know that

xk+1xk=αk|FT(zk)dk|F(zk)2F(zk)σαkdk2.

According to (3.4), we know that

limkαkdk=0.

Lemma 3.3  Assuming that the mapping F satisfies (H), the sequences{dk} and {Fk} are generated by Algorithm 2.1, then there exists a positive real number M such that

dkM,k0.

Proof For a certain solution x~ of Problem (1.1) satisfies

Fk=F(xk)F(x~)Lxkx~.

Then, applying the boundedness of sequence {xk}, we can see that there is a normal number μ, which makes

Fkμ,k0.

By using (2.2) and (3.1), as well as the Cauchy-Schwartz inequality and the definition of step αk, it can be obtained that

ykLxk+1xkLF(zk)2xkzkF(zk)2=LαkdkLβdk.

According to (2.6), (3.7), (3.8) and Cauchy-Schwartz inequality, it can be obtained that

|βk+1|Fk+1ykdkTωk+tyk2Fk+1dk(dkTωk)2LβFk+1dkdk2+tL2β2dk2Fk+1dkdk4Lβμdk+tβ2L2μdk.

|θk+1|Fk+1dkdk2μdk.

Based on the above two inequalities and (2.4), (3.8), it can be concluded that

dk+1Fk+1+|βk+1|dk+|θk+1|(dk+yk)μ+(βLμ+tβ2L2μ)+μ(1+Lβ).

Since d0=F0, take M=μ+(βLμ+tβ2L2μ)+μ(1+Lβ), the conclusion is proved.□

Theorem 3.1  If the mapping F satisfies (H), the sequences {dk} and {Fk} are generated by Algorithm 2.1, then

liminfkFk=0.

Proof Assuming (3.9) doesn’t hold, there exists a normal number ε such that

Fkε,k0.

From (3.2), (3.6) and (3.7), it can be found that

αkdkβFk2(L+σF(xk+β1αk)dk)βε2(L+σμ)M>0.

Obviously, it is contradictory with (3.5). Therefore, the hypothesis does not hold and the original proposition holds.

4 R-Linear convergence rate

To further prove the R-linear convergence rate of Algorithm 2.1, it is necessary to assume that the mapping F satisfies:

(G) Let x~Ω, there exist positive constants ρ(0,1) and ν>0 such that

ρdist(x,Ω)Fk2,xN(x~,ν),

where Ω is the set of solutions of Problem (1.1),N(x~,ν)={xRn|xx~ν}dist(x,Ω) denotes the distance of point x from the solution set Ω.

Theorem 4.1  Suppose that the mapping F satisfies (H) and (G), the sequence {xk} is generated by Algorithm 2.1, then the sequence {dist(xk,Ω)} is Q-linearly convergent to 0. Thus, the sequence {xk} is R-linearly convergent.

Proof Let ωk=argmin{xkω|ωΩ}. we can know ωk is the solution closest to point xk in solution set Ω, i.e.,

dist(xk+1,Ω)=xkωk.

By using (2.5) and Cauchy-Schwartz inequality, it can be obtained that

dkFk.

Known that ωkΩ, using (3.3), (4.1)‒(4.3) and Note 3.1, we can get

dist(xk+1,Ω)2xk+1ωk2xkωk2xk+1xk2=dist(xk,Ω)2xk+1xk2dist(xk,Ω)2σ2αkdk2dist(xk,Ω)2σ2αk4Fk4(1σ2αk4ρ2)dist(xk,Ω)2.

This shows that the sequence {dist(xk,Ω)} is Q-linearly convergent to 0. Since 1σ2αk4ρ2(0,1), the sequence {xk} is R-linearly convergent.

5 Numerical tests

Algorithm 2.1 is tested numerically below and compared numerically with the MPRP-based method [12] and the MHS-based method [18]. The program code is written using MATLAB 7.0, running on an Intel Core i5-4590 processor (quad-core, 3.30 GHz), 8.0 GB of internal memory, and the Windows 7 operating system. The selected test problems were taken as follows, from the literature [22, 3, 10], respectively.

Problem 5.1 [22]  F is defined by the following equation:

F(x)=A(x)+g(x)1,

where g(x)=(ex1,ex2,,exn)T and

A=(21121112).

Problem 5.2 [3]  F is defined by the following equation:

F(x)=A(x)1,

where

A=(2.5112.51112.5).

Problem 5.3 [10]  F is defined by the following equation:

F(x)=(F1(x),F2(x),,Fn(x))T,

where

F1(x)=x1+ecos(x1+x2n+1),Fi(x)=xi+ecos(xi1+xi+xi+1n+1),i=2,3,,n1.Fn(x)=xn+ecos(xn1+xnn+1).

The parameter value in Algorithm 2.1 is σ=0.01,t=2,β=0.5. The termination conditions for all programs are F(xk)1.0×105 or F(zk)1.0×105.

In numerical tests, the rand function in MATLAB software is used to randomly generate a fixed initial point within (1,0),(0,1),(1,1),(2,0),(0,2),(5,5),(10,10), and calculate three test problems. The numerical results are shown in Tab.1‒Tab.7. In Tab.1‒Tab.7, “Dim” represents the dimension of the test problem, “NI” represents the number of iterations, and “NF” represents the number of calculations for F, “CPU ” represents CPU time (in seconds).

Tab.1‒Tab.7 indicate that Algorithm 2.1, PRP-based method, and HS-based method can successfully solve the given test problem starting from randomly generated initial points. For Problems 5.1 and 5.2, the numerical effects of the three test methods are comparable; However, for Problem 5.3, the performance of Algorithm 2.1 is significantly better than that of PRP-based and HS-based methods.

References

[1]

Ahookhosh M, Amini K, Bahrami S. Two derivative-free projection approaches for systems of largescale nonlinear monotone equations. Numer Algorithms 2013; 64(1): 21–42

[2]

Andrei N. On three-term conjugate gradient algorithms for unconstrained optimization. Appl Math Comput 2013; 219: 6316–6317

[3]

Cheng W Y. A PRP type method for systems of monotone equations. Math Comput Modelling 2009; 50: 15–20

[4]

Dennis J E, More J J. A characterization of superlinear convergence and its application to quasi-Newton methods. Math Comp 1974; 28(126): 549–560

[5]

Dennis J E, More J J. Quasi-Newton method, motivation and theory. SIAM Rev 1997; 19(1): 46–89

[6]

Dirkse S P, Ferris M C. MCPLIB: A collection of nonlinear mixed complementarity problems. Optim Methods Softw 2002; 5(4): 319–345

[7]

Hager W W, Zhang H. A new conjugate gradient method with guaranteed descent and an efficient line search. SIAM J Optim 2005; 16(1): 170–192

[8]

Hestens M R, Stiefel E L. Methods of conjugate gradients for solving linear systems. J Research Nat Bur Standards 1952; 49(6): 409–436

[9]

Iusem A N, Solodov M V. Newton-type methods with generalized distances for constrained optimization. Optimization 1997; 41(3): 257–278

[10]

La Cruz M, Martinez J M, Raydan M. Spectral residual method without gradient information for solving large-scale nonlinear systems of equations. Math Comp 2006; 75(255): 1429–1448

[11]

Li D H, Fukushima M. A global and superlinear convergent Gauss-Newton-based BFGS method for symmetric nonlinear equations. SIAM J Numer Anal 1999; 37(1): 152–172

[12]

Li Q N, Li D H. A class of derivative-free methods for large-scale nonlinear monotone equations. IMA J Numer Anal 2011; 31(4): 1625–1635

[13]

Liu J K, Li S J. A projection method for convex constrained monotone nonlinear equations with applications. Comput Math Appl 2015; 70(10): 2442–2453

[14]

Meintjes K, Morgan A P. A methodology for solving chemical equilibrium systems. Appl Math Comput 1987; 22(4): 333–361

[15]

Polyak B T. The conjugate gradient method in extremal problems. USSR Comput Math and Math Phys 1969; 9(4): 94–112

[16]

Polak E, Ribière G. Note sur la convergence de méthodes de directions conjuguées. Rev Francaise Informat Recherche Opérationnelle 1969; 3(16): 35–43

[17]

SolodovM VSvaiterB F. A globally convergent inexact Newton method for systems of monotone equations. In: Reformulation: Non-smooth, Piecewise Smooth, Semismooth and Smoothing Methods. Dordrecht: Springer, 1999, 355–369

[18]

Wu X Y, Sai N E Z, Zhang H L. On three-term HS projection algorithm for solving nonlinear monotone equations. Southwest China Normal Univ Natur Sci Ed 2016; 41(5): 41–47

[19]

Xiao Y H, Zhu H. A conjugate gradient method to solve convex constrained monotone equations with applications in compressive sensing. J Math Anal Appl 2013; 405(1): 310–319

[20]

Zhao Y B, Li D H. Monotonicity of fixed point and normal mapping associated with variational inequality and its application. SIAM J Optim 2000; 11(4): 962–973

[21]

Zhou G, Toh K C. Superline convergence of a Newton-type algorithm for monotone equations. J Optim Theory Appl 2005; 125(1): 205–221

[22]

Zhou W J, Li D H. Limited memory BFGS method for nonlinear monotone equations. J Comput Math 2007; 25(1): 89–96

[23]

Zhou W J, Li D H. A globally convergent BFGS method for nonlinear monotone equations without any merit functions. Math Comp 2008; 77(264): 2231–2240

RIGHTS & PERMISSIONS

Higher Education Press 2023

AI Summary AI Mindmap
PDF (410KB)

681

Accesses

0

Citation

Detail

Sections
Recommended

AI思维导图

/