Processing math: 100%
Research article

Convergence properties of a family of inexact Levenberg-Marquardt methods

  • Received: 17 April 2023 Revised: 13 May 2023 Accepted: 18 May 2023 Published: 02 June 2023
  • MSC : 90C33, 65K05

  • We present a family of inexact Levenberg-Marquardt (LM) methods for the nonlinear equations which takes more general LM parameters and perturbation vectors. We derive an explicit formula of the convergence order of these inexact LM methods under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian. Moreover, we develop a family of inexact LM methods with a nonmonotone line search and prove that it is globally convergent. Numerical results for solving the linear complementarity problem are reported.

    Citation: Luyao Zhao, Jingyong Tang. Convergence properties of a family of inexact Levenberg-Marquardt methods[J]. AIMS Mathematics, 2023, 8(8): 18649-18664. doi: 10.3934/math.2023950

    Related Papers:

    [1] Dingyu Zhu, Yueting Yang, Mingyuan Cao . An accelerated adaptive two-step Levenberg–Marquardt method with the modified Metropolis criterion. AIMS Mathematics, 2024, 9(9): 24610-24635. doi: 10.3934/math.20241199
    [2] Lin Zheng, Liang Chen, Yanfang Ma . A variant of the Levenberg-Marquardt method with adaptive parameters for systems of nonlinear equations. AIMS Mathematics, 2022, 7(1): 1241-1256. doi: 10.3934/math.2022073
    [3] Xiaorui He, Jingyong Tang . A smooth Levenberg-Marquardt method without nonsingularity condition for wLCP. AIMS Mathematics, 2022, 7(5): 8914-8932. doi: 10.3934/math.2022497
    [4] Linsen Song, Gaoli Sheng . A two-step smoothing Levenberg-Marquardt algorithm for real-time pricing in smart grid. AIMS Mathematics, 2024, 9(2): 4762-4780. doi: 10.3934/math.2024230
    [5] Panjie Tian, Zhensheng Yu, Yue Yuan . A smoothing Levenberg-Marquardt algorithm for linear weighted complementarity problem. AIMS Mathematics, 2023, 8(4): 9862-9876. doi: 10.3934/math.2023498
    [6] Iftikhar Ahmad, Hira Ilyas, Muhammad Asif Zahoor Raja, Tahir Nawaz Cheema, Hasnain Sajid, Kottakkaran Sooppy Nisar, Muhammad Shoaib, Mohammed S. Alqahtani, C Ahamed Saleel, Mohamed Abbas . Intelligent computing based supervised learning for solving nonlinear system of malaria endemic model. AIMS Mathematics, 2022, 7(11): 20341-20369. doi: 10.3934/math.20221114
    [7] Xiangtuan Xiong, Wanxia Shi, Xuemin Xue . Determination of three parameters in a time-space fractional diffusion equation. AIMS Mathematics, 2021, 6(6): 5909-5923. doi: 10.3934/math.2021350
    [8] Miao Guo, Qingbiao Wu . Two effective inexact iteration methods for solving the generalized absolute value equations. AIMS Mathematics, 2022, 7(10): 18675-18689. doi: 10.3934/math.20221027
    [9] Khalil Ur Rehman, Wasfi Shatanawi, Zead Mustafa . Artificial intelligence (AI) based neural networks for a magnetized surface subject to tangent hyperbolic fluid flow with multiple slip boundary conditions. AIMS Mathematics, 2024, 9(2): 4707-4728. doi: 10.3934/math.2024227
    [10] Khalil Ur Rehman, Wasfi Shatanawi, Zeeshan Asghar, Haitham M. S. Bahaidarah . Neural networking analysis for MHD mixed convection Casson flow past a multiple surfaces: A numerical solution. AIMS Mathematics, 2023, 8(7): 15805-15823. doi: 10.3934/math.2023807
  • We present a family of inexact Levenberg-Marquardt (LM) methods for the nonlinear equations which takes more general LM parameters and perturbation vectors. We derive an explicit formula of the convergence order of these inexact LM methods under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian. Moreover, we develop a family of inexact LM methods with a nonmonotone line search and prove that it is globally convergent. Numerical results for solving the linear complementarity problem are reported.



    Consider the system of nonlinear equations

    F(x)=0, (1.1)

    where F(x):RnRn is a continuously differentiable function. In the paper, we assume that the solution set of (1.1) is nonempty and denote it by X. Moreover, we denote the Jacobian F(x) as J(x) and use the notations Fk=F(xk),Jk=J(xk) for simplification.

    The Levenberg-Marquardt (LM) method is one of the most important algorithms for solving (1.1). At every iteration, the LM method computes the trial step dk by solving the following linear system

    (JTkJk+μkI)dk=JTkFk, (1.2)

    where μk is the LM parameter which plays an important role in analyzing the convergence rate of the LM method. For example, Yamashita and Fukushima [13] proved that the LM method taking μk=Fk2 has quadratic convergence under the local error bound condition which is weaker than the nonsingularity. Fan and Yuan [6] proved that the LM method taking μk=Fkδ with δ[1,2] still achieves the quadratic convergence under the local error bound condition. More researches on the LM method can be found in [1,11,14,15,16] and references therein.

    The LM method solves the linear system (1.2) exactly at every iteration which may be very expensive when solving a large-scale nonlinear equation. The inexact approach is one way to overcome this difficulty. In the inexact LM method, the direction dk is given by the solution of the system

    (JTkJk+μkI)dk=JTkFk+pk, (1.3)

    where pkRn is a perturbation vector which measures how inexactly the linear system (1.2) is solved. Under the nonsingularity, Facchinei and Kanzow [3] proved that if μk0 and pko(JTkFk), then the inexact LM method has superlinear convergence rate and if μk=O(JTkFk) and pk=O(JTkFk2), then its convergence rate is quadratic. Suppose

    μk=O(Fkα)andpk=O(Fkα+θ),

    where α>0 and θ>0 are constants. Under the local error bound condition, many researchers (e.g., [2,4,5,7]) investigated the convergence rate of the inexact LM method for different values of α and θ respectively. Lately, Wang and Fan [12] studied the convergence rate of the inexact LM method taking μk=Fkα with pk=Fkα+θ and μk=JTkFkα with pk=JTkFkα+θ respectively under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian, which are more general than the local error bound condition and the Lipschitz continuity of the Jacobian used in [1,2,4,5,7].

    In this paper, we study the convergence rate of a family of inexact LM methods with more general LM parameters and perturbation vectors. We consider

    μk=σFkα+(1σ)JTkFkα, (1.4)
    pk=τFkα+θ+(1τ)JTkFkα+θ, (1.5)

    where σ,τ[0,1] and α,θ>0. We derive an explicit formula of the convergence order under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian. Moreover, we develop a family of inexact LM methods with a nonmonotone line search and prove its global convergence. We also investigate the numerical performances of these inexact LM methods by solving the nonlinear equations arising in the linear complementarity problem.

    The organization of this paper is as follows. In the next section, we investigate the convergence order under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian. In Section 3, we present a family of inexact LM methods with a nonmonotone line search and prove that it is globally convergent. In Section 4, we apply these inexact LM methods to solve the nonlinear equations arising from the linear complementarity problem and report some numerical results.

    In this section, we study the convergence rate of the inexact LM methods with the iteration

    xk+1=xk+dk, (2.1)

    where dk is obtained by (1.3) with μk,pk satisfying (1.4) and (1.5). We suppose that the generated sequence {xk} converges to the solution set X and lies in some neighbourhoods of xX.

    Assumption 2.1. (a) F(x) provides a H¨oderian local error bound of order γ(0,1] in some neighbourhoods of xX, i.e., there exist constants κ>0 and 0<r<1 such that

    κdist(x,X)F(x)γ,xN(x,r)={xRn|xxr}. (2.2)

    (b) J(x) is H¨oderian continuous of order υ(0,1], i.e., there exists a constant L>0 such that

    J(x)J(y)Lxyυ,x,yN(x,r). (2.3)

    It is worth pointing out that if γ=υ=1, then Assumption 2.1 (a) is the local error bound condition and Assumption 2.1 (b) is the Lipschitz continuity of the Jacobian. Moreover, by (2.3), we have (see [12])

    F(y)F(x)J(x)(yx)L1+υyx1+υ,x,yN(x,r). (2.4)

    Furthermore, there exists a constant M>0 such that

    F(y)F(x)Myx,x,yN(x,r). (2.5)

    In the following, we denote by ¯xk the vector in X that is closest to xk, i.e.,

    ¯xkxk=dist(xk,X). (2.6)

    Now we suppose the singular value decomposition (SVD) of J(ˉxk) is

    J(ˉxk)=ˉUkˉΣkˉVTk=(ˉUk,1,ˉUk,2)(ˉΣk,10)(ˉVTk,1ˉVTk,2)=ˉUk,1ˉΣk,1ˉVTk,1,

    where ˉΣk,1=diag(ˉσk,1,...,ˉσk,r) with ˉσk,1ˉσk,r>0. The corresponding SVD of Jk is

    Jk=UkΣkVTk=(Uk,1,Uk,2)(Σk,1Σk,2)(VTk,1VTk,2)=Uk,1Σk,1VTk,1+Uk,2Σk,2VTk,2,

    where Σk,1=diag(σk,1,...,σk,r) with σk,1σk,r>0 and Σk,2=diag(σk,r+1,...,σk,n) with σk,r+1σk,n0. For simplicity, we neglect the subscript k in Uk,i,Σk,i,Vk,i(i=1,2) and write J(ˉxk) and Jk as

    J(ˉxk)=ˉU1ˉΣ1ˉVT1,Jk=U1Σ1VT1+U2Σ2VT2.

    By the matrix perturbation theory [8] and (2.3), we have

    diag(Σ1ˉΣ1,Σ2)J(ˉxk)JkLˉxkxkυ,

    which gives

    Σ1ˉΣ1Lˉxkxkυ,Σ2Lˉxkxkυ. (2.7)

    Moreover, by (1.3) we have

    dk=(JTkJk+μkI)1JTkFk+(JTkJk+μkI)1pk=V1(Σ21+μkI)1Σ1UT1FkV2(Σ22+μkI)1Σ2UT2Fk+V1(Σ21+μkI)1VT1pk+V2(Σ22+μkI)1VT2pk, (2.8)

    and

    Fk+Jkdk=FkJk(JTkJk+μkI)1JTkFk+Jk(JTkJk+μkI)1pk=μkU1(Σ21+μkI)1UT1Fk+μkU2(Σ22+μkI)1UT2Fk+U1Σ1(Σ21+μkI)1VT1pk+U2Σ2(Σ22+μkI)1VT2pk. (2.9)

    In the following, we suppose without loss of generality that xk lies in N(x,r2).

    Lemma 2.1. Under the conditions of Assumption 2.1, if υ>2γ2, then there exist positive constants a1,a2,b1,b2 such that

    a1¯xkxk(2γ1)αμka2¯xkxkα, (2.10)
    b1¯xkxk(2γ1)(α+θ)pkb2¯xkxkα+θ. (2.11)

    Proof. Since ¯xkx¯xkxk+xkx2xkxr, we have ¯xkN(x,r). Hence, by (2.2) and (2.5),

    κ1γ¯xkxk1γFkM¯xkxk. (2.12)

    By (2.5), we have

    JTkFkJkFkF(¯xk)M2¯xkxk. (2.13)

    Let Tk:=FkF(¯xk)Jk(xk¯xk). Then,

    FTkJk(xk¯xk)=Fk2FTkTk. (2.14)

    It follows from (2.4), (2.12) and (2.14) that

    FTkJk¯xkxkFk2FTkTkFk2FkFkF(¯xk)Jk(xk¯xk)κ2γ¯xkxk2γLM1+υ¯xkxk2+υ,

    which together with υ>2γ2 gives

    FTkJkκ2γ¯xkxk2γ1LM1+υ¯xkxk1+υ(κ2γLM1+υ¯xkxk2+υ2γ)¯xkxk2γ1˜c¯xkxk2γ1, (2.15)

    where ˜c>0 is some constant. By (2.13) and (2.15), we have

    ˜c¯xkxk2γ1JTkFkM2¯xkxk. (2.16)

    Since μk=σFkα+(1σ)JTkFkα, by (2.12) and (2.16), we have

    a1¯xkxkmax{αγ,(2γ1)α}μka2¯xkxkα,

    where a1:=σκαγ+(1σ)˜cα and a2:=σMα+(1σ)M2α, which together with 2γ11γ gives

    a1¯xkxk(2γ1)αμka2¯xkxkα.

    This proves (2.10). Moreover, since pk=τFkα+θ+(1τ)JTkFkα+θ, according to (2.12) and (2.16), we have

    b1¯xkxkmax{α+θγ,(2γ1)(α+θ)}pkb2¯xkxkα+θ,

    where b1:=τκα+θγ+(1τ)˜cα+θ and b2:=τMα+θ+(1τ)M2(α+θ), which together with 2γ11γ gives

    b1¯xkxk(2γ1)(α+θ)pkb2¯xkxkα+θ.

    This proves (2.11).

    Lemma 2.2. Under the conditions of Assumption 2.1, if υ>2γ2, 0<α<2γ(1+υ)2γ and θ>(22γ)αγ, there exists a constant c>0 such that

    dkc¯xkxkmin{1,1+υ(2γ)α2γ,(2γ2)αγ+θ}. (2.17)

    Proof. Let

    ˉdk:=(JTkJk+μkI)1JTkFk. (2.18)

    Then ˉdk is the LM step computed by solving (1.2). Moreover, by (1.3) we have

    dk=ˉdk+(JTkJk+μkI)1pk. (2.19)

    Now we define

    φk(d):=Fk+Jkd2+μkd2. (2.20)

    Then, the LM step ¯dk defined by (2.18) is the minimizer of φk(d). By (2.4) and the left inequality in (2.10), we have

    ¯dk2φk(¯dk)μkφk(¯xkxk)μk=Fk+Jk(¯xkxk)2μk+¯xkxk2(L1+υ)2a11¯xkxk2+2υ(2γ1)α+¯xkxk2c1¯xkxk2min{1+υαγ+α2,1}, (2.21)

    where c1:=((L/(1+υ))2a11+1). Thus, by (2.19), (2.21) and the left inequality in (2.10) and the right inequality in (2.11), we have

    dk¯dk+dk¯dk=¯dk+(JTkJk+μkI)1pk¯dk+pkμkc1¯xkxkmin{1+υαγ+α2,1}+b2a1¯xkxkα+θ(2γ1)αc¯xkxkmin{1,1+υ(2γ)α2γ,(2γ2)αγ+θ}, (2.22)

    where c=c1+b2/a1.

    Lemma 2.3. [12, Lemma 2.3] Under the conditions of Assumption 2.1, we have

    (i)U1UT1FkM¯xkxk; (ii)U2UT2Fk2L¯xkxk1+υ.

    Theorem 2.1. Under the conditions of Assumption 2.1, if υ>2γ2, 0<α<2γ(1+υ)2γ and θ>(22γ)αγ, then the sequence {xk} converges to the solution set X with the order h(α,θ,γ,υ) where

    h(α,θ,γ,υ)=γmin{1+α,1+υ,α+θ,(2γ2)αγ+θ+υ,(1+υ)(1+υ(2γ)α2γ),(1+υ)((2γ2)αγ+θ)}. (2.23)

    Proof. Since xk converges to X, we assume that L¯xkxkvˉσ2 holds for all sufficiently large k. Then, it follows from (2.7) that

    (Σ21+μkI)1Σ211(ˉσL¯xkxkv)2<4ˉσ2. (2.24)

    On the other hand, by (2.7) and the left inequality in (2.10), for all sufficiently large k,

    Σ2(Σ22+μkI)1Σ2μkLa11¯xkxkυ(2γ1)α. (2.25)

    Hence, it follows from (2.9)–(2.11), (2.24), (2.25), (Σ22+μkI)1μ1k and Lemma 2.3 that

    Fk+JkdkμkU1(Σ21+μkI)1UT1Fk+μkU2(Σ22+μkI)1UT2Fk+U1Σ1(Σ21+μkI)1VT1pk+U2Σ2(Σ22+μkI)1VT2pk4Ma2ˉσ2¯xkxk1+α+2L¯xkxk1+υ+2ˉσb2¯xkxkα+θ+La11b2¯xkxk(2γ2)αγ+θ+υˉc¯xkxkmin{1+α,1+υ,α+θ,(2γ2)αγ+θ+υ}, (2.26)

    where ˉc=4Ma2/ˉσ2+2L+2ˉσb2+La11b2. Therefore, by (2.2), (2.4), (2.26) and Lemma 2.2, we have

    ¯xk+1xk+11κFk+1γ1κ(Fk+Jkdk+L1+υdk1+υ)γ1κ(ˉc¯xkxkmin{1+α,1+υ,α+θ,(2γ2)αγ+θ+υ}+L1+υc1+υ¯xkxkmin{1+υ,(1+υ)(1+υ(2γ)α2γ),(1+υ)((2γ2)αγ+θ)})γO(¯xkxkh(α,θ,γ,υ),

    where h(α,θ,γ,υ) is given in (2.23).

    Corollary 2.1. Under the conditions of Assumption 2.1 and γ=υ=1, if 0<α<4, then the sequence {xk} converges to the solution set X with the order h(α,θ) where

    h(α,θ)={min{α+θ,4α,2θ},if0<θ<1,min{2,1+α,4α},ifθ1.

    More precisely,

    h(α,θ)={α+θifα(0,θ],2θifα(θ,42θ],if0<θ<1,4αifα(42θ,4],

    and

    h(α,θ)={1+αifα(0,1],2ifα(1,2],ifθ1,4αifα(2,4].

    As we know from [5] that for any α(0,2] and θ1, the sequence generated by the inexact LM method (1.3) converges to some solution of (1.1), which is a stronger result than the convergence to the solution set. We show that this convergence result also holds true for the inexact LM methods studied in this paper.

    Theorem 2.2. Under the conditions of Assumption 2.1 and γ=υ=1, if α(0,2] and θ1, then the sequence {xk} converges to a solution of (1.1) with the order g(α) where

    g(α)={1+αifα(0,1],2ifα(1,2]. (2.27)

    Proof. By Corollary 2.1, when α(0,2] and θ1, it holds that

    ¯xk+1xk+1O(¯xkxkg(α)), (2.28)

    where g(α) is defined by (2.27). Then, for all sufficiently large k,

    ˉxkxkxkˉxk+1xk+1ˉxk+1+dkO(¯xkxkg(α))+dk,

    which together with g(α)>1 gives

    ˉxkxk2dk. (2.29)

    Hence, we deduce from (2.17), (2.28) and (2.29) that

    dk+1=O(ˉxk+1xk+1min{1,2α2,θ})=O(ˉxk+1xk+1)=O(ˉxkxkg(α))=O(dkg(α)).

    This implies that the sequence {xk} converges to some solution of (1.1) with the order g(α).

    In this section, we study a family of globally convergent inexact LM methods. We define the merit function ψ:RnR as

    ψ(x):=12F(x)2. (3.1)

    Obviously, ψ(x) is continuously differentiable at any xRn with ψ(x)=J(x)TF(x). Our method is described as follows.

    Algorithm 3.1. Choose parameters σ,τ[0,1], α(0,4], θ>0, ρ,ξ,χ,δ,ζ(0,1) and x0Rn. Let Θ0:=ψ(x0). Set k:=0.

    Step 1: If ψ(xk)=0, then stop.

    Step 2: Set

    μk:=σFkα+(1σ)JTkFkα, (3.2)
    wk:=τFkα+θ+(1τ)JTkFkα+θ. (3.3)

    Find a search direction dkRn which satisfies

    (JTkJk+μkI)dk=ψ(xk)+pk, (3.4)

    where

    pkmin{ρψ(xk),wk}. (3.5)

    If dk satisfies

    F(xk+dk)ξF(xk), (3.6)

    then set λk:=1 and go to Step 4.

    Step 3: If the descent condition

    ψ(xk)Tdkχdk2 (3.7)

    is not satisfied, then set dk:=ψ(xk). Let lk be the smallest nonnegative integer l satisfying

    ψ(xk+δldk)Θkζδldk2. (3.8)

    Set λk:=δlk and go to Step 4.

    Step 4: Set xk+1:=xk+λkdk and

    Θk+1:=(Θk+1)ψ(xk+1)ψ(xk+1)+1. (3.9)

    Set k:=k+1 and go to Step 1.

    Algorithm 3.1 is designed based on the inexact LM method [2] and the nonmonotone smoothing Newton method [10]. The main feature of Algorithm 3.1 is that it takes more general LM parameter μk and perturbation vector pk and adopts a nonmonotone line search technique to ensure the global convergence.

    Theorem 3.1. Algorithm 3.1 generates an infinite sequence {xk} which satisfies ψ(xk)Θk for all k0.

    Proof. For some k, we assume that ψ(xk)Θk. If ψ(xk)0, then F(xk)0 and hence μk>0. So, the matrix JTkJk+μkI is positive definite and the search direction dk in Step 2 is always obtained. Notice that the obtained dk0. In fact, if dk=0, then by (3.4) we have pkψ(xk)=0. Since pkρψ(xk), it follows that ψ(xk)=pk=0 which contradicts ψ(xk)0. So, in Step 3, if the descent condition (3.7) holds, then ψ(xk)Tdkχdk2<0. Otherwise, dk=ψ(xk) which gives ψ(xk)Tdk=ψ(xk)2<0. Thus, the direction dk used in the line search (3.8) is always a descent direction of ψ. Next we show that there exists at least a nonnegative integer l satisfying (3.8). On the contrary, we suppose that for all nonnegative integer l,

    ψ(xk+δldk)>Θkζδldk2,

    which together with ψ(xk)Θk gives

    ψ(xk+δldk)ψ(xk)δl+ζδldk2>0. (3.10)

    By letting l in (3.10), we have ψ(xk)Tdk0 which contradicts ψ(xk)Tdk<0. So, we can obtain xk+1 in Step 4. Now we show ψ(xk+1)Θk+1. In fact, if the condition (3.6) holds, then

    ψ(xk+1)ξ2ψ(xk)<ψ(xk)Θk.

    Otherwise, by Step 3, we also have ψ(xk+1)Θk. Hence, from (3.9) it holds that

    Θk+1=(Θk+1)ψ(xk+1)ψ(xk+1)+1(ψ(xk+1)+1)ψ(xk+1)ψ(xk+1)+1=ψ(xk+1).

    Therefore, we can conclude that if ψ(xk)Θk and ψ(xk)0 for some k, then xk+1 can be generated by Algorithm 2.1 with ψ(xk+1)Θk+1. Since ψ(x0)=Θ0, by induction on k, we prove the theorem.

    Theorem 3.2. Every accumulation point x of a sequence {xk} generated by Algorithm 3.1 is a stationary point of ψ(x), i.e., ψ(x)=0.

    Proof. By Steps 2 and 3, we have ψ(xk+1)Θk for all k0. This and (3.9) yield that

    Θk+1=Θkψ(xk+1)+ψ(xk+1)ψ(xk+1)+1Θkψ(xk+1)+Θkψ(xk+1)+1=Θk.

    Thus there exists Θ0 such that limkΘk=Θ. Further, by (3.9) we have

    limkψ(xk)=limkΘkΘk1Θk+1=Θ,

    and so

    limkF(xk)=2Θ. (3.11)

    So, if there are infinitely many k for which F(xk+dk)ξF(xk) holds, then 2Θξ2Θ which together with ξ(0,1) yields Θ=0, i.e., limkF(xk)=0. By the continuity, we have the desired result. Next, we assume that there exists an index ˉk such that F(xk+dk)>ξF(xk) for all kˉk, i.e., λk is determined by (3.8) for all kˉk. Since x is the accumulation point of {xk}, there exists a subsequence {xk}kK where K{0,1,...} such that lim(K)kxk=x. We assume that ψ(x)0 and will derive a contradiction. Since ψ(x)=J(x)TF(x)0, we have J(x)>0 and F(x)>0. Moreover, by the continuity, we have

    lim(K)kμk=σF(x)α+(1σ)J(x)TF(x)α:=μ.

    Obviously, μ>0. So, there exists a positive constant ˉμ such that μkˉμ>0 for all kK. Since {ψ(xk)} is bounded on any convergent subsequence {xk}kK, for any kK, either

    dk(JTkJk+μkI)1(ψ(xk)+pk)(1+ρ)ψ(xk)ˉμ<,

    or dk=ψ(xk)<. Hence, the sequence {dk}kK is bounded. By passing to the subsequence, we suppose lim(K1)kdk=d where K1K is an infinite subset. In the following, we prove ψ(x)Td=0. By (3.8) we have

    ζλ2kdk2Θkψ(xk+1).

    This and limkψ(xk)=limkΘk=Θ yield limkλkdk=0. Hence, if λkˉλ>0 for any kK1 where ˉλ>0 is a constant, then lim(K1)kdk=d=0 and hence ψ(x)Td=0. Otherwise, {λk}kK1 has a subsequence converging to zero and we suppose lim(K2)kλk=0 where K2K1 is an infinite set. From (3.8), for all kˉk and kK2,

    ψ(xk+δ1λkdk)>Θkζδ1λkdk2ψ(xk)ζδ1λkdk2,

    which gives

    ψ(xk+δ1λkdk)ψ(xk)δ1λk>ζδ1λkdk2. (3.12)

    Since ψ is continuously differentiable at x, by letting k with kK2 in (3.12), we have ψ(x)Td0. On the other hand, since dk is a sufficient descent direction of ψ, we have ψ(x)Td=lim(K2)kψ(xk)Tdk0. These give ψ(x)Td=0. Hence, we can conclude that ψ(x)Td=0. Let ˉK:={kK1|dk=ψ(xk)}. If ˉK is an infinite set, then

    ψ(x)2=lim(ˉK)kψ(xk)2=lim(ˉK)kψ(xk)Tdk=ψ(x)Td=0,

    which contradicts the assumption ψ(x)0. Otherwise, ˉK is a finite set and dk satisfies (3.7) for all sufficiently large kK1. Then, by (3.7) we have

    χd2=lim(K1)kχdk2lim(K1)kψ(xk)Tdk=ψ(x)Td=0,

    which gives d=0. By (3.4), we have for all kK1,

    pkψ(xk)JTkJk+μkIdk. (3.13)

    Since lim(K1)k(JTkJk+μkI)=J(x)TJ(x)+μI, by (3.13) and d=0, we have

    lim(K1)kpkψ(xk)=0. (3.14)

    Since pkρψ(xk), we can deduce from (3.14) that

    ψ(x)=lim(K1)kψ(xk)=lim(K1)kpk=0,

    which also contradicts the assumption ψ(x)0. We complete the proof.

    Next, we analyze the convergence rete of Algorithm 3.1. Suppose that the generated iteration sequence {xk} has an accumulation point x such that F(x)=0 and Assumption 2.1 holds at x for γ=υ=1. We will show that the whole sequence {xk} converges to x at leat superlinearly for any α(0,2] and θ>1.

    Lemma 3.1. Assume that Assumption 2.1 holds for γ=υ=1. Let α(0,2] and θ>1. If xk,xk+dkN(x,r/2), then there exists ˆc>0 such that

    dist(xk+dk,X)ˆcdist(xk,X)min{α2+1,θ}. (3.15)

    Proof. Since ¯dk defined by (2.18) is the minimizer of φk(d) in (2.20), by (2.4) and (2.10),

    φk(ˉdk)φk(ˉxkxk)=Fk+Jk(ˉxkxk)2+μkˉxkxk2L2/4ˉxkxk4+a2¯xkxkα+2(L2/4+a2)ˉxkxkα+2. (3.16)

    It holds from (2.20) and (3.16) that

    Fk+Jkˉdkφk(ˉdk)L2/4+a2ˉxkxkα2+1. (3.17)

    Since dk=ˉdk+(JTkJk+μkI)1pk, we have from (2.5), (2.10), (2.11) and (3.17) that

    Fk+Jkdk=Fk+Jkˉdk+Jk(JTkJk+μkI)1pkFk+Jkˉdk+MpkμkL2/4+a2ˉxkxkα2+1+Mb2a1¯xkxkθ˜C¯xkxkmin{α2+1,θ}, (3.18)

    where ˜C=L2/4+a2+Mb2a11. Moreover, by (2.17), α(0,2] and θ>1, we have

    dkc¯xkxk. (3.19)

    Thus, by (2.4), (2.17), (3.18) and (3.19), we have

    F(xk+dk)Fk+Jkdk+L/2dk2˜Cˉxkxkmin{α2+1,θ}+Lc2/2ˉxkxk2(˜C+Lc2/2)ˉxkxkmin{α2+1,θ},

    which together with (2.2) gives

    dist(xk+dk,X)κ1F(xk+dk)(˜C+Lc2/2)κ1dist(xk,X)min{α2+1,θ}.

    Letting ˆc:=(˜C+Lc2/2)κ1, we complete the proof.

    Lemma 3.2. Under the same conditions in Lemma 3.1, there exists an index ˉk such that for all kˉk it holds: (i) xk,xk+dkN(x,r/2); (ii) F(xk+dk)ξF(xk).

    Proof. By Lemma 3.1, similarly as the proof of [9, Lemma 11], we can prove the result.

    Theorem 3.3. Under the same conditions in Lemma 3.1, the whole sequence {xk} converges to x with

    xk+1x=O(xkxmin{α2+1,θ}).

    Proof. By Lemma 3.2 and Step 2 of Algorithm 3.1, we have xk+1=xk+dk and F(xk+1)ξF(xk) for all kˉk. It follows that limkF(xk)=0, which together with (2.2) yields limkdist(xk,X)=0. Thus, from Lemma 3.1, for all sufficiently large k,

    dist(xk+1,X)ˆcdist(xk,X)min{α2+1,θ}=ˆcdist(xk,X)min{α2+1,θ}1dist(xk,X).

    Since min{α2+1,θ}>1 and limkdist(xk,X)=0, we have ˆcdist(xk,X)min{α2+1,θ}1<12 for all sufficiently large k. It follows that for all sufficiently large k,

    dist(xk+1,X)12dist(xk,X).

    This implies that for all sufficiently large k,

    dist(xk,X)xkˉxk+1=xk+1ˉxk+1dkdist(xk+1,X)+dk12dist(xk,X)+dk,

    that is

    dist(xk,X)2dk. (3.20)

    By (3.19) and Lemma 3.1, for all sufficiently large k,

    dk+1cdist(xk+1,X)cˆcdist(xk,X)min{α2+1,θ}, (3.21)

    which together with limkdist(xk,X)=0 gives limkdk=0 and

    dk+114dist(xk,X). (3.22)

    By (3.20) and (3.22), for all sufficiently large k,

    dk+112dk. (3.23)

    So, when k is sufficiently large, (3.23) gives

    xk+1x=j=k+1djj=k+1dj2dk+1. (3.24)

    This and limkdk=0 yield limkxk=x. Further, by (3.21) and (3.24) we have

    limkxk+1xxkxmin{α2+1,θ}limk2dk+1dist(xk,X)min{α2+1,θ}<.

    The proof is completed.

    We apply Algorithm 3.1 to solve the nonlinear equations arising in the well-known linear complementarity problem (LCP):

    (LCP)u0,v0,u=Mv+q,uTv=0, (4.1)

    where MRn×n and qRn are given matrix and vector. To reformulate the LCP into an equivalent system of equations, we define the function ϕ:R2R as

    ϕ(a,b)=a2+b2sgn(a+b)(a+b)2,(a,b)R2, (4.2)

    where sgn(t):={1ift>00ift=0.1ift<0

    Proposition 4.1. (i) ϕ(a,b)=0a0,b0,ab=0.

    (ii) ϕ is continuously differentiable at any (a,b)R2 whose gradient is given by

    ϕ(a,b)=2[a|a+b|b|a+b|].

    Proof. Let f(t):=sgn(t)t2. Since fq is a bijective function, it follows that

    ϕ(a,b)=0f(a2+b2)f(a+b)=0f(a2+b2)=f(a+b)a2+b2=a+ba0,b0,ab=0.

    The result (ⅱ) holds because f(t) is continuously differentiable everywhere with f(t)=2|t|.

    Let x:=(u,v). By using the function ϕ, we may have that solving the LCP is equivalent to computing a solution of the smooth nonlinear equations

    F(x)=(Mv+quϕ(u1,v1)ϕ(un,vn))=0. (4.3)

    In the following, we apply Algorithm 3.1 to solve the nonlinear equations (4.3). The parameters are chosen as ρ=103,ξ=0.5,γ=105,ζ=105,δ=0.8,θ=1,τ=0.5 and σ,α are given the specific experiments. In Step 2, GMRES is used as the linear solver to find the inexact direction dk. Moreover, we use F(xk)105 as the stopping criterion.

    We test two classes of LCPs defined as follows:

    LCP (ⅰ) Let M be the block diagonal matrix with NT1N1NT1N1,...,NT4N4NT4N4 as block diagonals, i.e., M=diag(NTiNiNTiNi) with Ni=rand(n4,n4) for i=1,...,4. Take q=rand(n,1). Obviously, the matrix M is positive semidefinite.

    LCP (ⅱ) Let M=diag(NiNieye(n/4)) with Ni=rand(n4,n4) for i=1,...,4. Take q=rand(n,1).

    We use v0=(1,0,...,0)T and u0=Mv0+q as the initial point. Tables 1 and 2 show numerical experimental results of Algorithm 3.1 with different values of σ and α, in which IT denotes the iteration number, CPU denotes the CPU time in seconds, Fx denotes the value of F(xk) at the final iteration point and "–" stands for that the algorithm fails to find the solution. These numerical results show that Algorithm 3.1 is efficient for solving LCPs. It can find a solution point meeting the desired accuracy in very few iteration numbers and in short CPU time. Moreover, from our numerical implementations, we may find that Algorithm 3.1 with σ=1, i.e., μk=Fkα, has advantage over it with σ=0, i.e., μk=JTkFkα. At last, we point out that we have tested Algorithm 3.1 with different values of τ and found that the numerical performances are same.

    Table 1.  Numerical results for LCP (ⅰ).
    σ=0 σ=0.5 σ=1
    α n IT CPU Fx IT CPU Fx IT CPU Fx
    1 1000 7 2.94 1.176e-06 7 2.86 1.668e-06 6 2.58 6.607e-06
    1300 7 5.65 2.905e-07 7 5.65 3.887e-07 6 4.56 2.253e-06
    1500 7 7.86 8.645e-08 7 7.58 1.247e-07 6 6.84 3.491e-06
    1700
    2000 6 13.93 9.715e-06 6 13.72 6.021e-06 6 13.59 1.957e-06
    2500 7 28.52 2.317e-07 6 23.35 7.880e-06 6 23.41 1.507e-06
    2 1000 6 2.73 1.753e-06 6 2.44 3.739e-07
    1300 6 4.76 1.936e-06 6 4.55 3.479e-09
    1500 7 8.62 2.622e-06 7 8.02 5.073e-10 5 5.75 8.161e-06
    1700
    2000 7 15.30 1.478e-07 7 16.38 2.749e-11 5 11.40 8.913e-06
    2500 6 24.52 1.018e-06 6 24.18 7.721e-10

     | Show Table
    DownLoad: CSV
    Table 2.  Numerical results for LCP (ⅱ).
    σ=0 σ=0.5 σ=1
    α n IT CPU Fx IT CPU Fx IT CPU Fx
    1 1000 5 2.77 2.895e-06 5 2.17 4.086e-06 5 2.55 4.250e-06
    1300 5 4.02 1.334e-07 5 4.11 5.072e-07 4 3.19 9.426e-06
    1500 4 4.82 3.402e-08 3 3.50 8.312e-06 3 3.34 5.988e-06
    1700 5 7.72 4.276e-06 5 7.67 2.346e-06 5 7.87 1.505e-06
    2000 4 10.24 5.374e-07 4 9.72 3.970e-07 3 8.73 9.130e-06
    2500 5 21.22 2.845e-07 5 20.35 4.013e-07 4 16.23 8.123e-06
    ~~
    2 1000 4 1.77 3.116e-06 4 1.88 3.262e-06 4 1.82 2.224e-06
    1300 4 3.27 2.005e-07 4 3.26 1.156e-07 3 3.52 1.381e-08
    1500 3 3.56 5.093e-08 3 3.42 3.387e-08 3 3.75 1.226e-08
    1700 4 6.02 7.728e-06 4 6.23 1.625e-06 4 6.27 1.532e-07
    2000 3 7.15 1.553e-07 3 7.18 8.826e-08 3 7.40 4.989e-08
    2500 4 16.81 1.474e-07 4 16.52 1.299e-07 4 16.78 1.053e-07

     | Show Table
    DownLoad: CSV

    We have presented a family of inexact Levenberg-Marquardt methods for the nonlinear equations. The presented LM method takes more general LM parameters and perturbation vectors which are convex combinations of Fkα and JTkFkα and Fkα+θ and JTkFkα+θ. Under the H¨oderian local error bound condition and the H¨oderian continuity of the Jacobian, we have derived an explicit formula of the convergence order of these inexact LM methods. Moreover, we have developed a family of globally convergent inexact LM methods and showed its effectiveness by some numerical experiments.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    Research of this paper was supported by Natural Science Foundation of Henan Province (222300420520) and Key scientific research projects of Higher Education of Henan Province (22A110020).

    All authors declare no conflicts of interest in this paper.



    [1] K. Amini, F. Rostami, G. Caristi, An efficient Levenberg-Marquardt method with a new LM parameter for systems of nonlinear equations, Optimization, 67 (2018), 637–650. https://doi.org/10.1080/02331934.2018.1435655 doi: 10.1080/02331934.2018.1435655
    [2] H. Dan, N. Yamashita, M. Fukushima, Convergence properties of the inexact Levenberg-Marquardt method under local error bound, Optim. Method. Softw., 17 (2002), 605–626. http://dx.doi.org/10.1080/1055678021000049345 doi: 10.1080/1055678021000049345
    [3] F. Facchinei, C. Kanzow, A nonsmooth inexact Newton method for the solution of large scale nonlinear complementarity problems, Math. Program., 76 (1997), 493–512. https://doi.org/10.1007/bf02614395 doi: 10.1007/bf02614395
    [4] J. Y. Fan, J. Y. Pan, Inexact Levenberg-Marquardt method for nonlinear equations, Discrete Cont. Dyn.-B, 4 (2004), 1223–1232. http://dx.doi.org/10.3934/dcdsb.2004.4.1223 doi: 10.3934/dcdsb.2004.4.1223
    [5] J. Y. Fan, J. Y. Pan, On the convergence rate of the inexact Levenberg-Marquardt method, J. Ind. Manag. Optim., 7 (2011), 199–210. http://dx.doi.org/10.3934/jimo.2011.7.199 doi: 10.3934/jimo.2011.7.199
    [6] J. Y. Fan, Y. X. Yuan, On the quadratic convergence of the Levenberg-Marquardt method without nonsingularity assumption, Computing, 74 (2005), 23–39. https://doi.org/10.1007/s00607-004-0083-1 doi: 10.1007/s00607-004-0083-1
    [7] A. Fischera, P. K. Shuklaa, M. Wang, On the inexactness level of robust Levenberg-Marquardt methods, Optimization, 59 (2010), 273–287. https://doi.org/10.1080/02331930801951256 doi: 10.1080/02331930801951256
    [8] G. W. Stewart, J. G. Sun, Matrix Perturbation Theory, San Diego: Academic Press, 1990.
    [9] J. Y. Tang, J. C. Zhou, Quadratic convergence analysis of a nonmonotone Levenberg-Marquardt type method for the weighted nonlinear complementarity problem, Comput. Optim. Appl., 80 (2021), 213–244. http://dx.doi.org/10.1007/S10589-021-00300-8 doi: 10.1007/S10589-021-00300-8
    [10] J. Y. Tang, H. C. Zhang, A nonmonotone smoothing Newton algorithm for weighted complementarity problems, J. Optim Theory Appl., 189 (2021), 679–715. http://dx.doi.org/10.1007/S10957-021-01839-6 doi: 10.1007/S10957-021-01839-6
    [11] H. Y. Wang, J. Y. Fan, Convergence rate of the Levenberg-Marquardt method under Hölderian local error bound, Optim. Methods Softw., 35 (2020), 767–786. http://dx.doi.org/10.1080/10556788.2019.1694927 doi: 10.1080/10556788.2019.1694927
    [12] H. Y. Wang, J. Y. Fan, Convergence properties of inexact Levenberg-Marquardt method under Hölderian local error bound, J. Ind. Manag. Optim., 17 (2021), 2265–2275. http://doi.org/10.3934/jimo.2020068 doi: 10.3934/jimo.2020068
    [13] N. Yamashita, M. Fukushima, On the rate of convergence of the Levenberg-Marquardt method, Computing, 15 (2001), 239–249. http://doi.org/10.1007/978-3-7091-6217-0-18 doi: 10.1007/978-3-7091-6217-0-18
    [14] M. Zeng, G. Zhou, Improved convergence results of an efficient Levenberg-Marquardt method for nonlinear equations, J. Appl. Math. Comput., 68 (2022), 3655–367. http://doi.org/10.1007/S12190-021-01599-6 doi: 10.1007/S12190-021-01599-6
    [15] L. Zheng, L. Chen, Y. F. Ma, A variant of the Levenberg-Marquardt method with adaptive parameters for systems of nonlinear equations, AIMS Math., 7 (2021), 1241–1256. http://doi.org/10.3934/math.2022073 doi: 10.3934/math.2022073
    [16] L. Zheng, L. Chen, Y. X. Tang, Convergence rate of the modified Levenberg-Marquardt method under Hölderian local error bound, Open Math., 20 (2022), 998–1012. http://dx.doi.org/10.1515/MATH-2022-0485 doi: 10.1515/MATH-2022-0485
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1508) PDF downloads(60) Cited by(0)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog