Research article

An adaptive simple model trust region algorithm based on new weak secant equations

  • Received: 09 January 2024 Revised: 10 February 2024 Accepted: 22 February 2024 Published: 28 February 2024
  • MSC : 90C06, 90C30

  • In this work, we proposed a new trust region method for solving large-scale unconstrained optimization problems. The trust region subproblem with a simple form was constructed based on new weak secant equations, which utilized both gradient and function values and available information from the three most recent points. A modified Metropolis criterion was used to determine whether to accept the trial step, and an adaptive strategy was used to update the trust region radius. The global convergence and locally superlinearly convergence of the new algorithm were established under appropriate conditions. Numerical experiments showed that the proposed algorithm was effective.

    Citation: Yueting Yang, Hongbo Wang, Huijuan Wei, Ziwen Gao, Mingyuan Cao. An adaptive simple model trust region algorithm based on new weak secant equations[J]. AIMS Mathematics, 2024, 9(4): 8497-8515. doi: 10.3934/math.2024413

    Related Papers:

    [1] Yiting Zhang, Chongyang He, Wanting Yuan, Mingyuan Cao . A novel nonmonotone trust region method based on the Metropolis criterion for solving unconstrained optimization. AIMS Mathematics, 2024, 9(11): 31790-31805. doi: 10.3934/math.20241528
    [2] Dingyu Zhu, Yueting Yang, Mingyuan Cao . An accelerated adaptive two-step Levenberg–Marquardt method with the modified Metropolis criterion. AIMS Mathematics, 2024, 9(9): 24610-24635. doi: 10.3934/math.20241199
    [3] Ke Su, Wei Lu, Shaohua Liu . An area-type nonmonotone filter method for nonlinear constrained optimization. AIMS Mathematics, 2022, 7(12): 20441-20460. doi: 10.3934/math.20221120
    [4] B. El-Sobky, M. F. Zidan . A trust-region based an active-set interior-point algorithm for fuzzy continuous Static Games. AIMS Mathematics, 2023, 8(6): 13706-13724. doi: 10.3934/math.2023696
    [5] Yulin Cheng, Jing Gao . An efficient augmented memoryless quasi-Newton method for solving large-scale unconstrained optimization problems. AIMS Mathematics, 2024, 9(9): 25232-25252. doi: 10.3934/math.20241231
    [6] Bothina El-Sobky, Yousria Abo-Elnaga, Gehan Ashry . A nonmonotone trust region technique with active-set and interior-point methods to solve nonlinearly constrained optimization problems. AIMS Mathematics, 2025, 10(2): 2509-2540. doi: 10.3934/math.2025117
    [7] B. El-Sobky, G. Ashry, Y. Abo-Elnaga . An active-set with barrier method and trust-region mechanism to solve a nonlinear Bilevel programming problem. AIMS Mathematics, 2022, 7(9): 16112-16146. doi: 10.3934/math.2022882
    [8] B. El-Sobky, G. Ashry . An interior-point trust-region algorithm to solve a nonlinear bilevel programming problem. AIMS Mathematics, 2022, 7(4): 5534-5562. doi: 10.3934/math.2022307
    [9] Xuejie Ma, Songhua Wang . A hybrid approach to conjugate gradient algorithms for nonlinear systems of equations with applications in signal restoration. AIMS Mathematics, 2024, 9(12): 36167-36190. doi: 10.3934/math.20241717
    [10] Jun Yang, Prasit Cholamjiak, Pongsakorn Sunthrayuth . Modified Tseng's splitting algorithms for the sum of two monotone operators in Banach spaces. AIMS Mathematics, 2021, 6(5): 4873-4900. doi: 10.3934/math.2021286
  • In this work, we proposed a new trust region method for solving large-scale unconstrained optimization problems. The trust region subproblem with a simple form was constructed based on new weak secant equations, which utilized both gradient and function values and available information from the three most recent points. A modified Metropolis criterion was used to determine whether to accept the trial step, and an adaptive strategy was used to update the trust region radius. The global convergence and locally superlinearly convergence of the new algorithm were established under appropriate conditions. Numerical experiments showed that the proposed algorithm was effective.



    Consider the following unconstrained optimization problem

    minxRn f(x), (1.1)

    where f(x): RnR is a continuous differentiable function. The problem has penetrated deeply into various fields, such as aerospace, engineering technology, economics and finance, etc [1,2,3]. The trust region method is an important method for solving (1.1) and it has attracted the attention of many researchers [4,5,6]. The trust region methods usually compute a trial step sk by solving the following quadratic subproblem

    min q(k)(xk+s)=fk+gTks+12sTBks,s.t.sΔk, (1.2)

    where

    fk=f(xk),gk=f(xk),

    BkRn×n is the Hessian matrix of the function at the current iteration point xk or its symmetric approximation, s is the trial step, Δk>0 is the trust region radius, and stands for the Euclidean norm. The trust region methods take the ratio of the actual reduction to the predicted reduction

    rk=f(xk)f(xk+sk)q(k)(xk)q(k)(xk+sk)

    to decide whether to accept the trial step and how to adjust the trust region radius. If rk is close to 1, the trial step sk should be accepted and Δk can be increased. If rk is too small, the trial step sk should be rejected, Δk should be decreased, and the subproblem (1.2) should be resolved. If rk is much larger than 1, the case of 'too successful iteration' might occur. In addition, the trust region methods have always been accepted as effective methods for dealing with small and medium scale optimization problems due to the cost of computation and storage on the matrix B1k at each iteration. Many researchers [7,8,9,10,11,12,13,14,15,16] considered the modification of the trust region methods to adapt large scale optimization. We devote to the construction of subproblem and the adjustment of the trust region radius.

    In recent years, some trust region methods [11,12,13,14,15] based on simple models for solving large-scale optimization problems were proposed. For example, Sun et al. [13] developed a nonmonotone trust region algorithm with simple quadratic models, in which Hessian matrix in the subproblem is a diagonal positive definite matrix. Li et al. [14] proposed a simulated annealing-based trust region Bazilai-Borwein (BB) method and [15] proposed nonmonotone trust region BB methods. They all used scalar matrix with the reciprocal of the BB-stepsize to approximate the Hessian matrix of the objective function f(x). In the above methods, the amount of computation and storage is greatly reduced.

    The matrix Bk in subproblem (1.2) usually satisfies the classic secant equation (see [17])

    Bk+1sk=yk, (1.3)

    where

    sk=xk+1xk,yk=gk+1gk.

    Ebadi et al. [18] provided two new secant equations

    Bk+1¯sk=¯yk,¯sk=32sk12sk1,¯yk=yk13yk1+νk¯sk2¯sk, (1.4)

    where

    νk=2(fkfk+1)+¯sTk(43gk13gk1)+12(sk+sk1)Tgk+1,

    so

    Bk+1¯sk=zk,zk=yk13yk1+ηk¯sk2¯sk, (1.5)

    where

    ηk=2fk12fk132fk+1+νk.

    Constructing the approximation of the Hessian matrix based on the formulas (1.4) and (1.5) can make the algorithm maintain the third order curvature information of the objective function at the current iteration point, and it can make use of both gradient and function values and information from the three most recent points. It would improve the efficiency of the algorithm, which has attracted our attention. We try to introduce these two secant equations into the trust region algorithm.

    It is usually difficult to satisfy the secant Eq (1.3) with a nonsingular scalar matrix. Many researchers [16,19,20] considered some alternative conditions that could maintain the accumulated curvature information along the negative gradient. For example, Dennis and Wolkowicz [20] introduced a weaker form by projecting the secant Eq (1.3) in the direction sk as follows

    sTkBk+1sk=sTkyk. (1.6)

    Zhou et al. [16] considered some generalization of the weak secant Eq (1.6) and proposed a new simple model trust region method with generalized BB parameter. Inspired by the above work, we try to introduce two new weak secant equations with more information.

    Updating strategy of the trust region radius may significantly affect the number of iterations. Many researchers proposed adaptive trust region methods [21,22,23,24,25,26] to adjust the trust region radius. For example, Zhang et al. [24] proposed an adaptive trust region radius

    Δk=cpˆBk1gk,

    where c(0,1), p is a nonnegative integer, and

    ˆBk=Bk+iI

    is a positive definite matrix, for some iN. Under the same parameters, Shi et al. [25] proposed another adaptive trust region radius

    Δk=cpgTkqkqTkˆBkqkqk,

    with the vector parameter qkRn satisfying the angle condition

    gTkqkgkqko,

    where o(0,1). Rezaee and Babaie-Kafaki [26] proposed an adaptive choice for the trust region radius based on an eigenvalue analysis conducted on the scaled memoryless quasi-Newton updating formulas

    Δk=2cpgk{1,if k=0,sk12|sTk1yk1|,if k>0, (1.7)

    where

    sk1=xkxk1andyk1=gkgk1.

    The adaptive trust region radius does not use the Hessian matrix explicitly, and comparing the trust region method with the adaptive radius to some adaptive trust region methods, this method is low-cost to update the trust region radius.

    Note that some trust region methods require monotone reduction of the objective function, which may slow the convergence rate in the presence of a narrow curved valley. Nonmonotone trust region methods [14,27,28,29,30] were proposed. Li et al. [14] proposed a simulated annealing-based trust region BB method. Their nonmonotone strategy was defined by a modified Metropolis criterion, which can dynamically control the acceptance probability of the solution of the subproblem by introducing adaptive parameters (such as temperature parameters) into the accept-reject strategy. In the early stage of iteration, a higher acceptance probability can help the algorithm jump out of the local optimal solution by a larger extent, while in the later stage of iteration, the acceptance probability was gradually reduced to converge to a better solution.

    Our research aims to propose an adaptive simple model trust region algorithm based on new weak secant equations for solving large-scale optimization problems. The contributions of our work are listed as follows:

    Two new weak secant equations are introduced, which make use of both gradient and function values and utilize information from the three most recent points. A simple trust region subproblem is also constructed.

    In order to enable the algorithm to accept more trial steps, the nonmonotone strategy is defined by a modified Metropolis criterion.

    To overcome the case of "too successful iteration", adaptive strategy is introduced to adjust the trust region radius.

    The rest of this paper is organized as follows. An adaptive simple trust region algorithm based on new weak secant equations is proposed in the next section. In Section 3, the global convergence and locally superlinearly convergence of the new algorithm are established under mild assumptions. Section 4 introduces numerical experiments to prove the effectiveness of the algorithm. Conclusions are made in the last section.

    In this section, two new weak secant equations are introduced based on the formulas (1.4) and (1.5). On this basis, a simple trust region subproblem is constructed and a new trust region method for solving large-scale optimization problems is proposed.

    Based on the formulas (1.4) and (1.5), we introduce the following new weak secant equations

    ¯sTkBk+1¯sk=¯sTk¯yk, (2.1)
    ¯sTkBk+1¯sk=¯sTkzk, (2.2)

    where ¯sk, ¯yk, and zk are the same as formulas (1.4) and (1.5). Let the matrix Bk in subproblem (1.2) be a scalar matrix γkI satisfying (2.1) or (2.2), then we get

    γk=¯sTk1¯yk1¯sk12 (2.3)

    or

    γk=¯sTk1zk1¯sk12. (2.4)

    A simple trust region subproblem could be constructed as follows

    min m(k)(xk+s)=fk+gTks+12sTγkIs,s.t.sΔk. (2.5)

    Suppose gk0. The solution of subproblem (2.5) is given by:

    (i) If

    gkγkΔk,

    then

    sk=gkγk.

    (ii) If

    gkγk>Δk,

    then

    sk=Δkgkgk.

    The ratio rk can be rewritten as

    rk=f(xk)f(xk+sk)m(k)(xk)m(k)(xk+sk). (2.6)

    In our algorithm, the following modified Metropolis criterion is used to determine whether to accept the trial step

    pk={1,if rk>τ,exp{τrkTk},otherwise, (2.7)

    where Tk is the temperature at the k-iteration, 0<τ<1 is a sufficiently small real number, and the temperature Tk decreases to 0 as k. The modified Metropolis criterion is embedded into the trust region methods. Combine pk and rk to determine whether the algorithm is iterative. Different from the traditional trust region algorithm, when rkτ, the modified Metropolis criterion is used to accept more iterations with a certain probability, thereby reducing the amount of computation and improving the convergence rate of the algorithm.

    Based on the above analysis, we propose an adaptive simple model trust region algorithm (ASMTR) with new weak secant equations.

    ASMTR algorithm

    Step 0. Set ε>0, β(0,1), u(0,1), 0<τ<u<1, v>1, c(0,1), p=0, 0<κ1κ2. Give x0Rn, compute g0, and set s0=g0, x1=x0+s0. T1>0, Δ1>0, γ1=1, and k:=1.

    Step 1. Compute gk. If gk<ε, then stop.

    Step 2. If

    gkγk>Δk,

    then

    sk=Δkgkgk,

    otherwise

    sk=gkγk.

    Step 3. Compute rk and pk by (2.6) and (2.7), respectively, and let

    lk:=ev+(e1/vev)×rand(1).

    If pk>lk, then

    xk+1=xk+sk, (2.8)

    otherwise

    xk+1=xk. (2.9)

    Step 4. Compute γk+1 by (2.3) or (2.4). If γk+1κ1, set γk+1=κ1. If γk+1κ2, set γk+1=κ2.

    Step 5. If rk>u, set p=0; Otherwise, set p=p+1. Compute Δk+1 by (1.7).

    Step 6. Set Tk+1=βTk, k:=k+1, and go to Step 1.

    Remark 2.1. The formula (1.7) is used to update the trust region radius, i.e.,

    Δk+1=2cpgk+1{1,if k=0,sk2|sTkyk|,if k>0.

    Remark 2.2. From Step 4 of the ASMTR algorithm, we know that the sequence {γk} is uniformly bounded, i.e.,

    0<κ1γkκ2,k. (2.10)

    In this section, we will discuss the convergence of the ASMTR algorithm. We would like to make the following assumptions:

    Assumption (i) Let f: RnR be twice continuously differentiable and bounded below on the level set

    L(x0)={x|f(x)f(x0)}

    for x0Rn.

    (ii) Let the gradient g(x) be uniformly continuous on a compact convex set so that Ω contains the level set L(x0).

    Assumptions (i) and (ii) mean that 2f(x) is continuous and uniformly bounded on Ω, so there exists a positive constant L such that

    2f(x)L,xΩ.

    Therefore, from the mean value theorem, we have

    g(x)g(y)Lxy, x,yΩ,

    which ensures that g(x) is Lipschitz continuous on Ω.

    Lemma 1. Suppose that sk is the solution of subproblem (2.5) and the sequence {xk} is generated by the ASMTR algorithm. If gk0, then

    pred(sk)=m(k)(xk)m(k)(xk+sk)12δ1gkmin{Δk,gkγk},

    where δ1(0,1].

    Proof. The proof is similar to the proof of Lemma 4.1 in [16].

    Lemma 2. Suppose that Assumptions (i) and (ii) hold. The solution sk of the simple model (2.5) satisfies

    |f(xk+sk)m(k)(xk+sk)|12(κ2+L)sk2.

    Proof. According to the second-order Taylor expansion, we have

    f(xk+sk)=fk+sTkgk+10sTk[g(xk+tsk)g(xk)]dt.

    By the definition of m(k)(xk+s) in (2.5), we get

    |f(xk+sk)m(k)(xk+sk)|=|fk+sTkgk+10sTk[g(xk+tsk)g(xk)]dtfkgTksk12sTkγksk|=|12sTkγksk10sTk[g(xk+tsk)g(xk)]dt|12κ2sk2+|10sTk[g(xk+tsk)g(xk)]dt|12κ2sk2+12Lsk2=12(κ2+L)sk2.

    We complete the proof.

    Lemma 3. Suppose that Assumption (i) holds. Let the sequence {xk} be generated by the ASMTR algorithm, then there exists a sufficiently large N>0 such that

    fk+1fkτ4δ1gkmax{Δk,gkγk}

    holds for all k>N.

    Proof. If rkτ>0, then pk=1>lk holds. By Lemma 1, we have

    m(k)(xk)m(k)(xk+sk)12δ1gkmin{Δk,gkγk}

    and

    f(xk+1)=f(xk+sk)fk+τ(m(k)(xk+sk)m(k)(sk))fkτ2δ1gkmax{Δk,gkγk}.

    If rk<τ, then we accept xk+sk as the new iterate point when

    pk=exp{τrkTk}>lk. (3.1)

    By Step 6 of the ASMTR algorithm, it can be deduced that

    limkTk=0.

    Combining with lk[ev,e1/v], we have

    vlnlk1v,

    then, for a given ε1>0, there exists N>0 such that

    0<Tklnlk<ε1

    holds for all k>N. By a simple manipulation on (3.1), we get

    rk>τ+Tklnlk>τε1,

    then by the definition of rk, there is

    fk+1<fk+(τε1)(m(k)(xk+sk)m(k)(xk)).

    Set ε1=τ2. According to Lemma 1, we have

    fk+1fkτ4δ1gkmax{Δk,gkγk}.

    We complete the proof.

    Lemma 4. Suppose that Assumptions (i) and (ii) hold. Let the sequence {xk} be generated by the ASMTR algorithm. If xk is not the solution of the problem, i.e., gk0, then the iteration (2.9) will be terminated at a finite step.

    Proof. By contradiction that there is K1>0 such that the iteration (2.9) will cycle infinitely for kK1, then gkε. Thus, by Step 3 of the ASMTR algorithm, we have pk<lk for all k>K1, so

    rk<τ+Tklnlk<τTkv<τ<u. (3.2)

    By Step 5 of the ASMTR algorithm, we have

    limkΔk=0. (3.3)

    From Lemmas 1 and 2, we get

    |f(xk)f(xk+sk)pred(sk)1|=|[f(xk)f(xk+sk)][m(k)(xk)m(k)(xk+sk)]pred(sk)|(κ2+L)sk2/2δ1εmin{Δk,ε/κ2}/2=(κ2+L)Δk2δ1εmin{Δk,ε/κ2}. (3.4)

    Combining (3.3) and (3.4), we obtain

    limkrk=1.

    This implies that, for arbitrarily given ε2>0, there exists K2>0 such that rk>1ε2 holds for k>K2. Since 0<τ<u<1, by letting 0<ε2<1u, we get

    rk>u>τTkv. (3.5)

    Taking K=max{K1,K2}, we have that (3.2) and (3.5) simultaneously hold for k>K, leading to a contradiction.

    Theorem 1. Let the sequence {xk} be generated by the ASMTR algorithm, then we have

    limkgk=0.

    Proof. If the ASMTR algorithm terminates in a finite step, then the conclusion is obviously valid. Consider an infinite number of successful iterations. According to Assumption (i), it is known that the sequence {f(xk)} is bounded, i.e., there is an aR such that f(xk)a holds for all k.

    It is obtained from Lemma 3 that

    fk+1fkτ4δ1gkmax{Δk,gkγk}.

    By (2.10), we have

    max{Δk,gkγk}gkκ2,

    so

    fk+1fkτδ14κ2gk2. (3.6)

    Adding (3.6) with respect to k yields

    τδ14κ2N+Kk=Ngk2fNfN+K+1. (3.7)

    Noting that fN+K+1>a for any K>0 and taking limit on both sides of (3.7) as K, we have

    limK N+Kk=Ngk2<,

    which deduces

    limkgk=0.

    We complete the proof.

    Theorem 2. Suppose that Assumptions (i) and (ii) hold. Also, assume the ASMTR algorithm generates an infinite sequence {xk} converging to the optimal solution x, where the matrix 2f(x) is positive definite and 2f(x) is Lipschitz continuous in a neighborhood of x. If the following condition holds

    limkgk+2f(x)sksk=0, (3.8)

    then the sequence {xk} converges to x superlinearly.

    Proof. Since 2f(x) is positive definite and 2f(x) is continuous in a neighborhood of x, there exist positive scalars h and ψ such that

    sT2f(x)shs2,sRn, (3.9)

    for all

    x˜Ω={x|xxψ}.

    Also, there exists positive integer ¯k such that xk˜Ω, for all k¯k.

    From the Taylor expansion and inequality (3.9), for sufficiently large indices k, we have

    sTkyk=sTk(gk+1gk)=sTk2f(xk+ζksk)skhsk2,

    for some ζk(0,1). So, from (1.7), we get

    sk2cpsk12|sTk1yk1|gk2cphgk, (3.10)

    where c(0,1). Considering Lemma 4, p is finite in each iteration.

    On the other hand, from the Taylor expansion, we have

    gk+1=gk+2f(xk+ςksk)sk=gk+2f(x)sk+(2f(xk+ςksk)2f(x))sk,

    for some ςk(0,1). Thus,

    gk+1gk+2f(x)sk+2f(xk+ςksk)2f(x)sk.

    Dividing both sides by sk, we get

    gk+1skgk+2f(x)sksk+2f(xk+ςksk)2f(x).

    So, from Lipschitz continuity of 2f(x) on ˜Ω and (3.8), we have

    limkgk+1sk=0,

    which, from (3.10), yields

    limkgk+1gk=0,

    implying that the sequence {xk} converges to x superlinearly.

    In the current section, we show the numerical performance of the ASMTR algorithm. The test problems are unconstrained problems from CUTEr (a widely used testing environment for optimization software) library [31] and Andrei [32,33]. All codes are written on MATLAB R2015b and run on PC with a 1.19 GHz central processing unit (CPU) processor with 8.00 GB RAM memory. We write two new algorithms as

    (1) ASMTR1: the ASMTR algorithm with

    γk+1=¯sTk¯yk¯sk2.

    (2) ASMTR2: the ASMTR algorithm with

    γk+1=¯sTkzk¯sk2.

    Two new algorithms are compared with the following two algorithms. The first is the simulated annealing-based trust region BB method (SATRBB) [14], whose nonmonotone technique is defined by the modified Metropolis criterion; the second is the nonmonotone trust region BB methods (NTBB) [15]. The parameters are given by: T1=200, v=10, β=0.99, τ=0.1, c1=0.25, c2=0.75, Δ1=1, and ε=104 for the SATRBB algorithm, γ=0.1, η1=0.25, η2=0.75, and M=5 for the NTBB algorithm, and u=0.15, c=0.5, κ1=2, κ2=100, and γ1=1 for the ASMTR algorithm.

    All test algorithms are terminated when satisfying condition

    gk104.

    In addition, the algorithm is stopped if the number of iterations exceeds 500. In such a case, we claim fail of this algorithm. The values x0, x10, x20, x30, x40, and x50 in the second column associate with starting points with x0, 10x0, 10x0, 100x0, 100x0, and x0, where x0 is the same as [32,33]. The notations used in numerical results include the dimension of the problem (n), the initial point of the problem (x0), the number of iterations (k), the CPU time cost in seconds, the number of function evaluations (nf), and the number of gradient evaluations (ng). The sign "" means the algorithm fails because the number of iterations exceeds 500. Next, we present some of the numerical results in Examples 4.1–4.4.

    Example 4.1. Consider the Broyden banded function

    f(x)=ni=1(2xi+5x3i+1)2jJi(xj+x2j)2,

    where n is the variable,

    Ji={j:ji,max(1,iml)jmin(n,i+mu)},

    ml=5, and mu=1. The initial point of this function is x0=(1,,1)T, and the results are listed in Table 1.

    Table 1.  Numerical result of the Broyden banded function.
    SATRBB NTBB ASMTR1 ASMTR2
    n x0 k CPU nf ng k CPU nf ng k CPU nf ng k CPU nf ng
    10 x0 142 0.0010 282 143 131 0.0156 260 126 161 0.0010 320 162 111 0.0010 221 112
    x10 237 0.0010 472 238 242 0.0156 482 233 208 0.0156 414 209 149 0.0010 297 150
    x20 146 0.0010 290 147 145 0.0156 288 146 84 0.0156 166 85 82 0.0010 163 83
    x30 328 0.0156 654 328 333 0.0156 664 324 242 0.0469 464 243 228 0.0156 453 229
    x40 237 0.0156 472 238 236 0.0156 470 237 129 0.0313 256 130 127 0.0156 253 128

     | Show Table
    DownLoad: CSV

    Perform numerical experiments on the Broyden banded function for different initial points. Table 1 shows that ASMTR2 needs fewer iterations, function, and gradient evaluations. ASMTR1 and ASMTR2 are superior to SATRBB and NTBB.

    Example 4.2. Consider the Penalty function Ⅰ

    f(x)=105ni=1(xi1)2+(ni=1xi20.25)2,

    where n is the variable. The initial point of this function is x0=(1,2,,n)T, and the results are listed in Table 2.

    Table 2.  Numerical result of the Penalty function Ⅰ.
    SATRBB NTBB ASMTR1 ASMTR2
    n x0 k CPU nf ng k CPU nf ng k CPU nf ng k CPU nf ng
    10 x0 570.001011258 570.001011258 300.00105831 270.00105328
    20 720.078114273 720.015614273 370.00107238 320.00106333
    50 920.001018293 920.015618293 450.00108846 410.00108142
    100 1080.0010214109 1080.0010214109 520.001010253 480.00109549

     | Show Table
    DownLoad: CSV

    In the case of four dimensions, numerical experiments are performed from the same initial point. We find that ASMTR1 uses less iterations, function, and gradient evaluations than SATRBB and NTBB, and ASMTR2 is better than ASMTR1.

    Example 4.3. Consider the Broyden tridiagonal function

    f(x)=(3x12x21)2+n1i=2(3xi2x2ixi12xi+1+1)2+(3xn2x2nxn1+1)2,

    where n is the variable. The initial point of function is x0=(1,,1)T, and the results are listed in Table 3.

    Table 3.  Numerical result of the Broyden tridiagonal function.
    SATRBB NTBB ASMTR1 ASMTR2
    n x0 k CPU nf ng k CPU nf ng k CPU nf ng k CPU nf ng
    10000 x0 500.10949851 480.06259449 400.06257841 370.06257338
    x10 860.281317087 850.265616885 560.140611057 580.171911559
    x20 750.171914876 750.140614876 630.156312464 680.156313569
    x30 1160.2344230117 1210.1563240122 700.140613871 660.156313167
    x40 1110.3281220112 1100.2188218111 760.156315077 700.125013971
    20000 x0 490.18759650 480.23449449 510.125010049 910.312518189
    x10 830.296916484 870.406317287 510.156310052 540.265610755
    x20 770.328115278 760.250015077 690.203113670 660.234413167
    x30 1230.4063244124 1210.2969240122 760.203115077 660.265613167
    x40 1100.5000218111 1100.500218111 720.171914273 790.171915780
    50000 x0 440.39068645 470.37509248 580.343811455 580.484411556
    x10 880.750017489 870.703117288 570.546911258 520.437510353
    x20 780.687515479 780.703115479 990.7500196100 640.671912765
    x30 1220.9531242123 1210.9531240122 710.593814072 700.656313971
    x40 1121.3281222113 1110.9219220112 710.500014072 780.562515579

     | Show Table
    DownLoad: CSV

    For 10000, 20000, and 50000 dimensions, respectively, five initial points are selected to test the numerical results. SATRBB and NTBB win in approximately 13.4% of performed testing problems concerning the number of iterations, and ASMTR1 and ASMTR2 win in nearly 86.6% of performed testing problems. In addition, ASMTR1 and ASMTR2 need shorter CPU time for most problems. This means the new algorithm is very effective for large-scale optimization problems.

    Example 4.4. Consider the nearly separable function

    f(x)=ni=1x2i+nj=1x6j+cos2x2+n1i=2cos2(xi1+xi+1)+cos2xn1,

    where n is the variable and 1jn. The initial point of this function is

    x0=(n2(n+1),n12(n+1),...,12(n+1))T,

    and the results are listed in Table 4.

    Table 4.  Numerical result of the Nearly separable function.
    SATRBB NTBB ASMTR1 ASMTR2
    n x0 k CPU nf ng k CPU nf ng k CPU nf ng k CPU nf ng
    5000 x0 14730.2656292132 12726.5000364114
    10000 13027.5781258116 180118.3750525169
    20000 134103.4375266123 142329.2190409129

     | Show Table
    DownLoad: CSV

    Table 4 shows the numerical results of the function under different dimensions of the same initial point. SATRBB and NTBB do not run results within 500 iterations, and ASMTR1 and ASMTR2 effectively solved within 180 iterations.

    For more insight, we use the performance profiles introduced by Dolan and Moré [34] to illustrate the numerical performance of the four algorithms based on the testing functions in Table 5. The 20 test functions are listed in Table 5, in which the dimensions vary from 2 to 50000. In Table 5, "No." represents the number of the functions. Here, we also add the nonmonotone adaptive trust region method based on the simple conic model nonmonotone adaptive conic trust region (NACTR) method [7] to compare. This method needs less memory and computational efforts.

    Table 5.  The test functions.
    No. Functions
    1 Brown badly scaled function
    2 Generalized tridiagonal 1 function
    3 Allgower function
    4 Brown and Dennis function
    5 Boundary value function
    6 Staircase S1 function
    7 Chebyquad function
    8 Staircase S2 function
    9 Broyden banded function
    10 Broyden tridiagonal function
    11 Penalty function I
    12 Nearly separable function
    13 Schittkowski function 302
    14 Variable dimension function
    15 Yang tridiagonal function
    16 Generalized Rosebrock function
    17 Diagonal full Border function
    18 DIAG-AUP1 function
    19 Separable cubic function
    20 LIARWHD function

     | Show Table
    DownLoad: CSV

    Based on the numerical results of all the test problems, we present the performance profiles (including the number of iterations, the CPU time, the number of function evaluations, and the number of gradient evaluations). In a performance profile plot, the horizontal axis gives the percentage (τ) of the test problems for which a method is the fastest (efficiency), while the vertical side gives the percentage (ψ) of the test problems that are successfully solved by each of the methods.

    Figures 14 plot the performance profiles for the number of iterations, the CPU time, the number of function evaluations, and the number of gradient evaluations, respectively. It can be observed that ASMTR1 and ASMTR2 grow up faster than the other algorithms. In a word, they show that the performance of ASMTR1 and ASMTR2 is superior to SATRBB, NACTR, and NTBB in all aspects. In the overall trend, the performance of ASMTR2 is slightly better than ASMTR1. We believe that ASMTR2 is more competitive. Specifically for large-scale problems, the ASMTR algorithm has a strong numerical stability. From above analysis, we can conclude that the algorithm proposed in our work turns out to be quite competitive.

    Figure 1.  Performance profile of the number of iterations.
    Figure 2.  Performance profile of the CPU time.
    Figure 3.  Performance profile of the number of function evaluations.
    Figure 4.  Performance profile of the number of gradient evaluations.

    In this paper, we propose an adaptive simple model trust region algorithm based on new weak secant equations. It is worth noting that the trust region subproblem of the algorithm is solved more simply in contrast to the many other trust region methods proposed in the literature. We discuss the benefits of constructing a simple model using the last three points of information, and the algorithm combines the nonmonotone strategy defined by a modified Metropolis criterion and adaptive strategy. The global convergence and locally superlinearly convergence of the new algorithm are established under appropriate conditions. Numerical experiments show that the proposed algorithm is effective. There are still many deficiencies in our research; For example, the more efficient and widely used adaptive trust region radius is not taken into account. Therefore, we will explore a new adaptive trust region radius in the future and obtain a more effective and robust method.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The key project of natural science foundation joint fund of Jilin Province (YDZJ202101ZYTS167, YDZJ202201ZYTS303, YDZJ202201ZYTS320, YDZJ202101ZYTS156); The graduate innovation project of Beihua University (2022001, 2022038).

    All authors declare no conflicts of interest in this paper.



    [1] Y. Ji, Y. Ma, The robust maximum expert consensus model with risk aversion, Inf. Fusion, 99 (2023), 101866. https://doi.org/10.1016/j.inffus.2023.101866 doi: 10.1016/j.inffus.2023.101866
    [2] S. Qu, S. Li, A supply chain finance game model with order-to-factoring under blockchain, Syst. Eng. Theory Pract., 43 (2023), 3570–3586. https://doi.org/10.12011/SETP2022-2888 doi: 10.12011/SETP2022-2888
    [3] Y. Ji, Y. Yuan, Z. Peng, A novel robust flexible minimum cost consensus model with consensus granule, Group Decis. Negot., 2024. https://doi.org/10.1007/s10726-023-09869-3 doi: 10.1007/s10726-023-09869-3
    [4] N. Eslami, B. Najafi, S. M. Vaezpour, A trust region method for solving multicriteria optimization problems on riemannian manifolds, J. Optim. Theory Appl., 196 (2022), 212–239. https://doi.org/10.1007/s10957-022-02142-8 doi: 10.1007/s10957-022-02142-8
    [5] V. A. Ramirez, G. N. Sottosanto, Nonmonotone trust region algorithm for solving the unconstrained multiobjective optimization problems, Comput. Optim. Appl., 81 (2022), 769–788. https://doi.org/10.1007/s10589-021-00346-8 doi: 10.1007/s10589-021-00346-8
    [6] H. H. Dwail, M. A. K. Shiker, Using a trust region method with nonmonotone technique to solve unrestricted optimization problem, J. Phys. Conf. Ser., 1664 (2020), 012128. https://doi.org/10.1088/1742-6596/1664/1/012128 doi: 10.1088/1742-6596/1664/1/012128
    [7] L. Zhao, W. Sun, R. J. B. de Sampaio, Nonmonotone adaptive trust region method based on simple conic model for unconstrained optimization, Front. Math. China, 9 (2014), 1211–1238. https://doi.org/10.1007/s11464-014-0356-8 doi: 10.1007/s11464-014-0356-8
    [8] M. Ahookhosh, K. Amini, M. R. Peyghami, A nonmonotone trust-region line search method for large-scale unconstrained optimization, Appl. Math. Modell., 36 (2012), 478–487. https://doi.org/10.1016/j.apm.2011.07.021 doi: 10.1016/j.apm.2011.07.021
    [9] X. T. Zhu, M. Xi, W. Y. Sun, A new nonmonotone BB-TR method based on simple conic model for large scale unconstrained optimization, Numer. Math. A J. Chin. Univ., 38 (2016), 173–192.
    [10] H. Zhu, Q. Ni, J. Jiang, C. Dang, A new alternating direction trust region method based on conic model for solving unconstrained optimization, Optimization, 70 (2020), 1555–1579. https://doi.org/10.1080/02331934.2020.1745793 doi: 10.1080/02331934.2020.1745793
    [11] Q. Zhou, C. Zhang, A new nonmonotone adaptive trust region method based on simple quadratic models, J. Appl. Math. Comput., 40 (2012), 111–123. https://doi.org/10.1007/s12190-012-0572-x doi: 10.1007/s12190-012-0572-x
    [12] Q. Zhou, J. Chen, Z. Xie, A nonmonotone trust region method based on simple quadratic models, J. Comput. Appl. Math., 272 (2014), 107–115. https://doi.org/10.1016/j.cam.2014.04.026 doi: 10.1016/j.cam.2014.04.026
    [13] Q. Sun, L. Duan, B. Cui, A nomonotone trust region algorithm with simple quadratic models, J. Syst. Sci. Math. Sci., 29 (2009), 470–483.
    [14] X. Li, W. Dong, Z. Peng, A new nonmonotone trust region Barzilai-Borwein method for unconstrained optimization problems, Acta Math. Appl. Sin. Engl. Ser., 37 (2021), 166–175. https://doi.org/10.1007/s10255-021-0997-9 doi: 10.1007/s10255-021-0997-9
    [15] Y. Liu, X. Liu, Trust region BB methods for unconstrained optimization, Math. Numer. Sin., 38 (2016), 96–112. https://doi.org/10.12286/jssx.2016.1.96 doi: 10.12286/jssx.2016.1.96
    [16] Q. Zhou, W. Sun, H. Zhang, A new simple model trust-region method with generalized Barzilai-Borwein parameter for large-scale optimization, Sci. China Math., 59 (2016), 2265–2280. https://doi.org/10.1007/s11425-015-0734-2 doi: 10.1007/s11425-015-0734-2
    [17] Q. Z. Yang, Optimization method, Beijing: Science Press, 2015.
    [18] M. J. Ebadi, A. Fahs, H. Fahs, R. Dehghani, Competitive secant (BFGS) methods based on modified secant relations for unconstrained optimization, Optimization, 72 (2023), 1691–1701. https://doi.org/10.1080/02331934.2022.2048381 doi: 10.1080/02331934.2022.2048381
    [19] J. Zhang, C. Xu, Properties and numerical performance of quasi-Newton methods with modified quasi-Newton equations, J. Comput. Appl. Math., 137 (2001), 269–278. https://doi.org/10.1016/S0377-0427(00)00713-5 doi: 10.1016/S0377-0427(00)00713-5
    [20] J. E. Dennis, H. Wolkowicz, Sizing and least-change secant methods, SIAM J. Numer. Anal., 30 (1993), 1291–1314. https://doi.org/10.1137/0730067 doi: 10.1137/0730067
    [21] S. Zhao, T. Yan, K. Wang, Y. Zhu, Adaptive trust-region method on Riemannian manifold, J. Sci. Comput., 96 (2023), 67. https://doi.org/10.1007/s10915-023-02288-1 doi: 10.1007/s10915-023-02288-1
    [22] S. Lior, E. Yonathan, M. Shie, Adaptive trust region policy optimization: global convergence and faster rates for regularized MDPs, Proceedings of the AAAI Conference on Artificial Intelligence, 34 (2020), 5668–5675. https://doi.org/10.1609/aaai.v34i04.6021 doi: 10.1609/aaai.v34i04.6021
    [23] A. Kamandi, K. Amini, A new nonmonotone adaptive trust region algorithm, Appl. Math., 67 (2022), 233–250. https://doi.org/10.21136/AM.2021.0122-20 doi: 10.21136/AM.2021.0122-20
    [24] X. Zhang, J. Zhang, L. Liao, An adaptive trust region method and its convergence, Sci. China Ser. A, 45 (2002), 620–631. https://doi.org/10.1360/02ys9067 doi: 10.1360/02ys9067
    [25] Z. Shi, J. Guo, A new trust region method with adaptive radius, Comput. Optim. Appl., 41 (2008), 225–242. https://doi.org/10.1007/s10589-007-9099-8 doi: 10.1007/s10589-007-9099-8
    [26] S. Rezaee, S. Babaie-Kafaki, An adaptive nonmonotone trust region algorithm, Optim. Methods Software, 34 (2017), 264–277. https://doi.org/10.1080/10556788.2017.1364738 doi: 10.1080/10556788.2017.1364738
    [27] N. Ghalavand, E. Khorram, V. Morovati, Two adaptive nonmonotone trust-region algorithms for solving multiobjective optimization problems, Optimization, 2023. https://doi.org/10.1080/02331934.2023.2234920 doi: 10.1080/02331934.2023.2234920
    [28] X. Ding, Q. Qu, X. Wang, A modified filter nonmonotone adaptive retrospective trust region method, PLoS ONE, 16 (2021), e0253016. https://doi.org/10.1371/journal.pone.0253016 doi: 10.1371/journal.pone.0253016
    [29] M. Yousefi, A. M. Calomardo, A stochastic nonmonotone trust-region training algorithm for image classification, International IEEE Conference on Signal-Image Technologies and Internet-Based System, 2022,522–529. https://doi.org/10.1109/SITIS57111.2022.00084 doi: 10.1109/SITIS57111.2022.00084
    [30] Q. Zhou, D. Hang, Nonmonotone adaptive trust region method with line search based on new diagonal updating, Appl. Numer. Math., 91 (2015), 75–88. https://doi.org/10.1016/j.apnum.2014.12.009 doi: 10.1016/j.apnum.2014.12.009
    [31] N. I. Gould, D. Orban, P. L. Toint, UTEr and SifDec: a constrained and unconstrained testing environment, revisited, ACM Trans. Math. Software, 29 (2003), 373–394. https://doi.org/10.1145/962437.962439 doi: 10.1145/962437.962439
    [32] N. Andrei, Introduction: overview of unconstrained optimization, In: Nonlinear conjugate gradient methods for unconstrained optimization, Springer, 2020. https://doi.org/10.1007/978-3-030-42950-8_1
    [33] N. Andrei, An unconstrained optimization test functions collection, Adv. Model. Optim., 10 (2008), 147–161.
    [34] E. D. Dolan, J. J. Moré, Benchmarking optimization software with performance profiles, Math. Program., 91 (2002), 201–213. https://doi.org/10.1007/s101070100263 doi: 10.1007/s101070100263
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1198) PDF downloads(59) Cited by(0)

Figures and Tables

Figures(4)  /  Tables(5)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog