Research article

An accelerated conjugate gradient method for the Z-eigenvalues of symmetric tensors

  • Received: 06 February 2023 Revised: 23 March 2023 Accepted: 30 March 2023 Published: 23 April 2023
  • MSC : 15A18, 15A69, 90C55

  • We transform the Z-eigenvalues of symmetric tensors into unconstrained optimization problems with a shifted parameter. An accelerated conjugate gradient method is proposed for solving these unconstrained optimization problems. If solving problem results in a nonzero critical point, then it is a Z-eigenvector corresponding to the Z-eigenvalue. Otherwise, we solve the shifted problem to find a Z-eigenvalue. In our method, the new conjugate gradient parameter is a modified CD conjugate gradient parameter, and an accelerated parameter is presented by using the quasi-Newton direction. The global convergence of new method is proved. Numerical experiments are listed to illustrate the efficiency of the proposed method.

    Citation: Mingyuan Cao, Yueting Yang, Chaoqian Li, Xiaowei Jiang. An accelerated conjugate gradient method for the Z-eigenvalues of symmetric tensors[J]. AIMS Mathematics, 2023, 8(7): 15008-15023. doi: 10.3934/math.2023766

    Related Papers:

    [1] Xueyong Wang, Gang Wang, Ping Yang . Novel Pareto $ Z $-eigenvalue inclusion intervals for tensor eigenvalue complementarity problems and its applications. AIMS Mathematics, 2024, 9(11): 30214-30229. doi: 10.3934/math.20241459
    [2] Tinglan Yao . An optimal $ Z $-eigenvalue inclusion interval for a sixth-order tensor and its an application. AIMS Mathematics, 2022, 7(1): 967-985. doi: 10.3934/math.2022058
    [3] Zhi-jie Jiang . Complex symmetric difference of the weighted composition operators on weighted Bergman space of the half-plane. AIMS Mathematics, 2024, 9(3): 7253-7272. doi: 10.3934/math.2024352
    [4] Shunjie Bai . A tighter M-eigenvalue localization set for partially symmetric tensors and its an application. AIMS Mathematics, 2022, 7(4): 6084-6098. doi: 10.3934/math.2022339
    [5] Yajun Xie, Changfeng Ma . The hybird methods of projection-splitting for solving tensor split feasibility problem. AIMS Mathematics, 2023, 8(9): 20597-20611. doi: 10.3934/math.20231050
    [6] Habibu Abdullahi, A. K. Awasthi, Mohammed Yusuf Waziri, Issam A. R. Moghrabi, Abubakar Sani Halilu, Kabiru Ahmed, Sulaiman M. Ibrahim, Yau Balarabe Musa, Elissa M. Nadia . An improved convex constrained conjugate gradient descent method for nonlinear monotone equations with signal recovery applications. AIMS Mathematics, 2025, 10(4): 7941-7969. doi: 10.3934/math.2025365
    [7] Hong-Mei Song, Shi-Wei Wang, Guang-Xin Huang . Tensor Conjugate-Gradient methods for tensor linear discrete ill-posed problems. AIMS Mathematics, 2023, 8(11): 26782-26800. doi: 10.3934/math.20231371
    [8] Yajun Xie, Minhua Yin, Changfeng Ma . Novel accelerated methods of tensor splitting iteration for solving multi-systems. AIMS Mathematics, 2020, 5(3): 2801-2812. doi: 10.3934/math.2020180
    [9] Hong-bin Bai, Zhi-jie Jiang, Xiao-bo Hu, Zuo-an Li . 2-complex symmetric weighted composition operators on Fock space. AIMS Mathematics, 2023, 8(9): 21781-21792. doi: 10.3934/math.20231111
    [10] Yixin Li, Chunguang Li, Wei Yang, Wensheng Zhang . A new conjugate gradient method with a restart direction and its application in image restoration. AIMS Mathematics, 2023, 8(12): 28791-28807. doi: 10.3934/math.20231475
  • We transform the Z-eigenvalues of symmetric tensors into unconstrained optimization problems with a shifted parameter. An accelerated conjugate gradient method is proposed for solving these unconstrained optimization problems. If solving problem results in a nonzero critical point, then it is a Z-eigenvector corresponding to the Z-eigenvalue. Otherwise, we solve the shifted problem to find a Z-eigenvalue. In our method, the new conjugate gradient parameter is a modified CD conjugate gradient parameter, and an accelerated parameter is presented by using the quasi-Newton direction. The global convergence of new method is proved. Numerical experiments are listed to illustrate the efficiency of the proposed method.



    Since Lim [1] and Qi [2] presented some theories of the eigenvalues and eigenvectors for higher order tensors, the related research has received much more attention (see [3,4,5,6,7,8,9,10,11,12,13], etc). The eigenvalues of symmetric tensors have been applied in blind source separation [2] and hypergraph theory [9], statistical data analysis [14] and high order Markov chains [15], etc. Moreover, various definitions of eigenvalues and eigenvectors for tensors have been introduced [2,10,16].

    There are many works for computing the eigenvalues of tensors, especially for Z-eigenvalues. Qi et al. [17] proposed an elimination method for finding all Z-eigenvalues, which is specific to the third-order tensors. Kolda and Mayo [6] presented a shifted power method (SPM) for calculating Z-eigenvalues, in which the shifted parameter is crucial. Han [18] provided an unconstrained optimization approach for even order symmetric tensors. Hao, Cui and Dai [4] found the extreme Z-eigenvalues and corresponding Z-eigenvectors by the sequential subspace projection method. Under certain assumptions, the global convergence and linear convergence were established for symmetric tensors. Hao, Cui and Dai [19] proposed a feasible trust region method for finding the extreme Z-eigenvalues of symmetric tensors. The global convergence and local quadratic convergence were established.

    Inspired by the idea of improved conjugate parameters proposed in the above works and the application of optimization methods in tensor eigenvalue calculation, our main work is to study the transformation of Z-eigenvalues of symmetric tensors into unconstrained optimization and to propose a new algorithm. The contributions of this article are listed as follows:

    For the case of different critical point, we transform the Z-eigenvalues of symmetric tensors into different unconstrained optimization problems which include shifted problem.

    We propose a new conjugate gradient method with a new conjugate gradient parameter and an accelerated parameter, which converges to a critical point. The found nonzero critical point is a Z-eigenvector associated with a Z-eigenvalue of symmetric tensors. When the zero critical point is obtained, a shifted problem is solved for finding a Z-eigenvalue.

    The global convergence of new method is established. We compare our method with conjugate gradient methods proposed in [20,21], for computing the Z-eigenvalues of symmetric tensors. The numerical results show that the proposed method is competitive.

    The rest of this paper is organized as follows. In Section 2, we transform the Z-eigenvalues problem into unconstrained optimizations, and propose an accelerated conjugate gradient method for solving it. Global convergence result is established in Section 3. Numerical experiments are shown in Section 4.

    Let R be the real field, m,n be positive integers and

    A=(ai1i2im),ai1i2imR,1i1,,imn

    be an mth-order n-dimensional real tensor. The set of mth-order n-dimensional real tensor is denoted by R[m,n]. Tensor A is symmetric if its entries are invariant under any permutation of their indices. The set of mth-order n-dimensional real symmetric tensor is denoted by S[m,n].

    If AR[m,n], an mth-degree homogeneous polynomial function with real coefficients is uniquely determined by

    Axm:=ni1,i2,,im=1ai1i2imxi1xim.

    For x=(x1,,xn)TRn, Axm1 denotes a n-dimensional column vector, i.e.,

    Axm1:=(ni2,,im=1aii2imxi2xim)1in.

    If A is symmetric, then the gradient of Axm satisfies

    (Axm)=mAxm1

    for all xRn. In our work, we consider the following Z-eigenvalues of symmetric tensors.

    Definition 1. [2] Let AR[m,n], if there exist λR and a vector xRn{0} satisfying

    Axm1=λx,    xTx=1, (2.1)

    then λ is called a Z-eigenvalue of A and x is called the corresponding Z-eigenvector.

    Motivated by the work of Auchmuty [22], we generalize the unconstrained variational principles to Z-eigenvalues of A. Consider the following unconstrained optimization

    minxRnf(x)=12m(xTx)m1mAxm. (2.2)

    The gradient and Hessian of f(x) are listed as follows

    g(x):=f(x)=(xTx)m1xAxm1, (2.3)
    G(x):=2f(x)=(xTx)m1I+2(m1)(xTx)m2xxT(m1)Axm2, (2.4)

    where

    Axm2:=(ni3,,im=1aiji3imxi3xim)1i,jn.

    Obviously, G(x) is a symmetric matrix. In order to research the properties of f(x) in (2.2), we cite the following definition and a nice feature of it.

    Definition 2. [23] A continuous function h:RnR is called coercive if it satisfies

    limxh(x)=+.

    If x satisfies the equation h(x)=0, then it is termed as a critical point of h(x).

    Theorem 1. [23] Let h:RnR be continuous. If h is coercive, then h has at least one global minimizer. In addition, if the first partial derivatives exist on Rn, then h attains its global minimizers at its critical points.

    Based on a similar argument, for the Z-eigenvalues of tensors, we have the following result.

    Theorem 2. Let AR[m,n] be symmetric tensors. Assume that λmax is the largest Z-eigenvalue of A. Denote the Z-spectrum of A by σZ(A):={λ:λis a Z-eigenvalue ofA}. We have

    (ⅰ) f(x) is coercive on Rn.

    (ⅱ) The critical points of f(x) are at x=0 and any Z-eigenvector x0 associated with a Z-eigenvalue λ>0 of A satisfying λ=(xTx)m1.

    (ⅲ) If λmax>0, then f(x) attains its global minimal value

    fmin=12m(λmax)mm1

    at any Z-eigenvector associated with the Z-eigenvalue λmax such that λmax=(xTx)m1.

    (ⅳ) If λmax0, then x=0 is the unique critical point of f(x). Moreover, it is the unique global minimizer of f(x) on Rn.

    Proof. (ⅰ) Since

    f(x)=12m(xTx)m1mAxm=12mx2m1mAxm

    and Axm is an mth-degree homogeneous polynomial function with real coefficients, then

    f(x)asx.

    That is, f(x) is coercive on Rn.

    (ⅱ) From the definition of the critical point of f(x), we have

    Axm1=(xTx)m1x. (2.5)

    It is obvious that x=0 is a critical point of f(x) as g(0)=0. The point xRn{0} satisfying (2.5) is a Z-eigenvector corresponding to the Z-eigenvalue λ=(xTx)m1>0, which is also a critical point of f(x).

    (ⅲ) At the critical point xRn{0}, a Z-eigenvector x associated with a Z-eigenvalue λ satisfies λ=(xTx)m1 and Axm=λxTx. Moreover,

    f(x)=12mλxTx1mλxTx=12mλxTx=12mλmm112m(λmax)mm1.

    By Theorem 2.1 and conclusion (ⅱ), we get the global minimum value 12m(λmax)mm1 at any Z-eigenvector x corresponding to the Z-eigenvalue λmax such that λmax=(xTx)m1.

    (ⅳ) Since λmax0 implies that λ0 for any λσZ(A), λ=(xTx)m1 does not hold for any Z-eigenvector x associated with a Z-eigenvalue λ, as (xTx)m1>0 for any xRn{0}. Therefore, from Theorem 2.1, x=0 is the unique critical point and the unique global minimize of f(x).

    Note that, if λmax0, then x=0 is the unique critical point, but it does not result in a Z-eigenvalue. In this case, we solve a following shifted problem

    minxRnft(x)=12m(xTx)m1mAxmt2(xTx), (2.6)

    where t>0 is a shifted parameter. It is obvious that, when t is sufficient large, for any x0, we have ft(x)<0. From ft(0)=0, we know that x=0 is the unique maximizer of ft(x). Denote the gradient of ft(x) as

    gt(x):=ft(x)=(xTx)m1xAxm1tx. (2.7)

    Obviously, x=0 is also a critical point of ft(x). The nonzero critical point of problem (2.6) is a Z-eigenvector corresponding to Z-eigenvalue λt=(xTtxt)m1t. In this case, a suitable descent algorithm for solving shifted problem (2.6) should converge to a nonzero critical point. Therefore, we can get Z-eigenvalues and its associated Z-eigenvectors by solving problem (2.2) or (2.6). The algorithm is described as follows.

    Algorithm 1: Step 0: Given AS[m,n],t1,ˉρ>1.

    Step 1: Solving problem (2.2) by using an algorithm to obtain xk and compute λk=(xTkxk)m1.

    Step 2: If xkε, stop, output xk and compute λk=(xTkxk)m1t; otherwise, let t:=ˉρt, go to Step 3.

    Step 3: Solve problem (2.6) by using an algorithm to obtain xk, go to Step 2.

    Remark 1. There is an inner loop between Step 2 and Step 3. Since descent algorithm for solving problem (2.6) will result in a nonzero critical point, then this inner loop can be terminated by finite iteration for sufficient large t. Therefore, Algorithm 1 is well defined.

    Remark 2. When executing algorithm, it should use the same unconstrained optimization method to solve problems (2.2) and (2.6). We will propose a new accelerated conjugate gradient method, especially for solving problem (2.2) or (2.6).

    In this section, we devote to giving an accelerated conjugate gradient method for solving unconstrained optimization problem, such as (2.2) or (2.6). Firstly, we consider the iterative formula of nonlinear conjugate gradient algorithm

    xk+1=xk+αkdk. (3.1)

    The stepsize αk>0 is determined by a line search and the direction dk are computed by

    dk+1=θk+1gk+1+βk+1sk,d0=g0, (3.2)

    where gk+1=g(xk+1),sk=αkdk.

    We first introduce a new conjugate parameter of our conjugate gradient method based on CD nonlinear conjugate gradient method ([20]). The new conjugate parameter is

    βk+1=gk+12sTkyk(gTksk)2 (3.3)

    is introduced. When using the exact line search, βk+1 in (3.3) reduces to CD conjugate gradient parameter. An accelerated parameter θk+1 is obtained by the quasi-Newton direction ([24,25]). Let dk+1=G1k+1gk+1, namely

    G1k+1gk+1=θk+1gk+1+βk+1sk, (3.4)

    where Gk+1 satisfies the following secant equation

    Gk+1sk=yk,yk=gk+1gk. (3.5)

    From (3.4), then

    gk+1=θk+1Gk+1gk+1βk+1Gk+1sk.

    Pre-multiplying at the both sides by sTk, we have

    sTkgk+1=θk+1sTkGk+1gk+1βk+1sTkGk+1sk. (3.6)

    Then, combined with (3.3)–(3.6), we have

    θk+1=1yTkgk+1[gk+12(sTkyk)2(gTksk)2+sTkgk+1]. (3.7)

    The stepsize is generated by the strong Wolfe line search conditions

    f(xk+αkdk)f(xk)ραkgTkdk, (3.8)
    |g(xk+αkdk)Tdk|σgTkdk, (3.9)

    where 0<ρ<σ<1. We prove the descent property of the direction (3.2) under (3.8) and (3.9) in the following. Multiply both sides of (3.9) by αk, from sk=αkdk, we can easily obtain (gTk+1sk)2σ2(gTksk)2.

    Theorem 3. If θk+112+2σ2, then the direction determined by (3.2) satisfies the sufficient descent condition

    gTk+1dk+112gk+12. (3.10)

    Proof. Multiplying (3.2) by gTk+1, we obtain

    gTk+1dk+1=θk+1gk+12+gTk+1skgTkskgk+12+(gTk+1sk)2gk+12(gTksk)2. (3.11)

    Using the inequality aTb12(a2+b2), where a,bRn, we have

    gTk+1skgk+12gTksk=[(gTksk)gk+1/2]T[2(gTk+1sk)gk+1](gTksk)212[12(gTksk)2gk+12+2(gTk+1sk)2gk+12](gTksk)2=14gk+12+(gTk+1sk)2gk+12(gTksk)2. (3.12)

    Substituted (3.12) into (3.11), we have

    gTk+1dk+1θk+1gk+12+14gk+12+2(gTk+1sk)2gk+12(gTksk)2. (3.13)

    Then from (gTk+1sk)2σ2(gTksk)2 and θk+112+2σ2, we have

    gTk+1dk+1θk+1gk+12+14gk+12+2σ2gk+12=(θk+12σ214)gk+1212gk+12<0. (3.14)

    The proof is completed.

    Now, we describe an accelerated conjugate gradient algorithm (ACG).

    ACG algorithm

    Step 0: Given x0Rn,ε0and0<ρ<σ<1. Compute g0, let d0=g0. Set k:=0.

    Step 1: If gkε, stop, output xk. Otherwise, calculate αk from (3.8) and (3.9). Let xk+1=xk+αkdk and sk=αkdk.

    Step 2: Compute gk+1,yk, βk+1 by (3.3) and θk+1 by (3.7). Let θk+1:=max{θk+1,2σ2+12}.

    Step 3: Using (3.2) to obtain dk+1. Set k:=k+1 and go to step 1.

    Now, we establish the convergence result of ACG algorithm. Let the level set Ω={xRn|f(x)f(x0)} be a bounded closed set, i.e., there exists a constant γ>0 such that xγ for all xΩ. To facilitate analyzing, denote Ai1=(ai1i2im)1i2,i3,,imn and Ai1i2=(ai1i2im)1i3,i4,,imn.

    Lemma 1. Consider tensors AS[m,n], then Axm is Lipschitz continuous on Ω.

    Proof. Let p(x)=Axm, mathematical induction is adopted. If m=1, p(x)=ni=1aixi, utilizing equivalence of vector norm, for all x, yΩ, we have

    p(x)p(y)=|ni=1aixini=1aiyi|=|ni=1ai(xiyi)|ni=1|ai||xiyi|maxi=1,2,,n|ai|(ni=1|xiyi|)maxi=1,2,,n|ai|xyP2xy.

    Assume that the statements satisfy for all order mk1. When m=k, we have

    p(x)p(y)=|AxkAyk|=|ni1,i2,,ik=1ai1i2ikxi1xi2xikni1,i2,,ik=1ai1i2ikyi1yi2yik|=|ni1=1(xi1Ai1xk1yi1Ai1yk1)|=|ni1=1(xi1yi1)Ai1xk1+yi1(Ai1xk1Ai1yk1)|ni1=1(|xi1yi1|Ai1xk1+|yi1|Ai1xk1Ai1yk1).

    Since Ω is a bounded close set, then Ai1xk1 is bounded on Ω and zP1. From ni1=1|xi1yi1|xy and ni1=1|yi1|z, there exists a positive constant P3 such that

    p(x)p(y)P3xy.

    Namely, Axm is Lipschitz continuous on Ω. The proof is completed.

    Lemma 2. If AS[m,n], then Axm1 and Axm2 are Lipschitz continuous on Ω.

    Proof. For all x, yΩ, using Lemma 3.1 and equivalence of norm, we have

    Axm1Aym1M1(Ai1xm1Ai1xm1)1i1n=M1max1i1n|Ai1xm1Ai1xm1|M1max1i1nPi1xy=P4xy, (3.15)

    where P4 depends on tensor A and set Ω.

    Similarly, for all x, yΩ, using Lemma 3.1 and equivalence of norm, we have

    Axm2Aym2M2(Ai1i2xm2Ai1i2ym2)1i1,i2n=M2max1i1nni2=1|Ai1i2xm2Ai1i2ym2|M2max1i1,i2nni2=1Pi1i2xy=P5xy, (3.16)

    where P5 depends on tensor A and set Ω.

    Lemma 3. If AS[m,n], then g(x) is Lipschitz continuous in a neighbourhood N of Ω, namely

    g(x)g(y)Lxy (3.17)

    holds for any x,yN, where L is a positive number.

    Proof. There are two cases of gradient g(x) to consider. One case is computed by (2.3) and the other case is computed by (2.7).

    For (2.3): since N is a bound closed set, from Lemma 3.1, for all x,yN, we have

    g(x)g(y)=(xTx)m1xAxm1(yTy)m1x+Aym1Aym1Axm1+(xTx)m1x(yTy)m1xP4xy+P6xy=(P4+P6)xy. (3.18)

    For (2.7): since N is a bound closed set, from Lemma 3.1, for all x,yN, we have

    g(x)g(y)=(xTx)m1xAxm1tx(yTy)m1x+Aym1+tyAym1Axm1+(xTx)m1x(yTy)m1x+tytxP4xy+P6xy+P7xy=(P4+P6+P7)xy. (3.19)

    The proof is completed.

    We can easily get that there exists a constant M>0 such that g(x)M. The following useful lemma was essentially which proved by Zoutendijk [26]. From Theorem 3.1, we can obtain that the sequence {dk} generated by ACG algorithm satisfies the following Lemmas.

    Lemma 4. Let the sequences {xk} and {dk} be generated by ACG algorithm, we have

    k=0(gTkdk)2dk2<.

    Lemma 5. Let the sequence {xk} be generated by ACG algorithm. If

    k01dk2=,

    then

    lim infkgk=0.

    From Theorem 3.1, (3.17) and Lemma 3.4, the result can be proved, which is omitted here.

    Theorem 4. Let the sequence {xk} be generated by ACG algorithm. Then, we have

    lim infkgk=0. (3.20)

    Proof. From Theorem 3.1, there exists a constant c<0 satisfying gTkskcgksk<0, i.e., gTkskcgksk. Then, we have

    βk+1gk+12gTksk(1+σ)gk+12cgksk(1+σ)M(1+σ)csk=ξsk,

    where ξ=M(1+σ)c. According to |gTk+1dk|σgTkdk and sk2γ, we have

    θk+1=1yTkgk+1[gk+12(sTkyk)2(gTksk)2+sTkgk+1]=gk+12(sTkyk)2(gTksk)2)yTkgk+1+sTkgk+1yTkgk+1=gk+12[sTk(gk+1gk)]2(gTksk)2yTkgk+1+sTkgk+1yTkgk+1gk+12(σgTkskgTksk)2(gTksk)2yTkgk+1+sTkgk+1yTkgk+1=(σ+1)2gk+12(gTksk)2(gTksk)2yTkgk+1+sTkgk+1yTkgk+1=(σ+1)2gk+12+sTkgk+1yTkgk+1.

    Because of θk+12σ2+12, so yTkgk+1>0. Without losing generality, let yTkgk+1>κ>0, then we have

    θk+1<(σ+1)2gk+12+2γgk+1κ<(σ+1)2M2+2γMκδ. (3.21)

    Therefore,

    dk+1|θk+1|gk+1+|βk+1|skδM+|ξ|sksk=δM+|ξ|.

    We have

    k01dk21(δM+|ξ|)2k01=.

    From Lemma 3.5, it follows that (3.20) is derived. The proof is completed.

    Theorem 5. Let problems (2.2) and (2.6) are solved by ACG algorithm. ACG algorithm is well defined.

    Proof. In ACG algorithm, we obtain the Z-eigenvalues of symmetric tensors by solving problem (2.2) or (2.6). When problem (2.6) is solved, it means that solving problem (2.2) results in a zero critical point. Then ACG algorithm turn to solve problem (2.6) which converges to a nonzero critical point. Moreover, since the convergence of our algorithm has been guaranteed, so the termination criteria condition always holds. That is, ACG algorithm is well defined. The proof is completed.

    In this section, we report some numerical performance of ACG algorithm for solving problems (2.2) and (2.6). For convenience, we provide a table of abbreviations for the methods in Table 1.

    Table 1.  The abbreviations for methods.
    Abbreviation Method
    ACG Accelerated conjugate gradient method
    HS HestenesStiefel method
    PRP PolakRibiˊrePolyak method
    SPM Shifted power method
    QN QuasiNewton method
    ATTCG Accelerated threeterm conjugategradient method
    ADL Accelerated DaiLiao projection method

     | Show Table
    DownLoad: CSV

    We compare ACG with HS and PRP, which have been reported to be very efficient for unconstrained optimization. All experiments are done on a PC with CPU 2.40GHz and 2.00GB RAM using MATLAB R2013a. In the implementation of ACG algorithm, we set parameters ε=105,ρ=0.1,σ=0.5,t=1,ˉρ=2. In Table 2, Ex is the number of example, n is the dimension, k is the number of iterations, CPU stands for the time costed by algorithms (in seconds), λ stands for Z-eigenvalue outputted by algorithms. All algorithms share the same start points and stopping criteria. In the following examples, the tensors A are originally from [27].

    Table 2.  Some numerical results of examples for Z-eigenvalues.
    Ex n λ gk k/CPU(ACG) k/CPU(SPM)
    1 n=5 9.9873 1.6731e007 6/0.2188 10/0.3178
    1 n=10 17.7657 2.0318e006 4/0.3594 7/0.4815
    1 n=50 81.6402 9.0295e006 6/1.7031 11/2.8750
    1 n=100 158.1895 1.0076e007 4/12.9219 8/18.0159
    1 n=200 311.3130 5.3215e006 7/98.8125 12/115.5250
    2 n=100 132.1072 2.9073e006 9/30.0761 14/42.0918
    2 n=200 405.2981 6.9826e007 15/123.0050 12/145.0629
    3 n=5 13.0791 3.9828e006 4/0.2188 7/0.2983
    3 n=10 49.4905 2.3016e006 4/0.5705 8/0.9417
    3 n=50 154.9351 7.6239e006 3/31.6406 6/47.2781
    4 n=3 0.8893 1.1075e005 7/0.2188 12/0.4106
    5 n=10 43.2760 3.6133e006 5/0.2656 8/0.3875
    5 n=30 136.2817 2.8531e006 3/3.0469 6/6.2769
    6 n=5 34.5317 1.7063e006 6/0.2031 10/0.3385
    6 n=30 164.9089 6.7357e007 6/10.7969 12/18.7291

     | Show Table
    DownLoad: CSV

    Example 1. Let AS[3,n] defined by

    Ai1,i2,i3=(1)i1i1+(1)i2i2+(1)i3i3.

    Example 2. Let AS[3,n] defined by

    Ai1,i2,i3=tan(i1)+tan(i2)+tan(i3).

    Example 3. Let AS[4,n] defined by

    Ai1,i2,i3,i4=arctan((1)i1i1n)++arctan((1)i4i4n).

    Example 4. Let AS[4,3] defined by

    a1111=0.2883,a1112=0.0031,a1113=0.1973,a1122=0.2485,a1223=0.1862,a1133=0.3847,a1222=0.2972,a1123=0.2939,a1233=0.0919,a1333=0.3619,a2222=0.1241,a2223=0.3420,a2233=0.2127,a2333=0.2727,a3333=0.3054.

    Example 5. Let AS[4,n] defined by

    Ai1,i2,i3,i4=(1)i1i1++(1)i4i4.

    Example 6. Let AS[4,n] defined by

    Ai1,i2,i3,i4=tan(i1)++tan(i4).

    In Tables 2 and 3, we compare the numerical results of ACG algorithm and SPM algorithm (in [6]), PRP, HS and CD methods. Although the computed Z-eigenvectors x associated to λ are not shown, the values of gk are listed which are satisfy the given precision. It implies that (λ,x) are considered as true solutions of problem (2.1). Two methods reach the same Z-eigenvalues for problems with same dimensions. ACG algorithm requires less iterations and CPU time than that of SPM algorithm. That is, we can see that ACG algorithm is competitive for computing Z-eigenvalues of symmetric tensors.

    Table 3.  Some numerical results of examples for Z-eigenvalues.
    Ex n λ gk k/CPU(PRP) k/CPU(HS) k/CPU(QN)
    1 n=5 9.9872 2.1905e006 10/0.4713 8/0.2964 13/0.5025
    1 n=10 17.7657 1.0948e005 6/0.3855 6/0.4311 8/0.4903
    1 n=50 81.6401 6.0182e007 7/2.1065 7/2.0021 9/2.3250
    1 n=100 158.1896 2.7328e005 11/25.4710 10/21.1207 18/30.6260
    1 n=200 311.3130 1.0823e007 10/85.5025 8/70.1025 13/100.4183
    2 n=100 132.1072 8.2012e005 11/45.7262 9/39.1840 11/50.2500
    2 n=200 405.2981 3.2909e006 18/135.3550 16/123.0931 20/150.2125
    3 n=5 13.0792 3.9828e005 6/0.1750 4/0.1558 5/0.2910
    3 n=10 49.4905 1.7293e006 7/0.7023 5/0.6500 7/0.6250
    3 n=50 154.9351 4.2950e006 5/35.9826 5/30.8674 7/43.0050
    4 n=3 0.8893 9.0482e007 12/0.4587 10/0.6039 10/0.8214
    5 n=10 43.2760 3.9281e006 8/0.4980 7/0.4058 11/0.5709
    5 n=30 136.2817 2.4615e006 5/4.2160 6/4.0156 8/0.5096
    6 n=5 34.5315 2.9053e006 7/0.3900 5/0.3215 9/0.5600
    6 n=30 164.9086 7.3155e007 9/15.9872 5/12.0060 12/13.4900

     | Show Table
    DownLoad: CSV

    To show the numerical performance of a given optimal method, that the number of iterations (k) and CPU time (CPU) are important factors. So, we employ the profiles introduced by Dolan and Morˊe [28] to analyze the efficiency of ACG, PRP, HS [20], QN [29] ATTCG and ADL[30,31] methods, with the following conjugate gradient parameters, respectively,

    βPRPk+1=gTk+1ykgk2,βHSk+1=gTk+1ykdTkyk,βCDk+1=gk+12dTkgk.

    Let Y and W be the set of methods and test problems, ny, nw be the number of methods and test problems, respectively. The performance profile ψ:R[0,1] is for each yY and wW defined that aw,y>0 is k or CPU required to solve problems w by method y. Furthermore, the performance profile is obtained by

    ψy(τ)=1nwsize{wW:rw,yτ},

    where τ>0, size{} is the number of the elements in a set, and rw,y is the performance ratio defined as

    rw,y=aw,ymin{aw,y:yY}.

    In a performance profile plot, the top curve is a method that solved most problems in a time that is within a factor of best time. The horizontal axis gives the percentage (τ) of the test problems for which a method is the fastest (efficiency), while the vertical side gives the percentage (ψ) of the test problems that are successfully solved by each of the methods. If program runs failure, or the number of iterations can reach more than 500, it is regarded as failed. And we denote the number of iterations by 500 and CPU time by 200 seconds. In this way, only ACG algorithm can solve all test problems.

    As can be seen from Figure 1 shows the CPU time performance of the ACG algorithm and the other algorithms. It can be seen from the figure that when τ>3, the curves of the ACG algorithm and the ATTCG algorithm are similar, but when τ>3.5, both the ACG algorithm and ADL algorithm tend to be stable and coincide. Figure 2, the ACG algorithm is better than other algorithms in terms of the number of iterations, especially when τ>2, the curve of ACG algorithm becomes stable, which indicates that ACG algorithm can solve the problem only with fewer iterations. Therefore, Figures 1 and 2 show that the ACG algorithm proposed in this paper converge to the solution quickly.

    Figure 1.  The performance profile for the CPU time.
    Figure 2.  The performance profile for the number of iterations.

    We constructed the unconstrained optimization problems with a shifted parameter. Based on the shifted unconstrained optimization problems, we presented an accelerated conjugate gradient method by using the quasi-Newton direction for solving them. Furthermore, we showed the global convergence analysis of the proposed algorithm. Numerical experiments demonstrated that our method has good numerical performance. We further highlight that the proposed algorithm can be used in other fields, such as the symmetric system of nonlinear equations. It is vital to note that some new methods with random technology will be taken into account in our future work.

    National Natural Science Foundation of China (12061087), the Scientific and Technological Developing Scheme of Jilin Province (YDZJ202101ZYTS167, YDZJ202201ZYTS303), the Yunnan Provincial Ten Thousands Plan Young Top Talents, the Project of education department of Jilin Province (JJKH20210030KJ).

    The authors declare no conflicts of interest.



    [1] L. Lim, Singular values and eigenvalues of tensors: a variational approach, In: Computational Advances in Multi-Sensor Adaptive Processing, 2005, 8912515. https://doi.org/10.1109/CAMAP.2005.1574201
    [2] L. Qi, Eigenvalues of a real supersymmetric tensor, J. Symb. Comput., 40 (2005), 1302–1324. https://doi.org/10.1016/j.jsc.2005.05.007 doi: 10.1016/j.jsc.2005.05.007
    [3] M. Cao, Q. Huang, Y. Yang, A self-adaptive trust region method for extreme B-eigenvalues of symmetric tensors, Numer. Algor., 81 (2019), 407–420. https://doi.org/10.1007/s11075-018-0554-7 doi: 10.1007/s11075-018-0554-7
    [4] C. Hao, C. Cui, Y. Dai, A sequential subspace projection method for extreme Z-eigenvalues of supersymmetric tensors, Numer. Linear Algebra Appl., 22 (2015), 283–298. https://doi.org/10.1002/nla.1949 doi: 10.1002/nla.1949
    [5] S. Aji, P. Kumam, A. Awwal, M. M. Yahaya, K. Sitthithakerngkiet, An efficient DY-type spectral conjugate gradient method for system of nonlinear monotone equations with application in signal recovery, AIMS Math., 6 (2021), 8078–8106. http://dx.doi.org/10.3934/math.2021469 doi: 10.3934/math.2021469
    [6] T. Kolda, J. Mayo, Shifted power method for computing tensor eigenpairs, SIAM J. Matrix Anal. Appl., 32 (2011), 1095–1124. https://doi.org/10.1137/100801482 doi: 10.1137/100801482
    [7] T. Kolda, J. Mayo, An adaptive shifted power methods for computing generalized tensor eigenpairs, SIAM J. Matrix Anal. Appl., 35 (2014), 1563–1581. https://doi.org/10.1137/140951758 doi: 10.1137/140951758
    [8] Y. Chen, M. Cao, Y. Yang, Q. Huang, An adaptive trust-region method for generalized eigenvalues of symmetric tensors, J. Comput. Math., 39 (2021), 533–549.
    [9] G. Li, L. Qi, G. Yu, Semismoothness of the maximum eigenvalue function of a symmetric tensor and its application, Linear Algebra Appl., 438 (2013), 813–833. https://doi.org/10.1016/j.laa.2011.10.043 doi: 10.1016/j.laa.2011.10.043
    [10] M. Cao, Y. Yang, T. Hou, C. Li, A nonmonotone accelerated Levenberg-Marquardt method for the B-eigenvalues of symmetric tensors, Inter. Trans. Oper. Res., 29 (2022), 113–129. https://doi.org/10.1111/itor.12954 doi: 10.1111/itor.12954
    [11] M. Ng, L. Qi, G. Zhou, Finding the largest eigenvalue of a nonnegative tensor, SIAM J. Matrix Anal. Appl., 31 (2009), 1090–1099. https://doi.org/10.1137/09074838X doi: 10.1137/09074838X
    [12] J. Bai, W. Hager, H. Zhang, An inexact accelerated stochastic ADMM for separable composite convex optimization, Comput. Optim. Appl., 81 (2022), 479–518. https://doi.org/10.1007/s10589-021-00338-8 doi: 10.1007/s10589-021-00338-8
    [13] J. Bai, D. Han, H. Sun, H. Zhang, Convergence on a symmetric accelerated stochastic ADMM with larger stepsizes, CSIAM Trans. Appl. Math., 3 (2022), 448–479.
    [14] L. Qi, W. Sun, Y. Wang, Numerical multilinear algebra and its applications, Front. Math. China, 2 (2017), 501–526. https://doi.org/10.1007/s11464-007-0031-4 doi: 10.1007/s11464-007-0031-4
    [15] T. Wei, P. Goldbart, Geometric measure of entanglement and applications to bipartite and multipartite quantum states, Phys. Rev. A, 68 (2003), 042307. https://doi.org/10.1103/PhysRevA.68.042307 doi: 10.1103/PhysRevA.68.042307
    [16] L. Qi, G. Yu, E. Wu, Higher order positive semi-definite diffusion tensor imaging, SIAM J. Imaging Sci., 3 (2010), 416–433. https://doi.org/10.1137/090755138 doi: 10.1137/090755138
    [17] L. Qi, F. Wang, Y. Wang, Z-eigenvalue methods for a global polynomial optimization problem, Math. Program., 118 (2009), 301–316. https://doi.org/10.1007/s10107-007-0193-6 doi: 10.1007/s10107-007-0193-6
    [18] L. Han, An unconstrained optimization approach for finding real eigenvalues of even order symmetric tensors, Numer. Algebr. Control Optim., 3 (2013), 583–599.
    [19] C. Hao, C. Cui, Y. Dai, A feasible trust-region method for calculating extreme Z-eigenvalues of symmetric tensors, Pac. J. Optim., 11 (2015), 291–307.
    [20] J. Nocedal, S. J. Wright, Numerical optimization, New York: Springer, 1999.
    [21] Y. Yuan, W. Sun, Optimization theory and methods, Beijing: Science Press, 1997.
    [22] G. Auchmuty, Globally and rapidly convergent algorithm for symmetric eigenproblems, SIAM J. Matrix Anal. Appl., 12 (1991), 690–706. https://doi.org/10.1137/0612053 doi: 10.1137/0612053
    [23] A. Peressini, F. Sullivan, J. Uhl, The mathematics of nonlinear programmming, Berlin: Springer, 1988.
    [24] A. Neculai, A scaled BFGS preconditioned conjugate gradient algorithm for unconstrained optimization, Appl. Math. Lett., 20 (2007), 645–650. https://doi.org/10.1016/j.aml.2006.06.015 doi: 10.1016/j.aml.2006.06.015
    [25] A. Neculai, Scaled memoryless BFGS preconditioned conjugate gradient algorithm for unconstrained optimization, Optim. Meth. Soft., 22 (2007), 561–571. https://doi.org/10.1080/10556780600822260 doi: 10.1080/10556780600822260
    [26] G. Zoutendijk, Nonlinear programming, computational method, Integer Nonlinear Program., 1970, 37–86.
    [27] C. Cui, Y. Dai, J. Nie, All real eigenvalues of symmetric tensors, SIAM J. Matrix Anal. Appl., 4 (2014), 1582–1601. https://doi.org/10.1137/140962292 doi: 10.1137/140962292
    [28] E. Dolan, J. J. Moré, Benchmarking optimization software with performance profiles, Math. Program., 91 (2002), 201–213. https://doi.org/10.1007/s101070100263 doi: 10.1007/s101070100263
    [29] M. Zeng, Q. Ni, Quasi-Newton method for computing Z-eigenvalues of a symmetric tensor, Pac. J. Optim., 2 (2015), 279–290.
    [30] X. Dong, D. Han, Z. Dai, L. Li, J. Zhu, An accelerated three-term conjugate gradient method with sufficient descent condition and conjugacy condition, J. Optim. The. Appl., 179 (2018), 944–961. https://doi.org/10.1007/s10957-018-1377-3 doi: 10.1007/s10957-018-1377-3
    [31] B. Ivanov, G. Milovanoviˊc, P. Stanimiroviˊc, Accelerated Dai-Liao projection method for solving systems of monotone nonlinear equations with application to image deblurring, J. Glob. Optim., 85 (2023), 377–420. https://doi.org/10.1007/s10898-022-01213-4 doi: 10.1007/s10898-022-01213-4
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1731) PDF downloads(76) Cited by(0)

Figures and Tables

Figures(2)  /  Tables(3)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog