Processing math: 40%
Research article

A two-step Ulm-Chebyshev-like Cayley transform method for inverse eigenvalue problems with multiple eigenvalues

  • Received: 27 May 2024 Revised: 12 July 2024 Accepted: 15 July 2024 Published: 25 July 2024
  • MSC : 15A18, 65F15, 65F18

  • Our focus in this study was on examining the convergence problem of a novel method, inspired by the Ulm-Chebyshev-like Cayley transform method, which was designed to solve the inverse eigenvalue problems (IEPs) with multiple eigenvalues. Compared with other existing methods, the proposed method has higher convergence order and/or requires less operations. Under the assumption that the relative generalized Jacobian matrices at a solution are nonsingular, the proposed method was proved to be convergent with cubic convergence. Experimental findings demonstrated the practicality and efficiency of the suggested approaches.

    Citation: Wei Ma, Zhenhao Li, Yuxin Zhang. A two-step Ulm-Chebyshev-like Cayley transform method for inverse eigenvalue problems with multiple eigenvalues[J]. AIMS Mathematics, 2024, 9(8): 22986-23011. doi: 10.3934/math.20241117

    Related Papers:

    [1] Wei Ma, Ming Zhao, Jiaxin Li . A multi-step Ulm-Chebyshev-like method for solving nonlinear operator equations. AIMS Mathematics, 2024, 9(10): 28623-28642. doi: 10.3934/math.20241389
    [2] Batirkhan Turmetov, Valery Karachik . On solvability of some inverse problems for a nonlocal fourth-order parabolic equation with multiple involution. AIMS Mathematics, 2024, 9(3): 6832-6849. doi: 10.3934/math.2024333
    [3] Jiao Xu, Yinlan Chen . On a class of inverse palindromic eigenvalue problem. AIMS Mathematics, 2021, 6(8): 7971-7983. doi: 10.3934/math.2021463
    [4] Mingyuan Cao, Yueting Yang, Chaoqian Li, Xiaowei Jiang . An accelerated conjugate gradient method for the Z-eigenvalues of symmetric tensors. AIMS Mathematics, 2023, 8(7): 15008-15023. doi: 10.3934/math.2023766
    [5] Shixian Ren, Yu Zhang, Ziqiang Wang . An efficient spectral-Galerkin method for a new Steklov eigenvalue problem in inverse scattering. AIMS Mathematics, 2022, 7(5): 7528-7551. doi: 10.3934/math.2022423
    [6] Liangkun Xu, Hai Bi . A multigrid discretization scheme of discontinuous Galerkin method for the Steklov-Lamé eigenproblem. AIMS Mathematics, 2023, 8(6): 14207-14231. doi: 10.3934/math.2023727
    [7] Latifa I. Khayyat, Abdullah A. Abdullah . The onset of Marangoni bio-thermal convection in a layer of fluid containing gyrotactic microorganisms. AIMS Mathematics, 2021, 6(12): 13552-13565. doi: 10.3934/math.2021787
    [8] Lingling Sun, Hai Bi, Yidu Yang . A posteriori error estimates of mixed discontinuous Galerkin method for a class of Stokes eigenvalue problems. AIMS Mathematics, 2023, 8(9): 21270-21297. doi: 10.3934/math.20231084
    [9] Yalçın Güldü, Ebru Mişe . On Dirac operator with boundary and transmission conditions depending Herglotz-Nevanlinna type function. AIMS Mathematics, 2021, 6(4): 3686-3702. doi: 10.3934/math.2021219
    [10] Erdal Bas, Ramazan Ozarslan, Resat Yilmazer . Spectral structure and solution of fractional hydrogen atom difference equations. AIMS Mathematics, 2020, 5(2): 1359-1371. doi: 10.3934/math.2020093
  • Our focus in this study was on examining the convergence problem of a novel method, inspired by the Ulm-Chebyshev-like Cayley transform method, which was designed to solve the inverse eigenvalue problems (IEPs) with multiple eigenvalues. Compared with other existing methods, the proposed method has higher convergence order and/or requires less operations. Under the assumption that the relative generalized Jacobian matrices at a solution are nonsingular, the proposed method was proved to be convergent with cubic convergence. Experimental findings demonstrated the practicality and efficiency of the suggested approaches.



    Let A0, A1,,An be n+1 real symmetric n-by-n matrices. For any c=(c1,c2,,cn)T Rn such that the eigenvalues {λi(A(c))}ni=1 of the matrices

    A(c)A0+ni=1ciAi (1.1)

    with the order λ1(A(c))λ2(A(c))λn(A(c)). In this note, the inverse eigenvalue problem (IEP) defined here is, for the given n real numbers {λi}ni=1 with the order λ1λ2λn, to find a vector cRn such that

    λi(A(c))=λi,i=1,,n. (1.2)

    The IEP is utilized in a wide range of fields including the inverse Toeplitz eigenvalue problem [1,2,3], structural dynamics [4], molecular spectroscopy [5], the pole assignment problem [6], the inverse Sturm-Liouville problem [7], and also problems in mechanics applications [8,9], structural integrity assessments [10], geophysical studies [11], particle physics research [12], numerical analysis [13], and dynamics systems [14]. For further insights into the diverse practical uses, underlying mathematical principles, and computational techniques of general IEPs, readers may consult the comprehensive review articles [15,16] and the relevant literature [17,18].

    The IEP (1.2) can be represented mathematically through a set of non-linear equations:

    f(c):=(λ1(A(c))λ1,λ2(A(c))λ2,,λn(A(c))λn)T=0. (1.3)

    In situations where the given eigenvalues are distinct, i.e.,

    λ1<λ2<<λn. (1.4)

    Newton's method can be applied to nonlinear equation (1.3) with (1.4). However, as noted in [19,20,21], Newton's method has the following two disadvantages: (ⅰ) It requires computing the exact eigenvectors at each iteration; (ⅱ) It requires solving a Jacobian equation at each iteration. These two facts make it inefficient from the point of numerical computations especially when the problem size n is large. Thus, a focus was placed on avoiding on the disadvantages: (ⅰ) The Newton-like method was proposed in [22,23] which computed the approximate eigenvectors instead of the exact eigenvectors. The quadratic convergence rate of this type of Newton-like method was re-proved in [24]. To alleviate the over-solving problem, Chen et al. proposed in [25] an inexact Newton-like method, which stopped the inner iterations before convergence. (ⅱ) Shen et al. proposed in [26,27] an Ulm-type method, which avoided solving approximate Jacobian equation at each outer iteration and hence could reduce the instability problem caused by the possible ill-conditioning in solving an approximate Jacobian equation.

    Note that all of the methods mentioned above are quadratically convergent. In order to speed up the convergence rate of the methods, Chen et al. [28] proposed a super quadratic convergent two-step Newton-type method where the approximate Jacobian equations are solved by inexact methods. In view of this difficulty, Wen et al. proposed, in [29], a two-step inexact Newton-Chebyshev-like method with cubic root-convergence rate, in which the approximate eigenvectors were obtained by applying the one-step inverse power method and avoided solving the approximate Jacobian equations by using the Chebyshev method to approach the inverse of the Jacobian matrix. In 2022, Wei Ma designed a two-step Ulm-Chebyshev-like Cayley transform method [30] which utilized a Cayley transform to find the approximate eigenvectors. However, the convergence analysis for the above methods became ineffective in the absence of distinct eigenvalues, due to the breakdown of fs differentiability and the eigenvectors' continuity for multiple eigenvalues [22]. When multiple eigenvalues are present, all of the numerical methods in the mentioned references above are quadratic convergent, which extends to the case of multiples.

    In this paper, motivated by [30], we propose a two-step Ulm-Chebyshev-like Cayley transform method for solving the IEP (1.2). Further exploration involves analyzing the performance of the newly introduced two-step Ulm-Chebyshev-like method in the presence of multiple eigenvalues. Under the assumption similar to the one used in a previous study, by the Rayleigh quotient as an approximate eigenvalue of the symmetric matrix and the estimates of eigenvalues, eigenvectors, and the relative generalized Jacobian, we show that the proposed method is still cubically convergent. Numerical experiments show the efficiency of our method and comparisons with some known methods are made.

    The structure of this paper is as follows. We give some notations and preliminary results of the relative generalized Jacobian and some useful lemmas in Section 2. A novel method, the two-step Ulm-Chebyshev-like Cayley transform approach, is introduced in Section 3 and our main convergence theorems are established for the new method in Section 4. Experimental results are presented in the final section.

    Let n be a positive integer. Let Rn represent an n-dimensional Euclidean space, S be a subset of Rn, and clS represent the closure of S. Usually, we use B(x,δ1) to represent the empty sphere of Rn center xRn and radius δ1>0. Let and F represent the Euclidean vector norms or their corresponding induced matrix norms and Frobenius norms, respectively. I is the identity matrix of appropriate dimensions. Then, by (2.3.7) in [31], we have

    AAFnA,for eachARn.

    We define

    K=6β2λ,N=(n2t2)maxi[1,n1]1λi+1λi,H1=8n32ξβρ0max1jnAj1(δτ)2, (2.1)
    C=max{2+2β+12Nβλ, 2Nmax1jnAj}, ρ=max{2n(2β+β2+2NK+12βC), 3nC}, (2.2)
    α1=8n32βρ0max1jnAj, γ=K1H1δ, α2=1+8n32γKρ20, (2.3)
    α3=ρ(α2+4nρ20),α4=α3+ρ0α2,α6=1+2γα1,α5=6nβ2(nmax1jnAj+Kn+λ)α23,α7=2γα5+α2α6,δ2=min{ϵ0,1ρ,1β}, (2.4)
    τ=min{1,1α7,nρ0ρ3(α7+α23),1(1+2γα1)3,2H1} (2.5)

    and

    0<δ=min{μ, δ0, δ2, τ2, δ22nρ0, δ2α2, δ2α3, δ2α7, 1γα1}, (2.6)

    where δ0 and ρ0 are defined in Lemma 2.1, β and ϵ0 are defined in Lemma 2.2, and λ is defined in (2.9).

    A locally Lipschitz continuous function h:RnRm is considered. The Jacobian of h, denoted as h, whenever it exists, and Dh represents the set of differentiable points of h. Moreover, the B-differential Jacobian of h at xRn is denoted according to [32].

    Bh(x):={URm×n|U=limxkxh(xk), xkDh}.

    Considering the composite nonsmooth function h:=φψ, in which φ:RtRm is nonsmooth and ψ:RnRt is continuously differentiable, the generalized Jacobian Qh(x) [33] and relative generalized Jacobian Q|Sh(x) [34] are respectively defined by

    Qh(x):=B(φ(ψ(x)))ψ(x)

    and

    Q|Sh(x):={U|U is a limit of UiQh(yi), yiS, yix}.

    For cRn, write

    Λ(c):=diag(λ1(c),...,λn(c)),

    and define

    Q(c):={Q(c)|Q(c)TQ(c)=I and Q(c)TA(c)Q(c)=Λ(c)}. (2.7)

    By (1.3) and the concept of a generalized Jacobian to f, we have [34]

    Qf(c)={J|[J]ij=qi(c)TAjqi(c), where  [q1(c),...,qn(c)]Q(c)}.

    In particular, if J(c) is a singleton, we write Qf(c)={J(c)}. Let

    S:={cRn|A(c) has distinct eigenvalues}.

    Then, let cS and f be continuously differentiable at c. Moreover,

    Qf(c)={J(c)}={f(c)}.

    Thus, we get the following relative generalized Jacobian [34]:

    Q|Sf(c)={J|J=limk+J(yk) with {yk}S and ykc}.

    Throughout this paper, let the given eigenvalues {λi}ni=1 satisfy λ1λ2λn. For simplicity, without loss of generality, we assume that

    λ1=λ2==λt<λt+1<<λn, (2.8)

    where 1tn. Write

    Λ=diag(λ1,λ2,...,λn)andλ=(λ1,λ2,...,λn)T. (2.9)

    Then a solution of the IEP (1.2) can be written by c and Q(c) as

    Q(c)TA(c)Q(c)=Λ, (2.10)

    where Q(c) is an orthogonal matrix. Recall that Q(c) can be defined by (2.7). Let Q(c)Q(c) and write Q(c)=[Q(1)(c),Q(2)(c)] in which Q(1)(c)Rn×t and Q(2)(c)Rn×(nt). Let c be the solution of the IEPs with (2.8). Define

    Π=Q(1)(c)Q(1)(c)T.

    Clearly, Π is the eigenprojection of A(c) for λ1 in (2.8). Given an orthogonal matrix P=[P(1),P(2)], where P(1)Rn×t and P(2)Rn×(nt), we obtain the QR factorization of ΠP(1) by

    ΠP(1)=˜Q(1)(c)R, (2.11)

    where R is a t×t nonsingular upper triangular matrix and ˜Q(1)(c) is an n×t matrix whose columns are orthonormal. Let

    ˜Q(c):=[˜Q(1)(c),Q(2)(c)]. (2.12)

    Obviously, ˜Q(c)Q(c). Moreover, we define the error matrix

    E:=[P(1)ΠP(1),P(2)Q(2)(c)].

    Now, we state the following two lemmas, which are usefull for our proof.

    Lemma 2.1. [35,36] Let cRn and the eigenvalues of the matrix A(c) satisfy (2.8). Then, there exist two positive numbers δ0 and ρ0 such that, for each cB(c,δ0) and [Q(1)(c), Q(2)(c)]Q(c), we get

    A(c)A(c)max1jnAjcc,
    Λ(c)Λρ0cc,
    Q(2)(c)Q(2)(c)ρ0cc

    and

    (IΠ)Q(1)(c)ρ0cc.

    Lemma 2.2. There exist two positive numbers ϵ0 and β such that, for any orthogonal matrix P=[P(1),P(2)], if E=P(1)ΠP(1), P(2)Q(2)(c)ϵ0 and the skew-symmetric matrix X defined by eX=PT˜Q(c), then we get

    XFβEandX(11)FβE2, (2.13)

    in which X(11) is the t-by-t leading block of X. Moreover, if XF<1, then

    l=2Xl2l!F1, l=2(X)l2l!F1, l=1(X)l1l!F2, and l=0(X)ll!F3. (2.14)

    Proof. (2.13) can be found in [22,35,36]. Noting that

    l=21l!l=21l(l1)=1.

    If XF1, we get

    l=2Xl2l!F1andl=2(X)l2l!F1,

    which implies that

    l=1(X)l1l!F2andl=0(X)ll!F3.

    We first recall the given eigenvalues in (2.8). Suppose that Pk is the current estimate of Q(c) and Yk is a skew-symmetric matrix, i.e., YTk=Yk. Let us write Q(c)=PkeYk. Then, by using the Taylor series of the exponential function, we can express (2.10) as

    PTkA(c)Pk=eYkΛeYk=(I+Yk+12Y2k+)Λ(1Yk+12Y2k+).

    The vector ck is updated as ck+1 by neglecting the second–order term of the above equality in Yk as

    PTkA(ck+1)Pk=Λ+YkΛΛYk. (3.1)

    We obtain ck+1 by equating the diagonal elements in (3.1) as

    Jkck+1=λbk,

    in which, Jk and bk are defined by

    [Jk]ij=(pki)TAjpki,1i,jnand[b]ki=(pki)TA0pki,1in.

    On the other hand, equating the off-diagonal in (3.1),

    (pki)TA(ck+1)pkj=[Yk]ij(λjλi),  for each i,j[1,n] and ij, (3.2)

    and, assuming that the given eigenvalues are defined in (2.8), we obtain the skew-symmetric matrix Yk as

    [Yk]ij={0,                  for each 1i,jt, or i=j;(pki)TA(ck+1)pkjλjλi,  for each t+1in or t+1jn and ij. (3.3)

    Furthermore, by using the Cayley transform, we calculate the matrix Pk+1 as

    Pk+1=Pk(I+12Yk)(I12Yk)1. (3.4)

    Finally, by (3.3), (3.4), and the two-step Ulm-Chebyshev iterative procedure[30], we can propose the following two-step Ulm-Chebyshev-like Cayley transform method for solving the IEP with multiple eigenvalues.

    Algorithm Ⅰ: The two-step Ulm-Chebyshev-like Cayley transform method

    Step 1. Given c0Rn, calculate the orthogonal eigenvectors {qi(c0)}ni=1 of A(c0). Let

    P0=[p01,p02,,p0n]=[q1(c0),q2(c0),,qn(c0)],

    and J0=J(c0) and the vector b0 are defined as follows:

    [J0]ij=(p0i)TAjp0i,1i,jn,[b]0i=(p0i)TA0p0i,1in. (3.5)

    Let B0Rn×n satisfy

    IB0J(c0)μ,

    where μ is a positive constant.

    Step 2. For k=0, until convergence, do:

    (a) Calculate yk by

    yk=ckBk(Jkck+bkλ). (3.6)

    (b) Form the skew-symmetric matrix Yk:

    [Yk]ij={0,for 1i,jt, or i=j;(pki)TA(yk)pkjλjλi,   for t+1in or t+1jn and ij, (3.7)

    where the matrix A(yk) is defined by (1.1).

    (c) Calculate P(yk)=[p1(yk),p2(yk),,pn(yk)]T=[vk1,vk2,,vkn]T by solving

    (I+12Yk)vkj=hkj,for 1jn, (3.8)

    where hkj is the jth column of Hk=(I12Yk)PTk.

    (d) Calculate the approximate eigenvalues of A(yk) via

    ˆλi(yk)=(pi(yk))TA(yk)pi(yk),for 1in.

    (e) Calculate ck+1 by

    ck+1=ykBk(ˆλ(yk)λ). (3.9)

    (f) Form the skew-symmetric matrix ˆYk:

    [ˆYk]ij={0, for 1i,jt, or i=j;(pi(yk))TA(ck+1)(pj(yk))λjλi,for t+1in or t+1jn and ij,

    where the matrix A(ck+1) is defined by (1.1).

    (g) Calculate Pk+1=[pk+11,pk+12,,pk+1n]T=[ˆvk1,ˆvk2,,ˆvkn]T by solving

    (I+12ˆYk)ˆvkj=ˆhkj,for 1jn, (3.10)

    where ˆhkj is the jth column of ˆHk=(I12ˆYk)(P(yk))T.

    (h) Form the matrix Jk+1 and the vector bk+1:

    [Jk+1]ij=(pk+1i)TAjpk+1i,1i,jn,
    [b]k+1i=(pk+1i)TA0pk+1i,1in. 

    (i) Calculate the Chebyshev matrices Bk+1 by

    Bk+1=Bk+Bk(2IJ(ck+1)Bk)(IJ(ck+1)Bk).

    Remark 3.1. For k=0,1,2,, from (c) and (g) in Step 2 in Algorithm, we have

    P(yk)=Pk(I+12Yk)(I12Yk)1 (3.11)

    and

    Pk+1=P(yk)(I+12ˆYk)(I12ˆYk)1. (3.12)

    Remark 3.2. Without the distinction of the given eigenvalues, the convergence analysis of the two-step Ulm-Chebyshev-like method in [30] cannot work properly due to the differentiability of f and the discontinuity of the eigenvectors corresponding to multiple eigenvalues [22]. Based on the relative generalized Jacobian of eigenvalue function [32], we propose the improved method for solving the IEP (1.2) with multiple eigenvalues. Clearly, in the case when t=1, the method presented below is reduced to the two-step Ulm-Chebyshev-like method proposed in [30] for the distinct case.

    In this section, we shall analyze the convergence of Algorithm Ⅰ. To ensure the cubical convergence, it is assumed that all JQf(c) are nonsingular for robustness. Yet, a suitable choice of eigenvectors can render J nonsingular in a general manner. Therefore, we assume that all JQ|Sf(c) are nonsingular.

    Let ck, yk, Yk, ˆYk, Pk, P(yk), Jk and Bk be generated by Algorithm Ⅰ with initial point c0. For k=0,1,2,, let

    Ek:=[P(1)kΠP(1)kP(2)kQ(2)(c)] (4.1)

    and

    E(yk):=[P(yk)(1)ΠP(yk)(1)P(yk)(2)Q(2)(c)]. (4.2)

    Then, we can get the following lemmas.

    Lemma 4.1. Let the given eigenvalues {λi}ni=1 be defined as in (2.8). Then, in Algorithm Ⅰ, there exists a number 0<δ21 such that for k=0,1,2,, if ykcδ2, ck+1cδ2, Ekδ2 and E(yk)δ2, then

    Λ+XkΛΛXkPTkA(c)PkKEk2, (4.3)
    P(yk)Pkρ(ykc+Ek), (4.4)
    E(yk)ρ(ykc+Ek2), (4.5)
    Λ+YkΛΛYkP(yk)TA(c)P(yk)<KE(yk)2, (4.6)
    Pk+1P(yk)ρ(ck+1c+E(yk)) (4.7)

    and

    Ek+1ρ(ck+1c+E(yk)2), (4.8)

    where K and ρ are defined by (2.1) and (2.2), respectively.

    Proof. Let eXk:=PTk˜Q(c), where Xk is the skew-symmetric matrix and ˜Q(c) is defined by (2.11) and (2.12) with P=Pk. By Ekδ2ϵ0 and Lemma 2.2, we have

    XkFβEkandX(11)kFβEk2, (4.9)

    where β is a positive number and X(11)k is the t-by-t leading block of Xk. On the other hand, by ˜Q(c)Q(c), we derive

    eXkΛeXk=PTkA(c)Pk. (4.10)

    Thus, by the fact of eXk=l=0(Xlkl!), we can express (4.10) as

    Λ+XkΛΛXk=PTkA(c)Pk+R(Xk), (4.11)

    where

    R(Xk)=X2kl=2(Xl2kl!)Λ(l=0(Xk)ll!)+ΛX2kl=2(Xk)l2l!XkΛXkl=1(Xk)l1l!.

    By (2.1) and Lemma 2.2, we get

    R(Xk)F6Xk2FΛF=6Xk2Fλ6β2λEk2KEk2. (4.12)

    Thus, (4.3) is seen to hold by (4.11) and (4.12).

    In order to prove (4.4) and (4.5), assume that ykcδ2. We note by (4.11) that

    [Xk]ij=1λjλi(pki)TA(c)pkj+[R(Xk)]ij,

    where t+1in, 1jn,i>j. Combining this with (3.7), we have

    [Xk]ij[Yk]ij=1λjλi(pki)T[A(c)A(yk)]pkj+[R(Xk)]ij,

    in which t+1in, 1jn,i>j. By (4.12), Lemma 2.1, and the fact that {pki}ni=1 are orthogonal, we get

    maxt+1in, 1jn,ij|[Xk]ij[Yk]ij|max1in11λi+1λi×(max1jnAjykc+KEk2).

    In addition, by the fact that [Yk]ij=0, for each i,j[1,t], we have

    XkYkXkYkFX(11)kF+(n2t2)maxi[t+1,n],j[1,n],ij|[Xk]ij[Yk]ij|.

    Then, it follows from (2.1) and (4.9) that

    XkYkβEk2+N(max1jnAjykc+KEk2), (4.13)

    and so

    YkβEk+βEk2+N(max1jnAjykc+KEk2).

    Thus, thanks to the fact that βEkβδ21 and (2.2), one has

    YkNmax1jnAjykc+(1+β+6Nβλ)EkC2(ykc+Ek)ρ2(ykc+Ek). (4.14)

    Since Ekδ2 and ykcδ2, it follows from (2.4) and (4.14) that Yk1. Consequently,

    (I12Yk)11112Yk2. (4.15)

    Therefore, in the following, we estimate P(yk)Pk and E(yk). Indeed, by (3.11),

    P(yk)Pk=Pk[(I+12Yk)(I12Yk)](I12Yk)1=PkYk(I12Yk)1.

    This together with (4.14), (4.15), as well as the orthogonality of Pk indicate that (4.4) holds.

    As for (4.5), we note by (3.11) and Xk that

    P(yk)˜Q(c)=Pk[(I+12Yk)(I12Yk)1eXk]=Pk[(I+12Yk)eXk(I12Yk)](I12Yk)1.

    Combining this with eXk=l=0(Xlkl!), we get

    P(yk)˜Q(c)=Pk[YkXk+12XkYk(X2km=2Xm2km!)(I12Yk)](I12Yk)1=Pk(YkXk+12XkYk)(I12Yk)1PkX2km=2Xm2km!.

    Since Pk is orthogonal, note by (2.14) and (4.15) that

    P(yk)˜Q(c)2YkXk+XkYk+Xk2.

    Thus, we derive by using (2.13), (4.13), and (4.14),

    P(yk)˜Q(c)2βEk2+2N(max1jnAjykc+KEk2)+12βCEk(ykc+Ek)+β2Ek2(2β+β2+2NK+12βC)Ek2+(2Nmax1jnAj+12C)ykcmax1jnAjykc+KE(yk)2,1in, (4.16)

    where the second inequality holds because of the fact that βEk1 while the last inequality holds because of the definition of C. Write P(yk)=[P(yk)(1) P(yk)(2)], where P(yk)(1)Rn×t and P(yk)(2)Rn×(nt). Since (IΠ)˜Q(1)(c)=0, where 0 is a zero matrix, we have

    (IΠ)P(yk)(1)=(IΠ)(P(yk)(1)˜Q(1)(c)+˜Q(1)(c))=(IΠ)(P(yk)(1)˜Q(1)(c))P(yk)˜Q(c)

    and

    P(yk)(2)˜Q(2)(c)P(yk)˜Q(c).

    Hence, by (4.2) and (4.16), we obtain

    E(yk)(IΠ)P(yk)(1)+P(yk)(2)˜Q(2)(c)2nP(yk)˜Q(c)2n[(2β+β2+2NK+12βC)Ek2+32Cykc]. (4.17)

    Therefore, (4.5) is proved by (2.2) and (4.17). We defined eYk:=P(yk)T˜Q(c), where Yk is the skew-symmetric matrix. Similarly, (4.6)–(4.8) also hold.

    Lemma 4.2. Let ρ0 and δ0 be defined in Lemma 2.1. If ykcδ0 and ˆλ(yk)=(ˆλ1(yk),ˆλ2(yk),,ˆλn(yk))T, in which ˆλi(yk)=(pi(yk))TA(yk)pi(yk), 1in, then

    ˆλ(yk)nmax1jnAjykc+KnE(yk)2+λ. (4.18)

    Proof. From the diagonal elements of Λ+YkΛΛYkP(yk)TA(c)P(yk), we obtain from (4.6) that

    |(pi(yk))TA(c)pi(yk)λi|KE(yk)2,for 1in,

    which together with Lemma 3.1 gives

    |(pi(yk))TA(yk)pi(yk)λi|=|(pi(yk))T(A(yk)A(c))pi(yk)+(pi(yk))TA(c)pi(yk)λi||(pi(yk))T(A(yk)A(c))pi(yk)|+|(pi(yk))TA(c)pi(yk)λi|max1jnAjykc+KE(yk)2,1in.

    Therefore,

    ˆλ(yk)λnmax1jnAjykc+KnE(yk)2,

    which together with the fact that ˆλ(yk)ˆλ(yk)λ+λ, we can get (4.18).

    Lemma 4.3. Let the Jacobian matrix of ˜J(c) and the vector ˜b be defined as follows:

    [˜J(c)]ij=˜qi(c)TAj˜qi(c),1i,jnand[˜b]i=(˜qi(c))TA0˜qi(c),1in.

    Then, we have

    ˜J(c)yk+˜bˆλ(yk)6nβ2ˆλ(yk)E(yk)2. (4.19)

    Proof. Let eYk:=P(yk)T˜Q(c) where Yk is the skew-symmetric matrix and ˜Q(c) is defined by (2.11) and (2.12) with P=P(yk). By E(yk)δ2ϵ0 and Lemma 2.2, we get

    YkFβE(yk),

    where β is a positive number. By ˆΛ(yk)=P(yk)TA(yk)P(yk) and eYkˆΛ(yk)eYk=˜Q(c)TA(yk)˜Q(c), we have

    ˆΛ(yk)YkˆΛ(yk)+ˆΛ(yk)Yk=˜Q(c)TA(yk)˜Q(c)+R(Yk),

    where

    R(Yk)=(Yk)2l=2((Yk)l2l!)ˆΛ(yk)(l=0(Yk)ll!)ˆΛ(yk)(Yk)2l=2(Yk)l2l!+YkˆΛ(yk)Ykl=1(Yk)l1l!.

    By Lemma 2.2, we get

    R(Yk)F6Yk2FˆΛ(yk)F=6Yk2Fˆλ(yk)6β2ˆλ(yk)E(yk)2.

    Thus,

    ˜Q(c)TA(yk)˜Q(c)+YkˆΛ(yk)ˆΛ(yk)YkˆΛ(yk)6β2ˆλ(yk)E(yk)2. (4.20)

    By the diagonal entries of ˜Q(c)TA(yk)˜Q(c)+YkˆΛ(yk)ˆΛ(yk)YkˆΛ(yk) and (4.20), we get

    |(˜qi(c))TA(yk)˜qi(c)ˆλi(yk)|<6β2ˆλ(yk)E(yk)2,for 1in.

    Therefore, by the definitions of ˜J(c),ˆλ(yk),˜b and A(yk), we can get (4.19).

    Lemma 4.4. [35] Let J0 be defined in (3.5). Suppose that J0 is invertible. Let k1 such that

    2nJ10max1jnAjPkP0<1. (4.21)

    Then, the matrix Jk is nonsingular and

    J1kJ1012nJ10max1jnAjPkP0.

    Lemma 4.5. Let the vector cclS and the given eigenvalues of the matrix A(c) satisfy (2.8). Let all JQ|Sf(c) be nonsingular. If for c0B(c,δ)S where δ>0 and k=0,1,2,, then there exist three numbers 0<τ1, 0<δ<τ and 0μδ, that if c0cδ, the conditions

    Ek2nρ0τ(δτ)3k, (4.22)
    ckcτ(δτ)3k, (4.23)

    and

    IBkJkτ(δτ)3k (4.24)

    imply

    J1kγ,Bk2γ,andck+1cτ(δτ)3k+1. (4.25)

    Proof. Since all JQ|Sf(c) are nonsingular, and from Theorem 3.2 in [34], for each c0B(c,δ0)S, we have

    supJQ|Sf(c0)J1ξ,

    where ξ>0 and δ0>0. From (2.3), (2.5), and (2.6), we know that

    τ1. (4.26)

    By (2.6) and (4.22), we get

    Ek2nρ0τ(δτ)3k2nρ0δδ2,

    and then, from Lemma 2.2, we obtain

    XkFβEk, (4.27)

    where β>0. By the definitions of Xk and the Taylor series of the exponential function eXk, we have

    Pk˜Q(c)=Pk(IeXk)=Pk(Xk)l=1(X)l1l!.

    Since Pk is orthogonal, note by (2.14) and (4.27) that

    Pk˜Q(c)2Xk2XkF2βEk. (4.28)

    Similarly, we also have

    Pk1˜Q(c)2Xk12Xk1F2βEk1.

    Thus, by (2.5), (4.22), and (4.28), we have

    PkPk1Pk˜Q(c)+Pk1˜Q(c)2βEk+2βEk12β(2nρ0τ(δτ)3k+2nρ0τ(δτ)3k1)4nβρ0τ(δτ)3k1. (4.29)

    Therefore, we further have

    PmP0mk=1PkPk14nβρ0τ[(δτ)+(δτ)3+(δτ)32++(δτ)3m1]. (4.30)

    Since 3n2n+1 for each n0, we obtain from (4.30) that

    PmP04nβρ0τ[(δτ)+(δτ)3+(δτ)5++(δτ)2m1]4nβρ0δ[1(δτ)2m]1(δτ)2,

    which together with (2.1) and (2.6), we obtain

    2nξmax1jnAjPmP08n32ξβρ0δmax1jnAj1(δτ)2=H1δ<12H1τ<1.

    Consequently, using Lemma 4.4, we can derive that Jk is nonsingular and moreover

    J1mξ12nξmax1jnAjPmP0ξ1H1δ=γ. (4.31)

    Furthermore, by (2.5), (2.6), and (4.24), we have

    BkBkJkJ1k(I+IBkJk)J1k(1+τ(δτ)3k)γ(1+τ)γ2γ. (4.32)

    On the other hand, considering the diagonal elements of Λ+XkΛΛXkPTkA(c)Pk, we obtain from (4.3) that

    |(pki)TA(c)pkiλi|KEk2, for 1in.

    Therefore, by the definitions of λ, Jk, bk and A(c), we have

    Jkcλ+bknKEk2. (4.33)

    From (3.6), we get

    ykc=Bk(λbkJkc)+(IBkJk)(ckc).

    It follows that

    ykcBkJkcλ+bk+IBkJkckc,

    which together with (4.22)–(4.24), (4.32), and (4.33) gives

    ykc2γnKEk2+τ(δτ)3kτ(δτ)3k(1+8γKn32ρ20)τ2(δτ)23k:=α2τ2(δτ)23k. (4.34)

    By (2.6), (4.22), and (4.34), we have

    ykcα2δδ21 (4.35)

    and

    Ek2nρ0τ(δτ)3k2nρ0δδ21, (4.36)

    which together with (4.22) and (4.28), we obtain

    pki˜qi(c)Pk˜Q(c)4nβρ0τ(δτ)3k,1in.

    This together with the orthogonality of Pk and ˜Q(c) and the Cauchy-Schwarz inequality indicates that

    |[Jk]ij[˜J(c)]ij|=|(pki)TAjpki˜qi(c)TAj˜qi(c)|=|(pki˜qi(c))TAjpki˜qi(c)TAj(˜qi(c)pki)|2Ajpki˜qi(c)8nβρ0Ajτ(δτ)3k,1i,jn.

    Thus, we get

    Jk˜J(c)Jk˜J(c)F8n32βρ0max1jnAjτ(δτ)3k:=α1τ(δτ)3k. (4.37)

    By (4.24), (4.32), and (4.37), we have

    \begin{eqnarray} \|I-B_{k}\tilde{J}( {\bf c}^{*})\| &\leq& \|I-B_{k}J_{k}\|+\|B_{k}\|\|J_{k}-\tilde{J}( {\bf c}^{*})\| \\ &\leq& \tau\Big(\frac{\delta}{\tau}\Big)^{3^{k}}+2\gamma\alpha_{1}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{k}}\\ &\leq& (1+2\gamma\alpha_{1})\tau\Big(\frac{\delta}{\tau}\Big)^{3^{k}}\\ &: = &\alpha_{6}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{k}}. \end{eqnarray} (4.38)

    By (4.34)–(4.36) and Lemma 4..1, we obtain

    \begin{eqnarray} \|E( {\bf y}^{k})\| &\leq&\rho\Big(\alpha_{2}\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{2\cdot3^{k}}+4n\rho_{0}^{2}\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{2\cdot3^{k}}\Big)\\ &\leq&\rho(\alpha_{2}+4n\rho_{0}^{2})\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{2\cdot3^{k}}\\&: = &\alpha_{3}\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{2\cdot3^{k}}, \end{eqnarray} (4.39)

    which together with (2.6), (4.22), and (4.34), we have

    \begin{eqnarray*} \|E( {\bf y}^{k})\|\leq \alpha_{3}\delta\leq\delta_{2}\leq 1, \end{eqnarray*}

    which together with (4.35) and Lemmas 4.2 and 4..3, we get

    \begin{eqnarray} \|\tilde{J}( {\bf c}^{*}) {\bf y}^{k}+\tilde{ {\bf b}}-\hat{\boldsymbol\lambda}( {\bf y}^{k})\| &\leq& 6\sqrt{n}\beta^{2}\:\|{\hat{\boldsymbol\lambda}}( {\bf y}^{k})\|\|E( {\bf y}^{k})\|^{2}\\ &\leq& 6\sqrt{n}\beta^{2}\Big(\sqrt{n}\max\limits_{1\leq j\leq n}\|A_{j}\|\| {\bf y}^{k}- {\bf c}^{*}\|+ K\sqrt{n}\|E( {\bf y}^{k})\|^{2}+\|\boldsymbol\lambda^{*}\|\Big)\|E( {\bf y}^{k})\|^{2} \\ &\leq& 6\sqrt{n}\beta^{2}\Big(\sqrt{n}\max\limits_{1\leq j\leq n}\|A_{j}\|+ K\sqrt{n}+\|\boldsymbol\lambda^{*}\|\Big)\alpha_{3}^{2}\tau^{4}\Big(\frac{\delta}{\tau}\Big)^{4\cdot3^{k}} \\ &\leq& 6\sqrt{n}\beta^{2}\Big(\sqrt{n}\max\limits_{1\leq j\leq n}\|A_{j}\|+ K\sqrt{n}+\|\boldsymbol\lambda^{*}\|\Big)\alpha_{3}^{2}\tau^{3}\Big(\frac{\delta}{\tau}\Big)^{3^{k+1}} \\ &: = &\alpha_{5}\tau^{3}\Big(\frac{\delta}{\tau}\Big)^{3^{k+1}}. \end{eqnarray} (4.40)

    Together with (3.9) and {\boldsymbol \lambda}^{*} = \tilde{J}({\bf c}^{*}) {\bf c}^{*}+\tilde{ {\bf b}} , we get

    {\bf c}^{k+1}- {\bf c}^{*} = B_{k}(\tilde{J}( {\bf c}^{*}) {\bf y}^{k}+ {\bf b}-\hat{\boldsymbol\lambda}( {\bf y}^{k}))+(I-B_{k}\tilde{J}( {\bf c}^{*}))( {\bf y}^{k}- {\bf c}^{*}).

    It follows from (4.26), (4.32), (4.34), (4.38), and (4.40) that

    \begin{eqnarray} \| {\bf c}^{k+1}- {\bf c}^{*}\| &\leq& \|B_{k}\|\|\tilde{J}( {\bf c}^{*}) {\bf y}^{k}+\tilde{ {\bf b}}-\hat{\boldsymbol\lambda}( {\bf y}^{k})\|+\|I-B_{k}\tilde{J}( {\bf c}^{*})\|\| {\bf y}^{k}- {\bf c}^{*}\| \\ &\leq& 2\gamma\alpha_{5}\tau^{3}\Big(\frac{\delta}{\tau}\Big)^{3^{k+1}}+\alpha_{6}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{k}}\cdot\alpha_{2}\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{2\cdot3^{k}} \\ &\leq& (2\gamma\alpha_{5}+\alpha_{2}\alpha_{6})\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{3^{k+1}} = \tau\Big(\frac{\delta}{\tau}\Big)^{3^{k+1}}. \end{eqnarray} (4.41)

    Finally, by (4.31), (4.32), and (4.41), we have (4.25).

    Next, we can analyze the convergence of Algorithm Ⅰ.

    Theorem 4.1. Let the vector {\bf c}^{*}\in clS and the given eigenvalues \{\lambda_{i}^{*}\}_{i = 1}^{n} satisfy (2.8). All J\in \partial_{Q|S} {\bf f}({\bf c}^*) are nonsingular. Then Algorithm Ⅰ is locally cubic convergent.

    Proof. Let us start by mathematical induction that (4.22)–(4.24) are true for all k > 0 . Clearly, by assumptions \mu\leq\delta and \| {\bf c}^{0}- {\bf c}^{*}\|\leq\delta , (4.23) and (4.24) for k = 0 are trivial. From Lemma 2.1, we have

    \begin{eqnarray*} \|E_{0}\|\leq\|\|E_{0}\|_{F}\leq\|(I-\Pi)Q^{(1)}( {\bf c}^{0})\|_{F}+\|Q^{(2)}( {\bf c}^{0})-Q^{(2)}( {\bf c}^{*})\|_{F}\leq2\sqrt{n}\alpha_{1}\| {\bf c}^{0}- {\bf c}^{*}\|\leq2\sqrt{n}\alpha_{1}\delta, \end{eqnarray*}

    and this gives that (4.22) is true for k = 0 .

    Now assume that (4.22)–(4.24) are true for all k\leq m-1 . Recalling that (2.5) and (2.6), we get

    \|E_{m-1}\|\leq \sqrt{n}\rho_{0}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}\leq\sqrt{n}\rho_{0}\delta\leq\delta_{2},

    which together with Lemma 4.1, (2.5), (2.6), (4.34), and (4.39) with k = m-1 gives

    \|E( {\bf y}^{m-1})\|\leq\alpha_{3}\delta\leq\delta_{2}.

    It follows from Lemma 4.1, (2.5), (4.39), and (4.41) with k = m-1 that

    \begin{eqnarray*} \| {\bf c}^{m}- {\bf c}^{*}\|\leq\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m}}\leq\delta\leq\delta_{2} \end{eqnarray*}

    and

    \begin{eqnarray*} \|E_{m}\| \leq \rho_{3}\Big(\alpha_{7}\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{3^{m}}+\alpha_{3}^{2}\tau^{4}\Big(\frac{\delta}{\tau}\Big)^{4\cdot3^{m-1}}\Big) \leq \rho_{3}(\alpha_{7}+\alpha_{3}^{2})\tau^{2}\Big(\frac{\delta}{\tau}\Big)^{3^{m}} \leq 2\sqrt{n}\rho_{0}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m}}\leq2\sqrt{n}\rho_{0}\delta. \end{eqnarray*}

    Thus, (4.22) and (4.23) hold for k = m . From Lemma 4.5 with k = m-1 , we get \|I-B_{m-1}J_{m-1}\|\leq\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}} and \|B_{m-1}\|\leq2\gamma . Thanks to (4.29) (with k = m ), one can see that

    \begin{eqnarray*} \|P_{m}-P_{m-1}\|\leq 4\sqrt{n}\beta\rho_{0}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}, \end{eqnarray*}

    which implies that

    \begin{eqnarray*} \| {\bf p}_{i}^{m}- {\bf p}_{i}^{m-1}\|\leq 4\sqrt{n}\beta\rho_{0}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}},\quad 1\leq i\leq n. \end{eqnarray*}

    Consequently,

    \begin{eqnarray*} |[J_{m}]_{ij}-[J_{m-1}]_{ij}| & = &|( {\bf p}_{i}^{m}- {\bf p}_{i}^{m-1})^{T}A_{j} {\bf p}_{i}^{m}-( {\bf p}_{i}^{m})^{T}A_{j}( {\bf p}_{i}^{m}- {\bf p}_{i}^{m-1})|\\ &\leq& 2\|A_{j}\|\|( {\bf p}_{i}^{m}- {\bf p}_{i}^{m-1})\|\\ &\leq& 8\|A_{j}\|\sqrt{n}\beta\rho_{0}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}},\ 1\leq i,j\leq n. \end{eqnarray*}

    Hence,

    \begin{eqnarray*} \|J_{m}-J_{m-1}\|\leq\|J_{m}-J_{m-1}\|_{F}\leq 8n^{\frac{3}{2}}\beta\rho_{0}\max\limits_{1\leq j\leq n}\|A_{j}\|\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}: = \alpha_{1}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}. \end{eqnarray*}

    It follows that

    \begin{eqnarray} \|I-B_{m-1}J_{m}\| &\leq& \|I-B_{m-1}J_{m-1}\|+\|B_{m-1}\|J_{m-1}-J_{m}\|\\ &\leq& \tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}+2\gamma\cdot\alpha_{1}\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}\leq(1+2\gamma\alpha_{1})\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m-1}}. \end{eqnarray} (4.42)

    Notice that B_{m} = B_{m-1}+B_{m-1}(2I-J({\bf c}^{m})B_{m-1})(I-J({\bf c}^{m})B_{m-1}) , and we obtain

    I-B_{m}J_{m} = I-(B_{m-1}+B_{m-1}(2I-J( {\bf c}^{m})B_{m-1})(I-J( {\bf c}^{m})B_{m-1}))J_{m} = (I-B_{m-1}J_{m})^{3}.

    Together with (2.5), (4.26), and (4.42), one has

    \begin{eqnarray*} \|I-B_{m}J_{m}\|\leq\|I-B_{m-1}J_{m}\|^{3}\leq(1+2\gamma\alpha_{1})^{3}\tau^{3}\Big(\frac{\delta}{\tau}\Big)^{3^{m}}\leq\tau\Big(\frac{\delta}{\tau}\Big)^{3^{m}}, \end{eqnarray*}

    and therefore, (4.24) is true for k = m and the proof is complete.

    In this section, we present the computational performance of Algorithm Ⅰ in addressing the IEP with several multiple eigenvalues. Algorithm Ⅰ is contrasted with the two-step Ulm-Chebyshev-like Cayley transform method (TUCT method), inexact Cayley transform method (ICT method), and Ulm-like Cayley transform method (UCT method) as presented in [30,35,36], respectively. The tests were carried out in MATLAB 7.10 running on a PC Intel Pentium Ⅳ of 3.0 GHz CPU. We will now consider the problem of finding the eigenvalues for a matrix with a Toeplitz structure, as previously investigated in references [35,36].

    The QMR method [37] was utilized to solve all linear systems in all algorithms, using the MATLAB QMR function and setting the maximum number of iterations to 1000 . Notably, in the context of solving approximate Jacobian equations within the framework of the inexact Cayley transform method, an advanced approach was employed. This approach involved leveraging the preconditioned QMR method with a specific stopping tolerance, and integrating the MATLAB incomplete LU factorization as the designated preconditioner. Specifically, the drop tolerance in LUINC( A , drop-tolerance) is fixed at 0.001 . Furthermore, the precision level for the linear systems involved in all computational methods is adjusted to match the machine accuracy, ensuring the attainment of the desired solutions. The termination criterion for the outer (Newton) iterations is met in every algorithm when

    \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\|\leq10^{-12}.

    Example 5.1. (See [35,36]) Referring to the Toeplitz matrices stated in [35,36], consider \{A_{i}\}_{i = 0}^{n} as

    A_0 = O,\ \ A_1 = I,\ \ A_2 = \left[ \begin{array}{rrrrr} 0 & 1 & 0 & \cdots & 0\\ 1 & 0 & 1 & \ddots &\vdots\\ 0 & 1 & \ddots&\ddots&0\\ \vdots&\ddots&\ddots&0 &1\\ 0 &\cdots& 0 & 1 & 0\\\end{array} \right],\cdots, A_n = \left[ \begin{array}{rrrrr} 0 & 0 & \cdots &0 & 1\\ 0 & \ddots & \ddots & \ddots &0\\ \vdots & \ddots & \ddots&\ddots&\vdots\\ 0&\cdots&\ddots&\ddots &0\\ 1 &0& \cdots & 0 & 0\\\end{array} \right].

    Hence, the matrix A({\bf c}) can be characterized as a symmetric Toeplitz matrix where the first column is identical to the vector {\bf c} .

    Here, we consider three cases: n = 100,200,300 . For any n , we generate a vector \tilde{ {\bf c}}^{*} such that |\lambda_{k+1}(\tilde{ {\bf c}}^{*})-\lambda_{k}(\tilde{ {\bf c}}^{*})| < \eta with 1\leq l\leq n-1 , where

    \eta = \begin{cases} 5\times 10^{-5},\quad&\text{ n = 100 };\\ 1\times 10^{-5},\quad&\text{ n = 200 };\\ 1\times 10^{-6},\quad&\text{ n = 300 }. \end{cases}

    Set

    \lambda_{i}^{*} = \begin{cases} \lambda_{k}(\tilde{ {\bf c}}^{*}),\quad&\text{ i = k, k+1 };\\ \lambda_{i}(\tilde{ {\bf c}}^{*}),\quad&\text{otherwise}. \end{cases}

    Subsequently, we choose \{\lambda_{i}^{*}\}_{i = 1}^{n} as the prescribed eigenvalues. It is clear that, in this way of selecting \{\lambda_{i}^{*}\}_{i = 1}^{n} , multiple eigenvalues are present. Since all of the algorithms are locally convergent, the initial guess {\bf c}_{0} is formed by chopping the components of \tilde{ {\bf c}}^{*} to six decimal places for n = 100, \ 200, \ 300 . The information in Table 1 presents the average values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| , and "it." denotes the the averaged numbers of outer iterations. The data in Table 2 presents the average total numbers of outer iterations N_{0} throughout ten test scenarios and the average total numbers of inner iterations N_{i} essential for solving the IEPs. A comparison of the averaged CPU time of all of the algorithms for the ten tests with different n is shown in Table 3.

    Table 1.  Averaged values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| for the ten tests.
    n \mathtt{it.} ICT method UCT method Algorithm Ⅰ TUCT method
    \beta=1.5 \beta=1.8
    100 0 1.8973e-5 1.8973e-5 1.8973e-5 1.8973e-5 1.8973e-5
    1 1.0006e-6 9.5329e-5 3.4389e-5 7.2881e-9 2.3548e+3
    2 3.4623e-7 5.2316e-7 1.5623e-7 4.6864e-14 2.5421e+15
    3 1.5456e-9 2.5695e-9 9.2635e-9
    4 8.6994e-12 9.9999e-11 8.2312e-12
    5 9.5684e-15 1.0001e-14 6.2351e-15
    200 0 2.6051e-5 2.6051e-5 2.6051e-5 2.6051e-5 2.6051e-5
    1 1.0003e-7 4.1858e-6 3.1889e-6 9.9999e-8 4.2151e+4
    2 9.9998e-8 3.9856e-8 4.4668e-8 6.2392e-13 8.2357e+21
    3 6.1254e-9 2.2254e-9 3.5563e-9
    4 1.5684e-12 9.5695e-11 1.5623e-12
    5 2.3564e-14 3.2356e-13 8.8563e-14
    300 0 4.0904e-5 4.0904e-5 4.0904e-5 4.0904e-5 4.0904e-5
    1 9.1364e-7 4.2864e-7 7.2789e-7 1.9604e-9 3.2546e+3
    2 3.3874e-8 2.9856e-8 3.3668e-8 2.2275e-13 6.2534e+22
    3 4.2356e-9 8.5695e-9 4.5564e-9
    4 1.0012e-12 9.9898e-11 2.5873e-11
    5 6.5684e-14 1.2325e-13 8.2344e-13

     | Show Table
    DownLoad: CSV
    Table 2.  Averaged total numbers of outer and inner iterations for the ten tests.
    n ICT method Algorithm Ⅰ UCT method
    \beta=1.5 \beta=1.8
    100 N_{0} 4.9 4.9 2 4.9
    N_{i} 32.1 32.9 29.2 32.5
    200 N_{0} 5.1 5.1 2 5.1
    N_{i} 47.2 47.8 38.9 47.3
    300 N_{0} 5.1 5.1 2 5.1
    N_{i} 71.2 71.8 60.1 71.6

     | Show Table
    DownLoad: CSV
    Table 3.  Averaged CPU time in seconds of all algorithms for the ten tests.
    n 50 100 150 200 250 300
    UCT method 0.66 2.14 7.23 18.11 40.56 98.88
    ICT method with \beta=1.5 0.64 1.91 8.12 17.68 32.43 86.37
    Algorithm Ⅰ 0.33 1.18 5.53 12.52 28.99 55.67

     | Show Table
    DownLoad: CSV

    From the data presented in Table 1, we see that Algorithm Ⅰ requires a lower number of iterations compared to the ICT and UCT methods, and the TUCT method fails to converge. It is evident from the data presented in Table 2 that Algorithm Ⅰ outperforms the inexact Cayley transform method and Ulm-like Cayley transform method. Table 3 shows that Algorithm Ⅰ demonstrates a more cost-effective CPU time compared to the other approaches.

    Example 5.2. [22] This is an inverse problem with multiple eigenvalues and n = 8 . We are given that B = I+WW^T , where

    W = \left[ \begin{array}{rrrrr} 1 & -1 & -3 & -5 & -6 \\ 1 & 1 & -2 & -5 & -17 \\ 1 & -1 & -1 & 5 & 18 \\ 1 & 1 & 1 & 2 & 0 \\ 1 & -1 & 2 & 0 & 1 \\ 1 & 1 & 3 & 0 & -1 \\ 2.5 & 0.2 & 0.3 & 0.5 & 0.6 \\ 2 & -0.2 & 0.3 & 0.5 & 0.8 \\ \end{array} \right]_{8\times 5}.

    Define

    A_0 = 0,\quad A_i: = b_{ii} {\bf e}_i {\bf e}_i^T+\sum\limits_{j = 1}^{i-1}b_{ij} ( {\bf e}_i {\bf e}_j^T+ {\bf e}_j {\bf e}_i^T),\quad i = 1,\ldots,8.
    \boldsymbol \lambda^* = (1,1,1,2.120754,9.218868,17.28137,35.70822,722.6808)^T,
    {\bf c}^* = (1,1,1,1,1,1,1,1)^T.

    We report our numerical results for different starting points: (a) {\bf c}^0 = 10^{-5}\cdot(1, 1, 1, 1, 1, 1, 1, 1)^T ; (b) {\bf c}^0 = (0, 0, 0, 0, 0, 0, 0, 0)^T .

    Table 4 presents the average values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| , and "it." denotes the the averaged numbers of outer iterations.

    Table 4.  Averaged values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| for the ten tests.
    \mathtt{it.} ICT method UCT method Algorithm Ⅰ TUCT method
    \beta=1.5 \beta=1.8
    (a) 0 2.2113e-5 2.2113e-5 2.2113e-5 2.2113e-5 2.2113e-5
    1 3.2154e-6 4.2568e-6 8.2135e-5 9.889e-14 4.2356e+3
    2 8.3549e-8 4.2648e-8 2.1350e-8 3.2589e+15
    3 3.2111e-10 1.2309e-10 8.2156e-10
    4 8.2543e-13 5.2236e-13 9.9998e-14
    (b) 0 7.2383e+2 7.2383e+2 7.2383e+2 7.2383e+2 7.2383e+2
    1 5.2469e+5 3.2699e+6 5.3216e+6 8.3269e+18 3.2156e+14
    2 5.2148e+12 7.2584e+12 5.2318e+13

     | Show Table
    DownLoad: CSV

    It can be observed from Table 4 that Algorithm Ⅰ requires a lower number of iterations compared to the ICT and UCT methods and Algorithm Ⅰ has global non-convergence and local cubic convergence.

    To further illustrate the effectiveness of Algorithm Ⅰ, we present a practical engineering application in vibrations [15,16]. We consider the vibration of a taut string with n beads. Figure 1 shows such a model for the case where n = 4 . Here, we assume that the n beads are placed along the string, where the ends of the string are clamped. The mass of the j th bead is denoted by m_j . The horizontal lengths between masses m_j and m_{j+1} (and between each bead at each end and the clamped support) are set to be a constant L . The horizontal tension is set to be a constant T . Then the equation of motion is governed by

    \begin{equation} m_jy_j''(t) = T\frac{y_{j+1}-y_j}{L}-T\frac{y_{j}-y_{j-1}}{L},\quad j = 1,\ldots,n, \end{equation} (5.1)
    Figure 1.  A string with n = 4 beads.

    where y_0 = y_{n+1} = 0 . That is, the ends of the string are fixed. The matrix form of (5.1) is given by

    \begin{equation} y''(t) = -CJy(t), \end{equation} (5.2)

    where y(t) = (y_1(t), y_2(t), \ldots, y_n(t))^T , C = {{\rm{diag}}}(c_1, c_2, \ldots, c_n) with c_j = \frac{T}{m_jL} , and J is the discrete Laplacian matrix

    J = \left[ \begin{array}{ccccc} 2 & -1 & & & \\ -1 & 2 & -1 &&\\ & \ddots & \ddots & \ddots & \\ && -1 & 2 & -1\\ &&& -1 & 2\\ \end{array} \right]\in\mathcal{S}^{n}.

    The general solution of (5.2) is given in terms of the eigenvalue problem

    CJy = \lambda y,

    where \lambda is the square of the natural frequency of the vibration system and the nonzero vector y accounts for the interplay between the masses. The inverse problem for the beaded string is to compute the masses \{m_j\}_{j = 1}^n so that the resulting system has a prescribed set of natural frequencies.

    It is easy to check that the eigenvalues of J are given by

    \lambda_j(J) = 4\left(\sin\frac{j\pi}{n+1}\right)^2,\quad j = 1,2,\ldots,n.

    Thus, J is symmetric and positive definite and CJ is similar to L^TCL , where L is the Cholesky factor of J = LL^T [31]. Then, the inverse problem is converted into the form of the IEP where A_0 = 0 and A_j = L^TE_jL with E_j = {{\rm{diag}}}(e_j) for j = 1, 2, \ldots, n . The beaded string data in Examples 5.3 and 5.4 comes from the website \mathtt{http://www.caam.rice.edu/{\sim}beads} .

    Example 5.3. This is an inverse problem for the beaded string with n = 4 beads, where

    \begin{array}{c} (m_1,m_2,m_3,m_4) = (0.030783, 0.017804, 0.017804, 0.030783) \, (kg = kilogram),\\ (n+1)L = 1.12395 \,(meter),\quad T = 191.8199\, (Newton),\\ \boldsymbol \lambda^* = (15041.90, 42344.26, 88328.78, 15041.90)^T,\\ c^* = (27720.80, 47929.08, 47929.08, 27720.80)^T. \end{array}

    We report our numerical result for the starting points: c^0 = 10^{-5}\cdot(27720.80, 47929.08, 47929.08, 27720.80)^T .

    Example 5.4. This is an inverse problem for the beaded string with n = 6 beads, where

    \begin{array}{c} (m_1,m_2,m_3,m_4,m_5,m_6) = (0.017804, 0.030783, 0.017804, 0.017804, 0.030783, 0.017804) (kg),\\ (n+1)L = 1.12395 \,(meter),\quad T = 166.0370\, (Newton),\\ \boldsymbol \lambda^* = (9113.978, 30746.32, 83621.69, 148694.4, 148694.4, 193537.0)^T,\\ c^* = (58081.57, 33592.71, 58081.57, 58081.57, 33592.71, 58081.57)^T. \end{array}

    We report our numerical results for the starting points: c^0 = 10^{-5}\cdot(58081.57, 33592.71, 58081.57, 58081.57, 33592.71, 58081.57)^T .

    The information in Table 5 presents the average values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| , and Table 6 displays the computed masses for the beaded string.

    Table 5.  Averaged values of \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| in the ten tests for Examples 5.3 and 5.4.
    Example 5.3
    \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| -value (last 3 iterations)
    ICT method (6.0235\times 10^{-8}, 3.6564\times 10^{-10}, 9.3356\times 10^{-13})
    UCT method (1.9587\times 10^{-8}, 8.3985\times 10^{-10}, 2.8872\times 10^{-14})
    Algorithm Ⅰ (5.9254\times 10^{-5}, 2.8123\times 10^{-11}, 6.5865\times 10^{-21})
    Example 5.4
    \|P_{k}^{T}A(\mathbf{c}^{k})P_{k}-\Lambda^{*}\| -value (last 3 iterations)
    ICT method (1.0562\times 10^{-7}, 5.0420\times 10^{-9}, 7.4001\times 10^{-13})
    UCT method (2.9125\times 10^{-7}, 9.0859\times 10^{-9}, 3.7523\times 10^{-13})
    Algorithm Ⅰ (1.9862\times 10^{-5}, 2.9546\times 10^{-11}, 5.5555\times 10^{-21})

     | Show Table
    DownLoad: CSV
    Table 6.  Recovered masses for Examples 5.3 and 5.4.
    Example 5.3
    m_1 m_2 m_3 m_4
    true 0.030783 0.017804 0.017804 0.030783
    recovered 0.030783 0.017804 0.017804 0.030783
    Example 5.4
    m_1 m_2 m_3 m_4 m_5 m_6
    true 0.017804 0.030783 0.017804 0.017804 0.030783 0.017804
    recovered 0.017804 0.030783 0.017804 0.017804 0.030783 0.017804

     | Show Table
    DownLoad: CSV

    From Table 5, we know that Algorithm Ⅰ requires a lower number of iterations compared to the ICT and UCT methods, and the TUCT method fails to converge. Table 6 shows that the desired masses are recovered. All of these numerical observations agree with our prediction and further validate our theoretical results.

    In this paper, we have proposed a two-step Ulm-Chebyshev-like Cayley transform method for solving the IEP (1.2) with multiple eigenvalues, which avoids solving (approximate) Jacobian equations in each outer iteration. Furthermore, the proposed algorithm is proved to have cubical convergence under the following nonsingular condition in terms of the relative generalized Jacobian evaluated at a solution {\bf c}^{*} : Each J\in \partial_{Q|S} {\bf f}({\bf c}^*) is nonsingular. Furthermore, this kind of method can be worked with larger problems for the Toeplitz inverse eigenvalue problem and is efficient when working with small and medium-sized problems for a more general perspective on this problem. Nevertheless, the effective methods for general larger inverse eigenvalue problems should be studied in the future.

    Wei Ma, Zhenhao Li and Yuxin Zhang: Algorithms, Software, Numerical examples, Writing-original draft, Writing-review & editing. All authors of this article have been contributed equally. All authors have read and approved the final version of the manuscript for publication.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This work was supported by the Henan Province College Student Innovation and Entrepreneurship Training Program Project (No: 202410481003).

    The authors declare no conflicts of interest.



    [1] F. Diele, T. Laudadio, N. Mastronardi, On some inverse eigenvalue problems with Toeplitz-related structure, SIAM J. Matrix Anal. Appl., 26 (2004), 285–294. https://doi.org/10.1137/S0895479803430680 doi: 10.1137/S0895479803430680
    [2] J. Peinado, A. M. Vidal, A new parallel approach to the Toeplitz inverse eigenproblem using Newton-like methods, In: Vector and parallel processing–VECPAR 2000, Springer, 2001. https://doi.org/10.1007/3-540-44942-6_29
    [3] W. F. Trench, Numerical solution of the inverse eigenvalue problem for real symmetric Toeplitz matrices, SIAM J. Sci. Comput., 18 (1997), 1722–1736. https://doi.org/10.1137/S1064827595280673 doi: 10.1137/S1064827595280673
    [4] N. Wagner, Inverse eigenvalue problems in structural dynamics, Proc. Appl. Math. Mech., 6 (2006), 339–340. https://doi.org/10.1002/pamm.200610151 doi: 10.1002/pamm.200610151
    [5] P. J. Brussard, P. W. M. Glaudemans, Shell model applications in nuclear spectroscopy, North-Holland, 1977.
    [6] M. S. Ravi, J. Rosenthal, X. A. Wang, On decentralized dynamic pole placement and feedback stabilization, IEEE Trans. Automat. Control, 40 (1995), 1603–1614. https://doi.org/10.1109/9.412629 doi: 10.1109/9.412629
    [7] O. Hald, On discrete and numerical Sturm-Liouville problems, New York University, 1972.
    [8] G. M. L. Gladwell, Inverse problems in vibration, Appl. Mech. Rev., 39 (1986), 1013–1018. https://doi.org/10.1115/1.3149517 doi: 10.1115/1.3149517
    [9] G. M. L. Gladwell, Inverse problems in vibration-Ⅱ, Appl. Mech. Rev., 49 (1996), S25–S34. https://doi.org/10.1115/1.3101973 doi: 10.1115/1.3101973
    [10] K. T. Joseph, Inverse eigenvalue problem in structural design, AIAA J., 30 (1992), 2890–2896. https://doi.org/10.2514/3.11634 doi: 10.2514/3.11634
    [11] R. L. Parker, K. A. Whaler, Numerical methods for establishing solutions to the inverse problem of electromagnetic induction, J. Geophys. Res., 86 (1981), 9574–9584. https://doi.org/10.1029/JB086iB10p09574 doi: 10.1029/JB086iB10p09574
    [12] N. Li, A matrix inverse eigenvalue problem and its application, Linear Algebra Appl., 266 (1997), 143–152. https://doi.org/10.1016/S0024-3795(96)00639-8 doi: 10.1016/S0024-3795(96)00639-8
    [13] M. Müller, An inverse eigenvalue problem: computing B-stable Runge-Kutta methods having real poles, BIT Numer. Math., 32 (1992), 676–688. https://doi.org/10.1007/BF01994850 doi: 10.1007/BF01994850
    [14] S. Elhay, Y. M. Ram, An affine inverse eigenvalue problem, Inverse Problems, 18 (2002), 455. https://doi.org/10.1088/0266-5611/18/2/311 doi: 10.1088/0266-5611/18/2/311
    [15] M. T. Chu, Inverse eigenvalue problems, SIAM Rev., 40 (1998), 1–39. https://doi.org/10.1137/S0036144596303984 doi: 10.1137/S0036144596303984
    [16] M. T. Chu, G. H. Golub, Structured inverse eigenvalue problems, Acta Numer., 11 (2002), 1–71. https://doi.org/10.1017/S0962492902000016 doi: 10.1017/S0962492902000016
    [17] M. T. Chu, G. H. Golub, Inverse eigenvalue problems: theory, algorithms, and applications, Oxford: Oxford University Press, 2005. https://doi.org/10.1093/acprof:oso/9780198566649.001.0001
    [18] S. F. Xu, An introduction to inverse algebric eigenvalue problems, Beijing: Peking University Press, 1998.
    [19] Z. J. Bai, Inexact Newton methods for inverse eigenvalue problems, Appl. Math. Comput., 172 (2006), 682–689. https://doi.org/10.1016/j.amc.2004.11.023 doi: 10.1016/j.amc.2004.11.023
    [20] V. N. Kublanovskaja, On one approach to the solution of the inverse eigenvalue problem, In: Automatic programming and numerical methods of analysis, Springer, 1972, 80–86. https://doi.org/10.1007/978-1-4615-8588-6_10
    [21] Y. Wang, W. Shen, An extended two-step method for inverse eigenvalue problems with multiple eigenvalues, Numer. Math. Theor. Methods Appl., 16 (2023), 968–992.
    [22] S. Friedland, J. Nocedal, M. L. Overton, The formulation and analysis of numerical methods for inverse eigenvalue problems, SIAM. J. Numer. Anal., 24 (1987), 634–667. https://doi.org/10.1137/0724043 doi: 10.1137/0724043
    [23] Z. J. Bai, R. H. Chan, B. Morini, An inexact Cayley transform method for inverse eigenvalue problems, Inverse Problems, 20 (2004), 1675. https://doi.org/10.1088/0266-5611/20/5/022 doi: 10.1088/0266-5611/20/5/022
    [24] R. H. Chan, S. F. Xu, H. M. Zhou, On the convergence rate of a quasi-Newton method for inverse eigenvalue problems, SIAM J. Numer. Anal., 36 (1999), 436–441. https://doi.org/10.1137/S0036142997327051 doi: 10.1137/S0036142997327051
    [25] R. H. Chan, H. L. Chung, S. F. Xu, The inexact Newton-like method for inverse eigenvalue problem, BIT Numer. Math., 43 (2003), 7–20. https://doi.org/10.1023/a:1023611931016 doi: 10.1023/a:1023611931016
    [26] W. P. Shen, C. Li, X. Q. Jin, A Ulm-like method for inverse eigenvalue problems, Appl. Numer. Math., 61 (2011), 356–367. https://doi.org/10.1016/j.apnum.2010.11.001 doi: 10.1016/j.apnum.2010.11.001
    [27] W. P. Shen, C. Li, An Ulm-like Cayley transform method for inverse eigenvalue problems, Taiwanese J. Math., 16 (2012), 367–386. https://doi.org/10.11650/twjm/1500406546 doi: 10.11650/twjm/1500406546
    [28] X. S. Chen, C. T. Wen, H. W. Sun, Two-step Newton-type methods for solving inverse eigenvalue problems, Numer. Linear Algebra Appl., 25 (2018), e2185. https://doi.org/10.1002/nla.2185 doi: 10.1002/nla.2185
    [29] C. T. Wen, X. S. Chen, H. W. Sun, A two-step inexact Newton-Chebyshev-like method for inverse eigenvalue problems, Linear Algebra Appl., 585 (2020), 241–262. https://doi.org/10.1016/j.laa.2019.10.004 doi: 10.1016/j.laa.2019.10.004
    [30] W. Ma, Two-step Ulm-Chebyshev-like Cayley transform method for inverse eigenvalue problems, Int. J. Comput. Math., 99 (2022), 391–406. https://doi.org/10.1080/00207160.2021.1913728 doi: 10.1080/00207160.2021.1913728
    [31] G. H. Golub, C. F. Van Loan, Matrix computations, 3 Eds., Johns Hopkins University Press, 1996.
    [32] L. Q. Qi, Convergence analysis of some algorithms for solving nonsmooth equations, Math. Oper. Res., 18 (1993), 227–244. https://doi.org/10.1287/moor.18.1.227 doi: 10.1287/moor.18.1.227
    [33] F. A. Potra, L. Q. Qi, D. F. Sun, Secant methods for semismooth equations, Numer. Math., 80 (1998), 305–324. https://doi.org/10.1007/s002110050369 doi: 10.1007/s002110050369
    [34] D. F. Sun, J. Sun, Strong semismoothness of eigenvalues of symmetric matrices and its application to inverse eigenvalue problems, SIAM J. Numer. Anal., 40 (2003), 2352–2367. https://doi.org/10.1137/s0036142901393814 doi: 10.1137/s0036142901393814
    [35] W. P. Shen, C. Li, X. Q. Jin, An inexact Cayley transform method for inverse eigenvalue problems with multiple eigenvalues, Inverse Problems, 31 (2015), 085007. https://doi.org/10.1088/0266-5611/31/8/085007 doi: 10.1088/0266-5611/31/8/085007
    [36] W. P. Shen, C. Li, X. Q. Jin, An Ulm-like Cayley transform method for inverse eigenvalue problems with multiple eigenvalues, Numer. Math. Theor. Methods Appl., 9 (2016), 664–685. https://doi.org/10.4208/nmtma.2016.y15030 doi: 10.4208/nmtma.2016.y15030
    [37] R. W. Freund, N. M. Nachtigal, QMR: a quasi-minimal residual method for non-Hermitian linear systems, Numer. Math., 60 (1991), 315–339. https://doi.org/10.1007/BF01385726 doi: 10.1007/BF01385726
  • This article has been cited by:

    1. Wei Ma, Ming Zhao, Jiaxin Li, A multi-step Ulm-Chebyshev-like method for solving nonlinear operator equations, 2024, 9, 2473-6988, 28623, 10.3934/math.20241389
    2. Rui Guo, Weiping Shen, Xinge Yang, A Two‐Step Method for Parameterized Generalized Inverse Eigenvalue Problem, 2025, 0170-4214, 10.1002/mma.11073
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1083) PDF downloads(53) Cited by(2)

Figures and Tables

Figures(1)  /  Tables(6)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog