Research article Special Issues

Solving quaternion nonsymmetric algebraic Riccati equations through zeroing neural networks

  • Many variations of the algebraic Riccati equation (ARE) have been used to study nonlinear system stability in the control domain in great detail. Taking the quaternion nonsymmetric ARE (QNARE) as a generalized version of ARE, the time-varying QNARE (TQNARE) is introduced. This brings us to the main objective of this work: finding the TQNARE solution. The zeroing neural network (ZNN) technique, which has demonstrated a high degree of effectiveness in handling time-varying problems, is used to do this. Specifically, the TQNARE can be solved using the high order ZNN (HZNN) design, which is a member of the family of ZNN models that correlate to hyperpower iterative techniques. As a result, a novel HZNN model, called HZ-QNARE, is presented for solving the TQNARE. The model functions fairly well, as demonstrated by two simulation tests. Additionally, the results demonstrated that, while both approaches function remarkably well, the HZNN architecture works better than the ZNN architecture.

    Citation: Houssem Jerbi, Izzat Al-Darraji, Saleh Albadran, Sondess Ben Aoun, Theodore E. Simos, Spyridon D. Mourtas, Vasilios N. Katsikis. Solving quaternion nonsymmetric algebraic Riccati equations through zeroing neural networks[J]. AIMS Mathematics, 2024, 9(3): 5794-5809. doi: 10.3934/math.2024281

    Related Papers:

    [1] Chunjuan Du, Tongxin Yan . SOR-based alternately linearized implicit iteration method for nonsymmetric algebraic Riccati equations. AIMS Mathematics, 2023, 8(9): 19876-19891. doi: 10.3934/math.20231013
    [2] Vladislav N. Kovalnogov, Ruslan V. Fedorov, Denis A. Demidov, Malyoshina A. Malyoshina, Theodore E. Simos, Vasilios N. Katsikis, Spyridon D. Mourtas, Romanos D. Sahas . Zeroing neural networks for computing quaternion linear matrix equation with application to color restoration of images. AIMS Mathematics, 2023, 8(6): 14321-14339. doi: 10.3934/math.2023733
    [3] Vladislav N. Kovalnogov, Ruslan V. Fedorov, Denis A. Demidov, Malyoshina A. Malyoshina, Theodore E. Simos, Spyridon D. Mourtas, Vasilios N. Katsikis . Computing quaternion matrix pseudoinverse with zeroing neural networks. AIMS Mathematics, 2023, 8(10): 22875-22895. doi: 10.3934/math.20231164
    [4] Sondess B. Aoun, Nabil Derbel, Houssem Jerbi, Theodore E. Simos, Spyridon D. Mourtas, Vasilios N. Katsikis . A quaternion Sylvester equation solver through noise-resilient zeroing neural networks with application to control the SFM chaotic system. AIMS Mathematics, 2023, 8(11): 27376-27395. doi: 10.3934/math.20231401
    [5] Vladislav N. Kovalnogov, Ruslan V. Fedorov, Igor I. Shepelev, Vyacheslav V. Sherkunov, Theodore E. Simos, Spyridon D. Mourtas, Vasilios N. Katsikis . A novel quaternion linear matrix equation solver through zeroing neural networks with applications to acoustic source tracking. AIMS Mathematics, 2023, 8(11): 25966-25989. doi: 10.3934/math.20231323
    [6] R. Sriraman, R. Samidurai, V. C. Amritha, G. Rachakit, Prasanalakshmi Balaji . System decomposition-based stability criteria for Takagi-Sugeno fuzzy uncertain stochastic delayed neural networks in quaternion field. AIMS Mathematics, 2023, 8(5): 11589-11616. doi: 10.3934/math.2023587
    [7] R. Sriraman, P. Vignesh, V. C. Amritha, G. Rachakit, Prasanalakshmi Balaji . Direct quaternion method-based stability criteria for quaternion-valued Takagi-Sugeno fuzzy BAM delayed neural networks using quaternion-valued Wirtinger-based integral inequality. AIMS Mathematics, 2023, 8(5): 10486-10512. doi: 10.3934/math.2023532
    [8] Chen Wang, Hai Zhang, Hongmei Zhang, Weiwei Zhang . Globally projective synchronization for Caputo fractional quaternion-valued neural networks with discrete and distributed delays. AIMS Mathematics, 2021, 6(12): 14000-14012. doi: 10.3934/math.2021809
    [9] Călin-Adrian Popa . Synchronization of Clifford-valued neural networks with leakage, time-varying, and infinite distributed delays on time scales. AIMS Mathematics, 2024, 9(7): 18796-18823. doi: 10.3934/math.2024915
    [10] Ailing Li, Mengting Lv, Yifang Yan . Asymptotic stability for quaternion-valued BAM neural networks via a contradictory method and two Lyapunov functionals. AIMS Mathematics, 2022, 7(5): 8206-8223. doi: 10.3934/math.2022457
  • Many variations of the algebraic Riccati equation (ARE) have been used to study nonlinear system stability in the control domain in great detail. Taking the quaternion nonsymmetric ARE (QNARE) as a generalized version of ARE, the time-varying QNARE (TQNARE) is introduced. This brings us to the main objective of this work: finding the TQNARE solution. The zeroing neural network (ZNN) technique, which has demonstrated a high degree of effectiveness in handling time-varying problems, is used to do this. Specifically, the TQNARE can be solved using the high order ZNN (HZNN) design, which is a member of the family of ZNN models that correlate to hyperpower iterative techniques. As a result, a novel HZNN model, called HZ-QNARE, is presented for solving the TQNARE. The model functions fairly well, as demonstrated by two simulation tests. Additionally, the results demonstrated that, while both approaches function remarkably well, the HZNN architecture works better than the ZNN architecture.



    Since Kalman demonstrated how common algebraic Riccati equations (ARE) are in the theory of filtering and optimal control [1], AREs have drawn a great deal of interest in applied mathematics and a variety of engineering tasks. Among these topics are wheeled inverted pendulums [2], multi-agent linear systems [3], and managing wind generators [4]. The linear-quadratic regulator [5], Kalman filtering [6], linear-quadratic-Gaussian and H2/H control [7,8], coprime and spectral factorizations [9,10] all depend on ARE. Furthermore, several issues like losing controllability and numerical precision in direct and iterative methods are resolved through closed-form solutions of ARE [11]. In this paper, taking the quaternion nonsymmetric ARE (QNARE) as a generalized version of ARE, the time-varying QNARE (TQNARE) is introduced and solved.

    Recently, there has been a greater focus in the study of dynamic topics using time-varying quaternion matrices (TQM). These include the constrained TQM least-squares problem [12], the inversion of TQM [13], and the pseudoinversion of TQM [14]. Moreover, practical applications of TQMs contain synchronization of chaotic systems [15], kinematic control of redundant manipulators [16], mobile manipulator control [17], and tracking acoustic sources [18]. All of these research topics have one thing in common: they use the zeroing neural network (ZNN) approach to determine the appropriate response. Remarkably, Zhang {et al.} presented the ZNN approach in [19] for real-time handling of time-varying problems. Particularly, ZNNs are recurrent neural networks with remarkable parallel processing performance. Among its applications are dynamical systems for computing the time-varying Moore-Penrose inverse [20,21]. Linear/nonlinear equations systems [22,23,24,25,26], quadratic/linear programming [27,28,29], and generalized inversion [30,31] are among the issues that they are presently employed for.

    In 1843, Hamilton [32] made the first proposal regarding quaternions. They are very important in many fields, including mathematical physics [33], computer modeling [34], quantum mechanics [35], navigation [36], electromagnetism [37], and robotics [38,39], because they are useful in computations associated with 3-dimensional rotations in applied and theoretical mathematics [40]. Keep in mind that quaternions are an extension of complex numbers into a non-commutative number system. Let

    H={β1+β2ı+β3ȷ+β4k | ı2=ȷ2=k2=ıȷk=1,β1,β2,β3,β4R}

    be the quaternions set and Hm×n be the set of m×n matrices on H [41]. In this work, a recurrent neural network, named high-order ZNN (HZNN), is used to solve the next TQNARE:

    ˜D(t)˜X(t)+˜X(t)˜A(t)˜X(t)˜B(t)˜X(t)+˜C(t)=0m×n, (1.1)

    where ˜D(t)Hm×m,˜A(t)Hn×n,˜B(t)Hn×m, ˜C(t)Hm×n are the block coefficients, ˜X(t)Hm×n is the unknown matrix to be obtained, and 0m×n represents a zero m×n matrix. It is significant to mention that (1.1) is in its standard form without assumption on the symmetry of the coefficient matrices, as indicated by the incorrect use of the word "nonsymmetric". The all ones and all zeros u×1 matrices, respectively, will be denoted by 1u and 0u, while the identity u×u matrix will be denoted by Iu throughout this work. Additionally, and will denote the Kronecker and Hadamard products, respectively, and vec() will denote the vectorization procedure. Finally, the matrix transposition, inversion, and pseudoinversion, respectively, will be denoted by the superscripts ()T, ()1, and (), while F will denote the matrix Frobenius norm.

    From one perspective, a ZNN model is usually built in two main processes. First, the error matrix equation (EME), E(t), needs to be declared. Second, the ZNN architecture that follows should be applied:

    ˙E(t)=λE(t). (1.2)

    Notice that t[0,tf)[0,+) denotes the time, and (˙) is the derivative operator in terms of time. In addition, the model's convergence speed can be adjusted by changing the parameter λR+. For instance, any ZNN model converges even faster at larger values of λ [42]. The ZNN design is based on setting E(t) to 0, where t. This is achieved by utilizing the real-time learning rule that emerges from the building of the EME in (1.2). Thus, the EME can be thought of as a tracking tool for ZNN model learning.

    On the other hand, the family of hyperpower iterations has been the subject of extensive research and development in the last few years [43]. Nevertheless, many HZNN models were introduced and investigated in [23] because iterative approaches can be applied to discrete-time models and these methods usually require initial conditions that are approximated and sometimes can not be easily provided. An HZNN model is typically constructed through a pair of primary steps. First, we need to define the function of the high-order error matrix equation (HEME), EpH(t), as follows:

    EpH(t)=p1i=1Ei(t), (1.3)

    where Ei(t)Rn×n. Second, the HZNN architecture:

    ˙E(t)λEpH(t) (1.4)

    should be applied for finding the online solution to a time-varying issue. In addition, the model's convergence rate can be adjusted by changing the λR+ parameter and the order p2. For example, any HZNN model converges even faster at larger values of p [23]. It is significant to note that every HZNN model reduces to a ZNN model in the case where p=2 since the ZNN dynamical system of (1.2) and the HZNN dynamical system of (1.4) will match.

    In this work, a new HZNN model, termed HZ-QNARE, is introduced for solving the TQNARE. Two simulation trials demonstrate the excellent performance of the model. It is important to mention that the HZ-QNARE model has the advantage of being able to solve any equation that arises from TQNARE. Additionally, the results demonstrate that, while both approaches function remarkably well, the HZNN architecture works better than the ZNN architecture. Notice that, by theoretical examination of the HZ-QNARE model, this work adds to the corpus of literature. Finally, the list below displays the contributions made to the paper:

    ● A new HZNN model, named HZ-QNARE, for solving the TQNARE is introduced.

    ● The HZ-QNARE model is validated by a theoretical examination.

    ● Simulation trials are carried out to supplement the theoretical study.

    The paper is formatted as follows: Section 2 presents the preliminary results of the quaternion and the TQNARE reformulation. The HZ-QNARE model is presented in Section 3 and its theoretical examination is provided in Section 4. Observe that Section 3 contains information on the computational complexity of the HZ-QNARE model. Section 5 displays simulation trials. Finally, some remarks and closing comments are included in Section 6.

    In this section, the TQNARE is reformulated, and the fundamentals of TQM are described. Observe that the TQNARE (1.1) is being reformulated to reduce the computational complexity of the HZNN method.

    Employing the coefficients Du(t)Rm×m for u=1,,4, let

    ˜D(t)=D1(t)+D2(t)ı+D3(t)ȷ+D4(t)kHm×m

    to be a TQM. Similarly, employing the coefficients Au(t)Rn×n,Bu(t)Rn×m, and Cu(t),Xu(t)Rm×n for u=1,,4, let ˜A(t)Hn×n,˜B(t)Hn×m, and ˜C(t),˜X(t)Hm×n. The result of multiplying ˜B(t) by ˜X(t) is:

    ˜B(t)˜X(t)=~BX(t)=BX1(t)+BX2(t)ı+BX3(t)ȷ+BX4(t)kHn×n, (2.1)

    where

    BX1(t)=B3(t)X3(t)+B1(t)X1(t)B2(t)X2(t)B4(t)X4(t),BX2(t)=B1(t)X2(t)+B3(t)X4(t)+B2(t)X1(t)B4(t)X3(t),BX3(t)=B1(t)X3(t)+B3(t)X1(t)B2(t)X4(t)+B4(t)X2(t),BX4(t)=B4(t)X1(t)+B1(t)X4(t)B3(t)X2(t)+B2(t)X3(t). (2.2)

    Notice that BXu(t)Rn×n for u=1,,4. The products ~DX(t)=˜D(t)˜X(t), ~XA(t)=˜X(t)˜A(t) and ~XBX(t)=˜X(t)~BX(t) can be produced in a manner similar to that of (2.1).

    Using the previously provided information, (1.1) can be rewritten as:

    ~DX(t)+~XA(t)~XBX(t)+˜C(t)=0m×n, (2.3)

    where

    {DX1(t)+XA1(t)XBX1(t)+C1(t)=0m×n,DX2(t)+XA2(t)XBX2(t)+C2(t)=0m×n,DX3(t)+XA3(t)XBX3(t)+C3(t)=0m×n,DX4(t)+XA4(t)XBX4(t)+C4(t)=0m×n. (2.4)

    Then, setting

    Q(t)=[X1(t)X2(t)X3(t)X4(t)X2(t)X1(t)X4(t)X3(t)X3(t)X4(t)X1(t)X2(t)X4(t)X3(t)X2(t)X1(t)],A(t)=[A1(t)A2(t)A3(t)A4(t)],C(t)=[C1(t)C2(t)C3(t)C4(t)],M(t)=[B1(t)B2(t)B3(t)B4(t)B2(t)B1(t)B4(t)B3(t)B3(t)B4(t)B1(t)B2(t)B4(t)B3(t)B2(t)B1(t)],B(t)=[B1(t)B2(t)B3(t)B4(t)],X(t)=[X1(t)X2(t)X3(t)X4(t)],K(t)=[D1(t)D2(t)D3(t)D4(t)D2(t)D1(t)D4(t)D3(t)D3(t)D4(t)D1(t)D2(t)D4(t)D3(t)D2(t)D1(t)],D(t)=[D1(t)D2(t)D3(t)D4(t)], (2.5)

    where Q(t)R4m×4n,M(t)R4n×4m,K(t)R4m×4m,D(t)R4m×m,A(t)R4n×n,B(t)R4n×m, and X(t),C(t)R4m×n, (2.3) might be reshuffled as below:

    K(t)X(t)+Q(t)A(t)Q(t)M(t)X(t)+C(t)=04m×n, (2.6)

    where the coefficients Xu(t),u=1,,4, are included both in X(t) and Q(t). Recall that the anticipated solution to the TQNARE (1.1) is ˜X(t). Furthermore, take note of the fact that (2.6) gives four TQMs, whereas (1.1) yields only one TQM.

    To address the TQNARE, we shall construct a HZNN model in this section named HZ-QNARE. Let ˜D(t)Hm×m,˜A(t)Hn×n,˜B(t)Hn×m, and ˜C(t),˜X(t)Hm×n be differentiable TQMs. According to the analysis of Section 2, (2.6) is a reformulation of (1.1). Following (2.5), we formulate the matrices Q(t)R4m×4n,M(t)R4n×4m,K(t)R4m×4m,D(t)R4m×m,A(t)R4n×n,B(t)R4n×m, and C(t)R4m×n to construct the next EME:

    E(t)=Q(t)A(t)+K(t)X(t)Q(t)M(t)X(t)+C(t), (3.1)

    where Q(t)R4m×4n and X(t)R4m×n are unknown matrices. Because the HZNN design requires powers of E(t), it is applicable only to square input matrices M(t),A(t), and C(t). In order to satisfy this requirement, we right multiply (3.1) with ˉI=[In,0n×(4mn)]Rn×4m. Therefore, without loss of generality, we convert (3.1) as follows:

    EZ(t)=(Q(t)A(t)+K(t)X(t)Q(t)M(t)X(t)+C(t))ˉI=Q(t)A(t)ˉI+K(t)X(t)ˉIQ(t)M(t)X(t)ˉI+C(t)ˉI. (3.2)

    The following is (3.2)'s first derivative:

    ˙EZ(t)=Q(t)˙A(t)ˉI+˙Q(t)A(t)ˉI+˙K(t)X(t)ˉI+K(t)˙X(t)ˉI˙Q(t)M(t)X(t)ˉIQ(t)˙M(t)X(t)ˉIQ(t)M(t)˙X(t)ˉI+˙C(t)ˉI. (3.3)

    Further, the following HEME is defined:

    EpH(t)=p1i=1EiZ(t), (3.4)

    while its derivative is ˙EpH(t)˙EZ(t) according to the HZNN design.

    Next, we treat the HZNN dynamical system with regard to ˙X(t) and ˙Q(t). This is accomplished by superseding the aforementioned EpH(t) and ˙EpH(t) in (1.4). The outcome is as follows:

    K(t)˙X(t)ˉI+˙Q(t)A(t)ˉI˙Q(t)M(t)X(t)ˉIQ(t)M(t)˙X(t)ˉI=λp1i=1EiZ(t)˙K(t)X(t)ˉIQ(t)˙A(t)ˉI+Q(t)˙M(t)X(t)ˉI˙C(t)ˉI. (3.5)

    Afterwards, we simplify the dynamics of (3.5) by using the Kronecker product and vectorization:

    (ˉITK(t)ˉIT(Q(t)M(t)))vec(˙X(t))+((A(t)ˉI)TI4m(M(t)X(t)ˉI)TI4m)vec(˙Q(t))=vec(λp1i=1EiZ(t)˙K(t)X(t)ˉIQ(t)˙A(t)ˉI+Q(t)˙M(t)X(t)ˉI˙C(t)ˉI). (3.6)

    It is important to note that the components of vec(˙Q(t)) and vec(˙X(t)) are the same, but they are located differently. Stated differently, (3.6) can be further reduced by stating vec(˙Q(t)) in relation to vec(˙X(t)). Because of this, vec(˙Q(t)) can be substituted in (3.6) to obtain the following equation:

    vec(˙Q(t))=Rvec(˙X(t)). (3.7)

    Note that the algorithmic process outlined in Algorithm 1 can be used to compute the operational matrix RR16mn×4mn. Take into account that the pseudocode in Algorithm 1 adhere to standard MATLAB practices [44].

    Algorithm 1 Algorithmic process for calculating the operational matrix.
    Input: The matrix X(t) dimensions, m×n.
    1: procedure OM_R(m,n)
    2:  Put R=zeros(nm16,nm4), c=(1:nm4), Z=reshape(c,m4,n)
    3:  Put V1=Z(1:m,:), V2=Z(1+m:m2,:), V3=Z(1+m2:m3,:), V4=Z(1+m3:end,:)
    4:  Put V=[V1,V2,V3,V4;V2,V1,V4,V3;V3,V4,V1,V2;V4,V3,V2,V1]
    5:  Set F=reshape(V,[],1)
    6:  for u=1:nm16 do
    7:    R(u,abs(F(u)))=sign(F(u))
    8:  end for
    9: end procedure
    Output: R

    The (3.6) may be simplified even more by utilizing (3.7):

    (ˉITK(t)ˉIT(Q(t)M(t))+((A(t)ˉI)TI4m(M(t)X(t)ˉI)TI4m)R)vec(˙X(t))=vec(λp1i=1EiZ(t)˙K(t)X(t)ˉIQ(t)˙A(t)ˉI+Q(t)˙M(t)X(t)ˉI˙C(t)ˉI). (3.8)

    Additionally, upon the establishment of:

    U(t)=ˉITK(t)ˉIT(Q(t)M(t))+((A(t)ˉI)TI4m(M(t)X(t)ˉI)TI4m)RR16m2×4mn,g(t)=vec(λp1i=1EiZ(t)˙K(t)X(t)ˉIQ(t)˙A(t)ˉI+Q(t)˙M(t)X(t)ˉI˙C(t)ˉI)R16m2,˙x(t)=vec(˙X(t))R4mn,x(t)=vec(X(t))R4mn, (3.9)

    we arrive at the HZNN model below:

    ˙x(t)=U(t)g(t). (3.10)

    The recommended HZNN model to be used in solving the TQNARE of (1.1) is Eq (3.10)'s dynamic model, also known as HZ-QNARE. Note that the algorithmic process of the HZ-QNARE model outlined in Algorithm 2 can be used for solving the TQNARE.

    Algorithm 2 Algorithmic process of the HZ-QNARE model for solving the TQNARE.
    Input: The coefficient matrices ˜D(t),˜A(t),˜B(t), and ˜C(t) of the TQNARE (1.1).
    1: procedure HZ_QNARE(˜D(t),˜A(t),˜B(t),˜C(t))
    2:  Create Q(t),A(t),C(t),M(t),B(t),X(t),K(t), and D(t) based on (2.5)
    3:  Create EZ(t) according to (3.2)
    4:  Create EpH(t) according to (3.4)
    5:  Using Alg. 1, create U(t) and g(t) based on (3.9)
    6:  Set ˙x(t)=U(t)g(t) and use the ode15s solver of MATLAB to find x(t)
    7: end procedure
    Output: The TQNARE's solution x(t).

    Creating and solving (3.10) is computationally demanding in addition to the HZ-QNARE itself. In this regard, the complexity of formulating (3.10) is O((4nm)2) operations since, at each iteration, we do (4nm)2 multiplications and 4nm additions/subtractions. Besides, the ode15s solver of MATLAB is used to solve the linear system of equations step-by-step. A 4nm×4nm matrix is needed to address (3.10), adding O((4nm)3) to the complexity. As a result, O((4nm)3) is the HZ-QNARE's computational complexity.

    This section presents the stability analysis and convergence of the HZ-QNARE (3.10) model.

    Theorem 4.1. Let M(t)R4n×4m,K(t)R4m×4m,A(t)R4n×n, and C(t)R4m×n be differentiable. According to Lyapunov, the dynamics of (3.5) in accordance with the HZNN method (1.4) lead to the theoretical solution (TSOL) XTH(t), which is stable.

    Proof. The supersession XB(t):=X(t)+XTH(t) yields X(t)=XTH(t)XB(t), while the TSOL is XTH(t). The components of X(t) are reshuffled in Q(t), according to (2.5). Consequently, since QTH(t) and QB(t) represent a reshuffling of the components of XTH(t) and XB(t), Q(t)=QTH(t)QB(t). Also, the 1st derivatives of Q(t) and X(t) are ˙Q(t)=˙QTH(t)˙QB(t) and ˙X(t)=˙XTH(t)˙XB(t), respectively.

    Let

    K(t)XTH(t)ˉI+QTH(t)A(t)ˉIQTH(t)M(t)XTH(t)ˉI+C(t)ˉI=04m×4m, (4.1)

    and its 1st derivative

    K(t)˙XTH(t)ˉI+˙K(t)XTH(t)ˉI+˙QTH(t)A(t)ˉI+QTH(t)˙A(t)ˉI˙QTH(t)M(t)XTH(t)ˉIQTH(t)˙M(t)XTH(t)ˉIQTH(t)M(t)˙XTH(t)ˉI+˙C(t)ˉI=04m×4m. (4.2)

    After substituting (3.4) for X(t)=XTH(t)XB(t) and Q(t)=QTH(t)QB(t), the following holds:

    ETH(t)=K(t)XTH(t)ˉIK(t)XB(t)ˉI+QTH(t)A(t)ˉIQB(t)A(t)ˉIQTH(t)M(t)XTH(t)ˉI+QB(t)M(t)XTH(t)ˉI+QTH(t)M(t)XB(t)ˉIQB(t)M(t)XB(t)ˉI+C(t)ˉI, (4.3)

    and

    EpTH(t)=p1i=1EiTH(t), (4.4)

    with the following results from (1.4):

    ˙ETH(t)=K(t)˙XTH(t)ˉI+˙K(t)XTH(t)ˉIK(t)˙XB(t)ˉI˙K(t)XB(t)ˉI+˙QTH(t)A(t)ˉI+QTH(t)˙A(t)ˉI˙QB(t)A(t)ˉIQB(t)˙A(t)ˉI+˙C(t)ˉI˙QTH(t)M(t)XTH(t)ˉIQTH(t)˙M(t)XTH(t)ˉIQTH(t)M(t)˙XTH(t)ˉI+˙QB(t)M(t)XTH(t)ˉI+QB(t)˙M(t)XTH(t)ˉI+QB(t)M(t)˙XTH(t)ˉI+˙QTH(t)M(t)XB(t)ˉI+QTH(t)˙M(t)XB(t)ˉI+QTH(t)M(t)˙XB(t)ˉI˙QB(t)M(t)XB(t)ˉIQB(t)˙M(t)XB(t)ˉIQB(t)M(t)˙XB(t)ˉIλEpTH(t). (4.5)

    Afterwards, we pick the Lyapunov function indicated below to verify convergence:

    R(t)=12EpTH(t)2F=12tr(EpTH(t)(EpTH(t))T). (4.6)

    The following can then be verified:

    ˙R(t)=2tr((EpTH(t))T˙EpTH(t))2=tr((EpTH(t))T˙EpTH(t))=λtr((EpTH(t))TEpTH(t)). (4.7)

    Consequently, the following holds true:

    ˙R(t){<0,EpTH(t)0=0,EpTH(t)=0,˙R(t){<0,K(t)XTH(t)ˉIK(t)XB(t)ˉI+QTH(t)A(t)ˉIQB(t)A(t)ˉIQTH(t)M(t)XTH(t)ˉI+QB(t)M(t)XTH(t)ˉI+QTH(t)M(t)XB(t)ˉIQB(t)M(t)XB(t)ˉI+C(t)ˉI0=0,K(t)XTH(t)ˉIK(t)XB(t)ˉI+QTH(t)A(t)ˉIQB(t)A(t)ˉIQTH(t)M(t)XTH(t)ˉI+QB(t)M(t)XTH(t)ˉI+QTH(t)M(t)XB(t)ˉIQB(t)M(t)XB(t)ˉI+C(t)ˉI=0,˙R(t){<0,XB(t)0  QB(t)0=0,XB(t)=0  QB(t)=0,˙R(t){<0,XB(t)0=0,XB(t)=0. (4.8)

    Seeing that Q(t) is a components' regrouping of X(t), QB(t)0 in the case XB(t)0 and QB(t)=0 in the case XB(t)=0. Furthermore, because EpH(0)=0 and XB(t) are the equilibrium points of (4.5), the following holds:

     XB(t)0,˙R(t)0. (4.9)

    It is found that the state of equilibrium XB(t)=X(t)+XTH(t)=0 is stable according to the theory of Lyapunov. Afterwards, X(t)XTH(t) as t. Notice that, because Q(t) is a components' regrouping of X(t), Q(t)QTH(t) as t.

    Theorem 4.2. Let ˜D(t)Hm×m,˜A(t)Hn×n,˜B(t)Hn×m, and ˜C(t)Hm×n be differentiable. For each t[0,tf)[0,+), the HZ-QNARE model (3.10) converges exponentially to TSOL xTH(t) under any selection of initial value x(0).

    Proof. First, the TQNARE of (1.1) is converted into (2.6) using the analysis given in Section 2. In particular, we utilize (2.5) to construct the matrices M(t)R4n×4m,K(t)R4m×4m,A(t)R4n×n, and C(t)R4m×n employing the TQMs ˜D(t),˜A(t),˜B(t), and ˜C(t). Consequently, we substitute (1.1) for (2.6). Second, in order to solve (2.6), the HEME of (3.4) is declared. Model (3.5) is applied in compliance with the HZNN technique (1.4) for zeroing (3.4). In line with Theorem 4.1, Q(t)QTH(t) and X(t)XTH(t) when t for any selection of initial value. Remember that Q(t) is created by rearranging the X(t) parts. Model (3.5) converges to the TSOL equation (1.1) since (2.6) is a reorganization of the TQNARE of (1.1). Third, (3.5) is simplified to construct the HZ-QNARE model of (3.10) using the operation matrix of (3.7), Kronecker product, and vectorization. The HZ-QNARE is an alternative formulation of (3.5); for any beginning value x(0), it converges to TSOL xTH(t) when t. At that point, the proof is finished.

    This section exhibits two simulation trials (STs). A few significant arguments should be mentioned. For simplicity, we allocated both η(t)=cos(t) and ζ(t)=sin(t) in the STs. Additionally, all calculations are performed employing the ode15s solver of MATLAB, with a time interval set to [0,10]. Employing the ode15s and its regular double precision arithmetic, you will notice that most errors have minimum values that are close to 105 in this section's figures. Last, the associated TQNARE error, i.e., ||˜D(t)˜X(t)+˜X(t)˜A(t)˜X(t)˜B(t)˜X(t)+˜C(t)||F, is displayed as ||TQNARE||F in the figures of this section.

    Example 5.1. In this ST, the ˜D(t) coefficients are set up as below:

    D1(t)=I4(1+η(t)),D2(t)=[0001000000001000],D3(t)=[0100100000000000],D4(t)=[0000000100000100],

    the ˜A(t) coefficients are set up as below:

    A1(t)=[6ζ(t)+145η(t)+26],A2(t)=[02ζ(t)+24ζ(t)20],A3(t)=[055ζ(t)+10],A4(t)=[02ζ(t)+320],

    the ˜B(t) coefficients are set up as below:

    B1(t)=[η(t)ζ(t)+301ζ(t)+3ζ(t)+200],B2(t)=[06106011],B3(t)=[05+η(t)005η(t)000],B4(t)=[0η(t)41η(t)000],

    and ˜C(t) coefficients are set up as below:

    C1(t)=[ζ(t)+5η(t)η(t)η(t)+20110],C2(t)=[05+ζ(t)5ζ(t)05130],C3(t)=[06600111],C4(t)=[0ζ(t)ζ(t)00310].

    Hence, ˜D(t)H4×4,˜A(t)H2×2,˜B(t)H2×4, and ˜C(t)H4×2. Furthermore, the HZ-QNARE model's λ parameter is used, with values of 10 and 100, and the beginning conditions (BCs) are established as below:

    ● BC1: x(0)=032,

    ● BC2: x(0)=100132.

    You will see that we employ two BCs to confirm Theorem 4.2's findings. The results of the HZ-QNARE model are shown in Figure 1.

    Figure 1.  (1.1) error, solutions trajectories, and HEMEs of HZ-QNARE in ST 5.1.

    Example 5.2. Assume the next 10×10 real matrix:

    M=[3+η(t)1+ζ(t)/21+ζ(t)/101+ζ(t)/23+η(t)1+ζ(t)/91+ζ(t)/101+ζ(t)/93+η(t)].

    In this ST, the coefficients of ˜A(t) are set up as below:

    A1(t)=M+I10ζ(t),A2(t)=2M+I10η(t),A3(t)=3M+I10ζ(t)2,A4(t)=4M+I10η(t)2,

    and matrices ˜D(t),˜B(t), and ˜C(t) are set up as below:

    ˜D(t)=˜B(t)=010×10,˜C(t)=I10.

    Therefore, ˜D(t),˜A(t),˜B(t),˜C(t)H10×10. In order to confirm the convergence characteristics of the HZNN method, the beginning state of the HZ-QNARE model is set to x(0)=1001400. Also, the λ parameter of the HZ-QNARE model is employed with values 10 and 100, and the order p is employed with prices 2 and 4. The results of the HZ-QNARE model are shown in Figure 2.

    Figure 2.  (1.1) error, solutions trajectories, and HEMEs of HZ-QNARE in ST 5.2.

    Throughout Examples 5.1 and 5.2, the effectiveness of the HZ-QNARE (3.10) model in solving the TQNARE of (1.1) is tested. Each ST has a unique TQNARE due to the suitable matrices ˜D(t),˜B(t),˜C(t), and ˜A(t).

    It is noticeable that ˜C(t)H4×2 in ST 5.1. In other words, m=4 and n=2 in (1.1)'s TQNARE. Under λ=10, λ=100, p=2, and p=4, we have gathered the following findings regarding the HZ-QNARE model: Figure 1a exhibits the HZ-QNARE model's HEMEs. The two BC's instances in this exhibition begin at t=0 from a high error value. Notably, the beginning error value of BC2 is substantially higher than BC1's. Nevertheless, when λ=10, the HEMEs attain a negligible error value within the [104,102] interval in the BC1 paradigm at t=1 and in the BC2 paradigm at t=2. Additionally, when λ=100, the HEMEs attain a negligible error value within the [106,103] interval in the BC1 paradigm at t=0.1 and in the BC2 paradigm at t=0.2. Put another way, the HZ-QNARE model, which converges to a low value for two different BCs, validates Theorem 4.2. Figure 1b exhibits the HZ-QNARE model's (1.1) error. The outcomes shown there support those found in Figure 1a, showing that there is no difference in the solutions to (2.6) and (1.1). Trajectories representing the model's solutions are displayed in Figure 1c1f. The three imaginary and real components of the solutions are shown in these figures. It is shown that the models offer distinct solutions for a range of BCs and that their convergence to TSOL is in line with the convergence trend of the associated HEMEs.

    Further, it is noticeable that ˜D(t),˜A(t),˜B(t),˜C(t)H10×10 in ST 5.2. That is, m=n=10 in (1.1)'s TQNARE. It is crucial to notice that by setting ˜B(t)=˜D(t)=010×10 and ˜C(t)=I10, the TQNARE is reduced into a TQM inversion problem. As a result, the TSOL of the TQNARE is the inverse of ˜A(t), i.e., ˜A1(t). Under BC1 and BC2 for λ=10, we have gathered the next findings regarding the HZ-QNARE model. Figure 2a exhibits the HZ-QNARE model's HEMEs. The four instances in this exhibition begin at t=0 from a high error value and end up with a negligible value. Notably, in the paradigm of λ=10, they end up in the interval [105,103] with a negligible error value when p=2 at t=2, and at t=1.5 when p=4. In the paradigm of λ=100, they end up in the interval [107,104] with a negligible error value when p=2 at t=0.2, and at t=0.15 when p=4. Put another way, the HEME of the HZ-QNARE model, which is affected by the values of λ and p, verifies the convergence qualities of the HZNN method. Figure 2b exhibits the HZ-QNARE model's (1.1) error. The outcomes shown there support those found in Figure 2a, showing that there is no difference in the solutions to (2.6) and (1.1). Trajectories representing the model's solutions are displayed in Figure 2c2f. The three imaginary and real components of the solutions are shown in these figures. It is shown that the models offer distinct solutions for a range of BCs, and that their convergence to TSOL is in line with the convergence trend of the associated HEMEs.

    After all is said and done, the HZ-QNARE model handles two different TQNAREs fairly well. Notably, the above explanation validates the convergence properties of the HZNN approach and the conclusions of Theorem 4.2. It should be noted that the ZNN model is the HZNN under p=2 since the HZNN dynamical system of (1.4) and the ZNN dynamical system of (1.4) match for p=2. Therefore, the HZNN architecture performs better than the ZNN architecture because, at p=2, the HZ-QNARE model turns into a ZNN model, while at p>2, the HZ-QNARE model achieves better convergence. Additionally, the HZ-QNARE computational complexity is O((4nm)3), which is comparable to other ZNN models that tackle different TQM equations (see [14,15,17,18,45]).

    In this work, a new HZNN model, named HZ-QNARE, is introduced for solving the TQNARE. In order to validate the accuracy and effectiveness of the HZ-QNARE model, theoretical research and simulation experiments were carried out. These experiments' outcomes demonstrate that the HZNN design outperforms the ZNN architecture in terms of effectiveness.

    It must be noted that the recommended HZNN model has the disadvantage of being noise-intolerant because noise of any kind has a major impact on the accuracy of the suggested HZNN approach. Thus, the primary goal of future research may be to convert this model to a noise-tolerant HZNN architecture. Other than that, the suggested HZNN model may be used for a variety of additional technology-related issues in the future, such as network and power systems [46,47] and secure communications [48].

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    This research was funded by Research Deanship of Hail University-KSA Project Number (RG-23 072).

    The authors declare no conflicts of interest.



    [1] R. E. Kalman, Contributions to the theory of optimal control, Bol. Soc. Mat. Mexicana, 5 (1960), 102–119.
    [2] G. Rigatos, K. Busawon, J. Pomares, M. Abbaszadeh, Nonlinear optimal control for the wheeled inverted pendulum system, Robotica, 38 (2020), 29–47. https://doi.org/10.1017/S0263574719000456 doi: 10.1017/S0263574719000456
    [3] X. Dong, G. Hu, Time-varying formation tracking for linear multiagent systems with multiple leaders, IEEE Trans. Autom. Control, 62 (2017), 3658–3664. https://doi.org/10.1109/TAC.2017.2673411 doi: 10.1109/TAC.2017.2673411
    [4] B. Qin, H. Sun, J. Ma, W. Li, T. Ding, Z. Wang, et al., Robust H control of doubly fed wind generator via state-dependent Riccati equation technique, IEEE Trans. Power Syst., 34 (2019), 2390–2400. https://doi.org/10.1109/TPWRS.2018.2881687 doi: 10.1109/TPWRS.2018.2881687
    [5] L. T. Aguilar, Y. Orlov, L. Acho, Nonlinear H-control of nonsmooth time-varying systems with application to friction mechanical manipulators, Automatica, 39 (2003), 1531–1542. https://doi.org/10.1016/S0005-1098(03)00148-1 doi: 10.1016/S0005-1098(03)00148-1
    [6] A. J. Laub, A schur method for solving algebraic Riccati equations, IEEE Trans. Autom. Control, 24 (1979), 913–921. https://doi.org/10.1109/TAC.1979.1102178 doi: 10.1109/TAC.1979.1102178
    [7] A. Ferrante, L. Ntogramatzidis, The generalized continuous algebraic Riccati equation and impulse-free continuous-time LQ optimal control, Automatica, 50 (2014), 1176–1180. https://doi.org/10.1016/j.automatica.2014.02.014 doi: 10.1016/j.automatica.2014.02.014
    [8] T. Ohtsuka, A recursive elimination method for finite-horizon optimal control problems of discrete-time rational systems, IEEE Trans. Autom. Control, 59 (2014), 3081–3086. https://doi.org/10.1109/TAC.2014.2321231 doi: 10.1109/TAC.2014.2321231
    [9] Y. Oshman, I. Bar-Itzhack, Eigenfactor solution of the matrix Riccati equation–a continuous square root algorithm, IEEE Trans. Autom. Control, 30 (1985), 971–978. https://doi.org/10.1109/TAC.1985.1103823 doi: 10.1109/TAC.1985.1103823
    [10] P. Van Dooren, A generalized eigenvalue approach for solving Riccati equations, SIAM J. Sci. Stat. Comput., 2 (1981), 121–135. https://doi.org/10.1137/0902010 doi: 10.1137/0902010
    [11] R. C. Li, W. Kahan, A family of anadromic numerical methods for matrix Riccati differential equations, Math. Comp., 81 (2012), 233–265. https://doi.org/10.1090/S0025-5718-2011-02498-1 doi: 10.1090/S0025-5718-2011-02498-1
    [12] L. Xiao, P. Cao, W. Song, L. Luo, W. Tang, A fixed-time noise-tolerance ZNN model for time-variant inequality-constrained quaternion matrix least-squares problem, IEEE Trans. Neur. Net. Lear. Syst., 2023, 1–10. https://doi.org/10.1109/TNNLS.2023.3242313
    [13] L. Xiao, S. Liu, X. Wang, Y. He, L. Jia, Y. Xu, Zeroing neural networks for dynamic quaternion-valued matrix inversion, IEEE Trans. Ind. Inf., 18 (2022), 1562–1571. https://doi.org/10.1109/TII.2021.3090063
    [14] V. N. Kovalnogov, R. V. Fedorov, D. A. Demidov, M. A. Malyoshina, T. E. Simos, S. D. Mourtas, et al., Computing quaternion matrix pseudoinverse with zeroing neural networks, AIMS Math., 8 (2023), 22875–22895. https://doi.org/10.3934/math.20231164 doi: 10.3934/math.20231164
    [15] S. B. Aoun, N. Derbel, H. Jerbi, T. E. Simos, S. D. Mourtas, V. N. Katsikis, A quaternion Sylvester equation solver through noise-resilient zeroing neural networks with application to control the SFM chaotic system, AIMS Math., 8 (2023), 27376–27395. https://doi.org/10.3934/math.20231401 doi: 10.3934/math.20231401
    [16] N. Tan, P. Yu, F. Ni, New varying-parameter recursive neural networks for model-free kinematic control of redundant manipulators with limited measurements, IEEE Trans. Instrum. Meas., 71 (2022), 1–14. https://doi.org/10.1109/TIM.2022.3161713 doi: 10.1109/TIM.2022.3161713
    [17] R. Abbassi, H. Jerbi, M. Kchaou, T. E. Simos, S. D. Mourtas, V. N. Katsikis, Towards higher-order zeroing neural networks for calculating quaternion matrix inverse with application to robotic motion tracking, Mathematics, 11 (2023), 2756. https://doi.org/10.3390/math11122756 doi: 10.3390/math11122756
    [18] V. N. Kovalnogov, R. V. Fedorov, I. I. Shepelev, V. V. Sherkunov, T. E. Simos, S. D. Mourtas, et al., A novel quaternion linear matrix equation solver through zeroing neural networks with applications to acoustic source tracking, AIMS Math., 8 (2023), 25966–25989. https://doi.org/10.3934/math.20231323 doi: 10.3934/math.20231323
    [19] Y. Zhang, S. S. Ge, Design and analysis of a general recurrent neural network model for time-varying matrix inversion, IEEE Trans. Neur. Net., 16 (2005), 1477–1490. https://doi.org/10.1109/TNN.2005.857946 doi: 10.1109/TNN.2005.857946
    [20] Y. Chai, H. Li, D. Qiao, S. Qin, J. Feng, A neural network for Moore-Penrose inverse of time-varying complex-valued matrices, Int. J. Comput. Intell. Syst., 13 (2020), 663–671. https://doi.org/10.2991/ijcis.d.200527.001 doi: 10.2991/ijcis.d.200527.001
    [21] W. Wu, B. Zheng, Improved recurrent neural networks for solving Moore-Penrose inverse of real-time full-rank matrix, Neurocomputing, 418 (2020), 221–231. https://doi.org/10.1016/j.neucom.2020.08.026 doi: 10.1016/j.neucom.2020.08.026
    [22] W. Jiang, C. L. Lin, V. N. Katsikis, S. D. Mourtas, P. S. Stanimirović, T. E. Simos, Zeroing neural network approaches based on direct and indirect methods for solving the Yang-Baxter-like matrix equation, Mathematics, 10 (2022), 1950. https://doi.org/10.3390/math10111950 doi: 10.3390/math10111950
    [23] H. Jerbi, H. Alharbi, M. Omri, L. Ladhar, T. E. Simos, S. D. Mourtas, et al., Towards higher-order zeroing neural network dynamics for solving time-varying algebraic Riccati equations, Mathematics, 10 (2022), 4490. https://doi.org/10.3390/math10234490 doi: 10.3390/math10234490
    [24] V. N. Katsikis, P. S. Stanimirović, S. D. Mourtas, L. Xiao, D. Karabasević, D. Stanujkić, Zeroing neural network with fuzzy parameter for computing pseudoinverse of arbitrary matrix, IEEE Trans. Fuzzy Syst., 30 (2022), 3426–3435. https://doi.org/10.1109/TFUZZ.2021.3115969 doi: 10.1109/TFUZZ.2021.3115969
    [25] H. Alharbi, H. Jerbi, M. Kchaou, R. Abbassi, T. E. Simos, S. D. Mourtas, et al., Time-varying pseudoinversion based on full-rank decomposition and zeroing neural networks, Mathematics, 11 (2023), 600. https://doi.org/10.3390/math11030600 doi: 10.3390/math11030600
    [26] L. Xiao, Y. Zhang, W. Huang, L. Jia, X. Gao, A dynamic parameter noise-tolerant zeroing neural network for time-varying quaternion matrix equation with applications, IEEE Trans. Neural Networks Learn. Syst., 2022, 1–10. https://doi.org/10.1109/TNNLS.2022.3225309
    [27] S. D. Mourtas, V. N. Katsikis, Exploiting the Black-Litterman framework through error-correction neural networks, Neurocomputing, 498 (2022), 43–58. https://doi.org/10.1016/j.neucom.2022.05.036 doi: 10.1016/j.neucom.2022.05.036
    [28] V. N. Kovalnogov, R. V. Fedorov, D. A. Generalov, A. V. Chukalin, V. N. Katsikis, S. D. Mourtas, et al., Portfolio insurance through error-correction neural networks, Mathematics, 10 (2022), 3335. https://doi.org/10.3390/math10183335 doi: 10.3390/math10183335
    [29] S. D. Mourtas, C. Kasimis, Exploiting mean-variance portfolio optimization problems through zeroing neural networks, Mathematics, 10 (2022), 3079. https://doi.org/10.3390/math10173079 doi: 10.3390/math10173079
    [30] S. Qiao, Y. Wei, X. Zhang, Computing time-varying ML-weighted pseudoinverse by the Zhang neural networks, Numer. Funct. Anal. Optim., 41 (2020), 1672–1693. https://doi.org/10.1080/01630563.2020.1740887 doi: 10.1080/01630563.2020.1740887
    [31] X. Wang, P. S. Stanimirovic, Y. Wei, Complex ZFs for computing time-varying complex outer inverses, Neurocomputing, 275 (2018), 983–1001. https://doi.org/10.1016/j.neucom.2017.09.034 doi: 10.1016/j.neucom.2017.09.034
    [32] W. R. Hamilton, On a new species of imaginary quantities, connected with the theory of quaternions, Proceedings of the Royal Irish Academy (1836–1869), 2 (1840), 424–434.
    [33] A. Szynal-Liana, I. Włoch, Generalized commutative quaternions of the Fibonacci type, Bol. Soc. Mat. Mex., 28 (2022), 1. https://doi.org/10.1007/s40590-021-00386-4 doi: 10.1007/s40590-021-00386-4
    [34] D. Pavllo, C. Feichtenhofer, M. Auli, D. Grangier, Modeling human motion with quaternion-based neural networks, Int. J. Comput. Vis., 128 (2020), 855–872. https://doi.org/10.1007/s11263-019-01245-6 doi: 10.1007/s11263-019-01245-6
    [35] S. Giardino, Quaternionic quantum mechanics in real Hilbert space, J. Geom. Phys., 158 (2020), 103956. https://doi.org/10.1016/j.geomphys.2020.103956 doi: 10.1016/j.geomphys.2020.103956
    [36] A. M. S. Goodyear, P. Singla, D. B. Spencer, Analytical state transition matrix for dual-quaternions for spacecraft pose estimation, AAS/AIAA Astrodynamics Specialist Conference, 2019, Univelt Inc., 2020,393–411.
    [37] M. E. Kansu, Quaternionic representation of electromagnetism for material media, Int. J. Geom. Meth. Mod. Phys., 16 (2019), 1950105. https://doi.org/10.1142/S0219887819501056 doi: 10.1142/S0219887819501056
    [38] E. Özgür, Y. Mezouar, Kinematic modeling and control of a robot arm using unit dual quaternions, Rob. Auton. Syst., 77 (2016), 66–73. https://doi.org/10.1016/j.robot.2015.12.005 doi: 10.1016/j.robot.2015.12.005
    [39] S. Stepień, P. Superczyńska, Modified infinite-time state-dependent Riccati equation method for nonlinear affine systems: quadrotor control, Appl. Sci., 11 (2021), 10714. https://doi.org/10.3390/app112210714 doi: 10.3390/app112210714
    [40] M. Joldeş, J. M. Muller, Algorithms for manipulating quaternions in floating-point arithmetic, 2020 IEEE 27th Symposium on Computer Arithmetic (ARITH), 2020, 48–55. https://doi.org/10.1109/ARITH48897.2020.00016
    [41] F. Zhang, Quaternions and matrices of quaternions, Linear Algebra Appl., 251 (1997), 21–57. https://doi.org/10.1016/0024-3795(95)00543-9 doi: 10.1016/0024-3795(95)00543-9
    [42] J. Dai, P. Tan, X. Yang, L. Xiao, L. Jia, Y. He, A fuzzy adaptive zeroing neural network with superior finite-time convergence for solving time-variant linear matrix equations, Knowl.-Based Syst., 242 (2022), 108405. https://doi.org/10.1016/j.knosys.2022.108405 doi: 10.1016/j.knosys.2022.108405
    [43] P. S. Stanimirović, S. Srivastava, D. K. Gupta, From Zhang Neural Network to scaled hyperpower iterations, J. Comput. Appl. Math., 331 (2018), 133–155. https://doi.org/10.1016/j.cam.2017.09.048 doi: 10.1016/j.cam.2017.09.048
    [44] A. K. Gupta, Numerical methods using MATLAB, CA: Apress Berkeley, 2014. https://doi.org/10.1007/978-1-4842-0154-1
    [45] V. N. Kovalnogov, R. V. Fedorov, D. A. Demidov, M. A. Malyoshina, T. E. Simos, V. N. Katsikis, et al., Zeroing neural networks for computing quaternion linear matrix equation with application to color restoration of images, AIMS Math., 8 (2023), 14321–14339. https://doi.org/10.3934/math.2023733 doi: 10.3934/math.2023733
    [46] H. Su, R. Luo, M. Huang, J. Fu, Robust fixed time control of a class of chaotic systems with bounded uncertainties and disturbances, Int. J. Control Autom. Syst., 20 (2022), 813–822. https://doi.org/10.1007/s12555-020-0782-1 doi: 10.1007/s12555-020-0782-1
    [47] J. Singer, Y. Wang, H. H. Bau, Controlling a chaotic system, Phys. Rev. Lett., 66 (1991), 1123. https://doi.org/10.1103/PhysRevLett.66.1123
    [48] W. He, T. Luo, Y. Tang, W. Du, Y. Tian, F. Qian, Secure communication based on quantized synchronization of chaotic neural networks under an event-triggered strategy, IEEE Trans. Neural Networks Learn. Syst., 31 (2020), 3334–3345. https://doi.org/10.1109/TNNLS.2019.2943548
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1746) PDF downloads(95) Cited by(0)

Figures and Tables

Figures(2)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog