1.
Introduction
In this work, we consider the following multi-linear system:
where A∈R[m,n] and y,b∈Rn. Here, R[m,n] and Rn denote the sets of the m-order n-dimensional tensors and n-dimensional real vectors, respectively. Besides, the tensor-vector product Aym−1 in (1.1) is defined by [1]:
The multi-linear system (1.1) arises in many scientific and engineering fields, such as the tensor complementarity problems, image processing, numerical partial differential equations and so on [2,3,4,5,6,7,8]. Moreover, it is demanded to compute the solutions of multi-linear system (1.1) in many practical problems, thus solving the multi-linear system (1.1) as one of the topics of very active research in computational mathematics. So in this paper, we focus on the numerical solution of the multi-linear system (1.1).
It is noteworthy that the coefficient tensors of the multi-linear systems have special structures and they are M-tensors in lots of practical applications. Thus, it is inevitable to solve the multi-linear systems with M-tensors in many practical problems. In view of this fact, designing the efficient methods for solving the multi-linear systems with M-tensors has been a hot research topic. Many researchers have established many results and numerous methods for the multi-linear systems with M-tensors recently. For instance, Ding and Wei [8] proved that the multi-linear system (1.1) has a unique positive solution when A is a strong M-tensor and b>0, and they also designed the Newton, Jacobi and Gauss-Seidel methods for (1.1). Han [9] proposed the homotopy method for the multi-linear systems with M-tensors. Xie et al. [10] derived the tensor methods for solving symmetric M-tensor systems. Wang et al. [11] proposed the neural networks based method for (1.1). He et al. [12] developed a Newton-type method to solve multi-linear systems with M-tensors. Besides, tensor splitting method [13,14] has been presented for solving (1.1) recently. Among these methods, the tensor splitting method is one of the most popular methods, because it is efficient and easy to be implemented in the practical calculation. To solve multi-linear systems effectively, Liu et al. [13] proposed the tensor splitting A=E−F, and then established a general tensor splitting iterative method in the following form:
where M(E)−1F is called the iteration tensor of the tensor splitting method (1.2), and M(E) stands for the majorization matrix of the tensor E (see Definition 3). As stated in [13], the spectral radius of the iteration tensor can be served as an approximate convergence rate for the tensor splitting iterative method (1.2).
It is well-known that preconditioning technique can decrease the convergence factors of the iteration methods so as to accelerate their convergence speeds. Thus, it has been widely applied to the tensor splitting methods and many efficient preconditioners have been designed for (1.1) in recent years. In [14], Li et al. constructed the preconditioner Pα=I+Sα with
and αq∈R (q=1,⋯,n−1) are parameters. They proposed the related preconditioned Gauss-Seidel type and successive over-relaxation (SOR) type methods for (1.1). Then, Cui et al. [15] put forward a preconditioner Pmax=I+Smax with
where ki=min{j|maxj|aij⋯j|,i<j,i<n}. It is shown in [15] that the preconditioner Pmax outperforms Pα from the perspective of both the theoretical and numerical results. Subsequently, Zhang et al. [16] derived a new preconditioner ˜Pα=I+Gα, where
and αw∈R (w=2,⋯,n) are parameters, and proposed three preconditioned Jacobi type methods. Then, by making use of the preconditioner ˜Pα, Liu et al. [17] presented a new preconditioned SOR method for the multi-linear system (1.1), and a new preconditioned Gauss-Seidel type method can be obtained as the relaxation factor τ=1. Very recently, Chen and Li [18] constructed a new preconditioner Pγβ=I+Sγβ with
where γj>0 and βj (j=2,⋯,n) are constants. Besides, Cui et al. [19] introduced a matrix Rβ into the preconditioner Pα=I+Sα [14] and proposed a more effective preconditioner ˆPβ=I+Sα+Rβ with
and βq∈R (j=2,⋯,n). Xie and Miao [20] stated that the preconditioning effect of Pα is not observed on the last horizontal slice of the system tensor A and derived a new preconditioner ˆPγ=I+Sα+Rγ and the corresponding preconditioned Gauss-Seidel method, where
and γt∈R (t=1,⋯,n−1) are parameters. Comparison theorems and test results in [20] indicate that the preconditioned Gauss-Seidel method in [20] is more efficient than the one in [14]. In the sequel, Nobakht-Kooshkghazi and Najafi-Kalyani [21] presented a preconditioner ˜P=I+Sα+Rγ+Cα, where (Cα)kn=−αkakn⋯n (k=1,⋯,n−1) and the remaining elements in Cα are zero. Based on it, the authors in [21] established a new preconditioned Gauss-Seidel method for solving (1.1) and compared it with the one in [20]. Very recently, An et al. [22] replaced the matrix Sα in ˜P=I+Sα+Rγ+Cα [21] by Smax and proposed a new preconditioner ˆP=I+Smax+Rγ+Cα. It is proved in [22] that the corresponding preconditioned Gauss-Seidel method has faster convergence rate than the one in [21].
In addition to the preconditioners for multi-linear systems with M-tensors, some researchers have extended the related results to the multi-linear systems with other or more general tensors. For example, Wang et al. [23] investigated and developed a preconditioned tensor splitting accelerated over-relaxation (AOR) for solving the multi-linear systems with nonsingular H-tensors. Moreover, Cui et al. [24] provided a general preconditioner that contains several existing preconditioners. And they constructed two different preconditioned SOR-type splitting methods for the multi-linear systems with more general Z-tensors. For more preconditioners of the multi-linear system (1.1), we can refer to [25,26] and the references therein.
It can be seen that the preconditioners developed in [14,15,16,17,18,19,24] have preconditioning effect on some but not all horizontal slices of the coefficient tensor A in (1.1) [27]. That is, the preconditioners in [14,15,19,24] have no preconditioning effect on the last horizontal slice of the system tensor A, and the preconditioners in [16,17,18] have no preconditioning effect on the first horizontal slice of A. Thus, in this work, we will construct a new preconditioner for the multi-linear system (1.1), which has preconditioning effect on all horizontal slices of A and ameliorates the numerical performances of the preconditioners Pα [14] and ˜Pα [16,17]. Inspired by the constructions of the preconditioners in [19,20,21,22], and based on the fact that utilizing more entries of the majorization matrix of coefficient tensor in the preconditioner can make it more efficient, we design a new preconditioner Pα,β=I+Gα+Rβ. It is constructed by introducing some elements of the first row of the majorization matrix of the system tensor A into the preconditioner ˜Pα=I+Gα [16], where
and β=(βt),α=(αt)∈Rn−1, αt,βt∈R (t=2,⋯,n). The new preconditioner Pα,β=I+Gα+Rβ has the form
For the preconditioned multi-linear system with the new preconditioner Pα,β, we propose three Gauss-Seidel type splittings of preconditioned system tensor, then establish three new preconditioned Gauss-Seidel type methods. We first prove that the preconditioned coefficient tensor is a strong M-tensor, then investigate the convergence properties of the three proposed preconditioned Gauss-Seidel type methods to show that they are convergent. In the meanwhile, we establish several comparison theorems, which indicate that the convergence factors of the proposed preconditioned Gauss-Seidel type methods are monotonically decreasing and the last one is the most efficient. Also, the third proposed preconditioned Gauss-Seidel type method has a faster convergence rate than the original Gauss-Seidel method and the one in [17]. In addition, we also propose the monotonic theorem to illustrate that for a given parameter α, the convergence factor of the third proposed preconditioned Gauss-Seidel type method decreases with the parameter β in [0,1], i.e., its convergence rate becomes faster with the parameter β in [0,1]. Finally, some numerical examples are given to validate the correctness of the theoretical results, and illustrate that the proposed preconditioned Gauss-Seidel type methods are efficient, and outperform the Gauss-Seidel method and preconditioned Gauss-Seidel ones in [14,17]. We also show the feasibilities of our preconditioner and methods by applying them to solve practical problems on higher-order Markov chain and directed graph.
The remainder of this paper is organized as follows. In Section 2, we review some basic notations, definitions, and conclusions, which will be used in the subsequent sections. In Section 3, we propose a new preconditioner and present the corresponding preconditioned Gauss-Seidel type methods for solving multi-linear system (1.1). We prove that the proposed methods are convergent in this section as well. In Section 4, we establish some comparison theorems and the monotonic theorem for the proposed preconditioned Gauss-Seidel type methods. In Section 5, numerical experiments are performed to show the efficiencies and superiorities of the proposed preconditioner and iteration methods. As applications, we also apply the preconditioned Gauss-Seidel type methods to solve the practical problems on higher-order Markov chain and directed graph in Section 6. Finally, in Section 7, some conclusions are given to end this paper.
2.
Preliminaries
In this section, we list some notations, definitions, and lemmas, which will be used throughout this paper.
Let 0, O, and O denote a zero vector, a zero matrix, and a zero tensor, respectively. Let A,B∈R[m,n], A≥B mean that each entry of A is no less than the corresponding one of A. If all elements of B are nonnegative, then B is called a nonnegative tensor. The set of all m-order n-dimensional nonnegative tensors is denoted by R[m,n]+. Let ρ(B) = max{|λ||λ∈σ(B)} be the spectral radius of B, where σ(B) is the set of all eigenvalues of B. Let m≥2 be an integer, and a unit tensor Im = (μj1⋯jm)∈C[m,n] is given by
Below some needed definitions are listed.
Definition 1. [28] Let T∈R[m,n]. A pair (ζ, z)∈C×(Cn∖0) is called an eigenpair of T if they satisfy the equation
where z[m−1]=(zm−11,⋯,zm−1n)T. Further, (ζ, z) is called an H-eigenpair of T if both ζ and z are real.
Definition 2. [7] Let U∈R[m,n]. U is called a Z-tensor if its non-diagonal entries are nonpositive, U is called an M-tensor if there exist a tensor Q≥O and a positive constant δ≥ρ(Q) such that
In addition, if δ>ρ(Q), then U is called a strong M-tensor.
Definition 3. [13] Let U∈R[m,n]. The majorization matrix M(U) of U is defined as an n×n matrix with its entries
Definition 4. [34] Let F∈R[2,n] and U=(ui1⋯im)∈R[m,n]. The matrix-tensor product T=FU∈R[m,n] is defined by
The above equation can be written as T(1)=(FU)(1)=FU(1), where T(1) and U(1) are the matrices obtained from T and U flattened along the first index [27].
Definition 5. [29] Let W∈R[m,n]. If M(W) is nonsingular and W=M(W)Im, we call M(W)−1 the order 2 left-inverse of W.
Definition 6. [13] Let W∈R[m,n]. If W has an order 2 left-inverse, then W is called a left-invertible tensor.
Definition 7. [13] Let W, E, F∈R[m,n]. W=G−H is said to be a splitting of W if G is left-invertible; a convergent splitting if ρ(M(G)−1H)≤1; a regular splitting of W if M(G)−1≥O and H≥O; and a weak regular splitting if M(G)−1≥O and M(G)−1H≥O.
Definition 8. [7] A tensor A∈R[m,n] is said to be a semi-positive tensor, if there exists a vector x>0 such that Axm−1>0.
Additionally, some significant lemmas are reviewed in the following.
Lemma 1. [13] If W∈R[m,n] is a Z-tensor, then the following conditions are equivalent:
(1) W is a strong M-tensor;
(2) There exists y≥0 such that Wym−1>0;
(3) W has a convergent (weak) regular splitting;
(4) All (weak) regular splittings of W are convergent.
Lemma 2. [14] Let T∈R[m,n]+, and it holds that
and
Lemma 3. [8] If T∈R[m,n] is a strong M-tensor, then for every b>0, the multi-linear system Tzm−1=b has a unique positive solution.
Lemma 4. [19] Let T∈R[m,n] be a strong M-tensor. If T=G−H is a splitting with G being a Z-tensor and H being nonnegative, then the splitting T=G−H is a convergence regular splitting.
Lemma 5. [14] Let T∈R[m,n] be a strong M-tensor and T=G1−H1=G2−H2 be two weak regular splittings with M(G2)−1≥M(G1)−1. If the Perron vector z of M(G2)−1H2 satisfies Tzm−1≥0, then
Lemma 6. [14] Let T∈R[m,n] and T=U1−V1=U2−V2 be a weak regular splitting and a regular splitting, respectively. If F2≤F1, F2≠O, then one of the following statements holds:
(1) ρ(M(U2)−1V2)≤ρ(M(U1)−1V1)<1.
(2) ρ(M(U2)−1V2)≥ρ(M(U1)−1V1)≥1. Furthermore, if U2<V1, U2≠O and ρ(M(U1)−1V1)>1, the first inequality is strict.
Lemma 7. [7] A semi-positive Z-tensor has all positive diagonal entries.
Lemma 8. [13] If A is a strong M-tensor, then M(A) is a nonsingular M-matrix.
Lemma 9. [37] Let A and B be two nonsingular M-matrices. If B≥A, then A−1≥B−1≥O.
3.
Three new preconditioned Gauss-Seidel type methods and their convergence analyses
In this section, we present a new preconditioner Pαβ, then establish three new preconditioned Gauss-Seidel type methods and their convergence properties.
We consider a new preconditioner Pαβ=I+Gα+Rβ, where
and α=(αj), β=(βj) (j=2,⋯,n) are parameters. By premultiplying multi-linear system (1.1) by the new preconditioner Pαβ=I+Gα+Rβ, we transform the original system (1.1) into the preconditioned form
with Aαβ=PαβA=(I+Gα+Rβ)A, bαβ=Pαβb. Without loss of generality, we assume that aj⋯j=1 (j=1,⋯,n) as in [13,14,15,16,18,19,20,21,22,23,24].
Let A=Im−L−F, where L=LIm, −L is the strictly lower triangular part of M(A) and Im is a unit tensor. According to GαL=O, by some computations, we can obtain
Let GαF=D1−L1−F1 and RβL=D2−F2, where D1=D1Im, D2=D2Im and L1=L1Im. Here, D1 and D2 are the diagonal parts of M(GαF) and M(RβL), respectively, and −L1 is the strictly lower triangular part of M(GαF).
In what follows, we will study the convergence of the new preconditioned Gauss-Seidel type methods. To this end, we need the following lemma, which shows that Aαβ is a strong M-tensor.
Lemma 10. Let A∈R[m,n] be a strong M-tensor and αj,βj∈[0,1] (j=2,⋯,n), then Aαβ is a strong M-tensor.
Proof. We first prove that Aαβ is a Z-tensor. We denote Aαβ=(˜aji2,⋯,im). By Aαβ=(I+Gα+Rβ)A,
Now we prove that the non-diagonal entries of Aαβ are nonpositive in the following cases:
(1) When j=1, (i2,⋯,im)=(k,⋯.k),k≠j,k∈{2,⋯,n}, it holds that
For (i2,⋯,im)≠(k,⋯.k), k∈{1,2,⋯,n}, we have
(2) When j=2,⋯,n and (i2,⋯,im)=(1,⋯,1), it follows that
For (i2,⋯,im)=(k,⋯,k),k∈{2,⋯,n}, it has
For (i2,⋯,im)≠(k,⋯.k),k∈{1,2,⋯,n}, we have
Thus, for (j,i2,⋯,im)≠(j,j,⋯,j), when 0≤αj≤1, and 0≤βj≤1 (j=2,⋯,n), we have ˜aji2,⋯,im≤0. Therefore, Aαβ is a Z-tensor. Now we prove that the Z-tensor Aαβ is a strong M-tensor. Having in mind that A is a strong M-tensor, according to Lemma 1, there exists x≥0 such that Axm−1>0. Since Gα≥O and Rβ≥O, we can conclude that Aαβxm−1=(I+Gα+Rβ)Axm−1=Axm−1+GαAxm−1+RβAxm−1>0. Therefore, again using Lemma 1, Aαβ is a strong M-tensor. □
Theorem 1. For Pαβ=I+Gα+Rβ with αj,βj∈[0,1] (j=2,⋯,n), the preconditioned multi-linear system PαβAxm−1=Pαβb has the same unique positive solution as in Axm−1=b.
Proof. The conclusion of this theorem is obtained by Lemma 3 and Lemma 10. □
Below, we construct three new preconditioned Gauss-Seidel type methods
in terms of the following three Gauss-Seidel type splittings of Aαβ
respectively. According to (3.2), the three preconditioned Gauss-Seidel type iteration tensors are:
In [17], when the relaxation factor is τ=1, the Gauss-Seidel type splitting of ˜A=(I+Gα)A is defined as:
If M(ˇE4)−1 exists, then we use ˇT=M(ˇE4)−1ˇF4 to denote the corresponding iteration tensor.
Next, we prove that the three Gauss-Seidel type splittings in (3.2) are convergent.
Theorem 2. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n), then Aαβ=ˇE1−ˇF1=ˇE2−ˇF2=ˇE3−ˇF3 are convergence regular splittings.
Proof. We divide our proof into three cases.
(1) It is not difficult to verify that L−GαIm≥O and F−RβIm≥O under the conditions αj,βj∈[0,1] (j=2,⋯,n), which means that the off-diagonal entries of ˇE1=Im−(L−GαIm) are non-positive, and, therefore, ˇE1 is a Z-tensor. Besides, ˇF1=F+GαF−RβIm+RβL+RβF is a nonnegative tensor in view of F−RβIm≥O. Then, according to Lemma 4, Aαβ=ˇE1−ˇF1 is a convergence regular splitting.
(2) Inasmuch as −L1 is the strictly lower triangular part of M(GαF) and GαF≥O, it has −L1≥O. It follows from L−GαIm≥O that the off-diagonal entries of ˇE2=(Im−D1)−(L−GαIm−L1) are non-positive, so ˇE2 is a Z-tensor. Owing to F−RβIm≥O, it holds that ˇF2=F−F1−RβIm+RβL+RβF≥O. By making use of Lemma 4, we have that Aαβ=ˇE2−ˇF2 is a convergence regular splitting.
(3) In the same way, we can prove that ˇE3=(Im−D1−D2)−(L−GαIm−L1) is a Z-tensor, and ˇF3=F−F1−RβIm−F2+RβF≥O, then Aαβ=ˇE3−ˇF3 is a convergence regular splitting by Lemma 4. □
4.
Comparison theorem
In this section, we first prove that the third proposed Gauss-Seidel splitting of Aαβ (the corresponding method is denoted by "PˇT3G-S" throughout this paper) is optimal. Then, we present several comparison theorems, which show that the PˇT3G-S method has faster convergence rate than the original Gauss-Seidel method and the one in [17]. Lastly, we prove that for a given α, the convergence speed of the PαβG-S method increases gradually with the parameters βj (j=2,⋯,n) in the interval [0, 1].
Theorem 3. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1 (j=2,⋯,n), then Aα=E4−F4 is a convergence regular splitting.
Proof. According to Theorem 2, we see that ˇE4=(Im−D1)−(L−GαIm−L1) is a Z-tensor. It is easy to verify that F4=F−F1 is a nonnegative tensor. Then, by Lemma 4, we have that Aα=E4−F4 is a convergence regular splitting. □
Next, we will prove that the PˇT3G-S method is optimal among the three proposed methods, and compare it with the preconditioned Gauss-Seidel type method according to the splitting ˜A=ˇE4−ˇF4 in [17].
Theorem 4. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n), then
Proof. It can be seen from Theorem 2 that Aαβ=ˇE1−ˇF1=ˇE2−ˇF2=ˇE3−ˇF3 are three convergence regular splittings. By straightforward computations, it holds that
i.e., ˇF2≤ˇF1, ˇF3≤ˇF1, and ˇF3≤ˇF2. Below, we distinguish two cases to discuss.
1) If ˇF2≠O, then it follows from Lemma 6 that ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1)<1. If ˇF3=O, then ρ(M(ˇE3)−1ˇF3)=0, which together with ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1) yields that ρ(M(ˇE3)−1ˇF3)≤ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1). If F3≠O, then it holds that ρ(M(ˇE3)−1ˇF3)≤ρ(M(ˇE2)−1ˇF2)<1 in terms of Lemma 6, which combines with ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1) and leads to ρ(M(ˇE3)−1ˇF3)≤ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1).
2) If ˇF2=O, then it follows from O≤ˇF3≤ˇF2 that ˇF3=O, and, hence, ρ(M(ˇE3)−1ˇF3)=ρ(M(ˇE2)−1ˇF2)=0≤ρ(M(ˇE1)−1ˇF1).
By summarizing the above discussions, we obtain the conclusion of this theorem. □
Remark 4.1 Theorem 4 shows that ρ(M(ˇE3)−1ˇF3)≤ρ(M(ˇE2)−1ˇF2)≤ρ(M(ˇE1)−1ˇF1), which implies that the convergence factors of the three proposed preconditioned Gauss-Seidel type methods are monotonically decreasing, and the PˇT3G-S method has the fastest convergence speed among the proposed methods.
Note that the three proposed different preconditioned Gauss-Seidel type methods are proved to be similar. From the results of numerical experiments in Sections 5 and 6, we see that there is not much difference between the three methods. Thus, in the following analyses and theorems, we mainly discuss the properties and results of the third proposed preconditioned Gauss-Seidel type method, which is denoted by "PˇT3G-S method".
Now, we give a comparison theorem to compare the convergence factors between the original Gauss-Seidel method and the PˇT3G-S method.
Theorem 5. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n). Assume that M(Im−L)−1F is irreducible, then we have
Proof. By assumptions, we can deduce that A=(Im−L)−F is a convergent regular splitting. Since M(Im−L)−1F is irreducible, by the strong Perron-Frobenius theorem [35], there exists a positive Perron vector x of M(Im−L)−1F such that
where 0<λ=ρ(M(Im−L)−1F)<1. Then it holds that
and
Since A=(Im−L)−F and M(Im−L)−1Fxm−1=λx[m−1], it has 1λFxm−1=(Im−L)xm−1 and, therefore,
By pre-multiplying equation (4.3) by Rβ, we have
Hence the combination of (4.1)–(4.4) results in
By Theorem 2, we know that Aαβ=ˇE3−ˇF3 is a convergence regular splitting, so M(ˇE3)−1≥O. Since D1,D2,−L1,Rβ, and F are nonnegative, it follows from (4.5) that ˇT3xm−1≤λxm−1. By Lemma 2, we have ρ(M(ˇE3)−1ˇF3)≤ρ(M(Im−L)−1F), which completes the proof of this theorem. □
Remark 4.2. Theorem 5 implies that the convergence rate of the PˇT3G-S method is faster than the original Gauss-Seidel method, which indicates that the new preconditioner Pα,β can improve the computational efficiency of the original Gauss-Seidel method.
In the sequel, we prove that the convergence rate of the PˇT3G-S method is faster than that of the preconditioned Gauss-Seidel type method in [17]. To this end, we start with a useful lemma below.
Lemma 11. Let A∈R[m,n] be a strong M-tensor. If 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n), then
Proof. According to Lemma 10, we know that Aαβ is a strong M-tensor. Then, by making use of Lemma 1, there exists x≥0 such that Aαβxm−1>0. Since ˇF3≥O, we can get ˇE3≥Aαβ, and, hence,
It is known from Theorem 4 that ˇE3 is a Z-tensor, thus, we can conclude from Lemma 1 that ˇE3 is a strong M-tensor. In the same way, we can prove that ˇE4 is a strong M-tensor. According to Lemma 8, both M(ˇE3) and M(ˇE4) are nonsingular M-matrices.
Besides, it follows from ˇE3=Im−L+GαIm−D1+L1−D2, ˇE4=Im−L+GαIm−D1+L1 that ˇE4≥ˇE3, which leads to M(ˇE4)≥M(ˇE3). Having in mind that M(ˇE3) and M(ˇE4) are nonsingular M-matrices, then
in view of Lemma 9. The proof is completed. □
Theorem 6. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n). Suppose that ¯x is the Perron vector of M(ˇE3)−1ˇF3, and if ¯x satisfies A¯xm−1≥0, then
Proof. For Aαβ=ˇE3−ˇF3 and ˜A=E4−F4, we consider the following two splittings of the tensor A:
where E′3=P−1α,βˇE3, F′3=P−1α,βˇF3, E′4=P−1αˇE4, and F′4=P−1αˇF4. Then, we can deduce that
and
Theorems 2 and 3 prove that both Aαβ=ˇE3−ˇF3 and ˜A=ˇE4−ˇF4 are convergence regular splittings, then it follows that
which shows that A=E′3−F′3=E′4−F′4 are two weak regular splittings of the strong M-tensor A.
From Lemma 11, we get M(ˇE3)−1≥M(ˇE4)−1, which together with Pα,β≥Pα leads to
Inasmuch as A=E′3−F′3=E′4−F′4 are two convergence weak regular splittings, M(E′3)−1≥M(E′4)−1, and the Perron vector ¯x of M(E′3)−1F′3 satisfies A¯xm−1≥0, we have
in view of Lemma 5, which is equivalent to
□
Finally, we will give a conclusion that for a given α, the spectral radius of the iteration tensor of the PˇT3G-S method decreases with β.
Theorem 7. Let A∈R[m,n] be a strong M-tensor and 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n). Suppose that ¯x is the Perron vector of M(E3αβ2)−1F3αβ2, if ¯x satisfies A¯xm−1≥0, αj,β1j,β2j∈[0,1] (j=2,⋯,n) and β1=(β12,⋯,β1n)≤β2=(β22,⋯,β2n), then the following inequality holds
Proof. Since ˇE3=Im−L+GαIm−D1+L1−D2 and β1≤β2, we can deduce that ˇE3αβ2≤ˇE3αβ1, which yields that M(ˇE3αβ2)≤M(ˇE3αβ1). According to Lemma 11, we see that ˇE3 is a strong M-tensor when 0≤αj≤1, 0≤βj≤1 (j=2,⋯,n). Then, from Lemma 8, we obtain that M(ˇE3αβ2) and M(ˇE3αβ1) are nonsingular M-matrices. Therefore, M(ˇE3αβ2)−1≥M(ˇE3αβ1)−1≥O in terms of Lemma 9.
Besides, we derive
which means that Aαβ2¯xm−1≥Aαβ1¯xm−1≥0. This combines with M(ˇE3αβ2)−1≥M(ˇE3αβ1)−1 and results in
i.e.,
So we can get M(ˇE3αβ1)−1ˇF3αβ1¯xm−1≥M(ˇE3αβ2)−1ˇF3αβ2¯xm−1. Inasmuch as ¯x≠0 is the nonnegative Perron vector of the nonnegative tensor M(ˇE3αβ2)−1ˇF3αβ2, it has
According to Lemma 5, we have
□
Remark 4.3 From Theorem 3.5 of [17], we observe that ρ(M(ˇE4)−1ˇF4) has the smallest value as αj=1 (j=2,⋯,n). Thus we adopt αj=1 (j=2,⋯,n) in ρ(M(ˇE3)−1ˇF3). Besides, it can be concluded from Theorem 7 that for given αi (i=2,⋯,n), ρ(M(ˇE3)−1ˇF3) is monotonically decreasing with βj (j=2,⋯,n) in the interval [0, 1]. Thus, we can choose αj=βj=1 (j=2,⋯,n) for the PˇT3G-S method in practical computations.
5.
Numerical experiments
This section provides three numerical examples to illustrate the correctness of the theoretical results and the efficiencies of the proposed methods. To show the advantages of the proposed preconditioned Gauss-Seidel type methods, we also compare the convergence behaviors of the proposed methods with those of the original Gauss-Seidel, and the existing preconditioned Gauss-Seidel ones in [14,17], with respect to the number of iteration steps (IT) and the computational time in seconds (CPU).
All numerical results are computed in MATLAB (version R2016b) on a personal computer with Intel (R) Pentium (R) CPU G3240T 2.870 GHz, 16.0 GB memory and Windows 10 system. We use the power method in [33] to compute the spectral radii of all tested preconditioned Gauss-Seidel type methods. Besides, all iterations are terminated once the residual(RES) satisfies RES:=‖Aym−1k−b‖2≤10−10, or IT exceeds the maximal number of iteration steps 10,000. According to Remark 3.7 in [17] and Remark 5.7 in [14], we see that ρ(Tα) and ρ(ˇT) attain minimum values as αj=1 (j=2,…,n). Moreover, based on Theorem 7 and Remark 4.3, we can choose αl=1 and βl=1 (l=2,…,n) in the proposed PˇT3G-S method. Then, we compare the numerical results of the preconditioned Gauss-Seidel type methods with all parameters being 1. Also, we list the numerical experimental results of the proposed PˇT3G-S method when the parameters α and β are not equal, and give a reasonable suggestion of the values of the parameters in the PˇT3G-S method. The product Axm−1 is computed by transforming into the matrix-vector product Axm−1=A(1)(x⊗⋯⊗x⏟m), where ⊗ is the Kronecker product. The product of a matrix U and a tensor V is computed by (UV)(1)=UV(1). In our computations, the preconditioned Gauss-Seidel type methods in [14] and [17] are denoted by PTαG-S and PˇTG-S, respectively. The three proposed preconditioned Gauss-Seidel type methods are denoted by PˇT1G-S, PˇT2G-S, and PˇT3G-S, respectively.
Example 1. [32] Let A∈R[3,n] and b∈Rn, where
and
This example comes from the high order bellman equations in [32] with few modifications. It can be verified that A is a strong M-tensor. Both the righthand side vector b and the initial vector y0 are chosen to be ones(n,1).
Example 2. [8,22] Consider the multi-linear system (1.1) with coefficient tensor A=sI3−B and righthand side vector b=(1,1,⋯,1)T, where B is generated randomly by Matlab, and s=1.01max1≤j≤n(Be2)j with e=(1,1,⋯,1)T.
Example 3. [23] Let A∈R[3,n] and b∈Rn with
and bi=1 (i=1,⋯,n).
This example comes from [23] with few modifications. We can verify that A is a strictly diagonally dominant Z-tensor with positive diagonal elements, then A is a strong M-tensor. The initial vector is taken as y0=ones(n,1).
We compare the convergence factors of the tested Gauss-Seidel and preconditioned Gauss-Seidel type methods for Examples 1, 2, and 3 in Tables 1, 5, and 9, respectively. As observed in Tables 1, 5, and 9, the convergence factors of the preconditioned Gauss-Seidel type methods are smaller than that of the original Gauss-Seidel one, which reveals that applying the preconditioning technique can accelerate the tensor splitting methods and is in accordance with the result in Theorem 5. Also, the convergence factors of the three proposed preconditioned Gauss-Seidel type methods satisfy ρ(ˇT3)≤ρ(ˇT2)≤ρ(ˇT1) for all cases. This coincides with the conclusion in Theorem 4. Moreover, ρ(ˇT2) and ρ(ˇT3) are smaller than ρ(Tα) [14] and ρ(ˇT) [17], which demonstrates the correctness of Theorem 6. We conclude that the proposed PˇT2G-S and PˇT3G-S methods have faster convergence rates than the existing ones in [14,17].
To reflect the changing of ρ(ˇT3) with respect to parameters, in Tables 2, 6, and 10, we report the convergence factor ρ(ˇT3) and the corresponding IT of the proposed PˇT3G-S method with α=β from 0.1 to 1 with step size 0.1 for Examples 1–3, respectively. The corresponding IT of the PˇT3G-S method are listed in brackets. Numerical results in Tables 2, 6, and 10 show that the convergence factor and IT of the PˇT3G-S method decrease with increasing of α=β in [0, 1] for Example 2. However, ρ(ˇT3) increases first with respect to α=β in [0, 0.9], and then decreases with respect to α=β in [0.9, 1]. The reason is that the assumption A¯xm−1≥0 in Theorem 7 is invalid. Besides, the IT and convergence factor of the PˇT3G-S method are monotonically increasing with the problem size n for Example 1, while it is opposite for Example 3.
Also, we list the convergence factor ρ(ˇT3) and IT of the proposed PˇT3G-S method with different values of (α,β) and α,β from 0.1 to 1 with step size 0.1 for Examples 1–3 in Tables 3, 7, and 11, respectively. The best results are marked in bold. It can be observed in Tables 3, 7, and 11 that for a given α, ρ(ˇT3) and IT of the proposed PˇT3G-S method become smaller with β, which coincides with the result in Theorem 7. While for a given β, ρ(ˇT3) and IT of the proposed PˇT3G-S method are not monotonically decreasing with α for all cases. In addition, the PˇT3G-S method has the highest computational efficiency as α=0.1 and β=1. This reveals that we can adopt small values of α and large values of β for the PˇT3G-S method in numerical experiments and practical computations.
Additionally, numerical results of the tested Gauss-Seidel and preconditioned Gauss-Seidel methods for Examples 1, 2, and 3 with different problem sizes are reported in Tables 4, 8, and 12, respectively. From Tables 4, 8, and 12, we observe that all tested methods can successfully compute approximate solutions satisfying the prescribed stopping criterion, and their IT and CPU time increase with increasing of the problem size n. For Examples 1 and 3, the proposed PˇT2G-S and PˇT3G-S methods have advantages over the other ones from the point of view of IT and CPU time, and the PˇT3G-S method performs the best among the tested methods. For Example 2, the three proposed methods outperform the other ones except for the case of n=2, and the proposed PˇT3G-S method is the most effective method in all tested methods, which is in accordance with the results in Theorems 5 and 6. Another observation is that for Examples 1 and 3, the proposed PˇT3G-S method is more stable than the other ones since the variation amplitude of its IT is the smallest among the tested methods.
In order to further show the efficiencies of the proposed preconditioned Gauss-Seidel type algorithms, we also apply the tested algorithms to solve the multi-linear system in Example 3 with larger scales n=100,150,200,250,300, and the obtained numerical results are listed in Table 13. From Table 13, it is observed that when n increases, the IT and CPU time of the tested methods increase. The proposed PˇT2G-S and PˇT3G-S methods return better numerical performances than the other ones in terms of IT and CPU time, and the PˇT3G-S method performs the best among the tested methods. The advantages of the PˇT2G-S and PˇT3G-S methods become more pronounced as the problem size n becomes larger. This shows that the proposed methods are feasible and efficient for solving multi-linear systems with large scales.
To investigate the dependence of the tested methods on the problem size and parameter, the comparisons for the convergence factors of the tested methods and the changes of ρ(ˇT3) with α=β for Examples 1–3 are depicted in Figures 2, 4, and 6, respectively. We plot the RES(log10) curves of the tested algorithms for Examples 1–3 in Figures 1, 3, and 5, respectively, to show the advantages of the proposed methods. Here, "PGS(Sα)", "PGS(Gα)", and "PGS-j" (j=1,2,3) denote the PTαG-S, PˇTG-S, and PˇTjG-S (j=1,2,3) methods, respectively. Their convergence factors are represented by ρ(Sα), ρ(Gα), and ρj (j=1,2,3), respectively. It can be seen from Figures 1, 3 and 5 that the PˇT2G-S and PˇT3G-S methods are superior to the other ones in view of IT, and the PˇT1G-S and PTαG-S [14] methods are comparable for Examples 1 and 3. In addition, for Example 2, the three proposed preconditioned Gauss-Seidel type methods outperform the other ones. Also, the PTαG-S method performs better than the PˇTG-S one, while it is opposite for Examples 1 and 3. This is in consistent with the results in Tables 4, 8 and 12. On the other hand, Figures 2, 4 and 6 clearly show that ρ(ˇT2) and ρ(ˇT3) are less than the other ones, and ρ(ˇT3) is the minimum one among all convergence factors except for n=2 in Example 2. Meanwhile, ρ(Tα) and ρ(ˇT1), and ρ(ˇT) and ρ(ˇT2) are close for Example 1. Furthermore, all convergence factors become closer with n for Example 2. Lastly, ρ(ˇT3) decreases first and then increases as α=β increase in [0, 1] for Examples 1 and 3, while it decreases with the increasing of α=β in [0, 1] for Example 2. The reason may be that the assumption A¯xm−1≥0 in Theorem 7 is invalid.
6.
Applications in the higher-order Markov chain and directed graph
As applications of the proposed preconditioned Gauss-Seidel type methods for solving the multi-linear system (1), we consider the higher-order Markov chain [4] and directed graph [5,6]. Recently, Li and Ng [4] presented an approximate tensor model for higher-order Markov chain. The approximate tensor model for higher-order Markov chain is given as follows:
where P=(pj1j2⋯jm) is an m-th order n-dimensional stochastic tensor (or transition probability tensor), that is:
and z=(zj) is called a stochastic vector with zj≥0, n∑j=1zj=1. The following example relating to the higher-order Markov chain comes from [36].
Example 4. [36] Let
This example comes from occupational mobility of physicists date given in [36]. Let A=Im−θP, with a positive real number θ. In [13], Liu et al. gave an equivalent equation with (6.1):
We test and compare the preconditioned Gauss-Seidel type methods in this paper and in [14,17] for this example. According to the numerical experiments in [19,24], the initial vector is adopted to be z0=1nones(n,1), and all runs stopped once the RES satisfies RES:=‖Pzm−1−z‖2≤10−10 or the IT exceeds the maximal number of iteration steps 10,000.
Example 5. [5,6] Let Γ be a directed graph with V={1,2,3,4,5}, where P={1,2} and D={3,4,5}. Then, by definition, we can get the stochastic tensor ˆQ=(ˆqi1i2i3)∈R[3,5], where ˆq221=ˆq421=ˆq521=ˆq312=ˆq412=ˆq512=0, ˆq122=ˆq133=ˆq144=ˆq155=ˆq121=ˆq212 = ˆq211=ˆq311=ˆq411=ˆq321=ˆq112=ˆq511=1, and other elements are 1/6. We apply the tested preconditioned Gauss-Seidel type methods to solve the following multi-linear PageRank:
where V=(vi1i2⋯im) with vi1i2⋯im=vi1, ∀i2,⋯,im, and v=(vi) is a stochastic vector, ξ∈(0,1) is a parameter. This example comes from [5,6,24] with few modifications. As in Example 4, we take the initial vector as z0=1nones(n,1), and all runs stopped once the RES satisfies RES:=‖Pzm−1−z‖2≤10−10 or the IT exceeds the maximal number of iteration steps 10,000.
In Tables 14–16, we list the numerical results of the tested methods for Example 4 with five different values of θ. For Example 5, in Tables 17–19, we list the same items as those in Tables 14–16. In Figures 7 and 9, we plot the RES(log10) curves of the tested methods for Example 4 and Example 5, respectively. The comparisons for convergence factors of the tested methods and the variation curves of ρ(ˇT3) with α=β for Examples 4 and 5 are displayed in Figures 8 and 10, respectively. As in Section 5, we use "PGS(Sα)", "PGS(Gα)", and "PGS-j" (j=1,2,3) to denote the PTαG-S, PˇTG-S and PˇTjG-S (j=1,2,3) methods, respectively. Their convergence factors are represented by ρ(Sα), ρ(Gα) and ρj (j=1,2,3), respectively.
By the numerical results presented in Tables 14–19, we can conclude some observations: First, all tested algorithms are convergent for all cases. Second, the IT and CPU time of all tested algorithms increase with θ(ξ). Third, the PˇT2G-S and PˇT3G-S methods have higher computational efficiencies than the other ones, except the PTαG-S method is slightly better than the PˇT2G-S one in Example 4, and the advantages become more pronounced with θ(ξ). Fourth, numerical performances of the PTαG-S and the PˇT1G-S methods are comparable, and they are more efficient than the Gauss-Seidel and PˇTG-S ones with respect to IT and CPU time. Fifth, all convergence factors increase with θ(ξ), and ρ(ˇT3) is less than the other ones. Also, ρ(ˇT3)<ρ(ˇT2)<ρ(ˇT1) holds for all cases, which coincides with the conclusion in Theorem 4. Finally, ρ(ˇT3) decreases gradually with increasing of α=β.
From Figures 7 and 9, we observe that PˇT2G-S and PˇT3G-S methods have faster convergence speeds than the other ones. The convergence behaviors of the PTαG-S and PˇT1G-S methods are comparable, and they perform better than the remaining two in view of IT. Additionally, it follows from Figures 8 and 10 that all convergence factors become larger and closer with θ(ξ), and ρ(ˇT3) is the smallest in all convergence factors. Finally, ρ(ˇT3) increases with θ(ξ) and decreases with α=β. The above results are in consistent with Theorems 5–7 and Remark 4.3.
7.
Conclusions
In this paper, by introducing some elements in the first row of the majorization matrix associated with the coefficient tensor into the preconditioner ˜Pα [16,17], we propose a new preconditioner for the multi-linear system (1.1). Compared with preconditioners in [14,17], the new preconditioner has preconditioning effect on all horizontal slices of the coefficient tensor and improves their computational efficiencies. Then we construct three new preconditioned Gauss-Seidel type methods for (1.1) according to three different Gauss-Seidel type tensor splittings of the preconditioned system tensor. Convergence and comparison theorems have been given to show that the proposed methods are convergent. Meanwhile, the last proposed method has a faster convergence rate than the original Gauss-Seidel method and the preconditioned Gauss-Seidel method in [17]. Also, it is proved that for a given α, the convergence factor of the proposed PˇT3G-S method decreases monotonically with β in [0, 1]. Finally, numerical experiments are performed to illustrate that the theoretical results are correct, and the proposed methods are efficient and have better numerical performances than the Gauss-Seidel method and preconditioned Gauss-Seidel methods in [14,17]. As applications, we apply the proposed methods to solve the multi-linear system (1.1) arising from the higher-order Markov chain and directed graph to further show their feasibilities and superiorities.
It can be observed that comparison and monotonic theorems (Theorems 6 and 7) are valid under the condition A¯xm−1≥0, which is not easy to be verified in the practical calculation; thus, we will work to weaken this assumption in the future. Last but not least, it also makes sense to study more efficient preconditioners and the related preconditioned tensor splitting methods in our future work.
Use of AI tools declaration
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Acknowledgments
This work was supported by the National Natural Science Foundation of China (No. 12361078), and the Guangxi Natural Science Foundation (No. 2024GXNSFAA010498). We would like to express our sincere thanks to the editor and the anonymous reviewers for their valuable suggestions and constructive comments which greatly improved the presentation of this paper.
Conflict of interest
The authors declare there is no conflicts of interest.