
A two-step randomized Gauss-Seidel (TRGS) method is presented for large linear least squares problem with tall and narrow coefficient matrix. The TRGS method projects the approximate solution onto the solution space by given two random columns and is proved to be convergent when the coefficient matrix is of full rank. Several numerical examples show the effectiveness of the TRGS method among all methods compared.
Citation: Yimou Liao, Tianxiu Lu, Feng Yin. A two-step randomized Gauss-Seidel method for solving large-scale linear least squares problems[J]. Electronic Research Archive, 2022, 30(2): 755-779. doi: 10.3934/era.2022040
[1] | Fang Wang, Weiguo Li, Wendi Bao, Li Liu . Greedy randomized and maximal weighted residual Kaczmarz methods with oblique projection. Electronic Research Archive, 2022, 30(4): 1158-1186. doi: 10.3934/era.2022062 |
[2] | Xuefei He, Kun Wang, Liwei Xu . Efficient finite difference methods for the nonlinear Helmholtz equation in Kerr medium. Electronic Research Archive, 2020, 28(4): 1503-1528. doi: 10.3934/era.2020079 |
[3] | Jia-Min Luo, Hou-Biao Li, Wei-Bo Wei . Block splitting preconditioner for time-space fractional diffusion equations. Electronic Research Archive, 2022, 30(3): 780-797. doi: 10.3934/era.2022041 |
[4] | Yijun Chen, Yaning Xie . A kernel-free boundary integral method for reaction-diffusion equations. Electronic Research Archive, 2025, 33(2): 556-581. doi: 10.3934/era.2025026 |
[5] | Wenya Shi, Xinpeng Yan, Zhan Huan . Faster free pseudoinverse greedy block Kaczmarz method for image recovery. Electronic Research Archive, 2024, 32(6): 3973-3988. doi: 10.3934/era.2024178 |
[6] | Yan Chang, Yukun Guo . Simultaneous recovery of an obstacle and its excitation sources from near-field scattering data. Electronic Research Archive, 2022, 30(4): 1296-1321. doi: 10.3934/era.2022068 |
[7] | Qian He, Wenxin Du, Feng Shi, Jiaping Yu . A fast method for solving time-dependent nonlinear convection diffusion problems. Electronic Research Archive, 2022, 30(6): 2165-2182. doi: 10.3934/era.2022109 |
[8] | Dongmei Yu, Yifei Yuan, Yiming Zhang . A preconditioned new modulus-based matrix splitting method for solving linear complementarity problem of $ H_+ $-matrices. Electronic Research Archive, 2023, 31(1): 123-146. doi: 10.3934/era.2023007 |
[9] | Jian-Ying Rong, Xu-Qing Liu . Hybrid principal component regression estimation in linear regression. Electronic Research Archive, 2024, 32(6): 3758-3776. doi: 10.3934/era.2024171 |
[10] | Hua Yang, Xuan Geng, Heng Xu, Yichun Shi . An improved least squares (LS) channel estimation method based on CNN for OFDM systems. Electronic Research Archive, 2023, 31(9): 5780-5792. doi: 10.3934/era.2023294 |
A two-step randomized Gauss-Seidel (TRGS) method is presented for large linear least squares problem with tall and narrow coefficient matrix. The TRGS method projects the approximate solution onto the solution space by given two random columns and is proved to be convergent when the coefficient matrix is of full rank. Several numerical examples show the effectiveness of the TRGS method among all methods compared.
We consider the approximate solutions of a large linear least squares problem
minx∈Rn‖b−Ax‖22, | (1.1) |
where A∈Rm×n, m>n, is of full clolumn rank. A and b∈Rm are known, x∈Rn is unknown to be determined. Under the condition that the linear system is consistent or inconsistent, people are interested in finding the unique least squares solution x∗=A†b, where A† is the Moore-Penrose pseudoinverse of the matrix A (A†=(ATA)−1AT, where AT denotes the transpose of the matrix A). See also references [1,2,3,4,5,6]. As we know, the least squares problem arises widely in many fields such as tomography[7,8,9], protein structure[10], machine learning[11], biological feature selection[12], and so on[13,14,15]. For solving (1.1), there are many direct methods, such as QR decomposition and singular value decomposition (SVD) [1,16]. However, for large-scale system matrices, these methods are too expensive because they consume a lot of memory space. So, some iterative methods are applied to solve large-scale linear least squares problems.
As one of the famous iterative algorithms, the Gauss-Seidel method [17] selects a coordinate dk and a step αk=argminα∈Rf(xk+αdk) in each iteration. When αk=ATjk(b−Axk)/‖Ajk‖22 and dk=ejk are accurately given, it generates the following iterative process:
xk+1=xk+ATjk(b−Axk)‖Ajk‖22ejk,k=0,1,2,⋯, | (1.2) |
where jk=(kmodn)+1, (⋅)T represents the transpose of a matrix or a vector, and ejk represents the coordinate column vector, whose jk-th entry is 1 and zero otherwise. Inspired by the randomized Kaczmarz (RK) linear convergence characteristics of Strohmer and Vershynin[18], Leventhal and Lewis[19] obtained a randomized Gauss-Seidel (RGS) method, which is also known as randomized coordinate descent (RCD) method to solve (1.1). The RGS selects the update column index jk according to the appropriate probability. Theoretical analysis shows that RGS converges linearly to the unique least squares solution x∗=A†b. Many variants of RGS are receiving extensive attention recently due to its good performance. For example, the versions of block[20,21], random greedy[22,23,24]. Other versions, see literatures [4,17,25,26] and reference therein.
In 2018, Wu[27] obtained a randomized block Gauss-Seidel (RBGS) method, which can significantly improve the convergence speed. However, the good column pavings for the RBGS is difficult to find when the column norms of the coefficient matrix fluctuate in a large range. In this paper, we propose a two-step randomimzed Gauss-Seidel method (TRGS), which does not need any columns pavings. The convergence of the TRGS algorithm is proved for (1.1) with the coefficient matrix of full rank.
This paper is organized as follows. In Section 2, some symbols and a lemma related to RGS are introduced. The convergence of RGS2 is analyzed. In Section 3, a convergent upper bound of TRGS is obtained theoretically. Several numerical experiments are reported to verify the feasibility of our proposed algorithms in Section 4. The conclusions of this paper are given in Section 5 and the proof of Theorem 2 is shown in appendix.
We first introduce the notations and definitions as follows. For the matrix Q∈Rm×n, Qi represents the ith column of the Q, λmax(QTQ) and λmin(QTQ) represent the maximum and minimum positive eigenvalues of the QTQ, respectively, (⋅)T represents the transpose of a matrix or a vector, the Qτk represents a submatrix of the Q (where τk is a set of column indexes), ‖Q‖2 and ‖Q‖F represent the spectral norm and Frobenius norm of the Q, respectively, and R(Q) is the image space of matrix Q. For a vector p∈Rn or p∈Rm, pi is the ith component of p. For constant c∈R, [c] refers to the set consising of all positive integers not exceeding c. For the matrix in (1.1), assuming that every two columns are independent and identically distributed, the correlation coefficient parameters of the matrix A are defined as follows
δ=mins≠t|ATsAt|‖As‖2‖At‖2andΔ=maxs≠t|ATsAt|‖As‖2‖At‖2,s,t∈{1,2,⋯,n}. |
Then, we have 0≤δ≤Δ<1.
The randomized Gauss-Seidel method proposed by Leventhal and Lewis[19] consists of two parts. The RGS determines a column index jk according to the probability Pr(column=jk)=‖Ajk‖22‖A‖2F and then updates xk+1 by (1.2).
The following results in [19] summarized an upper bound for the error of the solution in expectation on the convergence of RGS algorithm.
Lemma 1. Assume the least squares problems (1.1) has tall and narrow coefficient matrix A∈Rm×n with full rank. Let x∗=A†b be a solution of (1.1). Given an initial guess x0∈Rn, then the sequence {xk}∞0 generated by RGS algorithm converges linearly to the x∗ in expectation. Moreover, it satisfies
E‖xk−x∗‖2ATA≤(1−λmin(ATA)‖A‖2F)k‖x0−x∗‖2ATA,k=1,2,⋯. |
Algorithm 1 RGS2 method |
Input: A∈Rm×n,b∈Rm and x0=0∈Rn |
Output: the approximation solution xk of (1.1) |
1. for k=0,1,… do until termination criterion is satisfied |
2. Set rk=b−Axk and sk=ATrk. |
3. Select jk1 and jk2 with probability by (2.1). |
4. Compute sjk1k=ATjk1rk. |
5. Update |
yk=xk+sjk1k‖Ajk1‖22ejk1 and xk+1=yk+ATjk2(b−Ayk)‖Ajk2‖22ejk2. |
6. end for |
Now, we propose a simplified two-step randomized Gauss-Seidel method (RGS2) to solve (1.1). Algorithm 1 lists the RGS2 method, which mainly consists of two stages. The first stage is to select two different working columns by
Pr(column=jk1)=‖Ajk1‖22‖A‖2F,Pr(column=jk2)=‖Ajk2‖22‖A‖2F−‖Ajk1‖22, | (2.1) |
where jk1∈{1,2,⋯,n}, jk2∈{1,2,⋯,n}/{jk1}. The second stage is to use (1.2) twice to update xk.
If the coefficient matrix A in (1.1) is tall and narrow with full column rank, the following result gives the convergence of the RGS2 algorithm.
Theorem 1. Assume the least squares problem (1.1) has tall and full-rank coefficient matrix A∈Rm×n. Let x∗=A†b be a solution of (1.1). Given an initial guess x0∈Rn, then the iterative sequence {xk}∞0 generated by RGS2 algorithm converges linearly to the x∗ in expectation. Moreover, it satisfies
Ek‖A(ˆxk+1−x∗)‖22≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)]⋅‖A(xk−x∗)‖22 |
and
E‖xk−x∗‖2ATA≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)]k⋅‖x0−x∗‖2ATA, | (2.2) |
where τmax=‖A‖2F−minp∈[n]‖Ap‖22.
Proof. Let ˆyk and ˆxk+1 be the first and the second iterative solutions of xk obtained by single-step continuous execution of Algorithm 1, respectively. The update process is divided into two steps as follows
ˆyk=xk+ATjk1(b−Axk)‖Ajk1‖22ejk1andˆxk+1=ˆyk+ATjk2(b−Aˆyk)‖Ajk2‖22ejk2. |
Let Pjki=Im−AjkiATjki‖Ajki‖22 satisfy P2jki=Pjki,PTjki=Pjki, where i=1,2. Then, Pjki is the projection matrix, and
A(ˆxk+1−x∗)=A(ˆyk−x∗)+Ajk2ATjk2(b−Aˆyk)‖Ajk2‖22=(Im−Ajk2ATjk2‖Ajk2‖22)A(ˆyk−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22)(Im−Ajk1ATjk1‖Ajk1‖22)A(xk−x∗)=Pjk2Pjk1A(xk−x∗), |
where the second equality follows from the normal equation ATAx∗=ATb, whose jk2-th equation gives ATjk2b=ATjk2Ax∗. Taking the expectation on the equality above, and by the expectation conditional upon jk1 (we fix the choice of jk1 and averagever the random index jk2), one can get
Ek‖A(ˆxk+1−x∗)‖22=Ek‖Pjk2Pjk1A(xk−x∗)‖22=Ek[(A(xk−x∗))TPjk1Pjk2Pjk1A(xk−x∗)]=n∑jk1=1‖Ajk1‖22‖A‖2Fn∑jk2=1,jk2≠jk1‖Ajk2‖22‖A‖2F−‖Ajk1‖22(A(xk−x∗))TPjk1Pjk2Pjk1A(xk−x∗)=n∑jk1=1‖Ajk1‖22‖A‖2F(A(xk−x∗))TPjk1(Im−AAT−Ajk1ATjk1‖A‖2F−‖Ajk1‖22)Pjk1A(xk−x∗)=n∑jk1=1‖Ajk1‖22‖A‖2F(A(xk−x∗))TPjk1(Im−AAT‖A‖2F−‖Ajk1‖22)Pjk1A(xk−x∗), |
in which the last equation holds because
Ajk1ATjk1‖A‖2F−‖Ajk1‖22Pjk1=Ajk1ATjk1‖A‖2F−‖Ajk1‖22(Im−Ajk1ATjk1‖Ajk1‖22)=Ajk1ATjk1‖A‖2F−‖Ajk1‖22−Ajk1(ATjk1Ajk1)ATjk1(‖A‖2F−‖Ajk1‖22)‖Ajk1‖22=0. |
Note that
‖ATu‖22≥λmin(ATA)‖u‖22 and τmax=maxp∈[n]{‖A‖2F−‖Ap‖22}, |
then ‖Im−AAT‖A‖2F−‖Ajk1‖22‖2≤1−λmin(ATA)τmax and
Ek‖A(ˆxk+1−x∗)‖22≤(1−λmin(ATA)τmax)n∑jk1=1‖Ajk1‖22‖A‖2F(A(xk−x∗))T(Im−Ajk1ATjk1‖A‖2F)⋅A(xk−x∗)≤(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)‖A(xk−x∗)‖22. |
Therefore,
Ek‖ˆxk+1−x∗‖2ATA≤(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)‖xk−x∗‖2ATA. | (2.3) |
(2.2) is obtained from the recurrence relation of (2.3) and the full expectation formula. This completed the proof.
The minimum positive eigenvalue of λmin(ATA) will become very small if the matrix A has high correlation parameters[28]. A weak bound of the convergence in Theorem 1 will appear. Under this condition, the angle of the unit coordinate direction as the search direction for two consecutive stages may be too small, which is the main reason for the slow convergence of RGS. Inspired by the work of Needell[28] and Wu[29], we iteratively update the solution by continuously seeking two more extensive directions, that is, determine the column pairs (r,s) in advance. So a two-step randomized Gauss-Seidel algorithm is obtained. In a two-step randomized algorithm, which mainly consists of three steps as follows: First, we randomly select the two columns r,s∈[n] according to the probability criterion, then estimate the optimal parameter λopt for the first iteration with yk=xk+λoptATrrk‖Ar‖22er, and finally perform the second iteration to update the iterative solution by xk+1=yk+ATs(b−Ayk)‖As‖22es.
We consider the convergence of the TRGS algorithm. We first need the following result presented in [28].
Lemma 2. For any ϵ∈R, ϕ,ψ∈Rm, the minimizer of ‖ϵϕ+ψ‖22 is ϵopt=−<ϕ,ψ>‖ϕ‖22.
Now, we note that
Ayk=Axk+λkArATrrk‖Ar‖22andAxk+1=Ayk+AsATs(b−Ayk)‖As‖22, |
then
Axk+1=Axk+λkArATrrk‖Ar‖22+AsATs(b−(Axk+λkArATrrk‖Ar‖22))‖As‖22=λk[(Ar‖Ar‖2−μkAs‖As‖2)ATrrk‖Ar‖2]+Axk+AsATsrk‖As‖22, | (3.1) |
where μk=(Ar)T(As)‖Ar‖2‖As‖2. Obviously,
‖Axk+1−b‖22=‖λk[(Ar‖Ar‖2−μkAs‖As‖2)ATrrk‖Ar‖2]+Axk−b+AsATsrk‖As‖22‖22. |
The selection of optimal parameter λopt aims to minimize ‖Axk+1−b‖22. By Lemma 2, one can get the optimal value of λk, that is,
λopt=−(ATr‖Ar‖2−μkATs‖As‖2)(Axk−b+AsATsrk‖As‖22)ATrrk‖Ar‖2‖Ar‖Ar‖2−μkAs‖As‖2‖22. |
Substituting λopt into (3.1), we obtain
Axk+1=(ATrb‖Ar‖2−μkATsb‖As‖2‖Ar‖Ar‖2−μkAs‖As‖2‖22)(Ar‖Ar‖2−μkAs‖As‖2)−((ATr‖Ar‖2−μkATs‖As‖2)(Axk+AsATsrk‖As‖22)‖Ar‖Ar‖2−μkAs‖As‖2‖22)(Ar‖Ar‖2−μkAs‖As‖2)+Axk+AsATsrk‖As‖22. |
So,
xk+1=(ATrb‖Ar‖2−μkATsb‖As‖2‖Ar‖Ar‖2−μkAs‖As‖2‖22)(er‖Ar‖2−μkes‖As‖2)−((ATr‖Ar‖2−μkATs‖As‖2)(A(xk+ATsrk‖As‖22es))‖Ar‖Ar‖2−μkAs‖As‖2‖22)(er‖Ar‖2−μkes‖As‖2)+xk+ATsrk‖As‖22es. |
Set r=jk1 and s=jk2 with the optimal parameter λopt, this process can be described as
yk=xk+ATjk2(b−Axk)‖Ajk2‖22ejk2andxk+1=yk+(βk−uTkAyk‖uk‖22)vk, | (3.2) |
where
μk=(Ajk1)T(Ajk2)‖Ajk1‖2‖Ajk2‖2and βk=ATjk1b‖Ajk1‖2−μkATjk2b‖Ajk2‖2, |
uk=Ajk1‖Ajk1‖2−μkAjk2‖Ajk2‖2and vk=ejk1‖Ajk1‖2−μkejk2‖Ajk2‖2. |
In order to simplify the computation, we rewrite the iterative process in Algorithm 2.
Algorithm 2 TRGS method |
Input: A∈Rm×n,b∈Rm andx0∈Rn |
Output: the approximation solution xk of (1.1) |
1. for k=0,1,… do until termination criterion is satisfied |
2. Set rk=b−Axk |
3. Select jk1 and jk2 with probability by (2.1) |
4. Compute |
μk=ATjk1Ajk2‖Ajk1‖2‖Ajk2‖2, rk1=ATjk1rk‖Ajk1‖2 and rk2=ATjk2rk‖Ajk2‖2 |
5. Updata |
xk+1=xk+rk1−μkrk2(1−|μk|2)‖Ajk1‖2ejk1+rk2−μkrk1(1−|μk|2)‖Ajk2‖2ejk2 |
6. end for |
The following Lemmas 3 and 4 will be used to prove the convergence of TRGS algorithm.
Lemma 3. For the column vector uk defined above, ‖uk‖22=1−μ2k.
Proof. By the definition of μk in (3.1) and uk in (3.2), one can obtain that
‖uk‖22=uTkuk=(ATjk1‖Ajk1‖2−μkATjk2‖Ajk2‖2)(Ajk1‖Ajk1‖2−μkAjk2‖Ajk2‖2)=1−2μ2k+μ2k=1−μ2k. |
Lemma 4. Define αs,t and βs,t as
αs,t=μ2k‖uk‖2=|ATsAt|2‖As‖22‖At‖22√1−|ATsAt|2‖As‖22‖At‖22andβs,t=μk‖uk‖2=ATsAt‖As‖2‖At‖2√1−|ATsAt|2‖As‖22‖At‖22, |
then, the existence of γ∈R subject to (|αs,t|−|βs,t|)2≥γ.
Proof. It is known that δ=mins≠t|ATsAt|‖As‖2‖At‖2 and Δ=maxs≠t|ATsAt|‖As‖2‖At‖2, Then,
(|αs,t|−|βs,t|)2=|ATsAt|2‖As‖22‖At‖22(|ATsAt|‖As‖2‖At‖2−1)2(1−|ATsAt|‖As‖2‖At‖2)(1+|ATsAt|‖As‖2‖At‖2)≥min{δ2(1−δ)1+δ,Δ2(1−Δ)1+Δ}=γ. |
When the coefficient matrix A in (1.1) is tall and narrow with full column rank, the following result gives the convergence of the TRGS algorithm.
Theorem 2. Assume that the tall and narrow coefficient matrix A∈Rm×n has full column rank. Let x∗=A†b be a solution of (1.1). Given an initial guess x0∈Rn, then the sequence {xk}∞0 generated by TRGS algorithm converges linearly to the x∗ in expectation. Moreover, it satisfies
Ek‖xk+1−x∗‖2ATA≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)−λmin(ATA)γτmin‖A‖2Fτmax]⋅‖xk−x∗‖2ATA |
and
E‖xk−x∗‖2ATA≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)−λmin(ATA)γτmin‖A‖2Fτmax]k⋅‖x0−x∗‖2ATA, |
where τmax=maxp∈[n]{‖A‖2F−‖Ap‖22}, τmin=minq∈[n]{‖A‖2F−‖Aq‖22} and γ=min{δ2(1−δ)1+δ,Δ2(1−Δ)1+Δ}.
Proof. See Appendix 1.
Remark 1. We remark that the upper bound of the convergence of RGS method from Lemma 1 is
ΨRGS=1−λmin(ATA)‖A‖2F. |
From Theorem 1, we remark the upper bound of the convergence of RGS2 method is
ΨRGS2=(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F). |
By Theorem 2, we can obtain an upper bound of the convergence of TRGS method
ΨTRGS=(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)−λmin(ATA)γτmin‖A‖2Fτmax. |
At each iteration, the TRGS method uses two columns of the matrix, while RGS utilizes only one. To be fair, we compare the upper bound on the convergence factor of one iteration of the TRGS method with that of two iterations, instead of one iteration, of the RGS method. Note that 0≤γ<1 and τmax≤‖A‖2F, we have 0≤γτmin/τmax<1 then ΨTRGS≤ΨRGS2≤Ψ2RGS<ΨRGS. Especially, when the column correlation coefficient δ or Δ=0, then γ=0. The RGS2 and TRGS methods have the same convergence factor in the upper bound.
Remark 2. If ‖Aj‖22,j=1,⋯,n, are precomputed, we discuss the computational cost of RGS, RGS2 and TRGS in each iteration step. The RGS costs 2m+2n+1 flopping operations (flops), the RGS2 method needs 4m+4n+2 flops, while the TRGS method requires 6m+4n+11 flops.
In this section, we give several examples to show the efficiency of our TRGS method. We compare TRGS with RGS2 and RGS. In addition, randomized Kaczmarz (RK) in [18], randomized extended Kaczmarz (REK) in [2], partially randomized extended Kaczmarz (PREK) in [30], generalized two-subspace randomized Kaczmarz (GTRK) and two-subspace randomized extended Kaczmarz (TREK) in [29], as other iterative methods, are considered for solving consistent or inconsistent linear systems. All experiments are carried out with the Matlab 2020b on a computer with 3.00 GHz processing unit and 16 GB RAM.
We measure the efficiency of TRGS and other methods by the relative solution error
RSE:=‖xk−x∗‖22‖x∗‖22. |
The initial vector is set as x0=(0,0,⋯,0)T for all methods. When the set maximum number of iterations kmax=106 or RSE<10−6, we terminate the iteration process of each method. The '-' means that the number of iteration steps of the algorithm reaches kmax.
Example 4.1. In this example, for conherent matrix A∈R2000×100 in least-squares problem (1.1). The entries of the A are the independent identically distributed uniform random variables in the interval (t,1), where the t∈[0,1]. We remark the average column correlation index as follows
ˉμk=2n2+nn∑q=1;q>p|ATpAq|‖Ap‖2‖Aq‖2,p,q∈{1,2,⋯,n}. |
When t increases from 0 to 1, the change of the ˉμk with t and the relationship between ΨRGS, Ψ2RGS, ΨRGS2 and ΨTRGS versus t are plotted in Figure 1.
As can be seen from Figure 1, the left subfigure shows that as t approaches 1, the average column correlation index ˉμk is highly correlated. From the right subfigure, the convergence upper bound of RGS2 and TRGS are always lower than RGS, and further we find that the theoretical upper bound of convergence of the TRGS will not exceed the RGS2.
Example 4.2 We use the RGS2 and TRGS methods to test the consistent least squares problem (1.1) with different size and compare them with the RK, RGS and the GTRK methods. The size of A is m×n with m=103∗k (k=1,2,⋯,5) and n=50. The entries of the matrix A are the independent identically distributed uniform random variables in the interval (t,1). The vector b=Ax∗, where x∗ is generated randomly with the MATLAB function randn.
In Tables 1–3, we list the IT and the CPU(s) for the RK, RGS, RGS2, GTRK, and TRGS methods with t=0.1,0.5 and 0.8, and the Euclidean condition number Cond(A) of the matrix is reported in each table. Figure 2 shows the plots of m versus IT (left), and m versus CPU (right) of Algorithm 2 applied to solve (1.1) with different coefficient matrix A listed in Tables 1–3, respectively.
name | t | m×n | 1000×50 | 2000×50 | 3000×50 | 4000×50 | 5000×50 |
0.1 | Cond(A) | 1A90 | 17.69 | 16.83 | 16.94 | 16.50 | |
RK | IT | 2742 | 2532 | 2461 | 2274 | 2282 | |
CPU(s) | 6.233e-02 | A149e-02 | 1.041e-01 | 1.650e-01 | 1.938e-01 | ||
RGS | IT | 2765 | 2252 | 2538 | 2259 | 2399 | |
CPU(s) | 6.663e-02 | 7.995e-02 | 1.170e-01 | 1.297e-01 | 1.574e-01 | ||
RGS2 | IT | 1390 | 1132 | 1083 | 1201 | 1087 | |
CPU(s) | 6.309e-02 | 7.624e-02 | 9.961e-02 | 1.351e-01 | 1.406e-01 | ||
GTRK | IT | 625 | 587 | 577 | 568 | 567 | |
CPU(s) | 2.931e-02 | 4.188e-02 | 5.464e-02 | A949e-02 | 1.063e-01 | ||
TRGS | IT | 483 | 539 | 533 | 486 | 466 | |
CPU(s) | 1.809e-02 | 2.678e-02 | 3.354e-02 | 3.921e-02 | 4.725e-02 |
name | |||||||
0.5 | Cond(A) | 47.91 | 43.13 | 41.39 | 41.19 | 40.64 | |
RK | IT | 13796 | 11153 | 11488 | 10563 | 10431 | |
CPU(s) | 3.017e-01 | 3.586e-01 | 5.073e-01 | 7.467e-01 | 9.254e-01 | ||
RGS | IT | 14074 | 11362 | 11162 | 10375 | 10714 | |
CPU(s) | 3.292e-01 | 3.956e-01 | 5.033e-01 | 5.988e-01 | 7.219e-01 | ||
RGS2 | IT | 6791 | 5733 | 5622 | 5474 | 5337 | |
CPU(s) | 3.050e-01 | 3.880e-01 | 5.036e-01 | 6.234e-01 | 7.082e-01 | ||
GTRK | IT | 726 | 611 | 703 | 713 | 687 | |
CPU(s) | 3.501e-02 | 4.567e-02 | 7.049e-02 | 1.151e-01 | 1.303e-01 | ||
TRGS | IT | 636 | 592 | 677 | 611 | 642 | |
CPU(s) | 2.53e-02 | 2.939e-02 | 4.328e-02 | 4.938e-02 | 6.810e-02 |
name | |||||||
0.8 | Cond(A) | 141.61 | 12A37 | 124.61 | 122.65 | 121.99 | |
RK | IT | 100928 | 96807 | 87951 | 90652 | 89026 | |
CPU(s) | 2.187e+00 | 3.089e+00 | 3.992e+00 | 6.870e+00 | 7.643e+00 | ||
RGS | IT | 116846 | 103915 | 89490 | 89978 | 87764 | |
CPU(s) | 2.690e+00 | 3.755e+00 | 4.087e+00 | 5.181e+00 | 5.902e+00 | ||
RGS2 | IT | 60650 | 50882 | 46398 | 44669 | 44784 | |
CPU(s) | 2.654e+00 | 3.428e+00 | 4.155e+00 | 5.091e+00 | 5.922e+00 | ||
GTRK | IT | 738 | 717 | 697 | 736 | 709 | |
CPU(s) | 3.463e-02 | 5.124e-02 | 6.969e-02 | 1.203e-01 | 1.356e-01 | ||
TRGS | IT | 696 | 665 | 658 | 658 | 683 | |
CPU(s) | 2.585e-02 | 3.253e-02 | 4.066e-02 | 5.334e-02 | A527e-02 |
From Tables 1–3, we can see that the TRGS method is better than other algorithms based on IT and CPU(s). We find that GTRK and TRGS are basically stable in both IT and CPU(s), while RK, RGS and RGS2 methods need more iterations and CPU time.
From Figure 2, it is not difficult to see that the curve of TRGS is much lower than that of RGS2 in terms of the IT and the CPU(s). In addition, RGS2 is sensitive to t, while TRGS is not affected by it. For example, in Figure 2, fix m=3000, when t=0.1,0.5 and 0.8, the IT of TRGS are basically steady at the level about 107, while the IT of RGS2 are steady at the level 108,1010 and 1012, respectively. Similar results also appear in the CPU(s).
Example 4.3 In this example, we apply the RGS2 and TRGS methods to solve the inconsistent least squares problem (1.1) and compare them with the REK, RGS, PREK and TREK methods. The entries of the matrix A are the independent identically distributed uniform random variables in the interval (t,1), and the vector b=Ax∗+r, where x∗ is one of the solutions of (1.1), which is generated randomly with the MATLAB function randn, and r is a nonzero vector in the null space of AT, which is generated by the MATLAB function null. The size of A is m×n with m=103∗k (k=1,2,⋯,5), and n=100.
Tables 4–6 list the iteration number (IT) and the CPU time when all methods stop and we also set t=0.1,0.5,0.8 in each case. Figure 3 shows the plots of m versus IT (left) and m versus CPU (right) of TRGS and RGS2 applied to solve all linear systems (1.1) in Tables 4–6.
name | t | m×n | 1000×100 | 2000×100 | 3000×100 | 4000×100 | 5000×100 |
0.1 | Cond(A) | 30.04 | 30.00 | 25.73 | 24.98 | 24.74 | |
REK | IT | 8246 | 7360 | 6383 | 6319 | 6135 | |
CPU(s) | 3.954e-01 | 5.279e-01 | 6.529e-01 | 1.080e-01 | 1.579e-01 | ||
RGS | IT | 6676 | 5821 | 5306 | 4710 | 4705 | |
CPU(s) | 2.334e-01 | 3.305e-01 | 4.174e-01 | 5.378e-01 | A937e-01 | ||
PREK | IT | - | - | - | 904195 | 857303 | |
CPU(s) | - | - | - | 1.385e+02 | 2.075e+02 | ||
RGS2 | IT | 3268 | 2914 | 2581 | 2383 | 2187 | |
CPU(s) | 2.224e-01 | 3.257e-01 | 3.954e-01 | 5.237e-01 | 7.610e-01 | ||
TREK | IT | 1641 | 1534 | 1439 | 1401 | 1486 | |
CPU(s) | 1.120e-01 | 1.410e-01 | 1.662e-01 | 2.582e-01 | 3.536e-01 | ||
TRGS | IT | 1402 | 1253 | 1201 | 1148 | 1118 | |
CPU(s) | 6.810e-02 | A920e-02 | 1.176e-01 | 1.738e-01 | 2.846e-01 |
name | |||||||
0.5 | Cond(A) | 74.36 | 66.12 | 63.36 | 60.93 | 60.49 | |
REK | IT | 45941 | 33128 | 31918 | 33464 | 29270 | |
CPU(s) | 2.207e+00 | 2.187e+00 | 2.993e+00 | 5.266e+00 | 7.963e+00 | ||
RGS | IT | 38943 | 24655 | 23342 | 24497 | 22516 | |
CPU(s) | 1.346e+00 | 1.332e+00 | 1.761e+00 | 2.605e+00 | 4.211e+00 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 18370 | 12110 | 11231 | 12188 | 11255 | |
CPU(s) | 1.230e+00 | 1.272e+00 | 1.668e+00 | 2.567e+00 | 3.854e+00 | ||
TREK | IT | 1896 | 1636 | 1714 | 1677 | 1617 | |
CPU(s) | 1.282e-01 | 1.501e-01 | 2.360e-01 | 2.885e-01 | 4.155e-01 | ||
TRGS | IT | 1607 | 1367 | 1381 | 1217 | 1396 | |
CPU(s) | 7.700e-02 | 1.022e-01 | 1.370e-01 | 1.647e-01 | 3.837e-01 |
name | |||||||
0.8 | Cond(A) | 223.66 | 197.26 | 18A00 | 183.47 | 17A91 | |
REK | IT | 405922 | 302370 | 277601 | 298566 | 268318 | |
CPU(s) | 1.818e+01 | 2.056e+01 | 2.700e+01 | 5.005e+01 | 7.072e+01 | ||
RGS | IT | 283262 | 226035 | 203500 | 208911 | 196200 | |
CPU(s) | 9.472e+00 | 1.211e+01 | 1.571e+01 | 2.304e+01 | 3.554e+01 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 144031 | 111424 | 103317 | 103665 | 100780 | |
CPU(s) | 9.268e+00 | 1.187e+01 | 1.565e+01 | 2.218e+01 | 3.378e+01 | ||
TREK | IT | 1888 | 1684 | 1613 | 1830 | 1804 | |
CPU(s) | 1.206e-01 | 1.477e-01 | 1.874e-01 | 3.337e-01 | 4.297e-01 | ||
TRGS | IT | 1725 | 1341 | 1207 | 1462 | 1340 | |
CPU(s) | A063e-02 | 9.848e-02 | 1.122e-01 | 2.184e-01 | 3.487e-01 |
From Tables 4–6, we can see that the TRGS method is better than other algorithms based on IT and CPU(s). We also find that TREK and TRGS are basically stable in both IT and CPU(s), while the REK, RGS, PREK and RGS2 need more iterations and CPU time.
From Figure 3, the curves of TRGS and RGS2 show that RGS2 needs more iterations and CPU time to reach the stopping criterion. In addition, RGS2 is sensitive to t, while TRGS is not. For example, in Figure 3, fix m=3000, when t=0.1,0.5 and 0.8, the IT of TRGS are basically steady at the level about 107, while the IT of RGS2 are steady at the level about 108,1010 and 1012 respectively. Similar results also appear in the CPU(s).
Example 4.4 In this example, we apply the TRGS method to solve the least squares problem (1.1) with the sparse coefficient matrix A taken from the Florida sparse matrix collection in [31]. Especially, we select the tall and narrow sparse matrix A with full column rank. Table 7 summarizes different sparse systems with density and condition number Cond(A), where the density of a matrix A means the ratio of the number of the nonzero elements of A to the total number of the elements of A. Algorithm 2 is compared with the RK, RGS, GTRK, REK, PREK, TREK and RGS2 methods.
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
m×n | 37932×331 | 50×9 | 55×46 | 85320×3240 | 315×100 | |
density | 1.09% | 50.00% | 53.04% | 0.09% | 23.87% | |
Cond(A) | 12.22 | 19.39 | 207.15 | 1.75 | 66.00 |
When the sparse least squares problem (1.1) is set to be consistent, the vector b=Ax∗, where x∗, one of the solutions of learst squares problem (1.1), is generated randomly with the MATLAB function randn. Table 8 lists the iteration number (IT) and the CPU time when RK, RGS, RGS2, GTRK and TRGS methods stop. Figure 4 shows the plot of RSE versus IT (left) and RSE versus CPU (right) of Algorithm 2 applied to solve (1.1) with the sparse coefficient matrix A named "divorce".
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
RK | IT | 150180 | 2244 | 320707 | 45633 | 37064 |
CPU(s) | 9.028e+01 | 3.642e-02 | 5.686e+00 | 6.894e+01 | 9.448e-01 | |
RGS | IT | 137190 | 2978 | 286699 | 35515 | 39585 |
CPU(s) | 2.913e+01 | 3.566e-02 | 3.779e+00 | 1.550e+01 | A348e-01 | |
RGS2 | IT | 76949 | 1340 | 145180 | 17935 | 20637 |
CPU(s) | 3.215e+01 | 2.958e-02 | 3.633e+00 | 1.477e+01 | A382e-01 | |
GTRK | IT | 78808 | 877 | 75523 | 22727 | 13788 |
CPU(s) | 7.695e+01 | 3.540e-02 | 3.232e+00 | 6.220e+01 | A000e-01 | |
TRGS | IT | 64956 | 139 | 56142 | 18351 | 11306 |
CPU(s) | 1.618e+01 | 4.470e-03 | 1.811e+00 | 9.819e+00 | 4.862e-01 |
When the sparse least squares problem (1.1) is set to be inconsistent, the b=Ax∗+r, where the r is a nonzero vector in the null space of AT. Due to the large dimension of A, the r cannot be generated by the Matlab function null, but it can be generated by the projection vector ˇr. The ˇr is generated by the MATLAB function randn and projected onto the null space of AT. That is to say, r=ˇr−AA†ˇr, where A†ˇr is obtained by the Matlab function lsqminnorm. Table 9 lists the IT and the CPU(s) when REK, RGS, PREK, RGS2, TREK and TRGS methods stop. Figure 5 shows the plot of RSE versus IT (left) and RSE versus CPU (right) of Algorithm 2 applied to solve (1.1) with the sparse coefficient matrix A named "divorce".
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
REK | IT | 194480 | 4450 | 403533 | 52384 | 58586 |
CPU(s) | 1.258e+02 | 1.636e-01 | 1.496e+01 | 9.001e+01 | 2.946e+00 | |
RGS | IT | 152939 | 3552 | 307945 | 35266 | 39632 |
CPU(s) | 3.198e+01 | 4.360e-02 | 4.202e+00 | 1.552e+01 | A494e-01 | |
PREK | IT | 166019 | 3890 | 347192 | 47165 | 65670 |
CPU(s) | 1.048e+02 | 9.710e-02 | 9.445e+00 | 7.933e+01 | 2.480e+00 | |
RGS2 | IT | 53481 | 1548 | 156941 | 18004 | 20030 |
CPU(s) | 2.210e+01 | 3.790e-02 | 4.086e+00 | 1.522e+01 | A309e-01 | |
TREK | IT | 88463 | 1075 | 103915 | 26109 | 20569 |
CPU(s) | A522e+01 | 9.452e-02 | 9.175e+00 | 7.641e+01 | 2.209e+00 | |
TRGS | IT | 77953 | 94 | 98361 | 18306 | 15423 |
CPU(s) | 1.957e+01 | 4.500e-03 | 3.260e+00 | 9.950e+00 | 6.633e-01 |
It can be seen from Table 8 that for sparse matrices listed in Table 7, TRGS achieves fast convergence with less IT and CPU(s) than that of RK, RGS, GTRK and RGS2 do. Similar results are shown in Table 9. Furthermore, from Figure 4, TRGS reaches the stop criteria with less iterations (left) and CPU time (right) than other methods for the "divorce" consistent sparse least squares problem (1.1). Similar results also appear in Figure 5.
Example 4.5 This example uses Algorithm 2 to restore a computer tomography (CT) image. We use the MATLAB function paralleltomo(N,θ,p) from Algebraic Iterative Reconstruction (ART) package in [31] to generate a large sparse matrix A and the exact solution x∗, where N=35, θ=0∘:1.5∘:178∘ and p=50, then the size of A is 5950×1225 and the condition number Cond(A)=352.32. We compute ˆb by ˆb=Ax∗ and b=ˆb+r, where the noise r is from the null space of the coefficient matrix AT, i.e., ATr=0. We set ordinary Gaussian white noise with noise levels δ=0.01 and the maximum iterative number is 5∗106. The TRGS is used to recover x∗ from b and compared with the REK, RGS, PREK, RGS2 and TREK methods.
Figure 6 shows the recovered images by REK, RGS, PREK, RGS2, TREK and TRGS together with the original image and noised image with δ=0.01. Figure 7 shows the convergence of RSE versus IT (left) and RSE versus CPU(s) (right) of TRGS compared with other methods when δ=0.01.
It is shown from Figure 6 that all methods obtain well restored image, and Figure 7 shows that TRGS converges much faster than REK, RGS, PREK, RGS2 and TREK do when δ=0.01.
Example 4.6 In this example, we use Algorithm 2 to solve the famous Phillips ill-posed problem in [32], which comes from the Fredholm integral equation of first kind
∫6−6K(s,t)ϕ(t)dt=f(s) |
on the square [−6,6]×[−6,6], where the kernel function is presented by K(s,t)=ϕ(s−t) with
ϕ(x)={1+cos(πx3),|x|<3,0,|x|≥3, |
and the right-hand side
f(s)=(6−|s|)(1+12cos(sπ3))+92πsin(|s|π3). |
The above problem is discretized to obtain the linear systems (1.1), where the coefficient matrix A∈Rn×n, exact solution vector x∗∈Rn and column vector b∈Rn are all generated by MATLAB function phillips(n) in [32]. As we know, if the system matrix A of (1.1) is ill-conditioned, one way to solve this problem is to use the Tikhonov regularization, that is
minx∈Rn{‖Ax−b‖22+λ‖x‖22}, with λ>0. |
We set n=1024, λ=0.01 and Gaussian white noise level denoted by δ=‖r‖2/||b||2=1% to obtain b=Ax∗+r. Here the condition number Cond(A)=4.1618e+10 and the rank is 1024, which means (1.1) is a several linear ill-posed problem. It is worth noting that the system matrix satisfies the conditions of Lemma 1, Theorems 1 and 2, so the above methods will converge. In Figure 8, the (a) displays the approximation solution derived by RGS, RGS2 and TRGS together with the exact solution for the Phillips test problem when δ=0.01. The (b) and (c) show the convergence of RSE versus IT and RSE versus CPU(s) of TRGS compared with the RGS and RGS2 methods when δ=0.01.
We can see from Figure 8 that all the recovered solution by RGS, RGS2 and TRGS are close to the exact solution in (a), (b) and (c) show that TRGS converges much faster than the other two methods.
A two-step randomized Guass-Seidel (TRGS) method for solving the linear least squares problems with tall and narrow coefficient matrix was presented. And the convergence analysis is provided when the coefficient matrix of (1.1) is of full column rank. This method does not need any columns pavings. Numerical examples for different cases show the superiority of the current method (TRGS) in this paper.
This work was supported by the Project of Department of Science and Technology of Sichuan Provincial (No. 2021ZYD0005), the Opening Project of Key Laboratory of Higher Education of Sichuan Province for Enterprise Informationalization and Internet of Things (No. 2020WZJ01), the Opening Project of Bridge Non-destruction Detecting and Engineering Computing Key Laboratory of Sichuan (Nos. 2021QYJ07, 2020QZJ03), the Scientific Research Project of Sichuan University of Science and Engineering (No. 2020RC24), and the Graduate student Innovation Fund (No. y2021101).
The authors declare no conflicts of interest regarding the publication of this paper.
Proof of Theorem 2. By the definition of TRGS, one can obtain that
xk+1=yk+(ATjk1(b−Ayk)‖Ajk1‖2−μkATjk2(b−Ayk)‖Ajk2‖2)ejk1‖Ajk1‖2−μkejk1‖Ajk2‖2‖uk‖22=xk+ATjk2(b−Axk)‖Ajk2‖2ejk2+(ATjk1(b−Ayk)‖Ajk1‖2−μkATjk2(b−Ayk)‖Ajk2‖2)⋅ejk1‖Ajk1‖2−μkejk1‖Ajk2‖2‖uk‖22. |
Then,
A(xk+1−xk)=ATjk2(b−Axk)‖Ajk2‖2Ajk2+(ATjk1(b−Ayk)‖Ajk1‖2−μkATjk2(b−Ayk)‖Ajk2‖2)⋅uk‖uk‖22. | (A1) |
The following will estimate ATjk1(b−Ayk)‖Ajk1‖2 and ATjk2(b−Ayk)‖Ajk2‖2,
(i)
ATjk1(b−Ayk)=ATjk1(b−A(xk+ATjk2(b−Axk)‖Ajk2‖2ejk2))=ATjk1(b−Axk)−ATjk1Ajk2ATjk2(b−Axk)‖Ajk2‖22=ATjk1(b−Axk)−μkATjk2(b−Axk)‖Ajk2‖2‖Ajk1‖2, |
then,
ATjk1(b−Ayk)‖Ajk1‖2=ATjk1(b−Axk)‖Ajk1‖2−μkATjk2(b−Axk)‖Ajk2‖2. | (A2) |
(ii)
ATjk2(b−Ayk)‖Ajk2‖2=ATjk2[b−A(xk+ATjk2(b−Axk)‖Ajk2‖2ejk2)]‖Ajk2‖2=ATjk2(b−Axk)−ATjk2Ajk2ATjk2(b−Axk)‖Ajk2‖22‖Ajk2‖2=0 | (A3) |
Substitute (A2) and (A3) into (A1), then,
A(xk+1−xk)=ATjk2(b−Axk)‖Ajk2‖2Ajk2+(ATjk1b‖Ajk1‖2−μkATjk2b‖Ajk2‖2−uTkAxk)uk‖uk‖22 |
Subtract Ax∗ from both sides of the equation, one can get
A(xk+1−x∗)=A(xk−x∗)+ATjk2(b−Axk)‖Ajk2‖2Ajk2+(ATjk1b‖Ajk1‖2−μkATjk2b‖Ajk2‖2−uTkAxk)uk‖uk‖22. |
Since
A(xk−x∗)+ATjk2(b−Axk)‖Ajk2‖2Ajk2=A(xk−x∗)+Ajk2ATjk2A(x∗−xk)‖Ajk2‖2=(Im−Ajk2ATjk2‖Ajk2‖22)A(xk−x∗) |
and
(ATjk1b‖Ajk1‖2−μkATjk2b‖Ajk2‖2−uTkAxk)uk‖uk‖22=[(ATjk1‖Ajk1‖2−μkATjk2‖Ajk2‖2)Ax∗−uTkAxk]uk‖uk‖22=ukuTkA(x∗−xk)‖uk‖22, |
then,
A(xk+1−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22−ukuTk‖uk‖22)A(xk−x∗) | (A4) |
Let ˆyk and ˆxk+1 be the first and second iterative solutions of xk obtained by single-step continuous execution of RGS2 algorithm, respectively. i.e.,
ˆyk=xk+ATjk1(b−Axk)‖Ajk1‖22ejk1and ˆxk+1=ˆyk+ATjk2(b−Aˆyk)‖Ajk2‖22ejk2. |
According to Theorem 1, one has
A(ˆxk+1−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22)(Im−Ajk1ATjk1‖Ajk1‖22)A(xk−x∗)=[Im−Ajk2ATjk2‖Ajk2‖22−(Im−Ajk2ATjk2‖Ajk2‖22)Ajk1ATjk1‖Ajk1‖22]A(xk−x∗)=[Im−Ajk2ATjk2‖Ajk2‖22−(Ajk1ATjk1‖Ajk1‖22−Ajk2ATjk2Ajk1ATjk1‖Ajk2‖22‖Ajk1‖22)]A(xk−x∗)=[Im−Ajk2ATjk2‖Ajk2‖22−(Ajk1ATjk1‖Ajk1‖22−μkAjk2ATjk1‖Ajk2‖2‖Ajk1‖2)]A(xk−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22−ukATjk1‖Ajk1‖2)A(xk−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22−ukuTk‖uk‖22+ukuTk‖uk‖22−ukATjk1‖Ajk1‖2)A(xk−x∗)=(Im−Ajk2ATjk2‖Ajk2‖22−ukuTk‖uk‖22)A(xk−x∗)+(ukuTk‖uk‖22−ukATjk1‖Ajk1‖2)A(xk−x∗)=A(xk+1−x∗)+(ukuTk‖uk‖22−ukATjk1‖Ajk1‖2)A(xk−x∗), |
the last equation is obtained from (A4), then
A(xk+1−x∗)=A(ˆxk+1−x∗)−(ukuTk‖uk‖22−ukATjk1‖Ajk1‖2)A(xk−x∗). |
Due to the orthogonality, one can get that,
‖A(xk+1−x∗)‖22=‖A(ˆxk+1−x∗)‖22−‖(ukuTk‖uk‖22−ukATjk1‖Ajk1‖2)A(xk−x∗)‖22=‖A(ˆxk+1−x∗)‖22−|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22. |
So,
Ek‖A(xk+1−x∗)‖22=Ek‖A(ˆxk+1−x∗)‖22−Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22. | (A5) |
Next, we estimate the two conditional expectations on the right side of (A5).
(I) The first expectation
According to Theorem 1,
Ek‖A(ˆxk+1−x∗)‖22≤(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)‖A(xk−x∗)‖22. | (A6) |
(II) The second expectation
By Lemma 3, ‖uk‖22−1=−μ2k. Then,
Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22=Ek|‖uk‖2ATjk1A(xk−x∗)‖Ajk1‖2−1‖uk‖2(ATjk1‖Ajk1‖2−μkATjk2‖Ajk2‖2)A(xk−x∗)|2=Ek|(‖uk‖2−1‖uk‖2)ATjk1A(xk−x∗)‖Ajk1‖2+μk‖uk‖2ATjk2A(xk−x∗)‖Ajk2‖2|2=Ek|μ2k‖uk‖2ATjk1A(xk−x∗)‖Ajk1‖2−μk‖uk‖2ATjk2A(xk−x∗)‖Ajk2‖2|2=n∑jk1=1‖Ajk1‖22‖A‖2Fn∑jk2=1jk2≠jk1‖Ajk2‖22‖A‖2F−‖Ajk1‖22⋅|μ2k‖uk‖2ATjk1A(xk−x∗)‖Ajk1‖2−μk‖uk‖2ATjk2A(xk−x∗)‖Ajk2‖2|2=n∑jk1=1n∑jk2=1jk2≠jk1‖Ajk1‖22‖Ajk2‖22‖A‖2Fτmax|μ2k‖uk‖2ATjk1A(xk−x∗)‖Ajk1‖2−μk‖uk‖2ATjk2A(xk−x∗)‖Ajk2‖2|2. |
By Lemma 4, it is established as follows
Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22≥n∑s=1n∑t=1t≠s‖As‖22‖At‖22‖A‖2Fτmax|αs,tATsA(xk−x∗)‖As‖2−βs,tATtA(xk−x∗)‖At‖2|2=1‖A‖2Fτmax∑s<t‖As‖22‖At‖22(|αs,tATsA(xk−x∗)‖As‖2−βs,tATtA(xk−x∗)‖At‖2|2+|αs,tATtA(xk−x∗)‖At‖2−βs,tATsA(xk−x∗)‖As‖2|2). |
For any α,β,θ,η∈R, we note the fact that
|αθ−βη|2+|αη−βθ|2≥(|α|−|β|)2(|θ|2+|η|2). |
Then,
Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22≥1‖A‖2Fτmax∑s<t‖As‖22‖At‖22(|αs,t|−|βs,t|)2(|ATtA(xk−x∗)‖At‖2|2+|ATsA(xk−x∗)‖As‖2|2)≥1‖A‖2Fτmax∑s<t(|αs,t|−|βs,t|)2(‖As‖22|ATtA(xk−x∗)|2+‖At‖22|ATsA(xk−x∗)|2). |
By Lemma 4, (|αs,t|−|βs,t|)2≥γ. Then
Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22≥γ‖A‖2Fτmax∑s<t(‖As‖22|ATtA(xk−x∗)|2+‖At‖22|ATsA(xk−x∗)|2)≥γ‖A‖2Fτmaxn∑s=1n∑t=1t≠s‖At‖22|ATsA(xk−x∗)|2=γ‖A‖2Fτmaxn∑s=1(‖A‖2F−‖As‖22)|ATsA(xk−x∗)|2)≥γτmin‖A‖2Fτmaxn∑s=1|ATsA(xk−x∗)|2≥λmin(ATA)γτmin‖A‖2Fτmax‖A(xk−x∗)‖2, |
i.e.,
Ek|ATjk1A(xk−x∗)‖Ajk1‖2−uTkA(xk−x∗)‖uk‖22|2‖uk‖22≥λmin(ATA)γτmin‖A‖2Fτmax‖A(xk−x∗)‖22 | (A7) |
Substitute (A6) and (A7) into (A5), one can obtain that
Ek‖A(xk+1−x∗)‖22≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)−λmin(ATA)γτmin‖A‖2Fτmax]‖A(xk−x∗)‖22. |
Thus,
Ek‖xk+1−x∗‖2ATA≤[(1−λmin(ATA)τmax)(1−λmin(ATA)‖A‖2F)−λmin(ATA)γτmin‖A‖2Fτmax]‖xk−x∗‖2ATA. |
[1] | A. Björck, Numerical Methods for Least Squares Problems, SIAM, Philadelphia, 1996. https://doi.org/10.1137/1.9781611971484 |
[2] |
A. Zouzias, N. M. Freris, Randomized extended Kaczmarz for solving least squares, SIAM J. Matrix Anal. Appl., 34 (2013), 773–793. https://doi.org/10.1137/120889897 doi: 10.1137/120889897
![]() |
[3] |
R. M. Gower, P. Richtarik, Randomized iterative methods for linear systems, SIAM J. Matrix Anal. Appl., 36 (2015), 1660-1690. https://doi.org/10.1137/15M1025487 doi: 10.1137/15M1025487
![]() |
[4] |
A. Ma, D. Needell, A. Ramdas, Convergence properties of the randomized extended Gauss-Seidel and Kaczmarz methods, SIAM J. Matrix Anal. Appl., 36 (2015), 1590–1604. https://doi.org/10.1137/15M1014425 doi: 10.1137/15M1014425
![]() |
[5] |
Z. Z. Bai, W. T. Wu, On greedy randomized Kaczmarz method for solving large sparse linear systems, SIAM J. Sci. Comput., 40 (2018), A592–A606. https://doi.org/10.1137/17M1137747 doi: 10.1137/17M1137747
![]() |
[6] |
A. Ma, D. Needell, A. Ramdas, Iterative methods for solving factorized linear systems, SIAM J. Matrix Anal. Appl., 39 (2018), 104–122. https://doi.org/10.1137/17M1115678 doi: 10.1137/17M1115678
![]() |
[7] |
C. Byrne, A unified treatment of some iterative algorithms in signal processing and image reconstruction, Inverse Prob., 20 (2004), 103–120. https://doi.org/10.1088/0266-5611/20/1/006 doi: 10.1088/0266-5611/20/1/006
![]() |
[8] |
C. A. Bouman, K. Sauer, A unified approach to statistical tomography using coordinate descent optimization, IEEE Trans. Image Process., 5 (1996), 480–492. https://doi.org/10.1109/83.491321 doi: 10.1109/83.491321
![]() |
[9] |
J. C. Ye, K. J. Webb, C. A. Bouman, R. P. Millane, Optical diffusion tomography by iterative-coordinate-descent optimization in a Bayesian framework, J. Opt. Soc. Am. A., 16 (1999), 2400–2412. https://doi.org/10.1364/JOSAA.16.002400 doi: 10.1364/JOSAA.16.002400
![]() |
[10] |
A. A. Canutescu, R. L. Dunbrack, Cyclic coordinate descent: A robotics algorithm for protein loop closure, Protein Sci., 12 (2003), 963–972. https://doi.org/10.1110/ps.0242703 doi: 10.1110/ps.0242703
![]() |
[11] |
K. W. Chang, C. J. Hsieh, C. J. Lin, Coordinate descent method for large-scale L2-loss linear support vector machines, J. Mach. Learn. Res., 9 (2008), 1369–1398. https://doi.org/10.1145/1390681.1442778 doi: 10.1145/1390681.1442778
![]() |
[12] |
P. Breheny, J. Huang, Coordinate descent algorithms for nonconvex penalized regression with applications to biological feature selection, Ann. Appl. Stat., 5 (2011), 232–253. https://doi.org/10.1214/10-AOAS388 doi: 10.1214/10-AOAS388
![]() |
[13] |
M. Elad, B. Matalon, M. Zibulevsky, Coordinate and subspace optimization methods for linear least squares with non-quadratic regularization, Appl. Comput. Harmonic Anal., 23 (2007), 346–367. https://doi.org/10.1016/j.acha.2007.02.002 doi: 10.1016/j.acha.2007.02.002
![]() |
[14] |
C. Wang, D. Wu, K. Yang, New decentralized positioning schemes for wireless sensor networks based on recursive least-squares optimization, IEEE Wirel. Commun. Lett., 3 (2014), 78–81. https://doi.org/10.1109/WCL.2013.111713.130734 doi: 10.1109/WCL.2013.111713.130734
![]() |
[15] |
J. A. Scott, M. Tuma, Sparse stretching for solving sparse-dense linear least-squares problems, SIAM J. Sci. Comput., 41 (2019), A1604–A1625. https://doi.org/10.1137/18M1181353 doi: 10.1137/18M1181353
![]() |
[16] | N. J. Higham, Accuracy and Stability of Numerical Algorithms, SIAM, Philadelphia, 2002. https://doi.org/10.2307/2669725 |
[17] | Y. Saad, Iterative Methods for Sparse Linear Systems, SIAM, Philadelphia, 2003. https://doi.org/10.1137/1.9780898718003.ch4 |
[18] |
T. Strohmer, R. Vershynin, A randomized Kaczmarz algorithm with exponential convergence, J. Fourier Anal. Appl., 15 (2009), 262–278. https://doi.org/10.1007/s00041-008-9030-4 doi: 10.1007/s00041-008-9030-4
![]() |
[19] | D. Leventhal, A. S. Lewis, Randomized methods for linear constraints: Convergence rates and conditioning, Math. Oper. Res., 35 (2010), 641–654. https://arXiv.org/abs/0806.3015 |
[20] |
Z. S. Lu, L. Xiao, On the complexity analysis of randomized block-coordinate descent methods, J. Math. Program, 152 (2015), 615–642. https://doi.org/10.1007/s10107-014-0800-2 doi: 10.1007/s10107-014-0800-2
![]() |
[21] |
Y. Liu, X. L. Jiang, C. Q. Gu, On maximum residual block and two-step Gauss-Seidel algorithms for linear least-squares problems, Calcolo, 58 (2021), 1–32. https://doi.org/10.1007/s10092-021-00404-x doi: 10.1007/s10092-021-00404-x
![]() |
[22] |
Z. Z. Bai, W. T. Wu, On greedy randomized coordinate descent methods for solving large linear least-squares problems, Numer. Linear Algebra Appl., 26 (2019), 22–37. https://doi.org/10.1002/nla.2237 doi: 10.1002/nla.2237
![]() |
[23] |
J. H. Zhang, J. H. Guo, On relaxed greedy randomized coordinate descent methods for solving large linear least-squares problems, J. Appl. Numer. Math., 157 (2020), 372–384. https://doi.org/10.1016/j.apnum.2020.06.014 doi: 10.1016/j.apnum.2020.06.014
![]() |
[24] |
Z. Z. Bai, L. Wang, W. T. Wu, On convergence rate of the randomized Gauss-Seidel method, Linear Algebra Appl., 611 (2021), 237–252. https://doi.org/10.1016/j.laa.2020.10.028 doi: 10.1016/j.laa.2020.10.028
![]() |
[25] |
J. Nutini, M. Schmidt, I. H. Laradji, M. Friedlander, H. Koepke, Coordinate descent converges faster with the Gauss-Southwell rule than random selection, Int. J. Technol. Manage., 43 (2015), 1632-1641. https://doi.org/10.1504/IJTM.200A019410 doi: 10.1504/IJTM.200A019410
![]() |
[26] |
K. Du, Tight upper bounds for the convergence of the randomized extended Kaczmarz and Gauss-Seidel algorithms, Numer. Linear Algebra Appl., 26 (2019), 22–33. https://doi.org/10.1002/nla.2233 doi: 10.1002/nla.2233
![]() |
[27] | W. Wu, Paving the Randomized Gauss-Seidel Method, BSc Thesis, Scripps College, Claremont, California, 2017. https://scholarship.claremont.edu/scripps_theses/1074 |
[28] |
D. Needell, R. Ward, Two-subspace projection method for coherent overdetermined systems, J. J. Fourier Anal. Appl., 19 (2013), 256–269. https://doi.org/10.1007/s00041-012-9248-z doi: 10.1007/s00041-012-9248-z
![]() |
[29] |
W. T. Wu, On two-subspace randomized extended Kaczmarz method for solving large linear least-squares problems, Numer. Algor., 89 (2022), 1–31. https://doi.org/10.1007/s11075-021-01104-x doi: 10.1007/s11075-021-01104-x
![]() |
[30] |
Z. Z. Bai, W. T. Wu, On partially randomized extended Kaczmarz method for solving large sparse overdetermined inconsistent linear systems, Linear Algebra Appl., 578 (2019), 225–250. https://doi.org/10.1016/j.laa.2019.05.005 doi: 10.1016/j.laa.2019.05.005
![]() |
[31] |
T. A. Davis, Y. Hu, The university of florida sparse matrix collection, ACM Trans. Math. Software, 38 (2011), 1–25. https://doi.org/10.1145/2049662.2049663 doi: 10.1145/2049662.2049663
![]() |
[32] |
P. C. Hansen, Regularization Tools: A Matlab package for analysis and solution of discrete ill-posed problems, Numer. Algor., 06 (1994), 1–35. https://doi.org/10.1007/BF02149761 doi: 10.1007/BF02149761
![]() |
1. | Yimou Liao, Tianxiu Lu, 2022, On Greedy Kaczmarz Method with Uniform Sampling for Consistent Tensor Linear Systems Based on T-Product, 9781450397155, 158, 10.1145/3577117.3577127 | |
2. | Fan Sha, Jianbing Zhang, Randomized symmetric Gauss-Seidel method for solving linear least squares problems, 2024, 9, 2473-6988, 17453, 10.3934/math.2024848 | |
3. | Yimou Liao, Tianxiu Lu, On greedy randomized block Gauss–Seidel method with averaging for sparse linear least-squares problems, 2023, 60, 0008-0624, 10.1007/s10092-023-00549-x | |
4. | Barış Yaylak, Fuat Polat, Cuma Süleymanoğlu, The Dynamic Relationship Among Atrial Pacing, Premature Atrial Beats, Mode Switching, and Atrial High-Rate Episodes in Patients with Sick Sinus Syndrome, 2025, 30620392, 10.32596/jucvm.galenos.2025.2024-7-62 |
name | t | m×n | 1000×50 | 2000×50 | 3000×50 | 4000×50 | 5000×50 |
0.1 | Cond(A) | 1A90 | 17.69 | 16.83 | 16.94 | 16.50 | |
RK | IT | 2742 | 2532 | 2461 | 2274 | 2282 | |
CPU(s) | 6.233e-02 | A149e-02 | 1.041e-01 | 1.650e-01 | 1.938e-01 | ||
RGS | IT | 2765 | 2252 | 2538 | 2259 | 2399 | |
CPU(s) | 6.663e-02 | 7.995e-02 | 1.170e-01 | 1.297e-01 | 1.574e-01 | ||
RGS2 | IT | 1390 | 1132 | 1083 | 1201 | 1087 | |
CPU(s) | 6.309e-02 | 7.624e-02 | 9.961e-02 | 1.351e-01 | 1.406e-01 | ||
GTRK | IT | 625 | 587 | 577 | 568 | 567 | |
CPU(s) | 2.931e-02 | 4.188e-02 | 5.464e-02 | A949e-02 | 1.063e-01 | ||
TRGS | IT | 483 | 539 | 533 | 486 | 466 | |
CPU(s) | 1.809e-02 | 2.678e-02 | 3.354e-02 | 3.921e-02 | 4.725e-02 |
name | |||||||
0.5 | Cond(A) | 47.91 | 43.13 | 41.39 | 41.19 | 40.64 | |
RK | IT | 13796 | 11153 | 11488 | 10563 | 10431 | |
CPU(s) | 3.017e-01 | 3.586e-01 | 5.073e-01 | 7.467e-01 | 9.254e-01 | ||
RGS | IT | 14074 | 11362 | 11162 | 10375 | 10714 | |
CPU(s) | 3.292e-01 | 3.956e-01 | 5.033e-01 | 5.988e-01 | 7.219e-01 | ||
RGS2 | IT | 6791 | 5733 | 5622 | 5474 | 5337 | |
CPU(s) | 3.050e-01 | 3.880e-01 | 5.036e-01 | 6.234e-01 | 7.082e-01 | ||
GTRK | IT | 726 | 611 | 703 | 713 | 687 | |
CPU(s) | 3.501e-02 | 4.567e-02 | 7.049e-02 | 1.151e-01 | 1.303e-01 | ||
TRGS | IT | 636 | 592 | 677 | 611 | 642 | |
CPU(s) | 2.53e-02 | 2.939e-02 | 4.328e-02 | 4.938e-02 | 6.810e-02 |
name | |||||||
0.8 | Cond(A) | 141.61 | 12A37 | 124.61 | 122.65 | 121.99 | |
RK | IT | 100928 | 96807 | 87951 | 90652 | 89026 | |
CPU(s) | 2.187e+00 | 3.089e+00 | 3.992e+00 | 6.870e+00 | 7.643e+00 | ||
RGS | IT | 116846 | 103915 | 89490 | 89978 | 87764 | |
CPU(s) | 2.690e+00 | 3.755e+00 | 4.087e+00 | 5.181e+00 | 5.902e+00 | ||
RGS2 | IT | 60650 | 50882 | 46398 | 44669 | 44784 | |
CPU(s) | 2.654e+00 | 3.428e+00 | 4.155e+00 | 5.091e+00 | 5.922e+00 | ||
GTRK | IT | 738 | 717 | 697 | 736 | 709 | |
CPU(s) | 3.463e-02 | 5.124e-02 | 6.969e-02 | 1.203e-01 | 1.356e-01 | ||
TRGS | IT | 696 | 665 | 658 | 658 | 683 | |
CPU(s) | 2.585e-02 | 3.253e-02 | 4.066e-02 | 5.334e-02 | A527e-02 |
name | t | m×n | 1000×100 | 2000×100 | 3000×100 | 4000×100 | 5000×100 |
0.1 | Cond(A) | 30.04 | 30.00 | 25.73 | 24.98 | 24.74 | |
REK | IT | 8246 | 7360 | 6383 | 6319 | 6135 | |
CPU(s) | 3.954e-01 | 5.279e-01 | 6.529e-01 | 1.080e-01 | 1.579e-01 | ||
RGS | IT | 6676 | 5821 | 5306 | 4710 | 4705 | |
CPU(s) | 2.334e-01 | 3.305e-01 | 4.174e-01 | 5.378e-01 | A937e-01 | ||
PREK | IT | - | - | - | 904195 | 857303 | |
CPU(s) | - | - | - | 1.385e+02 | 2.075e+02 | ||
RGS2 | IT | 3268 | 2914 | 2581 | 2383 | 2187 | |
CPU(s) | 2.224e-01 | 3.257e-01 | 3.954e-01 | 5.237e-01 | 7.610e-01 | ||
TREK | IT | 1641 | 1534 | 1439 | 1401 | 1486 | |
CPU(s) | 1.120e-01 | 1.410e-01 | 1.662e-01 | 2.582e-01 | 3.536e-01 | ||
TRGS | IT | 1402 | 1253 | 1201 | 1148 | 1118 | |
CPU(s) | 6.810e-02 | A920e-02 | 1.176e-01 | 1.738e-01 | 2.846e-01 |
name | |||||||
0.5 | Cond(A) | 74.36 | 66.12 | 63.36 | 60.93 | 60.49 | |
REK | IT | 45941 | 33128 | 31918 | 33464 | 29270 | |
CPU(s) | 2.207e+00 | 2.187e+00 | 2.993e+00 | 5.266e+00 | 7.963e+00 | ||
RGS | IT | 38943 | 24655 | 23342 | 24497 | 22516 | |
CPU(s) | 1.346e+00 | 1.332e+00 | 1.761e+00 | 2.605e+00 | 4.211e+00 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 18370 | 12110 | 11231 | 12188 | 11255 | |
CPU(s) | 1.230e+00 | 1.272e+00 | 1.668e+00 | 2.567e+00 | 3.854e+00 | ||
TREK | IT | 1896 | 1636 | 1714 | 1677 | 1617 | |
CPU(s) | 1.282e-01 | 1.501e-01 | 2.360e-01 | 2.885e-01 | 4.155e-01 | ||
TRGS | IT | 1607 | 1367 | 1381 | 1217 | 1396 | |
CPU(s) | 7.700e-02 | 1.022e-01 | 1.370e-01 | 1.647e-01 | 3.837e-01 |
name | |||||||
0.8 | Cond(A) | 223.66 | 197.26 | 18A00 | 183.47 | 17A91 | |
REK | IT | 405922 | 302370 | 277601 | 298566 | 268318 | |
CPU(s) | 1.818e+01 | 2.056e+01 | 2.700e+01 | 5.005e+01 | 7.072e+01 | ||
RGS | IT | 283262 | 226035 | 203500 | 208911 | 196200 | |
CPU(s) | 9.472e+00 | 1.211e+01 | 1.571e+01 | 2.304e+01 | 3.554e+01 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 144031 | 111424 | 103317 | 103665 | 100780 | |
CPU(s) | 9.268e+00 | 1.187e+01 | 1.565e+01 | 2.218e+01 | 3.378e+01 | ||
TREK | IT | 1888 | 1684 | 1613 | 1830 | 1804 | |
CPU(s) | 1.206e-01 | 1.477e-01 | 1.874e-01 | 3.337e-01 | 4.297e-01 | ||
TRGS | IT | 1725 | 1341 | 1207 | 1462 | 1340 | |
CPU(s) | A063e-02 | 9.848e-02 | 1.122e-01 | 2.184e-01 | 3.487e-01 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
m×n | 37932×331 | 50×9 | 55×46 | 85320×3240 | 315×100 | |
density | 1.09% | 50.00% | 53.04% | 0.09% | 23.87% | |
Cond(A) | 12.22 | 19.39 | 207.15 | 1.75 | 66.00 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
RK | IT | 150180 | 2244 | 320707 | 45633 | 37064 |
CPU(s) | 9.028e+01 | 3.642e-02 | 5.686e+00 | 6.894e+01 | 9.448e-01 | |
RGS | IT | 137190 | 2978 | 286699 | 35515 | 39585 |
CPU(s) | 2.913e+01 | 3.566e-02 | 3.779e+00 | 1.550e+01 | A348e-01 | |
RGS2 | IT | 76949 | 1340 | 145180 | 17935 | 20637 |
CPU(s) | 3.215e+01 | 2.958e-02 | 3.633e+00 | 1.477e+01 | A382e-01 | |
GTRK | IT | 78808 | 877 | 75523 | 22727 | 13788 |
CPU(s) | 7.695e+01 | 3.540e-02 | 3.232e+00 | 6.220e+01 | A000e-01 | |
TRGS | IT | 64956 | 139 | 56142 | 18351 | 11306 |
CPU(s) | 1.618e+01 | 4.470e-03 | 1.811e+00 | 9.819e+00 | 4.862e-01 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
REK | IT | 194480 | 4450 | 403533 | 52384 | 58586 |
CPU(s) | 1.258e+02 | 1.636e-01 | 1.496e+01 | 9.001e+01 | 2.946e+00 | |
RGS | IT | 152939 | 3552 | 307945 | 35266 | 39632 |
CPU(s) | 3.198e+01 | 4.360e-02 | 4.202e+00 | 1.552e+01 | A494e-01 | |
PREK | IT | 166019 | 3890 | 347192 | 47165 | 65670 |
CPU(s) | 1.048e+02 | 9.710e-02 | 9.445e+00 | 7.933e+01 | 2.480e+00 | |
RGS2 | IT | 53481 | 1548 | 156941 | 18004 | 20030 |
CPU(s) | 2.210e+01 | 3.790e-02 | 4.086e+00 | 1.522e+01 | A309e-01 | |
TREK | IT | 88463 | 1075 | 103915 | 26109 | 20569 |
CPU(s) | A522e+01 | 9.452e-02 | 9.175e+00 | 7.641e+01 | 2.209e+00 | |
TRGS | IT | 77953 | 94 | 98361 | 18306 | 15423 |
CPU(s) | 1.957e+01 | 4.500e-03 | 3.260e+00 | 9.950e+00 | 6.633e-01 |
name | t | m×n | 1000×50 | 2000×50 | 3000×50 | 4000×50 | 5000×50 |
0.1 | Cond(A) | 1A90 | 17.69 | 16.83 | 16.94 | 16.50 | |
RK | IT | 2742 | 2532 | 2461 | 2274 | 2282 | |
CPU(s) | 6.233e-02 | A149e-02 | 1.041e-01 | 1.650e-01 | 1.938e-01 | ||
RGS | IT | 2765 | 2252 | 2538 | 2259 | 2399 | |
CPU(s) | 6.663e-02 | 7.995e-02 | 1.170e-01 | 1.297e-01 | 1.574e-01 | ||
RGS2 | IT | 1390 | 1132 | 1083 | 1201 | 1087 | |
CPU(s) | 6.309e-02 | 7.624e-02 | 9.961e-02 | 1.351e-01 | 1.406e-01 | ||
GTRK | IT | 625 | 587 | 577 | 568 | 567 | |
CPU(s) | 2.931e-02 | 4.188e-02 | 5.464e-02 | A949e-02 | 1.063e-01 | ||
TRGS | IT | 483 | 539 | 533 | 486 | 466 | |
CPU(s) | 1.809e-02 | 2.678e-02 | 3.354e-02 | 3.921e-02 | 4.725e-02 |
name | |||||||
0.5 | Cond(A) | 47.91 | 43.13 | 41.39 | 41.19 | 40.64 | |
RK | IT | 13796 | 11153 | 11488 | 10563 | 10431 | |
CPU(s) | 3.017e-01 | 3.586e-01 | 5.073e-01 | 7.467e-01 | 9.254e-01 | ||
RGS | IT | 14074 | 11362 | 11162 | 10375 | 10714 | |
CPU(s) | 3.292e-01 | 3.956e-01 | 5.033e-01 | 5.988e-01 | 7.219e-01 | ||
RGS2 | IT | 6791 | 5733 | 5622 | 5474 | 5337 | |
CPU(s) | 3.050e-01 | 3.880e-01 | 5.036e-01 | 6.234e-01 | 7.082e-01 | ||
GTRK | IT | 726 | 611 | 703 | 713 | 687 | |
CPU(s) | 3.501e-02 | 4.567e-02 | 7.049e-02 | 1.151e-01 | 1.303e-01 | ||
TRGS | IT | 636 | 592 | 677 | 611 | 642 | |
CPU(s) | 2.53e-02 | 2.939e-02 | 4.328e-02 | 4.938e-02 | 6.810e-02 |
name | |||||||
0.8 | Cond(A) | 141.61 | 12A37 | 124.61 | 122.65 | 121.99 | |
RK | IT | 100928 | 96807 | 87951 | 90652 | 89026 | |
CPU(s) | 2.187e+00 | 3.089e+00 | 3.992e+00 | 6.870e+00 | 7.643e+00 | ||
RGS | IT | 116846 | 103915 | 89490 | 89978 | 87764 | |
CPU(s) | 2.690e+00 | 3.755e+00 | 4.087e+00 | 5.181e+00 | 5.902e+00 | ||
RGS2 | IT | 60650 | 50882 | 46398 | 44669 | 44784 | |
CPU(s) | 2.654e+00 | 3.428e+00 | 4.155e+00 | 5.091e+00 | 5.922e+00 | ||
GTRK | IT | 738 | 717 | 697 | 736 | 709 | |
CPU(s) | 3.463e-02 | 5.124e-02 | 6.969e-02 | 1.203e-01 | 1.356e-01 | ||
TRGS | IT | 696 | 665 | 658 | 658 | 683 | |
CPU(s) | 2.585e-02 | 3.253e-02 | 4.066e-02 | 5.334e-02 | A527e-02 |
name | t | m×n | 1000×100 | 2000×100 | 3000×100 | 4000×100 | 5000×100 |
0.1 | Cond(A) | 30.04 | 30.00 | 25.73 | 24.98 | 24.74 | |
REK | IT | 8246 | 7360 | 6383 | 6319 | 6135 | |
CPU(s) | 3.954e-01 | 5.279e-01 | 6.529e-01 | 1.080e-01 | 1.579e-01 | ||
RGS | IT | 6676 | 5821 | 5306 | 4710 | 4705 | |
CPU(s) | 2.334e-01 | 3.305e-01 | 4.174e-01 | 5.378e-01 | A937e-01 | ||
PREK | IT | - | - | - | 904195 | 857303 | |
CPU(s) | - | - | - | 1.385e+02 | 2.075e+02 | ||
RGS2 | IT | 3268 | 2914 | 2581 | 2383 | 2187 | |
CPU(s) | 2.224e-01 | 3.257e-01 | 3.954e-01 | 5.237e-01 | 7.610e-01 | ||
TREK | IT | 1641 | 1534 | 1439 | 1401 | 1486 | |
CPU(s) | 1.120e-01 | 1.410e-01 | 1.662e-01 | 2.582e-01 | 3.536e-01 | ||
TRGS | IT | 1402 | 1253 | 1201 | 1148 | 1118 | |
CPU(s) | 6.810e-02 | A920e-02 | 1.176e-01 | 1.738e-01 | 2.846e-01 |
name | |||||||
0.5 | Cond(A) | 74.36 | 66.12 | 63.36 | 60.93 | 60.49 | |
REK | IT | 45941 | 33128 | 31918 | 33464 | 29270 | |
CPU(s) | 2.207e+00 | 2.187e+00 | 2.993e+00 | 5.266e+00 | 7.963e+00 | ||
RGS | IT | 38943 | 24655 | 23342 | 24497 | 22516 | |
CPU(s) | 1.346e+00 | 1.332e+00 | 1.761e+00 | 2.605e+00 | 4.211e+00 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 18370 | 12110 | 11231 | 12188 | 11255 | |
CPU(s) | 1.230e+00 | 1.272e+00 | 1.668e+00 | 2.567e+00 | 3.854e+00 | ||
TREK | IT | 1896 | 1636 | 1714 | 1677 | 1617 | |
CPU(s) | 1.282e-01 | 1.501e-01 | 2.360e-01 | 2.885e-01 | 4.155e-01 | ||
TRGS | IT | 1607 | 1367 | 1381 | 1217 | 1396 | |
CPU(s) | 7.700e-02 | 1.022e-01 | 1.370e-01 | 1.647e-01 | 3.837e-01 |
name | |||||||
0.8 | Cond(A) | 223.66 | 197.26 | 18A00 | 183.47 | 17A91 | |
REK | IT | 405922 | 302370 | 277601 | 298566 | 268318 | |
CPU(s) | 1.818e+01 | 2.056e+01 | 2.700e+01 | 5.005e+01 | 7.072e+01 | ||
RGS | IT | 283262 | 226035 | 203500 | 208911 | 196200 | |
CPU(s) | 9.472e+00 | 1.211e+01 | 1.571e+01 | 2.304e+01 | 3.554e+01 | ||
PREK | IT | - | - | - | - | - | |
CPU(s) | - | - | - | - | - | ||
RGS2 | IT | 144031 | 111424 | 103317 | 103665 | 100780 | |
CPU(s) | 9.268e+00 | 1.187e+01 | 1.565e+01 | 2.218e+01 | 3.378e+01 | ||
TREK | IT | 1888 | 1684 | 1613 | 1830 | 1804 | |
CPU(s) | 1.206e-01 | 1.477e-01 | 1.874e-01 | 3.337e-01 | 4.297e-01 | ||
TRGS | IT | 1725 | 1341 | 1207 | 1462 | 1340 | |
CPU(s) | A063e-02 | 9.848e-02 | 1.122e-01 | 2.184e-01 | 3.487e-01 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
m×n | 37932×331 | 50×9 | 55×46 | 85320×3240 | 315×100 | |
density | 1.09% | 50.00% | 53.04% | 0.09% | 23.87% | |
Cond(A) | 12.22 | 19.39 | 207.15 | 1.75 | 66.00 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
RK | IT | 150180 | 2244 | 320707 | 45633 | 37064 |
CPU(s) | 9.028e+01 | 3.642e-02 | 5.686e+00 | 6.894e+01 | 9.448e-01 | |
RGS | IT | 137190 | 2978 | 286699 | 35515 | 39585 |
CPU(s) | 2.913e+01 | 3.566e-02 | 3.779e+00 | 1.550e+01 | A348e-01 | |
RGS2 | IT | 76949 | 1340 | 145180 | 17935 | 20637 |
CPU(s) | 3.215e+01 | 2.958e-02 | 3.633e+00 | 1.477e+01 | A382e-01 | |
GTRK | IT | 78808 | 877 | 75523 | 22727 | 13788 |
CPU(s) | 7.695e+01 | 3.540e-02 | 3.232e+00 | 6.220e+01 | A000e-01 | |
TRGS | IT | 64956 | 139 | 56142 | 18351 | 11306 |
CPU(s) | 1.618e+01 | 4.470e-03 | 1.811e+00 | 9.819e+00 | 4.862e-01 |
name | abtaha2 | divorce | Cites | bibd-81-3T | WorldCities | |
REK | IT | 194480 | 4450 | 403533 | 52384 | 58586 |
CPU(s) | 1.258e+02 | 1.636e-01 | 1.496e+01 | 9.001e+01 | 2.946e+00 | |
RGS | IT | 152939 | 3552 | 307945 | 35266 | 39632 |
CPU(s) | 3.198e+01 | 4.360e-02 | 4.202e+00 | 1.552e+01 | A494e-01 | |
PREK | IT | 166019 | 3890 | 347192 | 47165 | 65670 |
CPU(s) | 1.048e+02 | 9.710e-02 | 9.445e+00 | 7.933e+01 | 2.480e+00 | |
RGS2 | IT | 53481 | 1548 | 156941 | 18004 | 20030 |
CPU(s) | 2.210e+01 | 3.790e-02 | 4.086e+00 | 1.522e+01 | A309e-01 | |
TREK | IT | 88463 | 1075 | 103915 | 26109 | 20569 |
CPU(s) | A522e+01 | 9.452e-02 | 9.175e+00 | 7.641e+01 | 2.209e+00 | |
TRGS | IT | 77953 | 94 | 98361 | 18306 | 15423 |
CPU(s) | 1.957e+01 | 4.500e-03 | 3.260e+00 | 9.950e+00 | 6.633e-01 |