Research article Special Issues

A two-step randomized Gauss-Seidel method for solving large-scale linear least squares problems

  • A two-step randomized Gauss-Seidel (TRGS) method is presented for large linear least squares problem with tall and narrow coefficient matrix. The TRGS method projects the approximate solution onto the solution space by given two random columns and is proved to be convergent when the coefficient matrix is of full rank. Several numerical examples show the effectiveness of the TRGS method among all methods compared.

    Citation: Yimou Liao, Tianxiu Lu, Feng Yin. A two-step randomized Gauss-Seidel method for solving large-scale linear least squares problems[J]. Electronic Research Archive, 2022, 30(2): 755-779. doi: 10.3934/era.2022040

    Related Papers:

    [1] Fang Wang, Weiguo Li, Wendi Bao, Li Liu . Greedy randomized and maximal weighted residual Kaczmarz methods with oblique projection. Electronic Research Archive, 2022, 30(4): 1158-1186. doi: 10.3934/era.2022062
    [2] Xuefei He, Kun Wang, Liwei Xu . Efficient finite difference methods for the nonlinear Helmholtz equation in Kerr medium. Electronic Research Archive, 2020, 28(4): 1503-1528. doi: 10.3934/era.2020079
    [3] Jia-Min Luo, Hou-Biao Li, Wei-Bo Wei . Block splitting preconditioner for time-space fractional diffusion equations. Electronic Research Archive, 2022, 30(3): 780-797. doi: 10.3934/era.2022041
    [4] Yijun Chen, Yaning Xie . A kernel-free boundary integral method for reaction-diffusion equations. Electronic Research Archive, 2025, 33(2): 556-581. doi: 10.3934/era.2025026
    [5] Wenya Shi, Xinpeng Yan, Zhan Huan . Faster free pseudoinverse greedy block Kaczmarz method for image recovery. Electronic Research Archive, 2024, 32(6): 3973-3988. doi: 10.3934/era.2024178
    [6] Yan Chang, Yukun Guo . Simultaneous recovery of an obstacle and its excitation sources from near-field scattering data. Electronic Research Archive, 2022, 30(4): 1296-1321. doi: 10.3934/era.2022068
    [7] Qian He, Wenxin Du, Feng Shi, Jiaping Yu . A fast method for solving time-dependent nonlinear convection diffusion problems. Electronic Research Archive, 2022, 30(6): 2165-2182. doi: 10.3934/era.2022109
    [8] Dongmei Yu, Yifei Yuan, Yiming Zhang . A preconditioned new modulus-based matrix splitting method for solving linear complementarity problem of $ H_+ $-matrices. Electronic Research Archive, 2023, 31(1): 123-146. doi: 10.3934/era.2023007
    [9] Jian-Ying Rong, Xu-Qing Liu . Hybrid principal component regression estimation in linear regression. Electronic Research Archive, 2024, 32(6): 3758-3776. doi: 10.3934/era.2024171
    [10] Hua Yang, Xuan Geng, Heng Xu, Yichun Shi . An improved least squares (LS) channel estimation method based on CNN for OFDM systems. Electronic Research Archive, 2023, 31(9): 5780-5792. doi: 10.3934/era.2023294
  • A two-step randomized Gauss-Seidel (TRGS) method is presented for large linear least squares problem with tall and narrow coefficient matrix. The TRGS method projects the approximate solution onto the solution space by given two random columns and is proved to be convergent when the coefficient matrix is of full rank. Several numerical examples show the effectiveness of the TRGS method among all methods compared.



    We consider the approximate solutions of a large linear least squares problem

    minxRnbAx22, (1.1)

    where ARm×n, m>n, is of full clolumn rank. A and bRm are known, xRn is unknown to be determined. Under the condition that the linear system is consistent or inconsistent, people are interested in finding the unique least squares solution x=Ab, where A is the Moore-Penrose pseudoinverse of the matrix A (A=(ATA)1AT, where AT denotes the transpose of the matrix A). See also references [1,2,3,4,5,6]. As we know, the least squares problem arises widely in many fields such as tomography[7,8,9], protein structure[10], machine learning[11], biological feature selection[12], and so on[13,14,15]. For solving (1.1), there are many direct methods, such as QR decomposition and singular value decomposition (SVD) [1,16]. However, for large-scale system matrices, these methods are too expensive because they consume a lot of memory space. So, some iterative methods are applied to solve large-scale linear least squares problems.

    As one of the famous iterative algorithms, the Gauss-Seidel method [17] selects a coordinate dk and a step αk=argminαRf(xk+αdk) in each iteration. When αk=ATjk(bAxk)/Ajk22 and dk=ejk are accurately given, it generates the following iterative process:

    xk+1=xk+ATjk(bAxk)Ajk22ejk,k=0,1,2,, (1.2)

    where jk=(kmodn)+1, ()T represents the transpose of a matrix or a vector, and ejk represents the coordinate column vector, whose jk-th entry is 1 and zero otherwise. Inspired by the randomized Kaczmarz (RK) linear convergence characteristics of Strohmer and Vershynin[18], Leventhal and Lewis[19] obtained a randomized Gauss-Seidel (RGS) method, which is also known as randomized coordinate descent (RCD) method to solve (1.1). The RGS selects the update column index jk according to the appropriate probability. Theoretical analysis shows that RGS converges linearly to the unique least squares solution x=Ab. Many variants of RGS are receiving extensive attention recently due to its good performance. For example, the versions of block[20,21], random greedy[22,23,24]. Other versions, see literatures [4,17,25,26] and reference therein.

    In 2018, Wu[27] obtained a randomized block Gauss-Seidel (RBGS) method, which can significantly improve the convergence speed. However, the good column pavings for the RBGS is difficult to find when the column norms of the coefficient matrix fluctuate in a large range. In this paper, we propose a two-step randomimzed Gauss-Seidel method (TRGS), which does not need any columns pavings. The convergence of the TRGS algorithm is proved for (1.1) with the coefficient matrix of full rank.

    This paper is organized as follows. In Section 2, some symbols and a lemma related to RGS are introduced. The convergence of RGS2 is analyzed. In Section 3, a convergent upper bound of TRGS is obtained theoretically. Several numerical experiments are reported to verify the feasibility of our proposed algorithms in Section 4. The conclusions of this paper are given in Section 5 and the proof of Theorem 2 is shown in appendix.

    We first introduce the notations and definitions as follows. For the matrix QRm×n, Qi represents the ith column of the Q, λmax(QTQ) and λmin(QTQ) represent the maximum and minimum positive eigenvalues of the QTQ, respectively, ()T represents the transpose of a matrix or a vector, the Qτk represents a submatrix of the Q (where τk is a set of column indexes), Q2 and QF represent the spectral norm and Frobenius norm of the Q, respectively, and R(Q) is the image space of matrix Q. For a vector pRn or pRm, pi is the ith component of p. For constant cR, [c] refers to the set consising of all positive integers not exceeding c. For the matrix in (1.1), assuming that every two columns are independent and identically distributed, the correlation coefficient parameters of the matrix A are defined as follows

    δ=minst|ATsAt|As2At2andΔ=maxst|ATsAt|As2At2,s,t{1,2,,n}.

    Then, we have 0δΔ<1.

    The randomized Gauss-Seidel method proposed by Leventhal and Lewis[19] consists of two parts. The RGS determines a column index jk according to the probability Pr(column=jk)=Ajk22A2F and then updates xk+1 by (1.2).

    The following results in [19] summarized an upper bound for the error of the solution in expectation on the convergence of RGS algorithm.

    Lemma 1. Assume the least squares problems (1.1) has tall and narrow coefficient matrix ARm×n with full rank. Let x=Ab be a solution of (1.1). Given an initial guess x0Rn, then the sequence {xk}0 generated by RGS algorithm converges linearly to the x in expectation. Moreover, it satisfies

    Exkx2ATA(1λmin(ATA)A2F)kx0x2ATA,k=1,2,.

    Algorithm 1 RGS2 method
    Input: ARm×n,bRm and x0=0Rn
    Output: the approximation solution xk of (1.1)
    1. for k=0,1, do until termination criterion is satisfied
    2.    Set rk=bAxk and sk=ATrk.
    3.    Select jk1 and jk2 with probability by (2.1).
    4.    Compute sjk1k=ATjk1rk.
    5.    Update
    yk=xk+sjk1kAjk122ejk1 and xk+1=yk+ATjk2(bAyk)Ajk222ejk2.
    6. end for

    Now, we propose a simplified two-step randomized Gauss-Seidel method (RGS2) to solve (1.1). Algorithm 1 lists the RGS2 method, which mainly consists of two stages. The first stage is to select two different working columns by

    Pr(column=jk1)=Ajk122A2F,Pr(column=jk2)=Ajk222A2FAjk122, (2.1)

    where jk1{1,2,,n}, jk2{1,2,,n}/{jk1}. The second stage is to use (1.2) twice to update xk.

    If the coefficient matrix A in (1.1) is tall and narrow with full column rank, the following result gives the convergence of the RGS2 algorithm.

    Theorem 1. Assume the least squares problem (1.1) has tall and full-rank coefficient matrix ARm×n. Let x=Ab be a solution of (1.1). Given an initial guess x0Rn, then the iterative sequence {xk}0 generated by RGS2 algorithm converges linearly to the x in expectation. Moreover, it satisfies

    EkA(ˆxk+1x)22[(1λmin(ATA)τmax)(1λmin(ATA)A2F)]A(xkx)22

    and

    Exkx2ATA[(1λmin(ATA)τmax)(1λmin(ATA)A2F)]kx0x2ATA, (2.2)

    where τmax=A2Fminp[n]Ap22.

    Proof. Let ˆyk and ˆxk+1 be the first and the second iterative solutions of xk obtained by single-step continuous execution of Algorithm 1, respectively. The update process is divided into two steps as follows

    ˆyk=xk+ATjk1(bAxk)Ajk122ejk1andˆxk+1=ˆyk+ATjk2(bAˆyk)Ajk222ejk2.

    Let Pjki=ImAjkiATjkiAjki22 satisfy P2jki=Pjki,PTjki=Pjki, where i=1,2. Then, Pjki is the projection matrix, and

    A(ˆxk+1x)=A(ˆykx)+Ajk2ATjk2(bAˆyk)Ajk222=(ImAjk2ATjk2Ajk222)A(ˆykx)=(ImAjk2ATjk2Ajk222)(ImAjk1ATjk1Ajk122)A(xkx)=Pjk2Pjk1A(xkx),

    where the second equality follows from the normal equation ATAx=ATb, whose jk2-th equation gives ATjk2b=ATjk2Ax. Taking the expectation on the equality above, and by the expectation conditional upon jk1 (we fix the choice of jk1 and averagever the random index jk2), one can get

    EkA(ˆxk+1x)22=EkPjk2Pjk1A(xkx)22=Ek[(A(xkx))TPjk1Pjk2Pjk1A(xkx)]=njk1=1Ajk122A2Fnjk2=1,jk2jk1Ajk222A2FAjk122(A(xkx))TPjk1Pjk2Pjk1A(xkx)=njk1=1Ajk122A2F(A(xkx))TPjk1(ImAATAjk1ATjk1A2FAjk122)Pjk1A(xkx)=njk1=1Ajk122A2F(A(xkx))TPjk1(ImAATA2FAjk122)Pjk1A(xkx),

    in which the last equation holds because

    Ajk1ATjk1A2FAjk122Pjk1=Ajk1ATjk1A2FAjk122(ImAjk1ATjk1Ajk122)=Ajk1ATjk1A2FAjk122Ajk1(ATjk1Ajk1)ATjk1(A2FAjk122)Ajk122=0.

    Note that

    ATu22λmin(ATA)u22 and τmax=maxp[n]{A2FAp22},

    then ImAATA2FAjk12221λmin(ATA)τmax and

    EkA(ˆxk+1x)22(1λmin(ATA)τmax)njk1=1Ajk122A2F(A(xkx))T(ImAjk1ATjk1A2F)A(xkx)(1λmin(ATA)τmax)(1λmin(ATA)A2F)A(xkx)22.

    Therefore,

    Ekˆxk+1x2ATA(1λmin(ATA)τmax)(1λmin(ATA)A2F)xkx2ATA. (2.3)

    (2.2) is obtained from the recurrence relation of (2.3) and the full expectation formula. This completed the proof.

    The minimum positive eigenvalue of λmin(ATA) will become very small if the matrix A has high correlation parameters[28]. A weak bound of the convergence in Theorem 1 will appear. Under this condition, the angle of the unit coordinate direction as the search direction for two consecutive stages may be too small, which is the main reason for the slow convergence of RGS. Inspired by the work of Needell[28] and Wu[29], we iteratively update the solution by continuously seeking two more extensive directions, that is, determine the column pairs (r,s) in advance. So a two-step randomized Gauss-Seidel algorithm is obtained. In a two-step randomized algorithm, which mainly consists of three steps as follows: First, we randomly select the two columns r,s[n] according to the probability criterion, then estimate the optimal parameter λopt for the first iteration with yk=xk+λoptATrrkAr22er, and finally perform the second iteration to update the iterative solution by xk+1=yk+ATs(bAyk)As22es.

    We consider the convergence of the TRGS algorithm. We first need the following result presented in [28].

    Lemma 2. For any ϵR, ϕ,ψRm, the minimizer of ϵϕ+ψ22 is ϵopt=<ϕ,ψ>ϕ22.

    Now, we note that

    Ayk=Axk+λkArATrrkAr22andAxk+1=Ayk+AsATs(bAyk)As22,

    then

    Axk+1=Axk+λkArATrrkAr22+AsATs(b(Axk+λkArATrrkAr22))As22=λk[(ArAr2μkAsAs2)ATrrkAr2]+Axk+AsATsrkAs22, (3.1)

    where μk=(Ar)T(As)Ar2As2. Obviously,

    Axk+1b22=λk[(ArAr2μkAsAs2)ATrrkAr2]+Axkb+AsATsrkAs2222.

    The selection of optimal parameter λopt aims to minimize Axk+1b22. By Lemma 2, one can get the optimal value of λk, that is,

    λopt=(ATrAr2μkATsAs2)(Axkb+AsATsrkAs22)ATrrkAr2ArAr2μkAsAs222.

    Substituting λopt into (3.1), we obtain

    Axk+1=(ATrbAr2μkATsbAs2ArAr2μkAsAs222)(ArAr2μkAsAs2)((ATrAr2μkATsAs2)(Axk+AsATsrkAs22)ArAr2μkAsAs222)(ArAr2μkAsAs2)+Axk+AsATsrkAs22.

    So,

    xk+1=(ATrbAr2μkATsbAs2ArAr2μkAsAs222)(erAr2μkesAs2)((ATrAr2μkATsAs2)(A(xk+ATsrkAs22es))ArAr2μkAsAs222)(erAr2μkesAs2)+xk+ATsrkAs22es.

    Set r=jk1 and s=jk2 with the optimal parameter λopt, this process can be described as

    yk=xk+ATjk2(bAxk)Ajk222ejk2andxk+1=yk+(βkuTkAykuk22)vk, (3.2)

    where

    μk=(Ajk1)T(Ajk2)Ajk12Ajk22and βk=ATjk1bAjk12μkATjk2bAjk22,
    uk=Ajk1Ajk12μkAjk2Ajk22and vk=ejk1Ajk12μkejk2Ajk22.

    In order to simplify the computation, we rewrite the iterative process in Algorithm 2.

    Algorithm 2 TRGS method
    Input: ARm×n,bRm andx0Rn
    Output: the approximation solution xk of (1.1)
    1. for k=0,1, do until termination criterion is satisfied
    2.    Set rk=bAxk
    3.    Select jk1 and jk2 with probability by (2.1)
    4.    Compute
    μk=ATjk1Ajk2Ajk12Ajk22, rk1=ATjk1rkAjk12 and rk2=ATjk2rkAjk22
    5.    Updata
    xk+1=xk+rk1μkrk2(1|μk|2)Ajk12ejk1+rk2μkrk1(1|μk|2)Ajk22ejk2
    6. end for

    The following Lemmas 3 and 4 will be used to prove the convergence of TRGS algorithm.

    Lemma 3. For the column vector uk defined above, uk22=1μ2k.

    Proof. By the definition of μk in (3.1) and uk in (3.2), one can obtain that

    uk22=uTkuk=(ATjk1Ajk12μkATjk2Ajk22)(Ajk1Ajk12μkAjk2Ajk22)=12μ2k+μ2k=1μ2k.

    Lemma 4. Define αs,t and βs,t as

    αs,t=μ2kuk2=|ATsAt|2As22At221|ATsAt|2As22At22andβs,t=μkuk2=ATsAtAs2At21|ATsAt|2As22At22,

    then, the existence of γR subject to (|αs,t||βs,t|)2γ.

    Proof. It is known that δ=minst|ATsAt|As2At2 and Δ=maxst|ATsAt|As2At2, Then,

    (|αs,t||βs,t|)2=|ATsAt|2As22At22(|ATsAt|As2At21)2(1|ATsAt|As2At2)(1+|ATsAt|As2At2)min{δ2(1δ)1+δ,Δ2(1Δ)1+Δ}=γ.

    When the coefficient matrix A in (1.1) is tall and narrow with full column rank, the following result gives the convergence of the TRGS algorithm.

    Theorem 2. Assume that the tall and narrow coefficient matrix ARm×n has full column rank. Let x=Ab be a solution of (1.1). Given an initial guess x0Rn, then the sequence {xk}0 generated by TRGS algorithm converges linearly to the x in expectation. Moreover, it satisfies

    Ekxk+1x2ATA[(1λmin(ATA)τmax)(1λmin(ATA)A2F)λmin(ATA)γτminA2Fτmax]xkx2ATA

    and

    Exkx2ATA[(1λmin(ATA)τmax)(1λmin(ATA)A2F)λmin(ATA)γτminA2Fτmax]kx0x2ATA,

    where τmax=maxp[n]{A2FAp22}, τmin=minq[n]{A2FAq22} and γ=min{δ2(1δ)1+δ,Δ2(1Δ)1+Δ}.

    Proof. See Appendix 1.

    Remark 1. We remark that the upper bound of the convergence of RGS method from Lemma 1 is

    ΨRGS=1λmin(ATA)A2F.

    From Theorem 1, we remark the upper bound of the convergence of RGS2 method is

    ΨRGS2=(1λmin(ATA)τmax)(1λmin(ATA)A2F).

    By Theorem 2, we can obtain an upper bound of the convergence of TRGS method

    ΨTRGS=(1λmin(ATA)τmax)(1λmin(ATA)A2F)λmin(ATA)γτminA2Fτmax.

    At each iteration, the TRGS method uses two columns of the matrix, while RGS utilizes only one. To be fair, we compare the upper bound on the convergence factor of one iteration of the TRGS method with that of two iterations, instead of one iteration, of the RGS method. Note that 0γ<1 and τmaxA2F, we have 0γτmin/τmax<1 then ΨTRGSΨRGS2Ψ2RGS<ΨRGS. Especially, when the column correlation coefficient δ or Δ=0, then γ=0. The RGS2 and TRGS methods have the same convergence factor in the upper bound.

    Remark 2. If Aj22,j=1,,n, are precomputed, we discuss the computational cost of RGS, RGS2 and TRGS in each iteration step. The RGS costs 2m+2n+1 flopping operations (flops), the RGS2 method needs 4m+4n+2 flops, while the TRGS method requires 6m+4n+11 flops.

    In this section, we give several examples to show the efficiency of our TRGS method. We compare TRGS with RGS2 and RGS. In addition, randomized Kaczmarz (RK) in [18], randomized extended Kaczmarz (REK) in [2], partially randomized extended Kaczmarz (PREK) in [30], generalized two-subspace randomized Kaczmarz (GTRK) and two-subspace randomized extended Kaczmarz (TREK) in [29], as other iterative methods, are considered for solving consistent or inconsistent linear systems. All experiments are carried out with the Matlab 2020b on a computer with 3.00 GHz processing unit and 16 GB RAM.

    We measure the efficiency of TRGS and other methods by the relative solution error

    RSE:=xkx22x22.

    The initial vector is set as x0=(0,0,,0)T for all methods. When the set maximum number of iterations kmax=106 or RSE<106, we terminate the iteration process of each method. The '-' means that the number of iteration steps of the algorithm reaches kmax.

    Example 4.1. In this example, for conherent matrix AR2000×100 in least-squares problem (1.1). The entries of the A are the independent identically distributed uniform random variables in the interval (t,1), where the t[0,1]. We remark the average column correlation index as follows

    ˉμk=2n2+nnq=1;q>p|ATpAq|Ap2Aq2,p,q{1,2,,n}.

    When t increases from 0 to 1, the change of the ˉμk with t and the relationship between ΨRGS, Ψ2RGS, ΨRGS2 and ΨTRGS versus t are plotted in Figure 1.

    Figure 1.  Relationship between t and ˉμk (left). Comparison of ΨRGS, Ψ2RGS, ΨRGS2 and ΨTRGS with t (right).

    As can be seen from Figure 1, the left subfigure shows that as t approaches 1, the average column correlation index ˉμk is highly correlated. From the right subfigure, the convergence upper bound of RGS2 and TRGS are always lower than RGS, and further we find that the theoretical upper bound of convergence of the TRGS will not exceed the RGS2.

    Example 4.2 We use the RGS2 and TRGS methods to test the consistent least squares problem (1.1) with different size and compare them with the RK, RGS and the GTRK methods. The size of A is m×n with m=103k (k=1,2,,5) and n=50. The entries of the matrix A are the independent identically distributed uniform random variables in the interval (t,1). The vector b=Ax, where x is generated randomly with the MATLAB function randn.

    In Tables 13, we list the IT and the CPU(s) for the RK, RGS, RGS2, GTRK, and TRGS methods with t=0.1,0.5 and 0.8, and the Euclidean condition number Cond(A) of the matrix is reported in each table. Figure 2 shows the plots of m versus IT (left), and m versus CPU (right) of Algorithm 2 applied to solve (1.1) with different coefficient matrix A listed in Tables 13, respectively.

    Table 1.  IT, CPU of all methods for the consistent system with n=50 and different m.
    name t m×n 1000×50 2000×50 3000×50 4000×50 5000×50
    0.1 Cond(A) 1A90 17.69 16.83 16.94 16.50
    RK IT 2742 2532 2461 2274 2282
    CPU(s) 6.233e-02 A149e-02 1.041e-01 1.650e-01 1.938e-01
    RGS IT 2765 2252 2538 2259 2399
    CPU(s) 6.663e-02 7.995e-02 1.170e-01 1.297e-01 1.574e-01
    RGS2 IT 1390 1132 1083 1201 1087
    CPU(s) 6.309e-02 7.624e-02 9.961e-02 1.351e-01 1.406e-01
    GTRK IT 625 587 577 568 567
    CPU(s) 2.931e-02 4.188e-02 5.464e-02 A949e-02 1.063e-01
    TRGS IT 483 539 533 486 466
    CPU(s) 1.809e-02 2.678e-02 3.354e-02 3.921e-02 4.725e-02

     | Show Table
    DownLoad: CSV
    Table 2.  IT, CPU of all methods for the consistent system with n=50 and different m.
    name t m×n 1000×50 2000×50 3000×50 4000×50 5000×50
    0.5 Cond(A) 47.91 43.13 41.39 41.19 40.64
    RK IT 13796 11153 11488 10563 10431
    CPU(s) 3.017e-01 3.586e-01 5.073e-01 7.467e-01 9.254e-01
    RGS IT 14074 11362 11162 10375 10714
    CPU(s) 3.292e-01 3.956e-01 5.033e-01 5.988e-01 7.219e-01
    RGS2 IT 6791 5733 5622 5474 5337
    CPU(s) 3.050e-01 3.880e-01 5.036e-01 6.234e-01 7.082e-01
    GTRK IT 726 611 703 713 687
    CPU(s) 3.501e-02 4.567e-02 7.049e-02 1.151e-01 1.303e-01
    TRGS IT 636 592 677 611 642
    CPU(s) 2.53e-02 2.939e-02 4.328e-02 4.938e-02 6.810e-02

     | Show Table
    DownLoad: CSV
    Table 3.  IT, CPU of all methods for m-by-n consistent system with n=50 and different m.
    name t m×n 1000×50 2000×50 3000×50 4000×50 5000×50
    0.8 Cond(A) 141.61 12A37 124.61 122.65 121.99
    RK IT 100928 96807 87951 90652 89026
    CPU(s) 2.187e+00 3.089e+00 3.992e+00 6.870e+00 7.643e+00
    RGS IT 116846 103915 89490 89978 87764
    CPU(s) 2.690e+00 3.755e+00 4.087e+00 5.181e+00 5.902e+00
    RGS2 IT 60650 50882 46398 44669 44784
    CPU(s) 2.654e+00 3.428e+00 4.155e+00 5.091e+00 5.922e+00
    GTRK IT 738 717 697 736 709
    CPU(s) 3.463e-02 5.124e-02 6.969e-02 1.203e-01 1.356e-01
    TRGS IT 696 665 658 658 683
    CPU(s) 2.585e-02 3.253e-02 4.066e-02 5.334e-02 A527e-02

     | Show Table
    DownLoad: CSV
    Figure 2.  Pictures of log10(IT) (left) and log10(CPU) (right) versus for RGS2 and TRGS for consistent system when n=100. RGS2 for t=0.1:"- * -", RGS2 for t=0.5: "- o -", RGS2 for t=0.8:"", TRGS for t=0.1: "-", TRGS for t=0.5: "- -" and TRGS for t=0.8: "--".

    From Tables 13, we can see that the TRGS method is better than other algorithms based on IT and CPU(s). We find that GTRK and TRGS are basically stable in both IT and CPU(s), while RK, RGS and RGS2 methods need more iterations and CPU time.

    From Figure 2, it is not difficult to see that the curve of TRGS is much lower than that of RGS2 in terms of the IT and the CPU(s). In addition, RGS2 is sensitive to t, while TRGS is not affected by it. For example, in Figure 2, fix m=3000, when t=0.1,0.5 and 0.8, the IT of TRGS are basically steady at the level about 107, while the IT of RGS2 are steady at the level 108,1010 and 1012, respectively. Similar results also appear in the CPU(s).

    Example 4.3 In this example, we apply the RGS2 and TRGS methods to solve the inconsistent least squares problem (1.1) and compare them with the REK, RGS, PREK and TREK methods. The entries of the matrix A are the independent identically distributed uniform random variables in the interval (t,1), and the vector b=Ax+r, where x is one of the solutions of (1.1), which is generated randomly with the MATLAB function randn, and r is a nonzero vector in the null space of AT, which is generated by the MATLAB function null. The size of A is m×n with m=103k (k=1,2,,5), and n=100.

    Tables 46 list the iteration number (IT) and the CPU time when all methods stop and we also set t=0.1,0.5,0.8 in each case. Figure 3 shows the plots of m versus IT (left) and m versus CPU (right) of TRGS and RGS2 applied to solve all linear systems (1.1) in Tables 46.

    Table 4.  IT, CPU of all methods for the inconsistent system with n=100 and different m.
    name t m×n 1000×100 2000×100 3000×100 4000×100 5000×100
    0.1 Cond(A) 30.04 30.00 25.73 24.98 24.74
    REK IT 8246 7360 6383 6319 6135
    CPU(s) 3.954e-01 5.279e-01 6.529e-01 1.080e-01 1.579e-01
    RGS IT 6676 5821 5306 4710 4705
    CPU(s) 2.334e-01 3.305e-01 4.174e-01 5.378e-01 A937e-01
    PREK IT - - - 904195 857303
    CPU(s) - - - 1.385e+02 2.075e+02
    RGS2 IT 3268 2914 2581 2383 2187
    CPU(s) 2.224e-01 3.257e-01 3.954e-01 5.237e-01 7.610e-01
    TREK IT 1641 1534 1439 1401 1486
    CPU(s) 1.120e-01 1.410e-01 1.662e-01 2.582e-01 3.536e-01
    TRGS IT 1402 1253 1201 1148 1118
    CPU(s) 6.810e-02 A920e-02 1.176e-01 1.738e-01 2.846e-01

     | Show Table
    DownLoad: CSV
    Table 5.  IT, CPU of all methods for the inconsistent system with n=100 and different m.
    name t m×n 1000×100 2000×100 3000×100 4000×100 5000×100
    0.5 Cond(A) 74.36 66.12 63.36 60.93 60.49
    REK IT 45941 33128 31918 33464 29270
    CPU(s) 2.207e+00 2.187e+00 2.993e+00 5.266e+00 7.963e+00
    RGS IT 38943 24655 23342 24497 22516
    CPU(s) 1.346e+00 1.332e+00 1.761e+00 2.605e+00 4.211e+00
    PREK IT - - - - -
    CPU(s) - - - - -
    RGS2 IT 18370 12110 11231 12188 11255
    CPU(s) 1.230e+00 1.272e+00 1.668e+00 2.567e+00 3.854e+00
    TREK IT 1896 1636 1714 1677 1617
    CPU(s) 1.282e-01 1.501e-01 2.360e-01 2.885e-01 4.155e-01
    TRGS IT 1607 1367 1381 1217 1396
    CPU(s) 7.700e-02 1.022e-01 1.370e-01 1.647e-01 3.837e-01

     | Show Table
    DownLoad: CSV
    Table 6.  IT, CPU of all methods for the inconsistent system with n=100 and different m.
    name t m×n 1000×100 2000×100 3000×100 4000×100 5000×100
    0.8 Cond(A) 223.66 197.26 18A00 183.47 17A91
    REK IT 405922 302370 277601 298566 268318
    CPU(s) 1.818e+01 2.056e+01 2.700e+01 5.005e+01 7.072e+01
    RGS IT 283262 226035 203500 208911 196200
    CPU(s) 9.472e+00 1.211e+01 1.571e+01 2.304e+01 3.554e+01
    PREK IT - - - - -
    CPU(s) - - - - -
    RGS2 IT 144031 111424 103317 103665 100780
    CPU(s) 9.268e+00 1.187e+01 1.565e+01 2.218e+01 3.378e+01
    TREK IT 1888 1684 1613 1830 1804
    CPU(s) 1.206e-01 1.477e-01 1.874e-01 3.337e-01 4.297e-01
    TRGS IT 1725 1341 1207 1462 1340
    CPU(s) A063e-02 9.848e-02 1.122e-01 2.184e-01 3.487e-01

     | Show Table
    DownLoad: CSV
    Figure 3.  Pictures of log10(IT) (left) and log10(CPU) (right) versus for RGS2 and TRGS for inconsistent system when n=100. RGS2 for t=0.1:"- * -", RGS2 for t=0.5: "- o -", RGS2 for t=0.8:"", TRGS for t=0.1: "-", TRGS fort=0.5: "- -" and TRGS for t=0.8: "--".

    From Tables 46, we can see that the TRGS method is better than other algorithms based on IT and CPU(s). We also find that TREK and TRGS are basically stable in both IT and CPU(s), while the REK, RGS, PREK and RGS2 need more iterations and CPU time.

    From Figure 3, the curves of TRGS and RGS2 show that RGS2 needs more iterations and CPU time to reach the stopping criterion. In addition, RGS2 is sensitive to t, while TRGS is not. For example, in Figure 3, fix m=3000, when t=0.1,0.5 and 0.8, the IT of TRGS are basically steady at the level about 107, while the IT of RGS2 are steady at the level about 108,1010 and 1012 respectively. Similar results also appear in the CPU(s).

    Example 4.4 In this example, we apply the TRGS method to solve the least squares problem (1.1) with the sparse coefficient matrix A taken from the Florida sparse matrix collection in [31]. Especially, we select the tall and narrow sparse matrix A with full column rank. Table 7 summarizes different sparse systems with density and condition number Cond(A), where the density of a matrix A means the ratio of the number of the nonzero elements of A to the total number of the elements of A. Algorithm 2 is compared with the RK, RGS, GTRK, REK, PREK, TREK and RGS2 methods.

    Table 7.  The properties of different matrices from the Florida sparse matrix collection in [31].
    name abtaha2 divorce Cites bibd-81-3T WorldCities
    m×n 37932×331 50×9 55×46 85320×3240 315×100
    density 1.09% 50.00% 53.04% 0.09% 23.87%
    Cond(A) 12.22 19.39 207.15 1.75 66.00

     | Show Table
    DownLoad: CSV

    When the sparse least squares problem (1.1) is set to be consistent, the vector b=Ax, where x, one of the solutions of learst squares problem (1.1), is generated randomly with the MATLAB function randn. Table 8 lists the iteration number (IT) and the CPU time when RK, RGS, RGS2, GTRK and TRGS methods stop. Figure 4 shows the plot of RSE versus IT (left) and RSE versus CPU (right) of Algorithm 2 applied to solve (1.1) with the sparse coefficient matrix A named "divorce".

    Table 8.  IT, CPU of all methods for m-by-n consistent system.
    name abtaha2 divorce Cites bibd-81-3T WorldCities
    RK IT 150180 2244 320707 45633 37064
    CPU(s) 9.028e+01 3.642e-02 5.686e+00 6.894e+01 9.448e-01
    RGS IT 137190 2978 286699 35515 39585
    CPU(s) 2.913e+01 3.566e-02 3.779e+00 1.550e+01 A348e-01
    RGS2 IT 76949 1340 145180 17935 20637
    CPU(s) 3.215e+01 2.958e-02 3.633e+00 1.477e+01 A382e-01
    GTRK IT 78808 877 75523 22727 13788
    CPU(s) 7.695e+01 3.540e-02 3.232e+00 6.220e+01 A000e-01
    TRGS IT 64956 139 56142 18351 11306
    CPU(s) 1.618e+01 4.470e-03 1.811e+00 9.819e+00 4.862e-01

     | Show Table
    DownLoad: CSV
    Figure 4.  The RK, RGS, RGS2, GTRK and TRGS methods for solving linear consistent systems named divorce. Left: the relationship between IT and RSE; Right: the relationship between CPU (s) and RSE.

    When the sparse least squares problem (1.1) is set to be inconsistent, the b=Ax+r, where the r is a nonzero vector in the null space of AT. Due to the large dimension of A, the r cannot be generated by the Matlab function null, but it can be generated by the projection vector ˇr. The ˇr is generated by the MATLAB function randn and projected onto the null space of AT. That is to say, r=ˇrAAˇr, where Aˇr is obtained by the Matlab function lsqminnorm. Table 9 lists the IT and the CPU(s) when REK, RGS, PREK, RGS2, TREK and TRGS methods stop. Figure 5 shows the plot of RSE versus IT (left) and RSE versus CPU (right) of Algorithm 2 applied to solve (1.1) with the sparse coefficient matrix A named "divorce".

    Table 9.  IT, CPU of all methods for m-by-n inconsistent system.
    name abtaha2 divorce Cites bibd-81-3T WorldCities
    REK IT 194480 4450 403533 52384 58586
    CPU(s) 1.258e+02 1.636e-01 1.496e+01 9.001e+01 2.946e+00
    RGS IT 152939 3552 307945 35266 39632
    CPU(s) 3.198e+01 4.360e-02 4.202e+00 1.552e+01 A494e-01
    PREK IT 166019 3890 347192 47165 65670
    CPU(s) 1.048e+02 9.710e-02 9.445e+00 7.933e+01 2.480e+00
    RGS2 IT 53481 1548 156941 18004 20030
    CPU(s) 2.210e+01 3.790e-02 4.086e+00 1.522e+01 A309e-01
    TREK IT 88463 1075 103915 26109 20569
    CPU(s) A522e+01 9.452e-02 9.175e+00 7.641e+01 2.209e+00
    TRGS IT 77953 94 98361 18306 15423
    CPU(s) 1.957e+01 4.500e-03 3.260e+00 9.950e+00 6.633e-01

     | Show Table
    DownLoad: CSV
    Figure 5.  The REK, RGS, PREK, RGS2, TREK and TRGS methods for solving linear inconsistent systems named divorce. Left: the relationship between IT and RSE; Right: the relationship between CPU (s) and RSE.

    It can be seen from Table 8 that for sparse matrices listed in Table 7, TRGS achieves fast convergence with less IT and CPU(s) than that of RK, RGS, GTRK and RGS2 do. Similar results are shown in Table 9. Furthermore, from Figure 4, TRGS reaches the stop criteria with less iterations (left) and CPU time (right) than other methods for the "divorce" consistent sparse least squares problem (1.1). Similar results also appear in Figure 5.

    Example 4.5 This example uses Algorithm 2 to restore a computer tomography (CT) image. We use the MATLAB function paralleltomo(N,θ,p) from Algebraic Iterative Reconstruction (ART) package in [31] to generate a large sparse matrix A and the exact solution x, where N=35, θ=0:1.5:178 and p=50, then the size of A is 5950×1225 and the condition number Cond(A)=352.32. We compute ˆb by ˆb=Ax and b=ˆb+r, where the noise r is from the null space of the coefficient matrix AT, i.e., ATr=0. We set ordinary Gaussian white noise with noise levels δ=0.01 and the maximum iterative number is 5106. The TRGS is used to recover x from b and compared with the REK, RGS, PREK, RGS2 and TREK methods.

    Figure 6 shows the recovered images by REK, RGS, PREK, RGS2, TREK and TRGS together with the original image and noised image with δ=0.01. Figure 7 shows the convergence of RSE versus IT (left) and RSE versus CPU(s) (right) of TRGS compared with other methods when δ=0.01.

    Figure 6.  The original "phantom" image (a), the noised image (b), the recovered images by REK (c), RGS (d), PREK (e), RGS2 (f), TREK (g) and TRGS (h).
    Figure 7.  Convergence of RSE versus IT (a) and RSE versus CPU (b) of TRGS compared with that of RGS2 and TREK for restoring the noised "phantom" image with δ=0.01.

    It is shown from Figure 6 that all methods obtain well restored image, and Figure 7 shows that TRGS converges much faster than REK, RGS, PREK, RGS2 and TREK do when δ=0.01.

    Example 4.6 In this example, we use Algorithm 2 to solve the famous Phillips ill-posed problem in [32], which comes from the Fredholm integral equation of first kind

    66K(s,t)ϕ(t)dt=f(s)

    on the square [6,6]×[6,6], where the kernel function is presented by K(s,t)=ϕ(st) with

    ϕ(x)={1+cos(πx3),|x|<3,0,|x|3,

    and the right-hand side

    f(s)=(6|s|)(1+12cos(sπ3))+92πsin(|s|π3).

    The above problem is discretized to obtain the linear systems (1.1), where the coefficient matrix ARn×n, exact solution vector xRn and column vector bRn are all generated by MATLAB function phillips(n) in [32]. As we know, if the system matrix A of (1.1) is ill-conditioned, one way to solve this problem is to use the Tikhonov regularization, that is

    minxRn{Axb22+λx22}, with λ>0.

    We set n=1024, λ=0.01 and Gaussian white noise level denoted by δ=r2/||b||2=1% to obtain b=Ax+r. Here the condition number Cond(A)=4.1618e+10 and the rank is 1024, which means (1.1) is a several linear ill-posed problem. It is worth noting that the system matrix satisfies the conditions of Lemma 1, Theorems 1 and 2, so the above methods will converge. In Figure 8, the (a) displays the approximation solution derived by RGS, RGS2 and TRGS together with the exact solution for the Phillips test problem when δ=0.01. The (b) and (c) show the convergence of RSE versus IT and RSE versus CPU(s) of TRGS compared with the RGS and RGS2 methods when δ=0.01.

    Figure 8.  The performance of RGS, RGS2 and TRGS for the nosied Phillips test problem.

    We can see from Figure 8 that all the recovered solution by RGS, RGS2 and TRGS are close to the exact solution in (a), (b) and (c) show that TRGS converges much faster than the other two methods.

    A two-step randomized Guass-Seidel (TRGS) method for solving the linear least squares problems with tall and narrow coefficient matrix was presented. And the convergence analysis is provided when the coefficient matrix of (1.1) is of full column rank. This method does not need any columns pavings. Numerical examples for different cases show the superiority of the current method (TRGS) in this paper.

    This work was supported by the Project of Department of Science and Technology of Sichuan Provincial (No. 2021ZYD0005), the Opening Project of Key Laboratory of Higher Education of Sichuan Province for Enterprise Informationalization and Internet of Things (No. 2020WZJ01), the Opening Project of Bridge Non-destruction Detecting and Engineering Computing Key Laboratory of Sichuan (Nos. 2021QYJ07, 2020QZJ03), the Scientific Research Project of Sichuan University of Science and Engineering (No. 2020RC24), and the Graduate student Innovation Fund (No. y2021101).

    The authors declare no conflicts of interest regarding the publication of this paper.

    Proof of Theorem 2. By the definition of TRGS, one can obtain that

    xk+1=yk+(ATjk1(bAyk)Ajk12μkATjk2(bAyk)Ajk22)ejk1Ajk12μkejk1Ajk22uk22=xk+ATjk2(bAxk)Ajk22ejk2+(ATjk1(bAyk)Ajk12μkATjk2(bAyk)Ajk22)ejk1Ajk12μkejk1Ajk22uk22.

    Then,

    A(xk+1xk)=ATjk2(bAxk)Ajk22Ajk2+(ATjk1(bAyk)Ajk12μkATjk2(bAyk)Ajk22)ukuk22. (A1)

    The following will estimate ATjk1(bAyk)Ajk12 and ATjk2(bAyk)Ajk22,

    (i)

    ATjk1(bAyk)=ATjk1(bA(xk+ATjk2(bAxk)Ajk22ejk2))=ATjk1(bAxk)ATjk1Ajk2ATjk2(bAxk)Ajk222=ATjk1(bAxk)μkATjk2(bAxk)Ajk22Ajk12,

    then,

    ATjk1(bAyk)Ajk12=ATjk1(bAxk)Ajk12μkATjk2(bAxk)Ajk22. (A2)

    (ii)

    ATjk2(bAyk)Ajk22=ATjk2[bA(xk+ATjk2(bAxk)Ajk22ejk2)]Ajk22=ATjk2(bAxk)ATjk2Ajk2ATjk2(bAxk)Ajk222Ajk22=0 (A3)

    Substitute (A2) and (A3) into (A1), then,

    A(xk+1xk)=ATjk2(bAxk)Ajk22Ajk2+(ATjk1bAjk12μkATjk2bAjk22uTkAxk)ukuk22

    Subtract Ax from both sides of the equation, one can get

    A(xk+1x)=A(xkx)+ATjk2(bAxk)Ajk22Ajk2+(ATjk1bAjk12μkATjk2bAjk22uTkAxk)ukuk22.

    Since

    A(xkx)+ATjk2(bAxk)Ajk22Ajk2=A(xkx)+Ajk2ATjk2A(xxk)Ajk22=(ImAjk2ATjk2Ajk222)A(xkx)

    and

    (ATjk1bAjk12μkATjk2bAjk22uTkAxk)ukuk22=[(ATjk1Ajk12μkATjk2Ajk22)AxuTkAxk]ukuk22=ukuTkA(xxk)uk22,

    then,

    A(xk+1x)=(ImAjk2ATjk2Ajk222ukuTkuk22)A(xkx) (A4)

    Let ˆyk and ˆxk+1 be the first and second iterative solutions of xk obtained by single-step continuous execution of RGS2 algorithm, respectively. i.e.,

    ˆyk=xk+ATjk1(bAxk)Ajk122ejk1and ˆxk+1=ˆyk+ATjk2(bAˆyk)Ajk222ejk2.

    According to Theorem 1, one has

    A(ˆxk+1x)=(ImAjk2ATjk2Ajk222)(ImAjk1ATjk1Ajk122)A(xkx)=[ImAjk2ATjk2Ajk222(ImAjk2ATjk2Ajk222)Ajk1ATjk1Ajk122]A(xkx)=[ImAjk2ATjk2Ajk222(Ajk1ATjk1Ajk122Ajk2ATjk2Ajk1ATjk1Ajk222Ajk122)]A(xkx)=[ImAjk2ATjk2Ajk222(Ajk1ATjk1Ajk122μkAjk2ATjk1Ajk22Ajk12)]A(xkx)=(ImAjk2ATjk2Ajk222ukATjk1Ajk12)A(xkx)=(ImAjk2ATjk2Ajk222ukuTkuk22+ukuTkuk22ukATjk1Ajk12)A(xkx)=(ImAjk2ATjk2Ajk222ukuTkuk22)A(xkx)+(ukuTkuk22ukATjk1Ajk12)A(xkx)=A(xk+1x)+(ukuTkuk22ukATjk1Ajk12)A(xkx),

    the last equation is obtained from (A4), then

    A(xk+1x)=A(ˆxk+1x)(ukuTkuk22ukATjk1Ajk12)A(xkx).

    Due to the orthogonality, one can get that,

    A(xk+1x)22=A(ˆxk+1x)22(ukuTkuk22ukATjk1Ajk12)A(xkx)22=A(ˆxk+1x)22|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22.

    So,

    EkA(xk+1x)22=EkA(ˆxk+1x)22Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22. (A5)

    Next, we estimate the two conditional expectations on the right side of (A5).

    (I) The first expectation

    According to Theorem 1,

    EkA(ˆxk+1x)22(1λmin(ATA)τmax)(1λmin(ATA)A2F)A(xkx)22. (A6)

    (II) The second expectation

    By Lemma 3, uk221=μ2k. Then,

    Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22=Ek|uk2ATjk1A(xkx)Ajk121uk2(ATjk1Ajk12μkATjk2Ajk22)A(xkx)|2=Ek|(uk21uk2)ATjk1A(xkx)Ajk12+μkuk2ATjk2A(xkx)Ajk22|2=Ek|μ2kuk2ATjk1A(xkx)Ajk12μkuk2ATjk2A(xkx)Ajk22|2=njk1=1Ajk122A2Fnjk2=1jk2jk1Ajk222A2FAjk122|μ2kuk2ATjk1A(xkx)Ajk12μkuk2ATjk2A(xkx)Ajk22|2=njk1=1njk2=1jk2jk1Ajk122Ajk222A2Fτmax|μ2kuk2ATjk1A(xkx)Ajk12μkuk2ATjk2A(xkx)Ajk22|2.

    By Lemma 4, it is established as follows

    Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22ns=1nt=1tsAs22At22A2Fτmax|αs,tATsA(xkx)As2βs,tATtA(xkx)At2|2=1A2Fτmaxs<tAs22At22(|αs,tATsA(xkx)As2βs,tATtA(xkx)At2|2+|αs,tATtA(xkx)At2βs,tATsA(xkx)As2|2).

    For any α,β,θ,ηR, we note the fact that

    |αθβη|2+|αηβθ|2(|α||β|)2(|θ|2+|η|2).

    Then,

    Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk221A2Fτmaxs<tAs22At22(|αs,t||βs,t|)2(|ATtA(xkx)At2|2+|ATsA(xkx)As2|2)1A2Fτmaxs<t(|αs,t||βs,t|)2(As22|ATtA(xkx)|2+At22|ATsA(xkx)|2).

    By Lemma 4, (|αs,t||βs,t|)2γ. Then

    Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22γA2Fτmaxs<t(As22|ATtA(xkx)|2+At22|ATsA(xkx)|2)γA2Fτmaxns=1nt=1tsAt22|ATsA(xkx)|2=γA2Fτmaxns=1(A2FAs22)|ATsA(xkx)|2)γτminA2Fτmaxns=1|ATsA(xkx)|2λmin(ATA)γτminA2FτmaxA(xkx)2,

    i.e.,

    Ek|ATjk1A(xkx)Ajk12uTkA(xkx)uk22|2uk22λmin(ATA)γτminA2FτmaxA(xkx)22 (A7)

    Substitute (A6) and (A7) into (A5), one can obtain that

    EkA(xk+1x)22[(1λmin(ATA)τmax)(1λmin(ATA)A2F)λmin(ATA)γτminA2Fτmax]A(xkx)22.

    Thus,

    Ekxk+1x2ATA[(1λmin(ATA)τmax)(1λmin(ATA)A2F)λmin(ATA)γτminA2Fτmax]xkx2ATA.


    [1] A. Björck, Numerical Methods for Least Squares Problems, SIAM, Philadelphia, 1996. https://doi.org/10.1137/1.9781611971484
    [2] A. Zouzias, N. M. Freris, Randomized extended Kaczmarz for solving least squares, SIAM J. Matrix Anal. Appl., 34 (2013), 773–793. https://doi.org/10.1137/120889897 doi: 10.1137/120889897
    [3] R. M. Gower, P. Richtarik, Randomized iterative methods for linear systems, SIAM J. Matrix Anal. Appl., 36 (2015), 1660-1690. https://doi.org/10.1137/15M1025487 doi: 10.1137/15M1025487
    [4] A. Ma, D. Needell, A. Ramdas, Convergence properties of the randomized extended Gauss-Seidel and Kaczmarz methods, SIAM J. Matrix Anal. Appl., 36 (2015), 1590–1604. https://doi.org/10.1137/15M1014425 doi: 10.1137/15M1014425
    [5] Z. Z. Bai, W. T. Wu, On greedy randomized Kaczmarz method for solving large sparse linear systems, SIAM J. Sci. Comput., 40 (2018), A592–A606. https://doi.org/10.1137/17M1137747 doi: 10.1137/17M1137747
    [6] A. Ma, D. Needell, A. Ramdas, Iterative methods for solving factorized linear systems, SIAM J. Matrix Anal. Appl., 39 (2018), 104–122. https://doi.org/10.1137/17M1115678 doi: 10.1137/17M1115678
    [7] C. Byrne, A unified treatment of some iterative algorithms in signal processing and image reconstruction, Inverse Prob., 20 (2004), 103–120. https://doi.org/10.1088/0266-5611/20/1/006 doi: 10.1088/0266-5611/20/1/006
    [8] C. A. Bouman, K. Sauer, A unified approach to statistical tomography using coordinate descent optimization, IEEE Trans. Image Process., 5 (1996), 480–492. https://doi.org/10.1109/83.491321 doi: 10.1109/83.491321
    [9] J. C. Ye, K. J. Webb, C. A. Bouman, R. P. Millane, Optical diffusion tomography by iterative-coordinate-descent optimization in a Bayesian framework, J. Opt. Soc. Am. A., 16 (1999), 2400–2412. https://doi.org/10.1364/JOSAA.16.002400 doi: 10.1364/JOSAA.16.002400
    [10] A. A. Canutescu, R. L. Dunbrack, Cyclic coordinate descent: A robotics algorithm for protein loop closure, Protein Sci., 12 (2003), 963–972. https://doi.org/10.1110/ps.0242703 doi: 10.1110/ps.0242703
    [11] K. W. Chang, C. J. Hsieh, C. J. Lin, Coordinate descent method for large-scale L2-loss linear support vector machines, J. Mach. Learn. Res., 9 (2008), 1369–1398. https://doi.org/10.1145/1390681.1442778 doi: 10.1145/1390681.1442778
    [12] P. Breheny, J. Huang, Coordinate descent algorithms for nonconvex penalized regression with applications to biological feature selection, Ann. Appl. Stat., 5 (2011), 232–253. https://doi.org/10.1214/10-AOAS388 doi: 10.1214/10-AOAS388
    [13] M. Elad, B. Matalon, M. Zibulevsky, Coordinate and subspace optimization methods for linear least squares with non-quadratic regularization, Appl. Comput. Harmonic Anal., 23 (2007), 346–367. https://doi.org/10.1016/j.acha.2007.02.002 doi: 10.1016/j.acha.2007.02.002
    [14] C. Wang, D. Wu, K. Yang, New decentralized positioning schemes for wireless sensor networks based on recursive least-squares optimization, IEEE Wirel. Commun. Lett., 3 (2014), 78–81. https://doi.org/10.1109/WCL.2013.111713.130734 doi: 10.1109/WCL.2013.111713.130734
    [15] J. A. Scott, M. Tuma, Sparse stretching for solving sparse-dense linear least-squares problems, SIAM J. Sci. Comput., 41 (2019), A1604–A1625. https://doi.org/10.1137/18M1181353 doi: 10.1137/18M1181353
    [16] N. J. Higham, Accuracy and Stability of Numerical Algorithms, SIAM, Philadelphia, 2002. https://doi.org/10.2307/2669725
    [17] Y. Saad, Iterative Methods for Sparse Linear Systems, SIAM, Philadelphia, 2003. https://doi.org/10.1137/1.9780898718003.ch4
    [18] T. Strohmer, R. Vershynin, A randomized Kaczmarz algorithm with exponential convergence, J. Fourier Anal. Appl., 15 (2009), 262–278. https://doi.org/10.1007/s00041-008-9030-4 doi: 10.1007/s00041-008-9030-4
    [19] D. Leventhal, A. S. Lewis, Randomized methods for linear constraints: Convergence rates and conditioning, Math. Oper. Res., 35 (2010), 641–654. https://arXiv.org/abs/0806.3015
    [20] Z. S. Lu, L. Xiao, On the complexity analysis of randomized block-coordinate descent methods, J. Math. Program, 152 (2015), 615–642. https://doi.org/10.1007/s10107-014-0800-2 doi: 10.1007/s10107-014-0800-2
    [21] Y. Liu, X. L. Jiang, C. Q. Gu, On maximum residual block and two-step Gauss-Seidel algorithms for linear least-squares problems, Calcolo, 58 (2021), 1–32. https://doi.org/10.1007/s10092-021-00404-x doi: 10.1007/s10092-021-00404-x
    [22] Z. Z. Bai, W. T. Wu, On greedy randomized coordinate descent methods for solving large linear least-squares problems, Numer. Linear Algebra Appl., 26 (2019), 22–37. https://doi.org/10.1002/nla.2237 doi: 10.1002/nla.2237
    [23] J. H. Zhang, J. H. Guo, On relaxed greedy randomized coordinate descent methods for solving large linear least-squares problems, J. Appl. Numer. Math., 157 (2020), 372–384. https://doi.org/10.1016/j.apnum.2020.06.014 doi: 10.1016/j.apnum.2020.06.014
    [24] Z. Z. Bai, L. Wang, W. T. Wu, On convergence rate of the randomized Gauss-Seidel method, Linear Algebra Appl., 611 (2021), 237–252. https://doi.org/10.1016/j.laa.2020.10.028 doi: 10.1016/j.laa.2020.10.028
    [25] J. Nutini, M. Schmidt, I. H. Laradji, M. Friedlander, H. Koepke, Coordinate descent converges faster with the Gauss-Southwell rule than random selection, Int. J. Technol. Manage., 43 (2015), 1632-1641. https://doi.org/10.1504/IJTM.200A019410 doi: 10.1504/IJTM.200A019410
    [26] K. Du, Tight upper bounds for the convergence of the randomized extended Kaczmarz and Gauss-Seidel algorithms, Numer. Linear Algebra Appl., 26 (2019), 22–33. https://doi.org/10.1002/nla.2233 doi: 10.1002/nla.2233
    [27] W. Wu, Paving the Randomized Gauss-Seidel Method, BSc Thesis, Scripps College, Claremont, California, 2017. https://scholarship.claremont.edu/scripps_theses/1074
    [28] D. Needell, R. Ward, Two-subspace projection method for coherent overdetermined systems, J. J. Fourier Anal. Appl., 19 (2013), 256–269. https://doi.org/10.1007/s00041-012-9248-z doi: 10.1007/s00041-012-9248-z
    [29] W. T. Wu, On two-subspace randomized extended Kaczmarz method for solving large linear least-squares problems, Numer. Algor., 89 (2022), 1–31. https://doi.org/10.1007/s11075-021-01104-x doi: 10.1007/s11075-021-01104-x
    [30] Z. Z. Bai, W. T. Wu, On partially randomized extended Kaczmarz method for solving large sparse overdetermined inconsistent linear systems, Linear Algebra Appl., 578 (2019), 225–250. https://doi.org/10.1016/j.laa.2019.05.005 doi: 10.1016/j.laa.2019.05.005
    [31] T. A. Davis, Y. Hu, The university of florida sparse matrix collection, ACM Trans. Math. Software, 38 (2011), 1–25. https://doi.org/10.1145/2049662.2049663 doi: 10.1145/2049662.2049663
    [32] P. C. Hansen, Regularization Tools: A Matlab package for analysis and solution of discrete ill-posed problems, Numer. Algor., 06 (1994), 1–35. https://doi.org/10.1007/BF02149761 doi: 10.1007/BF02149761
  • This article has been cited by:

    1. Yimou Liao, Tianxiu Lu, 2022, On Greedy Kaczmarz Method with Uniform Sampling for Consistent Tensor Linear Systems Based on T-Product, 9781450397155, 158, 10.1145/3577117.3577127
    2. Fan Sha, Jianbing Zhang, Randomized symmetric Gauss-Seidel method for solving linear least squares problems, 2024, 9, 2473-6988, 17453, 10.3934/math.2024848
    3. Yimou Liao, Tianxiu Lu, On greedy randomized block Gauss–Seidel method with averaging for sparse linear least-squares problems, 2023, 60, 0008-0624, 10.1007/s10092-023-00549-x
    4. Barış Yaylak, Fuat Polat, Cuma Süleymanoğlu, The Dynamic Relationship Among Atrial Pacing, Premature Atrial Beats, Mode Switching, and Atrial High-Rate Episodes in Patients with Sick Sinus Syndrome, 2025, 30620392, 10.32596/jucvm.galenos.2025.2024-7-62
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2718) PDF downloads(118) Cited by(4)

Figures and Tables

Figures(8)  /  Tables(9)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog