Research article Special Issues

The ADMM algorithm for audio signal recovery and performance modification with the dual Douglas-Rachford dynamical system

  • Practitioners employ operator splitting methods—such as alternating direction method of multipliers (ADMM) and its "dual" Douglas-Rachford method (DR)—to solve many kinds of optimization problems. We provide a gentle introduction to these algorithms, and illustrations of their duality-like relationship in the context of solving basis pursuit problems for audio signal recovery. Recently, researchers have used the dynamical systems associated with the iterates of splitting methods to motivate the development of schemes to improve performance. These developments include a class of methods that act by iteratively minimizing surrogates for a Lyapunov function for the dynamical system. An exemplar of this class is currently state-of-the-art for the feasibility problem of finding wavelets with special structure. Early experimental evidence has also suggested that, when implemented in a primal-dual (ADMM and DR) framework, this exemplar may provide improved performance for basis pursuit problems. We provide a reasonable way to compute the updates for this exemplar, and we study the application of this method to the aforementioned basis pursuit audio problems. We provide experimental results and visualizations of the dynamical system for the dual DR sequence. We observe that for highly structured problems with real data, the algorithmic behavior is noticeably different than for randomly generated problems.

    Citation: Andrew Calcan, Scott B. Lindstrom. The ADMM algorithm for audio signal recovery and performance modification with the dual Douglas-Rachford dynamical system[J]. AIMS Mathematics, 2024, 9(6): 14640-14657. doi: 10.3934/math.2024712

    Related Papers:

    [1] Bing Xue, Jiakang Du, Hongchun Sun, Yiju Wang . A linearly convergent proximal ADMM with new iterative format for BPDN in compressed sensing problem. AIMS Mathematics, 2022, 7(6): 10513-10533. doi: 10.3934/math.2022586
    [2] Regina S. Burachik, Bethany I. Caldwell, C. Yalçın Kaya . Douglas–Rachford algorithm for control- and state-constrained optimal control problems. AIMS Mathematics, 2024, 9(6): 13874-13893. doi: 10.3934/math.2024675
    [3] Liping Geng, Jinchuan Zhou, Zhongfeng Sun, Jingyong Tang . Compressive hard thresholding pursuit algorithm for sparse signal recovery. AIMS Mathematics, 2022, 7(9): 16811-16831. doi: 10.3934/math.2022923
    [4] Wenxue Sun, Huiyuan Zhao, Xiao Zhang, Yuchao Sun, Xiaoxin Liu, Xueling Lv, Di Fan . Zero-watermarking Algorithm for Audio and Video Matching Verification. AIMS Mathematics, 2022, 7(5): 8390-8407. doi: 10.3934/math.2022468
    [5] Min Wang, Jiao Teng, Lei Wang, Junmei Wu . Application of ADMM to robust model predictive control problems for the turbofan aero-engine with external disturbances. AIMS Mathematics, 2022, 7(6): 10759-10777. doi: 10.3934/math.2022601
    [6] Siting Yu, Jingjing Peng, Zengao Tang, Zhenyun Peng . Iterative methods to solve the constrained Sylvester equation. AIMS Mathematics, 2023, 8(9): 21531-21553. doi: 10.3934/math.20231097
    [7] Nipa Jun-on, Raweerote Suparatulatorn, Mohamed Gamal, Watcharaporn Cholamjiak . An inertial parallel algorithm for a finite family of G-nonexpansive mappings applied to signal recovery. AIMS Mathematics, 2022, 7(2): 1775-1790. doi: 10.3934/math.2022102
    [8] Aliyu Muhammed Awwal, Poom Kumam, Kanokwan Sitthithakerngkiet, Abubakar Muhammad Bakoji, Abubakar S. Halilu, Ibrahim M. Sulaiman . Derivative-free method based on DFP updating formula for solving convex constrained nonlinear monotone equations and application. AIMS Mathematics, 2021, 6(8): 8792-8814. doi: 10.3934/math.2021510
    [9] Yue Zhao, Meixia Li, Xiaowei Pan, Jingjing Tan . Partial symmetric regularized alternating direction method of multipliers for non-convex split feasibility problems. AIMS Mathematics, 2025, 10(2): 3041-3061. doi: 10.3934/math.2025142
    [10] Jun Yang, Prasit Cholamjiak, Pongsakorn Sunthrayuth . Modified Tseng's splitting algorithms for the sum of two monotone operators in Banach spaces. AIMS Mathematics, 2021, 6(5): 4873-4900. doi: 10.3934/math.2021286
  • Practitioners employ operator splitting methods—such as alternating direction method of multipliers (ADMM) and its "dual" Douglas-Rachford method (DR)—to solve many kinds of optimization problems. We provide a gentle introduction to these algorithms, and illustrations of their duality-like relationship in the context of solving basis pursuit problems for audio signal recovery. Recently, researchers have used the dynamical systems associated with the iterates of splitting methods to motivate the development of schemes to improve performance. These developments include a class of methods that act by iteratively minimizing surrogates for a Lyapunov function for the dynamical system. An exemplar of this class is currently state-of-the-art for the feasibility problem of finding wavelets with special structure. Early experimental evidence has also suggested that, when implemented in a primal-dual (ADMM and DR) framework, this exemplar may provide improved performance for basis pursuit problems. We provide a reasonable way to compute the updates for this exemplar, and we study the application of this method to the aforementioned basis pursuit audio problems. We provide experimental results and visualizations of the dynamical system for the dual DR sequence. We observe that for highly structured problems with real data, the algorithmic behavior is noticeably different than for randomly generated problems.



    Throughout, H denotes a Hilbert space. Additionally, Γ0(H) denotes the set of proper lower semicontinuous convex functions from H to [,+]. For readers unfamiliar with these notions, it suffices to know that the functions f,g,f,g from the problems described herein all belong to the class Γ0(Rr).

    The alternating direction method of multipliers (ADMM) solves problems of the form

    minimize(x,z)Rn×Rmf(x)+g(z)subject toJx+Bzν=0,(p)

    where f and g are convex functions, νRr, JRr×n, and BRr×m. The ADMM algorithm deals with the two functions f and g separately, and so it is said to be an operator splitting method. Depending upon how we choose the functions f and g, we can model many different problems in this very flexible form. These include regression and classification problems, image and signal denoising, probabilistic optimization problems like progressive hedging, and basis pursuit [13]. We can use the latter to recover an approximate audio signal from a compressed set of sampling data.

    This paper is outlined as follows. In Section 2 we introduce the basis pursuit audio problem. In Section 3 we introduce ADMM, motivating with basis pursuit. In Section 4, we describe the Douglas-Rachford method (DR) and use the basis pursuit problem to illustrate DR's duality-like relationship with ADMM. In Section 5, we describe the exemplar Lyapunov surrogate method with which we will work and provide (Theorem 2), a clean way to compute it. The remainder of the paper contains our experimental results, which we summarize in Section 7. This paper is intended to be widely accessible, including to researchers who are unfamiliar with ADMM. As such, we adopt a somewhat tutorial approach to the topic.

    When digital audio is recorded, a microphone's diaphragm vibrates in response to sound waves, and N measurements are taken of the diaphragm's displacement. A typical rate at which these samples are collected is 44,100 per second (44.1 kHz). Recording over a time interval of length L, we denote by vi the measurement value at sample time:

    ti:=(2i12)(LN). (2.1)

    We can define a continuous function

    V(t):=x0N+2Nk=1N1xkcos(kπtL)such thatV(ti)=vii{1,,N}.

    The unique constants xk can be computed as follows. First, writing down the equality for the ith sample yields

    vi=V(ti)=x0N+2Nk=1N1xkcos(kπtiL)(just the definition)=x0N+2Nk=1N1xkcos(kπL(2i1)L2N)(substituting (2.1))=2N(x02+k=1N1xkcos(kπ(2i1)2N)).

    Since we have N such equalities (one for each sample (ti,vi)) in N unknowns (x0,,xN1), we can write them as the following linear system:

    2N[12cos(π2N)cos(2π2N)cos((N1)π2N)12cos(3π2N)cos(23π2N)cos((N1)3π2N)12cos(5π2N)cos(25π2N)cos((N1)5π2N)12cos((2N1)π2N)cos(2(2N1)π2N)cos((N1)(2N1)π2N)] = :M[x0x1x2xN1]=[v1v2v3vN]

    The matrix M is the inverse discrete cosine transform matrix, and it is orthogonal. Its inverse MT—the discrete cosine transform matrix, allows us to compute the x values as x=MTv. The nonzero values in x determine the shifts and dilates of the cosine function (i.e., the frequencies) that are present in the signal v. For this reason, x is said to be the spectrum of v.

    To reduce the size of audio files, we can delete some of the samples. The deletion of a sample (ti,vi) is reflected in the system by the removal of row i from M and v. For example, removing samples (t2,v2) and (t4,v4), the previous linear system becomes

    2N[12cos(π2N)cos(2π2N)cos(3π2N)cos(4π2N)12cos(5π2N)cos(25π2N)cos(35π2N)cos(45π2N)]=:M~[x0x1x2x3]=:x~=[v1v3v5]v~.

    Of course, the solution x~ is no longer unique. The set of solutions is a subspace of dimension 2, since 2 is the number of samples we have deleted and therefore the number of free variables we have gained. We can still use the Moore-Penrose pseudo-inverse M~T to obtain a spectrum x~=M~Tv~. The signal V~ that corresponds to this spectrum x~ still matches the samples we have not thrown away. Assuming we have not thrown away many samples, V~ may still provide a good approximation of V. In practice, of course, we want to reduce file size appreciably. When we remove 90% of samples (as is typical), the spectrum M~Tv~ no longer provides a good approximation signal. Note that while we removed entries 2 and 4 for illustrative purposes, there is nothing special about removing even-numbered entries; in our experiments, we remove 90% of entries randomly.

    We would expect that the sparsest solution x~ consists of the frequencies that actually matter, while avoiding the superfluous ones. Finding the sparsest solution to a linear system Ax=c is equivalent to the optimization problem:

    minimize xRn||x||1subject to Axc=0,withARr×n,cRr. (2.2)

    This is called the basis pursuit problem. For our signal recovery application, we solve this optimization problem with A=M~ and c=v~. Using a Matlab code (P-signal) to generate a signal c, we compare in Figure 1* the solution from (2.2) to a solution obtained using the Moore-Penrose pseudo-inverse. For this problem, c has 15 frequencies (i.e., Freq = 15) denoted by b1,,b15{1,2,,500} with corresponding amplitudes a1,,a15[0,1]. Here c represents 1/10th of a second of audio, and so is defined over domain d=[0,144100,244100,...,441044100].

    *The specific signal of Figure 1 can be generated by fixing the random generators with the additional code rng(3).

    Figure 1.  Portion of the reconstructed signals (left) and their most significant spectra (right).
    Matlabcode1.The code used for signal generation.d=(0:(1/44100):(0.1(1/44100)));a=rand( Freq, 1)b=randi(500, Freq, 1)c=0fori=1: Freq c=c+(a(i)sin(p ib(i)d)); end (Psignal)

    We can use ADMM to solve the basis pursuit problem. We first define the following set and associated function:

    LetS:={x|M~xc=0}and defineδS:={0,if xS,, otherwise.

    We can use these to rewrite (2.2) in the following form:

    minimise(x,z)δS(x)+z1 subject to xz=0.(BPp)

    The function δS is called the indicator of the set S. We introduce it so that we can take advantage of ADMM's ability to treat the functions f(x)=δS(x) and g(z)=z1 separately. ADMM solves the problem

    min(x,z)RnmaxλRmLρ(x,z,λ)=δS(x)+||z||1+(λk)T(xz)+ρ2||xz||2, (3.1)

    by finding a triple (x,z,λ). This triple is a saddle point for L0 in the sense that (x,z) minimizes (x,z)L0(x,z,λ) while λ maximizes λL0(x,z,λ). See, for example, [9, Section 5.5] and [10, Section 4.3]. Here ρ is a chosen constant. The multiplier variable λ generalizes the classical notion of a Lagrange multiplier for a general convex optimization problem, and is also called a dual variable. The primal variables x and z have the property that the optimal pair (x,z) is a solution to (BP-p). The ADMM update steps are

    xk+1:=argminx{Lρ(x,zk,λk)}, (3.2a)
    zk+1:=argminz{Lρ(xk+1,z,λk)}, (3.2b)
    λk+1:=λk+ρ(xk+1zk+1). (3.2c)

    The multiplier update (3.2c) maximizes a concave quadratic approximation to λLρ(xk+1,zk+1,λ). ADMM draws its name from this iterative process of minimizing with respect to the primal variables x and z, and then maximizing over the dual variable λ.

    Let us look at how these updates are computed for the basis pursuit problem. We begin with the x update:

    xk+1=argminx{δS(x)+zk1+(λk)T(xzk)+ρ2xzk2}=argminx{δS(x)+xzk+1ρ(λk)T2}( completing the square &  dropping terms without x)=argminxSx(zk+1ρλk)2=:PS(zkλkρ). (3.3)

    Here PS is the closest-point projection operator for the set S. As S is closed and convex in a Hilbert space, PS has nonempty, single-valued images [3]. Moreover, since S is an affine subspace defined by M~x=c, PS has a closed form [11]. We can use that closed form to write (3.3) explicitly:

    xk+1=(zkλkρ)+M~T(M~M~T)1(cM~(zkλkρ))=(IM~T(M~M~T)1M~)(zkλkρ)+M~T(M~M~T)1c.

    We compute the z update as follows:

    zk+1=argminz{δS(xk+1)+z1+(λk)T(xk+1z)+ρ2xk+1z2}=argminz{z1+xk+1+1ρ(λk)Tz2}( completing the square &  omitting terms without z)=:S1ρ(xk+1+1ρλk).

    The operator S1ρ(u) is known as the soft thresholding function. It is computable as follows:

    (i{1,2,,n})S1ρ(u)i:={ui1ρ, if ui>1ρ,0, if |ui|1ρ,ui+1ρ, if ui<1ρ.

    Both the projection operator PS and the shrinkage operator S1ρ are examples of proximity operators. The proximity operator for a function hΓ0(H) is defined as follows:

    proxh:HH:wargminuH(f(u)+12||wu||2).

    Straight from the definition, one sees that the projection operator PS is proxδS, while the soft thresholding operator S1ρ is prox1. Such operators are used to compute the steps for operator splitting methods.

    We will next look at another operator splitting method, one that has a special duality-like relationship with ADMM. Let I be the identity map and Rh denote the reflected proximal mapping:

    Rh:=2proxhI. (4.1)

    The Douglas-Rachford algorithm (DR) minimizes f+g where f,gΓ0(H) by generating a sequence (yn)n=1 as

    yn+1Tg,f(yn) where Tg,f:=12(I+RfRg). (4.2)

    In [2], Attouch showed that if f and g are convex and 0 belongs to the interior of {x:f(x)<}{x:g(x)<}, then f+g satisfies the necessary conditions for the classical DR convergence result of Lions and Mercier.

    Attouch's actual requirements are less restrictive than what we have presented.

    Theorem 1. ([2,26]) If f,gΓ0(H) and Attouch's criterion holds, then yk from (4.2) converges weakly to some yH such that proxgy minimizes f+g.

    For our context, we are interested in the setting where the Douglas-Rachford method is applied to the problem

    minimizeyf(ATy)+g(y)(d)

    where A is a matrix of appropriate dimension. Here h denotes the Fenchel-Moreau-Rockafellar conjugate of h [3,24], defined as

    h(y)=sup{yTxh(x)}.

    We are particularly interested in the case when A=I, f=δS, and g=1. The conjugates are

    δS(y)=supyS{yTx}and1(y)={0, if ||x||1,, otherwise.

    Note that Attouch's condition is satisfied for δS and 1, as long as the system Ax=b is consistent. The general problem (d) is said to be dual to the general problem (p), while the specific problem

    minimizeyδS(y)+y1(BPd)

    is dual to the primal basis pursuit problem (BP-p). For fully convex problems, DR and ADMM are closely related, with ADMM for (p) being equivalent to DR for (d) [19,20,24,25]. The linear relationship between DR and ADMM variables is outlined in Table 1.

    Table 1.  The relationship between ADMM and DR variables.
    Primal Dual
    ρAxk+1=proxρf(AT)(Rρgyk)Rρg(yk) Tρg,ρf(AT)yk1=yk=λk+ρzk
    ρzk=ykproxρgyk Rρgyk=λkρzk
    λk=proxρgyk Rρf(AT)Rρgyk=λkρzk+2ρAxk+1

     | Show Table
    DownLoad: CSV

    From Theorem 1 and the relationships in the table, it is clear that the multiplier sequence λk converges to the solution of (d). For the basis pursuit problem (BP-p) and its dual problem (BP-d), the relationship in Table 1 is illustrated in Figure 2. Because the DR operator Tg,f has particularly desirable properties, DR convergence results are often cited as particularly elegant proofs of the convergence of ADMM, and the two algorithms are frequently studied together [22,23].

    Figure 2.  The ADMM sequence (xn,zn,λn)nN for solving (BP-p) and the associated dual DR sequence (yn)nN for solving (BP-d). They have the relationship in Table 1.

    Benoist, Dao, Tam, Giladi, and Rüffer used Lyapunov functions to describe the dynamics of DR in cases when it exhibited spiraling patterns similar to the one in Figure 2 [8,14,21]. All of their Lyapunov functions U shared the property that

    U(yn),yn1yn=0. (5.1)

    Borrowing inspiration from their constructions, the method introduced in [24] provides an update candidate that minimizes a spherical surrogate for such a Lyapunov function. Figure 2 shows an example of such a surrogate, and the update obtained from it. When yk,yk+1,yk+2 are collinear, the construction is not possible and so a normal algorithmic update would be accepted. Otherwise, this update, denoted LT(yk), is constructed to belong to the 2-dimensional affine subspace containing yk, yk+1, and yk+2, and to have the property that LTykyk+1,yk+1yk=LTykyk+2,yk+2yk+1=0. A candidate update for the multiplier is

    λLT:=proxρgLT(yk)=P{x|x1}(LT(yk)).

    We can propagate the update to the primal variables by zLTLTykλLT and xLTPS(zLTλLT). Then xLT is a candidate for updating x. There are various criteria we can use to decide whether to accept it or reject it in favor of a regular ADMM update. In this paper, the main criterion we use in our experiments is whether xLT has a smaller 1-norm value than a regular update would. Based on the symbol for the operator LT:RrRr, we refer to the full algorithm in this case as LT. For comparison, in some experiments we instead accept the xLT update always. We denote the algorithm in such cases by LTA. The operator LT may be computed efficiently, as we now show.

    Theorem 2. Let y0,y1,y2H. Let w1=y1y0 and w2=y2y0. If w12w22(w1,w2)2=0, then y0,y1,y2 are collinear. Otherwise, let w=y0+μ1w1+μ2w2 where

    δ:=w12w22(w1,w2)2,μ:=1δ[w22w1,w2w1,w2w12][w12w22w1,w2+w12],

    and w is the unique point in the 2-dimensional affine subspace of H containing {y0,y1,y2} that satisfies

    wy1,y1y0=wy2,y2y1=0.

    Proof. Letting u=wy0, the linear system rewrites to

    uw2,w2w1=0anduw1,w1=0.

    The second equality yields u,w1=w12, whereupon the first equality yields

    u,w2w1=w2,w2w1=w22w2,w1u,w2=w22w2,w1+u,w1=w22w2,w1+w12.

    As w belongs to the 2-dimensional affine subspace containing {y0,y1,y2}, there must exist μ1,μ2R such that u=μ1w1+μ2w2. Substituting this identity into the two equalities above yields the system:

    w1,μ1w1+μ2w2=w12,andw2,μ1w1+μ2w2=w22w2,w1+w12,[w12w1,w2w1,w2w22][μ1μ2]=[w12w22w1,w2+w12].

    The left matrix is invertible so long as its determinant δ is nonzero. Bearing in mind that w1w2cos(θ)=w1,w2 where θ is the angle between w1 and w2, we have that this determinant is zero exactly when the angle between w1 and w2 is 0 or π (i.e., y0,y1,y2 are collinear). When the determinant is nonzero, we take the inverse of the matrix on the left and obtain the claimed identity for μ.

    Remark 1 (Checking collinearity). In computational practice, to decide whether the determinant δ in Theorem 2 is nonzero, one must choose some ϵ>0 and replace the condition δ0 with |δ|>ϵ. In our experiments, we used the threshold ϵ=1010, and never encountered the case when |δ|<ϵ. If ϵ is chosen very small and the condition number is very large, the computed μ may suffer numerical inaccuracy. This is another important reason for the inclusion of the objective function check.

    Interestingly, for structured feasibility problems of the kind studied in [8,14,21], it was shown in [24] that the circumcentered reflections method (see, for example, [1,4,5,6,7,15,17]) is another example of an algorithm that returns minimizers for surrogates of a Lyapunov function for the DR dynamical system. For reasons described in [24], the natural generalization of circumcentered reflections method does not work for basis pursuit problems, and so we do not include it here. Indeed, a key motivation for the algorithm LT is that whenever a Lyapunov function has the property (5.1), LT and LTA return minimizers of spherical surrogates for that Lyapunov function [24]. This means that they may be useful for a wider variety of applications than those listed here. Algorithm 1 shows their implementation for basis pursuit specifically. In Table 2, we summarize the criteria that we use in our experiments to decide whether to accept an update candidate derived from LT(yk) or reject it in favor of a regular update.

    Algorithm 1 LT/LTA.
    Data: Am×n, absolute and relative tolerances such that ϵabs, ϵrel>0, a tolerance ϵdet>0, and a data vector b.
    Initialisation:
    i,x,z,z^,λ,yi0; r=xz; s=z^z;
    ϵpri=nϵabs+ϵrelmax{||x||,||z^||}; ϵdual=nϵabs+ϵrel||λ||;
    while: ||r||ϵpri,||s||ϵdual do

    end

    Table 2.  The criteria for LT and LTA in Algorithm 1.
    Algorithm variation Criterion used
    LT ||xLT||1<||x||1
    LTA true

     | Show Table
    DownLoad: CSV

    Performance profiles are a popular way of comparing algorithm performance; see [18], whose notation we closely follow. We use them to compare the performances of LT, LTA, and ADMM. For each problem p and solver s, we define

    tp,s= the number of iterations required to solve problem p by solver s.

    The performance of each algorithm is evaluated relative to the superior algorithm in each instance. This is done according to the following ratio, where S denotes the set of solvers

    rp,s=tp,smin{tp,s:sS}.

    Our performance profile graphs depict the cumulative distribution functions Φs, defined as

    Φs(τ)=1npsize{pP|rp,sτ},

    where np is the total number of problems solved by s. For example, if Φs(1.5)=0.8, then algorithm s solved 80% of problems using no more than 150% of the passes through (3.2) required by the algorithm that used the fewest passes.

    In [24], LT was found to reliably outperform ADMM for problem (2.2), where the problems were randomly generated according to the specifications in Boyd, Parikh, Chu, Peleato and Eckstein's Matlab scripts [12]. The code for these problems is (P-random). Here, AR500×5000, with each entry drawn from the standard normal distribution, while xR5000 is sparse with approximately 500 nonzero xiN(0,1). The data vector c is given by c=Ax. The performance profile for LT, LTA, and ADMM in solving 1000 problems of this kind is Figure 3b. For these problems, the algorithms stopped once locating solutions within ±104 of optimality (i.e., ϵabs=ϵrel=104).

    Figure 3.  Algorithm performance for basis pursuit problems generated using (P-random).
    Matlabcode2.The code used for randomly-generating basis pursuit problems.n=5000;m=500;A=randn(m,n);x=sprandn(n,1,0.1n);c=Ax.(Prandom)

    Following [24], we say that an algorithm shows signs of spiraling when proxgyk+1proxgyk (i.e., λk+1λk) exhibits a distinct "tombstone" pattern (observable in Figure 3a). This phenomenon is often associated with the existence of a Lyapunov function satisfying condition (5.1). Problems generated using (P-random) typically exhibit signs of spiraling.

    In Figure 4, we illustrate the behavior of the DR dual sequence (yk)kN for a problem generated using (P-random). We visualize (yk)'s behavior by projecting it onto 10 different 2-dimensional subspaces (i.e., plotting coordinate pairs). Namely, we plot (y1n,y2n) (red), (y1n,y3n) (neon green), (y1n,y4n) (blue), (y1n,y5n) (cyan), (y2n,y3n) (magenta), (y2n,y4n) (yellow), (y2n,y5n) (purple), (y3n,y4n) (teal), (y3n,y5n) (orange), and (y4n,y5n) (green). In Figure 4b and 4c, we zoom in on the tail of the sequence (y2n,y3n) and plot the sequence ((yLTn)2,(yLTn)3) of coordinate pairs (blue shades) as well as the sequence ((yLTAn)2,(yLTAn)3) (green shades). The accepted LT updates are shown as squares.

    For the example shown, the random seed state was rng(8).

    Figure 4.  A basis pursuit problem showing signs of spiralling. Squares (shaded lighter) denote accepted LT updates.

    Because LT and LTA afforded measurable improvement for solving basis pursuit problems with random structure, we next benchmarked their performance on the audio compression and recovery problem.

    To evaluate the performance of LT and LTA for audio recovery problems, we generated signals using (P-signal). d=[0,144100,244100,...,441044100] was used as the domain of each c, while c is the sum of shifted and dilated sine functions with amplitudes and frequencies specified by a1,...,aFreq[0,1] and b1,...,bFreq[1,2,...,500]. The number of frequencies (Freq) corresponds to the number of such functions composing c, with a higher number of frequencies resulting in more complex signals. To assess the algorithms on a variety of signals, we ran experiments with Freq=2,5 and 10. For each Freq value, 1000 problems were solved where A=M~441×4410. We include the performance profiles for these problems in Figure 5.

    Figure 5.  The performance profiles generated for LT, LTA, and ADMM in solving 1000 problems with A=M~441×4410 and c generated using (P-signal).

    We provide a sense of how the algorithms pursue a sparse solution in Figure 6, where we define the superfluous xi as having absolute values less than 0.7. This threshold was determined based on plotting the spectra for various solved problems (as in Figure 1). Figures 6ac depict Mx5, Mx30 and Mx (x=x158), as produced by ADMM solving (2.2) with A=M~ and c specified by (P-signal) with Freq=15§. We graph the percentage of xik that satisfy |xik|>0.7 for k=5,30,158, in Figure 6d. These signals were reconstructed with ϵabs=ϵrel=103 accuracy. Although not perfect, we found that lower ϵabs and ϵrel values did not result in better approximations for V, so this accuracy is used throughout our experiments.

    §For the example shown, the random seed state was rng(3).

    Figure 6.  The signals reproduced at different stages of ADMM solving basis pursuit with A=M~ and c generated using (P-signal).

    We likewise applied LT and LTA to recovering the waveform signal of a performance of Frédéric Chopin's Scherzo No. 2. We decomposed the 2 minute and 14 second performance into 1340 problems, each recovering 0.1 seconds of audio. Using the file's sampling rate of 44800 Hz, each problem was formulated with A=M~448×4480. The performance profile for this experiment is Figure 7.

    Figure 7.  The performances of LT, LTA, and ADMM in recovering Chopin's Op. 31. The profile depicts 1340 problems with A=M~448×4480.

    The meta algorithms exhibited noticeably less improvement relative to ADMM when the random problem structure of (P-random) was replaced by the structure of the simulated audio problem (P-signal), and even less improvement for a real audio problem. For problems we solved using M~, we typically did not observe signs of spiraling for the dual sequence yk; when we did observe signs of spiraling, we only observed it after the signal was recovered (i.e. when computing more than necessary). Figure 8 shows yk coordinate pairs for an example problem generated using (P-signal) with Freq = 10. Enhanced views of sequences (y1n,y3n) and (y1n,y4n) for this problem are included in Figure 9a. Figure 9b enhances the sequence (y2n,y3n), and also shows the sequence of coordinate pairs ((yLTn)2,(yLTn)3) generated by LT, as well as the sequence of coordinate pairs ((yLTAn)2,(yLTAn)3) generated by the variant LTA. For the latter two sequences, regular DR updates are circles while updates accepted from the surrogate are squares.

    For the example shown, the random seed state was rng(4)

    Figure 8.  Convergence plot (left) and plot of coordinate pairs (right) for a problem not displaying signs of spiraling.
    Figure 9.  Enhanced views of sequences for the problem in Figure 8. Squares (shaded lighter) denote accepted LT updates. All three coordinate pair sequences entered from bottom right and terminated at top right.

    Our results highlight the importance of testing algorithms on problems with real data. While the Lyapunov surrogate variant LTA is state-of-the-art (much faster than DR and equally reliable) for the real world problem of finding wavelets [16], its performance for the real world audio compression and recovery problem is less impressive than experiments on random problems might suggest.

    Additionally, our results are consistent with the hypothesis in [24] that Lyapunov surrogate methods are more likely to deliver improvement in situations when signs of spiraling are present, as we suspect that LT's inconsistent performance is due to differences in the dynamics of the underlying DR dual algorithm.

    Finally, the visualizations in Figures 4 and 9 are interesting in two ways. First, they support the hypothesis that signs of spiraling do, in fact, reflect the dynamic they purport to. The dynamics we observe for the simulated audio problem in Figure 9, when signs of spiraling were not present, appear much less regular than those we observed for the fully random problem in Figure 4b, when signs of spiraling were present. Second, the latter dynamics still exhibit patterns that may lend themselves to extrapolation methods or other dynamics-based schemes. We suggest this as future research.

    The authors declare that they have not used artificial intelligence tools in the creation of this article.

    The authors declare no conflict of interest.

    The code used to generate all examples in this paper is available at https://github.com/AndrewCalcan/Basis-Pursuit-via-LT-LTA-ADMM.git.



    [1] R. Arefidamghani, R. Behling, Y. Bello-Cruz, A. Iusem, L. Santos, The circumcentered-reflection method achieves better rates than alternating projections, Comput. Optim. Appl., 79 (2021), 507–530. http://dx.doi.org/10.1007/s10589-021-00275-6 doi: 10.1007/s10589-021-00275-6
    [2] H. Attouch, On the maximality of the sum of two maximal monotone operators, Nonlinear Anal.-Theor., 5 (1981), 143–147. http://dx.doi.org/10.1016/0362-546X(81)90039-0 doi: 10.1016/0362-546X(81)90039-0
    [3] H. Bauschke, P. Combettes, Convex analysis and monotone operator theory in Hilbert spaces, Cham: Springer, 2 Eds., 2017. http://dx.doi.org/10.1007/978-3-319-48311-5
    [4] H. Bauschke, H. Ouyang, X. Wang, On circumcenters of finite sets in Hilbert spaces, arXiv: 1807.02093.
    [5] R. Behling, Y. Bello-Cruz, L. Santos, Circumcentering the Douglas-Rachford method, Numer. Algor., 78 (2018), 759–776. http://dx.doi.org/10.1007/s11075-017-0399-5 doi: 10.1007/s11075-017-0399-5
    [6] R. Behling, Y. Bello-Cruz, L. Santos, On the linear convergence of the circumcentered-reflection method, Oper. Res. Lett., 46 (2018), 159–162. http://dx.doi.org/10.1016/j.orl.2017.11.018 doi: 10.1016/j.orl.2017.11.018
    [7] R. Behling, Y. Bello-Cruz, L. Santos, On the circumcentered-reflection method for the convex feasibility problem, Numer. Algor., 86 (2021), 1475–1494. http://dx.doi.org/10.1007/s11075-020-00941-6 doi: 10.1007/s11075-020-00941-6
    [8] J. Benoist, The Douglas-Rachford algorithm for the case of the sphere and the line, J. Glob. Optim., 63 (2015), 363–380. http://dx.doi.org/10.1007/s10898-015-0296-1 doi: 10.1007/s10898-015-0296-1
    [9] D. Bertsekas, Convex optimization theory, Melrose: Athena Scientific, 2009.
    [10] J. Borwein, A. Lewis, Convex analysis and nonlinear optimization: theory and examples, New York: Springer, 2 Eds., 2006. http://dx.doi.org/10.1007/978-0-387-31256-9
    [11] S. Boyd, L. Vandenberghe, Convex optimisation, Cambridge: Cambridge University Press, 2004. http://dx.doi.org/10.1017/CBO9780511804441
    [12] S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein, Matlab scripts for alternating direction method of multipliers, 2011. Available from: https://web.stanford.edu/~boyd/papers/admm/.
    [13] S. Boyd, N. Parikh, E. Chu, B. Peleato, J. Eckstein, Distributed optimisation and statistical learning via the alternating direction method of multipliers, New York: Now Foundations and Trends, 2010. http://dx.doi.org/10.1561/2200000016
    [14] M. Dao, M. Tam, A Luapunov-type approach to convergence of the Douglas-Rachford algorithm, J. Glob. Optim., 73 (2019), 83–112. http://dx.doi.org/10.1007/s10898-018-0677-3 doi: 10.1007/s10898-018-0677-3
    [15] N. Dizon, J. Hogan, S. Lindstrom, Circumcentered reflections method for wavelet feasibility problems, ANZIAM J., 62 (2020), C98–C111. http://dx.doi.org/10.21914/anziamj.v62.16118 doi: 10.21914/anziamj.v62.16118
    [16] N. Dizon, J. Hogan, S. Lindstrom, Centering projection methods for wavelet feasibility problems, In Current trends in analysis, its applications and computation, Cham: Birkhäuser, 2022. http://dx.doi.org/10.1007/978-3-030-87502-2_66
    [17] Neil Dizon, J. A. Hogan, and Scott B. Lindstrom, Circumcentering reflection methods for nonconvex feasibility problems. Set-Valued Var. Anal., 30 (2022), 943–973. http://dx.doi.org/10.1007/s11228-021-00626-9
    [18] E. Dolan, J. Moré, Benchmarking optimisation software with performance profiles, Math. Program., 91 (2022), 201–213. http://dx.doi.org/10.1007/s101070100263 doi: 10.1007/s101070100263
    [19] J. Eckstein, W. Yao, Understanding the convergence of the alternating direction method of multipliers: theoretical and computational perspectives, Pac. J. Optim., 11 (2015), 619–644.
    [20] D. Gabay, Applications of the method of multipliers to variational inequalities, Studies in Mathematics and its Applications, 15 (1983), 299–331. http://dx.doi.org/10.1016/S0168-2024(08)70034-1 doi: 10.1016/S0168-2024(08)70034-1
    [21] O. Giladi, B. Rüffer, A Lyapunov function construction for a non-convex Douglas-Rachford iteration, J. Optim. Theory Appl., 180 (2019), 729–750. http://dx.doi.org/10.1007/s10957-018-1405-3 doi: 10.1007/s10957-018-1405-3
    [22] B. He, X. Yuan, On the O(1/n) convergence rate of the Douglas-Rachford alternating direction method, SIAM J. Numer. Anal., 50 (2012), 700–709. http://dx.doi.org/10.1137/110836936 doi: 10.1137/110836936
    [23] B. He, X. Yuan, On the convergence rate of Douglas-Rachford operator splitting method, Math. Program., 153 (2015), 715–722. http://dx.doi.org/10.1007/s10107-014-0805-x doi: 10.1007/s10107-014-0805-x
    [24] S. B. Lindstrom, Computable centering methods for spiralling algorithms and their duals with motivations from the theory of Lyapunov functions, Comput. Optim. Appl., 83 (2022), 999–1026. http://dx.doi.org/10.1007/s10589-022-00413-8 doi: 10.1007/s10589-022-00413-8
    [25] S. B. Lindstrom, B. Sims, Survey: sixty years of Douglas-Rachford, J. Aust. Math. Soc., 110 (2021), 333–370. http://dx.doi.org/10.1017/S1446788719000570 doi: 10.1017/S1446788719000570
    [26] P. Lions, B. Mercier, Splitting algorithms for the sum of two nonlinear operators, SIAM J. Numer. Anal., 16 (1979), 964–979. http://dx.doi.org/10.1137/0716071 doi: 10.1137/0716071
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1309) PDF downloads(84) Cited by(0)

Figures and Tables

Figures(9)  /  Tables(2)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog