1.
Introduction
In this paper, we investigate the following convex minimization problem
where H is a real Hilbert space, g:H→(−∞,+∞] is proper, lower semicontinuous and covex and f:H→R is convex and differentiable with the Lipschitz continuous gradient denoted by ∇f. It is known that x∗ is a minimizer of f+g if and only if
where ∂g denotes the subdifferential of g.
The convex minimization problem is an important mathematical models which unify numerous issues in applied mathematics for example, signal processing, image reconstruction, machine learning and so on. See [1,3,8,9,11,22,31].
The most popular algorithm for solving the convex minimization problem is the so-called forward-backward algorithm (FB), which generates by a starting point x1∈H and
where proxg is the proximal operator of g and the stepsize λ∈(0,2/L), L is the Lipschitz constant of ∇f.
Polyak [21] first proposed the inertial idea to improve the convergence speed of the method. In recent years, many authors introduced various fast iterative methods via inertial technique, for example, [7,8,10,15,16,18,23,25,26,32].
In 2009, Beck and Teboulle [4] introduced the fast iterative shrinkage-thresholding algorithm for linear inverse problem (FISTA). Let t0=1 and x0=x1∈H. Compute
This improves the convergence speed for O(1/n2). However, the stepsize is established under the condition of the Lipschitz constant which is not known in general.
In 2000, Tseng [29] proposed a modified forward-backward algorithm (MFB) via the stepsize with linesearch technique as follows. Given σ>0, ρ∈(0,1), δ∈(0,1) and x1∈H. Compute
where λn is the largest λ∈{σ,σρ,σρ2,...} satisfying λ‖∇f(yn)−∇f(xn)‖≤‖yn−xn‖.
In 2020, Padcharoen et al. [20] proposed the modified forward-backward splitting method based on inertial Tseng method (IMFB). Given {λn}⊂(0,1L), {αn}⊂[0,α]⊂[0,1). Let x0,x1∈H and compute
They established weak convergence of the proposed method.
In 2015, Shehu et al. [24] introduced the modified split proximal method (MSP). Let r:H→H be a contraction mapping with constant α∈(0,1). Set φ(x)=√‖∇h(x)‖2+‖∇ℓ(x)‖2 with h(x) = 12‖(I−proxλg)Ax‖2, ℓ(x)=12‖(I−proxλμnf)x‖2. Given an initial point x1∈H and construct
where the stepsize μn=ψnh(xn)+ℓ(xn)φ2(xn) with 0<ψn<4. They proved strong convergence theorem for proximal split feasibility problems.
In 2016, Cruz and Nghia [5] presented a fast multistep forward-backward method (FMFB) with a linesearch. Given σ>0, μ∈(0,12), ρ∈(0,1) and t0=1. Choose x0,x1∈H and compute
where λn=σρmn and mn is the smallest nonnegative integer such that
Very recently, Malitsky and Tam [17] introduced the forward-reflected-backward algorithm (FRB). Given λ0>0, δ∈(0,1), γ∈{1,β−1} and β∈(0,1). Compute
where the stepsize λn=γλn−1βi with i being the smallest nonnegative integer satisfying λn‖∇f(xn+1)−∇f(xn)‖≤δ2‖xn+1−xn‖.
Very recently, Hieu et al. [13] proposed the modified forward-reflected-backward method (MFRB) with adaptive stepsize. Given x0,x1∈H, λ0,λ1>0, μ∈(0,12):
This stepsize allows the proposed method without knowing the Lipschitz constant to solve the problem.
Inspired and motivated by previous works, we propose based on the adaptive stepsize, the inertial proximal gradient algorithm for convex minimization problems. This method requires more flexible conditions than the fixed stepsize does. We then establish weak convergence of our scheme under some assumptions. Moreover, we present some numerical experiments in image deblurring. It reveals that our algorithm outperforms other methods.
2.
Basic definitions and lemmas
In this section, we provide some definitions and lemmas for proving our theorem.
Weak and strong convergence of a sequence {xn}⊂Ω to z∈Ω are denoted by xn⇀z and xn→z, respectively.
Let g:H→(−∞,+∞] be a proper, lower semicontinuous and convex function. We denote the domain of g by domg={x∈H|g(x)<+∞}. For any x∈domg, the subdifferential of g at x is defined by
Recall that the proximal operator proxg:dom(g)→H is given by proxg(x)=(I+∂g)−1(z), z∈H. It is known that the proximal operator is single-valued. Moreover, we have
Definition 2.1. Let S be a nonempty subset of H. A sequence {xn} in H is said to be quasi-Fejér convergent to S if and only if for all x∈S there exists a positive sequence {εn} such that ∞∑n=1εn<+∞ and ‖xn+1−x‖2≤‖xn−x‖2+εn for all n≥1. If {εn} is a null sequence, we say that {xn} is Fejér convergent to S.
Lemma 2.1. [6] The subdifferential operator ∂g is maximal monotone. Moreover, the graph of ∂g, Gph(∂g)={(x,v)∈H×H:v∈∂g(x)} is demiclosed, i.e., if the sequence {(xn,vn)}⊂Gph(∂g) satisfies that {xn} converges weakly to x and {vn} converges strongly to v, then (x,v)∈Gph(∂g).
Lemma 2.2. [19] Let {an}, {bn} and {cn} be real positive sequences such that
If Σ∞n=1cn<+∞ and Σ∞n=1bn<+∞, then limn→+∞an exists.
Lemma 2.3. [12] Let {an} and {θn} be real positive sequences such that
Then, an+1≤K⋅∏ni=1(1+2θi) where K=max{a1,a2}. Moreover, if ∑∞n=1θn<+∞, then {an} is bounded.
Lemma 2.4. [2,14] If {xn} is quasi-Fejér convergent to S, then we have:
(i) {xn} is bounded.
(ii) If all weak accumulation points of {xn} is in S, then {xn} weakly converges to a point in S.
3.
Main result
In this section, we assume that the following conditions are satisfied for our convergence analysis:
(A1) The solution set of the convex minimization problem (1.1) is nonempty, i.e., Ω=argmin(f+g)≠∅.
(A2) f,g:H→(−∞,+∞] are two proper, lower semicontinuous and convex functions.
(A3) f is differentiable on H and ∇f is Lipschitz continuous on H with the Lipschitz constant L>0.
We next introduce a new inertial forward-backward method for solving (1.1).
Algorithm 3.1. Inertial modified forward-backward method (IMFB)
Initialization: Let x0=x1∈H, λ1>0, θ1>0 and δ∈(0,1).
Iterative step: For n≥1, calculate xn+1 as follows:
Step 1. Compute the inertial step:
Step 2. Compute the forward-backward step:
Step 3. Compute the xn+1 step:
where
Set n=n+1 and return to Step 1.
Remark 3.1. It is easy to see that the sequence {λn} is non-increasing. From the Lipschitz continuity of ∇f, there exists L>0 such that ‖∇f(wn)−∇f(yn)‖≤L‖wn−yn‖. Hence,
By the definition of {λn}, it implies that the sequence {λn} is bounded from below by min{λ0,δL}. So, we obtain limn→∞λn=λ>0.
Lemma 3.1. Let {xn} be generated by Algorithm 3.1. Then
Proof. Let x∗∈Ω. Then
Note that
It follows that
Combining (3.7) and (3.9), we obtain
From (3.2), we see that wn−λn∇f(wn)∈(I+λn∂g)yn. Since ∂g is maximal monotone, then there is un∈∂g(yn) such that
This shows that
Since 0∈(∇f+∂g)(x∗) and ∇f(yn)+un∈(∇f+∂g)yn, we get
Substituting (3.12) into (3.13), we have
This implies that ⟨wn−λn∇f(wn)−yn+λn∇f(yn),yn−x∗⟩≥0. Using (3.10), we derive
Lemma 3.2. Let {xn} be generated by Algorithm 3.1. If ∑∞n=1θn<∞, then limn→∞‖xn−x∗‖ exists for all x∗∈Ω.
Proof. Let x∗∈Ω. From Lemma 3.1, we see that
So, we have
Hence
By Lemma 2.3, we conclude that
where K=max{‖x1−x∗‖,‖x2−x∗‖}. Since ∑∞n=1θn<+∞, by Lemma 2.3, we have {xn−x∗} is bounded. Hence ∑∞n=1θn‖xn−xn−1‖<+∞. By Lemma 2.2 and (3.17), we have limn→∞‖xn−x∗‖ exists.
Lemma 3.3. Let {xn} be generated by Algorithm 3.1. If ∑∞n=1θn<∞, then
Proof. We see that
From (3.15) and (3.20), we have
Note that θn‖xn−xn−1‖→0 and limn→∞‖xn−x∗‖ exists by Lemma 3.2. From (3.1) and (3.21), we have ‖wn−xn‖→0 and ‖wn−yn‖→0, respectively. It is easy to see that ‖xn−yn‖→0. Since ∇f is uniformly continuous, we obtain
From (3.3) and (3.22), we get
Thus, we have
Theorem 3.1. Let {xn} be generated by Algorithm 3.1. If ∑∞n=1θn<∞, then {xn} weakly converges to a point in Ω.
Proof. Since {xn} is bounded, there exists a subsequence {xnk} of {xn} such that xnk⇀ˉx∈H. From Lemma 3.3, we obtain xnk+1⇀ˉx. We note that
From (2.1), we obtain
Hence
Since ‖xn−yn‖→0, we also have ynk⇀ˉx. Letting k→∞ in (3.27) and using (3.22), by Lemma 2.1 and Remark 3.1, we get
So ˉx∈Ω. From (3.21) we see that {xn} is a quasi-Fejer sequence. Hence, by Lemma 2.4, we conclude that {xn} weakly converges to a point in Ω. This completes the proof.
4.
Numerical experiment in image deblurring
The image deblurring can be modeled by the following linear equation system:
where A∈RN×M is the blurring matrix, x∈RN the original image, b∈RM the degraded image and v∈RM is the noisy.
An approximation of the clean image can be found by the following LASSO problem [27]:
where τ is a positive parameter, ‖⋅‖1 is the ℓ1-norm, and ‖⋅‖2 is the Euclidean norm.
It is known that (4.1) can be written in the form (1.1) by defining f(x)=12‖b−Ax‖22 and g(x)=τ‖x‖1. We compare our algorithm (IMFB) with FISTA, MFB, FRB, MFRB, IMFB, MSP and FMFB.
In method IMFB, we set t0=1, tn=1+√1+4t2n−12 and
The regularization parameters are chosen by τ=10−5 and x0=x1=(1,1,1,...,1)∈RN. We set the following parameters in Table 1.
In this example, we set all parameters as in Table 1. For the experiments, we use the sizes 251×189 for RGB images which are blurred by the following blur types:
(ⅰ) Motion blur with motion length of 45 pixels and motion orientation 180∘.
(ⅱ) Gaussian blur of filter size 5×5 with standard deviation 5.
(ⅲ) Out of focus with radius 7.
We add Poisson noise and use a Fast Fourier Transform (FFT) for converting it to the frequency domain. Structural similarity index measure (SSIM) [30] is used for measuring the similarity between two images. Peak-signal-to-noise ratio (PSNR) in decibel (dB) [28] is defined by
where MSE=‖xn−x‖2 and x is the original image. It is noted that, a higher PSNR generally indicates that the reconstruction is of higher quality. The resultant SSIM index is a decimal value between 0 and 1, and value 1 is indicates perfect structural similarity.
The numerical experiments have been carried out in Matlab environment (version R2020b) on MacBook Pro M1 with ram 8 GB. For the results recovering the degraded RGB images, we limit the iterations to 1,000. We report the numerical results in Table 2.
In Table 2, we see that IMFB has a higher PSNR than FISTA, MFB, FRB, MFRB, IMFB, MSP, FMFB for the same number of iterations. Moreover, SSIM of IMFB is closer to 1 than other methods. This shows that our algorithm has a better convergence than other methods for this example. However, we observe that IMFB has a less CPU time than other methods.
We next show the different types of blurred RGB images with the PSNR in Figure 1.
We next show the restored images of RGB images for Motion blur with the PSNR in Figure 2.
We next show the restored images of RGB images for Gaussian blur with the PSNR in Figure 3.
We next show the restored images of RGB images for out of focus with the PSNR in Figure 4.
The results of the numerical experiments are summarized in Table 2. Figure 1 shows the original and blurred images for this experiment. In Figures 2–5, we report all results that include the recovered images via each algorithms. It is shown that IMFB outperforms FISTA, MFB, FRB, MFRB, IMFB, MSP and FMFB in terms of PSNR and SSIM.
5.
Conclusions
In this paper, we established the convergence theorem of the iterates generated by a new modified inertial forward-backward algorithm with adaptive stepsize under some suitable conditions for convex minimization problems. We applied our main result to image recovery. It was shown that our proposed method outperforms other methods in terms of PSNR and SSIM. In future work, we will study the convergence rate of the iteration.
Acknowledgments
This project is funded by National Research Council of Thailand (NRCT) under grant No. N41A640094. The authors wish to thank University of Phayao and Thailand Science Research and Innovation grant No. FF65-UOE001.
Conflict of interest
The authors declare no conflict of interest.