1.
Introduction
There have been lots of theoretical studies on the optimal control of PDEs with state constraints, which form a foundation for its numerical approximation. Existence and uniqueness of the solutions, Lagrange multipliers, optimality conditions, and important regularity results were derived for control problems with state constraints in the pioneer work [1]. More discussion on these topics can be found in [2,3]. In respect of numerical methods, the finite element approximation of state constrained control problems has been widely investigated, and we don't try to give a detailed introduction here, one can find more works [4,5], including pointwise constraints, integral constraint and so on. At the same time, many numerical strategies were developed to provide efficient approximation for these control problems. Bergounioux and Kunisch [6] used the primal-dual active set algorithm to solve state-constrained problems. A semi-smooth Newton method was proposed to compute state-constrained control problems by De los Reyes and Kunisch [7]. Gong and Yan [8] established a mixed variational scheme for control problems with pointwise state constrains, and a direct numerical algorithm was adopted without the optimality conditions.
In recent years, spectral method has been used to approximate control problems. Despite the restriction on the higher regularity of the approximated solutions, spectral method can provide fast convergence rate and high-order accuracy with a smaller number of unknowns, which is significant to successful applications of control problems. The spectral method was considered to solve the control problems with integral state constraint, and a priori error estimates was established in [9]. Chen et al. [10,11] derived both the a priori and a posteriori error estimates for optimal control with control constraints. However, they only provided the upper bound estimation for the a posteriori error indicators, and numerical tests didn't illustrate the performance of estimators. To investigate the efficiency of adaptive strategy in the spectral method, we have to establish some successful a posteriori error estimators as in the finite element framework. However, there are not much work on these aspects to the best of our knowledge. In this work, we derive a posteriori error estimator, and prove that it can be constructed as the upper and lower bounds of approximation error.
The plan of the article is as follows. Spectral approximation of the control problem is presented, and optimality conditions of the exact problem and discretized problem are provided in the next section. We establish the a posteriori error estimator, and construct it as the upper and lower bounds of the approximation error in section 3. Numerical example confirms the theoretical result, and shows the behaviour of the indicator in section 4.
In this paper, we let Ω=(−1,1)2 and denote U=L2(Ω)2,Y=H10(Ω)2,Q=L20(Ω)={q∈L2(Ω) ∣ ∫Ωq=0}, where H10(Ω) and Hm(Ω) with m being a positive integer are usual Sobolev spaces on Ω. Let C denote a positive constant independent of N, the order of the spectral method.
2.
Spectral approximation and optimality conditions
In this section, we state the Galerkin spectral approximation and optimality conditions for the control problem with state constraint. The model under consideration is as follows: find (y,r,u)∈Y×Q×U such that
where y0,f∈L2(Ω)2, Gad={v∈U ∣ ∥v∥0,Ω ≤d}, and d, υ are positive constants. It is necessary to introduce a weak formula of the state equations for spectral approximation of the optimal control. Let
Then there exist two positive constants σ and δ such that for any u, v∈Y, q∈Q,
By Poincare inequality, we can know that there exists a constant γ>0 such that
Furthermore, there exists a constant β>0 (see, e.g., [12]) such that
Then the standard weak formula for the state equation can be presented as: find (y(u),r(u))∈Y×Q such that
The problem (2.4) is well-posed by Babuška-Brezzi theorem and (2.2)-(2.3), and the control problem (2.1) can be restated as follows (OCP): find (y,r,u)∈Y×Q×U such that
We can prove existence and uniqueness of the solutions for the control problem (OCP) by reformulating it to a control-constrained problem. In fact, the control problem (\rm OCP) can be equivalent to the control-constrained problem
where constraint set Uad={u∈U:‖Gu‖0,Ω≤d} with the operator G:u→y(u). It is clear that Uad is a closed and convex subset in U, and the control problem (2.5) has a unique solution by the standard theorem [13].
2.1. Spectral method approximation
Considering the spectral approximation of the control problem, we introduce the finite dimensional spaces YN, MN−2, and UN to approximate spaces Y, Q, and U respectively, where YN=(Q0N)2 with Q0N={xN∈QN:xN|∂Ω=0}, MN−2=QN−2∩L20(Ω), UN=(QN)2, and QN denotes the space of all algebraic polynomials of degree less than or equal to N with respect to each single variable xi, (i=1,2). Letting Ln(x) be the nth-degree Legendre polynomial and ϕk(x)=1√4k+6(Lk(x)−Lk+2(x)) (see [14], for example), then it holds
We now approximate the state equations as follows
It can be known from [12] that there exists a constant βN=O(N−12) satisfying
It follows from Babuška-Brezzi's theory that problem (2.8) is well-posed, and the Legendre Galerkin spectral approximation of (OCP) can be stated as (OCP)N:
where GNad=YN∩Gad.
2.2. Optimality conditions
We first derive optimality conditions for the exact problem (OCP) by the techniques and refined results developed in [2], though we can complete the derivation by other strategies. Then the similar conclusion is presented for the discretized problem (OCP)N.
Lemma 2.1. The triplet (y,r,u) is the solution of control problem (OCP) if and only if there is a (y∗,r∗,λ)∈Y×Q×R such that
where
Proof. We denote G:u→y the operator to solve the state equation (2.4), D(G(u)) the Gâteaux derivative of G at u. Furthermore, let U=Z=K=L2(Ω)2, C=Gad respectively and u0=−f satisfy the Slater type condition in Theorem 5.2 in [2], then there exists μ∈L2(Ω)2 such that (u,μ) satisfying
where (G(u),u)∈Y×U satisfies the control problem (OCP).
In the next analysis, it can be derived that
where λ satisfies (2.12). In fact, we can complete the proof by two cases: ‖y‖0,Ω<d and ‖y‖0,Ω=d.
In the first case ‖y‖0,Ω<d, the following inequality holds by (2.13)
Thus we have for all ω∈U and ‖ω‖0,Ω=1
due to (d−‖y‖0,Ω)ω+y∈Gad. Similarly,
By (2.15) and (2.16)
In the second case ‖y‖0,Ω=d, it follows from (2.13) that
Then, we have
which implies that
and
Then, the identity (2.14) follows from (2.17)–(2.19).
It can be derived from (2.4) that for any v∈L2(Ω)2
By y−y0+λy∈L2(Ω)2, we can introduce the co-state equation as follows
Letting w=y∗ in (2.20) and q=y′(u)v in (2.21), we have that for all v∈L2(Ω)2
Then (2.11)-(2.12) follows from (2.4), (2.13), (2.19), (2.21)-(2.22). Furthermore, lemma 2.1 can be proved as soon as the uniqueness of the solution is derived for (2.11). In fact, letting both (y1,r1,u1,y∗1,r∗1,λ) and (y2,r2,u2,y∗2,r∗2,λ) satisfy (2.11), then we have
It follows that
By (2.13)-(2.14), it holds that for all v∈Gad
It follows from (2.23) and (2.24) that
which implies u1=u2,y1=y2. Furthermore, it can be deduced that r1=r2,y∗1=y∗2,r∗1=r∗2,λ1=λ2. Then we can complete the proof of lemma 2.1 as we argue above.
Similarly, we can derive optimality conditions for the discretized control problem (2.10), and obtain the following result.
Lemma 2.2. The control problem (OCP)N has a unique solution, and the triple (yN,rN,uN)∈YN×MN−2×UN is the solution of (OCP)N if and only if there is a (y∗N,r∗N,λN)∈YN×MN−2×R such that
where
Remark 2.1. (see, e.g., [10]). The optimal control u∈H2(Ω)2 if the initial data functions y0,f∈L2(Ω)2 by (2.11).
3.
A posteriori error estimator
In this section, the a posteriori error estimates are derived, and an error indicator is established for the spectral approximation of the control problem. We first recall two important spectral projection operators and more details can be found in [15].
Lemma 3.1. Let P01,N:H10(Ω)2→(Q0N)2 be the projection operator, satisfying for any w∈H10(Ω)2
If w∈(Hm(Ω)∩H10(Ω))2 with m≥1, then
Lemma 3.2. Define PN:L2(Ω)→QN being the L2 orthogonal projection operator, which satisfies for any r∈L2(Ω)
If r∈Hm(Ω) with m≥0, then
The following lemma is helpful for further analysis.
Lemma 3.3. Let (yN,uN,y∗N,λN) be the solution of (2.25), then
Proof. By (yN,uN,y∗N,λN) satisfying the optimality conditions (2.25), we have
associating with (2.25a) and (2.25e), it holds that
The discussion on estimating ‖y∗N‖1,Ω can be divided into two cases: ‖yN‖0,Ω<d and ‖yN‖0,Ω=d.
The first case ‖yN‖0,Ω<d, we have λN=0. Letting qN=y∗N in (2.25c), we have
The second case ‖yN‖0,Ω=d, let qN=y∗N−1d2(y∗N,yN)yN in (2.25c). Then it holds
It follows that
It follows from (3.3) and (3.4) that
Similarly, we can estimate |λN| by two cases: ‖yN‖0,Ω<d and ‖yN‖0,Ω=d.
If ‖yN‖0,Ω<d, then |λN|≤C.
If ‖yN‖0,Ω=d, letting qN=yN in (2.25c), we derive
Therefore, it can be derived that
Then we can complete the proof by (3.2), (3.5), and (3.6).
3.1. Upper error bound
In this subsection, we establish the a posteriori error estimator and prove that it provides an upper bound for the discretization errors. It is convenient to introduce an auxiliary system: find (y(uN),r(uN),y∗(uN),r∗(uN)) such that
By utilizing lemma 3.1–3.3, we can get the following lemmas.
Lemma 3.4. Let (y,r,u,y∗,r∗,λ) and (yN,rN,uN,y∗N,r∗N,λN) be the solutions of (2.11) and (2.25) respectively. Then we have that
Proof. By (2.11) and (3.7), we have
The following analysis is divided into three cases.
If ‖y‖0,Ω=d, let q=y in (3.9d). Then we have
Let w=y∗−y∗(uN) in (2.11a), we have
Thus it can be deduced that
If ‖yN‖0,Ω=d, let q=y(uN) in (3.9d) to obtain that
Let w=y∗−y∗(uN) in (3.7a) to derive that
It follows that
If ‖y‖0,Ω<d and ‖yN‖0,Ω<d, it follows from (2.12) and (2.26) that λ=λN=0, which implies the identity
Letting w=y−y(uN) in (3.9a) and ϕ=r−r(uN) in (3.9b), we have that
which implies that
Then, we can derive from (2.11e), (2.25e), (3.10)–(3.13), and lemma 3.3 that
which implies the lemma 3.4.
Lemma 3.5. Let (y,r,u,y∗,r∗,λ) be the solution of (2.11), and (yN,rN,uN,y∗N,r∗N,λN) be the solution of (2.25). Then we have that
and
Proof. By (2.25) and (3.7), we have
Noting that
It follows that
Similarly, we can derive from (2.25) and (3.7) that
and
Then, we can complete the proof by (3.16)–(3.19) and Cauchy's inequality with ϵ.
The main result of this subsection can now be stated as follows.
Theorem 3.1. Let (y,r,u,y∗,r∗,λ) and (yN,rN,uN,y∗N,r∗N,λN) be the solutions of (2.11) and (2.25) respectively. Then we have that
where the total approximation error ‖e‖ is defined by
the estimator η is given below
and θ is presented as follows
Proof. By (2.11e), (2.25e), and (3.9), we have
It is clear that by (2.12) and (2.26)
which implies that
It follows from (3.24) and (3.26) that
associating with lemma 3.1 and lemma 3.3, we have
It can be derived from (3.9), (3.13), and lemma 3.3 that
associating with (3.13), (3.27), and lemma 3.4, it can be derived that
Then the theorem follows from lemma 3.4, lemma 3.5, (3.13), and (3.28)-(3.29).
3.2. Lower error bound
A lower bound for the error ‖e‖ is established to investigate the sharpness of the indicator in this subsection. We first need some polynomial inverse estimates presented in the following lemma, which can be found in [16].
Lemma 3.6. Let α,β∈R satisfy −1<α<β and δ∈[0,1]. Then there exist constant C1,C2=C2(α,β), and C3=C3(δ) such that for all zN∈QN
where ΦΩ(x) is the distance function defined by
In the subsequent analysis, we establish the lower bound for discretization error ‖e‖ by using lemma 3.6. In fact, letting α=0,β=γ∈(12,1] in (3.30b), we can derive that
Denote ϑ=(νΔyN−∇rN+uN+PNf)ΦγΩ, then we have
Furthermore, it follows from (3.30b) and (3.30c) that
Thus, we have
It follows from (3.31)–(3.33) that
It is clear that
Then, we have
Similarly, letting α=0,β=γ∈(12,1] in (3.30b), it holds that
Let ϑ∗=(νΔy∗N+∇r∗N+(1+λN)yN−PNy0)ΦγΩ to obtain that
Furthermore, it can be derived from (3.30b) and (3.30c) that
which implies the inequality
It follows from (3.36)–(3.38) that
Similar to η2, we have
Then the estimation of lower error bound can be stated as following theorem.
Theorem 3.2. Let (y,r,u,y∗,r∗,λ) and (yN,rN,uN,y∗N,r∗N,λN) be the solutions of (2.11) and (2.25) respectively. Then we have that
where e and θ are defined in theorem 3.1.
Proof. The theorem follows from (3.34)-(3.35) and (3.39)-(3.40).
4.
Numerical experiment
In this section, we carry out a numerical example for the control problem (OCP) to investigate whether the indicator η tends to zero at the same rate as the error ‖e‖. We are interested in the following model: find (y,r,u)∈Y×Q×U such that
where Gad={v∈U ∣ ∥v∥0,Ω ≤d}. The data of this example are as follows
and d=√6π. The exact solutions are given by
We solve the discrete system (2.25) via Arrow-Hurwicz algorithm (see, for example, [17] and [18]).
Arrow-Hurwicz Algorithm: we describe the main steps of the algorithm as follows.
● Step 1: Let k=0, choose a step size ρ>0, give the initial values λ0N and u0N.
● Step 2: Let l=0 and uk,0N=ukN.
● Step 3: Solve the state equations
and the co-state equations
● Step 4: Let uk,l+1N=uk,lN−PN(y∗k,lN+αuk,lN). If
let l=l+1 and then we turn to Step 3.
● Step 5: λk+1N=max{0,λkN+ρ(‖yk,lN‖0,Ω−d)}.
● Step 6: Stop if ∣λk+1N−λkN∣ <Tolλ and output
else, let uk+1N=uk,l+1N, k=k+1, and turn to Step 2.
Denote yN=(yN1,yN2), y∗N=(y∗N1,y∗N2), and uN=(uN1,uN2), we have the following expressions by the property of (2.7)
The numerical results are presented in Table 1. The table shows that the errors decrease rapidly with a relatively small number of unknowns, which is important in a number of applications. We further plot the error ‖e‖ and the error indicator η versus the ploynomial degree N in Figure 1, where the longitudinal axis is in logarithmic scale. It can be observed from the figure that the two curves are very close to each other, which implies that the indicator is nearly equivalent to the error ‖e‖. Moreover, it seems that the error ‖e‖ and the indicator η decay with the rate 104100.625N, which shows that the proposed method for the control problem is very efficient and the spectral accuracy is achieved. Compared with the case of finite element method, it can be seen that the indicator developed in this work can be implemented more simply and can provide successful estimation for the errors with less computational load, which is helpful for developing the hp adaptive spectral element method for the optimal control problems.
5.
Conclusions
In this paper, the upper and lower bounds of approximation error are provided with the help of the a posteriori error indicator. The illustrative numerical experiment shows the performance of the error estimator. In our future work, we hope to extend these results to adaptive method in the hp spectral element framework, which will be compared with the adaptive hp finite element method for optimal control problems.
Acknowledgments
The authors are grateful to the referees for their useful comments and suggestions, which lead to improvements of the presentation. The authors were supported by the State Key Program of National Natural Science Foundation of China (11931003), the National Natural Science Foundation of China (Grant No. 11601466 and 41974133), the Nanhu Scholars Program for Young Scholars of XYNU, the Postgraduate Education Reform and Quality Improvement Project of Henan Province (Grant No. YJS2022ZX34), the National Natural Science Foundation of Henan (Grant No. 202300410343), the Training Plan of Young Key Teachers in Universities of Henan (Grant No. 2018GGJS096), and the Scientific Research Fund Project for Young Scholars of XYNU (Grant No. 2020-QN-050).
Conflict of interest
The authors declare there is no conflicts of interest.