Loading [MathJax]/jax/output/SVG/jax.js
Editorial Special Issues

Urban greening for low carbon cities—introduction to the special issue

  • As a measure to counteract the effects of urban sprawl, with the continued growth of cities worldwide, different modes of urban greening are being increasingly recognized. This special issue addresses current developments in the transition to low carbon cities employing a variety of urban greening techniques. The special issue consists of 10 papers, including four review papers on the topics of biophilic architecture; environmental versus marketable aesthetics; urban agriculture; and the rationale for mainstreaming. It also contains several original research articles, some (about half of the special issue) presenting case studies, as for green redevelopment in Trenton, USA; facade greening in Genoa, Italy; climatic effects (on air temperature) in Rosario, Argentina; a modeling study for Melbourne, Australia; and another Australian case study on the greening and “un”greening of Adelaide. In addition to a broadly scoped paper that examines American stormwater management, the special issue also contains an editorial on technologies for wastewater treatment. Together, these papers constitute a contribution to recognize the importance of retaining greenery in cities chiefly, although not solely, as a countermeasure to urban sprawl and its environmental impacts. Urban greening here represents a cost-effective (soft) approach that is an effective tool as part of sustainable development.

    Citation: Mary Thornbush. Urban greening for low carbon cities—introduction to the special issue[J]. AIMS Environmental Science, 2016, 3(1): 133-139. doi: 10.3934/environsci.2016.1.133

    Related Papers:

    [1] Meixin Xiong, Liuhong Chen, Ju Ming, Jaemin Shin . Accelerating the Bayesian inference of inverse problems by using data-driven compressive sensing method based on proper orthogonal decomposition. Electronic Research Archive, 2021, 29(5): 3383-3403. doi: 10.3934/era.2021044
    [2] Youjun Deng, Hongyu Liu, Xianchao Wang, Dong Wei, Liyan Zhu . Simultaneous recovery of surface heat flux and thickness of a solid structure by ultrasonic measurements. Electronic Research Archive, 2021, 29(5): 3081-3096. doi: 10.3934/era.2021027
    [3] Chunhua Chu, Yijun Chen, Qun Zhang, Ying Luo . Joint linear array structure and waveform design for MIMO radar under practical constraints. Electronic Research Archive, 2022, 30(9): 3249-3265. doi: 10.3934/era.2022165
    [4] Lot-Kei Chou, Siu-Long Lei . High dimensional Riesz space distributed-order advection-dispersion equations with ADI scheme in compression format. Electronic Research Archive, 2022, 30(4): 1463-1476. doi: 10.3934/era.2022077
    [5] Aliyu Muhammad Awwal, Sulaiman M. Ibrahim, Issam A. R. Moghrabi, Mahmoud M. Yahaya, Aishatu I. Ahmad, Semiu Oladipupo Oladejo, Karimov Javlon Kuzievich, Aceng Sambas . A matrix-free DFP-like optimization method for problems arising from compressive sensing. Electronic Research Archive, 2025, 33(7): 4091-4118. doi: 10.3934/era.2025183
    [6] Yiyuan Qian, Kai Zhang, Jingzhi Li, Xiaoshen Wang . Adaptive neural network surrogate model for solving the implied volatility of time-dependent American option via Bayesian inference. Electronic Research Archive, 2022, 30(6): 2335-2355. doi: 10.3934/era.2022119
    [7] Dengke Xu, Linlin Shen, Yuanyang Tangzhu, Shiqi Ke . Bayesian analysis of random effects panel interval-valued data models. Electronic Research Archive, 2025, 33(5): 3210-3224. doi: 10.3934/era.2025141
    [8] John Daugherty, Nate Kaduk, Elena Morgan, Dinh-Liem Nguyen, Peyton Snidanko, Trung Truong . On fast reconstruction of periodic structures with partial scattering data. Electronic Research Archive, 2024, 32(11): 6481-6502. doi: 10.3934/era.2024303
    [9] Zhengyu Liu, Yufei Bao, Changhai Wang, Xiaoxiao Chen, Qing Liu . A fast matrix completion method based on truncated $ {\mathit{L}}_{2, 1} $ norm minimization. Electronic Research Archive, 2024, 32(3): 2099-2119. doi: 10.3934/era.2024095
    [10] Mohamed A. Ramadan, Mahmoud A. A. Abd El-Latif, Mohammed Z. Alqarni . An integrated Laplace transform and accelerated Adomian decomposition approach for solving time-fractional nonlinear partial differential equations. Electronic Research Archive, 2025, 33(7): 4398-4434. doi: 10.3934/era.2025201
  • As a measure to counteract the effects of urban sprawl, with the continued growth of cities worldwide, different modes of urban greening are being increasingly recognized. This special issue addresses current developments in the transition to low carbon cities employing a variety of urban greening techniques. The special issue consists of 10 papers, including four review papers on the topics of biophilic architecture; environmental versus marketable aesthetics; urban agriculture; and the rationale for mainstreaming. It also contains several original research articles, some (about half of the special issue) presenting case studies, as for green redevelopment in Trenton, USA; facade greening in Genoa, Italy; climatic effects (on air temperature) in Rosario, Argentina; a modeling study for Melbourne, Australia; and another Australian case study on the greening and “un”greening of Adelaide. In addition to a broadly scoped paper that examines American stormwater management, the special issue also contains an editorial on technologies for wastewater treatment. Together, these papers constitute a contribution to recognize the importance of retaining greenery in cities chiefly, although not solely, as a countermeasure to urban sprawl and its environmental impacts. Urban greening here represents a cost-effective (soft) approach that is an effective tool as part of sustainable development.


    The inverse problems refer to exploring the inherent nature of things based on observable phenomena. There are many parameters of interest in scientific and engineering problems that cannot be directly observed. We often discuss these quantities indirectly through known data, that is, to formulate such problems in the form of inverse problems [14,7]. It can be seen from this that inverse problems are the products of the rapid development of science and engineering, which have been widely used in various fields such as geological engineering [25,10], medicine [4], environment [31], telemetry [28], control [2] and so on.

    Due to the fact that parameters are sensitive to observation data, which is usually impure, this leads to the inverse problems are ill-posed, which can be addressed with regularization [9,27]. However, regularization methods only provide point estimates of the unknown parameters and without quantifying the uncertainties of the solution. Then, the statistical inference method [14,26] entered the researchers' vision. In Bayesian inverse problems, the unknown parameters are treated as random vectors with known prior distribution, and we need to infer their conditional distribution under observation data, namely the posterior distribution. In general, it is difficult to obtain the analytical form of the posterior distribution, so we often consider using the Markov Chain Monte Carlo (MCMC) method [3] to explore the posterior space. In these sampling algorithms, the forward model solution needs to be calculated for each candidate sample. We know that evaluating forward models is expensive in many practical problems, so direct sampling algorithms are prohibitive. To overcome this difficulty, many scholars have proposed a series of surrogate models and effective sampling algorithms to reduce the computational cost of Bayesian inverse problems. The former achieves the goal by reducing the cost of a single forward model evaluation. For example, the use of generalized polynomial chaos (gPC) basis functions to express the solutions of forward problems is proposed in [21,22], the projection-based model reduction techniques can be seen in [19,15], and adaptive methods to construct surrogate models are presented in [32,17]. The latter achieves this purpose by reducing the number of samples required, which has been studied in [23,20].

    Here, we focus on the construction of effective surrogate models. Since unknown parameters are treated as random vector $ \mathit{\boldsymbol{\xi}}(\omega) $, the forward problems can be regarded as the stochastic problems on prior support $ \Gamma $. Although the stochastic problem is more complicated than the original one, we can calculate the approximate posterior probability corresponding to every candidate sample at a negligible cost once its solution expression obtained. In the physical domain $ D $ and image probability space of unknown parameters, we can use the deterministic and stochastic basis functions to represent the solutions of stochastic problems [8]. For instance, with the stochastic finite element methods, an approximation of solution $ u(\mathbf{x}, \mathit{\boldsymbol{\xi}}): D \times \Gamma \rightarrow \mathbb{R} $ has the form:

    $ ˇu(x,ξ)=ni=1nuj=1ˇcijφj(x)ψi(ξ), $ (1)

    where the coefficient $ \check{c}_{ij} $ is with respect to (w.r.t.) a finite element basis $ \varphi_j $ and a multivariate polynomial chaos (PC) basis $ \psi_i $.

    In the current work, we propose the data-driven compressive sensing method based on proper orthogonal decomposition for constructing the accurate approximate solutions of stochastic forward problems to accelerate the calculation of Bayesian inverse problems. The POD-DCS method is derived from the data-driven compressive sensing method proposed in [18]. But here, we use proper orthogonal decomposition (POD) basis functions instead of Karhunen-Loève basis, so as to generate a reduced order model (ROM) and avoid the recovery of covariance. According to the idea of compressive sensing (CS), an accurate approximate solution of the ROM can be obtained by calculating a basis pursuit (BP) problem [18,30] with a small amount of data. The advantage of our method is that it constructs the reduced order model using snapshots first, which causes the degree of freedom (DoF) of the forward problem to be reduced from the cardinality of the finite element to the that of POD, and then we only need to reconstruct the low-dimensional ROM at a lower cost. From the numerical results we can see that when using the same number of fully discrete finite element solutions, our scheme can improve accuracy and sparsity compared to the conventional CS method based on finite element basis. Moreover, the cost of evaluating a forward model using our method is only a fraction of that with the finite element method (FEM), so it can speed up the calculation of Bayesian inverse problems effectively. As we all know, many practical problems can be described by partial differential equations (PDE). Here, we concentrate only on elliptic PDE, which is widely used in the studies of oil reservoirs and groundwater [25], and its ill-posedness have been discussed in [16]. Of course, our method can be naturally extended to other partial differential equations. All the computations were performed using MATLAB R2014b on a personal computer with 1.60GHz CPU and 4GB RAM.

    The rest of this paper is organized as follows. In Section 2, we describe the model problem used as the background of our study, then introduce the framework of Bayesian inference and the stochastic surrogate model. In Section 3, we discuss how to construct the POD-DCS approximate solution for a stochastic forward problem, and provide some direct simulation results and the corresponding algorithm. The error and sparsity analyses of our scheme are conducted in Section 4. In Section 5, we compare our POD-DCS scheme with other methods, and use it to solve elliptic inverse problem. Finally, we give some conclusions and indicate possible future work in Section 6.

    Consider an underground steady state aquifers modelled by two-dimensional elliptic partial differential equation with Dirichlet boundary as

    $ {(a(x)u(x))=f,xD,u(x)=0,xD, $ (2)

    where domain $ D = [0, 1]^2 $. Let $ u(\mathbf{x}) $ be the hydraulic head, $ a(\mathbf{x}) $ be the permeability field, and apply a constant source term in the whole domain and satisfy $ f \equiv 2 $. The goal of this elliptic inverse problem is to estimate the unknown permeability field $ a(\mathbf{x}) $ from a set of observations $ \mathbf{z} = [ u(\mathbf{x}_1, \mathit{\boldsymbol{\xi}} ), \ldots, u(\mathbf{x}_{n_z}, \mathit{\boldsymbol{\xi}} )]^T \in \mathbb{R}^{n_z} $ of the solution $ u $ at points $ \{ \mathbf{x}_s \}_{s = 1}^{n_z} $. In this work, we consider the permeability field defined by $ n_p = 4 $ radial basis functions with centers $ \mathbf{x}_1, \cdots , \mathbf{x}_{n_p} $ as:

    $ a(x)=0.01+npi=1ξiexp(xxi220.02), $ (3)

    where $ \| \cdot \|_2 $ is Euclidean norm. These radial basis functions are shown in Figure 1, and what we need to do is determine these weights $ \{ \xi_i\}_{i = 1}^{n_p} $.

    Figure 1.  Radial basis functions used to define the permeability field.

    Let $ \mathit{\boldsymbol{\xi}} = [\xi_1, \cdots, \xi_{n_p}]^T $ be the weight vector, and the permeability field and model output depending on $ \mathit{\boldsymbol{\xi}} $ are denoted as $ a (\mathbf {x}; \mathit{\boldsymbol{\xi}}) $ and $ u (\mathbf{x}; \mathit{\boldsymbol{\xi}} ) $, respectively. Without loss of generality, here we consider the additive error, i.e.,

    $ zs=u(xs;ξ)+es,s=1,,nz $ (4)

    where $ \{ \mathbf{x}_s \}_{s = 1}^{n_z} $ are sensor locations, $ \mathit{\boldsymbol{\xi}} \in \mathbb{R}^{n_p} $ is the weight vector corresponding to the data $ \mathbf{z} = [z_1, \ldots, z_{n_s}]^T \in \mathbb{ R }^{n_z} $, and additive error $ \mathbf{e} = [e_1, \ldots, e_{n_z} ]^T \in \mathbb{R}^{n_z} $ comes from measurement error, etc. Then, the forward model can be written as

    $ z=G(ξ)+e, $ (5)

    where $ \mathcal{G}:\mathbb{R}^{n_p} \rightarrow \mathbb{R}^{n_z} $ is a mapping from input parameters $ \mathit{\boldsymbol{\xi}} $ to noise-free data. Let the noise $ \mathbf{e} $ and parameters $ \mathit{\boldsymbol{\xi}} $ are independent mutually. Suppose further that the components of $ \mathbf{e} $ are independent and identically distributed (i.i.d.), and satisfy $ e_i \sim N(0, \sigma_e^2) $ for $ i = 1, \cdots, n_z $.

    In Bayesian inverse problems, the unknown parameters $ \mathit{\boldsymbol{\xi}} $ is regarded as random vector in a properly defined probability space $ ( \Omega, \mathcal{F}, \mathcal{P} ) $, whose components are i.i.d.. Moreover, the prior probability density function of $ \mathit{\boldsymbol{\xi}} $ is known. In this work, we only consider the uniform prior in $ [0, 1]^{n_p} $.

    By Bayes' rule, we can infer the posterior probability density function $ \pi (\mathit{\boldsymbol{\xi}} | \mathbf{z}) $ of unknown parameters $ \mathit{\boldsymbol{\xi}} $ by combining the prior information $ \pi( \mathit{\boldsymbol{\xi}}) $, likelihood function $ \pi (\mathbf{z} | \mathit{\boldsymbol{\xi}}) $ and observations $ \mathbf{z} $, i.e.,

    $ π(ξ|z)π(z|ξ)π(ξ). $ (6)

    According the forward model (5) and the assumption of noise independence, likelihood function $ \pi (\mathbf{z} | \mathit{\boldsymbol{\xi}}) $ has the form

    $ π(z|ξ)=nzi=1πei(ziGi(ξ))exp(G(ξ)z222σ2e). $ (7)

    In many practical problems, the posterior distributions (6) are analytically intractable. Consequently, many sampling algorithms, e.g. MCMC, have been used to ascertain the posterior space. The framework of Bayesian inverse problems with direct sampling algorithm is shown in Figure 2. The form of the likelihood function (7) implies that for each candidate sample $ \mathit{\boldsymbol{\xi}} $, we need to evaluate the corresponding forward model $ \mathcal{G} (\mathit{\boldsymbol{\xi}}) $, which is expensive generally. Therefore, the naive sampling methods are prohibitive. A series of reduced order models and effective sampling algorithms have been developed to deal with this problem. In order to reduce the total computational cost, the former is by reducing the cost of a single evaluation of the forward model, and the latter is by reducing the number of samples required. In current work, we concentrate on constructing a surrogate model to reduce the cost of a single forward model evaluation.

    Figure 2.  Bayesian inverse problems framework with direct sampling algorithm.

    In probability space $ ( \Omega, \mathcal{F}, \mathcal{P} ) $, we denote the probability density function of random variable $ \xi_i( \omega ): \Omega \rightarrow \mathbb{R} $ as $ \pi_i (\xi_i) $, whose range is $ \Gamma_i = \xi_i ( \Omega ) \subset \mathbb{R} $. Because the components of $ \mathit{\boldsymbol{\xi}} $ are assumed to be independent, the joint probability density function of random vector $ \mathit{\boldsymbol{\xi}} $ has the form

    $ π(ξ)=npi=1πi(ξi) $ (8)

    with support $ \Gamma = \prod_{i = 1}^{n_p} \Gamma_i = [0, 1]^{n_p} $. Here, the joint probability density function $ \pi ( \mathit{\boldsymbol{\xi}}) $ is equivalent to the prior density function in Bayes' formula (6).

    For any $ \omega \in \Omega $, the realization of random vector $ \mathit{\boldsymbol{\xi}} $ is belong to $ \Gamma $. Therefore, in the Bayesian framework, we can restate the boundary value problem (2) as following stochastic forward problem in image probability space $ (\Gamma, \mathcal{B}(\Gamma), \pi(\mathit{\boldsymbol{\xi}})d \mathit{\boldsymbol{\xi}} ) $ rather than abstract probability space $ ( \Omega , \mathcal{F}, \mathcal{P}) $:

    $ {(a(x,ξ)u(x,ξ))=f,(x,ξ)D×Γ,u(x,ξ)=0,xD. $ (9)

    Here, $ \mathcal{B}(\Gamma) $ is the Borel $ \sigma $-algebra on $ \Gamma $, and $ \pi(\mathit{\boldsymbol{\xi}})d \mathit{\boldsymbol{\xi}} $ is the measure of random vector $ \mathit{\boldsymbol{\xi}} $. In fact, we can construct explicit expression of the solution $ u(\mathbf{x}, \mathit{\boldsymbol{\xi}}) $ to stochastic problem (9). Then for any $ \omega \in \Omega $, substituting the realization $ \mathit{\boldsymbol{\xi}}(\omega) $ into $ u(\mathbf{x}, \mathit{\boldsymbol{\xi}}) $, the function value is consistent with the solution $ u(\mathbf{x} ; \mathit{\boldsymbol{\xi}}) $ of the deterministic problem (2) which has the same input $ \mathit{\boldsymbol{\xi}} $. Using this solution expression, we can repeatedly evaluate the forward model (2) at a negligible cost. For this reason, the model (9) is called the stochastic surrogate model of problem (2).

    In this section, we propose a data-driven compressive sensing method based on proper orthogonal decomposition, which can be used to construct efficient and sparse solution for a stochastic forward model.

    It's well-known that the solution $ u(\mathbf{x}, \mathit{\boldsymbol{\xi}}): D \times \Gamma \rightarrow \mathbb{R} $ of stochastic forward model (9) can be represented by stochastic and deterministic basis functions. Here, we propose the POD-DCS method, which chooses a set of data-driven POD basis functions $ \lbrace \phi_j( \mathbf{x} )\rbrace_{j = 1}^m $ for domain $ D $ and a set of gPC basis functions $ \lbrace \psi_i ( \mathit{\boldsymbol{\xi}} ) \rbrace_{i = 1}^n $ for prior support $ \Gamma $, and the coefficients are determined by solving a BP problem. Next, we discuss how to construct the POD-DCS approximate solution for a stochastic forward problem.

    The corresponding degree of freedom of the finite element method is expressed by $ n_u $, and let $ S \in \mathbb{R}^{n_u \times K_s} $ be the snapshot matrix:

    $ S=[uh(x,ξ1),,uh(x,ξKs)]:=[uh1(x),,uhKs(x)], $ (10)

    where $ u^h (\mathbf{x}, \mathit{\boldsymbol{\xi}}_i) \in \mathbb{R}^{n_u} $ is the finite element solution of problem (9) associated with input parameter $ \mathit{\boldsymbol{\xi}}_i \in \Gamma $. These random inputs $ \{ \mathit{\boldsymbol{\xi}}_i \}_{i = 1}^{K_s} $ are sampling from $ \Gamma $ randomly. By solving eigenvalue problem

    $ RV=VΛ $ (11)

    with

    $ Rij=1Ks(uhi(x),uhj(x)),i,j=1,,Ks, $ (12)

    we can obtain the diagonal singular value matrix $ \Lambda $ with $ \Lambda_{jj} = \lambda_j $, and the matrix of singular vectors $ V = [\mathbf{v}_1, \cdots, \mathbf{v}_{K_s}] $ with the $ j $-th component

    $ vj=[v(j)1,,v(j)Ks]T,j=1,Ks. $ (13)

    Since $ R $ is positive semidefinite and symmetric, we can make the eigenvectors $ \{ \mathbf{v}_j\}_{j = 1}^{K_s} $ to be orthogonal and sort the eigenvalues in a descending order. After normalization, the POD basis can be represented as

    $ ϕj(x)=1KsλjKsi=1v(j)iuhi(x),j=1,,m. $ (14)

    Here, the dimension of POD basis, $ m $, is determined by the energy ratio $ \nu_m $ defined as

    $ νm=mj=1λj/Ksj=1λj. $ (15)

    Once the basis functions are computed, the solution $ u ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) $ can be approximated as

    $ ˆu(x,ξ)=mj=1αj(ξ)ϕj(x). $ (16)

    The coefficients $ \{ \alpha_j \}_{j = 1}^m $ are called reduced states and satisfy the following algebraic system

    $ Aα=F, $ (17)

    where

    $ α=[α1(ξ),,αm(ξ)]T,Fi=Df(x)ϕi(x)dx,i=1,,m,Aij=Da(x,ξ)ϕj(x)ϕi(x)dx,i,j=1,,m. $

    From the definition of the POD basis functions we known that $ \phi_i(\mathbf{x}) $ and $ \phi_j( \mathbf{x} ) $ are orthogonal for $ i, j = 1, \cdots, m $, and $ \{ \phi_j( \mathbf{x}) \}_{j = 1}^{m} $ minimize the error measure

    $ M=1KsKsi=1uhi(x)mj=1(uhi(x),ϕj(x))ϕj(x)2L2(D) $ (18)

    with minimum $ \sum_{j = m+1}^{K_s} \lambda_j $ [12]. When $ K_s $ is large enough, $ \mathcal{M} $ can be regarded as an approximation of the mean square error in the $ L ^ 2 (D) $-norm

    $ ¯M=E[uh(x,ξ)mj=1(uh(x,ξ),ϕj(x))ϕj(x)2L2(D)]. $ (19)

    For 4-dimensional stochastic problem (9), we draw $ K_s $ samples of input parameters $ \mathit{\boldsymbol{\xi}} $ from prior support $ \Gamma $ randomly, and evaluate the corresponding forward model (9) by using finite element method with mesh size $ h = 2^{-6} $ (i.e. $ n_u = 4225 $). These solutions compose the snapshot matrix $ S $. For $ K_{s} = 50, 100, 200 $ and 300, the first 15 eigenvalues of correlation matrix $ R $ are shown on the left of Figure 3. We can observe that the eigenvalues have very little differences when $ K_s \geq 100 $, but have large gaps between $ K_s = 50 $ and $ K_s = 100 $.

    Figure 3.  (Left): The first 15 eigenvalues of correlation matrix $ R $ for $ K_s = 50, 100, 200 $ and 300. (Right): The cumulative energy ratios of the first 15 POD basis functions with $ K_s = 100, 200 $ and 300.

    In order to explore the prior space fully and ensure the accuracy of the POD-based reduced order model, we need use $ K_s \geq 100 $ realizations to construct the snapshot matrix. The right side of the Figure 3 shows the cumulative energy ratios of the first 15 POD basis functions with $ K_s = 100, 200 $ and 300. We can see that the energy ratios of first 2 POD basis functions have slight differences for different $ K_s $. For each case, the first 3 basis functions can capture 99% energy of snapshot set, and the cumulative energy ratio of the first 9 basis functions is very close to 1.

    We define the expectation and variance of $ L^2(D) $-norm error between the full finite element solution $ u^h ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) $ and the POD-based approximate solution $ \widehat{ u} ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) $ as

    $ ˆE=E[uhˆu2L2(D)], $ (20)

    and

    $ ˆV=Var[uhˆu2L2(D)], $ (21)

    where $ \mathbb{E}[\cdot] $ means the expectation and $ Var [\cdot] $ represents the variance. Here, we approximate these statistics with $ K_r = 500 $ samples.

    Figure 4 displays these error estimates of the POD-based approximate solution associated with different $ K_s $ and different degrees of freedom $ m $. We observe that for a given $ m $, the expectations and variances in these three cases only have a little differences. And when $ m \leq 4 $, the errors decrease very quickly, while they are very small variations with $ m\geq 9 $. The results for $ m = 3, 6, 9 $ and 12 are shown in Table 1. It is clear that the change in energy from 9 to 12 POD basis functions is so small that it is almost negligible, and the improvements in the corresponding error statistics are also small. Moreover, for $ m = 9 $, the order of magnitude of the differences in $ \widehat{ \mathcal{E}} $ is -4, and in variance $ \widehat{ \mathcal{V}} $ is -5 between these three cases, while they are larger for $ m = 3 $ and $ m = 6 $. Therefore, we consider using $ K_s = 100 $ realizations to generate the POD basis functions and choose $ m $ to be 9 in the next experiment to ensure accuracy without consuming too much computing resources.

    Figure 4.  The statistics estimation of $ L^2(D) $-norm error between the full finite element solution and the POD-based ROM solution with $ K_s = 100, 200 $ and 300. (Left): Expectation. (Right): Variance.
    Table 1.  The statistics estimation of $ L^2(D) $-norm error between the finite element solution and the POD-based approximate solution with $ K_s = 100, 200, 300 $ and $ m = 3, 6, 9, 12 $.
    $ m $ 3 6 9 12
    $ K_s = 100 $ $ \nu_m $ 0.9906 0.9988 0.9998 0.9999
    $ \widehat{ \mathcal{E}} \times 10^{-2} $ 5.2424 0.9679 0.2840 0.1267
    $ \widehat{ \mathcal{V}} \times 10^{-3} $ 8.1367 0.8915 0.1475 0.0282
    $ K_s = 200 $ $ \nu_m $ 0.9908 0.9988 0.9998 0.9999
    $ \widehat{ \mathcal{E}} \times 10^{-2} $ 4.9558 1.0836 0.2422 0.0907
    $ \widehat{ \mathcal{V}} \times 10^{-3} $ 6.7935 0.9507 0.1023 0.0126
    $ K_s = 300 $ $ \nu_m $ 0.9900 0.9986 0.9997 0.9999
    $ \widehat{ \mathcal{E}} \times 10^{-2} $ 4.6537 0.8260 0.2229 0.0622
    $ \widehat{ \mathcal{V}} \times 10^{-3} $ 5.5292 0.4418 0.0898 0.0037

     | Show Table
    DownLoad: CSV

    In our POD-DCS algorithm, using the POD method first can greatly reduce the DoF of the problem in hand, and simplify the subsequent processing. Next, we utilize the idea of compressive sensing to express the reduced states $ \{ \alpha_j ( \mathit{\boldsymbol{\xi}}) \}_{j = 1}^m $ with gPC basis functions, and determine the expression coefficients by using a small amount of data and solving a $ l_1 $-minimization problem.

    Clearly, once the POD basis functions $ \left\{\phi_{j} ( \mathbf{ x}) \right\}_{j = 1}^{ m } $ are obtained, we only need to compute the reduced states $ \alpha_j (\mathit{\boldsymbol{\xi}}) : \Gamma \rightarrow \mathbb{R} $ in the expression (16) for $ j = 1, \cdots, m $. Here, we use the $ N $th-order generalized polynomial chaos to represent the reduced states as the form

    $ αj(ξ)ni=1cijψi(ξ),j=1,,m, $ (22)

    where $ \{\psi_i(\mathit{\boldsymbol{\xi}})\}_{i = 1}^n $ represents a set of multivariate Legendre polynomials with order $ N $ that are orthonormal w.r.t. the joint probability density function $ \pi ( \mathit{\boldsymbol{\xi}}) $, and $ \psi_1 ( \mathit{\boldsymbol{\xi}} ) = 1 $. The cardinality of the stochastic polynomial basis is

    $ n=(N+np)!N!np!. $ (23)

    The stochastic collocation method [29] based on polynomial interpolation can be utilized to determine the coefficients $ \{c_{ij}\}_{i = 1, j = 1}^{ n , \quad m} $. With collocation points $ \{ \mathit{\boldsymbol{\xi}}_i \}_{i = 1}^{K_{c}} $, we have a linear system

    $ (α1(ξ1)αm(ξ1)α1(ξKc)αm(ξKc))=(ψ1(ξ1)ψn(ξ1)ψ1(ξKc)ψn(ξKc))(c11c1mcn1cnm). $

    Simply expressed as

    $ ˆA=Ψc, $ (24)

    where $ \widehat{ A } = \left( \alpha_j ( \mathit{\boldsymbol{\xi}}_i ) \right) \in \mathbb{R}^{ {K_{c}} \times m} $ is the matrix of reduced state generated by solving system (17), $ \mathbf{c} = (c_{ij}) \in \mathbb{R}^{n \times m} $ is the coefficient matrix that needs to be determined, $ \Psi = (\psi_i ( \mathit{\boldsymbol{\xi_j}} )) \in \mathbb{R}^{ K_{c} \times n } $ represents the stochastic information matrix, and $ K_{c} $ is the number of stochastic collocation points. As shown in Equation (23), with the increase of random input dimension $ n_p $, the cardinality $ n $ of polynomial basis functions grows very quickly, which will cause linear system (24) to be underdetermined, i.e., $ K_c <n $.

    Based on the idea of compressive sensing, given a highly incomplete set of reduced states by solving algebraic system (17), an accurate approximate solution of linear system (24) can be obtained by solving the BP problem

    $ vec(˜c)=argminvec(c)vec(c)1,subject tovec(ˆA)=Θvec(c), $ (25)

    where dictionary matrix $ \Theta = I_m \otimes \Psi $ is the Kronecker product of an $ m \times m $ identity matrix $ I_m $ with stochastic information matrix $ \Psi $, $ vec ( \widehat{A} ) $ denotes the vectorization of matrix $ \widehat{A} $ by column, and $ \widetilde{ \mathbf{c} } = (\widetilde{c}_{ij}) $ represents the coefficient matrix we calculated. Then the reduced states can be approximated by

    $ ˜αj(ξ)=ni=1˜cijψi(ξ),j=1,,m. $ (26)

    Combining equations (16) and (26), we can get the POD-DCS approximate solution of stochastic elliptic problem (9) as

    $ ˜u(x,ξ)=ni=1mj=1˜cijψi(ξ)ϕj(x). $ (27)

    Remark 1. By using the orthogonality of multivariate Legendre polynomials, we have

    $ E[˜u(x,ξ)]=mj=1˜c1jϕj(x), $ (28)
    $ E[˜u2(x,ξ)]=ni=1(mj=1˜cijϕj(x))(mj=1˜cijϕj(x)). $ (29)

    The details of our POD-DCS method for stochastic forward problem is presented as following Algorithm 1.

    Algorithm 1 The POD-DCS algorithm for stochastic forward problem
    Input: Grid parameter $ h $, positive integer $ K_{s} $, energy ratio $ \nu $, collocation points $ \{ \mathit{\boldsymbol{\xi}}_i \}_{i=1}^{K_{c}} $, the order of polynomial $ N $.
    Output: POD basis function $ \left\{\phi_{j} ( \mathbf{x} )\right\}_{j=1}^{ m } $, and the coefficient matrix $ \widetilde{ \mathbf{c} } $.
    1: Draw a set of random inputs $ \{ \mathit{\boldsymbol{\xi}}_k \}_{k=1}^{K_{s} } $ from prior support $ \Gamma $ independently, and compute the corresponding solutions $ \{u^h ( \mathbf{x}, \mathit{\boldsymbol{\xi}}_k ) \}_{k=1}^{K_{ s }} $ of problem (9) by using FEM with mesh size $ h $, then compose a snapshot matrix $ S $.
    2: Generate the POD basis functions $ \left\{\phi_{ j }( \mathbf{x} ) \right\}_{j=1}^{ m } $ as (14) by solving eigenvalue problem (11), where $ m $ is the smallest positive integer such that $ \nu_m \geq \nu $ holds.
    3: Construct the reduced state matrix $ \widehat{A} $ as (24) by solving algebraic system (17) with collocation points $ \{ \mathit{\boldsymbol{\xi}}_i \}_{i=1}^{K_{c}} $.
    4: Select appropriate stochastic basis functions $ \{\psi_i (\mathit{\boldsymbol{\xi}}) \}_{i=1}^n $ with order $ N $.
    5: Obtain coefficient matrix $ \widetilde{ \mathbf{c} } $ by solving BP problem (25).
    6: Generate the approximate reduced states $ ( \widetilde{ \alpha}_j( \mathit{\boldsymbol{\xi}}) ) $ and the POD-DCS approximate solution $ \widetilde{ u } ( \mathbf{ x}, \mathit{\boldsymbol{\xi}}) $ of system (9) as
    $ ˜αj(ξ)=ni=1˜cijψi(ξ),j=1,,m,˜u(x,ξ)=ni=1mj=1˜cijψi(ξ)ϕj(x). $

     | Show Table
    DownLoad: CSV

    Table 1 illustrates that using 100 realizations to generate the POD-based approximate solution with dimension 9 can guarantee the accuracy of the model. Thus, we utilize 100 snapshots to construct ROM, and retain only the first 9 basis functions. In this case, $ \widehat{ \mathcal{E}} = 2.8400 \times 10^{-3} $ and $ \widehat{ \mathcal{V}} = 1.4750 \times 10^{-4} $.

    Now, we need to determine the coefficients $ \{ \widetilde{c}_{ij} \}_{i = 1, j = 1}^{n, \quad m} $ in the expression (26) by solving $ l_1 $-minimization problem (25). Here, the 7th-order Legendre PC basis functions $ \psi( \mathit{\boldsymbol{\xi}}) $ are used to represent the reduced states $ \alpha_j (\mathit{\boldsymbol{\xi}}) $ for $ j = 1, \cdots, m $, and the total DoF of matirx $ \widetilde{ \mathbf{c} } $ is $ n\times m = 2970 $. Without loss of generality, we only consider the coefficients with absolute values larger than a fixed threshold $ \tau $, i.e., set the entries of $ \widetilde{ \mathbf{c} } $ whose absolute value smaller than the threshold $ \tau $ to 0, and let $ \widetilde{ \mathcal{R}}_{\tau} $ be the proportion of non-zero coefficients in matrix $ \widetilde{ \mathbf{c}} $, that is

    $ ˜Rτ=#{|˜c|>τ}#{˜c}. $ (30)

    Like the definitions of $ \widehat{ \mathcal{E}} $ and $ \widehat{ \mathcal{V}} $, we denote the expectation and variance of $ L^2(D) $-norm error between the full finite element solution $ u^h (\mathbf{x}, \mathit{\boldsymbol{\xi}}) $ and the POD-DCS approximate solution $ \widetilde{u} (\mathbf{x}, \mathit{\boldsymbol{\xi}}) $ as $ \widetilde{ \mathcal{E}} $ and $ \widetilde{ \mathcal{V}} $ respectively, namely

    $ ˜E=E[uh˜u2L2(D)], $ (31)

    and

    $ ˜V=Var[uh˜u2L2(D)]. $ (32)

    The sparsity and error estimation of the POD-DCS method w.r.t. different number of collocation points sampled from the prior space randomly and different thresholds are drawn in Figure 5. Compared with $ \tau = 0 $, for each $ K_c $, the ratio of non-zero coefficients is less than $ 5\% $ with $ \tau = 0.1 $, but the mean of the $ L^2(D) $-norm error is much larger. This is due to the fact that the threshold is so large that some influential basis functions are ignored. For $ \tau = 0.01 $, the highest proportion of non-zero coefficients in matrix $ \widetilde{ \mathbf{c}} $ is about $ 33\% $, and the error statistics are almost consistent with those of $ \tau = 0 $. Therefore, we can take the threshold $ \tau = 0.01 $ to ensure the accuracy. Through regression, the expectation of our approach converges with order 1.58 and variance converges with order 2.97, that is, $ \widetilde{\mathcal{E}} \approx \mathcal{O}( K_c^{-3/2}) $ and $ \widetilde{\mathcal{V}} \approx \mathcal{O}( K_c^{-3}) $, which can be seen in Figure 6. Such a highly "incomplete" set of reduced states does contain enough information to accurately recover the reduced order model. When $ K_c \geq 200 $, the expectations $ \widetilde{ \mathcal{E}} $ and variances $ \widetilde{ \mathcal{V}} $ of $ L^2(D) $-norm error do not change much as $ K_c $ increase. So here we use 200 collocation points to determine the coefficients $ (\widetilde{ c}_{ij} ) $ of expression (27), that is, we need to calculate 200 algebraic systems (17). At this time, $ \widetilde{ \mathcal{R}}_{0.01} = 20.24\% $, $ \widetilde{ \mathcal{E}} = 4.1758 \times 10^{-3} $, and $ \widetilde{ \mathcal{V}} = 2.4931 \times 10^{-4} $

    Figure 5.  Sparsity and error estimations of the POD-DCS method with different $ K_c $ and different threshold $ \tau $. (Left): The proportion $ \widetilde{ \mathcal{R}}_{\tau} $ of non-zero coefficients in matrix $ \widetilde{ \mathbf{ c}} $; (Middle): The expectation $ \widetilde{ \mathcal{E}} $ of $ L^2(D) $-norm error; (Right): The variance $ \widetilde{ \mathcal{ V }} $ of $ L^2(D) $-norm error.

    The POD-DCS method has been described in the previous section. In order to introduce the conclusions of error analysis and coefficients estimation of our scheme, we first introduce several relevant properties of the compressive sensing method in this section.

    Definition 4.1 (see [11]). A vector $ \mathbf{ b } \in \mathbb{R}^n $ is called $ k $-sparse if $ \| \mathbf{ b} \|_0 \leq k $ holds, and its best $ k $-term approximation error in the $ l_p $-norm is defined as

    $ σk,p(b)=infx0kxbp, $ (33)

    where $ \| \mathbf{ b} \|_0 $ is the number of non-zero terms in vector $ \mathbf{b} $.

    Note that for $ 0<q<p \leq \infty $ and set $ s = 1/q-1/p>0 $, the prior estimation holds:

    $ σk,p(b)ksbq. $ (34)

    In order to ensure that the matrix $ \mathbf{c} $ can be reconstructed exactly by using $ l_1 $ minimization, we introduce the restricted isometry property (RIP).

    Lemma 4.2 (see [5,24]). If for any $ k $-sparse vector $ \mathbf{b} \in \mathbb{ R }^n $, there exists a constant $ \delta \in (0, 1) $ such that the inequality

    $ (1δ)b22Ab22(1+δ)b22 $ (35)

    holds. Then $ \delta_{k} ( A ) : = \min(\delta) $ is called the restricted isometry constant (RIC) of matrix $ A \in \mathbb{R}^{m \times n} $, and the matrix $ A $ is said to satisfy the RIP of order $ k $ with RIC $ \delta_{k}( A ) $. Similarly, denoting the RIC of matrix $ B \in \mathbb{R}^{p \times q} $ by $ \delta_{ k } ( B) $, we can obtain the inequality

    $ max{δk(A),δk(B)}δk(AB)δk(A)+δk(B)+δk(A)δk(B). $ (36)

    Based on Lemma 4.2 we know that the RIC of dictionary matrix $ \Theta $ in (25) satisfies $ \delta_k ( \Theta) = \delta_k ( \Psi) $.

    Lemma 4.3 (see [11]). Assuming that $ \delta_{3k} (\Theta) <1/3 $, then the solution of $ l_1 $-minimization problem (25) satisfies

    $ vec(c)vec(˜c)2Cδσk,1(vec(c))k $ (37)

    where the constant $ C_{\delta} >0 $ depends only on $ \delta_{3k} (\Theta) $.

    By Lemma 4.3, we give the error estimation of our POD-DCS method as following.

    Theorem 4.4. There exist constants $ C_1, C_2, C_3, C_4, \theta>0 $ and $ 0<q<1 $, assuming that the dictionary matrix $ \Theta $ in (25) satisfies $ \delta_{3k} (\Theta) <1/3 $. Then with probability close to one, the expectation (31) of $ L^2(D) $-norm error between the finite element solution and the POD-DCS approximate solution satisfies

    $ ˜EC1ˆVKs+C2Ksj=m+1λj+C3Nθ+C4k12qvec(c)2q, $ (38)

    where $ C_3, C_4 $ depend on the smoothness of the reduced states and $ \delta_k (\Theta) $ respectively, while constants $ C_1, C_2 $ are universal.

    Proof. By using the inequality $ 2ab \leq a^2+b^2 $, the error (31) follows that

    $ ˜E=E[uhˆu+ˆu˜u2L2(D)]2{E[uhˆu2L2(D)]+E[ˆu˜u2L2(D)]}. $

    Denote

    $ I1=E[uhˆu2L2(D)],I2=E[ˆu˜u2L2(D)]. $

    We first estimate $ I_1 $. The standard Monte Carlo Finite Element Method (MCFEM) approximates the expectation function $ I_1 $ by the average of samples. Note that the random inputs $ \{ \mathit{\boldsymbol{\xi}}_j \}_{j = 1}^{K_{s}} $ used to generate the snapshot set are sampling from prior support $ \Gamma $ randomly. Thus, we can use these samples to approximate $ I_1 $ as

    $ ˉI1=1KsKsj=1uhjˆuj2L2(D). $

    where $ u^h_j $ and $ \widehat{u}_j $ represent the finite element solution and POD-based solution with random input $ \mathit{\boldsymbol{\xi}}_j $ for $ j = 1, \cdots, K_s $, respectively. The number of realizations, $ K_{s} $, controls the statistical error $ \mathcal{E}_S = I_1 - \bar{ I }_1 $ [1], and

    $ ES=E[uhˆu2L2(D)]1KsKsj=1uhjˆuj2L2(D)=Var[uhˆu2L2(D)]Ks×KsE[uhˆu2L2(D)]Ksj=1uhjˆuj2L2(D)KsVar[uhˆu2L2(D)]. $

    According to the central limit theorem we have

    $ ESN(0,ˆV/Ks), $

    then choose a constant $ C_q \geq 1.65 $, such that

    $ |ES|CqˆV/Ks, $

    holds with probability close to one, where $ C_q $ is called quantile. With the error measure (18), $ \bar{I}_1 $ satisfies

    $ ˉI1=1KsKsj=1uhjˆuj2L2(D)=Ksj=m+1λj. $

    As (22), we use multivariate polynomials with order $ N $ to approximate reduced state $ \alpha_j( \mathit{\boldsymbol{\xi}}) $ for $ j = 1, \cdots, m $, then there exist constants $ C_N, \theta >0 $ such that

    $ Γ(αj(ξ)ni=1cijψi(ξ))2π(ξ)dξCNNθ,j=1,,m, $

    holds. The constant $ C_N $ depends on the smoothness of $ \{ \alpha_j (\mathit{\boldsymbol{\xi}}) \}_{j = 1}^m $. Using the biorthogonality of the POD-DCS solution and Lemma 4.3, we have

    $ I2=ΓD(mj=1(αj(ξ)˜αj(ξ))ϕj(x))2π(ξ)dxdξ=mj=1Γ(αj(ξ)˜αj(ξ))2π(ξ)dξ=mj=1Γ(αj(ξ)ni=1cijψi(ξ)+ni=1(cij˜cij)ψi(ξ))2π(ξ)dξ2mj=1Γ(αj(ξ)ni=1cijψi(ξ))2π(ξ)dξ+2mj=1ni=1(cij˜cij)2 $
    $ 2mCNNθ+2C2δσ2k,1(vec(c))k. $

    By the prior estimation (34), we can arrive the error estimate with $ 0<q<1 $

    $ ˜E2(CqˆVKs+Ksj=m+1λj+2mCNNθ+2C2δσ2k,1(vec(c))k)C1ˆVKs+C2Ksj=m+1λj+C3Nθ+C4k12qvec(c)2q, $

    which completes the proof.

    This theorem implies that the mean square error $ \widetilde{ \mathcal{E}} $ consists of four parts, including statistical error, truncation error, polynomial approximation error and sparse reconstruction error. Among them, the first two items are derived from the POD-based reduced order model, and the last two items are caused by the sparse reconstruction of the corresponding reduced order model. Therefore, we can balance these four errors and discrete error to save computing resources for a desired accuracy.

    The error analysis has been completed and we are now ready to show the sparsity of our POD-DCS solution.

    Theorem 4.5. The coefficient matrix $ \widetilde{ \mathbf{c} } $ in the POD-DCS solution expression (27) satisfies the estimation

    $ ni=1˜c2ij=λj,j=1,m. $ (39)

    Proof. According to the eigenvalue problem (11) and the POD basis function (14) we obtain

    $ λjϕj(x)=1KsλjKsi=1λjv(j)iuhi(x)=1KsλjKsi=1(1KsKsr=1(uhi(x),uhr(x))v(j)ruhi(x))=1KsKsi=1(uhi(x),ϕj(x))uhi(x). $

    Thus, we know that $ \lambda_j \phi_j (\mathbf{x}) $ can be regarded as an approximation of the expectation of $ \left( u^h ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) , \phi_j ( \mathbf{x}) \right) u^h ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) $. Using the POD-DCS solution to approximate finite element solution, and combining the biorthogonality of the POD-DCS solution, $ \lambda_j \phi_j ( \mathbf{x}) $ can be approximated by

    $ E[(˜u(x,ξ),ϕj(x))˜u(x,ξ)]=Γ(Dni=1mk=1˜cikψi(ξ)ϕk(x)ϕj(x)dx)ni=1mk=1˜cikψi(ξ)ϕk(x)π(ξ)dξ $
    $ =Γ(ni=1˜cijψi(ξ))ni=1mk=1˜cikψi(ξ)ϕk(x)π(ξ)dξ=mk=1(ni=1˜cij˜cik)ϕk(x), $

    which implies that (39) holds true.

    Figure 7 confirms the conclusion of Theorem 4.5 numerically. The error in this figure is due to the fact that we only use the average of 100 samples to approximate the expectation, and the error of POD-DCS solution. It is well-known that the eigenvalues decay rapidly in many practical problems, so Theorem 4.5 implies that the coefficients are compressible and illustrates the feasibility of using the POD-based ROM first.

    Figure 6.  Error estimations of the POD-DCS method w.r.t. different collocation point sizes and $ \tau = 0.01 $.
    Figure 7.  The eigenvalues and coefficients in the POD-DCS method with $ K_s = 100 $, $ m = 9 $ and $ K_{c} = 200 $.

    In this section, we compare the POD-DCS scheme with POD-based ROM and the conventional CS method, and use our method to solve the 4-dimensional elliptic inverse problem (2)-(3) to further describe its feasibility and advantages.

    The error estimates and sparsity of our POD-DCS scheme and other two different methods are shown in Table 2. All three methods are constructed with 100 full discrete finite element solutions. Among them, the POD-DCS method is constructed as above discussion with $ K_s = 100, m = 9, K_c = 200 $, POD-based ROM has same dimension $ m $, and CS represents the conventional compressive sensing method.

    Table 2.  Error estimates and sparsity for different methods.
    POD-DCS POD CS
    $ \mathcal{E} \times 10^{-3} $ 4.1758 2.8399 7.3896
    $ \mathcal{V} \times 10^{-4} $ 2.4931 1.4750 11.1484
    $ \mathcal{R}_{0.01} $ 0.2024 - 0.2182

     | Show Table
    DownLoad: CSV

    Obviously, the POD method has highest accuracy, while the reconstruction error of reduced states makes the accuracy of our POD-DCS method to be inferior to POD, and relatively little information leads to the worst accuracy of the conventional CS method. In terms of sparsity, the proposed POD-DCS scheme is slightly better than the CS method. However, the total DoF of our method is $ 0.2024 \times n \times m = 601 $, which is only $ 0.2\% $ of that in CS method, and its DoF is $ 0.2182 \times n \times n_u \approx 3.04 \times 10^{5} $. The reason for the sharply decrease in the DoF is due to the fact that we use the reduced order model, which leads to the DoF in physical space decline from $ n_u $ to $ m $. Therefore, compared to the CS method, the proposed method can not only improve the accuracy and sparsity, but also greatly reduce the degree of freedom.

    From the discussions in Section 3 and Theorem 4.4 we know that the accuracy of the POD-DCS method can be improved by increasing the number of snapshots, the dimension of POD basis, the number of collocation points and the order of polynomials. These quantities can be determined with the required accuracy.

    Here we compare the efficiency of different methods. Table 3 summarizes the cost of each stage in the construction of different models. From the perspective of model construction, the number of full finite element solutions used in the three methods is the same. For POD-based ROM, although it has high accuracy and does not need to solve the BP problem, the time required to evaluate the model once is about 5 times that of the other two methods, which is not conducive to the implementation of the sampling algorithm. It takes only 0.2095s to evaluate the solution expression obtained by CS method, but the size of the dictionary matrix $ \Theta $ in the corresponding BP problem (25) is $ (100 \times n_u) \times ( n \times n_u) $, which is usually so large that it will exceed the computing resources or the calculation speed will be very slow. According to the nature of the finite element basis functions, here we can decompose the BP problem into solving the each column of the coefficient matrix $ \mathbf{c} $, that is, transform the problem into solving $ n_u $ $ l_1 $-minimization problems with size of $ 100 \times n $. This process takes about 1.4 hours. Since the sparse reconstruction in our POD-DCS scheme is just for reduced states, and the size of dictionary matrix is $ (m \times K_c) \times (m \times n) $. It takes only 21s to solve this BP problem, which is 0.42$ \% $ of the time required by the CS method. For the POD-DCS method, we also need to evaluate the reduced order models w.r.t. $ K_c $ collocation points, but it is usually much cheaper than the full finite element method.

    Table 3.  Computational times, in seconds, given by different methods.
    $ \# $ of full FE solution Time for per FE solution $ \# $ of POD solution Time for per POD solution Time for BP Time for per model output
    POD-DCS 100 2.1970 200 1.0085 21 0.2002
    POD 100 2.1970 - - - 1.0085
    CS 100 2.1970 - - 5045 0.2095

     | Show Table
    DownLoad: CSV

    Therefore, from an efficiency perspective, the offline cost of POD-based ROM is relatively small, but the online cost is larger than the other two methods. While compared with the CS method, the online time of the POD-DCS method only has a little difference, but the offline time has a obvious advantage. Moreover, our method achieves 11 times acceleration when evaluating a forward model, and the time required to construct the solution expression of stochastic surrogate model can be offset by the repeated calculation of the forward model. It is well known that both POD and stochastic collocation methods can deal with highly nonlinear problems, so for such complex problems, our method will be more attractive due to its high efficiency.

    The accuracy and efficiency comparison of the POD-DCS method with other two methods has been completed. Now, we utilize this method to deal with the elliptic inverse problem (2)-(3). We use finite element method with mesh size $ h = 2^{-8} $ to generated noise-free data associated with input $ \mathit{\boldsymbol{\xi}}_o = [0.3, 0.3, 0.3, 0.3]^T $. Note that the mesh size is finer than that used in inversion in order to prevent the "inverse crime". The true permeability field $ a(\mathbf{x} ; \mathit{\boldsymbol{\xi}}_o) $ and noise-free output are shown in Figure 8. The $ n_z = 49 $ measurement sensors are uniformly distributed over $ D $ with grid spacing $ 2^{-3} $.

    Figure 8.  (Left): The true permeability field $ a(\mathbf{x}) $ used for generating the synthesize data $ \mathbf{z} $; (Right): The model outputs associated with true permeability field, where black dots are the measurement sensors.

    In this Bayesian inverse problem, the components of weight vector $ \mathit{\boldsymbol{\xi}} $ are assumed to be i.i.d. and have a prior $ \xi_i \sim U(0, 1) $ for $ i = 1, \cdots , n_p $, where $ n_p = 4 $. As described in section 3, we first construct the POD-DCS approximate solution $ \widetilde{ u} ( \mathbf{x}, \mathit{\boldsymbol{\xi}}) $ of the stochastic problem (9) in prior support $ \Gamma = [0, 1]^{n_p} $ with $ K_s = 100, m = 9, K_c = 200 $ and $ \tau = 0.01 $. Then substituting this approximate solution expression into (7) to get the approximate likelihood function, thereby obtaining the approximate posterior probability density function from Bayes' rule (6). The framework for solving the Bayesian inverse problems using the POD-DCS approximate solution of the stochastic surrogate model is shown in Figure 9. Denote the approximate likelihood function and the approximate posterior probability density function associated with POD-DCS solution as $ \widetilde{ \pi} (\mathbf{z}| \mathit{\boldsymbol{\xi}}) $ and $ \widetilde{ \pi} (\mathit{\boldsymbol{\xi}}| \mathbf{z} ) $ respectively. The Hellinger distance and Kullback-Leibler distance between the exact posterior distribution $ \pi (\mathit{\boldsymbol{\xi}}| \mathbf{z} ) $ and its approximation $ \widetilde{ \pi} (\mathit{\boldsymbol{\xi}}| \mathbf{z} ) $ have been shown in [26] and [30]. Here, we use delayed-rejection adaptive Metropolis (DRAM) algorithm to explore the approximate posterior distribution $ \widetilde{ \pi} (\mathit{\boldsymbol{\xi}}| \mathbf{z} ) $, whose computational cost is only a fraction of that in exploring the true posterior distribution $ \pi (\mathit{\boldsymbol{\xi}}| \mathbf{z} ) $. This is because the cost of evaluating a finite element solution is about 11 times that of a POD-DCS solution, which can be observed from Table 3.

    Figure 9.  Bayesian inverse problems framework with POD-DCS approximate solution.

    Here we consider that the observation data $ \mathbf{z} $ contains noise $ \mathbf{e} $, which satisfies zero-mean Gaussian distribution $ N( \mathbf{0}, \sigma_e^2 I_{n_z} ) $ with $ \sigma_e = 0.05 $. For the finite element method, conventional CS method and POD-DCS method, we use the DRAM algorithm to draw $ 3 \times 10^4 $ samples respectively. Among them, the first 5000 samples are discarded as burn-in periods and the last 25000 samples are used to estimate the posterior densities. The 95% confidence interval and posterior marginal probability density of each component of unknown input parameters $ \mathit{\boldsymbol{\xi}} = [\xi_1, \xi_2, \xi_3, \xi_4] $ are shown in Table 4 and Figures 10. Obviously, the 95% confidence intervals w.r.t. all these three methods contain true input parameters $ \mathit{\boldsymbol{\xi}}_o $, and the length of these confidence intervals does not exceed 0.03. But compared with the conventional CS method, the posterior marginal densities obtained by our scheme are closer to that obtained by FEM. In Figure 11, we display the estimated permeability field and its error, which correspond to the maximum a posterior probability (MAP) parameters generated by the POD-DCS method. We can see that the error between the estimated permeability field and the true permeability field is small enough. Therefore, the proposed method can deal with Bayesian inverse problems efficiently.

    Table 4.  The 95% confidence intervals of each input parameter obtained by different methods.
    FEM POD-DCS CS
    $ \xi_1 $ [0.2898, 0.3173] [0.2883, 0.3164] [0.2963, 0.3248]
    $ \xi_2 $ [0.2858, 0.3134] [0.2844, 0.3126] [0.2913, 0.3175]
    $ \xi_3 $ [0.2913, 0.3191] [0.2927, 0.3200] [0.2906, 0.3208]
    $ \xi_4 $ [0.2750, 0.3016] [0.2762, 0.3047] [0.2736, 0.3022]

     | Show Table
    DownLoad: CSV
    Figure 10.  The posterior marginal densities of unknown parameters with $ \sigma_e = 0.05 $ noise in the observations.
    Figure 11.  The estimated permeability field and error corresponding to the MAP parameters generated by the POD-DCS method. (Left): Estimated permeability field; (Right): Error.

    From the previous discussion, it is clear that the POD-DCS surrogate model speeds up the evaluation speed of model while ensuring the accuracy. Therefore, this method can be considered for Bayesian inverse problem, optimal control and other problems requiring repeated evaluation of forward model. Note that the BP problem (25) is solvable in polynomial time [6]. While the dictionary matrix $ \Theta $ we used depends on the numbers of POD bases and stochastic bases, i.e. $ m $ and $ n $. Therefore, for a specific problem, the dimension of POD-based ROM and the degree of the polynomial interpolation should be selected appropriately. It is further known that this method has limitations for the problem whose eigenvalue decays slowly, which is the inherent shortcoming of the POD method.

    In summary, for statistical inverse problems, we can regard the deterministic forward problem as a stochastic forward problem on the prior support of unknown parameters, and the solutions of these two problems with the same input are equal. Therefore, in this work, we propose a data-driven compressive sensing method based on proper orthogonal decomposition to construct the solution expression of stochastic surrogate model to accelerate the Bayesian inference of an inverse problem. The snapshot-based POD method is first used to construct the ROM of stochastic problem, then the stochastic collocation method based on gPC basis functions is adopted to represent the reduced states, and the coefficients are determined by solving an $ l_1 $-minimization problem. Substituting this approximate solution into the likelihood function can obtain an approximate likelihood function, thereby obtaining an approximate posterior. The approximate posterior space can be explored faster than the original one. We make error analysis and coefficient estimation for this method, and prove that as the number of snapshots, POD cardinality, polynomial order, and number of collocation points increase, the mean square error of approximate solution will decrease. In addition, the expression coefficients are related to the eigenvalues. A series of numerical experiments show that when using the same number of full finite element solutions, the POD-DCS method has better accuracy and sparsity than the conventional CS method. Although our method needs to solve the ROM w.r.t. the collocation points, the sparse reconstruction of low-dimensional reduced states is much cheaper than the CS method. Moreover, in inferring the permeability field for elliptic PDE, our method can not only get an accurate result, but also be much faster than the direct calculation.

    Accelerating the decline of eigenvalue in POD method is a problem worthy of consideration. And for complex problems, evaluating the forward model is costly. In order to construct an accurate reduced order model, we usually need many realizations, which can be prohibitive. Therefore, we plan to use a multi-fidelity scheme or select appropriate snapshots to overcome this difficulty. This is the subject of future work.

    [1] “Sustainable Urban Development” A special issue of Sustainability, 15 articles, 2014. Available from: http://www.mdpi.com/journal/sustainability/special_issues/sustainable_urban_development.
    [2] “Towards True Smart and Green Cities?” A special issue of Sustainability, 10 articles, 2015. Available from: http://www.mdpi.com/journal/sustainability/special_issues/true_smart.
    [3] “Low Carbon Cities” (45th ISOCARP World Congress Porto, Portugal 18-22 October 2009). A special issue of Cities, 28(6), 2011. Available from: http://www.journals.elsevier.com/cities/special-issues/.
    [4] “Green Cities” A special issue of Cities, 13(5), 1996. Available from: http://www.journals.elsevier.com/cities/special-issues/.
    [5] “From urban sprawl to compact green cities – indicators for multi-scale and multi-dimensional analysis” A special issue of the Journal of Ecological Indicators, 2015. Available from: http://www.journals.elsevier.com/ecological-indicators/call-for-papers/call-for-papers-of-special-issue-on-from-urban-sprawl-to-com/.
    [6] Breuste J, Artmann M, Li J, et al. (2015) Introduction: special issue on green infrastructure for urban sustainability. J Urban Plann Dev 141: 1-5.
    [7] Moloney S, Horne R (2015) Low carbon urban transitioning: from local experimentation to urban transformation? Sustainability 7: 2437-2453. doi: 10.3390/su7032437
    [8] Simon D (2013) Climate and environmental change and the potential for greening African cities. Local Econ 28: 203-217. doi: 10.1177/0269094212463674
    [9] Rogerson JM (2014) Green commercial property development in urban South Africa: emerging trends, emerging geographies. Bull Geogr Socio-eco Ser 26: 233-246.
    [10] Blok A (2012) Greening cosmopolitan urbanism? On the transnational mobility oof low-carbon formats in Northern European and East Asian cities. Environ Plann A 44: 2327-2343.
    [11] Hirano Y, Fujita T, Bunya S, et al. (2011) Examination of how various measures in urban districts affect the development of low carbon cities. Soc Environ Sci Japan 24: 255-268.
    [12] Ahmad S, Baiocchi G, Creutzig F (2015) CO2 emissions from direct energy use of urban households in India. Environ Sci Techol 49: 11312-11320. doi: 10.1021/es505814g
    [13] Kim G, Miller PA, Nowak DJ (2015) Assessing urban vacant land ecosystem services: urban vacant land as green infrastructure in the City of Roanoke, Virginia. Urban For Urban Green 14: 519-526. doi: 10.1016/j.ufug.2015.05.003
    [14] Sung CY, Cho W, Hong S-H (2015) Estimating the annual carbon budget of a weekend tourist resort in a temperate secondary forest in Korea. Urban For Urban Green 14: 413-419. doi: 10.1016/j.ufug.2015.04.008
    [15] Moser A, Rötzer T, Pauleit S, et al. (2015) Structure and ecosystem services of small-leaved lime (Tilia cordata Mill.) and black locust (Robinia pseudoacacia L.) in urban environments. Urban For Urban Green 14: 1110-1121.
    [16] Kotsiris G, Nektarios PA, Ntoulas N, et al. (2013) An adaptive approach to intensive green roofs in the Mediterranean climatic region. Urban For Urban Green 12: 380-392. doi: 10.1016/j.ufug.2013.05.001
    [17] Ntoulas N, Nektarios PA, Charalambous E, et al. (2013) Zoysia matrella cover rate and drought tolerance in adaptive extensive green roof systems. Urban For Urban Green 12: 522-531. doi: 10.1016/j.ufug.2013.07.006
    [18] Bigsby KM, Ambrose MJ, Tobin PC, et al. (2014) The cost of gypsy moth sex in the city. Urban For Urban Green 13: 459-468. doi: 10.1016/j.ufug.2014.05.003
    [19] Jorgensen A, Anthopoulou A (2007) Enjoyment and fear in urban woodlands – Does age make a difference? Urban For Urban Green 6: 267-278. doi: 10.1016/j.ufug.2007.05.004
    [20] Ren Z, Zheng H, He X, et al. (2015) Spatial estimation of urban forest structures with Landsat TM data and field measurements. Urban For Urban Green 14: 336-344. doi: 10.1016/j.ufug.2015.03.008
    [21] Baeumler A, Ijjasz-Vasquez E, Mehndiratta S., eds., (2012) Sustainable Low-Carbon City Development in China, Washington, DC: International Bank for Reconstruction and Development/ International Development Association or The World Bank, 516 pp.
    [22] Steffen Lehmann – Cities of the Future, TEDx Talks, YouTube video, 2011. Available from: https://www.youtube.com/watch?v=hRSVR12M-k8.
    [23] Lehmann S, ed., (2010) The Principles of Green Urbanism: Transforming the City for Sustainability, London: Routledge, 912 pp.
    [24] Lehmann S, ed., (2014) Low Carbon Cities: Transforming Urban Systems, London, Routledge, Earthscan Series on Sustainable Design, 446 pp.
    [25] Simpson R, Zimmermann M, eds. (2013) The Economy of Green Cities: A World Compendium on the Urban Green Economy, New York: Springer, 450 pp.
    [26] Seymoar N-K, Ballantyne E, Pearson CJ (2010) Empowering residents and improving governance in low income communities through urban greening. Int J Agr Sust 8: 26-39. doi: 10.3763/ijas.2009.0467
    [27] Egenhofer C, Alessi M, Núñez Ferrer J (2010) Greening EU Cities: The Emerging EU Strategy to Address Climate Change, CEPS Task Force Report, November 2010.
    [28] Pediaditi K, Doick KJ, Moffat AJ (2010) Monitoring and evaluation practice for brownfield, regeneration to greenspace initiatives: a meta-evaluation of assessment and monitoring tools. Landscape Urban Plan 97: 22-36. doi: 10.1016/j.landurbplan.2010.04.007
    [29] Randolph J (2012) Environmental Land Use Planning and Management, Washington, DC: Island Press, 746 pp.
    [30] Mell IC (2015) Establishing the rationale for green infrastructure investment in Indian cities: is the mainstreaming of urban greening an expanding or diminishing reality? AIMS Environ Sci 2: 134-153. doi: 10.3934/environsci.2015.2.134
    [31] Robinson GM, Liu Z (2015) Greening and “un”greening Adelaide, South Australia. AIMS Environ Sci 2: 511-532. doi: 10.3934/environsci.2015.3.511
    [32] Chen D, Thatcher M, Wang X, et al. (2015) Summer cooling potential of urban vegetation―a modeling study for Melbourne, Australia. AIMS Environ Sci 2: 648-667. doi: 10.3934/environsci.2015.3.648
    [33] Coronel AS, Feldman SR, Jozami E, et al. (2015) Effects of urban green areas on air temperature in a medium-sized Argentinian city. AIMS Environ Sci 2: 803-826. doi: 10.3934/environsci.2015.3.803
    [34] Thornbush M (2015) Urban agriculture in the transition to low carbon cities through urban greening. AIMS Environ Sci 2: 852-867. doi: 10.3934/environsci.2015.3.852
    [35] Dolowitz DP (2015) Stormwater management the American way: why no policy transfer? AIMS Environ Sci 2: 868-883. doi: 10.3934/environsci.2015.3.868
    [36] Magliocco A, Perini K (2015) The perception of green integrated into architecture: installation of a green facade in Genoa, Italy. AIMS Environ Sci 2: 899-909. doi: 10.3934/environsci.2015.4.899
    [37] Drake L, Ravit B, Dikidjieva I, et al. (2015) Urban greening supported by GIS: from data collection to policy implementation. AIMS Environ Sci 2: 910-934. doi: 10.3934/environsci.2015.4.910
    [38] Bowd D, McKay C, Shaw WS (2015) Urban greening: environmentalism or marketable aesthetics. AIMS Environ Sci 2: 935-949. doi: 10.3934/environsci.2015.4.935
    [39] Söderlund J, Newman P (2015) Biophilic architecture: a review of the rationale and outcomes. AIMS Environ Sci 2: 950-969. doi: 10.3934/environsci.2015.4.950
  • This article has been cited by:

    1. Yiyuan Qian, Kai Zhang, Jingzhi Li, Xiaoshen Wang, Adaptive neural network surrogate model for solving the implied volatility of time-dependent American option via Bayesian inference, 2022, 30, 2688-1594, 2335, 10.3934/era.2022119
    2. Sohail Reddy, Hillary R Fairbanks, Accelerating multilevel Markov Chain Monte Carlo using machine learning models, 2025, 100, 0031-8949, 056008, 10.1088/1402-4896/adb52a
  • Reader Comments
  • © 2016 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(6978) PDF downloads(1351) Cited by(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog