The aim of this article is to give new generalizations of both the Ostrowski's inequality and some of its new variants with the help of the F-convex function class, which is a generalization of the strongly convex functions. Young's inequality, which is well known in the literature, as well as Hölder's inequality, was used to obtain the new results. Also we obtain some results for convex and strongly convex functions by utilizing these inequalities.
Citation: Alper Ekinci, Erhan Set, Thabet Abdeljawad, Nabil Mlaiki. Generalizations of Ostrowski type inequalities via F-convexity[J]. AIMS Mathematics, 2022, 7(4): 7106-7116. doi: 10.3934/math.2022396
[1] | Jama Mohamed, Dahir Abdi Ali, Abdimalik Ali Warsame, Mukhtar Jibril Abdi, Eid Ibrahim Daud, Mohamed Mohamoud Abdilleh . Bayesian extreme value modelling of annual maximum monthly rainfall in Somalia from 1901 to 2022. AIMS Geosciences, 2024, 10(3): 598-622. doi: 10.3934/geosci.2024031 |
[2] | Shailesh Kumar Singh, Richard Ibbitt . Assessment of irrigation shortfall using WATHNET in the Otago region of New Zealand. AIMS Geosciences, 2018, 4(3): 166-179. doi: 10.3934/geosci.2018.3.166 |
[3] | Khalid Ahmad, Umair Ali, Khalid Farooq, Syed Kamran Hussain Shah, Muhammad Umar . Assessing the stability of the reservoir rim in moraine deposits for a mega RCC dam. AIMS Geosciences, 2024, 10(2): 290-311. doi: 10.3934/geosci.2024017 |
[4] | Elisabetta Dall’Ò . Historicizing vulnerability: place-names, risk and memory in the Mont Blanc area. AIMS Geosciences, 2019, 5(3): 493-508. doi: 10.3934/geosci.2019.3.493 |
[5] | Solve Hov, David Gaharia . Geotechnical characterization of index and deformation properties of Stockholm clays. AIMS Geosciences, 2023, 9(2): 258-284. doi: 10.3934/geosci.2023015 |
[6] | Alessia Nannoni, Federica Meloni, Marco Benvenuti, Jacopo Cabassi, Francesco Ciani, Pilario Costagliola, Silvia Fornasaro, Pierfranco Lattanzi, Marta Lazzaroni, Barbara Nisi, Guia Morelli, Valentina Rimondi, Orlando Vaselli . Environmental impact of past Hg mining activities in the Monte Amiata district, Italy: A summary of recent studies. AIMS Geosciences, 2022, 8(4): 525-551. doi: 10.3934/geosci.2022029 |
[7] | Konstantinos X Soulis, Evangelos E Nikitakis, Aikaterini N Katsogiannou, Dionissios P Kalivas . Examination of empirical and Machine Learning methods for regression of missing or invalid solar radiation data using routine meteorological data as predictors. AIMS Geosciences, 2024, 10(4): 939-964. doi: 10.3934/geosci.2024044 |
[8] | Yaping Xu, Lei Wang, Chengliang Liu, Cuiling Liu . Identifying Unlawful Constructions in Cultural Relic Sites Based on Subpixel Mapping—a Case Study in Mangshan Tombs, China. AIMS Geosciences, 2017, 3(2): 268-283. doi: 10.3934/geosci.2017.2.268 |
[9] | Carbone Fabio, Corinto Gian, Malek Anahita . New Trends of Pilgrimage: Religion and Tourism, Authenticity and Innovation, Development and Intercultural Dialogue: Notes from the Diary of a Pilgrim of Santiago. AIMS Geosciences, 2016, 2(2): 152-165. doi: 10.3934/geosci.2016.2.152 |
[10] | Joek Peuchen, Bart M.L. Meijninger, Daniël Brouwer . North Sea as geo database. AIMS Geosciences, 2019, 5(2): 66-81. doi: 10.3934/geosci.2019.2.66 |
The aim of this article is to give new generalizations of both the Ostrowski's inequality and some of its new variants with the help of the F-convex function class, which is a generalization of the strongly convex functions. Young's inequality, which is well known in the literature, as well as Hölder's inequality, was used to obtain the new results. Also we obtain some results for convex and strongly convex functions by utilizing these inequalities.
Data assimilation (DA) is the process of combining information from models, measurements, and priors - all with associated uncertainties - in order to better describe the the true state of a physical system. The ultimate goal of DA is to efficiently represent the posterior PDF that fully describes the uncertainty in the analysis. Two families of methods, variational and ensemble based methods, have proved very successful in real applications. Variational methods, rooted in control theory, require costly developments of tangent linear and adjoint models [32]. Ensemble-based sequential DA schemes are rooted in statistical estimation theory. The ensemble Kalman Filter was introduced by Evensen [17] and has undergone considerable developments since then. EnKF formulations fall in one of two classes, namely stochastic or deterministic formulations [55]. In the stochastic approach, each ensemble member is updated using a perturbed version of the observation vector [11,26]. In the deterministic formulation (which leads to square root ensemble filters [5,9,46,55,59] no observation noise is added, but transformations of the covariance matrix are applied such as to recover the correct analysis statistics.
All variants of the EnKF work well in case of linear observations [19], however in real applications the observation operators are in general nonlinear. EnKF can accommodate nonlinear observation operators using linearization, in the spirit of the extended Kalman filter [60]. An alternative approach to handle the non-linearity of observation operators is to use the difference between nonlinear operators evaluated at two states instead of the linearized version; this approach can result in mathematical inconsistencies [60]. A different approach to deal with nonlinear observations is to pose a nonlinear estimation problem in a subspace spanned by the ensemble members, and to compute the maximum a posteriori estimate in that subspace. This leads to the maximum likelihood ensemble filter (MLEF) proposed by Zupanski [60]. MLEF minimizes a cost function that depends on nonlinear observation operators. MLEF doesn’t require the observation operator to be differentiable and uses a difference approximation of the Jacobian of the observation operator. However, this approach may diverge if the observation operator is highly nonlinear or the observations are very sparse. In addition it is inherently assumed that the posterior distribution is Gaussian; the MLEF maximum a posteriori probability estimate may face difficulties in case of multimodal distributions. The iterative ensemble Kalman filter (IEnKF) [23,47] handles nonlinearity in observations and models by repeating analyses steps with the same observations. Each iteration of IEnKF still assumes that the underlying probability distributions are Gaussian and the analysis state is the mode of the posterior. The “running in place” (RIP) EnKF scheme [33] uses a no-cost EnKS and repeatedly assimilates the observations over each assimilation window several times. Another nonlinear extension of EnKF incorporates a non-linear change of variables (anamorphosis function) [50] and executes the analysis step in a space where the distributions of the transformed variables are Gaussian.
Particle filters(PF) [15,22,34,56] are another family of nonlinear and non-Gaussian methods. PF performs well with small dimensional systems. For large dimensional systems, or when the number of independent observations increases, PF is more likely to suffer from filter degeneracy [51], a situation in which only one particle gets the significant weight and the all the other weights become close to zero. Development of efficient particle filters for large scale nonlinear and non-Gaussian DA is an active area of research. For example, in the merging particle filter (MPF) [41], the posterior particles are formulated as linear combinations of the resampled particles from the prior at the measurement time, in an attempt to reduce the variance in the weights and to avoid filter degeneracy. Although the mean and the covariance of the posterior are preserved, this filter does not guarantee the shape of the posterior PDF is preserved in the filtered ensemble. As mentioned in [41] if the posterior PDF is significantly non-Gaussian, the MPF will likely produce rather poor estimates. The performance of EnKF, MLEF, and PF when observation operators are highly nonlinear was tested in [28,29].
A promising approach for sequential Monte Carlo sampling from the posterior density is the implicit particle filter [12]. This algorithm directs the sampling towards the regions of high density areas in the posterior and in this way it controls the number of particles in case of very high dimensional state spaces. The implicit sampling filter, however, is expensive: it requires solving a set of large algebraic equations for each particle. Another enhancement over the standard particle filter is the equivalentweights particle filter [1,57,58] which incorporates a relaxation proposal density to steer the particles towards observations and consequently towards high-probability regions of the posterior PDF, and also uses an equivalent-weights proposal density to insure that the retained particles are given similar relative weights thus avoiding filter degeneracy.
In addition to the need of fully non-Gaussian DA schemes, efficient ensemble replenishment methods are needed to support parallel implementations of the current ensemble-based DA algorithms. In parallel implementations of EnKF each ensemble member typically runs on a separate set of nodes, and when any of the nodes dies out the corresponding ensemble member is lost. The ensemble number of members is reduced and after several node failures the entire EnKF DA process is in jeopardy. To continue the EnKF process efficiently, it is necessary to replace the lost ensemble member with a new one that is drawn independently from the underlying probability distribution.
Markov Chain Monte Carlo (MCMC) is a family of algorithms designed to sample from a given probability distribution by generating a Markov chain whose invariant (stationary) distribution is the target PDF. The traditional wisdom considers MCMC sampling methods to be a gold standard [35] in DA, which is however impractical in case of high dimensions due the massive computational cost required to achieve convergence and explore the whole state space. Development of scalable MCMC algorithms to sample efficiently from high dimensional space is an active research area.
A very promising approach for large systems is Hybrid Monte Carlo (HMC), a variant of MCMC sampling that incorporates an auxiliary variable and takes advantage of the properties of Hamiltonian system dynamics [16]. HMC has been considered before in the context of DA. [6] solve a nonlinear inverse problem that finds a solution to the shallow water equations such as to minimize the residual between the trace of this solution and ill-posed boundary conditions. Once the minimum is reached posterior error statistics are approximated by sampling the nearby states using HMC. They use a simulated annealing framework, where at low temperatures a minimum is obtained, and at high temperatures a posterior sample is obtained. [19, Chapter 6] discusses the solution of weak-constraint 4D-Var problem using a gradient method and employing HMC to estimate the analysis errors. The use of simulated annealing with HMC random walk as an alternative to gradient-based minimization is discussed in [6,19].
Alexander, Eiyink, and Restrepo [2] advocate the use of HMC for solving smoothing problems. They employ a generalized version of HMC aimed at shortening the decorrelation, i.e., the number of chain steps needed to ensure independence of the generated states. Generalized HMC uses the dynamics of a modified system that is no longer Hamiltonian. There are several differences between this work and [2]. In [2] the model error, represented by additive random noise, is the only source of uncertainty; here we consider the state of the system to be uncertain due to uncertainties in initial conditions and model equations. A leap-frog discretization is employed in [2], while here we consider high order symplectic time discretizations. Finally, they test their method on a one-dimensional system. We note that the generalized HMC method employed in [2] uses the dynamics of a modified system that is no longer Hamiltonian. Finally, the tests performed in [2] focus on a one-dimensional system, while here we carry out the numerical experiments with a multidimensional nonlinear model.
The current advances in sampling algorithms make it feasible to solve inverse problems with large models by directly sampling from the posterior probability distribution. This work develops a sequential ensemble-based DA filtering technique that can accommodate non-Gaussian posterior distributions and can be efficiently applied in operational situations. Moreover, it can be used as an ensemble replenishment tool to support parallel implementations of other ensemble based DA schemes. The approach is based on applying HMC sampling in a recursive manner to solve the sequential filtering problem. The new fully nonlinear sampling filter can accommodate nonlinear observation operators and it does not require the target probability distribution to be Gaussian. The filter is open to further improvements such as the use of the generalized (nonlocal) version described in [2]. We note that a smoothing approach using MCMC to sample the posterior distribution is discussed in [14,35], however that approach is not based on HMC.
The paper is organized as follows. An overview of the DA problem and widely-used solution strategies is given in Section 2. Sampling MCMC and HMC algorithms are summarized in Section 3. The proposed sampling filter is presented in Section 4. Numerical experiments, and a comparison of the sampling filter against the traditional EnKF, MLEF, and IEnKF methods, are given in Section 5. Conclusions are drawn in Section 6.
This section provides a brief overview of the DA problem and of several solution strategies, and highlights the motivation behind the present research.
DA combines information from prior (background) knowledge, a numerical model, and observations, all with associated errors, to describe the posterior distribution of the state of a physical system.
The background represents the best estimate of the true state prior to any measurement being available. The background errors (uncertainties) are typically assumed to have a Gaussian distribution @{x^b - x^{\rm true}} \in \mathcal{N}(0, \mathbf{B})@, where @x^b@ is the background state, @x^{\rm true}@ is the unknown true state, and @\mathbf{B}@ is the background error covariance matrix. The Gaussian assumption is widely used and we will follow it as well. As discussed in Section 4 this assumption can be inaccurate, due to the recursive application of the sampling filter, and better approximations to the prior distribution might be beneficial.
The numerical model propagates the initial model state (initial condition) @x_0 \in \mathbb{R}^{N_{var}}@ at time @t_0@ to future states @x_k \in \mathbb{R}^{N_{var}}@ at times @t_k@:
@xk=Mt0→tk(x0) , t0≤tk≤tF, @
|
(1) |
Small perturbations @\delta x@ of the state of the system evolve according to the tangent linear model:
@δxk=Mt0→tk(x0)⋅δx0 , t0≤tk≤tF, @
|
(2) |
Observations of the true state are available at discrete time instants @t_k@, @t_0 \leq t_k \leq t_F@,
@yk=\textbf{y}(t_k) = \mathcal{H}_k(x_k) + \varepsilon_k, \quad k = 0, 1, \ldots, N_{obs} -1.@ |
Data assimilation combines the background estimate, the measurements , and the model to obtain an improved estimate @x^a@, called the “analysis” (or posterior), of the true state @x^{\rm true}@, together with the corresponding analysis probability distribution. Two approaches for solving the DA problem have gained widespread popularity, variational and ensemble-based methods. The sampling filter proposed in this paper belongs to the latter family.
Since EnKF variations are the most popular ensemble-based algorithms in practice, we compare the new methodology against EnKF, MLEF, IEnKF. These schemes are briefly reviewed in the following subsections.
Kalman filters (KF) [30,31] are sequential DA methodologies, where measurements are incorporated at the time moment when they become available. Sequential DA algorithms proceed in two steps, namely, forecast and analysis. In the forecast step, the state of the system is propagated forward by the model equations (1) to the next time point where observations are available, producing a forecast of the state of the system, and a forecast error covariance matrix is presented to quantify the uncertainty of the forecast.
The ensemble Kalman filter (EnKF) [11,17,18,26] takes a Monte-Carlo approach to representing the uncertainty. An ensemble of @N_{ens}@ states (@x^a_{k-1}(e)@, @e = 1, \ldots, N_{ens}@) is used to sample the analysis probability distribution at time @t_{k-1}@. Each member of the ensemble is propagated to @t_k@ using the nonlinear model (1) to obtain the ”forecast” ensemble
@xfk(e)=Mtk−1→tk(xak−1(e))+ηk(e), e=1,…,Nens. @
|
(3a) |
@\bar x^f_k = \frac{1}{N_{ens}} \sum_{e=1}^{N_{ens}}{x^f_k(e) } \, , @ | (3b) |
@\mathbf{X}^{\rm f}_k = [x^f_k(1)- \bar x^f_k, \ldots, x^f_k(N_{ens})- \bar x^f_k] \, , @ | (3c) |
@\mathbf{B}_k = \left( \frac{1}{N_{ens}-1} \left( \mathbf{X}^{\rm f}_k \left( \mathbf{X}^{\rm f}_k \right)^T \right) \right) \circ \rho.@ | (3d) |
Each member of the forecast (ensemble of forecast states @\{x^f_k(e) \}_{e=1,\ldots,N_{ens}}@ is analyzed separately
using the KF formulas [11,17]
@x^a_k(e) = x^f_k(e) + \mathbf{K}_k \left( y_k - \mathcal{H}_k(x^f_k(e)) + \zeta_k(e) \right), @ | (4a) |
@\mathbf{K}_k = \mathbf{B}_k \mathbf{H}^T_k { \left(\mathbf{H}_k \mathbf{B}_k \mathbf{H}^T_k + \mathbf{R}_k \right)}^{-1}.@ | (4b) |
Square root versions (deterministic formulations) of EnKF [55] avoid adding random noise to observations, and thus avoid additional sampling errors. The mean of the forecast ensemble is updated to produce the analysis state (posterior mean), and the posterior ensemble is induced after updating a matrix of state deviations from the ensemble mean.
The main limitation of the EnKF is due to the Gaussianity assumption on which the Kalman updates are based. The filter is optimal only when both the forecast and the observation errors are Gaussian, and the observations are linearly related to system state.
The maximum likelihood ensemble filter (MLEF) [60] seeks to alleviate the limitations of the Gaussian
assumptions by computing the maximum likelihood estimate of the state in the ensemble space.
Specifically, it maximizes the posterior probability density, or equivalently, minimizes the following
nonlinear objective function over the ensemble subspace [37,60]:
@x_k^{\rm opt} = \arg\min_x\, \mathcal{J}(x), @ | (5a) |
@\mathcal{J}(x) = \frac{1}{2} {( x - x^b_k )^T\, \mathbf{B}_k^{-1}\, ( x - x^b_k )} + \frac{1}{2} {( y_k - \mathcal{H}_k(x) )^T\, \mathbf{R}_k^{-1}\, ( y_k - \mathcal{H}_k(x) ) }\, , @ | (5b) |
As discussed in Section 2.3 MLEF follows an iterative minimization procedure of a penalty (cost) functional in the subspace spanned by the ensemble members. MLEF is specifically designed to handle nonlinear observations but may fail in the presence of strong nonlinearity in the governing model [47]. The ensemble iterative randomized maximum likelihood (EnRML) filter [23], and the iterative ensemble Kalman filter (IEnKF) [47] are both developed as extensions of MLEF. Following an iterative minimization of the cost functional with EnKF as linear solution, both EnRML and IEnKF aim at handling highly nonlinear models as well as nonlinear observation operators. A deterministic approach to the EnKF is used as base for IEnKF, while the stochastic (perturbed observations) approach is followed in the case of EnRML. IEnKF also rescales the ensemble anomalies with the ensemble transform matrix from the previous iteration, rather than using a linear regression to estimate the sensitivities between the ensemble observations and ensemble anomalies [47]. Although IEnKF is very powerful in handling nonlinearity, it is not expected to perform well if the posterior is multimodal. The optimal solution obtained can be a local minimum of the cost functional resulting in an inaccurate representation of the true state of the system. Details of IEnKF with a simple, yet detailed, pseudo code can be found in [47]. We use the formulation of IEnKF presented in [47] to check the validity of the analysis produced by the sampling filter in the cases where nonlinear observation operators are used. The IEnKF parameters (inflation factor and tolerance) are carefully tuned in order to give the best possible analysis state with minimum RMSE. This requires more iterations of the IEnKF optimization step, but it will also allow us to check how close sampling filter estimate can approach the optimal solution.
Markov Chain Monte Carlo (MCMC) algorithms [42], introduced by Metropolis et. al [40], can sample from distributions with complex probability densities @\pi(x)@. They generate a Markov chain @\{x(i)\}_{i\geq 0}@ for which @\pi(x)@ is the invariant (stationary) distribution, given that @\pi(x)@ is known up to a multiplicative constant [42]. MCMC methods work by generating a random walk using a proposal PDF and an “acceptance/rejection” criterion to decide whether proposed samples should be accepted as part of the Markov chain or should just be rejected. These algorithms are generally powerful, but may take a long time to explore the whole state space or even to converge [54]. This section starts with a review of the Hamiltonian Monte-Carlo sampling (HMC) then presents the sampling filter algorithm for DA.
Hybrid Monte Carlo (HMC) methods, also known as Hamiltonian Monte Carlo, originated in the physics literature [16]. They attempt to handle the drawbacks of MCMC algorithms by incorporating an auxiliary variable such as to reduce the correlation between successive samples, to explore the entire space in very few steps, and to ensure high probability of acceptance for proposed samples in high dimensions [42].
Hamiltonian dynamical systems operate in a phase space of points @(p, x) \in \mathbb{R}^{2N_{var}}@, where the individual variables are the position @x \in \mathbb{R}^{N_{var}}@ and the momentum @p \in \mathbb{R}^{N_{var}}@. The total energy of the system is described by the Hamiltonian function @H(p, x)@. The dynamics of the system in time is described by the following ordinary differential equations:
@dxdt=∇pH,dpdt=−∇xH. @
|
(6) |
@ST:R2Nvar→R2Nvar;ST(p(0),x(0))=(p(T),x(T)), @
|
(7) |
In practical computations the analytic flow @\mathcal{S}_T@ is replaced by a numerical solution using a time reversible and symplectic numerical integration method [49,48]. In this paper we use five different high order symplectic integrators based on Strang’s splitting formula: Verlet (Störmer, Leapfrog) algorithm (29) [2,42], higher order integrators namely, two-stage (30), three-stage (31), and four-stage (32) position splitting integrators from [10], and the Hilbert space integrator (33) from [8]. The methods are summarized in B. To approximate @\mathcal{S}_T@ the integrator at hand takes @m@ steps of size @h=T/m@. With a slight abuse of notation we will also denote by @\mathcal{S}_T@ the flow of the numerical solution.
In order to draw samples @\{x(e)\}_{e\geq 0}@ from a given probability distribution @\pi(x)@ HMC makes the following analogy with a Hamiltonian mechanical system (6). The state @x@ is viewed as a “position variable”, and an auxiliary “momentum variable” @p@ is included. The Hamiltonian function of the system is:
@H(p,x)=12pTM−1p−log(π(x))=12pTM−1p+J(x). @
|
(8) |
@dxdt=M−1p,dpdt=−∇xJ(x). @
|
(9) |
@exp(−H(p,x))=exp(−12pTM−1p−J(x))=exp(−12pTM−1p)⋅π(x). @
|
(10) |
The HMC sampling algorithm builds a Markov chain starting from an initial state @x_0=x(0)@. Algorithm 1 summarizes the transition from the current Markov chain state @x_k@ to a new state @x_{k+1}@ [48]. Practical issues are related to the choice of the numerical integrator, the time step, and the choice of the function @\mathcal{J}(x)@ that represents the PDF we wish to sample from. The construction of the mass matrix @\mathbf{M}@ does not impact the final distribution, but does affect the computational performance of the algorithm [21]. The mass matrix @\mathbf{M}@ is symmetric and positive definite and is a parameter that is tuned by the user. It can be for example, a constant multiple of the identity [43], or a diagonal matrix whose entries are the background error variances [8,36]. We found that the latter approach is more efficient for the current application and used it in all numerical experiments reported here. We note that the generalized HMC method employed in [2] uses the dynamics of a modified system that is no longer Hamiltonian.
Algorithm 1 HMC Sampling [48]. | ||||
1: Draw a random vector @p_k \sim \mathcal{N}(0, \mathbf{M})@. | ||||
2: Use a symplectic numerical integrator (from B) to advance the current state @(p_k, x_k )@ by a time increment @T@ to obtain a proposal state @( p^* , x^* )@:
|
||||
3: Evaluate the loss of energy based on the Hamiltonian function. For the standard Verlet (29), twostage (30), three-stage (31), and four-stage (32) integrators [10,42] the loss of energy is computed as:
|
||||
4: Calculate the probability:
|
||||
5: Discard both @p^*, \ p_k@. | ||||
6: (Acceptance/Rejection) Draw a uniform random variable @u^{(k)}\sim \mathcal{U}(0, 1)@: i- If @a^{(k)} > u^{(k)}@ accept the proposal as the next sample: @x_{k+1} := x^*@; ii- If @a^{(k)} \leq u^{(k)}@ reject the proposal and continue with the current state: @x_{k+1} := x_k@. |
||||
7: Repeat steps 1 to 6 until sufficiently many distinct samples are drawn. |
The goal of this filter is to replace the analysis step in the traditional EnKF with a resampling procedure that draws representative ensemble members from the posterior distribution @\pi(x) = \mathcal{P}^{\rm a}(x)@. Even if the posterior may in general be non-Gaussian we assume, as most of the current ensemblebased DA algorithms, that the posterior has the form:
@\pi(x) = \mathcal{P}^{\rm a}(x) \, \propto \, \exp{\Bigl( - \mathcal{J}(x) \Bigr)} \, , @ | (15a) |
@\mathcal{J}(x) = \frac{1}{2} {\left( x - x^b \right)^T \mathbf{B}^{-1} \left( x - x^b \right)} + \frac{1}{2} {\Bigl( y - \mathcal{H}(x) \Bigr)^T \mathbf{R}^{-1} \Bigl( y - \mathcal{H}(x) \Bigr) } \, , @ | (15b) |
For sampling at time @t_k@ the corresponding @\mathcal{J}(x)@ is:
@J(x)=−log(Pa(x))=12(x−xbk)TB−1k(x−xbk)+12(yk−Hk(x))TR−1k(yk−Hk(x)), @
|
(16) |
@∇xJ(x)=B−1k(x−xbk)−HTkR−1k(yk−Hk(x)), @
|
(17) |
Algorithm 1 is used to generate @N_{ens}@ ensemble members drawn from the posterior distribution @\{ x^a_k(e)\sim \mathcal{P}^{\rm a}(x)\}_{e=1, 2, \ldots , N_{ens}}@. The mean of this ensemble is an estimate of the analysis state, and the ensemble covariance estimates the analysis error covariance matrix. Note that the proposed sampling filter is not restricted to a specific form of the posterior PDF, and the Gaussian assumption (16) can in principle be removed. The remaining issue is to represent non-Gaussian probability density functions and their logarithm. In the next section we describe the proposed sampling filter as an alternative to the EnKF.
The sampling filter is described in Algorithm 2. Like most of the ensemble-based sequential DA algorithms the sampling filter consists of two stages, namely, the forecast step and the analysis step.
Start with an ensemble @\{ x^a_{k-1}{(e)}\}_{e=1, \ldots , N_{ens}}@ describing the analysis PDF at time @t_{k-1}@. In the forecast step each ensemble member is propagated by the full model to the next time @t_{k-1}@ where observations are available, resulting in the forecast ensemble. In the analysis step the HMC algorithm is simply used to sample from the posterior PDF of the state, providing the new analysis ensemble @\{ x^a_{k}{(e)}\}_{e=1, \ldots , N_{ens}}@.
As stated in step ii) of Algorithm 2, the explicit representation of the matrix @\mathbf{B}_k@ is not necessary - one only needs to apply its inverse to a vector in (16), (17). Typically @\mathbf{B}_k@ is formed as a linear combination between a fixed matrix @\mathbf{B}_0@ and the ensemble covariance. The calculation requires to evaluate the products
@u=B−1k(x−xbk)=(γB0+(1−γNens−1Nens∑e=1Δx(e)(Δx(e))T)∘ρ)−1(x−xbk), @
|
(18) |
@(γB0+(1−γNens−1Nens∑e=1Δx(e)(Δx(e))T)∘ρ)⋅u=x−xbk, @
|
(19) |
Algorithm 2 Sampling Filter | |
1: Forecast step: given an analysis ensemble @\{ x^a_{k-1}{(e)}\}_{e=1, 2, \ldots , N_{ens}}@ at time @t_{k-1}@; generate the forecast ensemble by via the model @\mathcal{M}@:
|
|
2: Analysis step: given the observation vector @y_k@ at time point @t_k@, follow the steps: | |
i- Set the initial state @x_0@ of the Markov Chain to be to the best estimate available, e.g., the mean of the forecast ensemble. One can use the EnKF analysis if the cost is acceptable, and this choice is expected to result in a faster convergence of the chain to the stationary distribution. | |
ii- Calculate the ensemble-based forecast error covariance matrix @\mathbf{B}_k@ (and possibly balance it by a fixed (or frequently updated) covariance matrix @\mathbf{B}_0@), and apply localization as in equation (3d). It is important to emphasize that explicitly building the full background error covariance matrix is not necessary for the current algorithm to work. | |
iii- Choose a positive definite diagonal mass matrix @\mathbf{M}@. One choice that favors the performance of the sampling algorithm is the diagonal of the matrix @\mathbf{B}^{-1}_k@ [43] which scales the components of the state vector vary. Ideally, @\mathbf{M}@ should be set to the diagonal of the inverse posterior covariance matrix. | |
iv- Apply Algorithm 1 with initial state @x_0@ and generate @N_{ens}@ ensemble members. In practice one starts accepting samples after a warm-up phase (of, say, 30 steps), to guarantee that selected members explore the entire state space. | |
v- Use the generated samples @\{ x^a_k(e)\}_{e=1, 2, \ldots , N_{ens}}@ as an analysis ensemble and calculate the best estimate of the state (e.g. the mean), and the analysis error covariance matrix. | |
3: Increase time @k := k+1@ and repeat steps 1 and 2. |
In our numerical experiments we build flow-dependent background error covariance matrices @\mathbf{B}_k@ at each time step. For experiments with Lorenz-96 model, a fully flow-dependent background error covariance matrix is used, i.e. @\gamma=0@. In the experiments with the shallow water model on the sphere we chose @\gamma=0.5@. We set @\mathbf{M}@ to be equal to the diagonal of Bk in case of Lorenz-96 model following [8,36]. Taking @\mathbf{M}@ equal to the diagonal of @\mathbf{B}^{-1}_k@ leads to similar results for the Lorenz-96 model. For the shallow-water model on the sphere we set @\mathbf{M}@ to be equal to the diagonal of @\mathbf{B}^{-1}_k@ .
The sampling filter calculations shown in Algorithm 2 are computationally expensive. It is important to keep in mind that the main goal here is not to replace the operational filters when they work (the majority of situations). Instead we seek to develop a fully non-Gaussian sequential MCMC-based filter that is robust and can solve DA problems when the operational filters fail, at the expense of an affordable increase in the computational cost.
Moreover, as pointed out before, the sampling methodology can be used as replenishment tool in conjunction with parallel implementations of any practical filter, where some ensemble members are lost due to hardware failures. The forward model is not required during the sampling process (or the analysis step).
The cost of the analysis step of the sampling filter is determined by the settings of the chain parameters: the number of burn-in steps, the internal step size and the number of internal steps of the Hamiltonian trajectory, the number of states dropped at stationarity, and the ensemble size. The number of burn-in steps, the number of dropped states at stationarity, along with the size of the ensemble, define the length of the Markov chain. The number of internal steps is different for each symplectic integrator, and consequently the computational cost of the filter also dependents on the choice of the symplectic discretization method.
The total computational cost of the analysis step can be expressed in terms of linear algebra operations. Since the mass matrix is generally assumed to be diagonal, the computational bottleneck of the symplectic integrator is the evaluation(s) of the gradient of the potential energy. From Equation (17), assuming the observation operator is linear, three matrix-vector products (of dimension equal to the number of observations) and one linear system solution (of dimension equal to the number of states) are required to evaluate the gradient.
The total number of gradient evaluations for each symplectic integrator depends on the Hamiltonian trajectory settings. Verlet (29) requires one evaluation of the gradient of the cost functional (17) per step. If the length of the Hamiltonian trajectory is @T = m\, h@, where @m@ is the number of steps of size @h@, then @3m@ matrix-vector products and @m@ linear systems are evaluated to generate a proposal. An additional evaluation of the cost functional (16) is carried out per proposal to evaluate the loss of energy (12). An evaluation of the cost functional (16) requires two additional dot-product evaluations. The evaluation of the kinetic energy term in the Hamiltonian requires matrix-vector multiplication by the diagonal mass matrix, which is the same as the cost of a vector dot-product. The cost of other integrators scales with the number of stages, i.e., is larger by a factor of two for (30), by a factor of three for (31), and by a factor of four for (30).
If the number of burn-in steps is @b@, and the number of states dropped at stationarity is @r-1@ (each @r^{th}@ proposal is retained), then the total number of proposal states is @(b+r N_{ens})@. The total cost of the analysis step is @(b+r N_{ens})@ times the cost per step.
In the forecast step the forward model is run once for each ensemble member. The total number of forward model runs per assimilation cycle is then equal the size of the ensemble @N_{ens}@. The linear system (19) is solved once in order to update the background term.
In addition to the above discussion of the computational complexity of the sampling filter, we report the CPU times of both the sampling filter and the traditional filters in Section 5.11.
Numerical tests are primarily performed using the 40-variables Lorenz-96 model [38] which is described by the equations:
@dxidt=xi−1(xi+1−xi−2)−xi+F, @
|
(20) |
We tested the performance of the sampling filter with several observation operators of different complexities and varying levels of non-linearity. In this section, we show the results of a linear, a discontinuous-quadratic, and an exponential observation operators. The discontinuous quadratic observation (quadratic with a threshold) operator used here was employed by Zupanski [60,61] in the simple case of one dimensional state space. For experiments with more observation operators with different levels of discontinuity and nonlinearity see [3].
a) Linear observation operator: in all the experiments conducted with Lorenz-96 model we assume that third components only are observed. The linear observation operator takes the form:
@H(x)=Hx=(x1, x4, x7, …, x37, x40)T∈R14, @
|
(21) |
b) Quadratic observation operator with a threshold: this observation operator is similar to the simple version used by Zupanski et al in [61]. The observation vector is:
@H(x)=(x′1, x′4, x′7, …, x′37, x′40)T∈R14, @
|
(22) |
@x′i={ x2i:xi≥0.5−x2i:xi<0.5. @
|
(23) |
c) Exponential observation operator: this is a highly nonlinear, differentiable observation operator:
@H(x)=(er⋅x1, er⋅x4, er⋅x7, …, er⋅x37, er⋅x40)T∈R14, @
|
(24) |
The sampling filter can be tuned to accommodate complex settings of the experiment in hand. The parameters of the Hamiltonian trajectory can be tuned empirically such as to achieve a specific acceptance rate. In general the step size should be chosen to ensure that the rejection rate falls between 25% and 30%, an approach we followed in our settings, and the number of steps should be large [43]. We chose the length of the Hamiltonian trajectory of the sampling filter empirically to be @T = 10@ with @h = 0.01@ and @m = 10@ for all symplectic integrators. This choice is made by trial and error such that the sampling filter with the standard position Verlet integrator yields seemingly satisfying results with the linear observation operator at moderate cost. More precisely, these settings lead to acceptance rates from 25% to 30% when a linear observation operator is used in the present experiments. In general, however, the time-stepping parameters of each symplectic integrator should be set individually to get the best performance of the sampling filter. The goal here is to show that even with a very simple setting of the sampling filter parameters one can obtain promising results for this experimental setting. The mass matrix @\mathbf{M}@ is set as a diagonal matrix which diagonal is the precisions of the background errors at each time step. To guarantee that the Markov chain reaches the stationary distribution before starting the sampling process a set of 50 steps are perform as burn-in stage. We noticed however that the chain always converges in a small number of burn-in steps, typically 10-20 steps. Stationarity tests will be given special attention in our future work. An alternative to the burn-in steps is to apply a three dimensional variational (3D-Var) step which will certainly result in a state belonging to the posterior distribution. That resulting state can be used to initialize the chain, and sampling can start immediately. After the burn-in stage an ensemble member is selected after each 10 generated states. dropping states between successive selection has the effect of decreasing correlation between generated ensemble members since the chain is not memoryless. The number of generated states (between successive selections) that are not retained during the sampling from stationarity will be referred to as the number of mixing steps. In our experiments the acceptance probability is high (usually over 0:9) with this sampling strategy. The number of mixing steps is a parameter that can be tuned by the user to control the performance of the sampling filter.
Stability requirements impose tight upper bounds on the step size @h@ of the Verlet integrator. The step size should be decreased with the increasing dimension of the system in order to maintain @\mathcal{O}(1)@ acceptance probability [7]. On the other hand long Hamiltonian trajectories (large steps of the symplectic integrator) are needed in order to explore the space efficiently. There is no precise rule available to select the optimal step size values [48] and consequently @h@, and @m@ should be tuned for each problem. The higher-order integrators (30), (31), (32) are expected to be more stable than Verlet for larger step sizes [10,43] and a smaller number of steps.
To guarantee ergodicity of the Markov chain, which is a property required for the chain to converge to its invariant distribution, we follow [10,43] and change the step length at the beginning of each Markov step (once at the beginning of the Hamiltonian trajectory) to @h = (1+r)\, h_{\rm ref}@ where @h_{\rm ref}@ is a reference step size and @r\sim \mathcal{U}(-0.2, 0.2)@ is a uniformly distributed random variable. Randomizing the step size of the symplectic integrator, in addition to other benefits, ensures that the results obtained are not entrusted with specific choice of the step size [43].
Each numerical experiment performs 100 realizations of the sampling filter. Each realization uses the same settings but the sequence of random numbers generated by the sampling filter, for both the potential variable and the acceptance/rejection rule, was different. The root mean squared error (RMSE) metric is used to compare the analyses against the reference solution at observation time points:
@RMSE=√1NvarNvar∑i=1(xi−xtruei)2, @
|
(25) |
Figure 1 shows the average RMSE of the 100 instances of the sampling filter and the RMSE obtained by applying EnKF (with inflation factor 1:09). Statistics of RMSE results obtained by applying MLEF and IEnKF are summarized in Table 1. Rank histograms of the first two components of the state vector along the 300 assimilation cycles are given in Figure 2. We present rank histograms only for the first two components of the state vector because we noticed that rank histograms of observed components look very similar, and unobserved components nearly have similar rank histograms. Using the standard position Verlet as the symplectic integrator results in occasional failures (Figure 1). These failures are mostly due to the untuned step settings. On the other hand the higher order integrators show quite good performance. From Figure 2 we also see that using position Verlet the observed components (e.g. @x_1@) and unobserved components (e.g. @x_2@) tend to have uniform rank histogram while the use of higher order integrators results in overdispersed ensemble on average. Although not shown here, rank histograms resulting from three-stage and four-stage integrators are almost identical to those obtained by using two-stage integrator. The truth however lies in the middle of the ensemble and one can conclude that shorter Hamiltonian trajectory can be used in the case of higher-order integrators while the standard Verlet integrator requires smaller step size and larger number of steps in order to maintain stability and to reach distant points in state space respectively. For example, if the length of the Hamiltonian trajectory @T = 0.1@ is kept fixed, and the step size @h@ is reduced to 0.005, the likelihood of filter diverging (see the maximum RMSE in Table 1 for this case) can be drastically reduced. This is confirmed by the results reported in Tables 1 and 2, where the standard deviation of RMSE is reduced from 0.546054 to 0.039302, and the RMSE average is slightly decreased as well. Further tuning of the filter parameters can lead to smaller RMSE, and consequently result in a better estimation of the true state. However, in nonlinear settings, one is concerned more about the distribution of the ensemble members under the posterior PDF, rather than estimating one specific state.
Observation | Statistics | Symplectic integrator used with sampling filter | Traditional Filters | ||||||
Operator | Verlet | Two-stage | Three-stage | Four-stage | Hilbert | EnKF | MLEF | IEnKF | |
Linear observation operator | Min | 0.282158 | 0.225048 | 0.221871 | 0.22929 | 1.126023 | 0.04631 | 0.027045 | 0.027835 |
Max | 4.553241 | 0.331516 | 0.275494 | 0.292533 | 1.257923 | 0.13634 | 0.124409 | 0.197427 | |
Mean | 0.433644 | 0.250042 | 0.249086 | 0.252403 | 1.179563 | 0.079809 | 0.069438 | 0.080403 | |
Std | 0.546054 | 0.014128 | 0.01077 | 0.012136 | 0.027446 | 0.020084 | 0.021752 | 0.027741 | |
Quadratic observation operator with threshold | Min | 3.51313 | 0.362567 | 0.370246 | 0.364221 | 1.203196 | 2.223534 | 0.048766 | 0.033307 |
Max | 5.285517 | 2.525502 | 0.607215 | 1.123681 | 1.889033 | 8.452283 | 0.175145 | 0.116735 | |
Mean | 4.344827 | 0.476498 | 0.444522 | 0.462515 | 1.394191 | 3.949765 | 0.094103 | 0.06193 | |
Std | 0.318406 | 0.218234 | 0.043667 | 0.086601 | 0.134929 | 1.184631 | 0.028429 | 0.017315 | |
Exponential observation operator with @r = 0.2@ | Min | 0.527899 | 0.478779 | 0.389757 | 0.500551 | 1.332802 | 3.598643 | 0.065779 | 0.05316 |
Max | 2.406581 | 2.138528 | 0.596158 | 1.874854 | 1.840191 | 7.734853 | 0.239167 | 0.240487 | |
Mean | 1.119951 | 0.902746 | 0.446232 | 0.887058 | 1.546304 | 5.381176 | 0.155157 | 0.132423 | |
Std | 0.45629 | 0.374384 | 0.034703 | 0.35285 | 0.099781 | 0.956734 | 0.041048 | 0.036759 |
The use of the infinite dimensional integrator (33) (defined on Hilbert space) results in underdispersive ensemble however it does not show signs of divergence. This is illustrated in Figures 1(e), 2(c), and 2(f).
Figure 3 shows the RMSE results with quadratic observation operator (22) with threshold @a = 0.5@. In this case EnKF failed to converge, while both MLEF and IEnKF show very good behavior. However, we noticed that to get MLEF to produce good results inflation was needed. The performance of MLEF in the current setting is sensitive to both the observation frequency (the dimension of the observation space) and the noise level.
For clarity the figures only report results of the sampling filter (HMC) along with results obtained using MLEF (with inflation factor 1.25) and IEnKF (with inflation factor 1.09). Statistics of RMSE results for EnKF, MLEF, IEnKF and the sampling filter are given in Table 1. Given the current untuned parameter settings, we believe that the sampling filter using Verlet integrator fails due to the high non-linearity of the observation operator and/or the uncertainty levels. The high order integrators show good results and the analysis RMSE is close to the level obtained in case of linear observation operator. The likelihood of outliers is small and decreases using higher-order integrators as can be seen from the maximum values of the RMSE shown in Table 1. The Hilbert integrator gives reasonable results. Further tuning of the filter parameters can result in much better results. Rank histograms (4), and standard deviations of RMS errors in Table 1, both show that higher order integrators, e.g. three-stage (Figures 4(a), 4(b)) produce acceptable spread around the truth. The integrator defined on Hilbert space continues to show underdispersive behaviour with larger variation of the RMSE. Careful tuning of the step size used in case of the position Verlet and the integrator defined on Hilbert space is required for the sampling filter to converge. For example, setting the length of the Hamiltonian trajectory to @T = 0.03@ with @h = 0.001@; @m = 30@ results in better performance as shown in Figures 5, 6, and in Table 2. The unobserved components have roughly uniform shapes except for the two spikes at the two extremes of the rank histogram in Figure 6(b). These results indicate that a longer Hamiltonian trajectory might be needed in order for the sampling filter to be able to sample from the tails of the posterior distribution.
Statistics | Observation operator and symplectic integrator used | |
Linear @\mathcal{H}@; | Discontinuous quadratic @\mathcal{H}@; | |
Verlet: @h = 0.001@, @m = 30@ | Verlet: @h = 0.005@, @m = 20@ | |
Min | 0.291365 | 0.580504 |
Max | 0.569277 | 1.710029 |
Mean | 0.362358 | 0.900346 |
Std | 0.039302 | 0.370047 |
The ensemble size is taken to be @N_{ens}=30@, however it is important to discuss the filter performance in the case where fewer ensemble members are desirable. Reducing the ensemble size definitely reduces the cost of the sampling filter but increases the sampling error introduced. The spread of the ensemble should not be destroyed where only few ensembles are used, otherwise the filter will be questionable. The sampling filter was tested in several settings where the ensemble size is varied. We pick the case where three-stage symplectic integrator (with @h = 0.01@, @m = 10@) is used in the presence of the discontinuous quadratic observation operator (22) to show results of the sampling filter with different ensemble sizes. As appreciated from Figure 7 the spread of the ensemble with 20; 10 members is preserved compared to results with 30 ensemble members shown in Figure 4(a), 4(b) giving a hope that in case of large dimensional systems, small number of ensemble members can produce sufficient information of the posterior distribution. When the ensemble size is varied, all other components of the state vector show similar behavior to the case where 30 ensemble members are used but we showed only the first two components to be consistent with results shown before.
The sampling filter can be used to replenish the lost ensemble members in parallel implementations of the traditional filters (e.g EnkF, IEnKF). The traditional approach to replenishing is to perturb the average of the remaining ensemble members with random noise generated from a Gaussian distribution with zero mean and covariance matrix approximated based on the remaining ensemble members. This approach leads to the new forecast states living in the subspace spanned by the remaining ensemble members, a fact that can lead to filter degradation in real applications. Moreover, this strategy is no longer be valid if the Gaussianity assumption of the prior is violated.
One way to address this problem is to generate ensembles from the analysis in the previous cycle using HMC sampling filter, and to propagate them forward to the current time instance. In this case the sampling filter chain can be started from the average of the analysis ensemble and no burn-in states are required. We believe that the use of the sampling filter for replenishing dead ensemble members can be advantageous in the cases where few ensemble members are used or if the probability of node failure is high. Generally speaking, replenishment using the sampling filter does not depend the probability of node failure.
To illustrate this, we perform a simulation with the Lorenz-96 model where each of the 30 ensemble members is killed with a (admittedly large) probability @P_f=0.25@ (Figure 8). For clarity we removed the RMSE results for the case when the ensemble members are not replenished at all whenever any member dies out.
When the ensemble is not replenished, the IEnKF diverges after approximately 10 to 15 cycles. Replenishing the ensemble by perturbing the average of the remaining ensemble members with Gaussian perturbations with zero mean and covariance matrix @\mathbf{B}_k@ still can lead to filter divergence, as shown in Figure 8. Even in this case the sampling filter successfully replenishes the ensemble and aids IEnKF to maintain its desired performance.
Figure 9 shows the RMSE results with the exponential observation operator (24) with factor @r = 0.2@. This observation operator is differentiable, however small perturbations in the state might result in relatively large changes in the measured values. Under strongly nonlinear conditions the sampling filter performs better than EnKF and is giving results comparable to results from MLEF and IEnKF. MLEF results are obtained by forcing inflation with factor 1.15 in this case. Based on RMSE results, it is quite clear that the best choice is the three-stage symplectic integrator for the given parameter settings (Figures 9(c), 10). The other integrators require longer Hamiltonian trajectories or smaller step size or both to sample the posterior properly.
We have also tested the sampling filter capabilities in a very challenging setting: the exponential observation operator (24) is considered with a factor of @r = 0.5@. This factor leads to a large range of observation values (from @e^{-3.7}@ to @e^{6.2}@). In addition, small perturbations in the state variables cause very large changes in the corresponding measurements. EnKF, MLEF, and IEnKF all diverged when applied to this test, and consequently their results are not reported here.
This test problem is challenging for the sampling filter as well and the symplectic integration step sizes need to be tuned to achieve convergence. For example, the number of steps taken by three-stage integrator had to be increased to @m = 60@ while keeping the step-size fixed to @h = 0.01@, to result in good performance as shown in Figure 11(a). The length of the trajectory of the Hamiltonian system has to be increased as well. For the infinite dimensional integrator, a shorter trajectory of the Hamiltonian system works well if the step size is sufficiently reduced, e.g., @h = 0.001@, and @m = 30@ as shown in Figure 11(b). Empirical tuning of the Verlet, two-stage, and four-stage integrators proved to be challenging with this observation operator. The computational cost associated with the tuned three-stage integrator in the present experiment is much higher than the cost resulting from the use of the infinite dimensional integrator. Despite the relatively large RMSE revealed by Figure 11(b), Figure 12 shows that the analysis produced by the HMC sampling filter using the integrator defined on Hilbert space follows the truth reasonably closely, which can justify the argument that this level of RMSE can be accepted if the traditional filters fail and the cost associated with the higher order integrators is prohibitive. The Hilbert integrator operates at a much lower cost, and can be used to periodically check the results obtained with the three-stage integrator to safeguard against outliers.
The statistics of the results with Lorenz-96 model are summarized in Tables 1, 2, and 3. Given the unified time step settings @h = 0.01@, @m = 10@ for all symplectic integrators, Table 1 summarizes results for the linear observation operator (21), the discontinuous quadratic observation operator (22), and the exponential observation operator (24) with factor @r = 0.2@. Table 2 summarizes results for the linear observation operator (21), and the discontinuous quadratic observation operator (22) where the step settings of the position Verlet integrator were tuned to give better results. Table 3 summarizes results for the exponential observation operator (24) with factor @r = 0.5@. The results shown are computed for 100 instances of the sampling filter, EnKF, MLEF, and IEnKF, over the last 20% of the assimilation cycles for each experiment.
Statistics | Symplectic integrator used with sampling filter | |
Three-stage; | Hilbert; | |
@h = 0.01@, @m = 60@ | @h = 0.001@, @m = 30@ | |
Min | 0.304178 | 1.234498 |
Max | 2.671971 | 2.350684 |
Mean | 0.439776 | 1.699096 |
Std | 0.274643 | 0.250088 |
Tuning the sampling filter parameters can prevent outliers (filter divergence) that can happen especially when nonlinear observation operators are used. In addition to selecting the mass matrix @\mathbf{M}@ and the number of burn-in steps, there are two more parameters to be tuned. The first parameter is the length of the Hamiltonian trajectory (including step size and the number of steps of the symplectic integrator). The second parameter to be tuned is the number of steps skipped between selected states at stationarity (referred to as mixing steps). A common and a simple approach is to tune the parameters of the symplectic integrator by monitoring the acceptance rate in a preprocessing step. For simplicity, we followed this strategy in the present work. Automatically tuned versions of HMC have been proposed recently as well. No-U-Turn sampler (NUTS) [25] is a version of HMC capable of automatically tuning its parameters by prohibiting the sampler from retracing its steps along the constructed Hamiltonian trajectory. Riemann manifold HMC (RMHMC) [21] is another HMC sampler that tunes its parameters automatically using third-order derivative information. Tuning the symplectic integrator step settings can enhance the filter performance and reduce the of filter divergence, however, it might not be sufficient to completely prevent outliers. Tuning the number of mixing steps can in principle greatly enhance both the performance of the filter and the reliability of the results. The number of mixing steps in all experiments with Lorenz-96 is empirically set to 10 but it is not optimal. Increasing the number of mixing steps might be tempting, however, we cannot conclude that skipping more states will increase the chance of filter convergence [3]. A careful tuning of both the step size and the number of mixing steps in the chain may overcome the problem of outliers and lead to the desired performance of the filter. Generalized HMC with circular matrix can also be considered to shorten the decorrelation time [2] and hence reduce the number of mixing steps required by the sampling filter. From a statistical perspective, the former parameters can be tuned for example by inspecting the acceptance rate of the proposed states and the effective size of the ensemble.
In addition to controlling the time step settings of the integrator, and tuning the number of steps of the chain, we can use the Hilbert integrator (with tuned step size) to periodically validate the ensembles obtained using other integrators, since the Hilbert integrator suffers less from outliers. A simple solution is to run the assimilation process several times and exclude outlier states by creating a combined ensemble. Care must be exercised, however, to not change the probability density. These alternatives will be inspected in depth in future work in the context of more complex models.
As a first step towards large models we test the proposed sampling filter on the shallow water model
on a sphere, using linear observation operator where all components are observed. The shallow water
equations provide a simplified model of the atmosphere which describes the essential wave propagation
mechanisms found in general circulation models (GCMs) [52]. The shallow water equations in
spherical coordinates are given as
@\frac{\partial u}{\partial t} + \frac{1}{a\cos \theta} \left( u \frac{\partial u}{\partial \lambda} + v \cos \theta \frac{\partial u}{\partial \theta} \right) - \left(f + \frac{u \tan \theta}{a} \right) v + \frac{g}{a \cos \theta} \frac{\partial h} {\partial \lambda} = 0, @ | (26a) |
@\frac{\partial v}{\partial t} + \frac{1}{a\cos \theta} \left( u \frac{\partial v}{\partial \lambda} + v \cos \theta \frac{\partial v}{\partial \theta} \right) + \left(f + \frac{u \tan \theta}{a} \right) u + \frac{g}{a} \frac{\partial h} {\partial \theta} = 0, @ | (26b) |
@\frac{\partial h}{\partial t} + \frac{1}{a \cos \theta} \left(\frac{\partial\left(hu\right)}{\partial \lambda} + \frac{\partial{\left(hv \cos \theta \right)}}{\partial \theta} \right) = 0\, .@ | (26c) |
@xk+1=Mtk→tk+1(xk,θ),k=0,…,N;x0=x0(θ). @
|
(27) |
The modeled version of the background error covariance, @\mathbf{B}_0@, that accounts for correlations between state variables is created as follows:
· Start with a diagonal background error covariance matrix with uncertainty levels as mentioned previously.
· Apply the ensemble Kalman filter with 500 ensemble members for 48 hours. Synthetic initial ensemble is created by adding zero-mean Gaussian noise to the reference initial condition with covariances set to the initial (diagonal) background error covariance matrix.
· Decorrelate the ensemble-based covariances using a decorrelation matrix @\rho@ with decorrelation distance @L=1000\, km@.
· Calculate @\mathbf{B}_0@ by averaging the covariances over the last 6 hours.
This method of creating a synthetic initial background error covariance matrix is totally empirical, but we found that the resulting background error covariance matrix performs well for several algorithms including 4D-Var. Moreover, the quality of the background error covariance matrix can be enhanced by the flow-dependent versions obtained by the used filters.
The assimilation time interval of the current experiment is 22 hours with observations available each hour. The number of burn-in steps in the Markov chain is set to 50. The size of the ensemble is chosen to be @N_{ens}=100@. Given the simple settings in this experiment, it is expected that Verlet (29) or any of the higher order symplectic integrators (30), (31), (32) will behave similarly with tuned parameters. Here we provide the results obtained using two-stage integrator with step size @h = 0.01@ and number of steps @m = 10@. The number of mixing steps is 5. Beginning with the second assimilation cycle, we noticed that the filter starts to show converge in a very small number of steps, typically 5 to 10 steps. Statistical tests of convergence should be considered to start the sampling process as soon as convergence is achieved and to avoid discarding states generated from the posterior. We noticed that if the position Verlet is used instead, longer Hamiltonian trajectories are required while step sizes should be lowered in order to achieve both stability and fast space exploration. It is important to note that the quality of the background error covariance matrix and the way it is updated both have big impact on the performance of the filter. The flow-dependent background error covariance matrix is used to update the modelled background version. A hybrid version of the background error covariance matrix is created using (19) as a linear combination of the modeled version with the linear coefficient @\gamma = 0.5@. The forecast and analysis states obtained by EnKF and the sampling filter at the last assimilation cycle are shown in Figure 13. The sampling filter results in an analysis state that is almost identical to the analysis obtained by EnKF. Figure 14 shows the RMSE of the analysis states, over the 22 cycles, obtained by both EnKF and the HMC sampling filter. Clearly, in the presence of linear observation operator, the sampling filter competes with EnKF which is known to be optimal given the current settings. It is clear from both Figures 14, and 13 that the HMC is capable of handling high-dimensional models as well as EnKF. Despite the additional cost, the performance of the sampling filter under the current settings encourage us to test the sampling filter with larger models under more challenging settings. A comparison between EnKF, MLEF, and PF applied to SWE on the sphere, in the presence of both linear and nonlinear observation operators, can be found in [29].
Rank histograms of the uncorrelated components of the state vector along the 22 assimilation cycles are given in Figure 15. The first column of panels shows rank histograms where the ensemble size is @N_{ens}=100@, while in the second column of panels the size of the ensemble is reduced to @N_{ens}=20@. We see that even with 20 ensemble members, the rank histograms are approximately uniform. This gives the sampling filter a chance to work with high dimensional model where only small number of ensemble members is affordable.
To complement the discussion provided in 4.1, we show the CPU usage, per assimilation cycle, of both the traditional filters and the HMC sampling filter.
The CPU-times for the experiments carried out using Lorenz-96 model are shown in Figure 16. As discussed in Section 4.1 the cost of the sampling filter depends on the settings of the Markov chain, and relies heavily on the choice of the parameters of the symplectic integrator of choice. Here we fix the parameters of all the symplectic integrators to @T=m\times h=0.1@, with @m = 10@ and @h = 0.01@ (units) for both models. For experiments carried out using Lorenz-96 model, the size of the ensemble is 30 members, the number of burn-in steps is 50, and the number of mixing steps is 10. The blue parts of the bars report the CPU-time spent during the burn-in process, while the green parts show the time spent during sampling (collecting ensemble members after the Markov chain has reached stationarity). The reason behind the notable high CPU usage of IEnKF here is that we tune the IEnKF tolerance iteratively so that we guarantee the best possible results against which we can compare the HMC sampling results. While the average CPU-time of HMC sampling filter with Verlet integrator is roughly 4:7 times the CPU-time reported by EnKF, the cost of the HMC sampling filter is proportional to the number of stages of the symplectic integrator. This additional computational cost is justifiable in cases where traditional methods fail, or when ensemble member replenishment is needed. For a linear observation operator we show the CPU-times for two cases with ensemble sizes of 100 and 20 members. The number of burn-in steps and the number of mixing steps are 50 and 5 respectively. With only 20 ensemble members collected the sampling filter with Verlet integrator takes approximately 2:16 longer than EnKF. As pointed out before, only the time spent during collecting sample points changes by varying the ensemble size, as seen by comparing results reported in Figures 17(b) and 17(a). To greatly reduce the time usage of the HMC sampler, one can run the burn-in process sequentially, to guarantee convergence to the right distribution, followed by a parallel run of the sampling process. The burn-in process itself can be replaced by a sub-optimal 3D-Var run to attain a local minimum of the negative-log of the posterior kernel.
This paper proposes a sampling filter for data assimilation where the analysis scheme is replaced by sampling directly from the posterior distribution. A Hamiltonian Monte-Carlo technique is employed to generate a representative analysis ensemble at each time. The sampling filter avoids the need to develop tangent linear or adjoint models of the model solution operator. The sampling filter can work with highly nonlinear observation operators and provides analysis ensembles that describe non-Gaussian posterior probability densities. The mean of the generated posterior ensemble provides the analysis (a minimum variance estimate of the state). The ensemble covariance offers an estimate of the analysis error covariance matrix and can be used to quantify the uncertainty associated with the analysis state. The implementation does not require the construction of full covariance matrices, which makes the method attractive for large scale data assimilation problems with operational models and complex observation operators.
Numerical experiments are carried out with the Lorenz-96 model with several observation operators with different levels of non-linearity and smoothness. The sampling filter provides results similar to EnKF for linear observations. For nonlinear observations the new filter outperforms EnKF, and its performance can be improved further with careful tuning of filter parameters. In addition, the sampling filter produces satisfactory results for test cases where EnKF, MLEF, and IEnKF fail.
Large scale ensemble filtering data assimilation problems are typically run on large parallel machines. One important challenge is the failure of subsets of nodes, which terminates some of the ensemble member runs, and leads to fewer ensemble members being available at the next time. Over several cycles the effective number of ensemble members can decrease considerably. The sampling strategy proposed herein can be used to replace dead ensemble members in any parallel implementation of the EnKF. In addition, the sampling filter can be used in combination with classical filters by building analysis ensembles that have members given by EnKF analysis (these members retain the history of the system) mixed with sampling members (which are consistent with the posterior probability density, but add new directions to explore and can therefore avoid filter divergence).
The computational performance of the sampling filter depends on tuning its parameters, especially the symplectic integration time step and the number of steps taken in the Markov chain between successive accepted ensemble members. Future work will focus on refining the strategies for parameter tuning in the context of large operational models at high-resolution, and on incorporating acceleration strategies for the HMC process. Use of non-Gaussian kernels such as Gaussian mixtures to represent the prior will also be investigated. A side by side comparison between the proposed sampling filter and the successful particle filters is planned. Balance is a concern for all filters including the sampling filter. A possible solution could be to restrict posterior PDF with physical information. This direction will be also investigated in future work.
The initial background error covariance matrix @\mathbf{B}_0@, and the observations error covariance matrix R used with Lorenz-96 model experiments are built as described below.
Let @\delta x = \sigma_{x_0} \, x_0^{\rm ref}\, , @ where @x_0^{\rm ref}@ is the reference initial condition. @\sigma_{x_0}@ is the noise level in initial background solution. The initial background error covariance matrix is given by:
@B0=0.1×I+(0.9×(δx(δx)T)∘ρ), @
|
(28) |
@δx=[0.2581,0.2262,0.2867,0.4257,0.6204,−0.0480,−0.0213,0.4307,0.2429,−0.3132,0.1184,0.3484,0.6099,−0.1823,0.1344,0.3489,0.6167,−0.3491,0.5768,0.1640,0.0068,0.4713,0.3250,0.0875,0.3577,0.6307,0.4373,0.1470,−0.0495,−0.1448,0.0189,0.5290,0.2887,−0.1785,0.2546,0.5911,−0.1673,0.2455,0.6292,0.7743]T, @
|
Observation error covariance matrices @\mathbf{R}@ are assumed to be diagonal and are created based on the observation operator in hand. Diagonals of @\mathbf{R}@ for the used observation operators are given as follows:
· Linear observation operator (21):
@[0.0273,0.0271,0.0263,0.0326,0.0314,0.0258,0.0283,0.0273,0.0323,0.0287,0.0294,0.0340,0.0223,0.0281]T. @
|
· Quadratic observation operator (22) with a threshold @a = 0.5@:
@[0.6901,0.6022,0.6442,0.8984,0.8009,0.6371,0.7297,0.6929,1.0260,0.7944,0.8087,1.1770,0.5506,0.7371]T. @
|
· Exponential observation operator (24) with factor @r = 0.2@:
@[0.0093,0.0090,0.0094,0.0109,0.0106,0.0092,0.0095,0.0093,0.0123,0.0089,0.0104,0.0136,0.0083,0.0089]T. @
|
· Exponential observation operator (24) with factor @r = 0.5@:
@[0.3096,0.2065,0.3227,0.4626,0.3911,0.2820,0.3281,0.3266,0.7467,0.4050,0.4228,1.1328,0.3087,0.3206]T. @
|
Here we present the five numerical integrators employed in this work. We start with the standard position Verlet integrator in B.1. The results of the standard Verlet are very sensitive to the choice of the time step. Three higher order integrators namely, two-stage (B.2), three-stage (B.3), and fourstage (B.4) position splitting integrators, are taken from [10]. These higher-order integrators lead to filters that are more stable and efficient than Verlet. The last integrator tested (B.5) is from [8] and is designed to work efficiently in infinite dimensional state spaces, and to avoid problems resulting from subtracting infinitely large numbers related to the total energy of the Hamiltonian system for infinite dimensional state spaces.
All the integrators are applied to a Hamiltonian system of the form (9) [48,49].
One step of the position Verlet algorithm advances the solution of the Hamiltonian equations (9) from time @t_k@ to time @t_{k+1} = t_k + h@ as follows [48]:
@xk+1/2=xk+h2M−1pk,pk+1=pk−h∇xJ(xk+1/2),xk+1=xk+1/2+h2M−1pk+1. @
|
The optimal time step @h@ is @h \propto {(1/{N_{var}})}^{1/4}@ [7]. The experiments show that the step size should be small (close to zero) to make this integrator stable. It may still fail for high dimensionality and whenever complications are present in the target distributions. The weakness of this simple integrator is illustrated in our numerical experiments with highly nonlinear observation operators.
One step of the two-stage algorithm advances the solution of the Hamiltonian equations (9) from time @t_k@ to time @t_{k+1} = t_k + h@ as follows [10]:
@x1=xk+(a1h)M−1pk,p1=pk−(b1h)∇xJ(x1),x2=x1+(a2h)M−1p1,pk+1=p1−(b1h)∇xJ(x2),xk+1=x2+(a2h)M−1pk+1, @
|
Linear stability analysis using a standard harmonic oscillator test problem [10] shows that the stability of this symplectic integrator is achieved for time step that lies in the interval @0 < h\omega < 2.6321480259@, where @\omega@ is the frequency of the oscillator.
One step of the three-stage algorithm advances the solution of the Hamiltonian equations (9) from time tk to time tk+1 = tk + h by the set of equations [10]:
@x1=xk+(a1h)M−1pk,p1=pk−(b1h)∇xJ(x1),x2=x1+(a2h)M−1p1,p2=p1−(b2h)∇xJ(x2),x3=x2+(a2h)M−1p2,pk+1=p2−(b1h)∇xJ(x3),xk+1=x3+(a1h)M−1pk+1, @
|
Following the same argument as in B.2, the stability interval of the time step associated with this time integrator is of length @\approx 4.67@, that is, @h@ should be chosen such that @0 < h\omega < 4.67@, where @\omega@ is the frequency of the oscillator [10].
One step of the four-stage algorithm advances the solution of the Hamiltonian equations (9) from time @t_k@ to time @t_{k+1} = t_k + h@ as follows [10]:
@x1=xk+(a1h)M−1pk,p1=pk−(b1h)∇xJ(x1),x2=x1+(a2h)M−1p1,p2=p1−(b2h)∇xJ(x2),x3=x2+(a3h)M−1p2,p3=p2−(b2h)∇xJ(x3),x4=x3+(a2h)M−1p3,pk+1=p3−(b1h)∇xJ(x4),xk+1=x4+(a1h)M−1pk+1, @
|
Again, as discussed in B.2, and B.3, this integrator has a stability interval @0 < h\omega < 5.35@ [10].
The time here has unspecified units. Generally speaking, the high order integrators (30, 31, 32), provide more favorable and wider stability ranges for the time step. For more on the stability intervals of the time step settings of these high-order integrators, see [10].
One step of the Hilbert integrator advances the solution of the Hamiltonian equations (9) from time @t_k@ to time @t_{k+1} = t_k + h@ as follows [8]:
@p1=pk−h2M−1∇xJ(xk),xk+1=cos(h)xk+sin(h)p1,p2=−sin(h)xk+cos(h)p1,pk+1=p2−h2M−1∇xJ(xk+1). @
|
This work was supported in part by awards NSF CCF-1218454 and AFOSR FA9550-12-1-0293-DEF, and by the Computational Science Laboratory at Virginia Tech.
[1] |
G. A. Anastassiou, Ostrowski type inequalities, Proc. Amer. Math. Soc., 123 (1995), 3775–3781. https://doi.org/10.1090/S0002-9939-1995-1283537-3 doi: 10.1090/S0002-9939-1995-1283537-3
![]() |
[2] |
M. Adamek, On a problem connected with strongly convex functions, Math. Inequal. Appl., 19 (2016), 1287–1293. https://doi.org/10.7153/mia-19-94 doi: 10.7153/mia-19-94
![]() |
[3] |
M. Adamek, On Hermite-Hadamard type inequalities for F-convex functions, J. Math. Inequal., 14 (2020), 867–874. https://doi.org/10.7153/jmi-2020-14-56 doi: 10.7153/jmi-2020-14-56
![]() |
[4] | M. W. Alomari, M. M. Almahameed, Ostrowski's type inequalities for functions whose first derivatives in absolute value are MN-convex, Turkish J. Ineq., 1 (2017), 53–77. |
[5] | M. Alomari, M. Darus, Some Ostrowski type inequalities for convex functions with applications, RGMIA, 13 (2010), 1–11. |
[6] |
M. Alomari, M. Darus, S. S. Dragomir, P. Cerone, Ostrowski type inequalities for functions whose derivatives are s-convex in the second sense, Appl. Math. Lett., 23 (2010), 1071–1076. https://doi.org/10.1016/j.aml.2010.04.038 doi: 10.1016/j.aml.2010.04.038
![]() |
[7] |
A. O. Akdemir, M. E. Özdemir, S. Varošanec, On some inequalities for h-concave functions, Math. Comput. Model., 55 (2012), 746–753. https://doi.org/10.1016/j.mcm.2011.08.051 doi: 10.1016/j.mcm.2011.08.051
![]() |
[8] | B. Bayraktar, M. Gürbüz, On some integral inequalities for (s,m)-convex functions, TWMS J. App. Eng. Math., 10 (2020), 288–295. |
[9] | S. S. Dragomir, T. M. Rassias, Ostrowski type inequalities and applications in numerical integration, Dordrecht: Kluwer Academics, 2002. |
[10] | M. Gürbüz, M. E. Özdemir, On some inequalities for product of different kinds of convex functions, Turkish J. Sci., 5 (2020), 23–27. |
[11] | M. A. Khan, F. Alam, S. Z. Ullah, Majorization type inequalities for strongly convex functions, Turkish J. Ineq., 3 (2019), 62–78. |
[12] |
N. Merentes, K. Nikodem, Remarks on strongly convex functions, Aequat. Math., 80 (2010), 193–199. https://doi.org/10.1007/s00010-010-0043-0 doi: 10.1007/s00010-010-0043-0
![]() |
[13] |
K. Nikodem, Z. Pales, Characterizations of inner product spaces by strongly convex functions, Banach J. Math. Anal., 5 (2011), 83–87. https://doi.org/10.15352/bjma/1313362982 doi: 10.15352/bjma/1313362982
![]() |
[14] | M. E. Özdemir, A. O. Akdemir, E. Set, On the Ostrowski-Grüss type inequality for twice differentiable functions, Hacet. J. Math. Stat., 41 (2012), 651–655. |
[15] |
M. E. Özdemir, A. Ekinci, Generalized integral inequalities for convex functions, Math. Inequal. Appl., 19 (2016), 1429–1439. https://doi.org/10.7153/mia-19-106 doi: 10.7153/mia-19-106
![]() |
[16] | B. T. Poljak, Existence theorems and convergence of minimizing sequences for extremal problems with constraints, Dokl. Akad. Nauk SSSR, 166 (1966), 287–290. |
[17] |
E. Set, M. E. Özdemir, M. Z. Sarıkaya, A. O. Akdemir, Ostrowski-type inequalities for strongly convex functions, Georgian Math. J., 25 (2018), 109–115. https://doi.org/10.1515/gmj-2017-0043 doi: 10.1515/gmj-2017-0043
![]() |
1. | Achraf Azanzal, Chakir Allalou, Said Melliani, Gevrey class regularity and stability for the Debye-H¨uckel system in critical Fourier-Besov-Morrey spaces, 2022, 41, 2175-1188, 1, 10.5269/bspm.62517 | |
2. | Shouming Zhou, Li Zhang, On the Cauchy problem for Keller‐Segel model with nonlinear chemotactic sensitivity and signal secretion in Besov spaces, 2023, 0170-4214, 10.1002/mma.9104 | |
3. | Nguyen Thi Loan, Van Anh Nguyen Thi, Tran Van Thuy, Pham Truong Xuan, Periodic solutions of the parabolic–elliptic Keller–Segel system on whole spaces, 2024, 297, 0025-584X, 3003, 10.1002/mana.202300311 | |
4. | Nguyen Thi Loan, Pham Truong Xuan, Almost periodic solutions of the parabolic-elliptic Keller–Segel system on the whole space, 2024, 123, 0003-889X, 431, 10.1007/s00013-024-02023-8 | |
5. | Tran Van Thuy, Nguyen Thi Van, Pham Truong Xuan, On asymptotically almost periodic solutions of the parabolic-elliptic Keller-Segel system on real hyperbolic manifolds, 2024, 1468-9367, 1, 10.1080/14689367.2024.2360204 |
Observation | Statistics | Symplectic integrator used with sampling filter | Traditional Filters | ||||||
Operator | Verlet | Two-stage | Three-stage | Four-stage | Hilbert | EnKF | MLEF | IEnKF | |
Linear observation operator | Min | 0.282158 | 0.225048 | 0.221871 | 0.22929 | 1.126023 | 0.04631 | 0.027045 | 0.027835 |
Max | 4.553241 | 0.331516 | 0.275494 | 0.292533 | 1.257923 | 0.13634 | 0.124409 | 0.197427 | |
Mean | 0.433644 | 0.250042 | 0.249086 | 0.252403 | 1.179563 | 0.079809 | 0.069438 | 0.080403 | |
Std | 0.546054 | 0.014128 | 0.01077 | 0.012136 | 0.027446 | 0.020084 | 0.021752 | 0.027741 | |
Quadratic observation operator with threshold | Min | 3.51313 | 0.362567 | 0.370246 | 0.364221 | 1.203196 | 2.223534 | 0.048766 | 0.033307 |
Max | 5.285517 | 2.525502 | 0.607215 | 1.123681 | 1.889033 | 8.452283 | 0.175145 | 0.116735 | |
Mean | 4.344827 | 0.476498 | 0.444522 | 0.462515 | 1.394191 | 3.949765 | 0.094103 | 0.06193 | |
Std | 0.318406 | 0.218234 | 0.043667 | 0.086601 | 0.134929 | 1.184631 | 0.028429 | 0.017315 | |
Exponential observation operator with @r = 0.2@ | Min | 0.527899 | 0.478779 | 0.389757 | 0.500551 | 1.332802 | 3.598643 | 0.065779 | 0.05316 |
Max | 2.406581 | 2.138528 | 0.596158 | 1.874854 | 1.840191 | 7.734853 | 0.239167 | 0.240487 | |
Mean | 1.119951 | 0.902746 | 0.446232 | 0.887058 | 1.546304 | 5.381176 | 0.155157 | 0.132423 | |
Std | 0.45629 | 0.374384 | 0.034703 | 0.35285 | 0.099781 | 0.956734 | 0.041048 | 0.036759 |
Statistics | Observation operator and symplectic integrator used | |
Linear @\mathcal{H}@; | Discontinuous quadratic @\mathcal{H}@; | |
Verlet: @h = 0.001@, @m = 30@ | Verlet: @h = 0.005@, @m = 20@ | |
Min | 0.291365 | 0.580504 |
Max | 0.569277 | 1.710029 |
Mean | 0.362358 | 0.900346 |
Std | 0.039302 | 0.370047 |
Statistics | Symplectic integrator used with sampling filter | |
Three-stage; | Hilbert; | |
@h = 0.01@, @m = 60@ | @h = 0.001@, @m = 30@ | |
Min | 0.304178 | 1.234498 |
Max | 2.671971 | 2.350684 |
Mean | 0.439776 | 1.699096 |
Std | 0.274643 | 0.250088 |
Observation | Statistics | Symplectic integrator used with sampling filter | Traditional Filters | ||||||
Operator | Verlet | Two-stage | Three-stage | Four-stage | Hilbert | EnKF | MLEF | IEnKF | |
Linear observation operator | Min | 0.282158 | 0.225048 | 0.221871 | 0.22929 | 1.126023 | 0.04631 | 0.027045 | 0.027835 |
Max | 4.553241 | 0.331516 | 0.275494 | 0.292533 | 1.257923 | 0.13634 | 0.124409 | 0.197427 | |
Mean | 0.433644 | 0.250042 | 0.249086 | 0.252403 | 1.179563 | 0.079809 | 0.069438 | 0.080403 | |
Std | 0.546054 | 0.014128 | 0.01077 | 0.012136 | 0.027446 | 0.020084 | 0.021752 | 0.027741 | |
Quadratic observation operator with threshold | Min | 3.51313 | 0.362567 | 0.370246 | 0.364221 | 1.203196 | 2.223534 | 0.048766 | 0.033307 |
Max | 5.285517 | 2.525502 | 0.607215 | 1.123681 | 1.889033 | 8.452283 | 0.175145 | 0.116735 | |
Mean | 4.344827 | 0.476498 | 0.444522 | 0.462515 | 1.394191 | 3.949765 | 0.094103 | 0.06193 | |
Std | 0.318406 | 0.218234 | 0.043667 | 0.086601 | 0.134929 | 1.184631 | 0.028429 | 0.017315 | |
Exponential observation operator with @r = 0.2@ | Min | 0.527899 | 0.478779 | 0.389757 | 0.500551 | 1.332802 | 3.598643 | 0.065779 | 0.05316 |
Max | 2.406581 | 2.138528 | 0.596158 | 1.874854 | 1.840191 | 7.734853 | 0.239167 | 0.240487 | |
Mean | 1.119951 | 0.902746 | 0.446232 | 0.887058 | 1.546304 | 5.381176 | 0.155157 | 0.132423 | |
Std | 0.45629 | 0.374384 | 0.034703 | 0.35285 | 0.099781 | 0.956734 | 0.041048 | 0.036759 |
Statistics | Observation operator and symplectic integrator used | |
Linear @\mathcal{H}@; | Discontinuous quadratic @\mathcal{H}@; | |
Verlet: @h = 0.001@, @m = 30@ | Verlet: @h = 0.005@, @m = 20@ | |
Min | 0.291365 | 0.580504 |
Max | 0.569277 | 1.710029 |
Mean | 0.362358 | 0.900346 |
Std | 0.039302 | 0.370047 |
Statistics | Symplectic integrator used with sampling filter | |
Three-stage; | Hilbert; | |
@h = 0.01@, @m = 60@ | @h = 0.001@, @m = 30@ | |
Min | 0.304178 | 1.234498 |
Max | 2.671971 | 2.350684 |
Mean | 0.439776 | 1.699096 |
Std | 0.274643 | 0.250088 |