
Citation: Gheorghe Craciun, Matthew D. Johnston, Gábor Szederkényi, Elisa Tonello, János Tóth, Polly Y. Yu. Realizations of kinetic differential equations[J]. Mathematical Biosciences and Engineering, 2020, 17(1): 862-892. doi: 10.3934/mbe.2020046
[1] | Wojciech Kuryłek . Can we profit from BigTechs' time series models in predicting earnings per share? Evidence from Poland. Data Science in Finance and Economics, 2024, 4(2): 218-235. doi: 10.3934/DSFE.2024008 |
[2] | Matthew Ki, Junfeng Shang . Prediction of minimum wages for countries with random forests and neural networks. Data Science in Finance and Economics, 2024, 4(2): 309-332. doi: 10.3934/DSFE.2024013 |
[3] | Lindani Dube, Tanja Verster . Interpretability of the random forest model under class imbalance. Data Science in Finance and Economics, 2024, 4(3): 446-468. doi: 10.3934/DSFE.2024019 |
[4] | Sami Mestiri . Credit scoring using machine learning and deep Learning-Based models. Data Science in Finance and Economics, 2024, 4(2): 236-248. doi: 10.3934/DSFE.2024009 |
[5] | Tahir Afzal, Muhammad Asim Afridi, Muhammad Naveed Jan . Integrating LSTM with Fama-French six factor model for predicting portfolio returns: Evidence from Shenzhen stock market China. Data Science in Finance and Economics, 2025, 5(2): 177-204. doi: 10.3934/DSFE.2025009 |
[6] | Esau Moyoweshumba, Modisane Seitshiro . Leveraging Markowitz, random forest, and XGBoost for optimal diversification of South African stock portfolios. Data Science in Finance and Economics, 2025, 5(2): 205-233. doi: 10.3934/DSFE.2025010 |
[7] | Nitesha Dwarika . Asset pricing models in South Africa: A comparative of regression analysis and the Bayesian approach. Data Science in Finance and Economics, 2023, 3(1): 55-75. doi: 10.3934/DSFE.2023004 |
[8] | Alexandra Piryatinska, Boris Darkhovsky . Retrospective technology of segmentation and classification for GARCH models based on the concept of the ϵ-complexity of continuous functions. Data Science in Finance and Economics, 2022, 2(3): 237-253. doi: 10.3934/DSFE.2022012 |
[9] | Kasra Pourkermani . VaR calculation by binary response models. Data Science in Finance and Economics, 2024, 4(3): 350-361. doi: 10.3934/DSFE.2024015 |
[10] | Wojciech Kurylek . Are Natural Language Processing methods applicable to EPS forecasting in Poland?. Data Science in Finance and Economics, 2025, 5(1): 35-52. doi: 10.3934/DSFE.2025003 |
Intervals analysis (see Bauch and Neumaier (1992); Moore (1966); Jaulin et al., (2001); Alefeld and Herzberger (2012)), initially developed in the 1960s to take into account in a rigorous way, different types of uncertainties (rounding errors due to finite precision calculations, measurement uncertainties, linearization errors), makes it possible to build supersets of the domain of variation of a real function. Coupled with the usual theorems of existence, for example, the Brouwer or Miranda theorems, the interval theory also makes it possible to rigorously prove the existence of solutions for a system of equations (see Goldsztejn et al., (2005)). With interval analysis, it was now possible to model interval data.
In recent years, more precisely since the end of the 1980s, interval modeling has caught the attention of a growing number of researchers. The advantage of an interval-valued time series over a point-valued time series lies in that it contains both the trend (or level) information and volatility information (e.g., the range between the boundaries), while some informational loss is encountered when one uses a conventional point-valued data set, e.g., the closing prices of a stock collected at a specific time point within each time period, since it fails to record the valuable intraday information. Higher-frequency point-valued observations could result in hardly discriminating information from noises. A solution is to analyze the information in an interval format by collecting the maximum and minimum prices in a day, which avoids undesirable noises in the intraday data and contains more information than point-valued observations Sun et al. (2018). For instance, in their work Lu et al. (2022) proposed a modified threshold autoregressive interval-valued models with interval-valued factors to analyze and forecast interval-valued crude oil prices, and proved that oil price range information is more valuable than oil price level information in forecasting crude oil prices.
Huge progress in the field of interval-valued time series has been done by Billard and Diday (2000, 2003), who first proposed a linear regression model for the center points of 37 interval-valued data. They have been followed by other authors (Maia et al., (2008); Hsu and Wu (2008); Wang and Li (2011); González-Rivera and Lin (2013); Wang et al., (2016)). To study interval data, all those references apply point-valued techniques on the center, the left bound, or the right bound. By so doing, they may not efficiently make use of the information contained in interval data. In 2016, Han et al. (2016) developed a minimum-distance estimator to match the interval model predictor with the observed interval data as much as possible. They proposed a parsimonious autoregressive model for a vector of interval-valued time series processes with exogenous explanatory interval variables in which an interval observation is considered as a set of ordered numbers. It is shown that their model can efficiently utilize the information contained in interval data, and thus provides more efficient inferences than point-based data and models Han et al. (2015). As recent development in the field, one can refer to the work of Dai et al., (2023) where a new set-valued GARCH model was constructed. We also advise readers to look at the work of Wu et al., (2023).
Despite all these advances, the classical theory of interval modeling has some inconveniences. We can enumerate two which are addressed in another work and in the present paper, respectively.
First, the set of random intervals (or more generally random sets) is not a vector space. Indeed, the set of intervals is not an abelian group for the classical addition of intervals. So, all the useful theorems obtained through orthogonal projection such as the Wold decomposition theorem cannot be extended to interval-valued processes. Second, in time series, interval-valued data does not take into account some specifications or details of the study period, for instance in financial markets where a movement in stock prices during a given trading period is an observation of bounded intervals by maximum and minimum daily prices (see Han et al. (2016)). One can use two concepts to address each of these inconveniences. One can consider the set of random intervals as a "pseudovector space" where vectors do not necessarily have opposites. This concept of a pseudovector space was developed in Kamdem et al., (2020) to address the first inconvenience stated above. The second inconvenience can be addressed by working with "extended intervals" instead of classical intervals, as in the present paper.
Indeed, it may be more relevant to consider extended intervals formed by the opening and closing prices, regarding stock prices. Also, for the daily temperature in meteorology, instead of taking the max and min, it would be better in some cases to take the morning and evening temperature, as well as for the systolic and diastolic blood pressures in medicine. For this last example of blood pressure, when plotting the blood pressure of somebody as extended intervals of morning and evening records, one can easily see days where the morning blood pressure was higher than the evening one, which can indicate illness or emotional issues.
Therefore, given the constraints imposed by classical interval theory and its application on time series, our approach is based on the concept of extended or generalized intervals for which the left bound is not necessarily less than the right one. This generalization makes our modeling approach relevant for time series analysis. This generalization guarantees the completeness of interval space and consistency between interval operations. Extended intervals are also used for time series analysis in Han et al. (2012), but their approach does not highlight the advantages of generalized interval-valued variables.
Our contribution is therefore both theoretical and empirical. In other words, we have conceptualized and redefined some of the specific characteristics of the set of extended intervals. More precisely, we define on the set of extended intervals, a topology which generalizes the natural topology on the set of classical interval, unlike the topology introduced by Ortolf (1969) on generalized intervals, which restricted on classical interval is different from the natural topology.
The rest of the work is organized as follows: The main purpose of Section 2 is to fix notations, and give a novel and consistent definition of extended intervals. In Section 3 we introduce a suitable class of distances on the set of random extended intervals, which solves a disadvantage of the Hausdorff. We use this new distance to define the variance and covariance of random extended intervals and we show that they share some useful properties with point-valued random variables, (see propositions 3 and 4). Section 4 is concerned with stationary extended interval-valued time series, and ARMA model are investigated. In Section 5, we prove the Wold decomposition version of extended interval-valued time series. Section 6 is about numerical studies. In this section we present an algorithm to convert efficiently point-valued data to extended interval-valued data. We make a simulation of an I-AR(1) process and illustrate the interpretation of a plot of extended intervals on a few data on blood pressure. We also do empirical analysis and forecasting of the French CAC 40 market index from June 1st to July 26, 2019. The paper ends with a short conclusion.
In this section, we first recall some basic concepts related to standard intervals. Next, we define what is meant by "extended interval", and we introduce the set R← of real numbers traveled in the reverse direction as a Cartesian product. At the end of this section, we present a novel representation of extended intervals.
Let Kkc(R) be the set of nonempty compact (and convex) intervals. For A=[a1,a2],B=[b1,b2]∈Kkc(R), and λ∈R, we recall the operations
A+B=[a1+b1,a2+b2] | (2.1) |
λA={[λa1,λa2] if λ≥0[λa2,λa1] if λ≤0. | (2.2) |
It is noteworthy that Kkc(R) is closed under those operations, but it is not a vector space, since A+(−1)A is not necessarily {0}, unless A={0}. The Hausdorff distance dH is defined for closed intervals [a1,a2] and [b1,b2] by
dH([a1,a2],[b1,b2])=max(|b1−a1|,|b2−a2|). |
It is well-known that (Kkc(R),dH) is a complete metric space (see Yang and Li (2005) for details). For A∈Kkc(R), the support function of A is the function s(⋅,A):R→R defined by
s(x,A)=sup{ax;a∈A}. | (2.3) |
Equivalently, if we set A=[a1,a2],
s(x,A)=max(xa1,xa2). |
Keep in mind that s(x,A) returns x times the left bound of A when x is negative, and x times the right bound of A when x is positive. This observation will be used to extend the support function on "extended closed intervals".
Definition 1. An extended interval is a range A of real numbers between A_ and ¯A, with A_,¯A∈R∪{±∞}, traveled through from A_ to ¯A.
The difference with standard intervals is that, for extended intervals, we do not impose that A_≤¯A, but the running direction is important. We say that A is an increasing extended interval or a proper interval when A_<¯A, a decreasing extended interval or an improper interval when A_>¯A, and a degenerate interval when A_=¯A. When A_ and ¯A are in A, we say that A is an extended closed interval and denote it by A=⌊A_,¯A⌋. We also have extended open intervals ⌋A_,¯A⌊, R=⌋−∞,∞⌊ and R←:=⌋∞,−∞⌊.
Every non-degenerate extended interval A represents the classical interval from min(A_,¯A) to max(A_,¯A) in the increasing direction (for an increasing extended interval) or in the decreasing direction (for a decreasing extended interval). We call A_ the left bound and ¯A the right bound of the extended interval A.
A bounded extended interval can be seen as a subset of the product set
R⇄:=R×Z2=R×{0,1}=:R×{+,−}. |
An element of R⇄ is then a pair (x,α) where x∈R, and the direction α∈{0,1}. In this structure, we have two kinds of degenerate extended intervals, namely {a}+:={a}×{0} and {a}−:={a}×{1}. A decreasing extended interval (when A_>¯A) is written as ⌋A_,¯A⌊:=[¯A,A_]×{1}, and an increasing interval (when A_<¯A) as ⌋A_,¯A⌊:=[A_,¯A]×{0}.
Thus, R⇄ is the set of real numbers R endowed with two directions represented by the elements of the Abelian group Z2. The direction 0 (or +) means you move on the real line from the left to the right, and the direction 1 (or −) means you move from the right to the left. Further, the product [2,4]×{0,1} is the subset of R⇄ in which one can move either from 2 to 4 or from 4 to 2. Equivalently, [2,4]×{0,1}=([2,4]×{0})∪([2,4]×{1}).
We denote [a,b]×{0} by [a,b]+, or just [a,b], and [a,b]×{1} by [a,b]−. Also, we denote (x,0) by x+ or just x, and (x,1) by x−. For instance, 3∈[2,4] and 3∉[2,4]−, while 3−∉[2,4] and 3−∈[2,4]−.
Practically, talking about the French CAC40 index, if we say that we got 4922− today, this will mean that we got a value of 4922 and the index was decreasing when we got this value. This is an example of how this new structure of extended intervals can be very useful in the context of the trading market, and more.
The best choice of topology on the second member {0,1} of R⇄ is the discrete topology: every subset is open. So, if we also endow R with its natural topology, the only compact and convex subset for the product topology in R⇄, are the closed extended intervals ⌊A_,¯A⌋.
We need now to clarify how to compute the intersection of extended intervals with our notations. First observe that A⊆B means that B_≤A_≤¯A≤¯B or B_≥A_≥¯A≥¯B. For instance, \lfloor1, 2\rfloor\nsubseteq\lfloor3, 1\rfloor . In fact, the elements of \lfloor1, 2\rfloor are 1^+, 1.2^+, 1.5^+ , and so on, and do not belong to \lfloor3, 1\rfloor = [1, 3]^- . The only obstruction for the inclusion to hold in this example is the difference in the running direction between both intervals.
Proposition 2.1. Let A and B be two compact extended intervals. If A and B are running in opposite directions, then A\cap B = \emptyset . Otherwise, the intersection A\cap B is the biggest extended interval C such that C\subseteq A and C\subseteq B . This is naturally extended to general subsets A and B .
Example 2.1. \lfloor0, 1\rfloor\cap\lfloor1, 2\rfloor = \{1\} , \qquad \lfloor1, 0\rfloor\cap\lfloor2, 1\rfloor = \{1\}_\leftarrow , \qquad \lfloor0, 1\rfloor\cap\lfloor2, 1\rfloor = \emptyset , \qquad \lfloor2, 1\rfloor\cap\lfloor3, 1\rfloor = \lfloor2, 1\rfloor , \qquad \lfloor3, 1\rfloor \cap\lfloor4, 2\rfloor = \lfloor3, 2\rfloor , \qquad {\mathbb R}\cap {\mathbb R}_\leftarrow = \emptyset .
Now that union and intersection are well defined for subsets of {\mathbb R}_\rightleftarrows , one can define topologies on the latter.
Definition 2. The natural topology of {\mathbb R}_\rightleftarrows is the topology generated by the set of extended open intervals.
The topology induced on {\mathbb R} by {\mathbb R}_\rightleftarrows coincides with the natural topology of {\mathbb R} . We denote by \mathcal{K}({\mathbb R}) the set of all compact extended intervals, except decreasing degenerate extended intervals. That means that all degenerate intervals in \mathcal{K}({\mathbb R}) are increasing. We extend the Hausdorff distance on \mathcal{K}({\mathbb R}) as
\begin{equation} d_H(A, B) = max(|\underline{A}-\underline{B}|, |\overline{A}-\overline{B}|). \end{equation} | (2.4) |
Example 2.2. In \mathcal{K}({\mathbb R}) , the extended closed intervals \lfloor \underline{A}, \overline{A}\rfloor and \lfloor \overline{A}, \underline{A}\rfloor are different, unless \underline{A} = \overline{A} , and d_H(\lfloor \underline{A}, \overline{A}\rfloor, \lfloor \overline{A}, \underline{A}\rfloor) = |\overline{A}-\underline{A}| . This distance can be viewed as the effort needed to turn \lfloor \underline{A}, \overline{A}\rfloor into \lfloor \overline{A}, \underline{A}\rfloor .
It is simple to see that each extended interval A\in \mathcal{K}({\mathbb R}) is uniquely defined by the restriction of its support function on \{-1, 1\} . Moreover, the map (\mathcal{K}({\mathbb R}), d_H)\to \left({\mathbb R}^{\{-1, 1\}}, d_{max}\right) is an isometry. (To be precise, d_{max} here is the maximum distance given by d_{max}(f, g) = \max(|g(-1)-f(-1)|, |g(1)-f(1)|) .) Thus, the following result is a consequence of the completeness of \left({\mathbb R}^{\{-1, 1\}}, d_{max}\right) .
Theorem 2.1. (\mathcal{K}({\mathbb R}), d_H) is a complete metric space.
We endow \mathcal{K}({\mathbb R}) with the topology induced by the Hausdorff distance d_H . We extend multiplication (2.2) on extended intervals in such a way that multiplication of an increasing extended interval by a negative number gives a decreasing extended interval and vice versa. This ensure the consistency of the extensions on \mathcal{K}({\mathbb R}) of the internal composition laws (2.1)–(2.2):
\begin{equation} \lambda A = \lfloor\lambda \underline{A}, \lambda \overline{A}\rfloor, \quad A+B = \lfloor \underline{A}+\underline{B}, \overline{A}+\overline{B}\rfloor, \qquad A-B = \lfloor \underline{A}-\underline{B}, \overline{A}-\overline{B}\rfloor, \qquad\forall\lambda\in {\mathbb R}. \end{equation} | (2.5) |
The operator - can be seen as an extension of the difference of Hukuhara defines for standard intervals by A-B = [\min(\underline{A}-\underline{B}, \overline{A}-\overline{B}), \max(\underline{A}-\underline{B}, \overline{A}-\overline{B})] . It is simple to see that (\mathcal{K}({\mathbb R}), +, \cdot) is a vector space and 0: = [0, 0] is the zero vector.
For extended closed intervals A and B the support function reads
\begin{equation} s_{A}(u) = \begin{cases} \sup\{ux;\; x\in A\} & \mbox{ if } \underline{A}\leq \overline{A}\\ \inf\{ux;\; x\in A\} & \mbox{ if } \overline{A} < \underline{A} \end{cases}. \end{equation} | (2.6) |
For instance, s_A(-1) = -\underline{A} and s_A(1) = \overline{A} . Hence, the support function from the vector space of extended closed intervals to the vector space {\mathbb R}^{\{-1, 1\}} of maps from \{-1, 1\} to {\mathbb R} , is linear. That is, for all compact extended intervals A, B ,
\begin{align*} s_{A+B}& = s_A+s_B\\s_{\lambda A}& = \lambda s_A, \; \; \quad\forall\lambda\in {\mathbb R}\\s_{A-B}& = s_A-s_B. \end{align*} |
For any extended interval A , we call the vector of s_A the column vector S_A = (-s_A(-1), s_A(1))' .
Let (\Omega, {\mathscr A}, P) be a probability space. For any A\in \mathcal{K}({\mathbb R}) , we set
hits(A) = \{B\in \mathcal{K}({\mathbb R});A\cap B\neq\emptyset\} |
as set of compact extended intervals that hit A . We endow the set \mathcal{K}({\mathbb R}) of compact extended intervals with the \sigma- algebra \mathfrak B(\mathcal{K}({\mathbb R})) generated by \{hits(A); \; A\in \mathcal{K}({\mathbb R})\} . For simplicity, we denote X^{-1}(hits(A)): = \{\omega\in\Omega; \; X(\omega)\cap A\neq\emptyset\} by X^{-1}(A) and call it the inverse image of A by X . This inverse image X^{-1}(A) is the collection of \omega\in\Omega such that X(\omega) hits A . The following three definitions are equivalent to the ones given in Han et al. (2012).
Definition 3. A random extended interval on a probability space (\Omega, {\mathscr A}, P) is a map X:\Omega\to \mathcal{K}({\mathbb R}) such that, for any A\in \mathcal{K}({\mathbb R}) , X^{-1}(A)\in {\mathscr A} .
So, a random extended interval is a measurable map X:\Omega\to \mathcal{K}({\mathbb R}) from the underlying probability space to \mathcal{K}({\mathbb R}) , endowed with the \sigma- algebra \mathfrak B(\mathcal{K}({\mathbb R})) . We denote by \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})] the set of random extended intervals. \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})] inherits from the vector space structure of \mathcal{K}({\mathbb R}) . The distribution of X\in \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})] is the map P_X:\mathfrak B(\mathcal{K}({\mathbb R}))\to[0, 1] defined on \mathcal{O}\in\mathfrak B(\mathcal{K}({\mathbb R})) by
P_X(\mathcal{O}): = P(X\in\mathcal{O}). |
Definition 4. A map f:\Omega\to {\mathbb R} is called a selection map for a random extended interval X when f(\omega)\in X(\omega) for almost every \omega\in\Omega .
Selection maps for X = \lfloor \underline{X}, \overline{X}\rfloor are then maps leaving between \underline{X} and \overline{X} . For instance, \underline{X} and \overline{X} are selection maps for X . The expectation of X is the set of expectations of measurable selection maps for X . More precisely:
Definition 5. The expectation of a random extended interval X on a probability space (\Omega, {\mathscr A}, P) is the extended interval
\begin{equation} E[X] = \lfloor E[\underline{X}], E[\overline{X}]\rfloor. \end{equation} | (3.7) |
Proposition 3.2. For any X, Y\in \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})] and \lambda\in {\mathbb R} , E[X+\lambda Y] = E[X]+\lambda E[Y] .
We denote by \mathcal S_X = \{f\in L^1(\Omega) \mbox{ such that } f \mbox{ is a selection map for }X\} the set of integrable selection maps for X and \mathcal S_X({\mathscr A}_0) = \{f\in L^1(\Omega, {\mathscr A}_0) \mbox{ such that } f \mbox{ is a selection map for }X\} the set of (\Omega, {\mathscr A}_0)- integrable selection maps for X , where {\mathscr A}_0 a sub -\sigma- field of {\mathscr A} . The expectation of X is the classical interval \{E[f]\; ;\; f\in S_X\} together with the running direction coming from X .
To quantify the variability of X , that is the dispersion of X around its expectation, we need a suitable distance measure on random extended intervals. The first distance that could come to mind is the Hausdorff distance. But, a disadvantage of the Hausdorff distance is for instance that d_H([0, 2], [5, 6]) = 5 = d_H([0, 2], [5, 7]) , while intuitively the distance between [0, 2] and [5, 6] should be less than the distance between [0, 2] and [5, 7] .
In Bertoluzza et al. (1995), the authors defined the squared distance d_\gamma^2(A, B) between two standard intervals as follow: For any interval A = [\underline{A}, \overline{A}] , we consider the one-to-one map \nabla_{\!\!A}:[0, 1]\to A , t\mapsto t\underline{A}+(1-t)\overline{A} . Then, the squared distance d_\gamma^2(A, B) is given by
\begin{equation} d_\gamma^2(A, B) = \int_{0}^{1}\left(\nabla_{\!\!A}(t)-\nabla_{\!\!B}(t)\right)^2\gamma(t)dt = \int_{0}^{1}\left(t(\underline{A}-\underline{B})+(1-t)(\overline{A}-\overline{B})\right)^2 \gamma(t)dt, \end{equation} | (3.8) |
where \gamma(t)dt is a Borel measure on [0, 1] such that:
\begin{align} \gamma(t)&\geq0 \mbox{ for every } t\in[0, 1]; \end{align} | (3.9a) |
\begin{align} \int_{0}^{1}\gamma(t)dt& = 1; \end{align} | (3.9b) |
\begin{align} \gamma(t)& = \gamma(1-t); \end{align} | (3.9c) |
\begin{align} \gamma(0)& > 0 \end{align} | (3.9d) |
We extend d_\gamma on extended intervals with the same formula (3.8) and assumptions (3.9a)-(3.9d). If d_\gamma^2(A, B) = 0 , then \nabla_{\!\!A}(t) = \nabla_{\!\!B}(t) for almost every t\in[0, 1] , which implies that \underline{A} = \underline{B} and \overline{A} = \overline{B} ; thus A = B . For triangle inequality, we first write
(\nabla_{\!\!A}(t)-\nabla_{\!\!C}(t))^2 = (\nabla_{\!\!A}(t)-\nabla_{\!\!B}(t))^2+(\nabla_{\!\!B}(t)- \nabla_{\!\!C}(t))^2+2(\nabla_{\!\!A}(t)-\nabla_{\!\!B}(t))(\nabla_{\!\!B}(t)-\nabla_{\!\!C}(t)). |
Hence,
\begin{equation} d_\gamma^2(A, C) = d_\gamma^2(A, B)+d_\gamma^2(B, C)+2\int_{0}^{1}(\nabla_{\!\!A}(t)- \nabla_{\!\!B}(t))(\nabla_{\!\!B}(t)-\nabla_{\!\!C}(t))\gamma(t)dt. \end{equation} | (3.10) |
From here, using Hölder's inequality, one gets the triangle inequality. Thus, d_\gamma is a distance on the set \mathcal{K}({\mathbb R}) of extended intervals. The two extended intervals A = \lfloor \underline{A}, \overline{A}\rfloor and \tilde A = \lfloor \overline{A}, \underline{A}\rfloor represent the same standard interval but are different in \mathcal{K}({\mathbb R}) , and d_\gamma(A, \tilde A) = |\underline{A}-\overline{A}|cst (with cst = \left(\int_{0}^{1}(2t-1)^2\gamma(t)dt\right)^{1/2}\neq0 ) vanishes if and only if \underline{A} = \overline{A} . This distance can be seen as the effort needed to turn \tilde A into A .
Conditions (3.9a)–(3.9b) are required if we want the distance d_\gamma on degenerate intervals [a, a] and [b, b] to give the usual distance |b-a| . On other hand, the distance d_\gamma is suitable for intervals since it does not share some disadvantages of the Hausdorff distance, see Bertoluzza et al. (1995) for more details.
The norm of an interval A is the distance between A and 0 : \|A\| = d_\gamma(A, 0) . Condition (3.9c) means that there is no preferable position between left and right bounds. More precisely, this condition implies that \|\lfloor a, 0\rfloor\| = \|\lfloor0, a\rfloor\| = |a|\left(\int_{0}^{1}t^2\gamma(t)dt\right)^{1/2} . The previous observation justifies the following definition.
Definition 6. We say that \gamma(t)dt is an adapted measure if, in addition to conditions (3.9a)–(3.9d) one has
\begin{equation} \int_{0}^{1}t^2\gamma(t)dt = 1 \end{equation} | (3.9f) |
Example 3.3. One can check that, with
\gamma(t) = t(1-t)\left(480-\frac{10240}{3\pi}\sqrt{t(1-t)}\right)+1, |
\gamma(t)dt is an adapted measure. We will refer to this as the standard adapted measure. It has been used in the software R Core Team (2021) to check Lemma 3.1.
Generally, for any c\in(0, \infty) , the formula
\gamma_c(t) = t(1-t)\left(a+b\sqrt{t(1-t)}\right)+c, |
defines an adapted measure for a = -30c+510 and b = \frac{512(c-21)}{3\pi} .
This d_\gamma distance can be related to the D_K distance measure developed by Körner and Näther (2002) as follows:
\begin{align} d_\gamma^2(A, B)& = (s_{A}(-1)-s_{B}(-1))^2K(-1, -1)+(s_{A}(1)-s_{B}(1))^2K(1, 1)\\ &-2(s_{A}(-1)-s_{B}(-1))(s_{A}(1)-s_{B}(1))K(-1, 1)\\ & = \begin{pmatrix} -s_{A}(-1)+s_{B}(-1)\\s_A(1)-s_B(1) \end{pmatrix}'\begin{pmatrix} K(-1, -1) & K(-1, 1)\\K(1, -1) & K(1, 1) \end{pmatrix}\begin{pmatrix} -s_A(-1)+s_B(-1)\\s_A(1)-s_B(1) \end{pmatrix}\\ d_\gamma^2(A, B)& = S_{A-B}'\mathcal K_\gamma S_{A-B} \end{align} | (3.11) |
where the kernel \mathcal K_\gamma = (K(i, j))_{i, j = -1, 1} introduced by Han et al. (2012) is given by
\begin{equation} \begin{cases} K(-1, -1) = \int_{0}^{1}t^2\gamma(t)dt\\ K(1, 1) = \int_{0}^{1}(1-t)^2\gamma(t)dt\\ K(-1, 1) = K(1, -1) = \int_{0}^{1}t(1-t)\gamma(t)dt \end{cases}. \end{equation} | (3.12) |
We will often denote \langle S_{A-B}, S_{A-B}\rangle_\gamma: = d_\gamma^2(A, B) . As observed before by Han et al. (2012), the kernel \mathcal{K}_\gamma is symmetric positive definite and defines an inner product on \mathcal{K}({\mathbb R}) . We use some properties of this inner product in order to perform the proofs of Lemma 2 and Theorem 3.1. The following lemma shows that there exists a unique distance d_\gamma with \gamma(t)dt an adapted measure. This lemma is also useful for numerical simulations.
Lemma 3.1. All adapted measures induce the same metric given by
\mathcal K_\gamma = \begin{pmatrix} 1 & -1/2\\-1/2 & 1 \end{pmatrix}\quad\mathit{\mbox{and}}\quad d_\gamma^2(A, B) = (\underline{A}-\underline{B})^2+ (\overline{A}-\overline{B})^2-(\underline{A}-\underline{B})(\overline{A}-\overline{B}). |
Proof. If \gamma(t)dt is an adapted measure, then K(1, 1) = K(-1, -1) = \int_{0}^{1}t^2\gamma(t)dt = 1 . Using conditions (3.9a)-(3.9d), one shows that K(-1, 1) = K(1, -1) = -1/2 .
Let X and Y be two random intervals. For any \omega\in\Omega , X(\omega) and Y(\omega) are two extended intervals and one can compute the distance d_\gamma(X(\omega), Y(\omega)) . We defined a new distance on random extended intervals by taking the square root of the mean of the squared distance d_\gamma^2(X(\omega), Y(\omega)) in (\Omega, {\mathscr A}, P) .
Definition 7. The D_\gamma distance is defined for two random extended intervals X, Y by
D_\gamma(X, Y) = \left(E[d_\gamma^2(X, Y)]\right)^{1/2} = \sqrt{\int_{\Omega}\int_{0}^{1} \left(\nabla_{\!\!X(\omega)}(t)-\nabla_{\!\!Y(\omega)}(t)\right)^2\gamma(t)dt\, dP(\omega)}, |
provided the integral converges.
We denote by \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] the set of random extended intervals X such that E\|X\|_\gamma^2 : = E(d^2_\gamma(X, 0)) = D_\gamma^2(X, 0) < \infty .
Lemma 3.2. \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] is a vector space under laws (2.5).
Proof. It is enough to show that \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] is a sub-vector space of \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})] . Let X, Y\in \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] and \lambda\in {\mathbb R} . Then, D_\gamma(\lambda X, 0) = |\lambda|D_\gamma(X, 0) and
\begin{align*} D^2_\gamma(X+Y, &0) = E\left[S_{X+Y}' \mathcal{K}_\gamma S_{X+Y}\right]\\ & = E\left[(S_{X}+S_{Y})' \mathcal{K}_\gamma(S_{X}+S_{Y})\right]\\ & = D^2_\gamma(X, 0)+D^2_\gamma(Y, 0)+2E\left[S_{X}' \mathcal{K}_\gamma S_{Y}\right]\\ &\leq2D^2_\gamma(X, 0)+2D^2_\gamma(Y, 0). \end{align*} |
The last inequality comes from the fact that, using Cauchy-Schwarz inequality,
\begin{align*} 2S_{X}' \mathcal{K}_\gamma S_{Y}& = 2\langle S_{X}, S_{Y}\rangle_\gamma \leq2\sqrt{\langle S_{X}, S_{X}\rangle_\gamma}\sqrt{\langle S_{Y}, S_{Y}\rangle_\gamma} \leq{\langle S_{X}, S_{X}\rangle_\gamma}+{\langle S_{Y}, S_{Y}\rangle_\gamma} \end{align*} |
It is simple to see that for any X, Y\in \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] , 0\leq D_\gamma(X, Y) < \infty and the triangle inequality for D_\gamma follows from the one of d\gamma . However, D_\gamma is not a metric on \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] since D_\gamma(X, Y) = 0 does not imply the strict equality X = Y , but that they are equal almost everywhere. We denote by L^2[\Omega, \mathcal{K}({\mathbb R})] the quotient set of \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] under the equivalent relation "being equal almost everywhere". Then, D_\gamma is a metric on L^2[\Omega, \mathcal{K}({\mathbb R})] . We will keep denoting any class in L^2[\Omega, \mathcal{K}({\mathbb R})] by a representative X\in \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] .
Theorem 3.2. \left(\mathcal{K}({\mathbb R}), d_\gamma\right) and \left(L^2[\Omega, \mathcal{K}({\mathbb R})], D_\gamma\right) are complete metric spaces.
Proof. Assume that (A_n = \lfloor \underline{A}_n, \overline{A}_n\rfloor)_n is a d_\gamma- Cauchy sequence in \mathcal{K}({\mathbb R}) . Then, (\underline{A}_n, \overline{A}_n)'_n is a Cauchy sequence in {\mathbb R}^2 and so converges, say, to (\underline{A}, \overline{A})' . In fact, that d_\gamma(A_p, A_q) = S_{A_p-A_q}'\mathcal K_\gamma S_{A_p-A_q} goes to 0 as p, q go to infinity implies that S_{A_p-A_q} = (-\underline{A}_p+\underline{A}_q, \overline{A}_p-\overline{A}_q)' goes to 0 as p, q go to infinity. Also, (A_n)_n converges to A = \lfloor \underline{A}, \overline{A}\rfloor since d_\gamma(A_n, A) = S_{A_{n}-A}'\mathcal K_\gamma S_{A_{n}-A} . Hence, \left(\mathcal{K}({\mathbb R}), d_\gamma\right) is a complete metric space. Now, assume that (X_n = \lfloor \underline{X}_n, \overline{X}_n\rfloor)_n is a D_\gamma- Cauchy sequence in L^2[\Omega, \mathcal{K}({\mathbb R})] . Then, from Fatou's lemma and Definition 7,
E[\liminf\limits_{p, q\to\infty} d_\gamma^2(X_p(\omega), X_q(\omega))]\leq\liminf\limits_{p, q\to\infty}E[d_\gamma^2 (X_p(\omega), X_q(\omega))] = 0. |
Hence, E[\liminf\limits_{p, q\to\infty}d_\gamma^2 (X_p(\omega), X_q(\omega))] = 0 , which implies that for almost every \omega\in\Omega , \liminf\limits_{p, q\to\infty}d_\gamma^2(X_p(\omega), X_q(\omega)) = 0 . Hence there exists a subsequence (X_{n_k}(\omega)) which is a Cauchy sequence in the complete metric space \left(\mathcal{K}({\mathbb R}), d_\gamma\right) . So, for almost every \omega , (X_{n_k}(\omega))_k d_\gamma -converges to X(\omega) = \lfloor \underline{X}(\omega), \overline{X}(\omega)\rfloor ; setting X(\omega) to be 0 for the remaining \omega , one obtains an random extended interval X . As \lim\limits_{k\to\infty}d_\gamma^2(X_{n_k}, X) = 0 , we also have that \lim\limits_{k\to\infty}d_\gamma^2(X_n, X_{n_k}) = d_\gamma^2(X_n, X) for any n . Using Fatou's lemma again,
\lim\limits_{n\to\infty}E[d_\gamma^2(X_n, X)] = \lim\limits_{n\to\infty}E[\liminf\limits_{k\to\infty} d_\gamma^2(X_n, X_{n_k})]\leq\lim\limits_{n\to\infty}\liminf\limits_{k\to\infty}E[d_\gamma^2(X_n, X_{n_k})] = 0, |
since \lim\limits_{p, q\to\infty}E[d_\gamma^2(X_p(\omega), X_q(\omega))] = 0 implies that \lim\limits_{n, k\to\infty}E[d_\gamma^2(X_n, X_{n_k})] = 0 .
Remark 3.1. It is clear that the space \mathcal{K}({\mathbb R}) of compact extended intervals can be identified as a 2- dimensional vector space, and the metric d_\gamma can be written as
d_\gamma(A, B) = \|US_Y-US_X\|, |
where U is a matrix such that K_\gamma = U'U . Thus, \left(L^2[\Omega, \mathcal{K}({\mathbb R})], D_\gamma\right) is identified as the 2- dimensional random vector space on (\Omega, {\mathscr A}, P) with D_\gamma(X, Y) = E(\|US_Y-US_X\|^2)^{1/2} , and the previous result follows from the completeness of the 2- dimensional random vector space.
Definition 8. We say that a sequence (X_n) of random extended intervals converges to X in probability under the metric d_\gamma when (d_\gamma^2(X_n, X)) converges to 0 in probability, that is
\forall\varepsilon > 0, \qquad\lim\limits_{n\to\infty}P(d_\gamma^2(X_n, X) \geq\varepsilon) = 0. |
Theorem 3.3. A sequence (X_n) such that \sup\limits_nE\|X_n\| < \infty , converges to X in (L^2[\Omega, \mathcal{K}({\mathbb R})], D_\gamma) if and only if (X_n) converges to X in probability under the metric d_\gamma .
Proof. Let us assume that (X_n) converges to X , that is (D^2_\gamma(X_n, X) = E[d_\gamma^2(X_n, X)]) converges to 0 . That means that (d_\gamma(X_n, X)) converges to 0 in norm L^2 in (\Omega, {\mathscr A}, P) , which implies that (d_\gamma^2(X_n, X)) converges to 0 in probability. Conversely, assume that (X_n) converges to X in probability under the metric d_\gamma . So, the inequality |d_\gamma(X_n, 0)-d_\gamma(X, 0)|\leq d_\gamma(X_n, X) implies that (\|X_n\|) converges to \|X\| in probability. By Fatou's Lemma,
E\|X\|\leq\liminf\limits_{n\to\infty}E\|X_n\|\leq\sup\limits_nE\|X_n\| < \infty. |
The inequality
d_\gamma^2(X_n, X)\leq2\|X_n\|^2+2\|X\|^2 |
implies that (d_\gamma(X_n, X)) is uniformly integrable. Finally, the dominated convergence theorem implies that (D_\gamma(X_n, X)) converges to 0 .
Corollary 3.1. Let (X_n) be a sequence of random extended intervals such that \sup\limits_nE\|X_n\| < \infty and (\lambda_n) a family of nonnegative real numbers such that \sum\lambda_n^2 < \infty . Then, (S_n = \sum_{i = 0}^{n}\lambda_iX_i) converges in probability under the metric d_\gamma .
Definition 9 (Han et al. (2012)). The covariance of two random extended intervals X , Y is the real
\begin{align} Cov(X, Y)&: = E\langle S_{X-E[X]}, S_{Y-E[Y]}\rangle_\gamma\\ & = \int_{\Omega}\int_{0}^{1}\left(\nabla_{\!\!X(\omega)}(t)- \nabla_{\!\!E[X]}(t)\right)\left(\nabla_{\!\!Y(\omega)}(t)-\nabla_{\!\!E[Y]}(t)\right) \gamma(t)dt\, dP(\omega). \end{align} | (3.13) |
The variance of X is the real
\begin{equation} Var(X) = Cov(X, X) = E\langle S_{X-{E[X]}}, S_{X-{E[X]}}\rangle_\gamma = D^2_\gamma(X, E[X]). \end{equation} | (3.14) |
The next proposition is the extended interval version of Theorem 4.1 of Yang and Li (2005).
Proposition 3.3. For all random extended intervals X, Y, Z , the following hold:
① Var(C) = 0 , for every constant interval C ;
② Var(X+Y) = Var(X)+2Cov(X, Y)+Var(Y) ;
③ Cov(X, Y) = Cov(Y, X) ;
④ Cov(X+Y, Z) = Cov(X, Z)+Cov(Y, Z) ;
⑤ Cov(\lambda X, Y) = \lambda Cov(X, Y) ;
⑥ Var(\lambda X) = \lambda^2Var(X) , for every \lambda\in {\mathbb R} ;
⑦ P(d_\gamma(X, E[X])\geq\varepsilon)\leq Var(X)/\varepsilon^2 for every \varepsilon > 0 (Chebyshev inequality).
Proof. For any constant extended interval C , one has E[C] = C and Var(C) = 0 follows. Using the linearity of S and the form (3.12) of the metric d_\gamma , one proves items ②-⑥. The Chebyshev inequality follows from the fact that P(d_\gamma(X, E[X])\geq\varepsilon)\leq E[d_\gamma(X, E[X])^2]/\varepsilon^2 .
In the particular case of adapted measures, we have the following results, which are very useful in numerical simulations.
Proposition 3.4. If \gamma(t)dt is an adapted measure, a, b are random variables, and X is a random extended interval, then
① Var(\lfloor a, 0\rfloor) = Var(\lfloor0, a\rfloor) = Var(a) ;
② Var(\lfloor a, a\rfloor) = Var(a) ;
③ Cov(\lfloor a, 0\rfloor, \lfloor0, b\rfloor) = -\frac12Cov(a, b) ;
④ Var(X) = Var(\underline{X})-Cov(\underline{X}, \overline{X})+Var(\overline{X}) ;
⑤ Cov(X, Y) = Cov(\underline{X}, \underline{Y})+Cov(\overline{X}, \overline{Y})- \frac12Cov(\underline{X}, \overline{Y})-\frac12Cov(\underline{Y}, \overline{X}) ;
⑥ E\|X\|^2 = E[\underline{X}^2]+E[\overline{X}^2]-E[\underline{X}\overline{X}] .
Item ⑤ of the above proposition is similar to the one obtained for classical intervals in Example 4.1 of Yang and Li (2005), but the two last terms -\frac12Cov(\underline{X}, \overline{Y})-\frac12Cov(\underline{Y}, \overline{X}) are not present in the formula of Yang and Li. This difference can be explained by the fact that, for our distance d_\gamma , there is no preference between the left and the right bound, which is not the case for the distance d_p used by Yang and Li (2005). From the formula of Yang, if the left bounds of X, Y are independent and their right bounds are also independent then Cov(X, Y) = 0 , which is not the case for our formula ⑤ above.
Let \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})]_0 = \{X\in \mathcal{U}[\Omega, \mathcal{K}({\mathbb R})], E[X] = 0, \mbox{ and } E[\|X\|_\gamma^2] < \infty\} , that is, the sub-vector space of \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})] made by random extended interval with mean zero. For a random extended interval X\in \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})]_0 , Cov(X, X) = 0 means that X = E[X] = 0 almost everywhere. Hence, formula (3.14) cannot define a scalar product on \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})]_0 . We denote by L^2[\Omega, \mathcal{K}({\mathbb R})]_0 the set of classes of zero mean random extended intervals equal almost everywhere. We will keep denoting any class in L^2[\Omega, \mathcal{K}({\mathbb R})]_0 by a representative X\in \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})]_0 . L^2[\Omega, \mathcal{K}({\mathbb R})]_0 inherits from the structure of the vector space of \mathcal L^2[\Omega, \mathcal{K}({\mathbb R})]_0 , and for X, Y\in L^2[\Omega, \mathcal{K}({\mathbb R})]_0 , the formula (3.13) reads
\begin{equation} Cov(X, Y) = E\langle S_X, S_Y\rangle_\gamma = \int_\Omega\int_{0}^{1}\nabla_{\!\!X(\omega)} \nabla_{\!\!Y(\omega)}\gamma(t)dtdP(\omega) \end{equation} | (3.15) |
and is a scalar product on L^2[\Omega, \mathcal{K}({\mathbb R})]_0 .
Theorem 3.4. (L^2[\Omega, \mathcal{K}({\mathbb R})]_0, Cov) is a Hilbert space.
Proof. From what is written above, Cov is a scalar product on L^2[\Omega, \mathcal{K}({\mathbb R})]_0 . For the completeness, use the fact that \langle, \rangle_\gamma defined a scalar product on {\mathbb R}^2 .
Example 3.4. Take \Omega = {\mathbb R} , {\mathscr A} the Borel topology, and P = dx the Borel measure. Let us consider the random extended interval
\begin{equation} X = \lfloor f(\omega), g(\omega)\rfloor, \end{equation} | (3.16) |
the left and right bounds respectively,
\begin{align*} f(\omega)& = (1/\sqrt{2\pi})\exp(-0.5\omega^2)\\ g(\omega)& = 0.3\exp(-0.3\omega). \end{align*} |
We may write X\rightsquigarrow\mathcal NE(0, 1, 0.3) to say that the left bound of X follows the standard normal distribution and its right bound follows the exponential distribution with parameter 0.3 . The density functions of those random variables have been plotted on Figure 1.
Let (X_t)_{t\in {\mathbb Z}} be an extended interval time series; that is, for any integer t , X_t is an random extended interval. We denote by A_t the expectation of X_t and by C_t(j) = Cov(X_t, X_{t-j}) the auto-covariance function.
Definition 10. We say that an extended interval time series (X_t) is stationary when neither A_t nor C_t(j) depends on t . In this case, we just denote them A and C(j) , respectively.
For any n\in {\mathbb Z}^+ , the auto-covariance matrix is given by
\begin{equation} \mathbf{C}_n = (C(i-j))_{1\leq i, j\leq n} = \begin{pmatrix} C(0) & C(1) & \cdots & C(n-1)\\ C(1) & C(0) & \cdots & C(n-2)\\ \vdots & \vdots & \vdots & \vdots\\ C(n-1) & C(n-2) & \cdots & C(0) \end{pmatrix}. \end{equation} | (4.17) |
The proof of the following theorem is similar to the one of Theorem 4 in Wang et al., (2016).
Theorem 4.5. The auto-covariance function of any stationary process satisfies:
① C(k) = C(-k) for all k\in {\mathbb Z} ;
② |C(k)|\leq C(0) for all k\in {\mathbb Z} ;
③ the auto-covariance matrix \mathbf{C}_n is positive semi-definite;
④ if C(0) > 0 and (C(k)) converges to 0 then \mathbf{C}_n is positives definite.
Let X_1, \ldots, X_T be a sample of a stationary extended interval time series (X_t) with expectation A . An unbiased estimator of A is given by
\begin{equation} mX = \frac{X_1+\cdots+X_T}{T} \end{equation} | (4.18) |
and the sample-covariance is given by
\begin{equation} \widehat{C}(k) = \frac1{T}\sum\limits_{i = 1}^{T-|k|}\int_{0}^{1}(\nabla_{\!\!X_{i+|k|}}(t)- \nabla_{\!\!mX}(t))(\nabla_{\!\!X_{i}}(t)-\nabla_{\!\!mX}(t))\gamma(t)dt. \end{equation} | (4.19) |
Theorem 4.6. Let (X_t) be a stationary extended interval-valued time series with expectation A and auto-covariance function C(k) such that (C(k)) converges to 0 . Then, mX is a consistent estimator of A ; that is, for any \varepsilon > 0 , \lim\limits_{T\to\infty}P(d_\gamma(mX, A) \geq\varepsilon) = 0 .
Proof. One has
\begin{align*} Var(mX)& = D_\gamma^2(mX, A) = E\langle S_{mX-A}, S_{mX-A}\rangle_\gamma = \frac{1}{T^2}\sum\limits_{i, j = 1}^{T}E\langle S_{X_i-A}, S_{X_j-A}\rangle_\gamma\\ & = \frac{1}{T^2}\sum\limits_{i, j = 1}^{T}C(i-j) = \frac{1}{T^2}\sum\limits_{i-j = -T}^{T}(T-|i-j|) C(i-j) = \frac{1}{T}\sum\limits_{k = -T}^{T}\left(1-\frac{k}{T}\right)C(k). \end{align*} |
So, Var(mX) goes to 0 as T goes to infinity since (C(k)) converges to 0 . By the Chebyshev inequality, \forall\varepsilon > 0 , P(d_\gamma(m, A) \geq\varepsilon)\leq Var(mX)/\varepsilon^2 goes to 0 as T goes to infinity.
As usual, \widehat{C}(k) is not an unbiased estimator of {C}(k) (unless mX = A ), but:
Theorem 4.7. If (C(k)) converges to 0 as k goes to infinity, then for any k , \widehat{C}(k) is an asymptotically unbiased estimator of C(k) , that is \lim\limits_{T\to\infty}E[\widehat{C}(k)] = C(k) .
Proof.
\begin{align*} \widehat{C}(k)& = \frac1{T}\sum\limits_{i = 1}^{T-|k|}\int_{0}^{1}(\nabla_{\!\!X_{i+|k|}}(t)- \nabla_{\!\!mX}(t)) (\nabla_{\!\!X_{i}}(t)-\nabla_{\!\!mX}(t))\gamma(t)dt\\ & = \frac1{T}\sum\limits_{i = 1}^{T-|k|}\int_{0}^{1}(\nabla_{\!\!X_{i+|k|}}(t)-\nabla_{\!\!A}(t)) (\nabla_{\!\!X_{i}}(t) -\nabla_{\!\!A}(t))\gamma(t)dt +\frac1{T}\sum\limits_{i = 1}^{T-|k|}\int_{0}^{1}(\nabla_{\!\!mX}(t)-\nabla_{\!\!A}(t))^2 \gamma(t)dt\\ &-\frac1{T}\sum\limits_{i = 1}^{T-|k|}\int_{0}^{1}(\nabla_{\!\!mX}(t)-\nabla_{\!\!A}(t)) (\nabla_{\!\!X_{i+|k|}}(t) +\nabla_{\!\!X_{i}}(t)-2\nabla_{\!\!A}(t))\gamma(t)dt\\ \end{align*} |
Hence,
\begin{align*} \lim\limits_{T\to\infty} E[\widehat{C}(k)]& = \lim\limits_{T\to\infty}\frac1{T}\sum\limits_{i = 1}^{T-|k|}E[C(k)] +\lim\limits_{T\to\infty} \frac1{T}\sum\limits_{i = 1}^{T-|k|}Var(mX)\\&-\lim\limits_{T\to\infty}\frac1{T}\sum\limits_{i = 1}^{T-|k|} \left(Cov(mX, X_{i+|k|})+Cov(mX, X_{i})\right)\\ & = C(k)-\lim\limits_{T\to\infty}\frac1{T^2}\sum\limits_{i = 1}^{T-|k|}\sum\limits_{j = 1}^{T}\left(Cov(X_j, X_{i+|k|})+Cov(X_j, X_{i})\right)\\ & = C(k)-\lim\limits_{T\to\infty}\frac1{T^2}\sum\limits_{j-i = -T}^{T}(T-|j-i|)\left(C(j-i-|k|)+C(j-i)\right)\\ & = C(k)-\lim\limits_{T\to\infty}\frac1{T}\sum\limits_{l = -T}^{T}\left(1-\frac{|l|}{T}\right)\left(C(l-|k|)+C(l)\right) = C(k) \end{align*} |
Let (X_{t}) be an extended interval-valued stationary time series with expectation A , and auto-covariance function C(k) . To capture the dynamics of (X_{t}) one can assume that it follows an interval autoregressive moving-average (I-ARMA) process of order (p, q) , that is
\begin{equation} X_t = K+\sum\limits_{i = 1}^{p}\theta_iX_{t-i}+\varepsilon_t+\sum\limits_{i = 1}^{q}\phi_i\varepsilon_{t-i}, \end{equation} | (4.20) |
Where K is a constant extended interval, \phi_i and \theta_i are the parameters of the model, (\varepsilon_t)\rightsquigarrow IID(\{0\}, \sigma^2) , and for each t , \varepsilon_t is uncorrelated with the past of X_t . This model was introduced and studied by Han et al. (2012). They call such a model an Autoregressive Conditional Interval Model, and they proposed a D_K- distance based estimation method to estimate the parameters. Our interest in this method is to do forecasting and we propose a different estimation method, based on the Yule-Walker equation.
By taking expectation at the both sides of (4.20) one finds
\begin{equation} \lambda A = K, \end{equation} | (4.21) |
where \lambda = 1-\theta_1-\cdots-\theta_p . So, as in the case of real random variables, the expectation \mu_t of X_t does not depend on t and the new series X'_t = X_t-\frac{1}{\lambda}K is a zero-mean I-ARMA process, i.e. Equation (4.20) with K = 0 . In what follows, until the numerical study section, we assume that K = 0 , that is, (X_{t}) is a zero-mean stationary process. When p = 0 , the process (X_t) is called an extended interval-valued moving-average time series process of order q , I-MA( q ), and when q = 0 , one obtains an extended interval-valued autoregressive time series process of order p , I-AR( p ).
Let L be the delay operator, thus LX_t = X_{t-1} . Setting \Theta(L) = 1-\theta_1L-\cdots-\theta_{p}L^p and \Phi(L) = 1+\phi_1L+\cdots+\phi_{q}L^q , equation (4.20) can be written as
\begin{equation} \Theta(L)X_t = \Phi(L)\varepsilon_t. \end{equation} | (4.22) |
The functions \Theta and \Phi are called the autoregressive and moving-average polynomials, respectively.
In particular, if (X_t) is an I-MA( 1 ) process: X_t = \varepsilon_t+\phi\varepsilon_{t-1} , then
\begin{equation} C(1) = \phi\sigma^2. \end{equation} | (4.23) |
In section 5 we show that any non-deterministic zero-mean stationary random extended interval process can be expressed as a MA(\infty) .
If the moving-average polynomial \Phi = 1 , then (4.22) leads to
\begin{equation} X_t = (1-\Theta(L))X_t+\varepsilon_t, \end{equation} | (4.24) |
which is an extended interval-valued autoregressive process of order p , I-AR( p ). In this case, the existence and the uniqueness of a stationary solution is not guaranteed. However, when a stationary solution exits, using Proposition 3.3 it is simple to show that its auto-covariance function satisfies
\begin{equation} C(k)-\sum\limits_{i = 1}^{p}\theta_iC(k-i) = 0, \mbox{ for any } 1\leq k\leq p. \end{equation} | (4.25) |
Hence, the parameters of an I-AR( p ) process satisfy the Yule-Walker equation
\begin{equation} \mathbf{C}_p\mathbf{\Theta} = \mathbf{c}_p, \end{equation} | (4.26) |
where \mathbf{c}_p = (C(1), \ldots, C(p))^T , \mathbf{\Theta} = (\theta_1, \ldots, \theta_p)^T and \mathbf{C}_p is the auto-covariance matrix (4.17).
Theorem 4.8. Any AR( 1 ) process X_t = \theta X_{t-1}+\varepsilon_t , with 0 < \theta < 1 and \sup\limits_tE\|\varepsilon_t\| < \infty , possesses a unique stationary solution given by X_t = \sum_{i = 0}^\infty\theta^i\varepsilon_{t-i} .
Proof. One has
X_t = \theta X_{t-1}+\varepsilon_t = \theta^2X_{t-2}+\theta\varepsilon_{t-1}+\varepsilon_t = \theta^{n+1}X_{t-n-1}+\sum\limits_{i = 0}^{n}\theta^i\varepsilon_{t-i}. |
As 0 < \theta < 1 one has that \sum\theta^{2i} < \infty . This together with \sup_tE\|\varepsilon_t\| < \infty implies that (S_n = \sum_{i = 0}^{n}\theta^i\varepsilon_{t-i}) converges in probability under the metric d_\gamma by Corollary 3.1. Since (X_t) is stationary, Var(X_t) = E\|X_t\|^2 is constant and
\begin{align*} E\left\Vert X_t-\sum\limits_{i = 0}^{n}\theta^i\varepsilon_{t-i}\right\Vert^2& = E\Vert\theta^{n+1} X_{t-n-1}\Vert^2 = \theta^{2(n+1)}E\Vert X_{t-n-1}\Vert^2 \end{align*} |
goes to 0 as n goes to infinity. Hence, E\left\Vert X_t-\sum_{i = 0}^{\infty} \theta^i\varepsilon_{t-i}\right\Vert^2 = 0 . This implies that X_t = \sum_{i = 0}^{\infty}\theta^i\varepsilon_{t-i} a.e. From this solution, we have
Cov(X_{t+k}, X_t) = \sigma^2\sum\limits_{i = k}^{\infty}\theta^k\theta^{i-k} = \sigma^2 \frac{\theta^k}{1-\theta^2}. |
Now, if (X_t) is an I-ARMA( 1, 1 ) process: X_t = \theta X_{t-1}+\varepsilon_t+ \phi\varepsilon_{t-1} , then
\begin{equation} C(2) = {\theta}{C}(1)\qquad\mbox{and}\qquad C(1) = {\theta}{C}(0)+{\phi}\sigma^2. \end{equation} | (4.27) |
Let (X_t)_{t\in {\mathbb Z}} be a zero-mean extended interval-valued stationary process. The sets S_t = \overline{Span(\{X_k\}_{k = -\infty}^{t})} and S_{-\infty} = \mathop \bigcap\limits_{t = -\infty}^\infty S_t are Hilbert spaces of L^2[\Omega, \mathcal{K}({\mathbb R})]_0 . For any j\geq0 , the projection P_{S_{t-j}}X_t of X_t on S_{t-j} is called the prediction of X_t on S_{t-j} . We shall say that an extended interval-valued process (X_t)_{t\in {\mathbb Z}} is deterministic if for any t\in {\mathbb Z} , X_t\in S_{t-1} . X_t- P_{S_{t-1}}X_t is called the error in the projection of X_t on S_{t-1} and when P_{S_{t-1}}X_t = X_t one says that (X_t)_{t\in {\mathbb Z}} is (perfectly) predictable. As (L^2[\Omega, \mathcal{K}({\mathbb R})]_0, Cov) is a Hilbert space, we have the following Wold decomposition for extended interval time series.
Theorem 5.9. Let (X_t)_{t\in {\mathbb Z}} be a non-deterministic extended interval-valued stationary time series process with expectation \{0\} and auto-covariance function (C(k)) . Then, X_t can be expressed as
\begin{equation} X_t = \sum\limits_{k = 0}^{\infty}\alpha_k\varepsilon_{t-k}+W_t\; \mathit{\mbox{a.s}} \end{equation} | (5.28) |
where:
(i) \alpha_k = \frac{1}{\sigma^2}Cov(X_t, \varepsilon_{t-k}), \quad\alpha_0 = 1 and \sum\limits_{k = 0}^{\infty}\alpha_k^2 < \infty ;
(ii) \{\varepsilon_t\}\rightsquigarrow WN(\{0\}, \sigma^2) , with \sigma^2 = Var(X_t-P_{S_{t-1}}X_t) ;
(iii) Cov(W_t, \varepsilon_s) = 0 for all t, s\in {\mathbb Z} ;
(iv) (W_t)_{t\in {\mathbb Z}} is zero-mean, stationary and deterministic.
Proof. For any t\in {\mathbb Z} , the application of Theorem 4 in Bierens (2012) to the regular sequence (X_{t-k})_{k = 0}^\infty gives that X_t can be expressed as
\begin{equation} X_t = \sum\limits_{k = 0}^{\infty}\theta_ke_{t-k}+W_t\; \mbox{ a.s}\qquad \end{equation} | (5.29) |
where \{e_{t-k}\}_{k = 0}^\infty is an uncorrelated process with Cov(e_i, e_j) = \delta_{ij} , \theta_k = Cov(X_t, e_{t-k}) , and \sum\limits_{k = 1}^{\infty}\theta_k^2 < \infty , W_t\in U_{t}^\perp with U_{t} = \overline{Span(\{e_k\}_{k = -\infty}^t)}\subset S_t . Since the process (X_t)_{t\in {\mathbb Z}} is non-deterministic, the residual \varepsilon_t = X_t-P_{S_{t-1}} X_t is different from 0 and \varepsilon_t = \|\varepsilon_t\|e_t , hence (5.28) holds with \alpha_k = \theta_k/\|\varepsilon_{t-k}\| , and (\varepsilon_t) is also uncorrelated. As W_t, \varepsilon_t\in L^2[\Omega, \mathcal{K}({\mathbb R})]_0 , and E[W_t] = 0 = E[\varepsilon_t] . W_t\in U_t^\perp implies that Cov(W_t, \varepsilon_s) = 0 for any s\leq t . For s > t , taking the scalar product of (5.29) with \varepsilon_s , one has Cov(W_t, \varepsilon_s) = Cov(X_t, \varepsilon_s) = 0 since \varepsilon_s\in S_{s-1}^\perp and X_t\in S_t\subset S_{s-1} for s > t . This proves (iii). Let X_{t, n} be the projection of X_t on S_{t, n} = span(\{X_{t-j}\}_{j = 1}^n) , and \varepsilon_{t, n} the residual. Then, X_{t, n} takes the form
X_{t, n} = \sum\limits_{j = 1}^{n}\beta_{j, n}X_{t-j}, |
where the scalars \beta_{k, n} do not depend on t , since they are solutions of the system of equations
\sum\limits_{j = 1}^{n}\beta_{j, n}C(j-k) = C(k), \qquad k = 1, \ldots, n. |
Hence, E[X_{t, n}] = 0 , E[\varepsilon_{t, n}] = 0 . Moreover,
\begin{align*} Var(\varepsilon_{t, n})& = \|X_t-X_{t, n}\|^2 = \left\|X_t-\sum\limits_{j = 1}^{n}\beta_{j, n}X_{t-j}\right\|^2\\ & = C(0)+\sum\limits_{i, j = 1}^{n}\beta_{i, n}\beta_{j, n}C(i-j)-2\sum\limits_{j = 1}^{n}\beta_{j, n}C(j). \end{align*} |
Hence, Var(\varepsilon_{t, n}) = \sigma_n does not depend on t and same for \sigma = \|\varepsilon_t\| = \lim\limits_{n\to\infty}\sigma_n . Also,
Cov(X_{t+k}, \varepsilon_{t, n}) = C(k)-\sum\limits_{j = 1}^{n}\beta_{j, n}C(k+j), |
which does not depend on t . Using the Cauchy-Schwarz inequality,
\lim\limits_{n\to\infty}|Cov(X_{t+k}, \varepsilon_{t, n}-\varepsilon_t)|\leq\sqrt{C(0)} \lim\limits_{n\to\infty}\|\varepsilon_{t, n}-\varepsilon_t\| = 0, |
which implies that Cov(X_{t+k}, \varepsilon_{t}) = \lim\limits_{n\to\infty}Cov(X_{t+k}, \varepsilon_{t, n}) and does not depend on t . So,
\alpha_k = \frac{1}{\|\varepsilon_t\|} Cov(X_{t+k}, e_k) = \frac{1}{\|\varepsilon_t\|^2}Cov(X_{t+k}, \varepsilon_t) |
does not depend on t . Moreover, \alpha_0 = \frac{Cov(X_t, \varepsilon_t)}{\|\varepsilon_t\|^2} = 1 . All this completes the proof of (i) and (ii). For k\geq0 ,
\begin{align*} Cov(W_t, W_{t-k})& = Cov\left(X_{t-k}-\sum\limits_{j = 0}^{\infty}\alpha_j\varepsilon_{t-k-j}, X_t -\sum\limits_{j = 0}^{\infty} \alpha_j\varepsilon_{t-j}\right)\\ & = C(k)-\sum\limits_{j = 0}^{\infty}\alpha_jCov(X_t, \varepsilon_{t-k-j})-\sum\limits_{j = k}^{\infty} \alpha_jCov(X_{t-k}, \varepsilon_{t-j})+\sigma^2\sum\limits_{j = 0}^{\infty}\alpha_{j+k}\alpha_{j}\\ & = C(k)-\sigma^2\sum\limits_{j = 0}^{\infty}\alpha_{j+k}\alpha_{j}, \end{align*} |
which does not depend on t . As W_t\in S_t , one can write W_t = \sum_{k = 0}^{\infty}a_kX_{t-k} . Taking the covariance with \varepsilon_t and using the fact that \varepsilon_t\perp Span(X_{t-1}, X_{t-2}, \ldots) , one gets Cov(W_t, \varepsilon_t) = a_0Cov(X_t, \varepsilon_t) = a_0\|\varepsilon_t\|^2 . Since Cov(W_t, \varepsilon_t) = 0 , one deduces that a_0 = 0 , hence W_t\in S_{t-1} , and thus (W_t) is deterministic from the past of (X_t) . This completes the proof of (iv).
Let (X_t) be an AR( 1 ) process:
\begin{equation} X_t = K+\theta X_{t-1}+\varepsilon_t. \end{equation} | (6.30) |
Then, from the Yule-Walker equation, the parameter \theta can be estimated by \widehat{\theta} = \frac{\widehat{C}(1)}{\widehat{C}(0)} with
\begin{align*} \widehat{C}(0)& = \frac1T\sum\limits_{i = 1}^{T}\int_{0}^{1}(\nabla_{\!\!X_{i}}-\nabla_{\!\!{mX}})^2 \gamma(t)dt = \frac1T\sum\limits_{i = 1}^{T}d_\gamma^2(X_i, mX), \\ \widehat{C}(1)& = \frac1T\sum\limits_{i = 1}^{T-1}\int_{0}^{1}(\nabla_{\!\!X_{i+1}}-\nabla_{\!\!{mX}}) (\nabla_{\!\!X_{i}}-\nabla_{\!\!m{X}})\gamma(t)dt\\& = \frac1{2T}\sum\limits_{t = 1}^{T-1}\left (d_\gamma^2(X_{i+1}, mX)+d_\gamma^2(X_i, mX)-d_\gamma^2(X_{i+1}, X_i)\right), \end{align*} |
where \widehat{C}(1) and \widehat{C}(0) are the sample-covariance.
More generally, if we assume that the I-AR( p ) process (4.24) is stationary, then from Theorem 4.5, when C(0) > 0 and (C(k)) converges to 0 , the Yule-Walker equation (4.26) is well-posed and from a large sample X_1, \ldots, X_T , the coefficients of the I-AR( p ) process can be estimated by
\widehat{\mathbf{\Theta}} = \widehat{\mathbf{C}}_p\widehat{\mathbf{c}}_p. |
Using (3.10) and (4.19), the sample-covariance can be written as
\begin{equation} \widehat{C}(k) = \frac1{2T}\sum\limits_{i = 1}^{T-|k|}\left(d_\gamma^2(X_{i+k}, mX)+ d_\gamma^2(X_i, mX)-d_\gamma^2(X_{i+k}, X_i)\right). \end{equation} | (6.31) |
It is natural to assume that \gamma(t)dt is an adapted measure and, in this case, the distance d_\gamma is given by Lemma 3.1 and is easy to numerically compute.
Extended intervals can be very useful for displaying data. The plot of just one extended interval A gives much informations: (a) the range of values of the considered index during the recording; (b) the direction of variation of the considered index : decreasing when the arrow is pointing down, and increasing when the arrow is pointing up.
Figure 2 displays systolic (in blue) and diastolic (in red) blood pressure of a person recorded in the morning (left bounds) and in the afternoon (right bounds), over 4 days in 2004. One sees easily that on the 11/03/04, the blood pressure recorded in the morning is higher than the one recorded in the afternoon, both for systolic and diastolic.
Now, we plot the model (6.30) with \theta = 0.2 , and K = [13.31, 14.2] , \overline{\varepsilon_t} and \underline{\varepsilon_t} following independent standard normal distributions. Figure 3(a) shows a sample for this model for T = 100 , when the interval standard normal distribution used is the one plotted on Figure 3(b). One sees that most of the outputs of this sample are standard intervals (71 standard intervals versus 29 decreasing ones) while for the error (interval standard normal distribution), they seem to be the same number (41 standard intervals versus 59 decreasing). Figure 4 displays the estimated auto-covariance function C(k) and shows that it goes to 0 as k becomes large. Also, K is estimated using the formula \widehat{K} = (1-\widehat{\theta})mX .
T | \widehat{K} | C(T-2) | \widehat{\theta} | Error |
100 | [13.31, 14.2] | -0.02807759 | 0.1747072 | 0.02529285 |
500 | [13.51569, 14.41001] | 0.01240641 | 0.1892873 | 0.01071265 |
In Figure 5, we have plotted as standard min-max intervals (in blue) and open-close extended intervals (in red), the CAC 40 Stock Index from January 2nd to May 31st, 2019 (105 trading days). Extended intervals are formed by the opening values (left bounds) and the closing values (right bounds). This figure shows that most often, neither opening nor closing values are the lowest or the highest value of the index for the day. Notice that, in such an index, what is important most often is not just the opening and closing values, but also to know how it has been fluctuating along the day. For instance, the plot shows many days where the opening value and closing value are the same with fluctuation throughout the day. Now, we wish to find the I-ARMA model which best fits this data. The first step is to test stationarity. The augmented Dickey–Fuller test shows that neither the data nor its first difference are stationaries, but its second difference is stationary. So, we take the second difference data and use AIC to determine the optimal order (p, q) . We define the AIC of the random interval to be the summation of the AIC of the bounds, and we assume that p, q = 1, 2, 3, 4 . Figure 6 shows that the optimal order is p = q = 1 . Finally, using equation (4.27), we estimated the coefficients of the I-ARMA model by \widehat{\theta} = \frac{\widehat{C}(2)}{\widehat{C}(1)}, \widehat{\phi} = \widehat{C}(1)-\widehat{\theta}\widehat{C}(0) and we found
\begin{equation} \widehat{\theta} = -0.2519991\qquad\mbox{and}\qquad\widehat{\phi} = -0.5326387. \end{equation} | (6.32) |
Figure 7 shows the forecast of the differentiated CAC 40 for the next 40 trading days. From this graph, it appears that the sense of variation of CAC 40 throughout the day has been well predicted for 25 days over 40. Also, the predicted arrow is most often on top of the real value. This prediction, for sure, can be improved by using extended intervals with non-linear estimation methods.
In this paragraph we present how data are usually pretreated and show that this process can be better performed when one wishes to use extended intervals.
Let us consider an index ID (for example the French CAC40 index) that we try to model for predicting future values. Let us assume that the values of this index are changing every minute and that we want to analyze it over one year. That will make a huge set of data to analyze if we consider every single value of the index.
What people do most often is to consider a frequency; in the case of ID, one can decide to analyze daily values. But, we have something like 1440 values every day and have to decide for the value of the day. In point-value analysis, people consider either the opening value or the closing value or the average value of the day as the value of that day. It is clear that a lot of values have been neglected and this could lead to an inconsistent analysis.
In an analysis with the standard interval, people most often consider the highest and lowest values of a day to form the interval representing the value of the index that day. (See for example, Wang and Li (2011).) By so doing, every interval contains all the values of the index that day. But, the interval can be irreasonably large and does not reflect the variations of the index during the day. One can still do better by using extended intervals.
With extended intervals, one can proceed as follows. The first value is the left bound of the first interval. If the next value is smaller (resp. bigger) then we keep looking for the next value until either the index is no longer significantly decreasing (resp. increasing), or we have passed 1440 values (the period cannot exceed 1 day). The right bound of the first interval is then the previous value recorded, and the actual value is the left bound of the second interval, and we repeat this process until the end of the data set. This process is summarized as Algorithm 1, which returns the sequence Res of extended intervals obtained and the corresponding sequence of time intervals. There is a need to explain when we say "corresponding sequence of time intervals". The left bound of the first time interval is the time when the left bound of the first extended interval of Res has been recorded, and so on.
By applying this algorithm, we do not have a regular period, which is needed for a time series analysis. The period can be taken here as the average of the periods of extended intervals obtained.
We have implemented Algorithm 1 in R and test it on the CAC 40 stock index recorded minute by minute during five days: from June 22, to June 26, 2020. After treating the 2169 data, we obtained 787 extended intervals as shown in Figure 8. The initial data was recorded everyday from 9:00 am to 6:05 pm, except the last day which ends at 10:52 am. So, the total time of recording was 38 hours and 12 minutes. As we obtained 787 extended intervals, we can take as period for time series analysis: 3 minutes. We then assume that every extended interval is recorded during a lap time of 3 minutes.
Observe that the minimum value per day as well as the maximum value of the CAC 40 during the five days we considered is the same. So, those data could not be analyzed with min-max standard intervals. In Figure 9 are have plotted the extended intervals that we obtained.
Algorithm 1 Transform point-values to extended intervals |
Require: data, time, \varepsilon , frequency = 1440 Res\leftarrow\{\} , ResTime\leftarrow\{\} N\leftarrow length(data) , i\leftarrow1 \underline{A}\leftarrow data[i] , \underline{T}\leftarrow time[i] while i < N do if data[i+1]\leq data[i] then i\leftarrow i+1 , j\leftarrow1 while the index is decreasing or is not significantly (use \varepsilon ) increasing do i\leftarrow i+1 , j\leftarrow j+1 if j > frequency then break end if end while \overline{A}\leftarrow data[i] , \overline{T}\leftarrow time[i] add A = \lfloor\underline{A}, \overline{A}\rfloor in Res and T = \lfloor\underline{T}, \overline{T}\rfloor in Restime i\leftarrow i+1 end if if data[i+1] > data[i] then i\leftarrow i+1 , j\leftarrow1 while the index is increasing or is not significantly (use \varepsilon ) decreasing do i\leftarrow i+1 , j\leftarrow j+1 if j > frequency then break end if end while \overline{A}\leftarrow data[i] , \overline{T}\leftarrow time[i] add A = \lfloor\underline{A}, \overline{A}\rfloor in Res and T = \lfloor\underline{T}, \overline{T}\rfloor in ResTime i\leftarrow i+1 end if end while return Res and ResTime |
In this work, we have redefined extended intervals in a more natural manner and written an algorithm to efficiently transform point-valued data to extended interval-valued data. An extended interval is a standard interval endowed with a direction \alpha , which is an element of the Abelian group {\mathbb Z}_2 = \{0, 1\} . The direction 0 means you move on the real line from the left to the right, and the direction 1 means you move from the right to the left. This process can be generalized on {\mathbb R}^n . For example, one could define extended rectangles on {\mathbb R}^2 with 4 directions represented by the Abelian group {\mathbb Z}_4 .
We have seen that by using extended intervals to record the values of a given index, every extended interval gives the value of the index and the direction of variation at the time of recording. We have proposed a language that we hope will be use in the future in the trading markets. Precisely, talking about the French CAC40 index, if we say that we got 4922^- today, this would mean that we got a value of 4922 and the index was decreasing when we got this value. This is an example of how this new structure of extended intervals can be very useful in the context of trading markets, and more. A suitable distance has been defined on extended intervals and used to define variance and covariance on random extended intervals, in a natural way. We have studied ARMA processes with extended intervals both theoretically and numerically. In the numerical part, we forecasted on CAC 40 stock index from January 2nd to July 26, 2019.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
All authors declare no conflicts of interest in this paper.
[1] | V. Hárs and J. Tóth, On the inverse problem of reaction kinetics, In M. Farkas, editor, Colloquia Mathematica Societatis János Bolyai, volume 30, pages 363-379. Qualitative Theory of Differential Equations, 1979. |
[2] | C. P. P. Arceo, E. C. Jose, A. Marin-Sanguino, et al., Chemical reaction network approaches to biochemical systems theory, Math. Biosci., 269 (2015), 135-152. |
[3] | F. Horn and R. Jackson, General mass action kinetics, Arch. Ratl. Mech. Anal., 47 (1972), 81-116. |
[4] | M. Feinberg, Complex balancing in general kinetic systems, Arch. Ratl. Mech. Anal., 49 (1972), 187-194. |
[5] | F. Horn, Necessary and sufficient conditions for complex balancing in chemical kinetics, Arch. Ratl. Mech. Anal., 49 (1972), 172-186. |
[6] | A. I. Volpert and S. I. Hudyaev, Analyses in Classes of Discontinuous Functions and Equations of Mathematical Physics, Martinus Nijhoff Publishers, Dordrecht, 1985. Russian original: 1975. |
[7] | D. F. Anderson, A proof of the global attractor conjecture in the single linkage class case, SIAM J. Appl. Math., 71 (2011), 1487-1508. |
[8] | C. Pantea, On the persistence and global stability of mass-action systems, SIAM J. Math. Anal., 44 (2012), 1636-1673. |
[9] | M. Gopalkrshnan, E. Miller and A. Shiu, A geometric approach to the global attractor conjecture, SIAM J. Appl. Dyn. Syst., 13 (2014), 758-797. |
[10] | G. Craciun, F. Nazarov and C. Pantea, Persistence and permanence of mass-action and power-law dynamical systems, SIAM J. Appl. Math., 73 (2013), 305-329. |
[11] | G. Craciun, Toric differential inclusions and a proof of the global attractor conjecture, arXiv:1501.02860, 2016. |
[12] | R. Aris, Prolegomena to the rational analysis of systems of chemical reactions, Archive Ration. Mech. An., 19 (1965), 81-99. |
[13] | R. Aris, Mathematical aspects of chemical reaction, IEEC Fundamentals, 61 (1969), 17-29. |
[14] | M. Dukarić, H. Errami, R. Jerala, et al., On three genetic repressilator topologies, React. Kinet. Mech. Cat., 126 (2019), 3-30. |
[15] | D. Lichtblau, Symbolic analysis of multiple steady states in a MAPK chemical reaction network, J. Symb. Comp., 2018. submitted. |
[16] | B. Boros, On the existence of positive steady states for weakly reversible mass-action systems, SIAM J. Math. Anal., 51 (2019), 435-449. |
[17] | M. Feinberg, Foundations of Chemical Reaction Network Theory, Springer International Publishing, New York, 2019. |
[18] | G. Craciun and P. Y. Yu, Mathematical analysis of chemical reaction systems, Isr. J. Chem., 50, 2018. |
[19] | J. Tóth, A. L. Nagy and D. Papp, Reaction Kinetics: Exercises, Programs and Theorems, Mathematica for Deterministic and Stochastic Kinetics, Springer-Verlag, New York, 2018. |
[20] | G. Lente, Deterministic kinetics in chemistry and systems biology: the dynamics of complex reaction networks, Springer, 2015. |
[21] | M. Feinberg and F. J. M. Horn, Chemical mechanism structure and the coincidence of the stoichiometric and kinetic subspaces, Arch. Ratl. Mech. Anal., 66 (1977), 83-97. |
[22] | G. Craciun and C. Pantea, Identifiability of chemical reaction networks, J. Math. Chem., 44 (2008), 244-259. |
[23] | G. Craciun, J. Jin and P. Y. Yu, An efficient characterization of complex-balanced, detailedbalanced, and weakly reversible systems, SIAM J. Appl. Math., 2019. To appear. |
[24] | P. Érdi and J. Tóth, Mathematical Models of Chemical Reactions. Theory and Applications of Deterministic and Stochastic models, Princeton University Press, Princeton, New Jersey, 1989. |
[25] | G. Szederkényi, Comment on "identifiability of chemical reaction networks" by G. Craciun and C. Pantea, J. Math. Chem., 45 (2009), 1172-1174. |
[26] | G. Szederkényi, Computing sparse and dense realizations of reaction kinetic systems, J. Math. Chem., 47 (2010), 551-568. |
[27] | G. Szederkényi, K. M. Hangos and T. Péni, Maximal and minimal realizations of reaction kinetic systems: computation and properties, MATCH Commun. Math. Comput. Chem., 65 (2011), 309-332. |
[28] | B. Ács, G. Szederkényi, Z. A. Tuza, et al., Computing linearly conjugate weakly reversible kinetic structures using optimization and graph theory, MATCH Commun. Math. Comput. Chem., 74 (2015), 489-512. |
[29] | G. Ács, G. Szlobodnyik and G. Szederkényi, A computational approach to the structural analysis of uncertain kinetic systems, Comput. Physics Commun., 228 (2018), 83-95. |
[30] | G. Szederkényi, A. Magyar and K. M. Hangos, Analysis and control of polynomial dynamic models with biological applications, Academic Press, London, San Diego, Cambridge, MA, Oxford, 2018. |
[31] | J. Tóth, A formális reakciókinetika globális determinisztikus és sztochasztikus modelljéröl (On the global deterministic and stochastic models of formal reaction kinetics with applications), MTA SZTAKI Tanulmányok, 129 (1981), 1-166. In Hungarian. |
[32] | G. Lipták, G. Szederkényi and K. M. Hangos, Computing zero deficiency realizations of kinetic systems, Syst. Control Lett., 81 (2015), 24-30. |
[33] | G. Szederkényi and K. M. Hangos, Finding complex balanced and detailed balanced realizations of chemical reaction networks, J. Math. Chem., 49 (2011), 1163-1179. |
[34] | M. Feinberg, Necessary and sufficient conditions for detailed balancing in mass action systems of arbitrary complexity, Chem. Eng. Sci., 44 (1989), 1819-1827. |
[35] | V. N. Orlov and L. I. Rozonoer, The macrodynamics of open systems and the variational principle of the local potential II. Applications, J. Franklin Ins., 318 (1984), 315-347. |
[36] | B. Joshi and A. Shiu, A survey of methods for deciding whether a reaction network is multistationary, Math. Model. Nat. Pheno., 10 (2015), 47-67. |
[37] | G. Szederkényi, K. M. Hangos and Z. Tuza, Finding weakly reversible realizations of chemical reaction networks using optimization, MATCH Commun. Math. Comput. Chem., 67 (2012), 193-212. |
[38] | M. D. Johnston, D. Siegel and G. Szederkényi, Computing weakly reversible linearly conjugate chemical reaction networks with minimal deficiency, Math. Biosci., 241 (2013), 88-98. |
[39] | S. Schuster and R. Schuster, Detecting strictly detailed balanced subnetworks in open chemical reaction networks, J. Math. Chem., 6 (1991), 17-40. |
[40] | M. D. Johnston, D. Siegel and G. Szederkényi, Dynamical equivalence and linear conjugacy of chemical reaction networks: new results and methods, MATCH Commun. Math. Comput. Chem., 68 (2012), 443-468. |
[41] | M. D. Johnston, D. Siegel and G. Szederkényi, A linear programming approach to weak reversibility and linear conjugacy of chemical reaction networks, J. Math. Chem., 50 (2012), 274-288. |
[42] | J. Rudan, G. Szederkényi, K. Hangos, et al., Polynomial time algorithms to determine weakly reversible realizations of chemical reaction networks, J. Math. Chem., 52 (2014), 1386-1404. |
[43] | D. Csercsik, G. Szederkényi and K. M. Hangos, Parametric uniqueness of deficiency zero reaction networks, J. Math. Chem., 50 (2012), 1-8. |
[44] | G. Craciun, J. Jin and P. Y. Yu, Uniqueness of kinetic realizations for weakly reversible deficiency zero networks, In preparation. |
[45] | B. Boros and J. Hofbauer, Permanence of weakly reversible mass-action systems with a single linkage class, arXiv:1903.03071, 2019. |
[46] | L. Cardelli, M. Tribastone and M. Tschaikowski, From electric circuits to chemical networks, arXiv:1812.03308, 2018. |
[47] | D. Csercsik, G. Szederkényi and K. M. Hangos, Parametric uniqueness of deficiency zero reaction networks. J. Math. Chem., 50 (2012), 1-8. |
[48] | J. Rudan, G. Szederkényi and K. M. Hangos, Efficient computation of alternative structures for large kinetic systems using linear programming, MATCH Commun. Math. Comput. Chem., 71 (2014), 71-92. |
[49] | J. Rudan, G. Szederkényi, K. M. Hangos, et al., Polynomial time algorithms to determine weakly reversible realizations of chemical reaction networks, J. Math. Chem., (2014), 1-19. |
[50] | G. Szederkényi, K. M. Hangos and D. Csercsik, Computing realizations of reaction kinetic networks with given properties, In A. N. Gorban and D. Roose, editors, Coping with Complexity: Model Reduction and Data Analysis, volume 75, pages 253-267. Springer, 2010. |
[51] | J. Rudan, G. Szederkényi and K. M. Hangos, Computing dynamically equivalent realizations of biochemical reaction networks with mass conservation, In ICNAAM 2013: 11th International Conference of Numerical Analysis and Applied Mathematics, 21-27 September, Rhodes, Greece, AIP Conference Proceedings, volume 1558, pages 2356-2359, 2013. ISBN: 978-0-7354-1184-5. |
T | \widehat{K} | C(T-2) | \widehat{\theta} | Error |
100 | [13.31, 14.2] | -0.02807759 | 0.1747072 | 0.02529285 |
500 | [13.51569, 14.41001] | 0.01240641 | 0.1892873 | 0.01071265 |