We study the existence of weak solutions to a Cahn-Hilliard-Darcy system coupled witha convection-reaction-diffusion equation through the fluxes, through the source terms and in Darcy’slaw. The system of equations arises from a mixture model for tumour growth accounting for transportmechanisms such as chemotaxis and active transport. We prove, via a Galerkin approximation, theexistence of global weak solutions in two and three dimensions, along with new regularity results forthe velocity field and for the pressure. Due to the coupling with the Darcy system, the time derivativeshave lower regularity compared to systems without Darcy flow, but in the two dimensional case weemploy a new regularity result for the velocity to obtain better integrability and temporal regularityfor the time derivatives. Then, we deduce the global existence of weak solutions for two variantsof the model; one where the velocity is zero and another where the chemotaxis and active transportmechanisms are absent.
1.
Introduction
Record statistics are interest and importance in many areas of real-life applications including data relating to meteorology and weather, economics, sport, athletic events, oil, mining surveys and life testing studies and so on. In the midst of practical life there are many situations that depend entirely on the record statistics (record values) involving: Guinness World Records where only record values are observed (lower or upper record values). Shortest ever tennis matches both in terms of number of games (number of runs) and duration in terms of time (lower record values). News items like fastest time taken to recite the periodic table of the elements (lower record values). Weightlifting competitions where the highest weight is lifted in the fewest number of attempts (upper record values). Swimming competitions where the swimmer travels a certain distance in the least time of swimming (lower record values). Mountaineering, fastest marathon, longest time to hop, etc are of great importance to sports federations and athletes.
From the above, we note that to create a record, several attempts are made, and when the attempt is successful, the record is created, and accordingly we do not get the data for all the attempts that were made to break records. The data that we have are records. Many scientists specially the statisticians have become interested in record values and discussed the statistical inferences for different models based on lower and upper record values including Chandler [1] studied the distribution and frequency of lower record values, Resnick [2] compared the behavior of the sequence of record values, Shorrock [3] proposed asymptotic results on record values and inter-record times, Glick [4] discussed the breaking records and breaking boards, Nevzorov [5] and Nagaraja [6] listed some theoretical characteristics for record values for various univariate model, Kumar [7] developed some recurrence relations for generalized logistic distribution based on lower record values, Mahmoud et al. [8] studied the estimation problem for the behavior of the coefficient of variation of Lomax distribution based on upper record values, Kumar and Saran [9] investigated the ratio and inverse moments of record values from Marshall-Olkin log-logistic distribution, and EL-Sagheer [10] studied the estimation of the parameters for generalized logistic distribution based on the lower record values. Arnold et al. [11] summarized the study of upper and lower record values and presented many of the basic properties of them as follows: Assume X1,X2,X3,... a sequence of independent and identically distributed random variables with probability density function (PDF) f(x) and cumulative distribution function (CDF) F(x). Setting Yj=max(X1,X2,X3,...,Xj), for j≥1, we say Xj an upper record and denoted by XU(j) if Yj > Yj−1 , j>1. Also, let Yj=min(X1,X2,X3,...,Xj), for j≥1 we say that Xj is a lower record and denoted by XL(j) if Yj < Yj−1 , j>1. Thus the joint density function of the first n upper record values is given by
and the joint density function of the first n lower record values is given by
Dagum [12] introduced a statistical distribution closely fitting income distributions which accommodating the heavy tails, called the Dagum distribution (DD). In the actuarial literature, is also called the inverse Burr. Furthermore, DD admits a mixture representation in terms of generalized Gamma and inverse Weibull distributions. The DD can also be obtained as a compound generalized Gamma distribution whose scale parameter follows an inverse Weibull distribution with identical shape parameters. Since Dagum introduced his model as income distribution, many authors have discussed its properties in economics and financial fields, reliability, and survival data, etc. See for instance, Kleiber and Kotz [13], Kleiber [14], Domma et al. [15], Domma and Condino [16], among others. Recently, Alotaib et al. [17] discussed parameter estimation for the DD based on progressive type-Ⅰ interval censored data. A random variable X has the Dagum distribution and denoted by X∼D(λ,θ,β), if it has PDF, CDF, reliability function r(x) and hazard rate function h(x), given by
and
where β, θ>0 are the shape parameters and λ>0 is the scale parameter.
In this article, we investigate the Bayes and maximum likelihood estimates for the unknown quantities of the DD using lower record values. The Lindley's approximation and the Metropolis-Hasting (M-H) algorithm within Gibbs sampler are proposed to construct the Bayes estimates. To this end, the results of the Bayes estimates are obtained under both squared error (SE) and linear exponential (LINEX) loss functions. Also, the corresponding credible intervals (CRIs) are constructed under MCMC technique.
The rest of this article is organized as follows. Section 2 deals with the maximum likelihood estimates and asymptotic confidence intervals. Bayesian estimates using Lindley's approximation and MCMC technique are provided in Section 3. In Section 4 a simulation study is provided to examine the performance of the proposed estimation methods. Section 5 discusses an application to real life data. Finally, a brief conclusion is given in Section 6.
2.
Maximum likelihood estimation
In this section, we discuss the maximum likelihood estimates (MLEs) and approximate confidence intervals (ACIs) of the parameters as well as reliability and hazard rate functions of DD given in (3) when the available data are lower record values. Suppose that x_=xL(1),xL(2),xL(3),...,xL(n) be the lower record values of size n from D(λ,θ,β), then the likelihood function for observed record x_ is given by
The logarithm of the likelihood function may then be written as
The estimators ˆλ, ˆθ and ˆβ of the parameters λ, θ, and β, respectively, can be then obtained as the solution likelihood equations
and
Since (9)–(11) cannot be solved analytically for ˆλ, ˆθ and ˆβ, some numerical methods such as Newton-Raphson method must be employed to obtain the desired MLEs in such situations. Once we get desired MLEs of λ, θ and β using the invariant property of MLEs, for given mission time t, the MLEs of r(t) and h(t) can be obtained after replacing λ, θ and β by ˆλ, ˆθ and ˆβ as
2.1. Asymptotic confidence intervals
From the log-likelihood function ℓ(λ,θ,β|x_) in (8), we have
The Fisher information matrix I(λ,θ,β) is then obtained by taking expectation of negative Eqs (13)–(17). Based on some mild regularity conditions, (ˆλ,ˆθ,ˆβ) is approximately bivariate normal with mean (λ,θ,β) and covariance matrix I−1(λ,θ,β). Practically, we usually estimate I−1(λ,θ,β) by I−1(ˆλ,ˆθ,ˆβ). Moreover, a simpler and equally valid procedure is to use the approximation
where
ACIs for λ, θ and β can be obtained by to be bivariate normal distributed with mean (λ,θ,β) and variance-covariance matrix I−1∘(λ,θ,β). Therefore, the 100(1−γ)% ACIs for λ, θ and β, respectively, are
where Λ11, Λ22 and Λ33 are the elements on the main diagonal of the variance-covariance matrix I−1∘(λ,θ,β), and zγ/2 is the percentile of the standard normal distribution with right-tail probability γ/2.
In order to find the approximate estimates of the variance of r(t) and h(t), we use the delta method, see Agresti [18]. Assume that G1=(∂r(t)∂λ,∂r(t)∂θ,∂r(t)∂β) and G2=(∂h(t)∂λ,∂h(t)∂θ,∂h(t)∂β), where ∂r(t)∂λ, ∂r(t)∂θ, ∂r(t)∂β, ∂h(t)∂λ, ∂h(t)∂θ, and ∂h(t)∂β are the first partial derivatives of r(t) and h(t) with respect to the parameters λ, θ and β, respectively. The approximate asymptotic variances of ˆr(t) and ˆh(t) are given by
where and GTi is the transpose of Gi, i=1,2. Thus, the (1−γ)100% ACIs for ˆr(t) and ˆh(t) can be written as
3.
Prior information and Bayesian estimation
Besides being an accurate analysis and not easy to handle mathematically, the use of the Bayesian method allows the incorporation of previous knowledge of the parameters through informative prior densities. In the Bayesian approach, the information referring to the model parameters is obtained through posterior marginal distributions. In this section, two Bayesian inference procedures (Lindley's approximation and MCMC technique) are proposed to estimate the parameters λ, θ and β as well as the reliability r(t) and hazard rate h(t) functions. For estimating these quantities, we assume mainly both SE and LINEX loss functions. However, any other loss function can be easily implemented. From the perspective of frequentist, a natural choice for the prior distributions of λ, θ and β would be to assume that the three quantities are independent gamma distributions G(a1,b1), G(a2,b2) and G(a3,b3) respectively, where ai and bi, i=1,2,3 reflect the knowledge of prior about (λ,θ,β) and they are assumed to be known and nonnegative hyper-parameters. Further, the joint prior distribution can be written as
Subsequently, via Bayes' theorem across combining the likelihood function given (7) with the joint prior given in (20), the joint posterior density function of λ, θ and β can be written as follows
Thus, the conditional posterior densities of λ, θ and β can be computed and written, respectively, as
and
Therefore, for any function, say u(λ,θ,β), the Bayes estimates under SE and LINEX loss functions can be written as
There is a difficulty to compute (25) analytically. Therefore, we propose to approximate it by using both Lindley's approximation and MCMC technique with Metropolis-Hasting algorithm within Gibbs sampler, to obtain the Bayes estimates of λ, θ, β, r(t) and h(t).
3.1. Lindley's approximation
There are various methods to approximate the ratio of integrals of the above form, may be the simplest one is Lindley approximation method, which was introduced by Lindley [19] to approximate the Bayes estimates into a form containing no integrals. This method has been used with various lifetime distributions by several authors for instance, Singh et al. [20], Ying and Gui [21], Ramin and Kohansal [22]. Briefly, it can be described as follows: If the sample size is sufficiently large, according to Lindley [19], any ratio of the integral of the form
where u(λ,θ,β) is the function of λ, θ and β only, ℓ(λ,θ,β) is the log-likelihood function and ρ(λ,θ,β)=logg(λ,θ,β). Hence, I(x) can be estimated as
where ˆλ, ˆθ,and ˆβ are the MLEs of λ, θ and β respectively, and
where ϕ1=λ, ϕ2=θ, ϕ3=β, i,j,l=1,2,3 and σij is (i,j)th elements of the matrix ˆI−1∘(ˆλ,ˆθ,ˆβ) in Subsection (2.1).
From (20), ρ(λ,θ,β) can be written as
Hence, we get
Now, under SE and LINEX loss function, the approximate Bayes estimators λ are given by
and
Similarly, we can obtain the approximate Bayes estimators of θ, β, r(t) and h(t).
3.2. MCMC technique
It is known that there are several procedures of MCMC technique available in which samples are generated from the conditional posterior densities. One of the simplest MCMC procedures is the Gibbs sampling procedure. Another procedure is considered the Metropolis-Hastings (M-H) algorithm and later extended by Hastings [23]. A more general procedure of MCMC procedures which we will use is considered the M-H within Gibbs sampling. Gibbs sampler is required to decompose the joint posterior distribution into full conditional distributions for each parameter and then sample from them. Since, g∗2(θ|λ,β,x_) follows gamma [n+a2,b2+log(1+λx−βL(n))], it is quite simple to generate samples of θ from g∗2(θ|λ,β,x_) by implementing any gamma generating routine. In addition, the conditional posterior densities g∗1(λ|θ,β,x_) and g∗3(β|θ,λ,x_) can't be reduced analytically to well-known distributions. So, according to Tierney [24] M-H algorithm within Gibbs sampling with normal proposal distribution is used to conduct the MCMC methodology. The hybrid M-H algorithm and Gibbs sampler works as follows:
(1) Use the MLEs as the initial value, denoted by ˆλ(0),ˆθ(0)and ˆβ(0).
(2) Set j=1.
(3) Generate θ(j) from Gamma(n+a2,b2+log(1+λx−βL(n))).
(4) Using M-H algorithm, generate λ(j) and β(j) from g∗1(λ(j−1)|θ(j),β(j−1),x_), and g∗3(β(j−1)|θ(j),λ(j−1),x_), respectively, with the normal distributions N(λ(j−1),Λ11) and N(β(j−1),Λ33).
(a) Generate λ∗ from N(λ(j−1),Λ11) and β∗ from N(β(j−1),Λ33).
(b) Evaluate the acceptance probabilities
(c) Generate a ρ1 and ρ2 from a Uniform [0,1] distribution.
(d) If ρ1<Ωλ accept the proposal and set λ∗=λ(j), else set λ(j)=λ(j−1).
(e) If ρ2<Ωβ accept the proposal and set β∗=β(j), else set β(j)=β(j−1).
(5) Compute r(t) and h(t) as
(6) Set j=j+1.
(7) Repeat Steps (3)−(6) M times. The first M0 simulated varieties are ignored to remove the affection of the selection of initial value and to guarantee the convergence. Then the selected sample are λ(j),θ(j),β(j),r(j)(t) and h(j)(t), j=M0+1,...,M, for sufficiently large M, forms an approximate posterior sample which can be used to develop the Bayesian inference.
(8) Under SE and LINEX Loss functions the approximate Bayes estimate of η (where η= λ, θ, β, r(t) and h(t)) can be obtained by
where M0 is the burn-in period and η(j)= λ(j), θ(j), β(j), r(j)(t) and h(j)(t).
(9) To compute the CRIs of η, order {ηM0+1,ηM0+2,…,ηM} as {η[1],η[2],…,η[M−M0]}. Then, the (1−γ)100% CRIs of η can be given by
4.
Simulation study and comparisons
In this section, through a simulation study using R language, we compare the performances of the MLEs and Bayes estimates of the unknown parameters as well as reliability and hazard rate functions of the DD under SE and LINEX loss functions proposed in the previous sections. Here, we consider the simulation results in the case of (λ,θ,β)=(3.2,5.9,1.75). Then, the true values of r(t) and h(t) at time t=1.3 are computed to be 0,963895 and 0,234171, respectively. The performance of estimators is evaluated in terms of mean squared error (MSE) which computed as
where ϖ1=λ, ϖ2=θ, ϖ3=β, ϖ4=r(t) and ϖ5=h(t) for the point estimates (MLE, Lindley, MCMC), also average lengths (ALs) and coverage probability (CPs), which computed as the number of CIs that covered the true values divided by 1000, for interval estimates (ACIs and CRIs). Bayes estimates and the highest posterior density CRIs are computed based on M=12000 MCMC samples and discard the first values M0=2000 as burn-in. In addition, we assume the informative gamma priors for λ, θ and β that is, when the hyper-parameters are ai=1.3 and bi=2.2, i=1,2,3. Moreover, 95% CRIs were computed for each simulated sample. In our study, we consider different sample sizes n=5, 7, 9, 12, 15, 18, 20, 23 and 25. The gamma distribution was chosen as prior distributions of the parameters because it is the most appropriate one that matches the maximum likelihood function. Moreover, they are from the same family. The evidence for this is that two of the full conditional posterior distributions of the parameters resulted in a gamma distribution, which proves the validity of the chose. In addition, choosing another prior distribution or dependent prior will increase the complexity and difficulty of mathematical equations. Gamma distribution is one of the rich distributions, as when changing its parameters (hyper-parameters), we get new data with new information, so it is the focus of attention of most statisticians. Gamma distribution is flexible model with any values of hyper-parameters. A special case: When all hyper-parameters of gamma distribution are zero, we obtain Jaffrey prior. The results of means, MSEs, ALs and CPs of estimates are displayed in Tables 1–5.
From the results in Tables 1–5, we observe the following:
(1) It is clear that from all Tables, as sample size n increases, the MSEs and average interval lengths decrease, also Bayes estimates are performed better than the MLEs of λ, θ, β, r(t) and h(t) in terms of MSEs and average interval lengths, as expected.
(2) Bayes estimates under MCMC technique perform better than Bayes estimates under Lindley approximation in the sense of having smaller MSEs.
(3) Bayes estimates under Lindley approximation perform better than MLEs in the sense of having smaller MSEs.
(4) Bayes estimates under LINEX loss function perform better than Bayes estimates under SE loss function in the sense of having smaller MSEs.
(5) Bayes estimate under LINEX with c=2 provides better estimates for λ, θ, β, r(t) and h(t) because of having the smallest MSEs.
(6) Bayes estimates under LINEX for the choice c=2 are performed better than their estimates for the choice c=−2 in the sense of having smaller MSEs.
(7) Both MLE and Bayesian methods have very close estimates and their ACIs have quite high CPs (around 0.95). Also, the Bayesian CRIs have the highest CPs.
(8) In general, we can conclude that the best estimation method is the Bayes method under MCMC technique with LINEX loss function, especially if the prior information about the problem under study is available.
5.
Applications
To illustrate the inferential procedures developed in the preceding sections. we use the data set from Almongy [25], see Table 6. A complete sample from the data represents a COVID-19 data belonging to the Netherlands of 30 days, which recorded from 31 March to 30 April 2020. This data formed of rough mortality rate. The data are as follows
For the goodness of fit test, we compute the Kolmogorov-Smirnov (K-S) distance. It was found that the K-S is 0.072956 and the associated p-value is 0.9936. Therefore, according to the p-value we can say that the DD fits quite well to the above data. Empirical, Q−Q and P−P plots are shown in Figure 1, which clear that the DD fits the data very well. Table 7 lists the lower record values from the observed data.
Based on these 12 lower record values, the MLEs and ACIs for λ, θ, β, r(10) and h(10) are computed to be as in Tables 6 and 7. Moreover, to compute the Bayesian estimates under Lindley's approximation and MCMC technique, the prior distributions of the parameters λ, θ and β are needed to specify. Since we do not have any prior information available, we assume that the non-informative gamma priors for λ, θ and β that is, when the hyperparameters are ai=0.0001 and bi=0.0001, i=1,2,3. Under MCMC technique, the posterior analysis was done across combining M-H algorithm within Gibbs sampler. To conduct the MCMC algorithm, which was described in Subsection 3.2, the initial values for the parameters λ, θ and β were taken to be their MLEs. In addition, we run the chain for 12000 times and discard the first 2000 values as 'burn-in' to avoid the effect of the initial values (starting point). The Bayesian estimates as well as 95% CRIs for λ, θ, β, r(10) and h(10) are presented in Tables 8 and 9.
6.
Conclusions
In this paper, we have estimated the reliability and hazard rate functions of the Dagum distribution based record statistics. To get the best estimators for this purpose, the maximum likelihood and Bayes estimations methods have been utilized. The MLEs and the asymptotic confidence intervals based on the observed Fisher information matrix have been discussed. For Bayesian estimation approach, we assume the gamma priors for the unknown parameters and provide the Bayes estimators under the assumptions of SE and LINEX loss functions. Since the Bayes estimates cannot be obtained in explicit form, we have applied both Lindley's approximation and MCMC technique with Metropolis–Hasting algorithm within Gibbs sampler to compute the approximate Bayes estimates and constructed the credible intervals. The performance of the proposed techniques has been compared via a Monte Carlo simulation study with different sample sizes n. It was found that if we have no prior information on the unknown parameters, then it is always better to use the MLEs rather than the Bayes estimators, because the Bayes estimators are computationally more expensive. Finally, for illustrative purpose, the details have been explained using a real-life data set of COVID-19 mortality rates.
Conflict of interest
The authors declare no conflict of interest.