Research article

Predictive analysis of doubly Type-Ⅱ censored models

  • Received: 24 July 2024 Revised: 28 September 2024 Accepted: 30 September 2024 Published: 08 October 2024
  • MSC : 62F10, 62N01, 62N02

  • The application of a doubly Type-Ⅱ censoring scheme, where observations are censored at both the left and right ends, is often used in various fields including social science, psychology, and economics. However, the observed sample size under this censoring scheme may not be large enough to apply a likelihood-based approach due to the occurrence of censoring at both ends. To effectively respond to this difficulty, we propose a pivotal-based approach within a doubly Type-Ⅱ censoring framework, focusing on two key aspects: Estimation for parameters of interest and prediction for missing or censored samples. The proposed approach offers two prominent advantages, compared to the likelihood-based approach. First, this approach leads to exact confidence intervals for unknown parameters. Second, it addresses prediction problems in a closed-form manner, ensuring computational efficiency. Moreover, novel algorithms using a pseudorandom sequence, which are introduced to implement the proposed approach, have remarkable scalability. The superiority and applicability of the proposed approach are substantiated in Monte Carlo simulations and real-world case analysis through a comparison with the likelihood-based approach.

    Citation: Young Eun Jeon, Yongku Kim, Jung-In Seo. Predictive analysis of doubly Type-Ⅱ censored models[J]. AIMS Mathematics, 2024, 9(10): 28508-28525. doi: 10.3934/math.20241383

    Related Papers:

    [1] Haiping Ren, Xue Hu . Estimation for inverse Weibull distribution under progressive type-Ⅱ censoring scheme. AIMS Mathematics, 2023, 8(10): 22808-22829. doi: 10.3934/math.20231162
    [2] Yahia Abdel-Aty, Mohamed Kayid, Ghadah Alomani . Generalized Bayesian inference study based on type-Ⅱ censored data from the class of exponential models. AIMS Mathematics, 2024, 9(11): 31868-31881. doi: 10.3934/math.20241531
    [3] Peihua Jiang, Xilong Yang . Reliability inference and remaining useful life prediction for the doubly accelerated degradation model based on Wiener process. AIMS Mathematics, 2023, 8(3): 7560-7583. doi: 10.3934/math.2023379
    [4] Young Eun Jeon, Suk-Bok Kang, Jung-In Seo . Pivotal-based inference for a Pareto distribution under the adaptive progressive Type-II censoring scheme. AIMS Mathematics, 2024, 9(3): 6041-6059. doi: 10.3934/math.2024295
    [5] Mazen Nassar, Refah Alotaibi, Ahmed Elshahhat . Reliability analysis at usual operating settings for Weibull Constant-stress model with improved adaptive Type-Ⅱ progressively censored samples. AIMS Mathematics, 2024, 9(7): 16931-16965. doi: 10.3934/math.2024823
    [6] Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Statistical analysis of stress–strength in a newly inverted Chen model from adaptive progressive type-Ⅱ censoring and modelling on light-emitting diodes and pump motors. AIMS Mathematics, 2024, 9(12): 34311-34355. doi: 10.3934/math.20241635
    [7] Bing Long, Zaifu Jiang . Estimation and prediction for two-parameter Pareto distribution based on progressively double Type-II hybrid censored data. AIMS Mathematics, 2023, 8(7): 15332-15351. doi: 10.3934/math.2023784
    [8] Tahani A. Abushal . Parametric inference of Akash distribution for Type-Ⅱ censoring with analyzing of relief times of patients. AIMS Mathematics, 2021, 6(10): 10789-10801. doi: 10.3934/math.2021627
    [9] Ahmed R. El-Saeed, Ahmed T. Ramadan, Najwan Alsadat, Hanan Alohali, Ahlam H. Tolba . Analysis of progressive Type-Ⅱ censoring schemes for generalized power unit half-logistic geometric distribution. AIMS Mathematics, 2023, 8(12): 30846-30874. doi: 10.3934/math.20231577
    [10] M. Nagy, H. M. Barakat, M. A. Alawady, I. A. Husseiny, A. F. Alrasheedi, T. S. Taher, A. H. Mansi, M. O. Mohamed . Inference and other aspects for $ q- $Weibull distribution via generalized order statistics with applications to medical datasets. AIMS Mathematics, 2024, 9(4): 8311-8338. doi: 10.3934/math.2024404
  • The application of a doubly Type-Ⅱ censoring scheme, where observations are censored at both the left and right ends, is often used in various fields including social science, psychology, and economics. However, the observed sample size under this censoring scheme may not be large enough to apply a likelihood-based approach due to the occurrence of censoring at both ends. To effectively respond to this difficulty, we propose a pivotal-based approach within a doubly Type-Ⅱ censoring framework, focusing on two key aspects: Estimation for parameters of interest and prediction for missing or censored samples. The proposed approach offers two prominent advantages, compared to the likelihood-based approach. First, this approach leads to exact confidence intervals for unknown parameters. Second, it addresses prediction problems in a closed-form manner, ensuring computational efficiency. Moreover, novel algorithms using a pseudorandom sequence, which are introduced to implement the proposed approach, have remarkable scalability. The superiority and applicability of the proposed approach are substantiated in Monte Carlo simulations and real-world case analysis through a comparison with the likelihood-based approach.



    In many practical studies, extreme sample values are often excluded when the observations are not well known or the presence of outliers is suspected. In addition, carelessness or lack of experimental preparation at the beginning of an experiment can lead to the censoring of the smallest few values in the observed sample, and limited time or cost issues can lead to experiment termination without observing the largest values. When both conditions occur simultaneously, to be more accurate, a scenario where the smallest and largest values in an ordered sample are censored at both ends is called a doubly Type-Ⅱ (DT-Ⅱ) censoring scheme. Under this censoring scheme, the observed sample size is not often sufficient to apply valid statistical inference, requiring a very challenging task. This challenge has made the DT-Ⅱ censoring scheme less studied than other censoring schemes, and most of the studies related to it have focused on likelihood-based and Bayesian inference. For example, Fernández [1] introduced a Bayesian inference for a Rayleigh distribution with a scale parameter. Raqab and Madi [2] discussed the prediction problem for the total amount of the remaining testing time from the same distribution and resolved it using Gibbs sampling. Khan et al. [3] provided a prediction for a single future response in a two-parameter Rayleigh distribution, using a Bayesian approach. Kim and Song [4] discussed parameter estimation for a generalized exponential distribution using the likelihood-based and Bayesian methods. Kotb and Raqab [5] derived maximum likelihood estimators (MLEs) as well as Bayes estimators for a modified Weibull distribution. Panahi [6] provided likelihood-based and Bayesian estimation methods for a Burr Type Ⅻ distribution. Sindhu and Hussain [7] dealt with the estimation problem of a power function Ⅱ distribution, using non-informative and informative priors in a Bayesian approach. Besides these methods, some studies have employed a pivotal quantity to obtain a joint confidence region for unknown parameters under the DT-Ⅱ censoring scheme. Wu [8] obtained a joint confidence region in an exponential distribution with two parameters. Wu [9] discussed interval inference for a Pareto distribution with two parameters.

    However, the likelihood-based approach has the drawback of being able to cause substantial bias and invalid inference results for a small sample size. In particular, this approach for interval inference yields approximate, rather than exact, confidence intervals (CIs) for unknown parameters, and these CIs may not meet a nominal level in small sample scenarios. On the other hand, the Bayesian approach has stress related to the elicitation of prior information for each parameter and requires very demanding calculations such as the expected Fisher information for objective Bayesian inference if there is not enough prior information.

    To provide solutions to these challenging problems in a frequentist approach, this study proposes a pivotal-based approach under the DT-Ⅱ censoring scheme, yielding valid statistical inference results even in a situation where the available sample size falls short of being sufficiently large. Notably, the proposed approach leads to not only exact CIs for unknown parameters but also closed-form results in prediction problems for missing or censored samples, unlike the likelihood-based approach. Furthermore, this study introduces novel algorithms using a pseudorandom sequence to implement the proposed approach, which is attractive due to its excellent scalability. The superiority and validity of the proposed approach are demonstrated in the simulation study and real-world case analysis, showing that its performance does not lag behind and rather performs better compared with the most popular likelihood-based approach for small and medium sample sizes.

    The study is outlined as follows: Section 2 briefly describes a distribution family used for inference and the DT-Ⅱ censoring scheme. Section 3 provides an estimation method for unknown parameters and a prediction method for missing or censored samples based on a likelihood function within the DT-Ⅱ censoring framework. Section 4 proposes the parameter estimation and prediction methods using pivotal quantities within the same framework. Section 5 demonstrates the superiority and validity of the proposed method by applying the theoretical results to simulated and real-life datasets. Finally, Section 6 concludes the study with a concise recapitulation of the results.

    Let Θ=(α,β) be an unknown parameter vector. To implement our argument, it is assumed that a continuous random variable X has the following family of distributions with Θ:

    F(x;Θ)=1[G(x;α)]h(β),x>0,α>0,β>0, (2.1)

    where G(x;α) denotes a function of X involving α, and h(β) denotes a function of β. The distribution (2.1) is a cumulative distribution function (CDF) of X, and its members are given in Appendix (Table 5).

    Suppose that X1,,Xn is a random sample with CDF (2.1), and X(1)X(n) is an order statistic from this sample. In experiments concerning life-testing or reliability, negligence or poor preparation at the start of the experiment may result in censoring the first few observations, while the limitation of the experiment duration may result in censoring the last few observations. Specifically, when the r smallest and the s largest observations are simultaneously censored in a situation where n identical components are placed in the test, it is called a DT-Ⅱ censoring scheme. In addition, the remaining X=(X(r+1),,X(ns)) can be regarded as a DT-Ⅱ censored sample.

    This section provides the most widely used likelihood-based approach for comparison with the pivotal-based approach to be proposed in Section 4.

    The likelihood and its logarithm functions for the DT-Ⅱ censored sample X from the probability distribution with CDF (2.1) are formulated as

    L(Θ)(1[G(x(r+1);α)]h(β))r[G(x(ns);α)]sh(β)[h(β)]nsrnsi=r+1[G(x(i);α)]h(β)1[g(x(i);α)], (3.1)

    and

    (Θ)rlog(1[G(x(r+1);α)]h(β))+sh(β)logG(x(ns);α)+(nsr)logh(β)+(h(β)1)nsi=r+1logG(x(i);α)+nsi=r+1log[g(x(i);α)], (3.2)

    respectively, where ()=logL(), and g(x(i);α)=G(x(i);α)/x(i). Then, the MLEs ˆα and ˆβ are obtained by optimizing the log-likelihood function (3.2) for α and β, respectively. In addition, by the asymptotic normality of the MLEs, the approximate 100(1γ)% CIs are formulated as

    ˆα±zγ/2^Var(ˆα),

    and

    ˆβ±zγ/2^Var(ˆβ),

    for 0<γ<1, where zγ/2 is the (γ/2)th upper percentile of a standard normal distribution, and ^Var() is the diagonal element of the asymptotic variance-covariance matrix of the MLEs

    ˆΣ=(^Var(ˆα)^Cov(ˆα,ˆβ)^Cov(ˆα,ˆβ)^Var(ˆβ)),

    which is obtained by inverting the observed Fisher information matrix with the MLEs

    I(ˆα,ˆβ)=(2(Θ)α22(Θ)αβ2(Θ)αβ2(Θ)β2)|(α=ˆα,β=ˆβ).

    In the presence of the DT-Ⅱ censored sample, the predictive likelihood function requires two conditional probability density functions (PDFs) because censoring occurs at both ends. One is the conditional PDF of X(l) given X(r+1)=x(r+1) for l=1,,r, and the other is the conditional PDF of X(l) given X(ns)=x(ns) for l=ns+1,,n. The former is formulated as

    f(x(l)X(r+1)=x(r+1))=r!(l1)!(rl)!h(β)[G(x(l);α)]h(β)1[g(x(l);α)]1[G(x(r+1);α)]h(β)×{[G(x(l);α)]h(β)[G(x(r+1);α)]h(β)1[G(x(r+1);α)]h(β)}rl{1[G(x(l);α)]h(β)1[G(x(r+1);α)]h(β)}l1. (3.3)

    The latter is formulated as

    f(x(l)X(ns)=x(ns))=s!(ln+s1)!(nl)!h(β)[G(x(l);α)]h(β)1[g(x(l);α)][G(x(ns);α)]h(β)×{[G(x(ns);α)]h(β)[G(x(l);α)]h(β)[G(x(ns);α)]h(β)}ln+s1[G(x(l);α)G(x(ns);α)](nl)h(β). (3.4)

    Then, under the DT-Ⅱ censoring scheme, prediction is made from the following predictive likelihood function:

    L(x(l),Θ){[h(β)]nsr+1[G(x(l);α)]h(β)1[g(x(l);α)][G(x(ns);α)]sh(β)nsi=r+1[G(x(i);α)]h(β)1[g(x(i);α)]×{1[G(x(l);α)]h(β)}l1{[G(x(l);α)]h(β)[G(x(r+1);α)]h(β)}rl,l=1,,r,[h(β)]nsr+1[G(x(l);α)](nl+1)h(β)1[g(x(l);α)]nsi=r+1[G(x(i);α)]h(β)1[g(x(i);α)]×{1[G(x(r+1);α)]h(β)}r{[G(x(ns);α)]h(β)[G(x(l);α)]h(β)}ln+s1,l=ns+1,,n, (3.5)

    which is derived by multiplying the likelihood function (3.1) with the conditional PDFs (3.3) and (3.4). The point prediction for X(l), denoted by ˆX(l), and the predictive MLEs for α and β are obtained by optimizing the natural logarithm of the predictive likelihood function (3.5) for X(l), α, and β, respectively. Here, our focus is on the prediction of X(l), so only the results related to it are reported in Section 5.

    The goal of this section is to propose an estimation method for α and β and a prediction method for missing or censored samples, using pivotal quantities within the DT-Ⅱ censoring framework. The foremost advantage of the proposed approach is its provision of exact CIs for α and β, unlike the likelihood-based approach in Section 3. Moreover, the proposed approach excels in resolving prediction problems for missing or censored samples by yielding closed-form results.

    Before introducing the essential pivotal quantities used for inference in the subsequent subsection, we define the following notations for some distributions to simplify their expressions.

    χ2v : Chi-squared distribution with v degrees of freedom.

    Be(a,b) : Beta distribution with parameters (a,b).

    Let

    Y(i)=log[1F(x(i);Θ)]=h(β)logG(x(i);α),i=r+1,,ns.

    Since a DT-Ⅱ censored sample Y(r+1)Y(ns) has a standard exponential distribution with a mean E(Y(i))=ij=1(nj+1)1, it induces the normalized spacings

    Si=(ni+1)(Y(i)Y(i1))=(ni+1)h(β)log(G(x(i1);α)G(x(i);α)),i=r+2,,ns,

    which are standard exponential random variables, all independent and identically distributed (iid). Then, the following pivotal quantity can be led by the independence of the spacings:

    Tj(Θ)=2r+1+ji=r+2Si=2(nr)Y(r+1)+2(nrj)Y(r+j+1)+2r+ji=r+1Y(i)=2h(β)M1,j(x;α),j=1,,nsr1,

    which are independent random variables from χ22j, where

    M1,j(x;α)=(nr)logG(x(r+1);α)(nrj)logG(x(r+j+1);α)r+ji=r+1logG(x(i);α).

    The subsequent lemma provides some pivotal quantities derived from the pivotal quantity Tj(Θ), which are instrumental in inference for α and β.

    Lemma 4.1. Let X(r+1)X(ns) be a DT-Ⅱ censored sample from the probability distribution with CDF (2.1). Then,

    (i)W1(Θ)=2h(β)M2(x;α),

    and

    (ii)W2(α)=2nsr2j=1log(M1,j(x;α)M2(x;α)),

    which have χ22(nsr1) and χ22(nsr2), respectively, where

    M2(x;α)=(nr)logG(x(r+1);α)(s+1)logG(x(ns);α)ns1i=r+1logG(x(i);α).

    Proof. W1(Θ) is clear from W1(Θ)=Tnsr1(Θ). In addition, since (Tj(Θ)/Tj+1(Θ))j (j=1,,nsr2) are standard uniform random variables, (ii) is derived as

    W2(α)=2nsr2j=1jlog(Tj(Θ)Tj+1(Θ))=2nsr2j=1log(Tj(Θ)Tnsr1(Θ)).

    This completes the proof.

    It is worth noting that the pivotal quantity W2(α) is a function of α only, not dependent on β. This quantity plays a very important role in inference not only for α itself but also for the parameters of interest or functions of interest, even in a situation where α is a nuisance parameter.

    The process of deriving the essential pivotal quantities used for inference under the DT-Ⅱ censoring scheme is summarized in Figure 1, where Exp(1) denotes a standard exponential distribution.

    Figure 1.  Flowchart of the essential pivotal quantities used for inference under the DT-Ⅱ censoring scheme.

    Using the pivotal quantities in Lemma 4.1, exact CIs for α and β are provided now. For α, since the pivotal quantity W2(α) in Lemma 4.1 has χ22(nsr2), it leads to

    1γ=P[χ21γ/2,2(nsr2)<W2(α)<χ2γ/2,2(nsr2)],

    for 0<γ<1, where χ2γ/2,v is the (γ/2)th upper percentile of χ2v. Then, an exact 100(1γ)% CI of α is formulated as

    [W12(χ21γ/2,2(nsr2)),W12(χ2γ/2,2(nsr2))],

    or

    [W12(χ2γ/2,2(nsr2)),W12(χ21γ/2,2(nsr2))],

    when the pivotal quantity W2(α) in Lemma 4.1 is the increasing or decreasing functions of α, respectively, where W12(k) is a solution of α for the equation W2(α)=k. In a similar manner, using the pivotal quantity W1(Θ) in Lemma 4.1, an exact 100(1γ)% CI of β for known α can be formulated as

    [h1(χ21γ/2,2(nsr1)2M2(x;α)),h1(χ2γ/2,2(nsr1)2M2(x;α))],

    or

    [h1(χ2γ/2,2(nsr1)2M2(x;α)),h1(χ21γ/2,2(nsr1)2M2(x;α))],

    for 0<γ<1 when h(β) is the increasing or decreasing functions of β, respectively, where h1(t) is the solution of β for the equation h(β)=t. However, if the nuisance parameter α is unknown, this CI cannot be used. To resolve this issue, we propose a method of constructing a generalized CI for β by defining a generalized pivotal quantity (GPQ) and introducing a new algorithm based on it.

    Let α be a unique solution of α for the equation W2(α)=ω, where ω is a realization from χ22(nsr2). Then, from the pivotal quantity W1(Θ) in Lemma 4.1, we define a GPQ for β as

    ψ1(βα)=h1(W1(α,β)2M2(x;α)),

    with the observed values x, according to the argument of Weerahandi [10,11]. To obtain its percentile required to construct a generalized CI for β, we propose utilizing a pseudorandom sequence from χ22 based on the principles used to derive the pivotal quantities W1(Θ) and W2(α) in Lemma 4.1. This approach is detailed in Algorithm 1.

    Algorithm 1.

    Step 1. Generate ζ1,ζ2,,ζn from χ22.

    Step 2. Compute υj=r+1+ji=r+2ζi for j=1,,nsr1 and ω=2nsr2j=1log(υjυnsr1).

    Step 3. Obtain α by solving the equation W2(α)=ω for α.

    Step 4. Compute ψ1(βα)=h1(υnsr12M2(x;α)).

    Step 5. Repeat N(10,000) times.

    Algorithm 1 generates a sequence ψ1={ψ1,1(βα),,ψ1,N(βα)}. Then, a generalized 100(1γ)% CI of β is formulated as

    (ψ1,([(γ/2)N])(βα),ψ1,([(1γ/2)N])(βα)),

    for 0<γ<1, where ψ1,([(γ/2)N])(βα) is the [(γ/2)N]th smallest of the sequence ψ1, and [] denotes the greatest integer function.

    Furthermore, the pivotal quantities W1(Θ) and W2(α) in Lemma 4.1 lead to a valid and statistically good estimation equation for α and β. When the parameter of interest is α, the pivotal quantity W2(α) plays a very important role in estimating α since it is independent of β, unlike the pivotal quantity W1(Θ). To derive an estimator of α, the following lemma is introduced.

    Lemma 4.2. Let V(α)=W2(α)/[2(nsr3)]. Then, V(α) converges to one in probability as nsr.

    Proof. Since V(α) has a gamma distribution with parameters (nsr2,nsr3), it follows that

    P(|V(α)1|>ϵ)4Var(V(α))ϵ20,

    for any ϵ>0 as nsr. This completes the proof.

    From Lemma 4.2, the estimation equation for α is derived as W2(α)=2(nsr3). In addition, the subsequent theorem establishes the uniqueness of the solution to the equation.

    Theorem 4.3. The equation W2(α)=2(nsr3) has a unique solution ˜α for α.

    This requires different proofs depending on the form of the probability distribution with CDF (2.1), and the proofs are given in Appendix.

    If the parameter of interest is β for known α, then the following estimator of h(β) can be used:

    ˜hβ(α)=nsr2M2(x;α), (4.1)

    which has an inverse gamma distribution with parameters (nsr1,(nsr2)h(β)) from the pivotal quantity W1(Θ) in Lemma 4.1. It is noted that the estimator (4.1) has unbiasedness. From it, an estimation equation for β is derived and the consequent estimator of β is formulated as

    ˜β(α)=h1(nsr2M2(x;α)). (4.2)

    For unknown α in the estimator (4.2), it can be substituted with the estimator ˜α. The validity of the estimators ˜α and ˜β(˜α) is examined by Monte Carlo simulations, and the excellence is demonstrated through a comparison with the MLEs ˆα and ˆβ in Section 5.

    To predict missing or censored samples using the pivotal quantity, the following Lemmas are first introduced.

    Lemma 4.4. Let UL=1[G(x(l);α)]h(β)1[G(x(r+1);α)]h(β) for l=1,,r in the conditional PDF (3.3). Then, UL has Be(l,rl+1).

    Lemma 4.5 Let UR=[G(x(l);α)G(x(ns);α)]h(β) for l=ns+1,,n in the conditional PDF (3.4). Then, UR has Be(nl+1,ln+s).

    Lemmas 4.4 and 4.5 are easily proved by the variable transformation in the conditional PDFs (3.3) and (3.4), respectively. From these Lemmas, the prediction for X(l) is formulated as

    X(l)={G1((1UL{1[G(x(r+1);α)]h(β)})1/h(β)),l=1,,r,G1(U1/h(β)RG(x(ns);α)),l=ns+1,,n, (4.3)

    where G1() is an inverse function of G(x;α). Note that the proposed approach induces a closed-form prediction result for X(l), unlike the likelihood-based approach in Section 3.

    To obtain the predicted value for X(l) in (4.3), the realizations of UL and UR are required. These can be obtained from the fact that if X1 and X2 are independent random variables from χ22a and χ22b, respectively, then X1/(X1+X2) has Be(a,b). In this study, as in the case of Algorithm 1, a pseudorandom sequence is used to generate these realizations, and the detailed procedure is provided in Algorithm 2. Moreover, α and β in (4.3) can be substituted with ˜α and ˜β(˜α), respectively, then the consequent point prediction is denoted by ˜X(l).

    Additionally, a predictive interval (PI) for X(l) can be constructed based on its GPQ ψ2(X(l)) which is defined by substituting α and β with α and ψ1(βα) in (4.3), respectively, unlike the likelihood-based approach in Section 3. The percentiles of ψ2(X(l)), required to construct the PI for X(l), are obtained along the lines of Algorithm 1, with the realizations of UL and UR generated from two independent χ2 random variables as in the case of ˜X(l). The implementation process is detailed step-by-step within Algorithm 2.

    Algorithm 2.

    Step 1. Generate ζ1,ζ2,,ζn from χ22.

    Step 2. Compute υj=r+1+ji=r+2ζi for j=1,,nsr1 and ω=2nsr2j=1log(υjυnsr1).

    Step 3. Obtain α by solving the equation W2(α)=ω for α.

    Step 4. Compute ψ1(βα)=h1(υnsr12M2(x;α)).

    Step 5. Substitute α and β with α and ψ1(βα) in (4.3), respectively.

    Step 6. Compute uL=li=1ζili=1ζi+r+1i=l+1ζi or uR=ni=lζini=lζi+l1i=nsζi.

    Step 7. Substitute UL and UR with uL and uR in (4.3), respectively, to compute ψ2(X(l)).

    Step 8. Repeat N(10,000) times.

    Algorithm 2 generates a sequence ψ2={ψ2,1(X(l)),,ψ2,N(X(l))}. Then, a generalized 100(1γ)% PI of X(l) is formulated as

    (ψ2,([(γ/2)N])(X(l)),ψ2,([(1γ/2)N])(X(l))),

    for 0<γ<1, where ψ2,([(γ/2)N])(X(l)) is the [(γ/2)N]th smallest of ψ2. It is worth noting that Algorithm 2 includes some steps of Algorithm 1. So, it can be viewed as an extended version of Algorithm 1. This means that if the function of interest contains α and β, then its percentile and the consequent generalized PI can be obtained without difficulty, as in the case of prediction using Algorithm 2.

    This section substantiates the superiority and validity of the proposed method by making a comparison with the likelihood-based method in Monte Carlo simulations and real-world case analysis.

    In the probability distribution with CDF (2.1), Weibull and bathtub-shaped distributions are prominent statistical probability distributions employed in life data analysis. In addition, most distributions in this family have an unknown parameter α as a shape parameter which describes the shape of the distributions. This parameter may be an important parameter of interest because it allows us to model various characteristics of lifetime distributions. For example, the hazard function of a Weibull distribution increases over time for α>1, and vice versa for α<1. A bathtub-shaped distribution has an increasing hazard function for α1, while it has a bathtub shape for α<1. Based on those characteristics, results for the Weibull and bathtub-shaped distributions with α=(0.5,1.5) and β=0.5 are only reported here. Additionally, to highlight the superiority of the proposed method for small and medium sample sizes, the following censoring schemes are used:

    Scheme I : n=20,r=2,s=2,II : n=20,r=3,s=3,III : n=30,r=3,s=3,IV : n=30,r=4,s=4.

    Monte Carlo simulations are conducted by generating 1,000 DT-Ⅱ censored samples for each censoring scheme, and the results are presented through various statistical measures. First, the coverage probabilities (CPs) of 95% CIs provided in Sections 3 and 4 are reported in Figure 2. Here, the ACI and PCI denote the approximate and proposed CIs, respectively. In the case of the PCI for β, it is computed based on N=10,000 in Algorithm 1. In addition, the mean squared errors (MSEs) and biases of the estimators are reported in Figures 3 and 4.

    Figure 2.  CPs of 95% CIs for α and β in (a) Weibull and (b) bathtub-shaped distributions.
    Figure 3.  (a) MSEs and (b) biases of estimators for α and β in a Weibull distribution.
    Figure 4.  (a) MSEs and (b) biases of estimators for α and β in a bathtub-shaped distribution.

    Through Figures 24, the following results are led: The CPs of the PCI are generally found to be more aligned with the desired nominal level of 0.95 than those of the ACI. Notably, for β, the PCI has a CP much closer to 0.95 than the ACI, indicating that the generalized CI obtained from the proposed Algorithm 1 yields superior results for small and medium sample sizes. In terms of the MSE and bias, ˜α shows superior MSE and bias results to the MLE counterpart when α=1.5. For a Weibull distribution, the biases of ˜β(˜α) are generally more efficient than those of the MLE counterpart when α=0.5, despite performing poorly in terms of MSE. On the other hand, for a bathtub-shaped distribution, ˜β(˜α) exhibits generally superior performance in terms of bias when α=0.5 and 1.5, compared to the MLE counterpart.

    Furthermore, to assess the validity of the pivotal-based prediction approach proposed, the prediction errors for censored samples are computed. Specifically, the prediction error is computed as X(l)ˆX(l) and X(l)˜X(l) for likelihood-based and pivotal-based predictions, respectively. The resulting prediction errors for the simulated 1,000 DT-Ⅱ censored samples are visualized in Figures 58 as box plots, which indicates that the medians of prediction errors obtained from the pivotal-based method are consistently closer to zero than those of prediction errors obtained from the likelihood-based method in all settings. This finding substantiates the superior predictive performance of the proposed method for small and medium sample sizes, revealing its competitiveness compared to the likelihood-based method.

    Figure 5.  Box plots of prediction errors for a Weibull distribution under (a) Scheme Ⅰ and (b) Scheme Ⅱ.
    Figure 6.  Box plots of prediction errors for a Weibull distribution under (a) Scheme Ⅲ and (b) Scheme Ⅳ.
    Figure 7.  Box plots of prediction errors for a bathtub-shaped distribution under (a) Scheme Ⅰ and (b) Scheme Ⅱ.
    Figure 8.  Box plots of prediction errors for a bathtub-shaped distribution under (a) Scheme Ⅲ and (b) Scheme Ⅳ.

    To illustrate the practical application of the proposed pivotal-based approach, an analysis is conducted on a fraction of COVID-19 mortality rates provided by Alsuhabi et al. [12]. The COVID-19 mortality rates are obtained by calculating the ratio of the daily number of deaths to daily new cases in the United States of America (USA) from June 18 to July 7, 2020, and reported in Table 1. For analytical purposes, the data are multiplied by 100, from which a DT-Ⅱ censored sample (Table 2) is generated through the setting of r=2 and s=2. Tables 3 and 4 provide the analysis results. Here, the PCIs for β are computed based on N=20,000 in Algorithm 1. From Tables 3 and 4, it can be seen that the length of the PCI for α and β is longer compared with the ACI counterpart. However, it may be more reasonable than the ACI because the proposed method yields satisfactory performance in terms of CP according to the simulation results in Section 5.1.

    Table 1.  COVID-19 mortality rates in the USA from June 18 to July 7, 2020.
    0.0259 0.0333 0.0318 0.0188 0.0172 0.0112 0.0155 0.0229 0.0184 0.0621
    0.0146 0.0114 0.0216 0.0103 0.0129 0.0134 0.0117 0.0143 0.0032 0.0054

     | Show Table
    DownLoad: CSV
    Table 2.  DT-Ⅱ censored sample generated from COVID-19 mortality rates.
    1.03 1.12 1.14 1.17 1.29 1.34 1.43 1.46
    1.55 1.72 1.84 1.88 2.16 2.29 2.59 3.18

     | Show Table
    DownLoad: CSV
    Table 3.  95% CIs and estimates of α and β for a Weibull distribution.
    95% CI Estimates
    ACI for α PCI for α ACI for β PCI for β ˆα ˜α ˆβ ˜β(˜α)
    (1.344,2.997) (0.000,2.773) (0.036,0.393) (0.103,5.543) 2.171 2.004 0.215 0.264

     | Show Table
    DownLoad: CSV
    Table 4.  95% CIs and estimates of α and β for a bathtub-shaped distribution.
    95% CI Estimates
    ACI for α PCI for α ACI for β PCI for β ˆα ˜α ˆβ ˜β(˜α)
    (0.663,1.135) (0.000,0.994) (0.049,0.281) (0.120,4.255) 0.899 0.787 0.165 0.266

     | Show Table
    DownLoad: CSV

    Additionally, we evaluate the effectiveness of the proposed prediction approach using the COVID-19 mortality rates. The generalized 95% PIs for X(l) (l=1,2,19,20) are computed based on N=20,000 in Algorithm 2, and the results are plotted in Figure 9. It shows that the observations for the censored samples and the predicted values obtained from the proposed method lie within a 95% interval well. In addition, the proposed method has superior predictive performance compared with the likelihood-based method because its prediction results are closer to the observations for the censored samples.

    Figure 9.  Generalized 95% PIs and predicted values for (a) left and (b) right censored samples.

    The observed sample size under the DT-Ⅱ censoring scheme is likely to be very small, which can yield highly biased results when using the likelihood-based method. Our proposed pivotal-based estimation and prediction methods overcome these obstacles by providing exact interval estimation and prediction results that are valid in small sample scenarios. The proposed prediction method is particularly valuable in that it yields closed-form results, unlike the likelihood-based method. In addition, novel algorithms utilizing a pseudorandom sequence, introduced to implement the proposed approach, offer outstanding scalability.

    The superiority and effectiveness of the proposed method were demonstrated through a comparison with the likelihood-based method in Monte Carlo simulations and real-world case analysis. According to simulation results, the proposed intervals achieve closer CP to the considered nominal level than the ACIs for small and medium sample sizes. In addition, the proposed prediction method outperforms the likelihood-based method, showing that the medians of prediction errors for the proposed method are close to zero. These results reveal that the proposed method has superior competitiveness compared to the likelihood-based method when focusing on interval estimation and prediction. Based on these findings, we strongly recommend the use of the proposed method to estimate intervals for unknown parameters and predict censored samples, particularly in situations where the available sample size is not sufficiently large. Furthermore, the applicability of the proposed method was confirmed by analyzing COVID-19 mortality rates in the USA.

    The proposed method is applicable to inference not only for unknown parameters but also for functions of interest involving these parameters, as demonstrated through Algorithm 2 in this study. For example, the proposed method can be extended to inference for more complex functions involving unknown parameters, such as the reliability function, entropy, and hazard rate under various censoring schemes. This scalability is highly valuable and presents promising opportunities for future research.

    Proof of Theorem 4.3.

    The term M2(x;α)/M1,j(x;α) in W2(α) can be written as

    M2(x;α)M1,j(x;α)=1+nsi=r+2+jQ(i),(r+j+1)+sQ(ns),(r+j+1)(nrj1)r+1+ji=r+1Q(i),(r+j+1)(nr)Q(r+1),(r+j+1)+(nrj1),

    where

    Q(i),(j)=logG(x(i);α)logG(x(j);α).

    This proof can then be completed by simply showing that Q(i),(j) is a strictly increasing or decreasing function of α for any α>0. The quantity Q(i),(j) for each distribution is given in Table 5. For the Weibull distribution, it is clear that Q(i),(j) is a strictly increasing function of α. For the bathtub-shaped and Gompertz distributions, see Lemma 2 in Chen [13] and Example 3.1 in Wang et al. [14], respectively.

    Table 5.  Members of distribution family (2.1).
    Distribution F(x;Θ) G(x;α) h(β) Parameter Q(i),(j)
    α β
    Weibull 1eβxα,x>0 exα β Shape Scale (xixj)α
    Bathtub-shaped 1eβ(1exα),x>0 e1exα β Shape 1exαi1exαj
    Lomax 1(1+αx)β,x>0 (1+αx)1 β Scale Shape log(1+αxi)log(1+αxj)
    Gompertz 1e(β/α)(eαx1),x>0 e(eαx1)/α β Shape Scale eαxi1eαxj1
    Burr-Ⅻ 1(1+xα)β,x>0 (1+xα)1 β Shape Shape log(1+xαi)log(1+xαj)

     | Show Table
    DownLoad: CSV

    Young Eun Jeon: Formal analysis, methodology, software, writing-original draft, writing-review and editing; Yongku Kim: Conceptualization, methodology, supervision, writing-review and editing; Jung-In Seo: Conceptualization, formal analysis, methodology, software, supervision, writing-original draft, writing-review and editing. All authors have read and approved the final version of the manuscript for publication.

    This research was supported by Global-Learning & Academic research institution for Master'sPhD students, and Postdocs (G-LAMP) Program of the National Research Foundation of Korea (NRF) grant funded by the Ministry of Education (No. RS-2023-00301914).

    All authors declare no conflict of interest in this paper.



    [1] A. J. Fernández, Bayesian inference from type Ⅱ doubly censored Rayleigh data, Stat. Probabil. Lett., 48 (2000), 393–399. https://doi.org/10.1016/S0167-7152(00)00021-3 doi: 10.1016/S0167-7152(00)00021-3
    [2] M. Z. Raqab, M. T. Madi, Bayesian prediction of the total time on test using doubly censored Rayleigh data, J. Stat. Comput. Sim., 72 (2002), 781–789. https://doi.org/10.1080/00949650214670 doi: 10.1080/00949650214670
    [3] H. M. R. Khan, S. B. Provost, A. Singh, Predictive inference from a two-parameter Rayleigh life model given a doubly censored sample, Commun. Stat.-Theor. M., 39 (2010), 1237–1246. https://doi.org/10.1080/03610920902871453 doi: 10.1080/03610920902871453
    [4] C. Kim, S. Song, Bayesian estimation of the parameters of the generalized exponential distribution from doubly censored samples, Stat. Pap., 51 (2010), 583–597. https://doi.org/10.1007/s00362-008-0142-3 doi: 10.1007/s00362-008-0142-3
    [5] M. S. Kotb, M. Z. Raqab, Inference and prediction for modified Weibull distribution based on doubly censored samples, Math. Comput. Simulat., 132 (2017), 195–207. https://doi.org/10.1016/j.matcom.2016.07.014 doi: 10.1016/j.matcom.2016.07.014
    [6] H. Panahi, Estimation for the parameters of the Burr Type Ⅻ distribution under doubly censored sample with application to microfluidics data, Int. J. Syst. Assur. Eng., 10 (2019), 510–518. https://doi.org/10.1007/s13198-018-0735-8 doi: 10.1007/s13198-018-0735-8
    [7] T. N. Sindhu, Z. Hussain, Objective Bayesian analysis for the power function Ⅱ distribution under doubly type Ⅱ censored data, Thail. Statist., 20 (2022), 710–731.
    [8] S. F. Wu, Interval estimation for the two-parameter exponential distribution based on the doubly type Ⅱ censored sample, Qual. Quant., 41 (2007), 489–496. https://doi.org/10.1007/s11135-006-9008-8 doi: 10.1007/s11135-006-9008-8
    [9] S. F. Wu, Interval estimation for a Pareto distribution based on a doubly type Ⅱ censored sample, Comput. Stat. Data An., 52 (2008), 3779–3788. https://doi.org/10.1016/j.csda.2007.12.015 doi: 10.1016/j.csda.2007.12.015
    [10] S. Weerahandi, Generalized confidence intervals, J. Am. Stat. Assoc., 88 (1993), 899–905. https://doi.org/10.2307/2290779 doi: 10.2307/2290779
    [11] S. Weerahandi, Generalized inference in repeated measures: Exact methods in MANOVA and mixed models, New York: John Wiley & Sons, 2004.
    [12] H. Alsuhabi, I. Alkhairy, E. M. Almetwally, H. M. Almongy, A. M. Gemeay, E. H. Hafez, et al., A superior extension for the Lomax distribution with application to Covid-19 infections real data, Alex. Eng. J., 61 (2022), 11077–11090. https://doi.org/10.1016/j.aej.2022.03.067 doi: 10.1016/j.aej.2022.03.067
    [13] Z. Chen, A new two-parameter lifetime distribution with bathtub shape or increasing failure rate function, Stat. Probabil. Lett., 49 (2000), 155–161. https://doi.org/10.1016/S0167-7152(00)00044-4 doi: 10.1016/S0167-7152(00)00044-4
    [14] B. X. Wang, K. Yu, M. C. Jones, Inference under progressively type Ⅱ right-censored sampling for certain lifetime distributions, Technometrics, 52 (2010), 453–460. https://doi.org/10.1198/TECH.2010.08210 doi: 10.1198/TECH.2010.08210
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(885) PDF downloads(45) Cited by(0)

Figures and Tables

Figures(9)  /  Tables(5)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog