Research article

Investigation the generalized extreme value under liner distribution parameters for progressive type-Ⅱ censoring by using optimization algorithms

  • Received: 30 December 2023 Revised: 27 March 2024 Accepted: 07 April 2024 Published: 28 April 2024
  • MSC : 60G70, 62G32

  • Several random phenomena have been modeled by using extreme value distributions. Based on progressive type-Ⅱ censored data with three different distributions (i.e., fixed, discrete uniform, and binomial random removal), the statistical inference of the generalized extreme value distribution under liner normalization (GEVL distribution) parameters is investigated in this study. Since there is no analytical solution, determining the maximum likelihood parameters for the GEVL distribution is considered to be a problem. Standard numerical methods are frequently insufficient for this dilemma, requiring the use of artificial intelligence algorithms to address this difficulty. Here, nonlinear minimization and a genetic algorithm have been used to tackle that problem. In addition, Lindley approximation and Monte Carlo estimation were implemented via Metropolis-Hastings algorithms to carry out the Bayesian point estimation based on both the squared error loss function and LINEX loss functions. Moreover, the highest posterior density intervals were applied. The proposed theoretical inference techniques have been applied in a numerical simulation and a real-life example.

    Citation: Rasha Abd El-Wahab Attwa, Shimaa Wasfy Sadk, Hassan M. Aljohani. Investigation the generalized extreme value under liner distribution parameters for progressive type-Ⅱ censoring by using optimization algorithms[J]. AIMS Mathematics, 2024, 9(6): 15276-15302. doi: 10.3934/math.2024742

    Related Papers:

    [1] Ahmed Elshahhat, Vikas Kumar Sharma, Heba S. Mohammed . Statistical analysis of progressively first-failure-censored data via beta-binomial removals. AIMS Mathematics, 2023, 8(9): 22419-22446. doi: 10.3934/math.20231144
    [2] Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Analysis of Weibull progressively first-failure censored data with beta-binomial removals. AIMS Mathematics, 2024, 9(9): 24109-24142. doi: 10.3934/math.20241172
    [3] Hassan Okasha, Mazen Nassar, Saeed A. Dobbah . E-Bayesian estimation of Burr Type XII model based on adaptive Type-Ⅱ progressive hybrid censored data. AIMS Mathematics, 2021, 6(4): 4173-4196. doi: 10.3934/math.2021247
    [4] Ahmed R. El-Saeed, Ahmed T. Ramadan, Najwan Alsadat, Hanan Alohali, Ahlam H. Tolba . Analysis of progressive Type-Ⅱ censoring schemes for generalized power unit half-logistic geometric distribution. AIMS Mathematics, 2023, 8(12): 30846-30874. doi: 10.3934/math.20231577
    [5] Haiping Ren, Jiajie Shi . A generalized Lindley distribution:Properties, estimation and applications under progressively type-Ⅱ censored samples. AIMS Mathematics, 2025, 10(5): 10554-10590. doi: 10.3934/math.2025480
    [6] A. M. Abd El-Raheem, Ehab M. Almetwally, M. S. Mohamed, E. H. Hafez . Accelerated life tests for modified Kies exponential lifetime distribution: binomial removal, transformers turn insulation application and numerical results. AIMS Mathematics, 2021, 6(5): 5222-5255. doi: 10.3934/math.2021310
    [7] Xue Hu, Haiping Ren . Statistical inference of the stress-strength reliability for inverse Weibull distribution under an adaptive progressive type-Ⅱ censored sample. AIMS Mathematics, 2023, 8(12): 28465-28487. doi: 10.3934/math.20231457
    [8] Haiping Ren, Ziwen Zhang, Qin Gong . Estimation of Shannon entropy of the inverse exponential Rayleigh model under progressively Type-Ⅱ censored test. AIMS Mathematics, 2025, 10(4): 9378-9414. doi: 10.3934/math.2025434
    [9] Bing Long, Zaifu Jiang . Estimation and prediction for two-parameter Pareto distribution based on progressively double Type-II hybrid censored data. AIMS Mathematics, 2023, 8(7): 15332-15351. doi: 10.3934/math.2023784
    [10] Haiping Ren, Xue Hu . Estimation for inverse Weibull distribution under progressive type-Ⅱ censoring scheme. AIMS Mathematics, 2023, 8(10): 22808-22829. doi: 10.3934/math.20231162
  • Several random phenomena have been modeled by using extreme value distributions. Based on progressive type-Ⅱ censored data with three different distributions (i.e., fixed, discrete uniform, and binomial random removal), the statistical inference of the generalized extreme value distribution under liner normalization (GEVL distribution) parameters is investigated in this study. Since there is no analytical solution, determining the maximum likelihood parameters for the GEVL distribution is considered to be a problem. Standard numerical methods are frequently insufficient for this dilemma, requiring the use of artificial intelligence algorithms to address this difficulty. Here, nonlinear minimization and a genetic algorithm have been used to tackle that problem. In addition, Lindley approximation and Monte Carlo estimation were implemented via Metropolis-Hastings algorithms to carry out the Bayesian point estimation based on both the squared error loss function and LINEX loss functions. Moreover, the highest posterior density intervals were applied. The proposed theoretical inference techniques have been applied in a numerical simulation and a real-life example.



    Extreme value theory (EVT) models hold significant relevance in a wide array of disciplines, including environmental sciences, engineering, finance, insurance, and various other fields. In finance, particularly, extreme price fluctuations of financial assets or market indices are defined as the highest and lowest prices within a specific time frame. For a more comprehensive exploration of this concept, refer to the work in [1].

    EVT studies the asymptotic behavior of extreme values that follow a well-defined pattern, regardless of the underlying return mechanism. It is primarily concerned with providing a probabilistic description of extreme occurrences within a sequence of random events. The foundational principles of EVT are elucidated in [2]. According to the fundamental theorem of EVT, the maxima of independent and identically distributed random variables can be categorized into one of three extreme value distributions. These include the Frechet distribution, characterized by an unbounded upper heavy tail, the Gumbel distribution, distinguished by an infinite upper tail that is lighter than that of the Frechet distribution, and the inverse Weibull distribution, which has a finite upper tail. These three distribution forms can be interconnected within the cumulative distribution function (CDF) family models known as generalized extreme value distributions under linear normalization (GEVL distributions). Numerous authors have introduced studies on GEVL distribution, detailed [3,4,5,6,7,8], and several others. A random variable denoted as X is said to follow a GEVL distribution if its probability density function (PDF) and CDF assume the following forms:

    f(x;μ,σ,ξ)={σ1[1+ξ(xμ)σ]1ξ1exp[[1+ξ(xμ)σ]1ξ] if ξ0,σ1exp((xμ)σ)exp[exp((xμ)σ)] if ξ0, (1.1)
    F(x;μ,σ,ξ)={exp[[1+ξ(xμ)σ]1ξ] if ξ0,exp[exp((xμ)σ)] if ξ0. (1.2)

    where σ>0 and <μ< are the scale and location parameters respectively, and ξ is the shape parameter that represents the behavior of the tail. Sub-models can be defined by ξ0,ξ>0, and ξ<0, corresponding respectively to the Gumbel, Frechét, and Weibull distributions, which are mentioned above.

    The hazard rate (HR) and reversed hazard rate (RHR) for GEVL distribution respectively obtained as follows:

    h(x;μ,σ,ξ)=f(x;μ,σ,ξ)1F(x;μ,σ,ξ)={[1+ξ(xμ)σ]1ξ1σ[exp[1+ξ(xμ)σ]1ξ1] if ξ0,exp((xμ)σ)σ[exp(exp((xμ)σ))1] if ξ0, (1.3)
    rh(x;μ,σ,ξ)=f(x;μ,σ,ξ)F(x;μ,σ,ξ)={σ1[1+ξ(xμ)σ]1ξ1 if ξ0,σ1exp((xμ)σ) if ξ0. (1.4)

    In this study, we only considered the distribution form at ξ0. Notice that, the GEVL distribution can be readily adapted to various models, as shown in plots of the PDF, CDF, HR, and RHR for some parameter values in Figure 1. These plots illustrate how the PDFs and CDFs respond to alterations in parameters, revealing their sensitivity and the resultant changes in the behavior of the parameter ξ configuration.

    Figure 1.  Plots of the PDF, CDF, HR, and RHR of the GEVL distribution for several values of (μ,σ,ξ).

    Over the years, both maximum likelihood estimation (MLE) and Bayesian methods have become widely recognized as crucial statistical techniques for parameter estimation. Moreover, the asymptotic properties associated with these statistical approaches have significantly contributed to their popularity.

    In MLE, the parameters are treated as fixed but unknown values, often representing a measure such as the mean. On the other hand, Bayesian methods adopt a different perspective, treating the parameters as random variables with known prior distributions. This distinction allows Bayesian analysis to incorporate prior knowledge or beliefs about the parameters, resulting in a flexible and powerful approach to statistical inference.

    For a comprehensive exploration of these methodologies and their applications, we recommend referring to the work of [9,10]. In the present study, we study both MLE and Bayesian methods to investigate parameter estimation within the context of the GEVL distribution under a type-Ⅱ progressive censored sample.

    According to [11,12], there is no analytical solution for the maximum likelihood equations (ML.eqs) by the classical root-finding algorithms, such as the Newton-Raphson, Broyden's methods, and thus, some artificial intelligence algorithms were used. Therefore the maximization of the log-likelihood function can be obtained by using a standard numerical optimization algorithm like Non-linear minimization (NLM). Moreover, we compare the result obtained via the NLM method with one of the most popular algorithms in artificial intelligence genetic algorithm (GA).

    The NLM function is a valuable tool for numerical optimization, particularly for solving nonlinear optimization problems. It provides flexibility, stability, and efficiency, making it an important function for researchers, analysts, and practitioners working with complex optimization tasks. Moreover, the NLM function can handle optimization problems with multiple variables, allowing one to optimize functions of several parameters simultaneously. This makes it applicable in a wide range of fields, including statistics, econometrics, machine learning, and scientific research more details see [13]. The GA belongs to the field of artificial intelligence and specifically to the sub-field of Evolutionary computation. Evolutionary Computation encompasses computational methods inspired by biological evolution, and the GA is one of the prominent techniques within this domain. For more details in GA see [14,15].

    In certain experiments, it can be challenging to observe the failures of all of the units under investigation. This difficulty may arise from constraints such as limited time, budget constraints, or other practical considerations. Consequently, censored schemes are frequently employed in such situations. Censoring allows researchers to make inferences about the data, even when some of the information about failure times is incomplete or unavailable.

    One of the commonly used censoring schemes in the investigations of parameter estimation for any distribution is the type-Ⅱ progressive censored scheme. Type-Ⅱ progressive censored data can be described as follows: Let N units being subjected to an experiment, and only a predetermined number of units, denoted as n (where n<N), be observed until their failure. After each failure represented by x1:n:N, a specific quantity of units, R1, is chosen at random from the remaining units (N1). Following this, with the second failure, x2:n:N, another batch of units, R2, is randomly taken from the remaining (NR12), and this sequence continues until the n failure, xn:n:N which is the endpoint of the experiment. The type-Ⅱ progressive censored scheme allows researchers to collect and analyze data even when they are unable to observe the failure times of all units in the study. For more comprehensive information on this scheme, you can refer to the work of [16].

    This paper is organized as follows: In Section 2, we consider an explanation of all three cases under investigation (i.e., fixed, discrete uniform, and binomial random removal). In Section 3, MLE of GEVL distribution parameters based on type-Ⅱ progressively censored samples, for the three cases considered, are discussed for both point and approximate Confidence intervals. Section 4, Bayesian estimation of GEVL distribution parameters based on type-Ⅱ progressively censored samples, using both Lindley's approximation and Metropolis-Hastings (MH) algorithms for both informative and non-informative priors that involve the use of squared error loss function (SELF) and LINEX loss functions are presented. Moreover, the Bayesian estimation of GEVL distribution parameters is carried out by adopting the Gibbs sampling method for both the point estimation and the highest posterior density (HPD) confidence intervals estimation. The paper includes a simulation analysis in Section 5 to demonstrate the practical application of the theoretical findings. Real data examples are discussed in Section 6 to provide an application of the methodology in real-life situations. Finally, Section 7 presents the study conclusions.

    The associated sub-distribution hazards model, an established method for the regression analysis of time-to-event data in the presence of conflicting risks, depends heavily on the censoring distribution. Also, the features of the estimates from such a model when the censoring distribution is misspecified are examined by generating competing risk data under a proportional sub-distribution hazards model with various patterns of censoring. Models that correctly described the censoring distribution outperformed those that did not, providing estimates of the sub-distribution hazard ratio with less bias and risk. Estimates from the model based on these weights may not reflect the correct likelihood structure and, as a result, may have poor accuracy, particularly when the covariate of interest does not influence the censoring distribution but serves to calculate the risk set weights. In this section, we will indicate the three cases that will be utilized in this paper for the progressive type Ⅱ censoring schemes given by R=(R1,R2,,Rn). Suppose that, X1:n:N,X2:n:N,,Xn:n:N be the progressively type-Ⅱ censored ordering data from a GEVL distribution with the censoring scheme R=(R1,R2,,Rn) and the observed data x=(x1,x2,,xn).

    Case 1. Fixed removal censoring scheme

    For this case, the censoring schemes donated by R=(R1,R2,,Rn) are assumed to be predetermined fixed numbers.

    Case 2. Removals with discrete uniform distribution

    For this case, we assumed that R is an independent random variable following a discrete uniform distribution. The joint likelihood function of R is given by:

    P(R=r)=P(Rn=rn|Rn1=rn1,Rn2=rn2,,R1=r1)P(R2=r2|R1=r1)P(R1=r1), (2.1)

    where

    P(R1=r1)=1Nn+1, (2.2)

    where 0<r1<Nn and for i=2,3,,n1,

    P(Ri=ri|Ri1=ri1,Ri2=ri2R1=r1)=1Nni1j=1rj+1, (2.3)

    where 0<ri<Nni1j=1rj and Rn=Nnn1j=1Rj.

    Apply, Eqs (2.2) and (2.3) in Eq (2.1). Then the joint distribution of R=(R1,R2,,Rn) can be easily obtained as follows:

    P(R=r)=1Nn1n1i=11Nn(i1j=1rj)+1. (2.4)

    It is noticed that the joint PDF of R is parameter-free.

    Case 3. Removals with binomial distribution

    In this scenario, if the censoring scheme R=(R1,R2,,Rn) is assumed to consist of independent random variables following binomial distributions with a probability of P, then the joint distribution of R can be expressed as follows:

    P(R=r|P)=P(Rn=rn|Rn1=rn1,Rn2=rn2,,R1=r1,P)P(R2=r2|R1=r1,P)P(R1=r1,P), (2.5)

    where

    P(R1=r1|P)=(Nnr1)Pr1(1P)(Nnr1), (2.6)

    where 0<r1<Nn and for i=2,3,,n1,

    P(Ri=ri|P)=(Nni1j=1rjri)Pri(1P)(Nnij=1rj), (2.7)

    where 0<ri<Nni1j=1rj and Rn=Nnn1j=1Rj. Then the joint distribution of R=(R1,R2,,Rn) can be obtained easily by applying with Eqs (2.6) and (2.7) in Eq (2.5) as follows:

    P(R=r|P)=(Nn)!(Nnn1j=1rj)!(n1j=1rj!)Pn1i=1ri(1P)[(Nn)(n1)n1i=1(ni)rj]. (2.8)

    MLE method, which is a classical point estimation technique, combines both the overall distribution information and the observed sample data to facilitate comprehensive inference analysis. In the context of the GEVL distribution and type-Ⅱ progressive censoring, we have discussed the MLE approach for parameter estimation in three specific cases.

    First, under the assumption that the sample size approaches infinity, we derived parameter estimates by using the MLE method. Subsequently, we recognized that traditional numerical optimization techniques such as the Newton-Raphson method, and Broyden's method, among others, are commonly employed. However, in our study, we employed both artificial intelligence algorithms and optimization techniques to numerically compute the corresponding maximum likelihood estimates.

    Finally, we computed asymptotic confidence intervals to provide a comprehensive assessment of our findings and their statistical significance.

    Case 1. Fixed removals

    For this case, the joint likelihood function could be given by

    L(μ,σ,ξ)=Cni=1f(xi;μ,σ,ξ)[1F(xi;μ,σ,ξ)]Ri.

    Then

    L(μ,σ,ξ)ni=1f(xi;μ,σ,ξ)[1F(xi;μ,σ,ξ)]Ri, (3.1)

    where x1<x2<<xn and C is defined as follows:

    C=N(NR11)(NR1R22)(Nn1i=1Rin+1).

    Case 2. Removals with discrete uniforms

    For this case, the joint likelihood function can be obtained as follows:

    L(μ,σ,ξ)=P(R=r)L(μ,σ,ξ).

    Then

    L(μ,σ,ξ)ni=1f(xi;μ,σ,ξ)[1F(xi;μ,σ,ξ)]Ri, (3.2)

    where P(R=r) and L(μ,σ,ξ) are as given in Eqs (2.4) and (3.1) respectively.

    Case 3. Removals with binomial distribution

    For this case, the joint likelihood function can be obtained as follows:

    L(μ,σ,ξ,P)=P(R=r|P)L(μ,σ,ξ). (3.3)

    Then

    L(μ,σ,ξ,P)ni=1f(xi;μ,σ,ξ)[1F(xi;μ,σ,ξ)]RiPn1i=1ri(1p)(n1)(Nn)n1i=1ri. (3.4)

    where, P(R=r|P) and L(μ,σ,ξ) are given as in Eqs (2.8) and (3.1) respectively.

    Clearly, the ML.eqs for parameters μ,σ, and ξ for the three cases correspond to the same system of equations and the difference between the three cases is the distribution of the censored scheme R=(R1,R2,,Rn). In addition, for Case 3, the MLE of P could be easily obtained as follows:,

    ˆP=n1i=1ri(Nn)(n1)n1i=1(ni1)ri. (3.5)

    When obtaining the ML.eqs or parameters μ,σ, and ξ by substituting this Eqs (1.1) and (1.2) in Eq (3.1), the likelihood function can be obtained as follows:

    L(μ,σ,ξ)ni=1σ1Z1ξ1iUi(1Ui)Ri, (3.6)

    where, Zi=[1+(ξ(xiμ)σ)] and Ui=exp(Z1ξi).

    Then the log-likelihood function is given by

    log(L)nlog(σ)ni=1(1ξ+1)logZini=1Z1ξi+ni=1Rilog(1Ui). (3.7)

    The following three equations could be easily derived by taking the derivative of Eq (3.7) with respect to μ,σ, and ξ respectively, and making them equal to zero:

    δlogLδμ=ni=1(ξ+1σ)Z1i+ni=1σ1Z(1ξ1)i+ni=1RiZ[1ξ1]iUiσ[1Ui], (3.8)
    δlogLδσ=nσ+ni=1(ξ+1)(xiμ)Ziσ2ni=1(xiμ)Z1ξ+1iσ2ni=1RiUi(xiμ)Zi1ξ+1σ2[1Ui], (3.9)

    and

    δlogLδξ=ni=1log(Zi)[1Z1ξi+RiZ1ξiUi1ui]ξ2ni=1(xiμ)[1Z1ξi+RiUiZ1ξi[1Ui]]σZiξ. (3.10)

    The maximum likelihood estimators ˆμ,ˆσ and ˆξ based on type-Ⅱ progressively censored data can be obtained by setting Eqs (3.8)–(3.10) equal to zero and solveing them. The nonlinear characteristics of those likelihood equations are considered to be a challenge due to their complexity and the classical techniques like the Newton-Raphson method, failed to solve them. In such cases, numerical optimization methods, and artificial intelligence algorithms are typically employed to find the maximum likelihood estimators that best fit the observed data and likelihood equations. So, NLM and one of the most popular algorithms in artificial intelligence, i.e., the GA, are employed as indicated in Tables 14. A quick tour of GA is provided in detail the following website: https://cran.r-project.org/web/packages/GA/vignettes/GA.html. NLM is explained on [17].

    Table 1.  Thebias (mean square error; MSE) for MLE and Bayesian estimation, results, obtained by using both Lindley and MH for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(5,5,0.5)) for N=1000 under Case 1.
    Method GA NLM
    At Beginning At End At Beginning At End
    μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7
    Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais
    (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE)
    MLE 0.0438 -0.0443 -0.0003 0.0455 -0.0438 -0.0042 0.0116 0.0075 -0.2005 -0.0243 -0.0695 -0.3364
    (0.0027) (0.0024) (0.0015) (0.00300) (0.0025) (0.0015) (0.0003) (0.0002) (0.0411) (0.0007) (0.0049) (0.1141)
    NON Lindley Sq 0.0438 -0.0443 -0.0003 0.0455 -0.0438 -0.0042 0.0123 0.0094 -0.2003 -0.0241 -0.0688 -0.3367
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0410) (0.0007) (0.0048) (0.1143)
    lx(β=0.5) 0.04382 -0.0443 -0.0003 0.0455 -0.0438 -0.0042 0.0122 0.0084 -0.2006 -0.0241 -0.0692 -0.3368
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0411) (0.0007) (0.0049) (0.1144)
    lx(β=5) 0.04382 -0.0443 -0.0003 0.0455 -0.0437 -0.0042 0.0126 0.0098 -0.1979 -0.0239 -0.0687 -0.3352
    (0.0027) (0.0025) (0.0015) (0.0029) (0.0025) (0.0015) (0.0003) (0.0003) (0.0400) (0.0007) (0.0048) (0.1133)
    lx(β=5) 0.04382 -0.0443 -0.0003 0.0455 -0.0438 -0.0042 0.0119 0.009 -0.2027 -0.0243 -0.0689 -0.3381
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0420) (0.0007) (0.0048) (0.1153)
    MH Sq -0.0375 -0.1311 -0.3019 -0.0362 -0.1298 -0.3075 -0.0531 -0.105 -0.4056 -0.0708 -0.1442 -0.4746
    (0.0026) (0.0182) (0.1057) (0.0025) (0.0182) (0.1098) (0.0036) (0.0128) (0.1697) (0.0057) (0.0216) (0.2280)
    lx(β=0.5) -0.0395 -0.1332 -0.3109 -0.038 -0.1319 -0.3166 -0.0548 -0.1078 -0.4115 -0.0721 -0.1461 -0.4782
    (0.0027) (0.0188) (0.1113) (0.0026) (0.0187) (0.1154) (0.0038) (0.0134) (0.1744) (0.0059) (0.0222) (0.2314)
    lx(β=5) -0.0183 -0.1103 -0.2178 -0.0168 -0.1091 -0.2246 -0.0375 -0.0784 -0.3508 -0.0586 -0.1264 -0.4401
    (0.0015) (0.0133) (0.0603) (0.0016) (0.0133) (0.0646) (0.0022) (0.0078) (0.1278) (0.0041) (0.0168) (0.1964)
    lx(β=5) -0.0571 -0.1522 -0.3868 -0.0561 -0.1511 -0.3928 -0.0692 -0.1323 -0.4622 -0.0833 -0.1624 -0.5098
    (0.0043) (0.0242) (0.1639) (0.0042) (0.0240) (0.1684) (0.0055) (0.0191) (0.2183) (0.0076) (0.0272) (0.2625)
    INF Lindley Sq 0.04382 -0.0443 -0.0003 0.0455 -0.0437 -0.0042 0.0114 0.0084 -0.2018 -0.0244 -0.0691 -0.3375
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0024) (0.0015) (0.0003) (0.0002) (0.0416) (0.0007) (0.0049) (0.1148)
    lx(β=0.5) 0.04382 -0.0443 -0.0003 0.0455 -0.0437 -0.0042 0.0114 0.0084 -0.202 -0.0245 -0.0691 -0.3376
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0417) (0.0007) (0.0049) (0.1149)
    lx(β=5) 0.04382 -0.0443 -0.0003 0.0452 -0.0438 -0.0042 0.0118 0.0088 -0.1994 -0.0242 -0.0690 -0.3360
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0406) (0.0007) (0.0048) (0.1138)
    lx(β=5) 0.04382 -0.0443 -0.0002 0.0455 -0.0438 -0.0042 0.0111 0.008 -0.2042 -0.0246 -0.0693 -0.339
    (0.0027) (0.0025) (0.0015) (0.0030) (0.0025) (0.0015) (0.0003) (0.0002) (0.0425) (0.0007) (0.0049) (0.1158)
    MH Sq -0.0384 -0.1296 -0.3031 -0.0354 -0.1306 -0.2925 -0.0542 -0.1054 -0.411 -0.0717 -0.1437 -0.4785
    (0.0025) (0.0180) (0.1069) (0.0025) (0.0183) (0.1012) (0.0037) (0.0127) (0.1746) (0.0057) (0.0215) (0.2316)
    lx(β=0.5) -0.0404 -0.1318 -0.3121 -0.037 -0.1327 -0.3016 -0.0558 -0.1082 -0.4168 -0.0730 -0.1455 -0.4821
    (0.0026) (0.0185) (0.1125) (0.0026) (0.0188) (0.1067) (0.0038) (0.0133) (0.1793) (0.0058) (0.0221) (0.2350)
    lx(β=5) -0.0191 -0.1094 -0.2196 -0.0159 -0.1099 -0.2098 -0.0385 -0.0785 -0.3567 -0.0595 -0.1261 -0.4442
    (0.0015) (0.0131) (0.0617) (0.0015) (0.0134) (0.0573) (0.0022) (0.0077) (0.1323) (0.0041) (0.0168) (0.1999)
    lx(β=5) -0.058 -0.1507 -0.3885 -0.0553 -0.1517 -0.3784 -0.0701 -0.1328 -0.466 -0.0842 -0.1617 -0.5134
    (0.0042) (0.02380) (0.1650) (0.0041) (0.0241) (0.1591) (0.0056) (0.0192) (0.2224) (0.0076) (0.0270) (0.2660)

     | Show Table
    DownLoad: CSV
    Table 2.  The bais and (MSE) for MLE and Bayesian estimation results obtained by (using both Lindley and MH for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(5,5,0.5)) for N=50 under Case 1.
    Method GA NLM
    At Beginning At End At Beginning At End
    μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7
    Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais
    (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE)
    MLE 0.0586 -0.0649 0.014 0.0592 -0.0647 0.0102 0.0178 -0.0214 -0.3621 -0.0244 -0.0848 -0.4005
    (0.0040)) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0032) (0.0030) (0.1514) (0.0025) (0.0085) (0.1763)
    NON Lindley Sq 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0104 0.0271 0.0072 -0.3579 -0.0196 -0.0716 -0.408
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0017) (0.0037) (0.0032) (0.1467) (0.0023) (0.0067) (0.1814)
    lx(β=0.5) 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0104 0.0266 -0.0076 -0.3637 -0.02 -0.0784 -0.4119
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0017) (0.0036) (0.0029) (0.1508) (0.0023) (0.0076) (0.1845
    lx(β=5) 0.0586 -0.0648 0.0138 0.0592 -0.0647 0.0104 0.0325 0.0102 -0.3079 -0.016 -0.0699 -0.371
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0017) (0.0042) (0.0034) (0.1150) (0.0023) (0.0065) (0.1538)
    lx(β=5) 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0103 0.0212 0.0028 -0.4096 -0.0235 -0.0737 -0.4419
    (0.0040)) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0032) (0.0030) (0.1858) (0.0024) (0.0069) (0.2098)
    MH Sq -0.0299 -0.141 -0.2873 -0.0298 -0.1409 -0.2844 -0.0505 -0.1192 -0.4901 -0.0717 -0.1499 -0.5085
    (0.0023) (0.0210) (0.0985) (0.0022) (0.0209) (0.0973) (0.0041) (0.0162) (0.2478) (0.0061) (0.0236) (0.2648)
    lx(β=0.5) -0.032 -0.1429 -0.2969 -0.0319 -0.1428 -0.2937 -0.0522 -0.1217 -0.4936 -0.073 -0.1516 -0.5114
    (0.0024) (0.0215) (0.1041) (0.0023) (0.0214) (0.1028) (0.0043) (0.0168) (0.2509) (0.0063) (0.0241) (0.2675)
    lx(β=5) -0.0089 -0.1228 -0.2006 -0.0089 -0.1226 -0.2002 -0.0338 -0.0956 -0.4576 -0.0592 -0.1336 -0.4809
    (0.0014) (0.0162) (0.0540) (0.0014) (0.0161) (0.0544) (0.0030) (0.0114) (0.2196) (0.0047) (0.0191) (0.2396)
    lx(β=5) -0.0511 -0.1596 -0.3778 -0.0511 -0.1596 -0.3728 -0.0674 -0.1433 -0.5234 -0.0845 -0.1667 -0.5366
    (0.0039) (0.0265) (0.1582) (0.0038) (0.0264) (0.1551) (0.0058) (0.0222) (0.2790) (0.0079) (0.0287) (0.2922)
    INF Lindley Sq 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0103 0.0177 -0.0031 -0.4095 -0.0232 -0.075 -0.4418
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0029) (0.0026) (0.1843) (0.0023) (0.0070) (0.2089)
    lx(β=0.5) 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0103 0.0171 -0.0036 -0.4145 -0.0235 -0.0752 -0.4451
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0028) (0.0026) (0.1885) (0.0023) (0.007) (0.2118)
    lx(β=5) 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0104 0.0235 0.0013 -0.3512 -0.0194 -0.073 -0.4026
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0033) (0.0028) (0.1411) (0.0022) (0.0068) (0.1766)
    lx(β=5) 0.0586 -0.0648 0.0139 0.0592 -0.0647 0.0101 0.0119 -0.0085 -0.4455 -0.027 -0.0772 -0.4669
    (0.0040) (0.0046) (0.0016) (0.0041) (0.0046) (0.0016) (0.0026) (0.0025) (0.2157) (0.0024) (0.0073) (0.2318)
    MH Sq -0.0302 -0.1399 -0.2877 -0.0293 -0.1425 -0.2891 -0.0492 -0.1195 -0.4891 -0.071 -0.1529 -0.507
    (0.0022) (0.0207) (0.1001) (0.0019) (0.0213) (0.0992) (0.0042) (0.0163) (0.2474) (0.0060) (0.0244) (0.2638)
    lx(β=0.5) -0.0323 -0.1418 -0.2969 -0.0315 -0.1443 -0.2986 -0.0509 -0.1219 -0.4926 -0.0723 -0.1545 -0.5099
    (0.0023) (0.0213) (0.1057) (0.0021) (0.0218) (0.1049) (0.0043) (0.0168) (0.2506) (0.0062) (0.0249) (0.2665)
    lx(β=5) -0.0094 -0.1219 -0.203 -0.0082 -0.1242 -0.2026 -0.0326 -0.0962 -0.4563 -0.0586 -0.1367 -0.4798
    (0.0014) (0.0160) (0.0562) (0.0012) (0.0165) (0.0544) (0.0031) (0.0115) (0.2191) (0.0046) (0.0199) (0.2390)
    lx(β=5) -0.0513 -0.1585 -0.375 -0.0509 -0.1611 -0.3774 -0.0663 -0.1433 -0.5228 -0.0837 -0.1693 -0.535
    (0.0038) (0.0262) (0.1580) (0.0035) (0.0269) (0.1584) (0.0058) (0.0222) (0.2789) (0.0078) (0.0295) (0.2910)

     | Show Table
    DownLoad: CSV
    Table 3.  The Bais and (MSE) for MLE and Bayesian estimation results obtained by using both Lindley and MH for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(5,5,0.5)) for N=1000 under Cases 2 and 3.
    Method GA NLM
    Case2 Case3 Case2 Case3
    μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7
    Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais
    (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE)
    MLE 0.0326 -0.0311 0.0104 0.0274 -0.0273 0.01229 -0.0009 -0.0061 -0.0308 0.0005 -0.0028 -0.0188
    (0.0038) (0.0032) (0.0027) (0.0040) (0.0034) (0.0030) (0.0001) (0.0002) (0.0023) (0.0001) (0.0002) (0.0016)
    NON Lindley Sq 0.0327 -0.0308 0.0112 0.0276 -0.0269 0.01347 -0.0001 -0.0044 -0.0304 0.0013 -0.0012 -0.0183
    (0.0038) (0.0033) (0.0029) (0.0039) (0.0035) (0.0016) (0.0001) (0.0002) (0.0023) (0.0001) (0.0002) (0.0016)
    lx(β=0.5) 0.0326 -0.0311 0.0107 0.0275 -0.0271 0.0134 -0.0002 -0.0053 -0.0307 0.0013 -0.0020 -0.0186
    (0.0038) (0.0032) (0.0028) (0.0039) (0.0035) (0.0015) (0.0001) (0.0002) (0.0023) (0.0001) (0.0002) (0.0016)
    lx(β=5) 0.0327 -0.031 0.0112 0.0276 -0.0216 -0.2964 0.0002 -0.0041 -0.0272 0.0013 -0.002 -0.0151
    (0.0038) (0.0033) (0.0029) (0.0039) (0.0035) (0.0017) (0.0001) (0.0002) (0.0021) (0.0001) (0.0002) (0.0015)
    lx(β=5) 0.0326 -0.0311 0.0106 0.0274 -0.0271 -0.2967 -0.0004 -0.0048 -0.0336 0.0017 -0.0008 -0.0216
    (0.0038) (0.0032) (0.0028) (0.0039) (0.0035) (0.0886) (0.0001) (0.0002) (0.0025) (0.0001) (0.0002) (0.0017)
    MH Sq 0.0327 -0.0307 0.0116 0.0276 -0.0268 -0.2881 -0.0585 -0.1128 -0.3109 0.001 -0.0015 -0.3080
    (0.0038) (0.0033) (0.0030) (0.0039) (0.0035) (0.0844) (0.0042) (0.0140) (0.1102) (0.0001) (0.0002) (0.1089)
    lx(β=0.5) 0.0326 -0.031 0.0111 0.0275 -0.027 -0.2898 -0.06 -0.1154 -0.3196 -0.0594 -0.1126 -0.3125
    (0.0038) (0.0033) (0.0029) (0.0039) (0.0035) (0.0851) (0.0044) (0.0146) (0.1158) (0.0043) (0.0141) (0.1114)
    lx(β=5) 0.0327 -0.0308 0.0107 0.0276 -0.0269 -0.0156 -0.0442 -0.0877 -0.231 -0.0609 -0.1153 -0.3214
    (0.0038) (0.0033) (0.0028) (0.0039) (0.0035) (0.0889) (0.0027) (0.0089) (0.0651) (0.0045) (0.0147) (0.1170)
    lx(β=5) 0.0326 -0.0311 0.0102 0.0274 -0.0271 -0.2972 -0.0732 -0.1384 -0.3934 -0.0448 -0.0871 -0.2312
    (0.0038) (0.0032) (0.0027) (0.0040) (0.0035) (0.0889) (0.0061) (0.0203) (0.1681) (0.0028) (0.0089) (0.0660)
    INF Lindley Sq -0.0439 -0.1242 -0.2941 -0.0444 -0.123 -0.2913 -0.001 -0.0056 -0.0322 -0.0743 -0.1385 -0.0183
    (0.0034) (0.0172) (0.1031) (0.0038) (0.0171) (0.1021) (0.0001) (0.0002) (0.0024) (0.0062) (0.0205) (0.0016)
    lx(β=0.5) -0.0418 -0.1234 -0.2875 -0.046 -0.1227 -0.2883 -0.001 -0.0057 -0.0325 0.0003 -0.0022 -0.0186
    (0.0035) (0.0171) (0.0988) (0.0039) (0.0171) (0.0992) (0.0001) (0.0002) (0.0024) (0.0001) (0.0002) (0.0016)
    lx(β=5) -0.0458 -0.1265 -0.3033 -0.0462 -0.1253 0.01282 -0.0007 -0.0053 -0.029 0.0003 -0.0023 -0.0150
    (0.0036) (0.0178) (0.1087) (0.0040) (0.0176) (0.1077) (0.0001) (0.0002) (0.0022) (0.0001) (0.0002) (0.0015)
    lx(β=5) -0.0437 -0.1257 -0.297 -0.0478 -0.1251 -0.2979 -0.0013 -0.006 -0.0353 0.0006 -0.0019 -0.0215
    (0.0036) (0.0177) (0.1044) (0.0040) (0.0177) (0.1049) (0.0001) (0.0002) (0.0026) (0.0001) (0.0002) (0.0017)
    MH Sq -0.0257 -0.102 -0.2092 -0.0269 -0.1002 -0.2049 -0.0585 -0.1124 -0.3176 -0.2135 -0.0026 -0.2991
    (0.0024) (0.0125) (0.0585) (0.0028) (0.0123) (0.0574) (0.0043) (0.0142) (0.1141) (0.0001) (0.0002) (0.0901)
    lx(β=0.5) -0.0235 -0.1009 -0.2015 -0.0283 -0.1003 -0.2009 -0.06 -0.115 -0.3263 -0.0597 -0.1124 -0.3104
    (0.0025) (0.0124) (0.0548) (0.0029) (0.0124) (0.0542) (0.0045) (0.0148) (0.1198) (0.0042) (0.0139) (0.1105)
    lx(β=5) -0.0625 -0.1469 -0.3808 -0.0623 -0.1462 -0.3797 -0.0441 -0.0876 -0.2372 -0.0612 -0.1151 -0.3194
    (0.0051) (0.0231) (0.1613) (0.0054) (0.0230) (0.1611) (0.0028) (0.0091) (0.0677) (0.0044) (0.0145) (0.1162)
    lx(β=5) -0.0605 -0.1464 -0.3766 -0.0641 -0.146 -0.3781 -0.0732 -0.1378 -0.3993 -0.0451 -0.087 -0.2275
    (0.0052) (0.0230) (0.1576) (0.0055) (0.0230) (0.1592) (0.0062) (0.0205) (0.1722) (0.0027) (0.0088) (0.0643)

     | Show Table
    DownLoad: CSV
    Table 4.  The Bais and (MSE) for MLE and Bayesian estimation results obtained by using both Lindley and MH for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(5,5,0.5)) for N=50 under Cases 2 and 3.
    Method GA NLM
    Case2 Case3 Case2 Case3
    Parameters μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7 μ=0.2 σ=0.3 ξ=0.7
    Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais Bais
    (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE) (MSE)
    MLE 0.0561 0.0561 0.0262 0.0514 -0.0571 0.0241 0.0037 -0.0351 -0.2288 0.0056 -0.0325 -0.2179
    (0.0043) (0.0041) (0.0029) (0.0041) (0.0045) (0.0030) (0.0025) (0.0035) (0.0836) (0.0026) (0.0033) (0.0825)
    NON Lindley Sq 0.0564 -0.0553 0.0296 0.0521 -0.0531 -0.2584 0.0148 -0.009 -0.2175 0.0174 -0.0064 -0.2551
    (0.0043) (0.0043) (0.0043) (0.0040) (0.0060) (0.0865) (0.0027) (0.0028) (0.0770) (0.0029) (0.0028) (0.0768)
    lx(β=0.5) 0.0564 -0.0557 0.0296 0.0521 -0.0551 0.2594 0.0143 -0.0225 -0.225 0.0169 -0.0199 -0.2624
    (0.0043) (0.0042) (0.0043) (0.0040) (0.0050) (0.0866) (0.0026) (0.0030) (0.0799) (0.0028) (0.0029) (0.0828)
    lx(β=5) 0.0564 -0.0553 0.0293 0.0521 -0.054 -0.2555 0.0197 -0.0061 -0.1575 0.0224 -0.0034 -0.1475
    (0.0043) (0.0043) (0.0041) (0.0040) (0.0055) (0.0866) (0.0030) (0.0029) (0.0582) (0.0032) (0.0029) (0.0363)
    lx(β=5) 0.0564 -0.0553 0.0278 0.0521 -0.0502 -0.2696 0.0094 -0.0129 -0.2826 0.0118 -0.0105 -0.3213
    (0.0043) (0.0043) (0.0034) (0.0040) (0.0087) (0.0854) (0.0024) (0.0028) (0.1070) (0.0025) (0.0027) (0.1122)
    MH Sq -0.0296 -0.1381 -0.2805 -0.033 -0.1376 -0.2822 -0.0571 -0.1274 -0.4224 -0.0554 -0.1259 -0.4085
    (0.0024) (0.0203) (0.0951) (0.0024) (0.0204) (0.0970) (0.0046) (0.0180) (0.1926) (0.0045) (0.0174) (0.1834)
    lx(β=0.5) -0.0317 -0.1401 -0.2901 -0.0351 -0.1395 -0.2917 -0.0587 -0.1297 -0.4279 -0.057 -0.1282 -0.4142
    (0.0025) (0.0208) (0.1006) (0.0026) (0.0209) (0.1027) (0.0047) (0.0186) (0.1968) (0.0047) (0.0179) (0.1867)
    lx(β=5) -0.0091 -0.1189 -0.1931 -0.0127 -0.1186 -0.1948 -0.042 -0.1055 -0.3706 -0.0401 -0.1036 -0.3549
    (0.0016) (0.0154) (0.0518) (0.0017) (0.0156) (0.0529) (0.0033) (0.0131) (0.1564) (0.0033) (0.0126) (0.1471)
    lx(β=5) -0.0507 -0.1577 -0.3708 -0.0537 -0.1571 -0.3717 -0.0726 -0.1497 -0.4749 -0.0712 -0.1486 -0.4631
    (0.0039) (0.0259) (0.1535) (0.0041) (0.0259) (0.1557) (0.0063) (0.0239) (0.2356) (0.0062) (0.0233) (0.2265)
    INF Lindley Sq 0.0562 -0.0561 0.0277 0.0516 -0.0542 -0.2603 0.0046 -0.0234 -0.2698 0.0001 -0.0224 -0.2408
    (0.0043) (0.0041) (0.0033) (0.0041) (0.0055) (0.0858) (0.0021) (0.0027) (0.0980) (0.0021) (0.0027) (0.0720)
    lx(β=0.5) 0.0562 -0.0562 0.0275 0.0516 -0.0541 -0.2613 0.0041 -0.0238 -0.2765 -0.0005 -0.0229 -0.248
    (0.0043) (0.0041) (0.0032) (0.0041) (0.0056) (0.0859) (0.0021) (0.0028) (0.1014) (0.0021) (0.0027) (0.0758)
    lx(β=5) 0.0563 -0.056 0.0286 0.0516 -0.0547 -0.2558 0.0101 -0.0189 -0.1978 0.0059 -0.0176 -0.1376
    (0.0043) (0.0041) (0.0037) (0.0041) (0.0052) (0.0862) (0.0024) (0.0027) (0.0687) (0.0023) (0.0027) (0.0334)
    lx(β=5) 0.0562 -0.0563 0.0255 0.0515 -0.0538 -0.2702 -0.0008 -0.0284 -0.3181 -0.0053 -0.0276 -0.3066
    (0.0043) (0.0041) (0.0028) (0.0041) (0.0060) (0.0854) (0.0020) (0.0028) (0.1262) (0.0021) (0.0027) (0.1035)
    MH Sq -0.0309 -0.1372 -0.2846 -0.0325 -0.1378 -0.2820 -0.0570 -0.1265 -0.4215 -0.0564 -0.1269 -0.4111
    (0.0025) (0.0200) (0.0980) (0.0026) (0.0204) (0.0963) (0.0048) (-0.0178) (0.1931) (0.0046) (0.0177) (0.1845)
    lx(β=0.5) -0.033 -0.1392 -0.2941 -0.0346 -0.1398 -0.2917 -0.0586 -0.1288 -0.4268 -0.058 -0.1292 -0.4169
    (0.0026) (0.0205) (0.1035) (0.0027) (0.0209) (0.1019) (0.0049) (0.0183) (0.1973) (0.0048) (0.0183) (0.1888)
    lx(β=5) -0.0104 -0.1178 -0.1978 -0.0121 -0.1187 -0.1942 -0.0419 -0.1044 -0.371 -0.0411 -0.1045 -0.3568
    (0.0017) (0.0151) (0.0544) (0.0018) (0.0156) (0.0526) (0.0035) (0.0129) (0.1571) (0.0034) (0.0128) (0.1480)
    lx(β=5) -0.0519 -0.157 -0.3744 -0.0533 -0.1573 -0.3733 -0.0725 -0.1491 -0.4726 -0.0719 -0.1497 -0.4672
    (0.0040) (0.0257) (0.1565) (0.0042) (0.0259) (0.1555) (0.0065) (0.0237) (0.2351) (0.0064) (0.0236) (0.2287)

     | Show Table
    DownLoad: CSV

    The observed Fisher information can be computed based on both the full likelihood and the approximate likelihood equations. These formulas serve as the foundation for the creation of pivotal quantities, which are used to explore the coverage probabilities in the context of the limiting normal distribution. To rigorously assess the performance of these pivotal quantities, Monte Carlo simulations were conducted. In situations in which the sample size is sufficiently large (i.e., N and n both exceed zero), we also consider the construction of approximate confidence intervals for the three specific cases that have been previously discussed. These confidence intervals provide valuable insights into the precision of parameter estimates and their statistical significance in various scenarios.

    Under the assumption of censoring scheme distribution, the Fisher information matrix, based on log-likelihood functions given in Eqs (3.1) and (3.2) for Cases 1 and 2 can be obtained as follows:

    I=[log(L)μ2log(L)μσlog(L)μξlog(L)σμlog(L)σ2log(L)σξlog(L)ξμlog(L)ξσlog(L)ξ2], (3.11)

    while, for Case 3 the fisher information matrix I(θ) is obtained as follows:

    I=[log(L)μ2log(L)μσlog(L)μξ0log(L)σμlog(L)σ2log(L)σξ0log(L)ξμlog(L)ξσlog(L)ξ20000δ(P)], (3.12)

    where,

    log(L)μ2=ni=1(1+1ξ)ξ2Z2ini=1(1+ξ)Z2+1ξi+ni=1Ri[1+ξZ1ξi(Z1ξiUi)(1Ui)]UiZ2+1ξi[1Ui], (3.13)
    log(L)μσ=log(L)σμ=ni=1(1+ξ)[1(xiμ)ξσZi]σ2Zini=1[(xiμ)(1+ξ)σZi]σ2Z2+1ξi+ni=1RiUi[(xiμ)σ(1+ξZ1ξiZ1ξiUi[1Ui])Zi]σ2Z2+1ξi[1Ui], (3.14)
    log(L)μξ=log(L)ξμ=ni=1[1(1+ξ)(xiμ)σZi]σZini=1[log(Zi)ξ2(xiμ)[1+1ξ]σZi]σZ1+1ξi+ni=1RiUi[log(Zi)ξ2(1Z1ξi[1+Ui[1Ui]])(xiμ)σZi(1+1ξZ1ξiξ[1+Ui[1Ui]])]σZ1+1ξi[1Ui], (3.15)
    log(L)σ2=nσ2ni=1(1+ξ)(xiμ)[2ξ2σZi]σ3Zini=1(xiμ)[(xiμ)(1+ξ)σZi2]σ3Z1+1ξi+ni=1Ri(xiμ)2Ui[22σZi(xiμ)Z1ξi[1+Ui[1Ui]]]σ4Z2+1ξi[1Ui], (3.16)
    log(L)σξ=log(L)ξσ=ni=1(xiμ)[1ξ(1+1ξ)[1+ξ(xiμ)]σZi]σ2Zini=1(xiμ)[log(Zi)ξ2(xiμ)(1+1ξ)σZi]σ2Z1+1ξi+ni=1Ri(xiμ)Ui[log(Zi)ξ2(1Z1ξiZ1ξiUi[1Ui])+(xiμ)σZi((1ξ+1)+Z1ξiξ+Z1ξiUiξ[1Ui])]σ2Z1+1ξi[1Ui], (3.17)
    log(L)ξ2=ni=1[(xiμ)σZi(2ξ2+(xiμ)(1+1ξ)σZi)2log(Zi)ξ3]ni=1Z1ξi[log(Zi)ξ2(log(Zi)ξ22(xiμ)σZiξ2)+(xiμ)σξZi(2ξ+(xiμ)(1+1ξ)σZi)]+ni=1RiZ1ξiUi[(log(Zi)ξ2)2(1Z1ξi)+(xiμ)2σ2Z2iξ2(1+ξZ1ξi)][1Ui]ni=1RiZ1ξiUi[2log(Zi)ξ2((xiμ)σZiξ+1(xiμ)ξσZ1+1ξi)][1Ui], (3.18)

    and

    δ(P)=log(L3)P2=ni=1riP2(n1)(Nn)n1i=1(ni)ri(1P)2, (3.19)

    where, Zi=[1+ξ(xiμ)σ] and Ui=exp(Z1ξi).

    Similarly, the observed Fisher information could be obtained by supplying the maximum likelihood values ofthe parameters for Eqs (3.8) and (3.10).

    The variance-covariance matrix could be obtained by applying the inverse of matrices I or I given by Eqs (3.11) and (3.12) which depend on the case under study. According to [18], the asymptotic distribution of parameter ν for the proposed cases (Cases 1–3) follows a normal distribution ˆνN(ˆν,I1 where I1 is the variance-covariance matrix. Therefore, the asymptotic 100(1ζ)% confidence interval for parameters ν with a significance level is given by

    [ˆνzν2I1,ˆν+zν2I1]. (3.20)

    Then the confidence intervals of distribution parameters are given by

    [ˆμzζ2I1,ˆμ+zζ2I1],[ˆσzζ2I1,ˆσ+zζ2I1],[ˆξzζ2I1,ˆξ+zζ2I1], (3.21)

    where

    I1={I1,For cases1and2,I1,For case3. (3.22)

    The Bayesian estimation method is distinct in that it does not solely rely on the observed sample data, as it also incorporates prior information about the distribution of samples. Consequently, Bayesian estimation leverages both the available population distribution information and prior probabilities. This approach allows for a more objective and rational description of unknown parameters. In the following section, we consider the estimations of parameters by using Bayesian methods and, two different loss functions: the square loss function and the LINEX loss function. These estimations are obtained under two scenarios: one utilizing informative priors and the other employing non-informative priors. The incorporation of these loss functions and prior information enables a comprehensive analysis of the parameter estimation in a Bayesian framework.

    Informative priors

    Let the unknown parameters for all cases considered to be independent of each other. The method of choosing parameter priors for the informative case depends on the parameter validation region, as introduced by [19,20]. Suppose that the parameters μ,σ,ξ follow an exponential distribution with the hyperparameters a1,a2, and a3, while the random variable P follows a beta distribution with the parameters [α,γ], Then the prior PDFs of parameters μ, σ, ξ, and P are given respectively by

    π1(μ)=1a1exp(a1μ),a1,μ>0,π2(σ)=1a2exp(a2σ),a2,σ>0,π3(σ)=1a3exp(a3ξ),a3,ξ>0, (4.1)

    and,

    π4(P)=1B(α,γ)Pα1(1p)1γ,α,γ,P>0, (4.2)

    where B(α,γ) is the beta function and all of the hyper-parameters (a1,a2,a3,α,γ) are estimated by using the same method given in [21].

    Non-informative priors

    For this case, we assume that all prior PDFs of the parameters μ, σ, ξ, and P are equal to 1.

    Based on type-Ⅱ progressively censored data, for all distributions of the censored scheme discussed above, Bayesian estimation is discussed in this section for both point and interval estimation.

    According to the informative prior functions of parameters μ,σ,ξ and P. The posterior PDFs are given by

    πj(μ,σ,ξ)=L(μ,σ,ξ)exp((a1μ+a2σ+a3ξ))000L(μ,σ,ξ)exp((a1μ+a2σ+a3ξ))dμdσdξ,j=1,2=ni=1σ1Z1ξ1iUi[1Ui]Rie(a1μ+a2σ+a3ξ)000ni=1σ1Z1ξ1iUi[1Ui]Rie(a1μ+a2σ+a3ξ)dμdσdξni=1σ1Z1ξ1iUi[1Ui]Rie(a1μ+a2σ+a3ξ) (4.3)
    π3(μ,σ,ξ,P)=Pn1i=1ri+α1(1P)(n1)(Nn)+n1i=1(ni)ri+γ1B(n1i=1ri+α,(n1)(Nn)+n1i=1(ni)ri+γ)πj(μ,σ,ξ),j=1,2, (4.4)

    where πj,j=1,2,3 depends on the case of the censoring scheme under study which was mentioned previously, Zi=[1+ξ(xiμ)σ] and Ui=exp(Z1ξi).

    Then the Bayesian estimation of parameters of the GEVL distribution based on progressive type-Ⅱ censoring θ=(μ,σ,ξ) for Case (1, 2), and θ=(μ,σ,ξ,P) for case 3 for the SELF loss function for the informative prior is given by

    ˆθS=θπj(θ)dθ,j=1,2,3. (4.5)

    Also, the Bayesian estimation of GEVL distribution parameters is considered for the LINEX loss function Llx(θ)=(exp(βθ)(βθ)1) for the informative prior which is given by, (see [22])

    ˆθlx=1βlog(exp(βθ)πj(θ)dθ),j=1,2,3 (4.6)

    where β is a shape parameter in which the negative value of β provides more weight to underestimation than the overestimation while for very (small or large) values of β the LINEX loss function is almost symmetric (see [22]).

    For non-informative priors, the posterior PDFs are given by

    πj(μ,σ,ξ)=L(μ,σ,ξ)000L(μ,σ,ξ)dμdσdξ,j=1,2=ni=1σ1Z1ξ1iUi[1Ui]Ri000ni=1σ1Z1ξ1iUi[1Ui]Ridμdσdξni=1σ1Z1ξ1iUi[1Ui]Ri (4.7)
    π3(μ,σ,ξ,P)=Pn1i=1ri(1P)(n1)(Nn)+n1i=1(ni)riB(n1i=1ri+1,(n1)(Nn)+n1i=1(ni)ri+1)πj(μ,σ,ξ),j=1,2, (4.8)

    where πj,j=1,2,3 depends on the case of the censoring scheme under study, as mentioned above, Zi=[1+ξ(xiμ)σ] and Ui=exp(Z1ξi).

    Then the Bayesian estimation of parameters of the GEVL distribution based on progressive type-Ⅱ censoring under the SELF and LINEX loss functions for the informative prior are given respectively by,

    ˆθS=θπj(θ)dθ,j=1,2,3, (4.9)
    ˆθlx=1βlog(exp(βθ)πj(θ)dθ),j=1,2,3, (4.10)

    Since the systems of equations given by Eqs (4.5), (4.6), (4.9) and (4.10) cannot be reduced analytically, we solve them numerically. The most popular numerical techniques for Bayesian estimation are Lindley's approximation method and Gibbs sampling with MH algorithms which are discussed in detail in the next subsection.

    1) Lindley's approximation method

    Among the various methods suggested to approximate the ratio of integrals given in the systems of equations given by Eqs (4.5), (4.6), (4.9) and (4.10), e.g., the Markov chain Monte Carlo method, Gibbs sampler, and Lindley's approximation, perhaps the simplest one is Lindley's approximation method. The Lindley approximation method can be described as, finding the following expectation:

    E(U(ϑ))=U(ϑ)exp(L(ϑ)+ρ(ϑ))d(ϑ)exp(L(ϑ)+ρ(ϑ))d(ϑ), (4.11)

    where ϑ=(ϑ1,ϑ2,ϑr), U(ϑ) is any function of ϑ, L(ϑ) is the log likelihood function of ϑ and ρ(ϑ) is the log of the joint prior of ϑ. Then Lindley's approximation of this integral is given by

    E(U(ϑ))=U(ˆϑ)+12ri=1rj=1[Ui,j(ˆϑ)+2Ui(ˆϑ)ρj(ˆϑ)]I1(ˆϑ)i,j+12ri=1rj=1rk=1rl=1Li,j,k(ˆϑ)Ul(ˆϑ)I1(ˆϑ)i,jI1(ˆϑ)k,l, (4.12)

    where Ui=Uϑi, ρj=ρϑj, Ui,j=2Uϑiϑj, Li,j,k=3Lϑiϑjϑk, and I1(ˆϑ)i,j is equal to the variance covariance matrix. All of the partial derivatives are evaluated by using the MLEs of ϑ (see [23]). For our case study, we considered U(ϑ) to be any parameter from θ for the proposed cases (Cases 1–3) and L(ϑ) is given by at Eq (3.6). Additionally,

    ρ(ϑ)={exp((a1μ+a2σ+a3ξ)),forCases1and2,Pn1i=1ri+α1(1P)(n1)(Nn)+n1i=1(ni)ri+γ1exp((a1μ+a2σ+a3ξ)),forCase3, (4.13)

    2) MH algorithms

    Gibbs sampling is a popular method and one of the commonly used Bayesian estimation methods for estimating specific distribution attributes. If it is not easy to generate samples directly from the posterior distribution, it is convenient to apply the Gibbs sampling method with MH algorithms. For more details on MH algorithms see [24].

    Let ϑ=(ϑ1,ϑ2,ϑr) be a vector of the parameter we want to estimate. The MH algorithm could be described as follows: suppose that our goal is to draw samples from some distribution f(ϑ|x)=νg(ϑ), where ν is the normalizing constant which may not be known or very difficult to compute. The MH algorithm provides a way of sampling from f(ϑ|x) without requiring any knowledge of ν.

    For any ϑjϑ, let q(ϑb|ϑaj) be the transition kernel (proposal distribution), that is the probability of jumping from the current state ϑaj to ϑbj. The following steps will generate a sequence of values (ϑ1j,ϑ2j,) which form a Markov chain with a stationary distribution given by f(ϑj|x).

    (1) Choose an arbitrary starting point ϑ0j for which f(ϑ0j|x)>0.

    (2) For i=1, sample a candidate point or proposal ϑj from the proposal distribution q(ϑj|ϑi1j).

    (3) Calculate the acceptance probability ρ(ϑi1j,ϑj)=min[1,f(ϑj|x)q(ϑi1j|ϑij)f(ϑj|x)q(ϑij|ϑi1j)].

    (4) Generate UU(0,1).

    (5) If U<ρ(ϑi1j,δ) accept the proposal and set ϑij=ϑj. Otherwise, reject the proposal and set δi=δi1.

    (6) Repeat Steps 2–5. If the proposal distribution is symmetric then the acceptance condition becomes ρ(δi1,δ)=min[1,f(ϑj|x)f(ϑj|x)].

    (7) We could get a set of Gibbs (MH) samplings of parameters ϑ1j,ϑ2j,,ϑ10000j. Now, the approximate expectation of reliability ϑj could be calculated as follows:

    E(ϑj)=1tbbtk=bb+1ϑkj.

    For our case of study, from Eq (4.3) the proper density function of parameters μ,σ and ξ can obtained as follows:

    f(μ|σ,ξ)ni=1[1+ξ(xiμ)σ]1ξ1exp((a1μ)[1+ξ(xiμ)σ]1ξ)[1exp([1+ξ(xiμ)σ]1ξ)]Ri,f(σ|μ,ξ)ni=1σ1[1+ξ(xiμ)σ]1ξ1exp((a2σ)[1+ξ(xiμ)σ]1ξ)[1exp([1+ξ(xiμ)σ]1ξ)]Ri,f(ξ|μ,σ)ni=1σ1[1+ξ(xiμ)σ]1ξ1exp((a3ξ)[1+ξ(xiμ)σ]1ξ)[1exp([1+ξ(xiμ)σ]1ξ)]Ri. (4.14)

    Moreover, these proper densities cannot as be reduced analytically to a well-known distribution but its plots show that as they are similar to a normal distribution. So, the Mh algorithm with normal proposal distribution is used for Bayesian estimation of every parameter as by following the Steps 1–7 mentioned above.

    To compute HPD credible interval of the parameters denoted by θ=(μ,σ,ξ), first order the importance of MH samplings

    (ˆμbb+1,ˆσbb+1,ξbb+1),(ˆμbb+2,ˆσbb+2,ˆξbb+2),,(ˆμt,ˆσt,ˆξt),

    Then we could obtain three sets of ascending order samplings:

    ˆμaa+1<ˆμaa+2<<ˆμt,ˆσaa+1<ˆσaa+2<<ˆσt,ˆξaa+1<ˆξaa+2<<ˆξt, (4.15)

    Therefore, the 100(1α)% HPD credible interval of θ=(μ,σ,ξ) can be calculated as follows:

    (ˆμ(α2)(taa),ˆμ(1α2)(taa)),(ˆσ(α2)(taa),ˆσ(1α2)(taa)),(ˆξ(α2)(taa),ˆξ(1α2)(taa)). (4.16)

    In this section, a Monte Carlo simulation presented to evaluate as well as compare the efficacy of different estimators in determining the parameters (μ,σ,ξ) of GEVL distribution, as introduced in Sections 3 and 4. This simulation was conducted to rigorously assess various estimation techniques and their performance in terms of accurately estimating these crucial parameters. To study the behavior of the GA and NLM methods under small and large samples, 1000 samples each of size 50 and 1000 were generated for X. The MLE of (μ,σ,ξ) was contrasted with Bayesian estimators utilizing both the SELF and the LINEX loss function. This comparison aims to assess the performance and robustness of these estimation techniques in terms of the accuracy of parameter estimation.

    we performed a comprehensive comparison between various estimation techniques by using Bayesian approaches and MLE, considering different loss functions like LINEX and numerical techniques such as The Lindley and MH. The method of selection of hyperparameters was explored and derived by using a method previously employed in [21]. The comparison is based on estimated values, the MSE, and 95% HPD for both informative and non-informative Bayesian estimation. The LINEX loss function was examined across three parameter values: a very large value β=5, a very small value β=0.5, and a positive value β=5, detailed in Tables 14. All computationsextcolorredwere executed by using the R program, assessing estimators with fixed random removals. The censoring schemes involve a 10 percent elimination either at the beginning or end of the sample. The algorithm for generating progressive censoring of type Ⅱ can be seen in [16].

    Clearly, from Tables 14, the Lindley method gives a better estimate than the MH method. The informative Bayesian estimation gives results that are very close to the non-informative Bayesian estimation in terms of Bais and (MSE). Bayesian estimation under the LINEX loss function gives results that are very close result to that under the SELF in term of Bias and MSE; the negative value of β provides more weight to underestimation than the overestimation while, for very (small or large) values of β the LINEX loss function is almost symmetric (see [22]). In most cases, the GA gives better results than the NLM method. Moreover, for Case 1, the elimination at the beginning gives a better estimate in terms of bias and MSE than the end. Concerning the sensitivity of the estimates for sample size, we find that the behavior of both methods (i.e., the GA and NLM)gives a better result for large sample sizes. In addition, MLE gives a better bais than Bayesian estimation. In Table 5, it's obvious that the 95% HPD credible interval of the Gibbs (MH) sampling method gives a better estimate for informative beginning elimination for Case 1. In general, the GA gives a better estimate than NLM in terms of the lower-length intervals containing the real value.

    Table 5.  95% HPD credible interval of Gibbs (MH) sampling method for Bayes estimation of GEVL distribution parameters in terms of lower (LB) and upper bounds (UB), and the length interval (IL).
    method parameters N=1000 N=50
    NON INF NON INF
    LB UB IL LB UB IL LB UB IL LB UB IL
    Case1 GA Beginning μ=0.2 0.0124 0.3024 0.29 0.0111 0.302 0.2909 0.0056 0.3216 0.3161 0.0049 0.3211 0.3163
    σ=0.3 0.0119 0.314 0.3021 0.0129 0.3139 0.301 0.005 0.2986 0.2936 0.0046 0.2988 0.2942
    ξ=0.7 0.0744 0.7118 0.6375 0.0702 0.711 0.6408 0.0464 0.7504 0.704 0.0504 0.7462 0.6958
    End μ=0.2 0.0038 0.3087 0.3049 0.0047 0.3088 0.3042 0.0045 0.3217 0.3171 0.0035 0.3224 0.3189
    σ=0.3 0.0045 0.3195 0.315 0.0038 0.3194 0.3156 0.0043 0.299 0.2947 0.0033 0.2985 0.2951
    ξ=0.7 0.0402 0.728 0.6878 0.0496 0.7344 0.6848 0.0479 0.7449 0.697 0.0489 0.7469 0.698
    NLM Beginning μ=0.2 0.0113 0.2716 0.2602 0.0101 0.2716 0.2615 0.0112 0.2771 0.2659 0.0121 0.2778 0.2657
    σ=0.3 0.016 0.3623 0.3463 0.0157 0.3627 0.347 0.0155 0.3354 0.3199 0.0149 0.3351 0.3202
    ξ=0.7 0.0321 0.5411 0.509 0.0325 0.5386 0.5062 0.0193 0.3886 0.3693 0.0199 0.3895 0.3696
    End μ=0.2 0.0089 0.2377 0.2287 0.0086 0.2374 0.2288 0.0087 0.237 0.2283 0.0096 0.2373 0.2277
    σ=0.3 0.0109 0.2898 0.2789 0.0115 0.2897 0.2783 0.0108 0.2754 0.2646 0.0101 0.2743 0.2643
    ξ=0.7 0.0216 0.4162 0.3946 0.0189 0.4147 0.3958 0.0176 0.3544 0.3368 0.0189 0.3546 0.3358
    Case2 GA μ=0.2 0.0107 0.2914 0.2807 0.0129 0.2918 0.2789 0.014 0.3142 0.3002 0.0129 0.3137 0.3008
    σ=0.3 0.0134 0.3257 0.3123 0.0141 0.3266 0.3126 0.0119 0.3023 0.2904 0.0122 0.3024 0.2902
    ξ=0.7 0.0777 0.72 0.6423 0.0788 0.7251 0.6463 0.0825 0.7369 0.6544 0.0814 0.7321 0.6507
    NLM μ=0.2 0.0107 0.2597 0.2490 0.0112 0.2599 0.2487 0.0107 0.2639 0.2531 0.011 0.2639 0.2529
    σ=0.3 0.0144 0.3498 0.3354 0.0156 0.349 0.3334 0.0142 0.3223 0.3081 0.0131 0.3228 0.3097
    ξ=0.7 0.0687 0.6901 0.6215 0.0649 0.6876 0.6227 0.036 0.5100 0.474 0.0386 0.5089 0.4702
    Case3 GA μ=0.2 0.0053 0.2914 0.2861 0.0034 0.2909 0.2875 0.0043 0.3149 0.3106 0.0051 0.3148 0.3097
    σ=0.3 0.0042 0.3352 0.331 0.0051 0.3353 0.3302 0.0038 0.3061 0.3023 0.004 0.3062 0.3022
    ξ=0.7 0.0488 0.7469 0.6981 0.0485 0.7506 0.7021 0.0536 0.7586 0.7050 0.0466 0.7577 0.7111
    NLM μ=0.2 0.0027 0.2647 0.262 0.0025 0.2647 0.2623 0.0038 0.2701 0.2663 0.0037 0.2697 0.266
    σ=0.3 0.0047 0.3587 0.354 0.0039 0.3588 0.3548 0.0029 0.3302 0.3274 0.0027 0.33 0.3272
    ξ=0.7 0.0383 0.7178 0.6795 0.0383 0.7201 0.6818 0.0203 0.5417 0.5214 0.0168 0.5407 0.5239

     | Show Table
    DownLoad: CSV

    In this section, we discuss how real lifetime data were used to illustrate the above algorithms. Extreme weather events like high temperatures are often associated with extreme weather events such as heatwaves, droughts, and wildfires. Moreover, extremely high temperatures can have serious health implications, leading to heat-related illnesses. After the recent global crisis due to the unexpected rise in temperature degree, it became necessary to understand the historical highs for each country to avoid some of the catastrophic events that occurred as a result of this increase, and to help in the assessment of the frequency, intensity, and impact of these events, enabling better preparedness and response strategies.

    So, we applied a real data set that included the highest ten daily temperatures for Queensland Australia from 19572022. It was obtained from the Australian government from the following website: http://www.bom.gov.au/cgi-bin/climate/extremes/annualextremes.cgi?period=%2Fcgibin%2Fclimate%2Fextremes%2Fannualextremes.cgi&climtab=tmaxhigh&area=qld&year=1957. This data set was fitted to the GEVL distribution by using the Kolmogorov-Smirnov (K-S) goodness of fit test at the level of significance of α=0.01. The K-S value was 0.0510 which is less than the tabulated value of 0.0681. Also, we constructed the Q-Q plot and returned the level for this data set, as in Figures 2 and 3. Figures 2, 3 and the K-S value verify that this data set isfit the GEVL distribution. In Table 6 some basic statistics are presented for this data set including the, mean, variance, median, and others.

    Figure 2.  Q-Q plots regarding the fitting of the GEVL distribution for Queensland Australia's extreme temperature data.
    Figure 3.  Returned level and empirical quantiles of the GEVL distribution for the Queensland Australia extreme temperaturedata.
    Table 6.  The basic statistics of the extreme temperature for Queensland (Australia) data sets.
    Country Mean Median VAR Standard deviation Minimum Maximum Range Quartiles (25%,50%,75%)
    Queensland(Australia) 45.3877 45.2 1.5019 1.2256 43 49.5 6.5 (44.5, 45.2, 46.1)

     | Show Table
    DownLoad: CSV

    The real data serve as a practical example of the simulation technique, and we have used the methodology previously described in the simulation section in the process of estimation of the parameters and selection of the hyperparameters. The maximum likelihood estimate and Bayesian estimates considered in Sections 3 and 4 for the GEVL distribution parameters were obtained as in Tables 7 and 8 for elimination for the three cases considered above by using the two optimization methods considered i.e., (NLM and GA) for MLE and Bayesian estimation. Bayesian estimation is considered for informative and non-informative prior for both the Lindley and MH methods under the SELF (sq) and LINEX loss function (lx(β=A) and A0.5,5,5).

    Table 7.  The MLE, Bayesian estimation, and 95% HPD credible interval of Gibbs (MH) sampling method for Bayesian estimation of GEVL distribution parameters for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(0.5,5,5)) for Queensland (Australia) data set For Case 1.
    GA NLM
    Beginning End Beginning Bnd
    μ σ ξ μ σ ξ μ σ ξ μ σ ξ
    MLE 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.949 1.1855 0.1186 44.6992 1.0072 0.1973
    NON Lindley Sq 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9512 1.1927 0.1206 44.7003 1.0094 0.1975
    lx(β=0.5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9506 1.1886 0.1203 44.6998 1.008 0.1973
    lx(β=5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9571 1.1974 0.1234 44.7051 1.0119 0.1989
    lx(β=5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9451 1.1877 0.1178 44.6955 1.0067 0.196
    MH Sq 44.6311 1.0327 0.1249 44.5174 0.4654 0.1404 44.6094 0.7503 0.077 44.3944 0.5559 0.1084
    lx(β=0.5) 44.6136 1.0259 0.1236 44.4957 0.4455 0.1391 44.5921 0.7149 0.0764 44.3887 0.5456 0.1071
    lx(β=5) 44.7616 1.0916 0.1375 44.6843 0.6845 0.1535 44.7345 0.9839 0.0829 44.4549 0.6784 0.122
    lx(β=5) 44.4584 0.964 0.1126 44.3081 0.3014 0.127 44.4039 0.4573 0.0714 44.3437 0.4741 0.0959
    UB 44.0901 0.6414 0.0015 43.8473 0.0035 0.0017 43.8255 0.0511 0.001 44.1044 0.1881 0.0011
    LB 44.9997 1.267 0.2582 44.9734 1.1611 0.2612 44.9956 1.2414 0.1836 44.7574 1.0552 0.2557
    LI 0.9095 0.6256 0.2567 1.1261 1.1576 0.2595 1.1701 1.1902 0.1826 0.653 0.8671 0.2546
    INF Lindley Sq 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9526 1.1823 0.1094 44.7016 1.0052 0.1938
    lx(β=0.5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.952 1.1818 0.1091 44.7011 1.0049 0.1937
    lx(β=5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9585 1.1873 0.1121 44.7063 1.0078 0.1952
    lx(β=5) 44.9414 1.2063 0.2059 44.9207 1.2081 0.2047 44.9465 1.1774 0.1069 44.6967 1.0026 0.1924
    MH Sq 44.3906 0.906 0.1641 44.0854 0.5644 0.1363 44.4551 0.898 0.0918 44.4891 0.6405 0.2666
    lx(β=0.5) 44.3717 0.8866 0.1625 44.0418 0.5541 0.1348 44.4406 0.8877 0.0911 44.4845 0.6078 0.2664
    lx(β=5) 44.5951 1.0206 0.1789 44.5823 0.6466 0.1506 44.6093 0.9814 0.0983 44.5306 0.8291 0.2689
    lx(β=5) 44.3717 0.8866 0.1625 44.0418 0.5541 0.1348 44.4406 0.8877 0.0911 44.4845 0.6078 0.2664
    UB 43.9698 0.2064 0.0021 43.4954 0.0544 0.0014 44.0197 0.4063 0.0014 44.1315 0.004 0.0652
    LB 45.0057 1.2466 0.2751 44.9853 0.9846 0.2686 44.9746 1.2358 0.1795 44.7534 1.0583 0.2752
    LI 1.0358 1.0402 0.273 1.4899 0.9302 0.2673 0.9549 0.8295 0.1782 0.6219 1.0543 0.21

     | Show Table
    DownLoad: CSV
    Table 8.  The MLE, Bayesian estimation and 95% HPD credible interval of Gibbs (MH) sampling method of Bayes estimation of GEVL distribution parameters for informative (INF) and non-informative priors (NON) under SELF (sq) LINEX loss function (lx) at (β=(0.5,5,5)) for Queensland (Australia) data sets for Cases 2 and 3.
    GA NLM
    Case2 Case3 Case2 Case3
    μ σ ξ μ σ ξ μ σ ξ μ σ ξ
    MLE 44.8468 1.09 0.2343 44.8081 1.1847 0.2417 44.9558 1.2257 0.1833 44.7978 1.2539 0.3541
    NON Lindley Sq 44.8488 1.0925 0.2342 44.8096 1.1882 0.2421 44.9588 1.2319 0.1845 44.7758 1.2564 0.3593
    lx(β=0.5) 44.8483 1.0909 0.2341 44.8089 1.1861 0.2419 44.9581 1.2282 0.1842 44.7743 1.255 0.3592
    lx(β=5) 44.8542 1.0951 0.2353 44.8162 1.1923 0.2443 44.9652 1.2374 0.1866 44.7925 1.258 0.36
    lx(β=5) 44.8434 1.0897 0.2332 44.8029 1.184 0.2399 44.9522 1.2261 0.1823 44.7624 1.2548 0.3585
    MH Sq 43.9561 0.7164 0.1517 44.2639 0.6625 0.1502 43.7289 1.2975 0.1221 44.7086 0.9467 0.1817
    lx(β=0.5) 43.933 0.7043 0.1497 44.2314 0.6441 0.1484 43.7015 1.2971 0.121 44.7057 0.9182 0.178
    lx(β=5) 44.184 0.8143 0.1721 44.4741 0.8319 0.1676 43.9399 1.3008 0.1325 44.7344 1.1113 0.2194
    lx(β=5) 43.7592 0.5964 0.1319 43.8342 0.4986 0.1331 43.4916 1.2905 0.1117 44.6783 0.6673 0.1477
    UB 43.2867 0.2697 0.0016 43.1805 0.0929 0.0016 44.8227 1.1008 0.1069 44.4412 0.2674 0.0021
    LB 44.5614 1.1159 0.3157 44.8466 1.2382 0.3089 43.1062 1.0118 0.002 44.8626 1.3128 0.4061
    LI 1.2747 0.8462 0.3142 1.6661 1.1454 0.3073 44.3877 1.3087 0.2471 0.4214 1.0454 0.404
    INF Lindley Sq 44.8679 1.0524 0.1981 44.8086 1.1858 0.2409 1.2815 0.297 0.2451 44.7756 1.2559 0.3591
    lx(β=0.5) 44.8674 1.0524 0.1983 44.808 1.1854 0.2406 44.9584 1.2245 0.1789 44.774 1.2557 0.3591
    lx(β=5) 44.8717 1.0517 0.1956 44.8153 1.1899 0.2431 44.9577 1.2239 0.1787 44.7923 1.2575 0.3599
    lx(β=5) 44.863 1.0532 0.2001 44.802 1.1817 0.2387 44.9648 1.2302 0.1811 44.7622 1.2543 0.3583
    MH Sq 44.7346 0.8012 0.1245 44.3627 0.8073 0.1525 44.9518 1.2188 0.1769 44.4404 1.0759 0.2183
    lx(β=0.5) 44.7315 0.7929 0.1226 44.3425 0.7947 0.1507 44.2519 0.4094 0.1331 44.4297 1.0679 0.2143
    lx(β=5) 44.7618 0.8802 0.1435 44.5391 0.9331 0.1699 44.1865 0.3988 0.1316 44.5452 1.1419 0.2545
    lx(β=5) 44.7315 0.7929 0.1226 44.3425 0.7947 0.1507 44.6945 0.5326 0.1477 44.4297 1.0679 0.2143
    UB 44.4411 0.3583 0.001 43.8676 0.3526 0.002 44.1865 0.3988 0.1316 43.999 0.6778 0.0016
    LB 44.9045 1.1519 0.2882 44.8653 1.2453 0.305 43.4544 0.0192 0.0013 44.8383 1.3196 0.4134
    LB 44.9045 1.1519 0.2882 44.8653 1.2453 0.305 43.4544 0.0192 0.0013 44.8383 1.3196 0.4134
    LI 0.4634 0.7936 0.2871 0.9978 0.8927 0.303 45.014 0.9599 0.2505 0.8394 0.6417 0.4118

     | Show Table
    DownLoad: CSV

    In this paper, parameter estimation for GEVL distributions was derived under a type-Ⅱ progressive censored scheme by using optimization methods (NLM); also, artificial intelligence (GA) for MLE parameters, and the Bayesian technique was studied. The Bayesian estimation of parameters has been considered for two loss functions (symmetric and asymmetric loss functions) for informative and non-informative priors. Furthermore, progressive censoring is considered for three cases of elimination (i.e., fixed, discrete uniform, and binomial), with random removal under 10% elimination from the sample size. For a fixed random removal censoring scheme two cases were considered i.e., beginning and end. A Monte Carlo simulation was performed to compare the performance of the different estimators of GEVL distribution parameters; the results revealed that the Lindley method outperforms the MH method in terms of estimation quality. Informative Bayesian estimation yielded results that are very close to those of non-informative Bayesian estimation in terms of the bias and MSE. When it comes to Bayesian estimation under the LINEX loss function, the results closely resembled those under the SELF, with the negative value of β favoring underestimation over overestimation. However, for very small or large values of β, the LINEX loss function exhibits nearly symmetrical behavior (refer to [22] for details). Moreover, in most cases, the GA method outperformed the NLM method. Furthermore, for Case 1, the initial elimination step provided more accurate estimates in terms of bias and MSE than the final elimination step. Regarding the sensitivity of the estimates to sample size, both the GA and NLM methods performed better with larger sample sizes. Additionally, MLE demonstrated superior bias compared to Bayesian estimation. The 95% HPD credible interval obtained via the Gibbs (MH) sampling method provided more accurate estimates with the informative initial elimination for Case 1. In general, the GA method yields more precise estimates than the NLM method, as indicated by narrower intervals containing the true values.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The authors extend their appreciation to Taif University, Saudi Arabia, for supporting this work through project number (TU-DSPP-2024-162).

    The authors state no conflict of interest.



    [1] M. Gilli, E. këllezi, An application of extreme value theory for measuring financial risk, Comput. Econ., 27 (2006), 207–228. https://doi.org/10.1007/s10614-006-9025-7 doi: 10.1007/s10614-006-9025-7
    [2] R. A. Fisher, L. H. C. Tippett, Limiting forms of the frequency distribution of the largest or smallest member of a sample, Math. Proc. Cambridge, 24 (1928), 180–190. https://doi.org/10.1017/S0305004100015681 doi: 10.1017/S0305004100015681
    [3] T. G. Bali, The generalized extreme value distribution, Econ. Lett., 79 (2003), 423–427. https://doi.org/10.1016/S0165-1765(03)00035-1 doi: 10.1016/S0165-1765(03)00035-1
    [4] J. Hosking, J. Wallis, E. Wood, Estimation of the generalized extreme-value distribution by the method of probability-weighted moments, Technometrics, 27 (1985), 251–261. https://doi.org/10.2307/1269706 doi: 10.2307/1269706
    [5] E. Bertin, M. Clusel, Generalized extreme value statistics and sum of correlated variables, J. Phys. A Math. Gen., 39 (2006), 7607. https://doi.org/10.1088/0305-4470/39/24/001 doi: 10.1088/0305-4470/39/24/001
    [6] S. Zhou, A. Xu, Y. Tang, L. Shen, Fast Bayesian inference of reparameterized gamma process with random effects, IEEE T. Reliab., 73 (2024), 399–412. https://doi.org/10.1109/TR.2023.3263940 doi: 10.1109/TR.2023.3263940
    [7] L. Zhuang, A. Xu, X. Wang, A prognostic driven predictive maintenance framework based on Bayesian deep learning, Reliab. Eng. Syst. Safe., 234 (2023), 109181. https://doi.org/10.1016/j.ress.2023.109181 doi: 10.1016/j.ress.2023.109181
    [8] W. Wang, Z. Cui, R. Chen, Y. Wang, X. Zhao, Regression analysis of clustered panel count data with additive mean models, Stat. Papers, 2023. https: /doi.org/10.1007/s00362-023-01511-3
    [9] S. Phoong, M. Ismail, A comparison between Bayesian and maximum likelihood estimations in estimating finite mixture model for financial data, Sains Malays., 44 (2015), 1033–1039. https://doi.org/10.17576/jsm-2015-4407-16 doi: 10.17576/jsm-2015-4407-16
    [10] S. Wang, W. Chen, M. Chen, Y. Zhou, Maximum likelihood estimation of the parameters of the inverse Gaussian distribution using maximum rank set sampling with unequal samples, Math. Popul. Stud., 30 (2023), 1–21. https://doi.org/10.1080/08898480.2021.1996822 doi: 10.1080/08898480.2021.1996822
    [11] S. Coles, J. Bawa, L. Trenner, P. Dorazio, An introduction to statistical modeling of extreme values, London: Springer, 2001.
    [12] H. Barakat, O. Khaled, E. Nigm, Statistical techniques for modeling extreme value data and related applications, Cambridge Scholars Publishing, 2019.
    [13] J. Dennis Jr, R. Schnabel, Numerical methods for unconstrained optimization and nonlinear equations, Society for Industrial and Applied Mathematics, 1996.
    [14] L. Haldurai, T. Madhubala, R. Rajalakshmi, A study on Genetic algorithm and its applications, Int. J. Comput. Sci. Eng., 4 (2016), 2347–2693.
    [15] L. Scrucca, GA: A package for genetic algorithm in R, J. Stat. Softw., 53 (2013), 1–37. https://doi.org/10.18637/jss.v053.i04 doi: 10.18637/jss.v053.i04
    [16] N. Balakrishnan, R. Aggarwala, Progressive censoring: theory, methods, and applications, Springer Science & Business Media, 2000. https://doi.org/10.1007/978-1-4612-1334-5
    [17] nlm: Non-linear minimization, stats (version 3.6.2), 2019. Available from: https://www.rdocumentation.org/packages/stats/versions/3.6.2/topics/nlm.
    [18] J. F. Lawless, Statistical models and methods for lifetime data, John Wiley & Sons, 2003. https://doi.org/10.1002/9781118033005
    [19] N. A. Mokhlis, E. J. Ibrahim, D. M. Gharieb, Stress-strength reliability with general form distributions, Commun. Stati. Theor. M., 46 (2017), 1230–1246. https://doi.org/10.1080/03610926.2015.1014110 doi: 10.1080/03610926.2015.1014110
    [20] N. A. Mokhlis, S. K. Khames, Estimation of stress-strength reliability for Marshall-Olkin extended Weibull family based on type-Ⅱ progressive censoring, J. Stat. Appl. Probab., 10 (2021), 385–396. https://doi.org/10.18576/jsap/100210 doi: 10.18576/jsap/100210
    [21] S. Ahn, C. Park, H. Kim, Hazard rate estimation of a mixture model with censored lifetimes, Stoch. Environ. Rese. Risk Assess., 21 (2007), 711–716. https://doi.org/10.1007/s00477-006-0082-1 doi: 10.1007/s00477-006-0082-1
    [22] N. Khatun, M. Matin, A study on LINEX loss function with different estimating methods, Open J. Stat., 10 (2020), 52–63. https://doi.org/10.4236/ojs.2020.101004 doi: 10.4236/ojs.2020.101004
    [23] D. V. Lindley, Approximate Bayesian methods, Trabajos de Estadistica Y de Investigacion Operativa, 31 (1980), 223–245. https://doi.org/10.1007/BF02888353 doi: 10.1007/BF02888353
    [24] C. Wang, M. H. Chen, E. Schifano, J. Wu, J. Yan, Statistical methods and computing for big data, Stat. interface, 9 (2016), 399–414. https://doi.org/10.4310/SII.2016.v9.n4.a1 doi: 10.4310/SII.2016.v9.n4.a1
  • This article has been cited by:

    1. Rasha Abd El-Wahab Attwa, Shimaa Wasfy Sadk, Taha Radwan, Estimation of Marshall–Olkin Extended Generalized Extreme Value Distribution Parameters under Progressive Type-II Censoring by Using a Genetic Algorithm, 2024, 16, 2073-8994, 669, 10.3390/sym16060669
    2. Rasha Abd El-Wahab Attwa, Shimaa Wasfy Sadk, Hassan M. Aljohani, Hilary Izuchukwu Okagbue, Estimating q−GEVL distribution parameters under Type II progressive censoring using particle swarm optimization, 2025, 20, 1932-6203, e0323897, 10.1371/journal.pone.0323897
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1100) PDF downloads(38) Cited by(2)

Figures and Tables

Figures(3)  /  Tables(8)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog