Loading [MathJax]/extensions/TeX/mathchoice.js
Research article

On residual cumulative generalized exponential entropy and its application in human health

  • Received: 05 February 2025 Revised: 05 March 2025 Accepted: 12 March 2025 Published: 20 March 2025
  • Numerous adaptations of traditional entropy concepts and their residual counterparts have emerged in statistical research. While some methodologies incorporate supplementary variables or reshape foundational assumptions, many ultimately align with conventional formulations. This study introduces a novel extension termed residual cumulative generalized exponential entropy to broaden the scope of residual cumulative entropy for continuous distributions. Key attributes of the proposed measure include non-negativity, bounds, its relationship to the continuous entropy measure, and stochastic comparisons. Practical implementations are demonstrated through case studies involving established probability models. Additionally, insights into order statistics are derived to characterize the measure's theoretical underpinnings. The residual cumulative generalized exponential entropy framework bridges concepts such as Bayesian risk assessment and excess wealth ordering. For empirical implementation, non-parametric estimation strategies are devised using data-driven approximations of residual cumulative generalized exponential entropy, with two distinct estimators of the cumulative distribution function evaluated. A practical application is showcased, using clinical diabetes data. The study further explores the role of generalized exponential entropy in identifying distributional symmetry, mainly through its application to uniform distributions to pinpoint symmetry thresholds in ordered data. Finally, the utility of generalized exponential entropy is examined in pattern analysis, with a diabetes dataset serving as a benchmark for evaluating its classification performance.

    Citation: Hanan H. Sakr, Mohamed S. Mohamed. On residual cumulative generalized exponential entropy and its application in human health[J]. Electronic Research Archive, 2025, 33(3): 1633-1666. doi: 10.3934/era.2025077

    Related Papers:

    [1] Sadia Anwar, Showkat Ahmad Lone, Aysha Khan, Salmeh Almutlak . Stress-strength reliability estimation for the inverted exponentiated Rayleigh distribution under unified progressive hybrid censoring with application. Electronic Research Archive, 2023, 31(7): 4011-4033. doi: 10.3934/era.2023204
    [2] Mahmoud El-Morshedy, Mohamed S. Eliwa, Mohamed El-Dawoody, Hend S. Shahen . A weighted hybrid discrete probability model: Mathematical framework, statistical analysis, estimation techniques, simulation-based ranking, and goodness-of-fit evaluation for over-dispersed data. Electronic Research Archive, 2025, 33(4): 2061-2091. doi: 10.3934/era.2025091
    [3] Chuanyao Li, Yichao Lu, Yuqiang Wang, Gege Jiang . Congestion behavior and tolling strategies in a bottleneck model with exponential scheduling preference. Electronic Research Archive, 2023, 31(2): 1065-1088. doi: 10.3934/era.2023053
    [4] Natália Bebiano, João da Providência, Wei-Ru Xu . Approximations for the von Neumann and Rényi entropies of graphs with circulant type Laplacians. Electronic Research Archive, 2022, 30(5): 1864-1880. doi: 10.3934/era.2022094
    [5] Dewang Li, Meilan Qiu, Jianming Jiang, Shuiping Yang . The application of an optimized fractional order accumulated grey model with variable parameters in the total energy consumption of Jiangsu Province and the consumption level of Chinese residents. Electronic Research Archive, 2022, 30(3): 798-812. doi: 10.3934/era.2022042
    [6] Yu Chen, Qingyang Meng, Zhibo Liu, Zhuanzhe Zhao, Yongming Liu, Zhijian Tu, Haoran Zhu . Research on filtering method of rolling bearing vibration signal based on improved Morlet wavelet. Electronic Research Archive, 2024, 32(1): 241-262. doi: 10.3934/era.2024012
    [7] Eduardo Paluzo-Hidalgo, Rocio Gonzalez-Diaz, Guillermo Aguirre-Carrazana . Emotion recognition in talking-face videos using persistent entropy and neural networks. Electronic Research Archive, 2022, 30(2): 644-660. doi: 10.3934/era.2022034
    [8] E. A. Abdel-Rehim . The time evolution of the large exponential and power population growth and their relation to the discrete linear birth-death process. Electronic Research Archive, 2022, 30(7): 2487-2509. doi: 10.3934/era.2022127
    [9] Caiwen Chen, Tianxiu Lu, Ping Gao . Chaotic performance and circuitry implement of piecewise logistic-like mapping. Electronic Research Archive, 2025, 33(1): 102-120. doi: 10.3934/era.2025006
    [10] Wangwei Zhang, Hao Sun, Bin Zhou . TBRAFusion: Infrared and visible image fusion based on two-branch residual attention Transformer. Electronic Research Archive, 2025, 33(1): 158-180. doi: 10.3934/era.2025009
  • Numerous adaptations of traditional entropy concepts and their residual counterparts have emerged in statistical research. While some methodologies incorporate supplementary variables or reshape foundational assumptions, many ultimately align with conventional formulations. This study introduces a novel extension termed residual cumulative generalized exponential entropy to broaden the scope of residual cumulative entropy for continuous distributions. Key attributes of the proposed measure include non-negativity, bounds, its relationship to the continuous entropy measure, and stochastic comparisons. Practical implementations are demonstrated through case studies involving established probability models. Additionally, insights into order statistics are derived to characterize the measure's theoretical underpinnings. The residual cumulative generalized exponential entropy framework bridges concepts such as Bayesian risk assessment and excess wealth ordering. For empirical implementation, non-parametric estimation strategies are devised using data-driven approximations of residual cumulative generalized exponential entropy, with two distinct estimators of the cumulative distribution function evaluated. A practical application is showcased, using clinical diabetes data. The study further explores the role of generalized exponential entropy in identifying distributional symmetry, mainly through its application to uniform distributions to pinpoint symmetry thresholds in ordered data. Finally, the utility of generalized exponential entropy is examined in pattern analysis, with a diabetes dataset serving as a benchmark for evaluating its classification performance.



    Shannon [1] first proposed the idea of entropy in 1948. It is a fundamental idea in the theories of information that measures the degree of uncertainty or knowledge contained in a given random variable. Shannon entropy is acquainted mathematically as

    Sn(P)=nk=1pklnpk,

    for discrete random variables, with noting that P=(p1,p2,...,pn) is the vector of mass function probability, or

    Sn(X)=f(x)lnf(x)dx, (1.1)

    for a randomly generated continuous variable X, and f(x) represents the probability density functional (PDF). It has proven to be a powerful tool in various fields, including communication systems, where it optimizes data transmission, and in data compression, where it measures the limits of compressibility. By capturing the average amount of information generated by random events, Shannon entropy provides a robust framework for analyzing uncertainty in diverse systems.

    Building on this seminal work, Rao et al. [2] established the conception of the cumulative residual entropy model as an extension of entropy designed for survival analysis and reliability theory. Unlike traditional entropy, which measures overall uncertainty, residual cumulative entropy focuses on the uncertainty remaining in a system or process beyond a given time or threshold. Formally, given a randomly variable that is not negative X, the residual cumulative function of entropy measurement is described as:

    RCEn(X)=0ˉF(x)lnˉF(x)dx, (1.2)

    in which the survival function is encapsulated by ˉF(x)=1F(x), and the cumulative distribution function (CDF) of X is encapsulated by F(x). This measure captures the tail behavior of the determined distribution, causing it to be quite helpful in applications requiring an understanding of the uncertainty associated with extreme events. Residual cumulative entropy has since been employed in numerous disciplines regarding reliability engineering, risk analysis form, and lifetime data analysis, where the quantification of remaining uncertainty is of critical importance.

    Following Campbell [3], Pal and Pal [4,5] used these concerns to establish a new measure, called the exponential entropy measure, through other descriptions parallel to Shannon entropy. In the discrete situation, the formulation of the exponential entropy model is described below:

    EXn(P)=nk=1Pk(e1Pk1). (1.3)

    They added the 1 term since it seems only logical that any measurement that contains data should be assigned 0 for the degenerative distribution of probabilities (0,...,0,1,0,...,0). The authors argued that exponential entropy offers distinct benefits over Shannon's formulation. For example, they noted that exponential entropy reaches a fixed upper value in the case of a distribution that is considered to be uniform, where each probability is given by pk=1n for k=1,2,,n as demonstrated by

    limnEXn(1n,1n,,1n)=e1;

    in contrast, Shannon's entropy does not exhibit this bounded behavior.

    Furthermore, Panjehkeh et al. [6] examined the features and attributes of the exponential measure of entropy, along with the Shannon measure of entropy model, in both discrete and continuous situations, including the asymptotic equipartition property, invariancy under monotone transformation, subadditivity, and the chain rule. The concept of a continuous exponential measure of entropy was first presented as

    EXn(X)=Dsf(x)(e1f(x))dx. (1.4)

    We may observe that Panjehkeh et al. [6] eliminated the term 1 from the measurement in order to make it inherently non-negative against Shannon entropy, given that X is a randomly variable that is continuous, following the PDF f, and the support Ds. Kvalseth [7] developed a more broad variant of the exponential entropy of Eq (1.2), known as generalized entropy, given by

    GEXn(P)=1βnk=1pk(e1pβk1), (1.5)

    where the parameter β is arbitrary and has a non-zero real value (i.e., βR{0}). Moreover, the generalized exponential entropy measure has been used in many topics. Alotaibi and Elaraby [8] created a generalized exponential entropy-based method for COVID-19 disease segmentation from computed tomography scans. Fuzzy c partitioning and generalized exponential entropy were combined in their suggested method. In multi-criteria decision-making, we can see, for example, Wei et al. [9] and Dinesh and Kumar [10].

    Work motivation: Since the PDF is calculated as the distribution's derivative form, the CDF appears more regularly than the density function. Furthermore, the CDF is what is relevant and/or quantifiable in practice. For instance, if the variable that is chosen randomly is the life span of a machine, the happenstance that is relevant is not whether the life span equals t, but rather whether it surpasses t. This is what prompted us to study it instead of the existing entropy based on the PDF. The question that arises is whether it can be generalized and if the traditional model can be made a special case of it. Moreover, studying the discrete case is not satisfactory without considering the continuous case. As we have mentioned that the exponential entropy in discrete case has been discussed in distinct fields; take into consider, for instance, Wei et al. [11], and Ye and Cui [12]. Therefore, it is essential to study and discuss it in the continuous case with clarification of the applications associated with the continuous side. Another important aspect to discuss is that the proposed model is based on the exponential function. Dose this function help solve some of the problems that appear in uncertainty issues?

    This article aimed to present a generalization of the continuous residual cumulative entropy, known as residual cumulative generalized exponential entropy. Applications, including non-parametric estimation, are provided. On the other hand, the continuous form of the generalized exponential entropy is used to discuss the characterization of symmetry via order statistics. Moreover, a classification problem involving the pattern recognition in diabetes data based on the generalized exponential entropy model is presented. The following is the arrangement of the remaining portions of this paper: in Section 2, the concept of residual cumulative generalized exponential entropy in the continuous setting is presented. Additionally, the properties, including bounds, non-negativity, the relationship with differential entropy, and stochastic orders, are explained. In Section 3, some consequences of the residual cumulative generalized exponential entropy expansion, such as the Bayes risk and the order of excess wealth, are discussed. In Section 4, the non-parametric estimation of the empirical residual cumulative generalized exponential entropy is applied using two methods. Finally, in Section 5, the symmetry characterization using order statistics with an example based on the symmetrical uniform distribution and the classification problem using pattern recognition based on the generalized exponential entropy are illustrated.

    In this section, we will establish the concept of the residual cumulative generalized exponential entropy. Inspired by Rao et al. [2], we can depend on the function of survival ˉF(x) to derive the residual cumulative generalized exponential entropy from the discrete case of the generalized exponential entropy in (1.5) according to the following definition.

    Definition 2.1. Consider the non-negativity continual randomly variable X following the CDF F. Then, we can realize the residual cumulative generalized exponential entropy by the following formula:

    RGEXnβ(X)=1β0ˉF(x)(e1ˉFβ(x)1)dx=1β[0ˉF(x)e1ˉFβ(x)dxμ], (2.1)

    where βR{0}, and the mean (expected value) μ=E(X)=0ˉF(x)dx.

    The following proposition shows the limitation of the residual cumulative generalized exponential entropy when β tends to zero, which returns to residual cumulative entropy in (1.2).

    Proposition 2.1. Consider the non-negativity continual random variable X following the CDF F. Then, from (2.1) and (1.2), we have

    limβ0RGEXnβ(X)=RCEn(X).

    Proof. From (2.1), utilizing the L'Hopital's rule, we have

    limβ0RGEXnβ(X)=limβ00ˉF(x)(e1ˉFβ(x)1)dxβ=limβ00ˉF1+β(x)e1ˉFβ(x)lnˉF(x)dx=0ˉF(x)lnˉF(x)dx=RCEn(X).

    In the following discussions, we will discuss the non-negativity and the bounds of the residual cumulative generalized exponential entropy when β>0.

    Proposition 2.2. Consider the non-negativity continual random variable X following the CDF F. Then, from (2.1), we can say that

    (1) The residual cumulative generalized exponential entropy is non-negative for all β>0.

    (2) To discuss the bounds of the residual cumulative generalized exponential entropy:

    (a) We get

    RGEXnβ(X)()RGEXn1(X),0<β1(β1). (2.2)

    (b) We get

    0RGEXnβ(X)μ(e1)β,β0. (2.3)

    Proof. (1) Its known that e1ˉFβ(x)1, for all β>0, and ˉF(x)[0,1]. Thus, ˉF(x)e1ˉFβ(x)ˉF(x), which implies that 0ˉF(x)e1ˉFβ(x)dx0ˉF(x)dx=μ. Then, RGEXnβ(X)0, for all β>0. Or, by another method, we can assume, by converse, that βRGEXnβ(X)<0, then we have

    0ˉF(x)e1ˉFβ(x)dx0ˉF(x)dx<00ˉF(x)e1ˉFβ(x)dx<0ˉF(x)e0dx.

    Therefore, we deduce that e1ˉFβ(x)<1, which implies 1ˉFβ(x)<0 or, equivalently, ˉFβ(x)>1. This contradicts the fact that ˉF(x)[0,1] for any x. Then, the result follows.

    (2) For 0<β1(β1), we have ˉFβ(x)()ˉF(x), which implies that ˉF(x)e1ˉFβ(x)()ˉF(x)e1ˉF(x). Then, 1β[0ˉF(x)e1ˉFβ(x)dxμ]()0ˉF(x)e1ˉF(x)dxμ, and the result follows.

    Lemma 2.1. If μ=E(X)<, then RGEXnβ(X)<, for all β>0.

    Proof. The result is obtained directly from (2.3).

    Theorem 2.1. If X is an absolutely continual non-negativity random variable following a PDF f(x), then

    RGEXn1(X)CeSn(X),

    with noting that

    C=exp{10ln|u(e1u1)|du}0.176192,

    and Sn(X) is defined in (1.1).

    Proof. Using the fact of the inequality of log-sum, it contends that

    0f(x)lnf(x)|¯F(x)(e1¯F(x)1)|dxln10|¯F(x)(e1¯F(x)1)|dxdx=lnRGEXn(X), (2.4)

    Moreover, the left-hand side of (2.4) can be expressed as

    0f(x)lnf(x)|¯F(x)(e1¯F(x)1)|dx=Sn(X)0f(x)ln|¯F(x)(e1¯F(x)1)|dx=Sn(X)10ln|u(e1u1)|du.

    Therefore, it follows that

    Sn(X)+10ln|u(e1u1)|dulnRGEXn(X).

    Applying the exponential function to both sides of the aforementioned relation, we derive

    RGEXn(X)CeSn(X),

    where

    C=exp{10ln|u(e1u1)|du}0.176192,

    thereby finalizing the proof.

    Proposition 2.3. If X is an absolutely continual non-negativity random variable with residual cumulative generalized exponential entropy RGEXnβ(X)< as given in (2.1), β>0. Then, we can obtain

    RGEXnβ(X)=1βE(ψβ(X)), (2.5)

    where

    ψβ(X)=x0(e1ˉFβ(t)1)dt. (2.6)

    Proof. From (2.1), and utilizing the theorem of Fubini, we can express the following:

    RGEXnβ(X)=1β0[tf(x)dx](e1ˉFβ(t)1)dt=1β0f(x)[x0(e1ˉFβ(t)1)dt]dx,

    and (2.5) is obtained by utilizing (2.6).

    In the following, we will show some examples of the residual cumulative generalized exponential entropy of well-known distributions.

    Example 2.1. Consider the non-negativity continual random variable X following the CDF F. Then,

    (1) Under the distribution of exponential (Exp(γ)) with F(x)=1eγx, we get

    RGEXnβ(X)=(βeΓ[1β]+eΓ[1β,1])γβ2,

    where Γ[1β] is the function of the Euler gamma, and Γ[1β,1] is the function of the incomplete gamma. Moreover, with β=1, we get

    RGEXn1(X)=2eγ.

    (2) Under the uniform distribution (Ud(t1,t2)) with F(x)=xt1t2t1, and β=1, we get

    RGEXn1(X)=e52.

    Figure 1 shows the plot of the residual cumulative generalized exponential entropy and the residual cumulative entropy defined in (1.2) of the Exp(γ) distribution. Therefore, we can see the assurance of Proposition 2.1 and Eq (2.2).

    Figure 1.  Plot of RGEXnβ(X), RGEXn1(X), and RCEn(X) of the Exp(2) distribution (left) and the Exp(5) distribution (right).

    The theorem outlined below establishes a characterization in terms of the residual cumulative generalized exponential entropy under the usual stochastic, dispersive, increasing convex, and hazard rate orders. In the usual stochastic, dispersive, increasing convex, and hazard rate orders, represented, respectively, by X1UstOrX2, X1DisOrX2, X1IcxOrX2, and X1HrOrX2, we mainly remember that the random variable X1 is smaller than X2 if

    (1)ˉF1(x)ˉF2(x) (for the usual stochastic order),

    (2)F12(v)F11(v) is increasing in v(0,1) (for the dispersive order),

    (3)E(ξ(X1))=E(ξ(X1)), with the existence of the expectations and for all convex increasing functions ξ (for the increasing convex order),

    (4)ˉF2(x)ˉF1(x) is increasing with respect to x (for the hazard rate order).

    where F11 and F12 are the right continually inverses of the CDF's F1 and F2, correspondingly; see Shaked and Shanthikumar [13].

    Theorem 2.2. Assume that two random variables that are continuously, X1 and X2, following distribution functions that are strictly increasing, (F1 and F2, respectively). If X1UstOrX2, then, we obtain:

    (1) RGEXnβ(X1)RGEXnβ(X2).

    (2) RGEXnβ(X1)RGEXnβ(X2)1β(μ1μ2), where μ1=0ˉF1(x)dx, and μ2=0ˉF2(x)dx.

    Proof. 1) Let ˉF1(x)ˉF2(x), and since e1ˉFβ(x)1, for all β>0. Then, the result follows.

    (2) Since ˉF1(x)ˉF2(x), then ˉFβ1(x)ˉFβ2(x), for all β>0, or equivalently e1ˉFβ1(x)e1ˉFβ2(x). Therefore, we have

    RGEXnβ(X1)RGEXnβ(X2)=1β0[ˉF1(x)(e1ˉFβ1(x)1)ˉF2(x)(e1ˉFβ2(x)1)]dx1β(0(ˉF1(x)ˉF2(x))dx)=1β(μ1μ2).

    Example 2.2. Suppose that two random variables that are continuously, X1 and X2, following Ud(0,t1) and Ud(0,t2) distributions with the CDFs F1(x)=xt1, 0xt1, and F2(x)=xt2, 0xt2, respectively. Moreover, the means are μ1=t12 and μ2=t22. If we let t2t1, then we have X1UstOrX2. With β=1, we obtain

    RGEXn1(X1)=t1(e52)t2(e52)=RGEXn1(X2),

    and

    RGEXn1(X1)RGEXn1(X2)=(e52)(t1t2)12(t1t2)=μ1μ2,

    where (t1t2)0. Which assures the results in Theorem 2.2.

    Theorem 2.3. Assume that two random variables that are continuously, X1 and X2, following distribution functions that are strictly increasing, (F1 and F2, respectively). If X1DisOrX2, and

    RGEXnβ(X1)=RGEXnβ(X2),

    for a fixed β>0. Consequently, up to a location parameter, the distributions of X1 and X2 are identical.

    Proof. Suppose that X1DisOrX2 (i.e., the function F12(v)F11(v) is decreasing in v), and RGEXnβ(X1)=RGEXnβ(X2). Then, by a change of variable v=F(x), we observe that (according to the equality given)

    RGEXnβ(X2)RGEXnβ(X1)=1β10((1v)(e1(1v)β1))d[F12(v)F11(v)]=0,

    for a fixed β>0. Since X1DisOrX2, we are aware of that F12(v)F11(v) is a function considered to be decreasing of v. We now assert that for every 0v1, F12(v)F11(v)=c (constant). Suppose, to the contrary, that there exists a subinterval (θ1,θ2) within [0,1] where F12(v)F11(v) fails to remain constant across (θ1,θ2). In this case,

    0=1β10((1v)(e1(1v)β1)d[F12(v)F11(v)]1βθ2θ1((1v)(e1(1v)β1)d[F12(v)F11(v)]>0,

    which is a contradiction. Consequently, for all 0v1, F12(v)F11(v)=c (constant), indicating that X1 and X2 have an equal distribution according to the location parameter. Alternatively, in more detail, the proof can be argued by contradiction. Suppose that F12(v)F11(v) is not constant on some subinterval (θ1,θ2). Then, because F12(v)F11(v) is decreasing, the integral over (θ1,θ2) will be positive:

    1βθ2θ1(1v)(e1(1v)β1)d[F12(v)F11(v)]>0.

    This contradicts the equality RGEXnβ(X1)=RGEXnβ(X2), implying that F12(v)F11(v) must be constant for all v[0,1].

    Remark 2.1. As a deeper explanation of Theorem 2.3, the phrase "Consequently, up to a location parameter, the distributions of X1 and X2 are identical" means that the distributions of X1 and X2 are the same except for a shift or translation along the real number line. In other words, X1 and X2 have the same shape and structure in their distributions, but one is a shifted version of the other. For example, let X1 and X2 be Exp(γ) distributed random variables with the rate parameter γ>0. If X1 and X2 both follow the same exponential distribution with the rate parameter γ, then their CDFs are:

    F1(x)=1eγx,F2(x)=1eγx.

    Here, X1 and X2 are identically distributed, and there is no shift (c=0). Suppose that X2 is a shifted version of X1 by a constant c>0. In this case, the CDF of X2 is:

    F2(x)=F1(xc)=1eγ(xc),xc.

    Here, X2 is the same as X1 but shifted to the right by c. The distributions are identical up to the location parameter c.

    The following assertion gives an alternate formula for the residual cumulative generalized exponential entropy of X. The sequel uses this formulation, which is in respect of an expanding convex function, to derive a number of findings.

    Lemma 2.2. The following is true for β>0 if X represents an entirely constantly non-negativity random variable with the limiting mean μ=E(X):

    RGEXnβ(X)ψβ(μ)β,

    with noting that the function ψβ() is given in (2.6).

    Proof. Given that ψβ() is a function to be convex, Jensen's inequality may be used as E(ψβ(X))ψβ(E(X)), which obtains the proof.

    The following characteristics of stochastic order are useful for comparing risk measurements and are also obtained by using Proposition 2.3.

    Proposition 2.4. For β>0, it is true that if X1 and X2 are completely continuously non-negative random variables that correspond to X1IcxOrX2, then

    ψβ(X1)IcxOrψβ(X2),

    where (2.6) defines the function ψβ(). X1IcxOrX2 specifically suggests

    RGEXnβ(X1)RGEXnβ(X2).

    Proof. According to Theorem 4.A.8 of Shaked and Shanthikumar [13], ψβ(X1)IcxOrψβ(X2) is an function which considered to be increasing convex for β>0 in the function ψβ(). Specifically, using Eq (2.5) and the concept of rising convex order, we obtain RGEXnβ(X1)RGEXnβ(X2).

    Proposition 2.5. Assume that X1,,Xm be m independent non-negativity absolutely continuous random variables with the collective CDF F, and Z1,,Zm are another set of m independent non-negative continuous random variables with the collective CDF F. If XiIcxOrZi for i=1,2,,m, then for all β>0, we obtain

    RGEXnβ(max{X1,X2,,Xm})RGEXnβ(max{Z1,Z2,,Zm}).

    Proof. Given that XiIcxOrZi for i=1,2,,m, using Shaked and Shanthikumar's [13] Corollary 4.A.16, we get

    max{X1,X2,,Xm}IcxOrmax{Z1,Z2,,Zm}.

    The result then follows directly from Proposition 2.3.

    Proposition 2.6. For β>0, the following is true if X1 and X2 are non-negative randomized variables that match X1HrOrX2:

    RGEXnβ(X1)E(X1)RGEXnβ(X1)E(X2).

    Proof. Shaked and Shanthikumar [13] determined that, assuming X1HrOrX2, the function ψβ(), described in (2.6), is an increasing function to be convex such that ψβ(0)=0:

    E(ψβ(X1))E(X1)E(ψβ(X2))E(X2).

    Therefore, Proposition 2.4 completes the evidence.

    In this subsection, we will examine the characterization of the residual cumulative generalized exponential entropy based on order statistics. Using the well-known Müntz-Szász theorem, we first review the idea of a full sequence of functions and a lemma; for further information, see [14,15,16].

    Lemma 2.3. (Müntz-Szász theorem; see Higgins [16], pp, 95–96). On a limited interval (θ1,θ2), for a function which considered integrable Θ(z), if

    θ2θ1ztiΘ(z)dz=0,i1,

    for nearly all z(θ1,θ2), Θ(z)=0, where {ti,i1} is a sequence of positive integers that increasing strictly fulfilling

    j=11tj=+.

    In functional analysis, the well-known Lemma 2.3 states that the set of values {zt1,zt2,:1t1<t2<} constitutes a complete sequence. It is important to note that Galambos [17] presents a straightforward version of the Müntz-Szász theorem along with a proof (see Theorem AlI.3). Furthermore, the Müntz-Szász theory for {ψti(z),ti1} was extended by Hwang and Lin [14], where ψ(z) is monotone and absolutely continuous on a range (θ1,θ2). We examine characterization using the first-order statistics in the following theorem. Let Z1,,Zt be completely continuously non-negativity random variables with the routine PDF f and CDF F, and let t be independent and distributed in an identical manner. The function that is considered to be the survival function of the first-order statistics is thus expressed as follows: ˉF1;t(z)=[ˉF(z)]t, z0.

    Theorem 2.4. Assume that X and Z are two completely continual, non-negativity random variables, each with a PDF of f and h and a CDF of F and H. Then, if and only if F and H are members of the same distribution family, albeit, with a different scale and location, we have

    RGEXnβ(X1;t)E(X1;t)=RGEXnβ(Z1;t)E(Z1;t),

    for every t=tk, k1, and for a given β>0, such that

    k=1t1k=.

    Proof. The necessary condition is inessential. For the sufficiency condition, after letting v=ˉF(x), we realize that

    E(X1;t)=0ˉF1;t(x)dx=10vtf(ˉF1(v))dv,

    and that,

    βRGEXnβ(X1;t)=10vt(e1vtβ1)f(ˉF1(v))dv.

    Consequently,

    RGEXnβ(X1;t)E(X1;t)=RGEXnβ(Z1;t)E(Y1;t),

    is equivalent to

    10vt(e1vtβ1)f(ˉF1(v))dv10vtf(ˉF1(v))dv=10vt(e1vtβ1)h(ˉH1(v))dv10vth(ˉH1(v))dv.

    Therefore,

    10vt(e1vtβ1)[1f(ˉF1(v))1Ψh(ˉH1(v))]dv,=10vtΘ(v)dv,

    where

    Θ(v)=(e1vtβ1)[1f(ˉF1(v))1Ψh(ˉH1(v))], (2.7)
    Ψ=10vtf(ˉF1(v))dv10vth(ˉH1(v))dv.

    For every t=tk,k1, the latter relation is hypothesized to exist, so that k=1t1k=. Applying the Müntz-Szász theorem to the whole sequence vt,t1 and Lemma 2.3, the relation (2.7) provides Θ(v)=0, or equivalently, f(ˉF1(v))=Ψh(ˉH1(v)), for every 0<v<1. We may observe that F1(v)=ΨH1(v)+c, for any 0<v<1 and a real constant c, if we remember that ddv¯F1(v)=1f(ˉF1(v)). By the same manner, in Psarrakos and Toomaj [18], the CDFs F and H are members of the same distribution family, with a different scale and location.

    Theorem 2.5. Consider two completely continuous, non-negativity random variables, X and Z, each with PDFs of f and h and CDFs of F and H. For a change in location, F and H are members of the same distribution family, if and only if

    RGEXnβ(X1;t)=RGEXnβ(Z1;t),

    for a fixed β>0 and for every t=tj, j1, where k=11tk=.

    Proof. For Ψ=1, the proof is comparable to Theorem 2.4.

    In this section, we will examine the expansion of the residual cumulative generalized exponential entropy and obtain some results. The definition of the residual cumulative generalized exponential entropy is:

    RGEXnβ(X)=1β0ˉF(x)(e1ˉFβ(x)1)dx.

    Using the Taylor series expansion for e1ˉFβ(x)1, we have:

    e1ˉFβ(x)1=(1ˉFβ(x))+(1ˉFβ(x))22!+(1ˉFβ(x))33!+=k=1(1ˉFβ(x))kk!.

    For β>0, ˉFβ(x)0 for all x, so 1ˉFβ(x) is finite and the series converges. Substituting into the integral:

    RGEXnβ(X)=1βk=11k!0ˉF(x)(1ˉFβ(x))kdx.

    For (1ˉFβ(x))k, we use the binomial theorem to obtain

    (1ˉFβ(x))k=kj=0(kj)(1)jˉFβj(x).

    Substituting this back yields

    RGEXnβ(X)=1βk=11k!kj=0(kj)(1)j0ˉF1+βj(x)dx. (3.1)

    For β>0, the survival function ˉF(x) typically decreases to 0 as x, ensuring the convergence of ˉF1+βj(x).

    The residual or excess of Z, assuming that it surpasses a threshold t, is represented by Zt=[ZtZ>t] if the random variable Z represents the lifespan of a component or a system. In contrast, [ZW] often indicates a random variable with a similar distribution as Z conditional on W. The PDF of Zt is obtained as follows:

    f(zt)=f(z)ˉF(t),z>t.

    The equation for the function of the average residual life of Z with a finite mean μ can potentially be calculated as

    M(t)=M(Z;t)=EZ>t[ZtZ>t],t0, (3.2)

    in this case, EZ>t denotes the expectation of the residual PDF f(zt). With Zβ in place of Z, the function of the mean residual life of Zβ, represented as Mβ(t), may be found using (3.2). We note that the best choice under the quadratic loss function is the function of the mean residual life of Zβ:

    Qls(δ,Zβt)=(Zβtδ)2,Zβ>t,

    for excess of prediction, i.e.,

    δ(t)=argminδEZβ>t[Qls(δ,Zβt)]=Mβ(t),β>0.

    Asadi et al. [19] noted that Mβ(t) is a risk local measure that depends on the threshold t. The Bayes risk is its global risk, as

    Br(Mβ)=EΠ[Mβ(t)],

    where Π(t) is the distribution of prior for the threshold t.

    Theorem 3.1. Let Z have the function of the mean residual life M, and assume the baseline prior Π(t)=f(t). Then, the residual cumulative generalized exponential entropy can be expressed by the Bayes risk of Mβ(t) as follows:

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jjBr(Mβj+1)=k=11k!kj=0(kj)(1)jjE[Mβj+1(Z)]. (3.3)

    Proof. Under the substitution Π(t)=f(t), and utilizing the theorem of Fubini, we can express the Bayes risk as

    Br(Mα)=0Mα(t)Π(t)dt=0Mα(t)f(t)dt=0(tˉFα(z)dzˉFα(t))f(t)dt=0ˉFα(z)(z0f(t)dtˉFα(t)dt)dz=0ˉFα(z)(1α1(ˉFα+1(z)1))dz=1α10(ˉF(z)ˉFα(z))dz=1α1(μ0ˉFα(z)dz),

    where 1α>0. Therefore, we can see that

    0ˉFα(z)dz=μ(α1)Br(Mα). (3.4)

    Substituting from (3.4) in (3.1), we obtain

    RGEXnβ(Z)=1βk=11k!kj=0(kj)(1)j[μ(βj)Br(Mβj+1)]=k=11k!kj=0(kj)(1)jjBr(Mβj+1),

    where the last line is obtained from noting that the inner summation:

    kj=0(kj)(1)j=(11)k=0,for k1.

    Theorem 3.2. Assume that Z is a non-negativity, exactly continually random variable following the PDF f(z). In this case

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jj(E(Zβj+1)+E(gβj+1(Z))),

    where E(Zβ)=0ˉFβ(z)dz, gβ(v)=v0M(u)ˉFβ1(u)du, and v>0.

    Proof. From (3.1), we can rewrite it as

    RGEXnβ(Z)=1βk=11k!kj=0(kj)(1)jj[1βj(μ0ˉF1+βj(z)dz)]. (3.5)

    Let η(z)=f(z)ˉF(z) be the function of the hazard rate. We then have the integration

    0M(z)η(z)ˉFβ(z)dz=0(zˉF(t)dtˉF2β(z))f(z)dz=0ˉF(z)(z0f(t)ˉFβ2(t)dt)dz=1β1(μ0ˉFβ(z)dz). (3.6)

    By substituting (3.6) in (3.5), and using the result M(u)η(u)=1+M(u), we get

    RGEXnβ(Z)=1βk=11k!kj=0(kj)(1)jj[0M(z)η(z)ˉFβj+1(z)dz]=1βk=11k!kj=0(kj)(1)jj[E(Zβj+1)+0M(z)ˉFβj+1(z)dz]=1βk=11k!kj=0(kj)(1)jj[E(Zβj+1)+0M(z)xf(u)[ˉF(z)]βjdudz]=1βk=11k!kj=0(kj)(1)jj[E(Zβj+1)+0f(u)u0M(z)[ˉF(z)]βjdzdu],

    then the result follows.

    This part examines the connection between the residual cumulative generalized exponential entropy and the wealth order excess, sometimes referred to as the spread right order. Examining the standard deviations of two distribution functions is usually the simplest method to compare their variability. However, Shaked and Shanthikumar [13] developed and thoroughly analyzed stochastic ordering and different transformations for comparing the variability, since numerical measurements alone may not always give adequate information. One of them is the order of excess wealth, which is used to evaluate spread. The wealth excess convert for a non-negativity random variable Z follows a CDF H and PDF h is given by (cf. Fernandez-Ponce et al. [20])

    ΔZ(v)=H1(v)¯H(z)dz=1v(1q)1h(H1(q))dq=1v(H1(q)H1(v))dq,

    with noting that H1(v)=inf{z:F(z)v},v(0,1), is the quantile function of H, and dz=ddqH1(q)dq=1h(H1(q))dq. Therefore, the difference H1(q)H1(v) measures the excess above the threshold F1(v) at a level q.

    This function and the function of the mean residual life are also connected in this manner by the following connection:

    MZ(F1(u))=ΔZ(v)1v,0<v<1. (3.7)

    Equation (3.7) is used to prove the following theorem.

    Theorem 3.3. If Z has a CDF of H and is a completely continuous, non-negative random variable, then

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jj[10MZ(H1(v))(1v)βjdv]. (3.8)

    for all β>0.

    Proof. The residual cumulative generalized exponential entropy given in (3.3), can be rewritten as

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jjE[Mβj+1(Z)]=k=11k!kj=0(kj)(1)jjβj+1E[MZ(Zβj+1)],

    and the result follows.

    Example 3.1. (1) Suppose that Z has a distribution of Ud(0,d). It is clear to see that

    MZ(F1(v))=d(1v)2.

    Consequently, using (3.8), we obtain

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jjd2[10(1v)βj+1dv]=k=11k!kj=0(kj)(1)jjd2(βj+2),

    β>0.

    (2) Let us examine the Pareto distribution, characterized by a scale parameter p>0 and a shape parameter s>0, where the function of survival is provided by ¯H(z)=ps(z+p)s for z0. It is straightforward to observe that

    MZ(F1(v))=p(1v)1s(s1).

    Consequently, using (3.8), we obtain

    RGEXnβ(Z)=k=11k!kj=0(kj)(1)jjp(s1)[10(1v)βj1sdv]=k=11k!kj=0(kj)(1)jjsp(s(βj+1)1)(s1),

    β>0.

    The residual cumulative generalized exponential entropy is estimated in this section using the empirical residual cumulative entropy. For any β>0, and the random sample X1,X2,,Xn, the empirical estimation of the residual cumulative generalized exponential entropy is expressed as

    RGEXnβ(ˆFn)=1β0ˆFn(x)[e1ˆFβn(x)1]dx=1βn1j=1Xj+1;nXj;n(1jn)[e1(1jn)β1]dx=1βn1j=1Ωj+1(1jn)[e1(1jn)β1], (4.1)

    with noting that X1;nX2;nXn;n are the associated order statistics of the random sample, and the sample spacings are Ωj+1=Xj+1;nXj;n, j=1,2,...,n1. For the sample that corresponds to F, the empirical distribution function is described by ˆFn(x)=n1j=1jnA[xj,xj+1](x),x0, with the indicator function, Aϖ(x)=1, xϖ. Moreover, with β>0, we can use the expansion form of the residual cumulative generalized exponential entropy in (3.1) to present its empirical expression as follows:

    RGEXnβ(ˆFn)=1βk=11k!kj=0(kj)(1)j0ˉF1+βj(x)dx=1βk=11k!kj=0(kj)(1)jn1l=1Ωl+1(1ln)1+βj. (4.2)

    Utilizing the expansion form of the empirical residual cumulative generalized exponential entropy in (4.2), we can now provide a central limit theorem for this measure, which is derived from a random sample with an exponential distribution.

    Theorem 4.1. A sample selected at random X1,X2,,Xn drawn from a common Exp(γ) distribution is considered. In this case,

    RGEXnβ(ˆFn)E[RGEXnβ(ˆFn)]Var[RGEXnβ(ˆFn)]standard normal distribution,

    where β>0.

    Proof. The empirically residual cumulative generalized exponential entropy measure can be written as a total of the independent exponential random variables Xl, l=1,2,...,n, using the expansion (4.2), where the variance and expected value are given by (noting that the spacing Ωl+1 are independent and distributed by Exp(γ(nl)))

    E[Xl]=1nγβk=11k!kj=0(kj)(1)j(1ln)βj, (4.3)

    and

    Var[Xl]=1n2γ2β2[k=11k!kj=0(kj)(1)j(1ln)βj]2. (4.4)

    Consider the following: Φl,q=E[|XlE(Xl)|q], q=2,3. From (4.4), one can derive the following estimates for n considered to be large, as follows

    nl=1Φl,2=nl=1E[|XlE(Xl)|2]=nl=1Var[Xl]=1n2γ2β2nl=1[k=11k!kj=0(kj)(1)j(1ln)βj]21n2γ2β210(e1xβ1)2dx=1n2γ2β2g2.

    Additionally, given whatever random variable Xl with an exponential distribution, the following result may be reached. From (4.3), we obtain (observing that E[|XlE(Xl)|3]=2(6e)[E(Xl)]3e, see, [21,22])

    nl=1Φl,3=2(6e)enl=1[E(Xl)]3=2(6e)e(nγβ)3nl=1[k=11k!kj=0(kj)(1)j(1ln)βj]32(6e)en3γ3β310(e1xβ1)3dx=2(6e)en3γ3β3g3.

    Taking note of that

    gq=10(e1xβ1)qdx,q=2,3,

    and the integrand (e1xβ1)q is bounded and continuous over the interval [0,1]. Therefore, it is true for large n, given an adequate function Gn, that

    nl=1Φl,3nl=1Φl,2Gnnn0.

    Thus, the central limit theorem's Lyapunov condition is satisfied (see, for example, [23]), thus completing the proof.

    In this subsection, a different non-parametric estimator can be developed as follows. The residual cumulative generalized exponential entropy given in (3.1) can be rewritten as

    RGEXnβ(X)=1βk=11k!kj=0(kj)(1)j0ˉF1+βj(x)dx=1β(k=11k!kj=0(kj)(1)j(1+βj)(0xf(x)ˉFβj(x)dx)). (4.5)

    The residual cumulative generalized exponential entropy was introduced as an L-functional by Zardasht [24]. Similarly, RGEXnβ(X) in (4.5) can be represented as

    RGEXnβ(X)=1β(k=11k!kj=0(kj)(1)j(1+βj)(0xˉFβj(x)dF(x)))=1β(k=11k!kj=0(kj)(1)j(1+βj)(0xLnβj(F(x))dF(x))), (4.6)

    where Lnβj(v)=(1u)βj, 0u1. In the follow-up, we can produce an estimate for RGEXnβ(X) using the following L-statistic by replacing F in (4.6) with ˆFn:

    RGEXnβ(ˆFn)=1β(k=11k!kj=0(kj)(1)j(1+βj)(0xLnβj(ˆFn)dˆFn))1β(k=11k!kj=0(kj)(1)j(1+βj)n(nl=1Xl;nLnβj(ln))). (4.7)

    Example 4.1. A random sample X1,X2,,Xn selected from the EXP(1) distribution is considered. Ωl+1=X(l+1)X(l) provides the sample spacing for l=1,2,,n1, which are independent. Each Ωj+1 has an exponential distribution with a parameter of (nl). Thus, from (4.1), we obtain the following:

    (1) The mean of RGEXnβ(ˆFn) is

    E[RGEXnβ(ˆFn)]=1nβn1l=1(e1(1ln)β1).

    (2) The variance of RGEXnβ(ˆFn) is

    Var[RGEXnβ(ˆFn)]=1n2β2n1l=1(e1(1ln)β1)2.

    It is possible to infer, from Eq (4.7) and the relations (4.6.6)–(4.6.8) provided by Arnold et al. [25], that

    (1) The mean of RGEXnβ(ˆFn) is

    E[RGEXnβ(ˆFn)]=1βk=11k!kj=0(kj)(1)j(1+βj)nnl=1Lnβj(ln)lp=11np+1.

    (2) The variance of RGEXnβ(ˆFn) is

    Var[RGEXnβ(ˆFn)]=1β2(nl=1((k=11k!kj=0(kj)(1)j(1+βj)nLnβj(ln))2lp=11(np+1)2)+2nl1=1nl2=1(k=11k!kj=0(kj)(1)j(1+βj)nLnβj(l1n))(k=11k!kj=0(kj)×(1)j(1+βj)nLnβj(l2n))l1p=11(np+1)2).

    Remark 4.1. The intervals of confidence for RGEXnβ(X) can possibly be computed using the results of Theorem 4.1 if the random variables have an exponential distribution. In particular, from (4.1) and (4.7), it is true for any specific β>0 that

    RGEXnβ(ˆFn)±χδ2Var[RGEXnβ(ˆFn)],
    RGEXnβ(ˆFn)±χδ2Var[RGEXnβ(ˆFn)],

    where the critical point of the standard normal distribution at δ2 is shown by χδ2.

    The mean and variance of the empirical residual cumulative generalized exponential function of entropy for the aforementioned relationships are shown in Table 1. Sample sizes of 10,20,30,40, and 50 were chosen, with varying values of the order β. The residual cumulative generalized exponential entropy's precise values are computed as follows: RGEXn1(X)=0.718282, RGEXn2(X)=0.515039, and RGEXn3(X)=0.398348. On the basis of the results in Table 1, we can conclude the following

    Table 1.  Empirical residual cumulative generalized exponential entropy's expected value and variance for the exponential distribution with a unit mean for β=1,2, and 3 and n=10,20,30,40, and 50 sample sizes.
    Sample size β E[RGEXnβ(ˆFn)] Var[RGEXnβ(ˆFn)] E[RGEXnβ(ˆFn)] Var[RGEXnβ(ˆFn)]
    n = 10 1 0.633799 0.0618118 0.636254 0.086091
    2 0.471249 0.0302985 0.57848 0.0651343
    3 0.368875 0.0174016 0.530634 0.056318
    n = 20 1 0.675683 0.0343049 0.027132 0.617761
    2 0.493352 0.0160719 0.55787 0.0203291
    3 0.38382 0.00911087 0.507678 0.017622
    n = 30 1 0.689803 0.023654 0.611674 0.0135907
    2 0.500628 0.0109196 0.550958 0.0101643
    3 0.388709 0.00616504 0.499906 0.0061073
    n = 40 1 0.696893 0.0180386 0.608645 0.00826977
    2 0.504248 0.0082666 0.547493 0.00618269
    3 0.391136 0.00465795 0.495995 0.00438061
    n = 50 1 0.701156 0.014575 0.606833 0.00560691
    2 0.506414 0.00665019 0.545411 0.00419222
    3 0.392587 0.00374276 0.493641 0.00365316

     | Show Table
    DownLoad: CSV

    (1) It is evident that the mean converges to the true value, and the variability of the empirical measurement approaches zero as the sample size increases.

    (2) For any fixed n and increasing β, the variance decreases.

    (3) For any large n, the second estimator provides a more accurate result (by decreasing the variance) compared with the first.

    Example 4.2. Reaven and Miller [26] investigated the connection between insulin and blood chemistry indicators of glucose tolerances in 145 non-fat individuals. They visualized the data in three dimensions using the Stanford Accelerator Linear Center's PRIM9 technology and found an odd pattern that resembled a big blob with two wings pointing in separate directions. Three categories were created from the 145 observations: Overt diabetes, Chemical diabetics, and Normals. Additionally, the 145 observations were divided into three groups: 33 for overt diabetes, 36 for chemical diabetes, and 76 for normal diabetes. Five factors for every single individual were examined as follows:

    (1) Relative of weight (Λ1),

    (2) Test plasma glucose level (Λ2),

    (3) Fasting plasma glucose level (Λ3),

    (4) Plasma insulin during the test (Λ4), and

    (5) Steady state plasma glucose (Λ5).

    For the Λ5 data set, we fit it to the exponential distribution with the parameter γ=0.00433. Figure 2 shows the histogram of the Λ5 data set with the exponential curve, along with the empirical and theoretical CDFs. To statistically validate this fit, we performed a Kolmogorov-Smirnov test, which yielded a p-value of 0.051032. This result supports the suitability of the EXP(0.00433) distribution for modeling these data.

    Figure 2.  Histogram of the Λ5 data set along with the exponential curve (left) and the empirical and theoretical CDFs (right).

    Figure 3 shows the theoretical residual cumulative generalized exponential entropy, which was computed using this exponential parameter distribution and the confidence interval (at δ=0.05) obtained from Eq (4.1). It is clear that the estimators' confidence intervals contain the theoretical value.

    Figure 3.  Theoretical measure of the residual cumulative generalized exponential entropy with 95% confidence intervals in Example 4.2.

    In this section, we will apply some properties and applications with the generalized exponential entropy including symmetry characterization and classification problem with pattern of recognition.

    Using (1.5), we can define the continuous case of the generalized exponential entropy of the continual random variable Z with the PDF f as follows:

    GEXn(Z)=1βf(z)(e1fβ(z)1)dz, (5.1)

    where βR{0}. Several interesting features of the extended exponential entropy of order statistics appear when the PDF of the underlying identical besides the independent of random variables is symmetric. With an underlying distribution Z containing the lth-order statistic Zl;t, 1lt, the PDF of a sample of size t is derived by

    fl;t(z)=1Betaf(l,tl+1)Fl1(z)¯Ftl(z)f(z), (5.2)

    with nothing that Betaf(l,tl+1)=Γ(l)Γ(tl+1)Γ(t+1). We begin with two lemmas discussed by Fashandi and Ahmadi [27] and Balakrishnan and Selvitella [28], respectively, the definition of fr;t in (5.2) and the symmetry assumption serve as the immediate foundation for the proof.

    Lemma 5.1. [27] The following result is supported by GZ, PDF f, and CDF F, and Z is a continuous random variable as

    f(F1(v))=f(F1(1v))for allv(0,1),

    which suggests the symmetry of F(z) with respect to a constant gnGZ.

    Lemma 5.2. [28] Assume that the parent distribution of the order statistic Zl;t, l=1,...,t, has a PDF f with noting f(μ+z)=f(μz), z0. We proceed with the following analysis:

    F(μ+z)=¯F(μz),fl;t(μ+z)=ftl+1;t(μz).

    Theorem 5.1. Let Z1,...,Zt be identical, including independent distributed observations over Z whose PDF is regarded as symmetric around its mean μ. Consequently, we have

    (1) In the event that t is deemed to be odd, GEXnβ(Zl;t)=GEXnβ(Ytl+1;t), l=1,...,t.

    (2) If and only if GEXnβ(Z1;t)=GEXnβ(Zt;t), t1, then Z has a symmetric PDF.

    Proof. (1) Lemma 5.2 and Eq (5.1) provide us with

    GEXnβ(Zl;t)=1βfl;t(z)(e1fβl;t(z)1)dz1βfl;t(μ+z)(e1fβl;t(μ+z)1)dz=1βftl+1;t(μz)(e1fβtl+1;t(μz)1)dz=1βftl+1;t(z)(e1fβtl+1;t(z)1)dz=GEXnβ(Ztl+1;t).

    (2) The first part of this theorem establishes the necessity. We now turn to the sufficiency. Suppose that GEXnβ(Zl;t)=GEXnβ(Zt;t) holds for all t1. Using Lemma 5.1, we derive the following, for every v(0,1),

    f(F1(1v))=f(F1(v)),

    which leads to dduF1(1v)=ddvF1(v). Consequently, F1(1v)=F1(v)+gn, and thus f(F1(v)gn)=f(F1(v)), where gn is a constant, which is valid for all v(0,1). Substituting F1(v)=gn2+z, we obtain f(gn2z)=f(gn2+z) for all zR, thereby completing the proof of the theorem.

    Corollary 5.1. In alignment with Theorem 5.1, let ΔGEXnβ(Zp;t)=GEXnβ(Zp+1;t)GEXnβ(Zp;t) represent the forward difference operator with respect to p, where 1pt1. It follows that ΔGEXnβ(Zl;t)=ΔGEXnβ(Ztl;t), for l=1,...,t.

    Remark 5.1. Define Ξt as GEXnβ(Z1;t)GEXnβ(Zt;t). The condition Ξt=0, for t=1,2,..., holds if and only if Z exhibits symmetry. Consequently, Ξt can serve as a fundamental measure of symmetry and as a statistic for testing symmetry.

    From the assumptions outlined in Corollary 5.1, it can be deduced that the fractional generalized entropy GEXnβ(Zl;n) attains either a local maximum at the median. This property is illustrated using the Ud(1,1) distribution. Specifically, for the generalized exponential entropy of the median (l=4) with t=7, the maximum values are 0.7269227 for β=1, 0.56026 for β=2, 0.4016726 for β=3, and 0.3036904 for β=4 (refer to Figure 4).

    Figure 4.  Generalized exponential entropy of the lth-order statistics of Ud(1,1) distribution.

    The probabilistic measures of information in the classification problems using pattern recognition will be covered in this section. For classification problems, uncertainty measurements are a helpful tool; for instance, one might refer to [29] and [30] in this context. We will use the data presented in Example 4.2 to discuss its relevant pattern of recognition using the generalized exponential entropy given in (1.5). The variables and their correlation are presented in Figure 5.

    Figure 5.  The variables and their correlation of Example 4.2 data (:p-value<0.001;:p-value<0.01).

    As indicated in Table 2(I), we select 30 samples for each data categorization in order to build an interval number model. We next identify a sample that has both the greatest and lowest values. An anonymous test case of every possible occurrence in the dataset is displayed. Assuming that the selected singleton data sample (1.18,108,486,297,220) originates from the chemical diabetic group (a singleton is a quantity that occurred only once in a dataset). To identify the singletons in a sample, we must first select a sample from a sequence of data and then note the occurrences of the values in that sample.

    Table 2.  (I) The three specified groups' interval numbers. (II) Distributions of probabilities according to the interval values.
    (I) Item Λ1 Λ2 Λ3 Λ4 Λ5
    Normal [0.74, 1.2] [74,112] [269,418] [81,267] [29,273]
    Chemical diabetic [0.83, 1.2] [75,114] [413,643] [109,748] [60,300]
    Overt diabetic [0.74, 1.2] [120,353] [538, 1520] [10,460] [150,458]
    (II) Item Λ1 Λ2 Λ3 Λ4 Λ5
    P(normal) 0.319397 0.45749 0.318917 0.395996 0.329123
    P(chemical diabetic) 0.361206 0.483345 0.603506 0.234745 0.405461
    P(overt diabetic) 0.319397 0.0591652 0.0775768 0.369259 0.265417

     | Show Table
    DownLoad: CSV

    Subsequently, we create five distinct probability distributions using Kang et al.'s [31] technique, which is based on the closeness between interval numbers. R1=[λ1,λ2] and R2=[λ1,λ2] are the two ranges taken into account. The distance between the ranges R1 and R2 is then calculated by

    I(R1,R2)=[(λ1+λ22)(λ1+λ22)]2+13[(λ1λ22)2+(λ1λ22)2].

    Furthermore, their similarity ρ(R1,R2) is explained as

    ρ(R1,R2)=11+ζI(R1,R2),

    in which ζ is the supporting coefficient; one example of its application is to set ζ to 5. For range the R1, we use the ranges listed in Table 2(I); for the range R2, we use individual values from the selected sample to generate the given probability distributions (for example, the range of for the value 1.18 in the Λ1 attribute is R2=[1.18,1.18]). Table 2(II) shows that each of the five evaluated criteria produces three similarity values. A probability distribution is then created using the normalized representation of this data. These probability distributions are then evaluated using our generalized exponential entropy measure (with β=1, β=2, and β=3) and are shown in Table 3(I). We also use W(T)=eT as the weighting foundation due to the monotonicity of the function that is considered to be exponential. The weights are then obtained by normalizing them. For example, when the generalized exponential entropy's Λ3 characteristic is used, the procedure produces

    W(Λ3)=eGEXn(Λ3)eGEXn(Λ1)+eGEXn(Λ2)+eGEXn(Λ3)+eGEXn(Λ4)+eGEXn(Λ5),
    Table 3.  (I) Generalized exponential entropy measurements (II) W(Λt), t=1,2,3,4,5, where the weights represent the five variables.
    (I) Item Λ1 Λ2 Λ3 Λ4 Λ5
    β=1 0.945856 0.748902 0.722495 0.922879 0.931814
    β=2 0.71489 0.604572 0.5662 0.699949 0.705139
    β=3 0.539139 0.488135 0.4556 0.531995 0.534171
    (II) Item W(Λ1) W(Λ2) W(Λ3) W(Λ4) W(Λ5)
    β=1 0.181642 0.221183 0.227102 0.185864 0.18421
    β=2 0.188615 0.210614 0.218853 0.191454 0.190463
    β=3 0.194114 0.204272 0.211027 0.195506 0.195081

     | Show Table
    DownLoad: CSV

    The weighted values W(Λt), t=1,2,3,4,5, corresponding to the five characteristics, are presented in Table 3(II). Therefore, the final probability distribution of the generalized exponential entropy measure is listed as:

    (1) Under β=1, we obtain

    P(normal)=0.36586,P(chemical diabetic)=0.427895,P(overt diabetic)=0.206244.

    (2) Under β=2, we obtain

    P(normal)=0.364894,P(chemical diabetic)=0.424175,P(overt diabetic)=0.210931.

    (3) Under β=3, we obtain

    P(normal)=0.364377,P(chemical diabetic)=0.421197,P(overt diabetic)=0.214426.

    Following that, it was concluded that the chosen sample is most likely to be in the chemical diabetic category. As a result, in this case, the conclusion was correct.

    Using this strategy, we examined all 145 samples via the entropy and generalized exponential entropy approaches over a range of β values, including 33 in the overt diabetic, 36 in the chemical diabetic, and 76 in the normal categories. The recognition rates are displayed in Table 4, where the entropy technique yields 72.03%. In the meanwhile, 59.23 is obtained using the extended exponential entropy technique. Moreover, when β=1, we can see the probabilities when we choose the 76 normal categories tested samples in Figure 6, the 36 chemical diabetic categories tested samples in Figure 7, and the 33 overt diabetic categories tested samples in Figure 8 (the higher probability indicates the right choice).

    Table 4.  The recognition rates of entropy and generalized exponential entropy approaches.
    Method Normal Chemical diabetic Overt diabetic In general
    Entropy method 100% 55.5% 60.6% 72.03%
    Generalized exponential entropy
    method 100% 44.4% 33.3% 59.23%

     | Show Table
    DownLoad: CSV
    Figure 6.  Probabilities of the 76 normal categories tested samples.
    Figure 7.  Probabilities of the 36 chemical diabetic categories tested samples.
    Figure 8.  Probabilities of the 33 overt diabetic categories tested samples.

    It must be emphasized that the results obtained in this section are based on the specific dataset used in our analysis and should not be interpreted as a general statement of preference. The performance of the proposed model may vary with different datasets, where its advantages could become more evident.

    In this study, we emphasized that the study of the continuous case must be carried out and not just the discrete case, as its importance appeared in many of the applications used, such as dealing with the order statistics, as dealing with them in the discrete case is not flexible and limited. In addition, it is important to study the model based on the distribution function because of its many advantages. Besides, we have seen that the residual cumulative generalized exponential entropy measure is a generalization of the original model presented by Rao et al. [2]; this is one of the reasons why we do not rely on the original model. Moreover, proving that this measure is bounded contributed to solving some problems that have appeared in some theories and made it flexible in dealing with them. In addition, the measure's reliance on the exponential function made us use its expansion in studying some topics, like the order of excess wealth, Bayes risk, and estimations. Moreover, the real data were selected to see the application of this measure in terms of non-parametric estimators and the extent of suitability of these estimators with the proposed measure. On the other hand, this data was used to solve the problem of classification, as is clear at the end in Table 4, where the extent of the efficiency of using this measure to solve the problem at hand was shown.

    Overall, the implications of the residual cumulative generalized exponential entropy as an extension of the residual cumulative entropy, which tends to it when β0, has been presented. Numerous findings have been examined, including non-negativity, limits, relationship to the measure of classical differential measure of entropy, and preservation features, with a few well-known and familiar stochastic comparisons. Besides, some of those features are verified with some well-known distributions. In addition, some characterization of our model based on the first-order statistics has been obtained. Moreover, under the Taylor series expansion, some results on the expansion of the residual cumulative generalized exponential entropy, such as Bayes risk and the connection with the transform of excess wealth, have been explained. On top of that, we have examined the issue of using its empirical CDF to estimate the residual cumulative generalized exponential entropy. In this context, we estimate this measure using two distinct empirical estimators of the CDF. A theorem of the centralized limitation for the empirical measurement of randomly samples drawn from a distribution that is considered to be exponential is developed for the first estimator. A theorem of the central limit solution for empirical measuring constructed from a randomly sample using an unknown distribution is provided; however, it is also provided for the second estimator. Both methods were used on the data, and we found that they are close in their results, as the average values give results close to the true values, but there is some slight advantage for the second estimator. Moreover, the continuous case of generalized exponential entropy is discussed to illustrate the symmetry characterization of order statistics, using an example of a symmetric uniform distribution that shows that the median is the point of symmetry. On the other hand, generalized exponential entropy has been discussed in many areas, such as multi-criteria decision-making. Therefore, we have applied this model to the classification issue by utilizing the pattern recognition of a diabetes dataset and comparing it with the classical entropy, which shows superiority to the classical entropy measure.

    In future work, we can implement the obtained measure to different topics like the concomitants of order statistics and tests of hypotheses and compare it against other existing entropy-based models in terms of interpretability, efficiency, or computational feasibility. Moreover, we aim to extend our analysis to additional common distributions, such as the normal and gamma distributions, which currently require numerical solutions. Investigating these distributions in greater detail will help illustrate the broader applicability of the measure and provide further insights into its practical utility. Furthermore, while the current study relies on closed-form expressions for the variance and mean (which eliminate the need for simulation-based data), future work will incorporate simulation studies to compute the mean square error. This will allow us to further validate the estimator's accuracy and assess the convergence of bias as the sample sizes increase. Moreover, we plan to explore the integration of the stochastic precedence order to evaluate its potential contributions and impact on our theoretical framework, thereby broadening the scope of stochastic ordering relations analyzed in this study.

    The authors declare they have not used artificial intelligence (AI) tools in the creation of this article.

    This study is supported via funding from Prince Sattam bin Abdulaziz University project number (PSAU/2025/R/1446).

    The authors declare there is no conflict of interest.

    The following Mathematica codes are provided to allow readers to reproduce the results presented in the paper.

    Beta = 2; Gamma = 1;

    fx = 1 - Exp[-Gamma x];

    RGEXn = N[1/Beta Integrate[(1 - fx) (Exp[1 - (1 - fx)^Beta] - 1),

    {x, 0, Infinity}]];

    n = 10;

    mean = N[1/(n Beta) Sum[(Exp[1 - (1 - l/n)^Beta] - 1), {l, 1, n - 1}]];

    var = N[1/(n Beta)^2 Sum[(Exp[1 - (1 - l/n)^Beta] - 1)^2,

    {l, 1, n - 1}]]

    y = Sort[{data}];

    n = Length[y];

    Beta = Beta;

    Gamma = 0.00433;

    fx = 1 - Exp[-Gamma x];

    RGEXn = N[1/Beta Integrate[(1 - fx) (Exp[1 - (1 - fx)^Beta] - 1),

    {x, 0, Infinity}]];

    s1 = N[1/Beta Sum[(y[[i + 1]] - y[[i]]) (1 - i/n)

    (Exp[1 - (1 - i/n)^Beta] - 1), {i, 1, n - 1}]];

    var1 = N[Variance[y]];

    c1 = s1 + 1.96 Sqrt[var1];

    c2 = s1 - 1.96 Sqrt[var1];

    Plot[{STMn, c1, c2}, {Beta, 0.001, 3},

    PlotLegends -> {"RGEXn", "Upper", "Lower"},

    AxesLabel -> Automatic,

    PlotStyle -> {Black, Red, {Dashed, Red, AbsoluteThickness[3]}},

    Frame -> True]

    rm(list = ls())

    # Load necessary libraries

    library(ggplot2)

    # Define parameters

    Theta < - 4

    n < - 7

    a < - -1

    b < - 1

    # Define the function for fyos1

    fyos1 < - function(y, r) {

    ffy < - (y - a) / (b - a)

    fy < - 1

    return((gamma(n + 1) * ffy^(r - 1) * (1 - ffy)^(n - r) * fy) /

    (gamma(r) * gamma(n - r + 1)))

    }

    # Define the generalized function GF

    GF < - function(r) {

    integrand < - function(y) {

    fyos1_val < - fyos1(y, r)

    return(fyos1_val*(exp(1-(fyos1_val)^(Theta))-1))

    }

    result < - integrate(integrand, a, b)$value

    return((1 / Theta) * result)

    }

    # Create the plot data

    r_values < - 1:n GF_values < - sapply(r_values, GF)

    # Create a data frame for ggplot

    df < - data.frame(r = r_values, GF = GF_values)

    # Plot using ggplot2

    ggplot(df, aes(x = r, y = GF)) +

    geom_point() +

    geom_line() +

    labs(x = "l", y = "", title = expression(paste(beta, " = 3"))) +

    theme_minimal(base_size = 15)

    # Clear the environment

    rm(list = ls())

    library(heplots)

    library(GGally)

    data("Diabetes")

    # Select only the numerical variables

    df < - Diabetes[, c("relwt", "glufast", "glutest", "instest", "sspg")]

    # Create a scatterplot matrix

    ggpairs(df, lower = list(continuous = "smooth"))



    [1] C. Shannon, A mathematical theory of communication, Bell Syst. Tech. J., 27 (1948), 379–423. https://doi.org/10.1002/j.1538-7305.1948.tb01338.x doi: 10.1002/j.1538-7305.1948.tb01338.x
    [2] M. Rao, Y. Chen, B. Vemuri, F. Wang, Cumulative residual entropy: A new measure of information, IEEE Trans. Inf. Theory, 50 (2004), 1220–1228. https://doi.org/10.1109/TIT.2004.828057 doi: 10.1109/TIT.2004.828057
    [3] L. L. Campbell, Exponential entropy as a measure of extent of distribution, Z. Wahrscheinlichkeitstheorie verw Gebiete, 5 (1966), 217–225. https://doi.org/10.1007/BF00533058 doi: 10.1007/BF00533058
    [4] N. R. Pal, S. K. Pal, Object background segmentation using new definitions of entropy, IEE Proc. E, 136 (1989), 284–295. https://doi.org/10.1049/ip-e.1989.0039 doi: 10.1049/ip-e.1989.0039
    [5] N. R. Pal, S. K. Pal, Entropy: A new definition and its applications, IEEE Trans. Syst. Man Cybern., 21 (1991), 1260–1270. https://doi.org/10.1109/21.120079 doi: 10.1109/21.120079
    [6] S. M. Panjehkeh, G. R. M. Borzadaran, M. Amini, Results related to exponential entropy, Int. J. Inf. Coding Theory, 4 (2017), 258–275. https://doi.org/10.1504/IJICOT.2017.086915 doi: 10.1504/IJICOT.2017.086915
    [7] T. O. Kvalseth, On exponential entropies, in Proceedings of the IEEE International Conference on Systems, Man and Cybernatics, 4 (2000), 2822–2826. https://doi.org/10.1109/ICSMC.2000.884425
    [8] S. S. Alotaibi, A. Elaraby, Generalized exponential fuzzy entropy approach for automatic segmentation of chest CT with COVID-19 infection, Complexity, 2022 (2022), 7541447. https://doi.org/10.1155/2022/7541447 doi: 10.1155/2022/7541447
    [9] A. P. Wei, D. F. Li, B. Q. Jiang, P. P. Lin, The novel generalized exponential entropy for intuitionistic fuzzy sets and interval valued intuitionistic fuzzy sets, Int. J. Fuzzy Syst., 21 (2019), 2327–2339. https://doi.org/10.1007/s40815-019-00743-6 doi: 10.1007/s40815-019-00743-6
    [10] S. Kumar, A new exponential knowledge and similarity measure with application in multi-criteria decision-making, Decis. Anal. J., 10 (2024), 100407. https://doi.org/10.1016/j.dajour.2024.100407 doi: 10.1016/j.dajour.2024.100407
    [11] C. Wang, G. Shi, Y. Sheng, H. Ahmadzade, Exponential entropy of uncertain sets and its applications to learning curve and portfolio optimization, J. Ind. Manage. Optim., 21 (2025), 1488–1502. https://doi.org/10.3934/jimo.2024134 doi: 10.3934/jimo.2024134
    [12] J. Ye, W. Cui, Exponential entropy for simplified neutrosophic sets and its application in decision making, Entropy, 20 (2018), 357. https://doi.org/10.3390/e20050357 doi: 10.3390/e20050357
    [13] M. Shaked, J. G. Shanthikumar, Stochastic Orders and Their Applications, San Diego, Academic Press, 1994.
    [14] J. S. Hwang, G. D. Lin, On a generalized moment problem. II, Proc. Am. Math. Soc., 91 (1984), 577–580. https://doi.org/10.1090/S0002-9939-1984-0746093-4 doi: 10.1090/S0002-9939-1984-0746093-4
    [15] U. Kamps, 10 Characterizations of distributions by recurrence relations and identities for moments of order statistics, in Order Statistics: Theory & Methods, Elsevier, 16 (1998), 291–311. https://doi.org/10.1016/S0169-7161(98)16012-1
    [16] J. R. Higgins, Completeness and Basis Properties of Sets of Special Functions, Cambridge University Press, 2004.
    [17] J. Galambos, The asymptotic theory of extreme order statistics, Technometric, 32 (1990), 110–111. https://doi.org/10.1080/00401706.1990.10484616 doi: 10.1080/00401706.1990.10484616
    [18] G. Psarrakos, A. Toomaj, On the generalized cumulative residual entropy with applications in actuarial science, J. Comput. Appl. Math., 309 (2017), 186–199. https://doi.org/10.1016/j.cam.2016.06.037 doi: 10.1016/j.cam.2016.06.037
    [19] M. Asadi, N. Ebrahimi, E. S. Soofi, Connections of Gini, Fisher, and Shannon by bayes risk under proportional hazards, J. Appl. Probab., 54 (2019), 1027–1050. https://doi.org/10.1017/jpr.2017.51 doi: 10.1017/jpr.2017.51
    [20] J. M. Fernandez-Ponce, S. C. Kochar, J. Muñoz-Perez, Partial orderings of distributions based on right spread functions, J. Appl. Probab., 35 (1998), 221–228.
    [21] A. Di Crescenzo, M. Longobardi, On cumulative entropies, J. Stat. Plann. Inference, 139 (2009), 4072–4087. https://doi.org/10.1016/j.jspi.2009.05.038
    [22] H. Xiong, P. Shang, Y. Zhang, Fractional cumulative residual entropy, Commun. Nonlinear Sci. Numer. Simul., 78 (2019), 104879. http://dx.doi.org/10.1016/j.cnsns.2019.104879 doi: 10.1016/j.cnsns.2019.104879
    [23] P. Billingsle, Probability and Measure, John Wiley & Sons, 2008.
    [24] V. Zardasht, Testing the dilation order by using cumulative residual Tsallis entropy, J. Stat. Comput. Simul., 89 (2019), 1516–1525. https://doi.org/10.1080/00949655.2019.1588270 doi: 10.1080/00949655.2019.1588270
    [25] B. Arnold, N. Balakrishnnan, H. N. Nagaraja, A First Course in Order Statistics, Society for Industrial and Applied Mathematics, 2008.
    [26] G. M. Reaven, R. G. Miller, An attempt to define the nature of chemical diabetes using a multidimensional analysis, Diabetologia, 16 (1979), 17–24. https://doi.org/10.1007/BF00423145 doi: 10.1007/BF00423145
    [27] M. Fashandi, J. Ahmadi, Characterizations of symmetric distributions based on Renyi entropy, Stat. Probabil. Lett., 82 (2012), 798–804. https://doi.org/10.1016/j.spl.2012.01.004 doi: 10.1016/j.spl.2012.01.004
    [28] N. Balakrishnan, A. Selvitella, Symmetry of a distribution via symmetry of order statistics, Stat. Probabil. Lett., 129 (2017), 367–372. https://doi.org/10.1016/j.spl.2017.06.023 doi: 10.1016/j.spl.2017.06.023
    [29] N. Balakrishnan, F. Buono, M. Longobardi, On Tsallis extropy with an application to pattern recognition, Stat. Probab. Lett., 180 (2022), 109241. https://doi.org/10.1016/j.spl.2021.109241 doi: 10.1016/j.spl.2021.109241
    [30] R. A. Aldallal, H. M. Barakat, M. S. Mohamed, Exploring weighted Tsallis extropy: Insights and applications to human health, AIMS Math., 10 (2025), 2191–2222. https://doi.org/10.3934/math.2025102 doi: 10.3934/math.2025102
    [31] B. Y. Kang, Y. Li, Y. Deng, Y. J. Zhang, X. Y. Deng, Determination of basic probability assignment based on interval numbers and its application, Acta Electron. Sin., 40 (2012), 1092–1096. https://doi.org/10.3969/j.issn.0372-2112.2012.06.004 doi: 10.3969/j.issn.0372-2112.2012.06.004
  • Reader Comments
  • © 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(485) PDF downloads(30) Cited by(0)

Figures and Tables

Figures(8)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog