Processing math: 100%
Research article

On the effectiveness of the new estimators obtained from the Bayes estimator

  • Received: 13 September 2024 Revised: 18 February 2025 Accepted: 26 February 2025 Published: 14 March 2025
  • MSC : 60E05, 62H10

  • Estimating the mean parameters in random variables, particularly within multivariate normal distributions, is a critical issue in statistics. Traditional methods, such as the maximum likelihood estimator, often struggle in high-dimensional or small-sample contexts, driving interest in shrinkage estimators that enhance accuracy by reducing variance. This study builds on the foundational work by Stein and others examining the minimax properties of shrinkage estimators. In this paper, we propose Bayesian estimation techniques that incorporate prior information within a balanced loss function framework, aiming to improve upon existing methods. Our findings demonstrate the advantages of using the balanced loss function for performance evaluation, which offers a robust alternative to conventional quadratic loss functions. In this paper, we present the theoretical foundations, a simulation study, and an application to real data.

    Citation: Amani Alahmadi, Abdelkader Benkhaled, Waleed Almutiry. On the effectiveness of the new estimators obtained from the Bayes estimator[J]. AIMS Mathematics, 2025, 10(3): 5762-5784. doi: 10.3934/math.2025265

    Related Papers:

    [1] Baria A. Helmy, Amal S. Hassan, Ahmed K. El-Kholy, Rashad A. R. Bantan, Mohammed Elgarhy . Analysis of information measures using generalized type-Ⅰ hybrid censored data. AIMS Mathematics, 2023, 8(9): 20283-20304. doi: 10.3934/math.20231034
    [2] Essam A. Ahmed, Laila A. Al-Essa . Inference of stress-strength reliability based on adaptive progressive type-Ⅱ censing from Chen distribution with application to carbon fiber data. AIMS Mathematics, 2024, 9(8): 20482-20515. doi: 10.3934/math.2024996
    [3] Hassan Okasha, Mazen Nassar, Saeed A. Dobbah . E-Bayesian estimation of Burr Type XII model based on adaptive Type-Ⅱ progressive hybrid censored data. AIMS Mathematics, 2021, 6(4): 4173-4196. doi: 10.3934/math.2021247
    [4] Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Statistical analysis of stress–strength in a newly inverted Chen model from adaptive progressive type-Ⅱ censoring and modelling on light-emitting diodes and pump motors. AIMS Mathematics, 2024, 9(12): 34311-34355. doi: 10.3934/math.20241635
    [5] Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Analysis of reliability index R=P(Y<X) for newly extended xgamma progressively first-failure censored samples with applications. AIMS Mathematics, 2024, 9(11): 32200-32231. doi: 10.3934/math.20241546
    [6] Ahmed Elshahhat, Refah Alotaibi, Mazen Nassar . Statistical inference of the Birnbaum-Saunders model using adaptive progressively hybrid censored data and its applications. AIMS Mathematics, 2024, 9(5): 11092-11121. doi: 10.3934/math.2024544
    [7] Riffat Jabeen, Aamir Sanaullah, Muhammad Hanif, Azam Zaka . Two-exponential estimators for estimating population mean. AIMS Mathematics, 2021, 6(1): 737-753. doi: 10.3934/math.2021045
    [8] Ehab M. Almetwally, Ahlam H. Tolba, Dina A. Ramadan . Bayesian and non-Bayesian estimations for a flexible reduced logarithmic-inverse Lomax distribution under progressive hybrid type-Ⅰ censored data with a head and neck cancer application. AIMS Mathematics, 2025, 10(4): 9171-9201. doi: 10.3934/math.2025422
    [9] Alanazi Talal Abdulrahman, Khudhayr A. Rashedi, Tariq S. Alshammari, Eslam Hussam, Amirah Saeed Alharthi, Ramlah H Albayyat . A new extension of the Rayleigh distribution: Methodology, classical, and Bayes estimation, with application to industrial data. AIMS Mathematics, 2025, 10(2): 3710-3733. doi: 10.3934/math.2025172
    [10] Rashad M. EL-Sagheer, Mohamed S. Eliwa, Khaled M. Alqahtani, Mahmoud El-Morshedy . Bayesian and non-Bayesian inferential approaches under lower-recorded data with application to model COVID-19 data. AIMS Mathematics, 2022, 7(9): 15965-15981. doi: 10.3934/math.2022873
  • Estimating the mean parameters in random variables, particularly within multivariate normal distributions, is a critical issue in statistics. Traditional methods, such as the maximum likelihood estimator, often struggle in high-dimensional or small-sample contexts, driving interest in shrinkage estimators that enhance accuracy by reducing variance. This study builds on the foundational work by Stein and others examining the minimax properties of shrinkage estimators. In this paper, we propose Bayesian estimation techniques that incorporate prior information within a balanced loss function framework, aiming to improve upon existing methods. Our findings demonstrate the advantages of using the balanced loss function for performance evaluation, which offers a robust alternative to conventional quadratic loss functions. In this paper, we present the theoretical foundations, a simulation study, and an application to real data.



    Estimation of the mean parameters in random variables is a significant problem that has garnered considerable attention from researchers in various fields, including statistics, economics, and social sciences. Accurately estimating these parameters is crucial for effective decision-making and inference in many applications, particularly when dealing with multivariate normal distributions (MNDs). The complexity of these distributions requires advanced estimation techniques that can yield reliable results, even in challenging scenarios. The motivations for this work stem from the limitations of traditional estimation methods, particularly the maximum likelihood estimator (MLE). While the MLE is widely used due to its desirable properties under certain conditions, it can fall short in high-dimensional settings or when the sample size is small. This has led to a growing interest in developing shrinkage estimators, which can improve upon the MLE by reducing variance and enhancing the overall estimation accuracy. Researchers have sought to create shrinkage estimators that not only outperform the MLE but also exhibit lower risks, making them more robust in practice.

    Numerous studies have contributed to the understanding of shrinkage estimators. Early foundational work by Stein [11], followed by contributions from James and Stein [6] and Yang and Berger [12], laid the groundwork for exploring the minimax properties of these estimators. Their research established that shrinkage can lead to improved performance in terms of mean squared error compared with traditional estimation methods.

    A comprehensive bibliographic review reveals further advancements in the field. For instance, Khan and Saleh [3] investigated the estimation problem of the mean in univariate normal distributions with unknown variance, incorporating uncertain non-sample prior information. This work highlights the potential benefits of utilizing prior knowledge in parameter estimation. Similarly, Singh [8] addressed the challenges of estimating variances and means in k-variate normal distributions when samples are subject to truncation or censoring on both sides for s variables (s<k). Both studies primarily employed the quadratic loss function to compute risk functions, emphasizing the need for more nuanced approaches.

    In more recent developments, Hamdaoui et al. [5] demonstrated the minimaxity of specific shrinkage estimators for the mean of an MNDs, focusing particularly on the risk ratios of the James-Stein estimator (JSE) and its positive-part version compared with the MLE under the balanced loss function (BLF). Their findings underscore the importance of using alternative loss functions to capture the performance of estimators more accurately.

    Benkhaled and Hamdaoui [1] further advanced this discussion by examining estimators that approximate the mean of the multivariate normal distribution, particularly when the variance is unknown. They proposed two categories of shrinkage estimators, with the first category focusing on the minimax properties of these estimators and identifying the optimal estimator, known as the JSE. Their second category included an estimator that exceeded the performance of the JSE, assessed through risk functions calculated using the BLF.

    Gomez-Deniz [10] investigated the application of the BLF in actuarial statistics, particularly within the framework of credibility theory. His work highlights the advantages of the BLF in Bayesian estimation, demonstrating its ability to balance accuracy and robustness. While his research primarily focused on actuarial applications, the underlying principles of the BLF are widely applicable in statistical estimation. Our study extends the application of the BLF to Bayesian estimation of the mean in the MNDs. Unlike Gomez-Deniz's actuarial focus, we explore how the BLF can enhance the estimation accuracy in multivariate settings, particularly when prior information is available. The novelty of this manuscript lies in its emphasis on the balanced loss function and the integration of prior information within a Bayesian framework, which can significantly enhance the estimation accuracy and robustness. By leveraging the strengths of both Bayesian and shrinkage approaches, we aim to provide a comprehensive solution to the estimation of mean parameters in MNDs. This shift not only broadens the applicability of the BLF but also enables more refined risk assessments by moving beyond the commonly used quadratic loss function.

    The structure of the paper is as follows: Section 2 presents the preliminary results and theoretical foundations that will be utilized throughout the study. In Section 3, we detail our primary findings and the performance of the proposed estimators. Section 4 focuses on a simulation study conducted to validate our results and assesses the practical implications of our findings. Section 4 presents an application of the estimators on a real-world problem. Finally, we conclude the paper with a discussion of our results and suggestions for future research directions.

    This manuscript addresses the estimation of an unknown parameter μ within the framework of the model Z|μNq(μ,σ2Iq). The prior distribution for μ is assumed to be μNq(η,ρ2Iq), where the value of σ2 is unknown and is estimated using the statistic S2σ2χ2n. The hyperparameters η and ρ2 may be known or unknown.

    To evaluate the performance of the introduced estimators, we employ the BLF, which can be defined as follows: The following holds for all estimators Λ of the parameter μ

    Lω(Λ,μ)=ωΛΛ02+(1ω)Λμ2, 0ω<1, (2.1)

    where Λ0 represents the target estimator of μ, ω corresponds to the weight assigned to the nearness value of Λ to Λ0, and 1ω represents the weighting factor attributed to the accuracy of the estimation component.

    We will denote the risk function of the estimator Λ under loss (2.1) as

    Rω(Λ,μ)=E(Lω(Λ,μ)), (2.2)

    and the Bayesian risk as

    Rω,b(Λ,σ2,η,ρ2)=Eμ(Rω(Λ,μ)). (2.3)

    The MLE of μ is commonly known to be Z:=Λ0. The risk function of the MLE with respect to the loss function (2.1) is equal to Rω(Z,μ)=(1ω)qσ2. Furthermore, the MLE is both minimax and inadmissible when q3. Hence, any estimator that improves upon it is also minimax.

    Under the model defined above, we recall some known results of the Bayes estimator. From Lindley and Smith [7], we assume that the posterior is considered to be Gaussian

    μ|ZNq(η+ρ2ρ2+σ2(Zη),σ2ρ2ρ2+σ2Iq).

    Thus, the Bayes estimator of μ is

    ΛB(Z)=E(μ|Z)=(1σ2ρ2+σ2)(Zη)+η. (2.4)

    In order to compute the expectation functions of a variable following a non central chi-square distribution, we recall the following definition:

    Definition 2.1. Let us assume Uχ2q(λ), representing a non central chi-square distribution with q degrees of freedom and a non centrality parameter λ. The probability density function of U is expressed as follows:

    f(z)=+k=0eλ2(λ2)kk!z(q/2)+k1ez/2Γ(q2+k)2(q/2)+k, 0<z<+.

    The expression on the right-hand side of this equation corresponds to the formula

    +k=0eλ2(λ2)kk!χ2q+2k.

    In light of this definition, we can infer that if Uχ2q(λ), where χ2q+2k denotes the density of the central χ2 distribution with q+2k degrees of freedom. Therefore for any integrable function ϕ:R+R, we have the following relationship:

    E[ϕ(U)]=Eχ2q(λ)[ϕ(U)]=+0ϕ(z)χ2q(λ)dz=+k=0[+0ϕ(z)χ2q+2k(0)dz]eλ2(λ2)kk!=+k=0[+0ϕ(z)χ2q+2kdz]P(λ2;dk). (2.5)

    In this context, where P(λ2;dk) represents the Poisson distribution with the parameter value being equal to λ2, and χ2q+2k denotes the central chi-square distribution with degrees of freedom equal to q+2k, we can introduce the subsequent lemma.

    Lemma 2.1. Let Uχ2q(λ) be a non central chi-square with q degrees of freedom and let λ be the non centrality parameter. Then for 0r<q2,

    E(Ur)=E[(χ2q(λ))r]=E[(χ2q+2K)r]=2rE(Γ(q2r+K)Γ(q2+K)),

    where K has a Poisson distribution with the mean λ2.

    We would like to recall the subsequent lemmas, provided by [2,9] and Hamdaoui et al. [4], which will be frequently utilized in the subsequent analysis.

    Lemma 2.2. Assume that Y is a real random variable following a standard normal distribution N(0,1), and let g:RR be an indefinite integral of a Lebesgue measurable function g, which can be considered to be the derivative of g. Furthermore, suppose that E|g(Y)|<+. Then the following statement holds true:

    E[Yg(Y)]=E(g(Y)).

    Lemma 2.3. For any real function h such that the expectation E(h(χ2q(λ))χ2q(λ)) exists, we can establish the following relationship:

    E{h(χ2q(λ))χ2q(λ)}=qE{h(χ2q+2(λ))}+2λE{h(χ2q+4(λ))}.

    The following lemma shows a lower bound and an upper bound of the expectation of the functions f1(u)=1u+α (respectively f2(u)=1(u+α)2), where α is a strictly positive real number, relative to the random variable χ2n+2 (respectively, to the χ2n+4 random variable).

    Lemma 2.4. Let Vχ2n+2 be a central chi-square with n+2 degrees of freedom and let Wχ2n+4 be a central chi-square with n+4 degrees of freedom. For any real α>0, we have

    1n+2+αE(1V+α)=Eχ2n+2(1u+α)1n+α, (2.6)

    and

    1(n+4+α)2E(1(W+α)2)=Eχ2n+4[1(u+α)2]1(n+α)2. (2.7)

    In this section, we introduce novel estimators for the mean parameter μ based on both the MLE and the Bayes estimator presented in (2.4). We then investigate their minimaxity properties and the asymptotic behavior of their risk ratios relative to the MLE when both the dimensionality of the parameter space q and the sample size n approach infinity. Our main results are presented in two distinct parts. First, we consider the same model as described above, assuming that the hyperparameters η and ρ2 are known. Second, we examine the same model, but with the hyperparameter η being known and the hyperparameter ρ2 being unknown.

    Now, let Z|μNq(μ,σ2Iq) and μNq(η,ρ2Iq), where the value of σ2 is unknown and is estimated by the statistic S2σ2χ2n. The hyperparameters η,ρ2 are known.

    Hamdaoui et al. [4] showed that the statistic S2S2+n ρ2  is an asymptotically unbiased estimator of the ratio σ2ρ2+σ2. If we substitute the ratio σ2σ2+ρ2  in Formula (2.4) with the estimator S2S2+nρ2, we can introduced a new estimator derived from the Bayes estimator, which can be expressed as

    ΛDB,γ(Z,S2)=(1γS2S2+nρ2)(Zη)+η. (3.1)

    Proposition 3.1. Under the BLF Lω, the Bayesian risk of the estimator ΛDB,γ(Z,S2) demonstrated in (3.1) is

    Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)=(1ω)qσ2[1+γ2n(n+2)1ω(1+ρ2σ2)Eχ2n+4(u+nρ2σ2)2]2γn(1ω)qσ2[Eχ2n+2(u+nρ2σ2)1]. (3.2)

    Proof. By utilizing the risk function linked to the BLF defined in (2.1), we derive the following expression:

    Rω(ΛDB,γ(Z,S2);μ)=ωE(ΛDB,γ(Z,S2)Z2)+(1ω)E(ΛDB,γ(Z,S2)μ2).

    By exploiting the independence between two random variables S2 and Z, we arrive at the following result:

    E(ΛDB,γ(Z,S2)Z2)=E(γS2S2+nρ2(Zη)2)=γ2E(S2S2+nρ2)2E(Zη2),

    and

    E(ΛDB,γ(Z,S2)μ2)=E(ZγS2S2+nρ2(Zη)μ2)=E(Zμ2)+γ2E(S2S2+nρ2)2EZη22γσ2E(S2S2+nρ2)E[Zμσ,Zησ].

    Thus,

    (Zη)|μNq(μη,σ2Iq)Zη2σ2|μχ2q(μη2σ2)E(Zη2)=σ2(q+μη2σ2),
    Zμσ2χ2qE(Zμσ2)=E(χ2q)1σ2E(Zμ2)=Eχ2q(u)E(Zμ2)=qσ2

    and

    E[Zμσ,Zησ]=E[Zμσ,Zμσ+μησ]=E[Zμσ,Zμσ]+E[Zμσ,μησ]=E(Zμ2σ2)+(μησ)tE(Zμσ)=E(χ2q)=q

    because E(Zμσ)=0.

    Therefore

    Rω(ΛDB,γ(Z,S2);μ)=ωγ2E(S2S2+nρ2)2(qσ2+μη2)+(1ω)[qσ2+γ2E(S2S2+nρ2)2E(Zη2)2γσ2E(S2S2+nρ2)q].

    Using Lemma 2.3 we get,

    E(S2S2+nρ2)2=Eχ2n(uu+nρ2σ2)2=n(n+2)Eχ2n+4(u+nρ2σ2)2

    and

    E(S2S2+nρ2)=Eχ2n(uu+nρ2σ2)=nEχ2n+2(u+nρ2σ2)1.

    Therefore,

    Rω(ΛDB,γ(Z,S2);μ)=(1ω)qσ2+γ2n(n+2)Eχ2n+4(u+nρ2σ2)2(qσ2+μη2)2γn(1ω)qσ2Eχ2n+2(u+nρ2σ2)1.

    Thus, the Bayesian risk of the estimator ΛDB,γ(Z,S2) under the BLF Lω is

    Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)=Eμ(Rω(ΛDB,γ(Z,S2);μ))=(1ω)qσ2+γ2n(n+2)Eχ2n+4(u+nρ2σ2)2(qσ2+Eμμη2)2γn(1ω)qσ2Eχ2n+2(u+nρ2σ2)1=(1ω)qσ2[1+γ2n(n+2)1ω(1+ρ2σ2)Eχ2n+4(u+nρ2σ2)2]2γn(1ω)qσ2[Eχ2n+2(u+nρ2σ2)1].

    Theorem 3.2. Assume the estimator ΛDB,γ(Z,S2) as defined in (3.1). If

    0γ2(1ω)nn+2,

    then under the BLF Lω given in (2.1), the estimator ΛDB,γ(Z,S2) dominates the MLE and thus it is minimax.

    Proof. From Proposition 3.1, a sufficient condition for the estimator ΛDB,γ(Z,S2) to dominate the MLE is

    γ2n(n+2)1ω(1+ρ2σ2)Eχ2n+4(u+nρ2σ2)22γnEχ2n+2(u+nρ2σ2)10, (3.3)

    which is equivalent to

    0γ2(1ω)(n+2)(1+ρ2σ2)Eχ2n+2(u+nρ2σ2)1Eχ2n+4(u+nρ2σ2)2.

    Using Lemma 2.4, we have the following inequality:

    Eχ2n+2(u+nρ2σ2)1Eχ2n+4(u+nρ2σ2)2n2(1+ρ2σ2)22+n(1+ρ2σ2).

    Therefore, we can deduce that if

    0γ2(1ω)(n+2)(1+ρ2σ2)n2(1+ρ2σ2)22+n(1+ρ2σ2)=2(1ω)nn+2,

    the inequality (3.3) is satisifed, and thus the estimator ΛDB,γ(Z,S2) has a risk smaller than that of the MLE. This last point indicates that ΛDB,γ(Z,S2) is a minimax estimator.

    In this section, we investigate the asymptotic behavior of the risk ratio of our estimator ΛDB,γ(Z,S2) in response to the MLE when 0γ2(1ω)nn+2.

    If we take the real constant α (0<α2), our aims is to show that for any γ, such as γ=α(1ω)nn+2, the risk ratio Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ) tends to a value inferior to one, when n tends to infinity.

    Theorem 3.3. Assume the estimator ΛDB,γ(Z,S2) defined in 3.1. If γ=α(1ω)nn+2 where 0<α2, then

    limnRω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)=1(2α)α(1ω)+ρ2σ21+ρ2σ21.

    Proof. From Proposition 3.1, we have

    Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)=1+α2(1ω)n3(n+2)(1+ρ2σ2)Eχ2n+4(u+nρ2σ2)22α(1ω)n2n+2Eχ2n+2(u+nρ2σ2)1.

    Using Lemma 2.4, we get

    Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)1+α2(1ω)n(n+2)(1+ρ2σ2)2α(1ω)n2n+21(2+n(1+ρ2σ2)),

    and

    Rω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)1+α2(1ω)n3n+2(1+ρ2σ2)(4+n(1+ρ2σ2))22α(1ω)n(n+2)(1+ρ2σ2).

    By passing to the limit, we get

    limnRω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)1+α2(1ω)1+ρ2σ22α(1ω)1+ρ2σ2=1(2α)α(1ω)+ρ2σ21+ρ2σ2,

    and

    limnRω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)1+α2(1ω)1+ρ2σ22α(1ω)(1+ρ2σ2)=1(2α)α(1ω)+ρ2σ21+ρ2σ2.

    Thus,

    limnRω,b(ΛDB,γ(Z,S2);η,ρ2,σ2)Rω(Z,μ)=1(2α)α(1ω)+ρ2σ21+ρ2σ21,

    because 0(2α)α1 and 0<1ω1.

    The algorithm of estimator Type 1 can be summarized as follows:

    Algorithm 1 Estimator Type 1
    1: Input: Sample data Z=(Z1,Z2,,Zn). from Nq(μ,σ2Iq)
    2: Known values for q, η, and ρ2.
    3: Step 1: Calculate sample mean
                            ˉZ=1qqi=1Zi
    4: Step 2: Calculate a sample variance
                        S2=1q1qi=1(ZiˉZ)2
    5: Step 3: Determine a shrinkage factor
                            k=1γS2S2+nρ2
      Note: This step assumes q>2 for minimax properties.
    6: Step 4: Compute estimator Type 1
                        ΛDB,γ(Z,S2)=k(ˉZη)+η
    7: Output: Return estimator Type 1

    Next, we consider the model Z|μNq(μ,σ2Iq) and μNq(η,ρ2Iq), where the parameter σ2 is unknown and is also estimated by the statistic S2σ2χ2n, and the hyper parameter η is known and the hyper parameter ρ2 is unknown.

    Hamdaoui et al. [4] showed that the statistic q2n+2S2Zη2 is an asymptotically unbiased estimator of the ratio σ2ρ2+σ2. Therefore, if we substitute the ratio σ2σ2+ρ2  in Formula (2.4) with the estimator q2n+2S2Zη2, we can consider the new estimator derived from the Bayes estimator, expressed as

    ΛDB,β(Z,S2)=(1βS2Zη2)(Zη)+η (3.4)

    where the positive real parameter β can depend on n and q.

    Proposition 3.4. The Bayesian risk of the estimator ΛDB,β(Z,S2) given in Eq (3.4), which has been derived from the BLF illustrated in (2.1), can be expressed as

    Rω,b(ΛDB,β(Z,S2);η,ρ2,σ2)=(1ω)qσ2+βnσ4ρ2+σ2[β(n+2)q22(1ω)]. (3.5)

    Proof. By utilizing the risk function linked to the BLF Lω defined in (2.1), we derive the following expression:

    Rω(ΛDB,β(Z,S2);μ)=ωE(ΛDB,β(Z,S2)Z2)+(1ω)E(ΛDB,β(Z,S2)μ2).

    In the one hand, since the random variables Z and S2 are independent, we can have:

    E(ΛDB,β(Z,S2)Z2)=E(βS2Zη2(Zη)2)=β2E(S2)2E(1Zη2)=β2σ4E(S2σ2)21ρ2+σ2E(1Zη2ρ2+σ2).

    As S2σ2χ2n, the marginal distribution of Z is :ZNq(η,(ρ2+σ2)Iq).By employing Definition 2.1, we get

    E(S2σ2)=n,  E(S2σ2)2=n(n+2)

    and

    E(1Zη2ρ2+σ2)=E(1χ2q)=1q2.

    Thus

    E(ΛDB,β(Z,S2)Z2)=n(n+2)β2q2σ4ρ2+σ2. (3.6)

    On the other hand, we have

    E(ΛDB,β(Z,S2)μ2)=E(ZβS2Zη2(Zη)μ2)=EZμ2+β2E(S2)2E(1Zη2)2βσ2E(S2σ2)E[Zμσ,1Zη2σ2Zησ]. (3.7)

    Let W=(W1,W2,...,Wq)t=Zμσ. It is clear that W|μNq(0,Iq), and thus

    E(Zμσ,1Zη2σ2(Zησ))=E(W,1W+μησ2(W+μησ))=qi=1E[Wi(1qj=1(Wj+μjηjσ)2(Wi+μiηiσ))].

    From Stein's Lemma 2.2, we have

    qi=1E[Wi(1qj=1(Wj+μjηjσ)2(Wi+μiηiσ))]=qi=1E[Wi(1j=1q(Wj+μjηjσ)2(Wi+μiηiσ))]=qi=1E[Wi(Wi+μiηiσ(W1+μ1η1σ)2++(Wi+μiηiσ)2++(Wq+μqηqσ)2)]=qi=1E[qj=1(Wj+μjηjσ)22(Wi+μiηiσ)(Wi+μiηiσ)(qj=1(Wj+μjηjσ)2)2]=qi=1E[W+μησ22(Wi+μiηiσ)2(W+μησ2)2]=E(qi=1[1W+μησ22(Wi+μiηiσ)2(W+μησ2)2])=E[qW+μησ22qi=1(Wi+μiηiσ)2(W+μησ2)2]=(q2)E[1W+μησ2]=(q2)σ2ρ2+σ2E[1Zη2ρ2+σ2]=σ2ρ2+σ2. (3.8)

    This is in line with Formulas (3.6)–(3.8) and the fact that

    EZμ2=σ2E(χ2q)=qσ2,  E(S2)2=n(n+2)σ4

    and

    E(1Zη2)=(ρ2+σ2)E(ρ2+σ2χ2q)=ρ2+σ2q2.

    Theorem 3.5. Relative to the BLF Lω given in (2.1), a sufficient condition for the estimator ΛDB,β(Z,S2) defined in (3.4) to be minimax is

    0β2(1ω)(q2)n+2.

    Proof. On the basis of Proposition 3.4, we can readily determine that a sufficient condition for the estimator ΛDB,β(Z,S2) to be minimax is,

    [β(n+2)q22(1ω)]0,

    which is equivalent to

    0β2(1ω)(q2)n+2.

    If we used the convexity of the risk function Rω,b(ΛDB,β(Z,S2);η,ρ2,σ2) with respect to β, it becomes apparent that the optimal value of β that minimizes the function is ˆβ=(1ω)(q2)n+2. By substituting the value of β with ˆβ in Eq (3.4), we then derive the best estimator in the class of estimators Λβ, which is defined as

    ΛDB,ˆβ(Z,S2)=(1(1ω)(q2)n+2S2Zη2)(Zη)+η. (3.9)

    Furthermore, its risk function related to the BLF is given by

    Rω,b(ΛDB,ˆβ(Z,S2))=(1ω)qσ2(1ω)2q2n+2nσ4ρ2+σ2Rω(Z,μ). (3.10)

    We can then deduce that the estimator ΛDB,ˆβ(Z,S2) dominates the MLE; thus, it is minimax.

    In this section, we examine the asymptotic behavior of the risk ratio of the estimator ΛDB,ˆβ(Z,S2) relative to the MLE as the parameter q approaches infinity while the parameter n remains fixed, and when both parameters q and n simultaneously tend to infinity. Consequently, we infer that the estimator ΛDB,ˆβ(Z,S2) exhibits a stable minimax property even in the scenario where the dimension of the parameter space q tends to infinity while the sample size n remains fixed, as well as when both the dimension of the parameter space q and the sample size n simultaneously tend to infinity.

    Theorem 3.6.

    1. limqRω,b(ΛDB,ˆβ(Z,S2))Rω(Z,μ)=(n+2)ρ2+(2+nω)σ2(n+2)(ρ2+σ2)1,
    2. limn,qRω,b(ΛDB,ˆβ(Z,S2))Rω(Z,μ)=ρ2+ωσ2ρ2+σ21.

    Proof. From Formula (3.10), we have

    Rω,b(ΛDB,ˆβ(Z,S2))Rω(Z,μ)=1(1ω)qσ2[(1ω)qσ2(1ω)2q2n+2nσ4ρ2+σ2]=1[(1ω)q2qnn+2σ2ρ2+σ2].

    We can then easily deduce that

    1. limqRω,b(ΛDB,ˆβ(Z,S2))Rω(Z,μ)=(n+2)ρ2+(2+nω)σ2(n+2)(ρ2+σ2)1,
    2. limn,qRω,b(ΛDB,ˆβ(Z,S2))Rω(Z,μ)=ρ2+ωσ2ρ2+σ21.

    The algorithm of estimator Type 2 can be summarized as follows:

    Algorithm 2 Estimator Type 2
    1: Input: Sample data Z=(Z1,Z2,,Zn). from Nq(μ,σ2Iq), where q, η are known.
    2: Step 1: Calculate sample mean and sample variance
                      ˉZ=1qqi=1Zi and S2=1q1qi=1(ZiˉZ)2
    3: Step 2: Determine shrinkage factor
                          k=1βS2Zη2
      Note: This step assumes q>2 for minimax properties.
    4: Step 3: Compute estimator Type 2
                       ΛDB,β(Z,S2)=k(ˉZη)+η
    5: Output: Return estimator Type 2

    The aim of this simulation was to prove the effectiveness of this study by comparing the estimators ΛDB,ˆγ(Z,S2), (ˆγ=(1ω)nn+2), and ΛDB,ˆβ(Z,S2) with the MLE Z.

    First, we apply the risk ratio of the estimator ΛDB,ˆγ(Z,S2) to the MLE Z, Rω(ΛDB,ˆγ(Z,S2);η,ρ2,σ2)Rω(Z,μ) as a function of ρ2σ2 for various values of n and ω.

    We see that in Figure 1, the risk ratio is less than 1, i.e., the shrinkage estimator ΛDB,ˆγ(Z,S2) dominates the natural estimator Z. We also see that if ω increases, the improvement decreases, and become negligible whenever ω is near to one. Rasing the value of n gives a small improvement.

    Figure 1.  Graph of the risk ratio Rω,b(ΛDB,ˆγ(Z,S2);η,ρ2,σ2)Rω(Z,μ)) as a function of x=ρ2σ2 for n=5,10,30; and ω=0.1,0.4,0.7.

    The same applies for the estimator ΛDB,ˆβ(Z,S2), except that here, the risk ratio depends on n, ω, and the dimension of the parameter space q. As shown in Figures 24 the shrinkage estimator ΛDB,ˆβ(Z,S2) is better than the natural estimator Z. We also see the same for the estimator ΛDB,ˆγ(Z,S2). Furthermore, we note that the influence of the parameter q on the improvement is the same as that of the parameter n.

    Figure 2.  Graph of the risk ratio Rω,b(ΛDB,ˆβ(Z,S2);η,ρ2,σ2)Rω(Z,μ)) as a function of x=ρ2σ2 for n=5; q=3,7,10; and ω=0.1,0.4,0.7.
    Figure 3.  Graph of the risk ratio Rω,b(ΛDB,ˆβ(Z,S2);η,ρ2,σ2)Rω(Z,μ) as a function of x=ρ2σ2 for n=10; q=3,7,10; and ω=0.1,0.4,0.7.
    Figure 4.  Graph of the risk ratio Rω,b(ΛDB,ˆβ(Z,S2);η,ρ2,σ2)Rω(Z,μ)) as a function of x=ρ2σ2 for n=30; q=3,7,10; and ω=0.1,0.4,0.7.

    Secondly, we plot the risk difference ΥRω=Rω,b(ΛDB,ˆγ(Z,S2);η,ρ2,σ2)Rω(Z,μ) and ΥRω=Rω,b(ΛDB,ˆβ(Z,S2);η,ρ2,σ2)Rω(Z,μ) of the estimators ΛDB,ˆγ(Z,S2) and ΛDB,ˆβ(Z,S2) to the MLE Z as a function of x=σ2 and y=ρ2 for n=7, q=15, and various values of ω (ω=0.1,0.3,0.6,0.9). In Figures 5 and 6, we see that the risk differences ΥRω and ΥRω are entirely negative. This indicates that the estimators ΛDB,ˆγ(Z,S2) and ΛDB,ˆβ(Z,S2) dominate the maximum likelihood estimator Z.

    Figure 5.  Plots of the risk difference ΥRω for various sets of ω, n=7, and q=15.
    Figure 6.  Plots of ths risk difference ΥRω for various sets of ω, n=7, and q=15.

    As in the simulation section, we demonstrate the effectiveness of the new estimators by comparing the estimators \(\Lambda_{DB, \widehat{\gamma}}(Z, S^{2}) \) (where \(\widehat{\gamma} = \frac{(1-\omega)n}{n+2} \)) and \(\Lambda_{DB, \widehat{\beta}}(Z, S^{2}) \) with the MLE \(Z \) on a real dataset. The dataset that we will use is the Batting dataset from the Lahman package in the R program, which contains baseball statistics. Our focus is on evaluating the effectiveness of the new estimators in estimating batting averages. We create a scenario where we sample only a limited number of observations from a group of players.

    We then compare how accurately the MLE and the proposed estimators predict the actual values. For this comparison, we analyze the risk ratio of each estimator to the MLE using the BLF in (2.1) as the loss function. Table 1 shows that for various values of size, dimension, and ω, both new estimators achieve a ratio that is less than one. This means that the estimations derived from estimator Type 1 and estimator Type 2 are more efficient than the MLE and reduce the variance through shrinkage.

    Table 1.  The risk ratio of each estimator to the MLE using the BLF in (2.1) as the loss function with different values of ω and n.
    Estimator Type 1 n=10 n=30 n=100
    ω=0.1 0.6760075 0.8234310 0.9208014
    ω=0.4 0.7840050 0.8822873 0.9472009
    ω=0.9 0.8920025 0.9411437 0.9736005
    Estimator Type 2 n=10, q=3 n=30, q=3 n=100, q=3
    ω=0.1 0.9920534 0.9987638 0.9998857
    ω=0.4 0.9947023 0.9991759 0.9999238
    ω=0.9 0.9973511 0.9995879 0.9999619
    Estimator Type 2 n=10, q=20 n=30, q=20 n=100, q=20
    ω=0.1 0.8420707 0.9757862 0.9979490
    ω=0.4 0.8947138 0.9838575 0.9986327
    ω=0.9 0.9473569 0.9919287 0.9993163

     | Show Table
    DownLoad: CSV

    In this research article, we conducted a thorough investigation into the estimation of the mean of a multivariate normal distribution from a Bayesian perspective using the BLF. Our primary objective was to assess the performance of the proposed estimator in comparison with the conventional MLE through a comprehensive simulation study. To begin with, we focused on establishing the minimaxity property of the modified Bayes estimator and analyzing the behavior of the risk ratios between this estimator and the MLE. Specifically, we examined the scenario where both the sample size n and the dimension of the parameter space p tend to infinity. By investigating the asymptotic behavior of these risk ratios, we gained valuable insights into the relative efficiency of the modified Bayes estimator and the MLE in settings with large samples. This analysis provided a deeper understanding of the estimator's performance and its robustness under different scaling conditions. Furthermore, we explored the domination of a class of estimators that encompassed the empirical modified Bayes estimator over the MLE. Through rigorous mathematical proofs and empirical evidence, we demonstrated the superiority of this class of estimators over the MLE. This dominance result further supported the efficacy and advantages of the proposed Bayesian approach within the context of estimating the mean of a multivariate normal distribution.

    Overall, our research shed light on the Bayesian estimation of the mean under the BLF framework and showcased the performance of the proposed modified Bayes estimator. The extensive simulation study, the application, and theoretical analysis provided valuable insights into the estimator's behavior and its superiority over the traditional MLE. These findings contribute to the existing literature on Bayesian estimation and offer practical implications for data analysis in scenarios where accurate estimation of the mean is of paramount importance. The inconvenience of our constructed estimators can be deduced from Figures 14: We see that if the values of the variance σ2 of the variable Z|μNq(μ,σ2Iq), exceed 1/4 of the variance ρ2 of the prior distribution μNq(η,ρ2Iq), the improvement of the our proposed estimators to the MLE becomes negligible, and this shows the poor performance of the suggested estimators. This limitation requires further investigation, which we will address in future work.

    A. Alahmadi and A. Benkhaled conceived the idea; A. Alahmadi developed the theory and performed the computations; A. Benkhaled and W. Almutiry verified the analytical methods and supervised the findings of this paper. All authors discussed the results and contributed to the final manuscript.

    The author would like to thank the Deanship of Scientific Research at Shaqra University for supporting this work.

    The authors declare that there is no conflict of interest regarding the publication of this paper.



    [1] A. Benkhaled, A. Hamdaoui, A study of minimax shrinkage estimators dominating the James-Stein estimator under the balanced loss function, Open Math., 20 (2022), 1–11. https://doi.org/10.1515/math-2022-0008 doi: 10.1515/math-2022-0008
    [2] G. Casella, J. T. Hwang, Limit expressions for the risk of James-Stein estimators, Can. J. Stat., 10 (1982), 305–309. https://doi.org/10.2307/3556196 doi: 10.2307/3556196
    [3] S. Khan, A. Saleh, On the comparison of the pre-test and shrinkage estimators for the univariate normal mean, Stat. Pap., 42 (2001), 451–473. https://doi.org/10.1007/s003620100073 doi: 10.1007/s003620100073
    [4] A. Hamdaoui, A. Benkhaled, N. Mezouar, Minimaxity and limits of risks ratios of shrinkage estimators of a multivariate normal mean in the bayesian case, Stat., Optim. Inf. Comput., 8 (2020), 507–520. https://doi.org/10.19139/soic-2310-5070-735 doi: 10.19139/soic-2310-5070-735
    [5] A. Hamdaoui, W. Almutiry, M. Terbeche, A. Benkhaled, Comparison of risk ratios of shrinkage estimators in high dimensions, Mathematics, 52 (2022), 1–14. https://doi.org/10.3390/math10010052 doi: 10.3390/math10010052
    [6] W. James, C. Stein, Estimation with quadratic loss, Proc 4th Berkeley Symp, Math. Statist.Prob., Univ of California Press, Berkeley, 1 (1961), 361–379.
    [7] D. V. Lindley, A. F. M. Smith, Bayes estimates for the linear model (with discussion), J. Roy. Statist. Soc., 34 (1972), 1–41.
    [8] N. Singh, Estimation of parameters of a multivariate normal population from truncated and censored samples, J. R. Stat. Soc. B, 22 (1960), 307–311. https://doi.org/10.1111/j.2517-6161.1960.tb00376.x doi: 10.1111/j.2517-6161.1960.tb00376.x
    [9] C. Stein, Estimation of the mean of a multivariate normal distribution, Ann. Statist., 9 (1981), 1135–1151. https://doi.org/10.1214/aos/1176345632 doi: 10.1214/aos/1176345632
    [10] E. Gómez-Déniz, A generalization of the credibility theory obtained by using the weighted balanced loss function, Insur. Math. Econ., 42 (2008), 850–854. https://doi.org/10.1016/j.insmatheco.2007.09.002 doi: 10.1016/j.insmatheco.2007.09.002
    [11] C. Stein, Inadmissibilty of the usual estimator for the mean of a multivariate normal distribution, Proc. 3th Berkeley Symp, Math. Statist. Prob. Univ. of California Press, Berkeley, 1 (1956), 197–206.
    [12] R. Yang, J. O. Berger, Estimation of a covariance matrix using the reference prior, Ann. Statist., 22 (1994), 1195–1211. https://doi.org/10.1214/aos/1176325625 doi: 10.1214/aos/1176325625
  • Reader Comments
  • © 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(384) PDF downloads(38) Cited by(0)

Figures and Tables

Figures(6)  /  Tables(1)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog