Processing math: 78%
Research article Special Issues

Monitoring river water quality through predictive modeling using artificial neural networks backpropagation

  • Received: 02 July 2024 Revised: 05 August 2024 Accepted: 07 August 2024 Published: 19 August 2024
  • Predicting river water quality in the Special Region of Yogyakarta (DIY) is crucial. In this research, we modeled a river water quality prediction system using the artificial neural network (ANN) backpropagation method. Backpropagation is one of the developments of the multilayer perceptron (MLP) network, which can reduce the level of prediction error by adjusting the weights based on the difference in output and the desired target. Water quality parameters included biochemical oxygen demand (BOD), chemical oxygen demand (COD), total suspended solids (TSS), dissolved oxygen (DO), total phosphate, fecal coliforms, and total coliforms. The research object was the upstream, downstream, and middle parts of the Oya River. The data source was secondary data from the DIY Environment and Forestry Service. Data were in the form of time series data for 2013–2023. Descriptive data results showed that the water quality of the Oya River in 2020–2023 was better than in previous years. However, increasing community and industrial activities can reduce water quality. This was concluded based on the prediction results of the ANN backpropagation method with a hidden layer number of 4. The prediction results for period 3 in 2023 and period 1 in 2024 are that 1) the concentrations of BOD, fecal coli, and total coli will increase and exceed quality standards, 2) COD and TSS concentrations will increase but will still be below quality standards, 3) DO and total phosphate concentrations will remain constant and still on the threshold of quality standards. The possibility of several water quality parameters increasing above the quality standards remains, so the potential for contamination of the Oya River is still high. Therefore, early prevention of river water pollution is necessary.

    Citation: Muhammad Andang Novianta, Syafrudin, Budi Warsito, Siti Rachmawati. Monitoring river water quality through predictive modeling using artificial neural networks backpropagation[J]. AIMS Environmental Science, 2024, 11(4): 649-664. doi: 10.3934/environsci.2024032

    Related Papers:

    [1] Sohaib Ahmad, Sardar Hussain, Javid Shabbir, Muhammad Aamir, M. El-Morshedy, Zubair Ahmad, Sharifah Alrajhi . Improved generalized class of estimators in estimating the finite population mean using two auxiliary variables under two-stage sampling. AIMS Mathematics, 2022, 7(6): 10609-10624. doi: 10.3934/math.2022592
    [2] Khazan Sher, Muhammad Ameeq, Sidra Naz, Basem A. Alkhaleel, Muhammad Muneeb Hassan, Olayan Albalawi . Developing and evaluating efficient estimators for finite population mean in two-phase sampling. AIMS Mathematics, 2025, 10(4): 8907-8925. doi: 10.3934/math.2025408
    [3] Tolga Zaman, Cem Kadilar . Exponential ratio and product type estimators of the mean in stratified two-phase sampling. AIMS Mathematics, 2021, 6(5): 4265-4279. doi: 10.3934/math.2021252
    [4] Xuechen Liu, Muhammad Arslan . A general class of estimators on estimating population mean using the auxiliary proportions under simple and two phase sampling. AIMS Mathematics, 2021, 6(12): 13592-13607. doi: 10.3934/math.2021790
    [5] Olayan Albalawi . Estimation techniques utilizing dual auxiliary variables in stratified two-phase sampling. AIMS Mathematics, 2024, 9(11): 33139-33160. doi: 10.3934/math.20241582
    [6] Amber Yousaf Dar, Nadia Saeed, Moustafa Omar Ahmed Abu-Shawiesh, Saman Hanif Shahbaz, Muhammad Qaiser Shahbaz . A new class of ratio type estimators in single- and two-phase sampling. AIMS Mathematics, 2022, 7(8): 14208-14226. doi: 10.3934/math.2022783
    [7] Yasir Hassan, Muhammad Ismai, Will Murray, Muhammad Qaiser Shahbaz . Efficient estimation combining exponential and ln functions under two phase sampling. AIMS Mathematics, 2020, 5(6): 7605-7623. doi: 10.3934/math.2020486
    [8] Hleil Alrweili, Fatimah A. Almulhim . Estimation of the finite population mean using extreme values and ranks of the auxiliary variable in two-phase sampling. AIMS Mathematics, 2025, 10(4): 8794-8817. doi: 10.3934/math.2025403
    [9] Sohaib Ahmad, Sardar Hussain, Muhammad Aamir, Faridoon Khan, Mohammed N Alshahrani, Mohammed Alqawba . Estimation of finite population mean using dual auxiliary variable for non-response using simple random sampling. AIMS Mathematics, 2022, 7(3): 4592-4613. doi: 10.3934/math.2022256
    [10] Saman Hanif Shahbaz, Aisha Fayomi, Muhammad Qaiser Shahbaz . Estimation of the general population parameter in single- and two-phase sampling. AIMS Mathematics, 2023, 8(7): 14951-14977. doi: 10.3934/math.2023763
  • Predicting river water quality in the Special Region of Yogyakarta (DIY) is crucial. In this research, we modeled a river water quality prediction system using the artificial neural network (ANN) backpropagation method. Backpropagation is one of the developments of the multilayer perceptron (MLP) network, which can reduce the level of prediction error by adjusting the weights based on the difference in output and the desired target. Water quality parameters included biochemical oxygen demand (BOD), chemical oxygen demand (COD), total suspended solids (TSS), dissolved oxygen (DO), total phosphate, fecal coliforms, and total coliforms. The research object was the upstream, downstream, and middle parts of the Oya River. The data source was secondary data from the DIY Environment and Forestry Service. Data were in the form of time series data for 2013–2023. Descriptive data results showed that the water quality of the Oya River in 2020–2023 was better than in previous years. However, increasing community and industrial activities can reduce water quality. This was concluded based on the prediction results of the ANN backpropagation method with a hidden layer number of 4. The prediction results for period 3 in 2023 and period 1 in 2024 are that 1) the concentrations of BOD, fecal coli, and total coli will increase and exceed quality standards, 2) COD and TSS concentrations will increase but will still be below quality standards, 3) DO and total phosphate concentrations will remain constant and still on the threshold of quality standards. The possibility of several water quality parameters increasing above the quality standards remains, so the potential for contamination of the Oya River is still high. Therefore, early prevention of river water pollution is necessary.



    In large scale surveys, it is a usual practice to prefer multi-stage sampling to estimate the population characteristics over single-stage sampling. The main purpose to use multi-stage sampling is the clear reduction in the cost of survey operations even if estimates derived from multi-stage sampling are likely to be less efficient than those of the single-stage sampling. Sukhatme et al. [1] advised some ratio and regression type estimators in two-stage sampling using single auxiliary variable when first stage units are of unequal or equal sizes. The suitable use of auxiliary information in estimation stage results as a considerable reduction in the mean square error of the estimator. By making use of auxiliary information, Srivastava and Garg [2] proposed separate-type estimator for the estimation of population mean in two-stage sampling design. Taking inspiration from Srivastava and Garg [2], Koyuncu and Kadilar [3] and Jabeen et al. [4] proposed separate type estimator under two-stage sampling. In the literature, the use of two-phase sampling under two-stage sampling design is not well documented. Saini and Bahl [5] proposed estimator under two-stage sampling design using double sampling for stratification and multi-auxiliary information. A generalized ratio cum product estimator for population mean in simple random sampling was developed by Singh et al. [6]. Shabbir [7] produces the estimators of population mean under stratified two phase sampling.

    In the literature ([8,9,10,11]), the use of two phase sampling under two stage sampling design is not well documented. Also the shrinkage estimators have been discussed several times in literature by considering the unbiasedness of the estimators but no one has discussed the situation when the property of unbiasedness is not fulfill that is very common in real life applications. In order to fill this gap, we are motivated to produce two-exponential estimators under two stage two phase sampling that is discussed in section 2. Also we will discuss general shrinkage estimator in section 3. We will compare both estimators mathematically and by using real population data in section 4 and 5. Finally the conclusion will be discussed in section 6.

    Let a population consists of N first stage units, each containing Mi second-stage units where i = 1, 2, …, N. Let a first-stage sample of size n (N) is selected and subsequently a second-stage sample of mi(1)(Mi) units is selected and information on some auxiliary variables say xij(1) is taken. Here it is assumed that each first stage unit/cluster is of different size so each cluster is assigned a weight ηi=MiˉM to it. Let a sub-sample (second-phase sample) of mi(2) units is selected from mi(1)(first-phase sample) such that mi(2)Mi. Let mi(2) units are observed so as to collect information regarding study variable yij(2) and auxiliary variablesxij(2). Let ˉXts=1NNi=11ˉMMiˉXi. be the mean in the population and ˉXi=1MiMij=1xij be the mean of ith first stage unit in the population. Let ˉxts(1)=1nni=1ηiˉxi(1) and ˉxi(2)=1mi(2)mi(2)j=1xij(2) respectively are the means of first-phase and second-phase sample in two-stage sampling where ˉxi(1) and ˉxi(2) be the sample means of first phase and second phase in ith stage. Let S2xb=1N1Ni=1(ηiˉXiˉXts)2 be the population variance between fsu’s and S2xwi=1Mi1ni=1(xijˉXi.)2 be the population variance within first stage units. Similarly these notations can be defined for other variables. Further we consider that the selection of units at each stage (or phase) has been made by simple random sampling without replacement.

    We propose a generalized estimator by considering the exponential relationship as:

    tG=ˉytsexp(α(1aˉxts(1)(ˉX+(a1)ˉxts(1))))exp(β(1bˉzts(2)(ˉZ+(b1)ˉzts(2)))), (1)

    Where (a, b) are constants to be determined such that the mean square error is minimum, (α, β) are known constants takes the value (0, 1, -1) to produce different ratio-type and product-type estimators as presented in Table B1 (Appendix B).

    To derive the bias and mean square error, we proceed as follows:

    ˉyts(2)ˉYtsˉYts=e0(2),ˉxts(1)ˉXtsˉXts=e1(1),ˉzts(2)ˉZtsˉZts=e2(2),ˉxts(2)ˉXtsˉXts=e1(2),ˉzts(1)ˉZtsˉZts=e2(1)

    Further we assume that E(e0(2))=E(e1(2))=E(e2(2))=E(e2(1))=E(e1(1))=0, and some expectations under two-stage sampling design are obtained in order to obtain the bias and mean square error as,

    E(e0(2)2)=C020(2),E(e1(2)2)=C200(2),E(e2(2)2)=C002(2),E(e1(1)2)=C200(1),E(e2(1)2)=C002(1),E(e1(2)e1(1))=C200(1)E(e0(2)e1(2))=C110(2),E(e1(2)e2(2))=C101(2),E(e0(2)e2(2))=C011(2),E(e0(2)e1(1))=C110(1),E(e1(2)e2(1))=C101(1),E(e0(2)e2(1))=C011(1)WhereC020(2)=1ˉY2ts{(1n1N)S2yb+1nNNi=1η2i(1mi(2)1Mi)S2ywi,},C200(1)=1ˉX2ts{(1n1N)S2xb+1nNNi=1η2i(1mi(1)1Mi)S2xwi}C200(2)=1ˉX2ts{(1n1N)S2xb+1nNNi=1η2i(1mi(2)1Mi)S2xwi},C002(1)=1ˉZ2ts{(1n1N)S2zb+1nNNi=1η2i(1mi(1)1Mi)S2zwi},C002(2)=1ˉZ2ts{(1n1N)S2zb+1nNNi=1η2i(1mi(2)1Mi)S2zwi},C110(1)=1ˉYtsˉXts{(1n1N)ρxybSybSxb+1nNNi=1η2i(1mi(1)1Mi)ρxywiSywiSxwi},C110(2)=1ˉYsˉXs{(1n1N)ρxybSybSxb+1nNNi=1η2i(1mi(2)1Mi)ρxywiSywiSxwi},C101(1)=1ˉZtsˉXts{(1n1N)ρxzbSzbSxb+1nNNi=1η2i(1mi(1)1Mi)ρxzwiSzwiSxwi},C101(2)=1ˉZtsˉXts{(1n1N)ρxzbSzbSxb+1nNNi=1η2i(1mi(2)1Mi)ρxzwiSzwiSxwi},C200(1)=1ˉZtsˉXts{(1n1N)SxbSxb+1nNNi=1η2i(1mi(1)1Mi)SxwiSxwi}C011(2)=1ˉZtsˉYts{(1n1N)ρyzbSzbSyb+1nNNi=1η2i(1mi(2)1Mi)ρyzwiSzwiSywi},C011(1)=1ˉZtsˉYts{(1n1N)ρyzbSzbSyb+1nNNi=1η2i(1mi(1)1Mi)ρyzwiSzwiSywi}} (2)

    Using (2) we can express (1) to derive the bias and mean square error as,

    tG=ˉYts(1+e0(2))exp[αae1(1)(1+(a1)ae1(1))1]exp[βbe2(2)(1+(b1)be2(2))1], (3)

    If |e1(1)|<1, we expand the series, (1+(a1)ae1(1))1 and (1+(b1)be2(2))1 up to the order n-1, we get,

    tG=ˉYts(1+e0(2))exp[αae1(1)(1(a1)ae1(1)+(a1)2a2e1(1)2+...)]exp[βbe2(2)(1(b1)be2(2)+(b1)2b2e2(2)2+...)], (4)

    If the contribution of terms involving powers in e0(2), e1(1) and e2(2) higher than two is negligible, we have;

    tG=ˉYts(1+e0(2))[1αae1(1)+α(a1)2a2e21(1)+12α2a2e21(1)][1βbe2(2)+β(b1)2b2e22(2)+12β2b2e22(2)], (5)

    or

    tGˉYts=ˉYts[e0(2)αae1(1)+α(a1)a2e21(1)+αβabe1(1)e2(2)+α22a2e21(1)βbe2(2)+β22b2e2(2)2+β(b1)b2e2(2)2βbe2(2)e0(2)αae1(1)e0(2)], (6)

    In order to get the bias, we take expectation on (6) and get,

    Bias(tG)=ˉYts[α22a2C200(1)+α(a1)a2C200(1)+β22b2C002(2)+β(b1)b2C002(2)βbC011(2)αaC110(1)+αaβbC101(1)], (7)

    To get the mean square error of the estimator, we take square and retain terms up to first order of e’s then we take expectation of (6) and we obtain,

    MSE(tG)=ˉY2ts(C020(2)+(αa)2C200(1)+(βb)2C002(2)2αaC110(1)2βbC011(2)+2αaβbC101(1)), (8)

    For the following optimal value of the constants a and b, we achieve the minimum Variance among the class of proposed generalized estimator,

    a=α(C200(1)C002(2)C101(2)2)C110(1)C002(2)C011(1)C101(2),andb=β(C200(1)C002(2)C101(2)2)C200(1)C011(1)C110(1)C101(2), (9)

    We obtain minimum mean square error as,

    minMSE(tG)=ˉY2ts(C020(2)C110(1)2C002(2)+C011(1)2C200(1)2C110(1)C101(2)C011(1)C200(1)C002(2)C101(2)2), (10)

    We observe from (10) that proposed generalized estimator gives us more precise results under the optimal conditions, as compare to its class of the estimators.

    From (7)–(10), we get expressions of the bias, mean square error, optimal values and minimum mean square error for the exponential-type estimators presented in Table B1 as class of estimators. Some of the results are discussed as,

    ⅰ). For α=1,β=1, we get a class of exponential-type ratio-cum-ratio estimators given in Table B1 and expressions of the mean square error and bias for these estimators are given as,

    MSE(tjerr)={ˉY2ts(C020(2)+1a2C200(1)+1b2C002(2)21aC110(1)21bC011(2)+21abC101(1))j(G)=1,3,5},andBias(tjerr)={ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)1aC110(1))j(G)=1,3,5},} (11)

    Substitution of the different values of a and b yield the mean square error and bias for the estimators belongs to the class of exponential-type ratio-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential ratio-cum-ratio estimators is obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C200(1)C002(2)C101(2)2C200(1)C011(1)C110(1)C101(2).

    ⅱ). For α=1,β=1, we get exponential-type product-cum-product estimators given in Table B1.The mean square error of tG are expressed using as,

    MSE(tjepp)={ˉY2ts(C020(2)+1a2C200(1)+1b2C002(2)+21aC110(1)+21bC011(2)+21abC101(1)),j(G)=2,4,6}andBiastjepp)={ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)+1aC110(1)),j(G)=2,4,6}) (12)

    Substitution of the different values of “a” andb” yield the Mean Square Error and bias for the estimators belongs to the class of exponential-type ratio-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential product-cum-product estimators is obtained as,

    a=C101(2)2C200(1)C002(2)C110(1)C002(2)C011(1)C101(2)andb=C101(2)2C200(1)C002(2)C200(1)C011(1)C110(1)C101(2).

    ⅲ). For α=1,β=1, we get exponential-type product-cum-ratio estimators given in Table B1.The mean square error of tG is expressed as,

    MSE(tjerp)={ˉY2ts(C020(2)+1aC200(1)+1bC002(2)+21aC110(1)21bC011(2)21abC101(1)),j(G)=7,9,11}andBias(tjerp)={ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)+1aC110(1)),j(G)=7,9,11}) (13)

    Substitution of the different values of “a” and “b” yield the mean square error and bias for the estimators belongs to the class of exponential-type ratio-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential product-cum-ratio estimators are obtained as,

    a=C101(2)2C200(1)C002(2)C110(1)C002(2)C011(1)C101(2)andb=C200(1)C011(1)C110(1)C101(2)C200(1)C002(2)C101(2)2.

    ⅳ). For α=1,β=1, we get exponential-type ratio-cum-product estimators given in Table A1. The mean square error of tG is expressed as,

    MSE(tjepr)={ˉY2ts(C020(2)+1aC200(1)+1bC002(2)21aC110(1)+21bC011(2)21a1bC101(1)),j(G)=8,10,12}andBias(tjepr)={ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)+1bC101(2)1aC110(1)),j(G)=8,10,12}) (14)

    Substitution of the different values of “a” and “b” yield the mean square error and bias for the estimators belongs to the class of exponential-type ratio-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential ratio-cum-product estimators, are obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C1012C200(1)C002(2)C200(1)C011(1)C110(1)C101(2).

    It is to mention that minimum mean square error of tjepr, tjepr, tjepr, and tjepr in (11)–(14) will be same but the conditions under which these attain minimum value are not unique.

    In literature, the shrinkage estimators have been discussed several times considering the property that the estimators are unbiased. In many practical situations the property of unbiasedness is not met so consequently the existing literature will not be helpful. Such deviations from the said property have motivated us to consider the shrinkage estimators following the property of biasedness. Important results are presented in theorem and subsequently proof is provided in Appendix A.

    For any general estimator t, let ts be a general shrinkage estimator for estimating any population characteristic T (e.g., population mean) which is unknown to us,

    ts=λt (15)

    A general form for the bias and mean square error of a general shrinkage estimator up to the first order of approximation may be given as,

    Bias(ts)=(λ1)T+Bias(t) (16)

    and

    MSE(ts)=T2(λ1)2+λ2MSE(t)+2Tλ(λ1)Bias(t) (17)

    And form of the optimal shrinkage estimator in (15) along with minimum bias and minimum mean square error will be as,

    ts=(T2+MSE(t)+2TBias(t))T(T+Bias(t)t (18)
    Bias(topts)=((T2+MSE(t)+2TBias(t))T(T+Bias(t)1)T+Bias(t) (19)

    and

    minMSE(ts)=T2T2(T+Bias(t)T2+MSE(t)+2TBias(t) (20)

    Proof: For details see Appendix A.

    By using (15), generalized form of a shrinkage class of estimator is,

    tGs=λtG

    where

    tG=ˉytsexp(α(1aˉxts(1)(ˉXts+(a1)ˉxts(1))))exp(β(1bˉzts(2)(ˉZts+(b1)ˉzts(2)))) (21)

    For different choices of the values of the constants α,β,a,b, we get different family of shrinkage estimators as given in Table B2 (See Appendix).

    Following the Theorem, we can directly write the mean square error and bias expressions of tGs using (7) and (8) respectively as,

    Bias(tGs)=λˉYts[α(a1)a2C200(1)+α22a2C200(1)+β(b1)b2C002(2)+β22b2C002(2)βbC101(2)αaC110(1)]+ˉYts(λ1),or
    Bias(tGs)=λBias(tGe)+ˉYts(λ1), (22)

    and

    MSE(tGs)=λ2ˉY2ts(C020(2)+(αa)2C200(1)+(βb)2C002(2)2αaC110(1)2βbC011(2)+2αaβbC101(1))+(λ1)2ˉYts2+2λ(λ1)ˉYts[α(a1)a2C200(1)+α22a2C200(1)+β(b1)b2C002(2)+β22b2C002(2)βbC101(2)αaC110(1)]

    or

    MSE(tGs)=λ2MSE(tGe)+(λ1)2ˉYts2+2ˉYtsλ(λ1)MSE(tGe) (23)

    The optimal value of λ, which minimize the expression (23) is obtained as,

    λ=(T2+MSE(tGe)+2TBias(tGe))T(T+Bias(tGe) (24)

    Using (23), the optimal estimator for the shrinkage family of estimator (tGs) is written as

    tGs=ˉyts(ˉYts2+MSE(tGe)+2ˉYtsBias(tGe))ˉYts(ˉYts+Bias(tGe)exp(α(1aˉxts(1)(ˉXts+(a1)ˉxts(1))))exp(β(1bˉzts(2)(ˉZts+(b1)ˉzts(2))))} (25)

    The expression for minimum value of min MSE(tGs) is obtained as,

    minMSE(tGs)=ˉY2tsˉY2ts(ˉYts+Bias(tGe))2[ˉY2ts+MSE(tGe)+2ˉYtsBias(tGe)]}minMSE(tGs)=ˉY2ts[1(1+Bias(ta)ˉYts)2[1+MSE(ta)ˉY2ts+2Bias(ta)ˉYts]] (26)

    From (22)–(26), we get expressions of the bias, mean square error, optimal values and minimum mean square error for the exponential-type estimators presented in Table B2 (See Appendix B) as class of estimators. Some of the results are discussed as,

    ⅰ). For α=1,β=1, we get a class of exponential-type ratio-cum-ratio estimators given in Table B2 (See Appendix) and expressions of the mean square error and bias for these estimators are given as,

    MSE(tjerrs)={ˉY2ts[1(1+Bias(tierr)ˉYts)2[1+MSE(tierr)ˉY2ts+2Bias(tierr)ˉY2ts]]j(G)=1,3,5}andBias(tjerrs)={ˉYts(λ(Bias(tjerr)ˉYts)+(λ1))j(G)=1,3,5} (27)

    Substitution of the different values of a and b produce the mean square error and bias for the estimators belongs to the class of exponential-type ratio-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential product-cum-ratio estimators are obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C200(1)C002(2)C101(2)2C200(1)C011(1)C110(1)C101(2),λ=(T2+MSE(tGe)+2TBias(tGe))T(T+Bias(tGe),whereMSE(tG)=ˉYts2[C020(2)+1a2C200(1)+1b2C002(2)21aC110(1)21bC011(2)+21a1bC101(1)].Bias(tG)=ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)1aC110(1))

    ⅱ). For α=1,β=1, we get exponential-type product-cum-product estimators given in Table B2 (See Appendix B).The mean square error of tGs is expressed using as,

    MSE(tjepps)={ˉY2ts[1(1+Bias(tierr)ˉYts)2[1+MSE(tierr)ˉY2ts+2Bias(tierr)ˉY2ts]]j(G)=2,4,6}andBias(tjepps)={ˉYts(λBias(tjepps)ˉYts+(λ1))j(G)=2,4,6} (28)

    Substitution of the different values of a and b yield the mean square error and bias for the estimators belongs to the class of exponential-type product-cum-product estimators. The optimal values which lead to minimum mean square error for the class of exponential product-cum-ratio estimators are obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C200(1)C002(2)C101(2)2C200(1)C011(1)C110(1)C101(2),λ=(T2+MSE(tGe)+2TBias(tGe))T(T+Bias(tGe),whereMSE(tG)=ˉYts2[C020(2)+1a2C200(1)+1b2C002(2)21aC110(1)21bC011(2)+21a1bC101(1)].BiastG)=ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)1aC110(1))

    ⅲ). For α=1,β=1, we get exponential-type product-cum-ratio estimators given in Table B2. The mean square error of tGs is expressed as,

    MSE(tjerps)={ˉY2ts[1(1+Bias(tierr)ˉYts)2[1+MSE(tierr)ˉY2ts+2Bias(tierr)ˉY2ts]]j(G)=7,9,11}andBias(tjerps)={ˉYts(λBias(tjerps)ˉYts+(λ1))j(G)=7,9,11} (29)

    Substitution of the different values of a and b produce the mean square error and bias for the estimators belongs to the class of exponential-type ratio-cum-product estimators. The optimal values which lead to minimum mean square error for the class of exponential ratio-cum-product estimators, are obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C200(1)C002(2)C101(2)2C200(1)C011(1)C110(1)C101(2),λ=(T2+MSE(tGe)+2TBias(tGe))T(T+Bias(tGe),whereMSE(tGe)=ˉYts2[C020(2)+1a2C200(1)+1b2C002(2)21aC110(1)21bC011(2)+21a1bC101(1)].BiastGe)=ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)1aC110(1))

    ⅳ). For α=1,β=1, we get exponential-type ratio-cum-product estimators given in Table B2. The mean square error of tGs is expressed as,

    MSE(tjeprs)={ˉY2ts[1(1+Bias(tierr)ˉYts)2[1+MSE(tierr)ˉY2ts+2Bias(tierr)ˉY2ts]]j(G)=8,10,12}andBias(tjeprs)={ˉYts(λBias(tjeprs)ˉYts+(λ1))j(G)=8,10,12} (30)

    Substitution of the different values of a and b yield the mean square error and bias for the estimators belongs to the class of exponential-type product-cum-ratio estimators. The optimal values which lead to minimum mean square error for the class of exponential product-cum-ratio estimators, are obtained as,

    a=C200(1)C002(2)C101(2)2C110(1)C002(2)C011(1)C101(2)andb=C200(1)C002(2)C101(2)2C200(1)C011(1)C110(1)C101(2),λ=(T2+MSE(tGe)+2TBias(tGe))T(T+Bias(tGe),whereMSE(tG)=ˉYts2[C020(2)+1a2C200(1)+1b2C002(2)21aC110(1)21bC011(2)+21a1bC101(1)].Bias(tG)=ˉYts(a1a2C200(1)+12a2C200(1)+b1b2C002(2)+12b2C002(2)1bC101(2)1aC110(1))

    It is to mention that minimum mean square error of tjeprs, tjeprs, tjeprs, and tjeprs in (27)–(30) will be same but the conditions under which these attain minimum value are not unique.

    In this section, some efficiency conditions have been derived in terms of mean square error and bias. The efficiency comparison of proposed two-exponential general shrinkage estimator (see section 3) and two-exponential estimator (see section 2) with unbiased estimator in two-stage sampling. In order to derive the efficiency conditions, consider the following notations,

    A1=C020(2)+(αa)2C200(1)+(βb)2C002(2)2(αa)C110(1)2(βb)C011(2)+2(αa)(βb)C101(1)
    A2=[α(a1)a2C200(1)+α22a2C200(1)+β(b1)b2C002(2)+β22b2C002(2)βbC101(2)αaC110(1)]
    A3=C020(2)+C200(1)+C002(2)2C110(1)2C011(2)+2C101(1)
    A4=(C020(2)C110(1)2C002(2)+C011(1)2C200(1)2C110(1)C101(2)C011(1)C200(1)C002(2)C101(2)2)
    A5=[λ2+2λ(λ1)(1αC200(1)+12C200(1))]
    A6=[λ2(C110(1)βbC101(!))+λ(λ1)(C200(1)C110(1))]
    A7=[λ2((βb)2C002(2)2βbC011(2))+(λ1)2+2λ(λ1)(β(b1)b2C002(2)+(βb)2C002(2)βbC101(1))]

    The efficiency conditions may be written as:

    ⅰ. MSE(tGs)MSE(ˉyts)0 if

    min((1A2)±((1A2)2(1+A1+2A2)(C020(2)))(1A2))λmax((1A2)±((1A2)2(1+A1+2A2)(C020(2)))(1A2)), (31)

    ⅱ. MSE(tGs)MSE(tr)0

    min((1A2)±((1A2)2(1+A1+2A2)(A3))(1A2))λmax((1A2)±((1A2)2(1+A1+2A2)(A3))(1A2)) (32)

    ⅲ. MSE(tGs)MSE(tG)0

    min((1A2)±((1A2)2(1+A1+2A2)(A1))(1A2))λmax((1A2)±((1A2)2(1+A1+2A2)(A1))(1A2)) (33)

    ⅳ. MSE(tGs)minMSE(tG)0

    min((1A2)±((1A2)2(1+A1+2A2)(A4))(1A2))λmax((1A2)±((1A2)2(1+A1+2A2)(A4))(1A2)) (34)

    ⅰ. MSE(tGs)MSE(ˉyts)0 if

    min(A6±A62A5(A7C020(2))αA5)1amax(A6±A62A5(A7C020(2))αA5) (35)

    ⅱ. MSE(tGs)MSE(tr)0

    min(A6±A62A5(A7A3)αA5)1amax(A6±A62A5(A7A3)αA5) (36)

    Substituting different values of (α,β,a,b) in above equations, we get different mathematical comparisons for the estimators given in Table B2 (See Appendix B). Also similar comparison is obtained if we assume b as unknown and λ,a is known.

    For the demonstration of the performance of proposed two-exponential general shrinkage estimator, we take real population consists of four clusters with unequal first stage units. The description about the populations is given in Table B1 in Appendix B. The mean square error and percentage relative efficiency values for each of the estimators are given in Table C2.For this population, correlations within the clusters are positive (see Table C1). It is therefore the population is applicable only for ratio/ratio-type estimators.

    The performance of the proposed two-exponential general shrinkage estimator and two-exponential estimator under two stage two phase sampling has been expressed in form of their mean square error and percentage relative efficiency values in Table C2. From Table C2 we see that the minimum mean square error of two-exponential general shrinkage estimator tGs that is equal to the mean square error of t5s. The percentage relative efficiency demonstrate the same result as the percentage relative efficiency of the proposed two-exponential general shrinkage is high among the family of proposed two-exponential general shrinkage and t5s is the same efficient as tGs.

    We also see that the two-exponential general shrinkage estimator (section 3) is better to be used as compare to two-exponential estimator (section 2). The performance of t1s is better than t1, similarly t3s performs better than t3. So it is concluded that for this particular population the use of two-exponential general shrinkage estimator in two stage two phase sampling produce better results.

    Finally from the above empirical results, it is concluded that the performance of the two-exponential general shrinkage estimator (section 3) is higher as compare to proposed two-exponential estimator in two stage two phase sampling. So the two-exponential general shrinkage estimator (section 3) is acceptable for the real life application in two stage two phase sampling design.

    We are grateful to the anonymous reviewers for their valuable comments and suggestions in improving the manuscript.

    The authors declare no conflict of interest.

    Proof of Theorem. Let t be any general estimator of some population parameter T in two stage sampling for which a general shrinkage estimator ts is defined as,

    ts=λt, (37)

    By definition the bias of ts is

    Bias(ts)=λE(t)T

    or

    =λE(tT)+(λ1)T

    or

    =λBias(t)+(λ1)T (38)

    By definition the MSE of ts may be defined as,

    MSE(ts)=E(tsT)2

    or

    =E(λtT)2

    consider t=T(1+eT), with E(eT)=0 and E(eT2)=Var(t)T2=MSE(t)T2.

    =E{λ(T(1+eT))T}2

    or

    =T2E{λ2(1+eT)2+12λ(1+eT)}

    or

    MSE(ts)=T2{λ2(1+MSE(t)T2)+12λ} (39)

    We may also write (B-3) as

    MSE(ts)=λ2MSE(t)+(λ1)2T2

    or

    MSE(ts)=λ2T2C2t+(λ1)2T2WhereC2t=MSE(t)T2 (40)

    Where Ct is a coefficient of variation of an estimator t.

    Partially differentiating (40) w.r.t λ and simplifying, we get:

    λ=11+T2MSE(t), (41)

    or

    λ=11+C2t, (42)

    Now using (41) in (15), (16) and (17) respectively, we may get,

    ts=t1+T2MSE(t) (43)
    Bias(ts)=11+T2MSE(t)[Bias(t)+T2MSE(t)T] (44)
    MSEmin(ts)=MSE(t)1+T2MSE(t) (45)

    Alternatively we may also produce the expressions in (44) to (46) as,

    ts=t1+C2t (46)
    Bias(ts)=11+C2t(Bias(t)+C2tT) (47)
    MSEmin(ts)=T2C2t1+C2t (48)
    Table B1.  Some members of the proposed two exponential estimator (tG).
    Ratio-cum-Ratio estimator
    α=1,β=1
    Product -cum-product estimator
    α=1,β=1
    a b
    t1err=ˉyts(2)exp(ˉXtsˉxts(1)ˉXts+ˉxts(1))exp(ˉZtsˉzts(2)ˉZts+ˉzts(2)) t2epp=ˉyts(2)exp(ˉxts(1)ˉXtsˉxts(1)+ˉXts)exp(ˉzts(2)ˉZtsˉzts(2)+ˉZts) 2 2
    t3err=ˉyts(2)exp(ˉXtsˉxts(1)ˉXts)exp(ˉZtsˉzts(2)ˉZts) t4epp=ˉyts(2)exp(ˉxts(1)ˉXtsˉXts)exp(ˉzts(2)ˉZtsˉZts) 1 1
    t5err=ˉyts(2)exp(ˉXtsˉxts(1)ˉXts+(a1)ˉxts(1))exp(ˉZtsˉzts(2)ˉZts+(b1)ˉzts(2)) t6epp=ˉyts(2)exp(ˉxts(1)ˉXtsˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) a b
    Product -cum-ratio estimator
    α=1,β=1
    Ratio-cum-product estimator
    α=1,β=1
    a b
    t7epr=ˉytsexp(ˉXtsˉxts(1)ˉXts+ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+ˉzts(2)) t8erp=ˉyts(2)exp(ˉxts(1)ˉXtsˉxts(1)+ˉXts)exp(ˉzts(2)ˉZtsˉzts(2)+ˉZts) 2 2
    t9epr=ˉyts(2)exp(ˉXtsˉxts(1)ˉXts)exp(ˉzts(1)ˉZtsˉZts) t10erp=ˉyts(2)exp(ˉxts(1)ˉXtsˉXts)exp(ˉzts(2)ˉZtsˉZts) 1 1
    t11epr=ˉyts(2)exp(ˉXtsˉxts(1)ˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) t12erp=ˉyts(2)exp(ˉxts(1)ˉXtsˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) a b

     | Show Table
    DownLoad: CSV
    Table B2.  Some members of the generalized shrinkage estimator (tGs).
    Ratio-cum-product estimator
    α=1,β=1
    Product -cum-product estimator
    α=1,β=1
    a λ b
    t1errs=λˉytsexp(ˉXtsˉxts(1)ˉXts+ˉxts(1))exp(ˉZtsˉzts(2)ˉZts+ˉzts(2)) t2epps=λˉytsexp(ˉxts(1)ˉXtsˉXts+ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+ˉzts(2)) 2 λ 2
    t3errs=λˉytsexp(ˉXtsˉxts(1)ˉXts)exp(ˉZtsˉzts(2)ˉZts) t4epps=λˉytsexp(ˉxts(1)ˉXtsˉXts)exp(ˉzts(2)ˉZtsˉZts) 1 λ 1
    t5erps=λˉytsexp(ˉXtsˉxts(1)ˉXts+(a1)ˉxts)exp(ˉZtsˉzts(2)ˉZts+(b1)ˉzts(2)) t6epps=λˉytsexp(ˉxts(1)ˉXtsˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) a λ b
    Product -cum-ratio estimator
    α=1,β=1
    Ratio-cum-product estimator
    α=1,β=1
    a λ b
    t7eprs=λˉytsexp(ˉXtsˉxts(1)ˉXts+ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+ˉzts(2)) t8erps=λˉyts(2)exp(ˉxts(1)ˉXtsˉxts(1)+ˉXts)exp(ˉzts(2)ˉZtsˉzts(2)+ˉZts) 2 λ 2
    t9eprs=λˉyts(2)exp(ˉXtsˉxts(1)ˉXts)exp(ˉzts(1)ˉZtsˉZts) t10erps=λˉyts(2)exp(ˉxts(1)ˉXtsˉXts)exp(ˉzts(2)ˉZtsˉZts) 1 λ 1
    t11eprs=λˉyts(2)exp(ˉXtsˉxts(1)ˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) t12erps=λˉyts(2)exp(ˉxts(1)ˉXtsˉXts+(a1)ˉxts(1))exp(ˉzts(2)ˉZtsˉZts+(b1)ˉzts(2)) a λ b

     | Show Table
    DownLoad: CSV
    Table C1.  Data Statistics for Population-Ⅰ.
    Population-Ⅰ (unequal fsu’s)
    Cluster 1 2 3 4
    Mi 18 14 12 20
    mi 9 7 6 10
    ˉYi 25.77722 22.79286 28.43500 23.0905
    ˉXi 51.06389 46.49700 67.00217 57.11855
    Cyi2 0.58025 0.39297 0.34783 0.31545
    Cxi2 0.43322 0.29984 0.41947 0.40689
    ρi1 0.88373 0.83895 0.82425 0.82113

     | Show Table
    DownLoad: CSV
    Table C2.  MSEs and PREs of the adapted and suggested class of estimators.
    Estimators MSE PRE With new fraction
    ¯ys 24.44649 100
    t1err 11.66857 209.51
    t3err 20.27969 120.5467
    t5err 11.46787 213.1738
    t1errs 11.58499 211.0229
    t3errs 18.92219 129.1974
    t5errs 11.25906 217.1317
    tGopts 11.25906 217.1317

     | Show Table
    DownLoad: CSV


    [1] Utama JP, Syafrudin, Nugraha WD (2015) Penentuan Daya tampung beban pencemaran BOD dan Fecal Coliform Sungai Plumbon Kota Semarang Dengan software QUAL2E. J Teknik Lingkungan 4: 1–9.
    [2] Warsito B, Sumiyati S, Yasin H, et al. (2021) Evaluation of river water quality by using hierarchical clustering analysis, IOP Conference Series: Earth and Environmental Science. https://doi.org/10.1088/1755-1315/896/1/012072 doi: 10.1088/1755-1315/896/1/012072
    [3] Zubaidah T, Karnaningroem N, Slamet A (2018) K-means method for clustering water quality status on the rivers of Banjarmasin, Indonesia. ARPN J Eng Appl Sci 13. https://doi.org/10.31227/osf.io/s9n2u doi: 10.31227/osf.io/s9n2u
    [4] Novianta MA, Syafrudin S, Warsito B (2023) K-means clustering for Grouping Rivers in DIY based on water quality parameters. JUITA J Inform 11: 155–163. https://doi.org/10.30595/juita.v11i1.16986 doi: 10.30595/juita.v11i1.16986
    [5] Di Z, Chang M, Guo P (2019) Water quality evaluation of the Yangtze River in China using machine learning techniques and data monitoring on different time scales. Water 11: 339. https://doi.org/10.3390/w11020339 doi: 10.3390/w11020339
    [6] Mohammadrezapour O, Kisi O, Pourahmad F (2020) Fuzzy c-means and K-means clustering with genetic algorithm for identification of homogeneous regions of groundwater quality. Neural Comput Appl 32: 3763–3775. https://doi.org/10.1007/s00521-018-3768-7 doi: 10.1007/s00521-018-3768-7
    [7] Widyastuti M, Marfai MA (2004) Kajian daya tampung sungai gajahwong thdp beban pencemaran. Majalah Geografi Indonesia. 8: 81–97.
    [8] Riza H, Santoso EW, Tejakusuma IG, et al. (2020) Utilization of artificial intelligence to improve flood disaster mitigation. Jurnal Sains dan Teknologi Mitigasi Bencana 30: 1–11. https://doi.org/10.29122/jstmb.v15i1.4145 doi: 10.29122/jstmb.v15i1.4145
    [9] Makubura R, Meddage DPP, Azamathulla HM, et al. (2022) A simplified mathematical formulation for water quality index (WQI): A case study in the Kelani River Basin, Sri Lanka. Fluids 7: 147. https://doi.org/10.3390/fluids7050147 doi: 10.3390/fluids7050147
    [10] Madhushani C, Dananjaya K, Ekanayake IU, et al. (2024) Modeling streamflow in non-gauged watersheds with sparse data considering physiographic, dynamic climate, and anthropogenic factors using explainable soft computing techniques. J Hydrol 631: 130846. https://doi.org/10.1016/j.jhydrol.2024.130846 doi: 10.1016/j.jhydrol.2024.130846
    [11] Perera UAKK, Coralage DTS, Ekanayake IU, et al. (2024) A new frontier in streamflow modeling in ungauged basins with sparse data: A modified generative adversarial network with explainable AI. Results Eng 21: 101920. https://doi.org/10.1016/j.rineng.2024.101920 doi: 10.1016/j.rineng.2024.101920
    [12] Pohan S, Warsito B, Suryono S (2020) Backpropagation artificial neural network for prediction plant seedling growth. J Phys 1524: 012147. https://doi.org/10.1088/1742-6596/1524/1/012147 doi: 10.1088/1742-6596/1524/1/012147
    [13] Purba RA, Samsir S, Siddik M, et al. (2020), The optimization of backpropagation neural networks to simplify decision making, IOP Conference Series: Materials Science and Engineering, 830: 022091. https://doi.org/10.1088/1757-899X/830/2/022091 doi: 10.1088/1757-899X/830/2/022091
    [14] Nurjaya IK, Estananto E, Murti A (2022) Pemodelan sistem kendali suhu otomatis pada Smart poultry farm menggunakan metode jaringan saraf tiruan. eProc Eng 9.
    [15] Haekal M, Wibowo WC (2023) Prediksi kualitas air sungai menggunakan metode pembelajaran mesin: Studi kasus Sungai Ciliwung: Prediction of river water quality using machine learning methods: Ciliwung River case study. Jurnal Teknologi Lingkungan 24: 273–282. https://doi.org/10.55981/jtl.2023.795 doi: 10.55981/jtl.2023.795
    [16] Mustafa HM, Mustapha A, Hayder G, et al. (2021) Applications of iot and artificial intelligence in water quality monitoring and prediction: A review, In 2021 6th International Conference on inventive computation technologies (ICICT), 968–975. https://doi.org/10.1109/ICICT50816.2021.9358675
    [17] DLHK DIY (2021) Dokumen Informasi Kinerja Pengelolaan Lingkungan Hidup Daerah (DIKPLHD) Provinsi DIY Tahun 2021, Yogyakarta.
    [18] Masruroh M (2020) Perbandingan metode regresi linear dan neural network backpropagation dalam prediksi nilai Ujian nasional siswa smp menggunakan software R. Joutica 5: 331–336. https://doi.org/10.30736/jti.v5i1.347 doi: 10.30736/jti.v5i1.347
    [19] Wanto A (2019) Prediksi produktivitas jagung di Indonesia sebagai upaya antisipasi impor menggunakan jaringan saraf tiruan backpropagation, SINTECH 2: 53–62. https://doi.org/10.31598/sintechjournal.v2i1.355 doi: 10.31598/sintechjournal.v2i1.355
    [20] Sudarsono A (2016) Jaringan syaraf tiruan untuk memprediksi laju pertambahan penduduk menggunakan metode backpropagation (studi Kasus di Kota Bengkulu). Jurnal Media Infotama 12. https://doi.org/10.37676/jmi.v12i1.273 doi: 10.37676/jmi.v12i1.273
    [21] Cynthia EP, Ismanto E (2017) Jaringan syaraf tiruan algoritma backpropagation dalam memprediksi ketersediaan komoditi pangan provinsi riau. RABIT Jurnal Teknologi Dan Sistem Informasi Univrab 2. https://doi.org/10.36341/rabit.v2i2.152 doi: 10.36341/rabit.v2i2.152
    [22] Aggarwal L, Sahoo BM (2018) Back propagation algorithm for computerized paper evaluation using neural network. IJSRST 4.
    [23] Bande S, Shete VV (2017) Smart flood disaster prediction system using IoT & neural networks, 2017 International Conference on Smart Technologies For Smart Nation (SmartTechCon), 189–194. https://doi.org/10.1109/SmartTechCon.2017.8358367 doi: 10.1109/SmartTechCon.2017.8358367
    [24] Mitra P, Ray R, Chatterjee R, et al. (2016) Flood forecasting using Internet of things and artificial neural networks, 2016 IEEE 7th Annual Information Technology, Electronics and Mobile Communication Conference (IEMCON), 1–5. https://doi.org/10.1109/IEMCON.2016.7746363 doi: 10.1109/IEMCON.2016.7746363
    [25] Alquisola GLV, Coronel DJA, Reolope BMF, et al. (2018) Prediction and visualization of the disaster risks in the Philippines using discrete wavelet transform (DWT), autoregressive integrated moving average (ARIMA), and artificial neural network (ANN), 2018 3rd International Conference on Computer and Communication Systems (ICCCS), 146–149. https://doi.org/10.1109/CCOMS.2018.8463238
    [26] Wu J, Ming H, Xu J, (2021) Research on intelligent disaster prevention and mitigation method in high flood risk area of river basin based on artificial neural network, 2021 7th International Conference on Hydraulic and Civil Engineering & Smart Water Conservancy and Intelligent Disaster Reduction Forum (ICHCE & SWIDR), 943–952. https://doi.org/10.1109/ICHCESWIDR54323.2021.9656352 doi: 10.1109/ICHCESWIDR54323.2021.9656352
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1411) PDF downloads(81) Cited by(2)

Figures and Tables

Figures(5)  /  Tables(3)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog