
Protein functions are closely related to their subcellular locations. At present, the prediction of protein subcellular locations is one of the most important problems in protein science. The evident defects of traditional methods make it urgent to design methods with high efficiency and low costs. To date, lots of computational methods have been proposed. However, this problem is far from being completely solved. Recently, some multi-label classifiers have been proposed to identify subcellular locations of human, animal, Gram-negative bacterial and eukaryotic proteins. These classifiers adopted the protein features derived from gene ontology information. Although they provided good performance, they can be further improved by adopting more powerful machine learning algorithms. In this study, four improved multi-label classifiers were set up for identification of subcellular locations of the above four protein types. The random k-labelsets (RAKEL) algorithm was used to tackle proteins with multiple locations, and random forest was used as the basic prediction engine. All classifiers were tested by jackknife test, indicating their high performance. Comparisons with previous classifiers further confirmed the superiority of the proposed classifiers.
Citation: Lei Chen, Ruyun Qu, Xintong Liu. Improved multi-label classifiers for predicting protein subcellular localization[J]. Mathematical Biosciences and Engineering, 2024, 21(1): 214-236. doi: 10.3934/mbe.2024010
[1] | Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Statistical analysis of stress–strength in a newly inverted Chen model from adaptive progressive type-Ⅱ censoring and modelling on light-emitting diodes and pump motors. AIMS Mathematics, 2024, 9(12): 34311-34355. doi: 10.3934/math.20241635 |
[2] | Mohamed A. H. Sabry, Ehab M. Almetwally, Osama Abdulaziz Alamri, M. Yusuf, Hisham M. Almongy, Ahmed Sedky Eldeeb . Inference of fuzzy reliability model for inverse Rayleigh distribution. AIMS Mathematics, 2021, 6(9): 9770-9785. doi: 10.3934/math.2021568 |
[3] | Amit Singh Nayal, Bhupendra Singh, Abhishek Tyagi, Christophe Chesneau . Classical and Bayesian inferences on the stress-strength reliability R=P[Y<X<Z] in the geometric distribution setting. AIMS Mathematics, 2023, 8(9): 20679-20699. doi: 10.3934/math.20231054 |
[4] | Neama Salah Youssef Temraz . Analysis of stress-strength reliability with m-step strength levels under type I censoring and Gompertz distribution. AIMS Mathematics, 2024, 9(11): 30728-30744. doi: 10.3934/math.20241484 |
[5] | Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat . Analysis of reliability index R=P(Y<X) for newly extended xgamma progressively first-failure censored samples with applications. AIMS Mathematics, 2024, 9(11): 32200-32231. doi: 10.3934/math.20241546 |
[6] | Hanan Haj Ahmad, Ehab M. Almetwally, Dina A. Ramadan . A comparative inference on reliability estimation for a multi-component stress-strength model under power Lomax distribution with applications. AIMS Mathematics, 2022, 7(10): 18050-18079. doi: 10.3934/math.2022994 |
[7] | Yichen Lv, Xinping Xiao . Grey parameter estimation method for extreme value distribution of short-term wind speed data. AIMS Mathematics, 2024, 9(3): 6238-6265. doi: 10.3934/math.2024304 |
[8] | Saurabh L. Raikar, Dr. Rajesh S. Prabhu Gaonkar . Jaya algorithm in estimation of P[X > Y] for two parameter Weibull distribution. AIMS Mathematics, 2022, 7(2): 2820-2839. doi: 10.3934/math.2022156 |
[9] | Naif Alotaibi, A. S. Al-Moisheer, Ibrahim Elbatal, Salem A. Alyami, Ahmed M. Gemeay, Ehab M. Almetwally . Bivariate step-stress accelerated life test for a new three-parameter model under progressive censored schemes with application in medical. AIMS Mathematics, 2024, 9(2): 3521-3558. doi: 10.3934/math.2024173 |
[10] | Essam A. Ahmed, Laila A. Al-Essa . Inference of stress-strength reliability based on adaptive progressive type-Ⅱ censing from Chen distribution with application to carbon fiber data. AIMS Mathematics, 2024, 9(8): 20482-20515. doi: 10.3934/math.2024996 |
Protein functions are closely related to their subcellular locations. At present, the prediction of protein subcellular locations is one of the most important problems in protein science. The evident defects of traditional methods make it urgent to design methods with high efficiency and low costs. To date, lots of computational methods have been proposed. However, this problem is far from being completely solved. Recently, some multi-label classifiers have been proposed to identify subcellular locations of human, animal, Gram-negative bacterial and eukaryotic proteins. These classifiers adopted the protein features derived from gene ontology information. Although they provided good performance, they can be further improved by adopting more powerful machine learning algorithms. In this study, four improved multi-label classifiers were set up for identification of subcellular locations of the above four protein types. The random k-labelsets (RAKEL) algorithm was used to tackle proteins with multiple locations, and random forest was used as the basic prediction engine. All classifiers were tested by jackknife test, indicating their high performance. Comparisons with previous classifiers further confirmed the superiority of the proposed classifiers.
The work of Markowitz [1] stands as one of the pioneering works on portfolio theory [2] and details Markowitz's major contributions to the rising modern portfolio theory. Besides providing criterious description of Markowitz's model of portfolio choice [3], the work in [1] traced future research directions explored by other scientists like the suggestion to replace variance with semi-variance as a measure of risk, recommendations on the maximization of the expected logarithmic utility of return, and the outlining of a market model developed in depth by Sharpe [4].
However, Markowitz's portfolio selection framework relies on the premise that one can measure asset's return variance as emphasized by the statement that "examples of rapidly increasing variances are of mostly academic interest" [1]. This restriction brings concerns related to the framework suitability in a scenario where heavy-tailed distributions better model assets return by allowing fast varying volatilities resulting from extreme events. We can find evidences in literature that financial data is better modelled by α-stable processes (heavy-tailed alternative to Brownian motion [5]) or by heavy-tailed time series models [6,7]. Although in this work we rely on the general hypothesis that logarithmic returns in financial data follow an α-stable process with parameter 0<α<2 [8] (which implies that variance of return is undefined), without loss of generality, we use results from extreme value theory (EVT) regarding generalized extreme value (GEV) distribution [9] as an alternative to α-stable distributions. This approach can be considered valid since the GEV distribution has fat-tailed behaviour and can be used as a proxy of various fat-tailed distributions.
From an economic point of view, it is well known that extreme share returns on stock markets can have important implications for financial risk management and several studies have successfully applied GEV to model financial data [10]. For example, Gettinby et al. [11] characterized the distribution of extreme returns for a UK share index over the years 1975 to 2000. They considered the suitability of several distributions, being the weekly maxima and minima of daily return best modelled by the GEV and the Generalised Logistic distributions. For the UK case, Generalised Logistic was a better choice overall. On the other hand, GEV has presented a similar modelling capability as well as presenting some important properties due to EVT. Also, Hussain & Li [12] studied the distribution of the extreme daily returns of the Shanghai Stock Exchange (SSE) Composite Index. They modelled the SSE Composite index returns based on the data from 1991 to 2013, which indicated that the Generalized Logistic distribution is a better fit for the minima series and that the GEV distribution is a better fit for the maxima series of the returns for the Chinese stock market.
EVT is a branch of probability and statistics that deals with the modeling of extreme events that are related to maximums and minimums of independent random samples. Applications of this theory are found in finance [13], natural catastrophes, and equipment failures, among others. The books [6,14,15,16] provide extensive coverage that allows for a detailed study of EVT.
Furthermore, EVT provides a theoretical basis and framework to deal with extreme deviations from the mean of distribution functions (DFs) by restricting the behavior of the DFs in the tails. It focuses on the study of the possible limiting distributions and their properties for the normalized maximum.
Specifically, let X1,X2,⋯,Xn be a sequence of independent and identically distributed (i.i.d.) random variables (RVs) with common distribution function F and set Mn=max{X1,⋯,Xn}. The theory is concerned with properties of F and of the possible non-degenerate distribution functions G satisfying
limn→∞P(Mn−bnan≤x)=limn→∞Fn(anx+bn)=G(x),∀x∈C(G), | (1.1) |
for sequences of constants an>0 and bn∈R (n=1,2,⋯), suitably chosen, where C(G) denotes the set of continuity points of G.
The possible distribution functions G satisfying (1.1) have been known for some time [17] and have been extensively studied by several authors from then on. They are also known as max-stable laws (or max-stable distributions) and can only be of three well-known types: Fréchet, Weibull or Gumbel.
Goncu et al. [18] used the EVT to model the extreme return behaviour of the Istanbul Stock Exchange (ISE), Turkey. They considered Gumbel, Fréchet and Weibull distributions for modelling extreme returns over different investment horizons. Their results indicate that when the Value at Risk (VaR) is computed with the proposed distributions, backtesting results indicate that the EVT provides superior risk management in all the sub-intervals considered compared to the VaR estimation under the assumption of a normal distribution.
For statistical applications, the max-stable distributions can be summarized in a single distribution function called generalized extreme value (GEV) distribution. Essentially, the GEV distribution has the cumulative distribution function (CDF) given by
G(x)=exp{−(1+γx)−1γ},1+γx>0,γ∈R, | (1.2) |
where γ is the shape parameter.
Our interest in this work is the property of stress-strength probability which, in general terms, consists of the study of the probability of failure of a system or component based on the comparison of the applied stress to the strength of the system. Let stress Y and strength X be independent continuous RVs with probability density function (PDF) fY and CDF FX, respectively. The stress-strength probability (or reliability) is defined as
R=P(X<Y)=∫∞−∞FX(x)fY(x)dx. | (1.3) |
There are several applications of this theory such as in engineering and manufacturing, aerospace and defense, automotive industry, energy sector, healthcare, and electronics, among others. See [19] for more details.
The stress-strength reliability framework is versatile and finds various applications in economics. Besides this contribution regarding financial data, previous works researched economic inequality [19,20]. Regarding financial data, the authors explored stress-strength reliability framework in [21,22,23,24].
The stress-strength probability for the extreme Fréchet, Weibull, and Gumbel distributions has been widely studied in the literature. Nadarajah [25] considered the class of extreme value distributions and derived the corresponding forms for the reliability R in terms of special functions. Confidence limits for R involving Weibull models were presented in [26]. Kundu & Raqab [27] proposed a modified maximum likelihood estimator of R and obtained the asymptotic distribution of the modified maximum likelihood estimators, which was used to construct the confidence interval of R. The previous results of R for Weibull distribution were generalized by Nojosa & Rathie [28], where R was expressed in terms of H-functions. Bayesian estimation of R for Fréchet and Weibull distributions has also been explored [29,30].
The goal of this paper is to present an asset selection approach based on the probability R=P(X<Y) when both X and Y present the distributions of returns of two assets. In particular, we seek to derive an expression of R when X and Y have three-parameter GEV distributions and to propose an estimation procedure of R by not using transformations in the data and with as fewer parameter restrictions as possible.
The paper is organized as follows: In Section 2, we define the H-function, H-function and the three-parameter GEV distribution. Section 3 deals with the derivation of R when X and Y are independent GEV RVs. The maximum likelihood estimation for R is presented in Section 4. In Section 5, we present Monte-Carlo simulations for the estimation of R and also deal with two real situations involving log-returns of stock prices and different-length carbon fibers. The last section deals with the conclusions and the Appendix presents the correlation matrices of the data set modeled in Section 5.
In this section, we present definitions and results on which our contributions are based.
Recently, Rathie et al. [31] introduced the extreme-value H-function as:
H(a1,a2,a3,a4,a5,a6):=∫∞0ya6exp{−a1y−(a2ya3+a4)a5}dy, | (2.1) |
where ℜ(a1),ℜ(a2),ℜ(a4)∈R+,a3,a5∈C, not both ℜ(a1) and ℜ(a2) can be equal to zero at the same time, ℜ(a6)>−1 when a1≠0 or a1=0 and sign(a3)=sign(a5), ℜ(a6)<−1 when a1=0 and sign(a3)≠sign(a5). In this paper, R, C and ℜ denote the real numbers, complex numbers and the real part of a complex number, respectively.
In this work, we are interested in the case a6=0. Thus, we omit such a parameter from the representation and denote only:
H(a1,a2,a3,a4,a5):=∫∞0exp{−a1y−(a2ya3+a4)a5}dy. | (2.2) |
In the following sections, we prove that all stress-strength probabilities involving three-parameter GEV distribution with shape parameters of equal sign can be written as H-functions.
Note that (2.1) generalizes some important cases of the H-function (cf. [32]) defined by
Hm,np,q[z|(a1,A1),⋯,(ap,Ap)(b1,B1),⋯,(bq,Bq)]=12πi∫L∏mk=1Γ(bj+Bjs)∏nj=1Γ(1−aj−Ajs)∏qk=m+1Γ(1−bj−Bjs)∏pj=n+1Γ(aj+Ajs)z−sds, | (2.3) |
where 0≤m≤q, 0≤n≤p (not both m and n simultaneously zero), Aj>0 (j=1,⋯,p), Bk>0 (k=1,⋯,q), aj and bk are complex numbers such that no poles of Γ(bk+Bks) (k=1,⋯,m) coincide with poles of Γ(1−aj−Ajs) (j=1,⋯,n). L is a suitable contour w−i∞ to w+i∞, w∈R, separating the poles of the two types mentioned above. For more details, see [32].
An important special case of this function is obtained by taking a4=0, which represents an upper (or lower) bound for its value depending on the sign of a5. This case is, therefore, an extreme value of the function and can be written in terms of the H-function as [31]:
H(a1,a2,a3,0,a5,a6)=∫∞0ya6exp{−a1y−aa52ya3a5}dy=1a(1+a6)/a32a3a5H1,11,1[a1a−1/a32|(1−(1+a6)a3a5,1a3a5)(0,1)]=1aa6+11H1,11,1[(a2aa31)a5|(−a6,a3a5)(0,1)], | (2.4) |
when sign(a3)=sign(a5) and:
H(a1,a2,a3,0,a5,a6)=1a(1+a6)/a32|a3a5|H2,00,2[a1a−1/a32|−(0,1),((1+a6)a3a5,1|a3a5|)]=1aa6+11H2,00,2[(a2aa31)a5|−(0,1),(1+a6,|a3a5|)], | (2.5) |
otherwise.
The three-parameter GEV distribution is obtained by taking CDF of the same type of the standard GEV G defined in (1.2). That means, G(x;μ,σ,γ)=G(x−μσ). We denote by X∼GEV(μ,σ,γ), μ,γ∈R and σ∈R+, an RV with CDF given by
G(x;μ,σ,γ)=exp{−[1+γσ(x−μ)]−1γ},1+γσ(x−μ)>0, | (2.6) |
where μ is the location parameter, σ is the scale parameter and γ is the shape parameter. The corresponding probability density function (PDF) is given by
g(x;μ,σ,γ)=G(xi,γi,μ,σ)1σ[1+γσ(x−μ)]−1γ−1,1+γσ(x−μ)>0. | (2.7) |
Figure 1 shows the behavior of g for some parameter choices. Note that the location parameter shifts the curve, the scale controls dispersion, and the density changes according to the sign of the shape parameter.
In this section, the reliability of two independent three-parameter GEV RVs is derived in terms of the H-function. In addition, with suitable parameter restrictions, representations of R as an H-function and an explicit form are obtained. We consider the case of two independent GEV distributions with different shape parameters (but of the same sign). Cases of opposite signs of shape parameters are not normally of interest, as it would indicate that random variables with incompatible support are being compared. Therefore, these cases are not treated in the present paper.
Theorem 3.1. Let Y and X be independent RVs, respectively, with distribution GEV(μ1,σ1,γ1) and GEV(μ2,σ2,γ2), μ∈R, σj∈R+, γj∈R(γj≠0), j=1,2. Then
● When γj>0, j=1,2:
R=P(X<Y)=H(1,γ2σ1σ2γ1,−γ1,1+γ2σ2(μ1−μ2−σ1γ1),−1γ2), | (3.1) |
provided that μ1−σ1γ1≥μ2−σ2γ2. When μ1−σ1γ1≤μ2−σ2γ2:
R=P(X<Y)=1−H(1,γ1σ2σ1γ2,−γ2,1+γ1σ1(μ2−μ1−σ2γ2),−1γ1). | (3.2) |
● When γj<0, j=1,2:
R=P(X<Y)=H(1,γ2σ1σ2γ1,−γ1,1+γ2σ2(μ1−μ2−σ1γ1),−1γ2), | (3.3) |
provided that μ1−σ1γ1≤μ2−σ2γ2. When μ1−σ1γ1≥μ2−σ2γ2:
R=P(X<Y)=1−H(1,γ1σ2σ1γ2,−γ2,1+γ1σ1(μ2−μ1−σ2γ2),−1γ1). | (3.4) |
In particular, if μ1−σ1γ1=μ2−σ2γ2, we have
R=γ2γ1(γ2σ1γ1σ2)1/γ1H1,11,1[(γ2σ1γ1σ2)1/γ1|(γ1−γ2γ1,γ2γ1)(0,1)]. | (3.5) |
Proof. Set μj∈R, σj,γj∈R+ (j=1,2). Then
R=P(X<Y)=∫∞−∞G(x;μ2,σ2,γ2)g(x;μ1,σ1,γ1)dx=∫+∞Mexp{−[1+γ2σ2(x−μ2)]−1γ2−[1+γ1σ1(x−μ1)]−1γ1}[1+γ1σ1(x−μ1)]−1γ1−1dxσ1, | (3.6) |
where M=max{μ1−σ1γ1,μ2−σ2γ2}. Substituting y=[1+γ1σ1(x−μ1)]−1γ1 and taking M=μ1−σ1γ1, we can rewrite (3.6) as
R=∫+∞0exp{−y−[1+γ2σ2(μ1−μ2−σ1γ1)+γ2σ2σ1γ1y−γ1]−1/γ2}dy. | (3.7) |
Hence, (3.1) follows from (2.1) and (3.7). For the case where γj>0, j=1,2 and μ1−σ1γ1≤μ2−σ2γ2, it suffices to notice that P(X<Y)=1−P(Y<X) and the result in (3.1) is applied with interchanged sub-indices. For the cases where γj<0, j=1,2, the same rationale can be applied, just noticing that in such cases x mostly takes negative values. The case where γj=0, j=1,2, can be obtained as a limiting procedure and shall be explicitly explored later on the present paper. In addition, applying (2.4) with μ1−σ1γ1=μ2−σ2γ2, we obtain (3.5).
Remark 3.2. In a practical scenario, the estimates (ˆμ1,ˆσ1,ˆγ1,ˆμ2,ˆσ2,ˆγ2) should be obtained. Then, if sign(ˆγ1)=sign(ˆγ2), the conditions μ1−σ1γ1≥μ2−σ2γ2 or μ1−σ1γ1≤μ2−σ2γ2 must be verified and the corresponding R expression should be used.
Next, we consider some special cases of two independent GEV random variables. We have the following immediate consequence of Theorem 3.1:
Corollary 3.3. Let Y∼GEV(μ1,σ1,γ1) and X∼GEV(μ2,σ2,γ2) be independent RVs, with γ1=γ2=γ∈R,γ≠0, μ1,μ2∈R, σ1,σ2∈R+ and μ1−μ2=σ1γ−σ2γ. Then, we have
R=P(X<Y)=(σ1/γ1σ1/γ1+σ1/γ2). | (3.8) |
Lastly, we consider the cases of two independent GEV with γ1=γ2=0.
Theorem 3.4. Let Y and X be independent RVs, respectively, with distribution GEV(μ1,σ1,0) and GEV(μ2,σ2,0), μj∈R, σj∈R+, j=1,2. Then
R=exp{μ1σ1}H(exp{μ1σ1},exp{μ2σ2},σ1σ2,0,1)=exp{μ1−μ2σ1}σ2σ1H1,11,1[exp{μ1−μ2σ1}|(σ1−σ2σ1,σ2σ1)(0,1)]. | (3.9) |
Proof. Set μj∈R and σj∈R+ (j=1,2). Then
R=P(X<Y)=∫∞−∞G(x;μ2,σ2,0)g(x;μ1,σ1,0)dx=∫∞−∞exp{−exp{−x−μ2σ2}−exp{−x−μ1σ1}}exp{−x−μ1σ1}dxσ1. | (3.10) |
Substituting y=exp{−x/σ1}, we can rewrite (3.10) as
R=exp{μ1σ1}∫+∞0exp{−exp{μ1σ1}y−exp{μ2σ2}yσ1/σ2}dy. | (3.11) |
Hence, (3.9) follows from (2.1) and (3.11).
We have the immediate consequence of Theorem 3.4.
Corollary 3.5. Let Y∼GEV(μ1,σ1,0) and X∼GEV(μ2,σ2,0) be independent RVs, with σ1=σ2=σ∈R+. Then, we have
R=P(X<Y)=exp{μ1/σ}exp{μ1/σ}+exp{μ2/σ}. | (3.12) |
The results presented in Theorems 3.1 and 3.4 are more general than that presented in the literature. The H-function allows us to write the probability R with as little parameter restrictions as possible. Table 1 lists related studies and their parameter restrictions.
sign(γ) | Distribution | Reference | Parameter restriction |
0 | Gumbel | [25] | σ1=σ2 or σ1=2σ2 or σ2/σ1>1 |
1 | Fréchet | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[30] | μ1=μ2=0 | ||
[29] | μ1=μ2=0 and γ1=γ2 | ||
[33] | μ1=μ2=0 | ||
−1 | Weibull (min) | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[26] | μ1=μ2=0 and γ1=γ2=γ | ||
[27] | μ1=μ2=μ and γ1=γ2=γ | ||
[28] | μ1=μ2=0 | ||
*p and q are coprime integers. |
Remark 3.6. The particular case of GEV while sign(γ)=−1 (cf. [14]) is called reversed Weibull. The Weibull distribution studied by the authors cited in Table 1 is obtained as the limit of a normalized minimum of i.i.d. RVs. That is, the Weibull distribution is obtained by
limn→∞P(min{X1,⋯,Xn}−bnan≤x)=limn→∞(1−(1−F(anx+bn))n)=1−exp{−(x−μσ)γ},x≥μ, | (3.13) |
where X1,⋯,Xn are i.i.d. RVs of F and an and bn are suitable sequences of constants (see Theorem 2.1.5 in [14]).
Let X1,⋯,Xn be i.i.d. RVs with distribution GEV(μ2,σ2,γ2) and Y be an independent RV with distribution GEV(μ1,σ1,γ1). Set Mn=max{X1,⋯,Xn}. Then, P(Mn≤u)=Gn(u;μ2,σ2,γ2) and we have
P(X1<Y,⋯,Xn<Y)=P(Mn≤Y)=∫∞−∞Gn(u;μ2,σ2,γ2)g(u;μ1,σ1,γ1)du=:In. | (3.14) |
Closed expressions for (3.14) are presented in the following result. Its proof follows the same steps of Theorems 3.1 and 3.4 and it will be omitted.
Theorem 3.7. Let X1,⋯,Xn be i.i.d. RVs with distribution GEV(μ2,σ2,γ2) and Y be an independent RV with distribution GEV(μ1,σ1,γ1). Then
● When γj>0, j=1,2:
P(X1<Y,⋯,Xn<Y)=H(1,γ2σ1σ2γ1n−γ2,−γ1,[1+γ2σ2(μ1−μ2−σ1γ1)]n−γ2,−1γ2), | (3.15) |
provided that μ1−σ1γ1≥μ2−σ2γ2. When μ1−σ1γ1≤μ2−σ2γ2:
P(X1<Y,⋯,Xn<Y)=1−H(1,γ1σ2σ1γ2n−γ1,−γ2,[1+γ1σ1(μ2−μ1−σ2γ2)]n−γ1,−1γ1). | (3.16) |
● When γj<0, j=1,2:
P(X1<Y,⋯,Xn<Y)=H(1,γ2σ1σ2γ1n−γ2,−γ1,[1+γ2σ2(μ1−μ2−σ1γ1)]n−γ2,−1γ2), | (3.17) |
provided that μ1−σ1γ1≤μ2−σ2γ2. When μ1−σ1γ1≥μ2−σ2γ2:
P(X1<Y,⋯,Xn<Y)=1−H(1,γ1σ2σ1γ2n−γ1,−γ2,[1+γ1σ1(μ2−μ1−σ2γ2)]n−γ1,−1γ1). | (3.18) |
● When γ1=γ2=0:
P(X1<Y,⋯,Xn<Y)=exp{μ1σ1}H(exp{μ1σ1},exp{μ2σ2}n,σ1σ2,0,1). | (3.19) |
Remark 3.8. In a broader k-out-of-n multicomponent reliability context, consider independent RVs Y,X1,⋯,Xk with Y∼GEV(μ1,σ1,γ1) and Xj∼GEV(μ2,σ2,γ2), for j=1,⋯,k (X′js are i.i.d.). The reliability for this kind of model is given by
Rs,k=P(atleastsoutof(X1,⋯,Xk)exceedY)=k∑j=s(kj)∫∞−∞(1−G(u;μ2,σ2,γ2))j(G(u;μ2,σ2,γ2))k−jg(u;μ1,σ1,γ1)du. |
Using a binomial expansion, we obtain
Rs,k=k∑j=sj∑r=0(kj)(jr)(−1)j−r∫∞−∞(G(u;μ2,σ2,γ2))k−rg(u;μ1,σ1,γ1)du. | (3.20) |
Note that the integral terms in (3.20) are particular cases of (3.14) provided that n=k−r. Therefore,
Rs,k=k∑j=sj∑r=0(kj)(jr)(−1)j−rIk−r. |
This section deals with parameter estimation for R=P(X<Y) given two independent GEV RVs. The literature presents some maximum likelihood estimators for R considering explicit forms of R obtained by strong parameter restrictions on extreme value distributions (such as [27,29,30]). Those approaches require the estimation of the parameters to be done jointly in the two samples. In our case, we release any requirements about having the same parameters between different samples, since we deal with expressions of R in terms of functions H.
Consider the PDF g(⋅;μ,σ,γ) defined in (2.7). Take X=(X1,⋯,Xn) as a sample of n observations. The likelihood function for the GEV(⋅;μ,σ,γ) is given by:
L(μ,σ,γ;X)=n∏i=1g(Xi;μ,σ,γ)1[1+γ(Xi−μ)/σ>0], | (4.1) |
where 1A denotes the indicator function on the set A. Note that ∏ni=11[1+γ(Xi−μ)/σ>0]>0 if and only if xi∈suppg(⋅;μ,σ,γ) for all i=1,⋯,n. Here, suppg denotes the support of the function g. Then, if γ≠0, we are not able to obtain the MLE explicitly, so an additional numeric procedure is required in the likelihood maximization (see [6] for a more detailed discussion).
Remark 4.1. Set X=(X1,⋯,Xn), a random sample of GEV(μ2,σ2,γ2), and Y=(Y1,⋯,Ym), a random sample of GEV(μ1,σ1,γ1), with γj>0, j=1,2 and μ1−σ1/γ1≥μ2−σ2/γ2 (or γj<0, j=1,2 and μ1−σ1/γ1≤μ2−σ2/γ2). Let ˆμi,ˆσi,ˆγi (i=1,2) be the estimates of μi,σi,γi. We are able to estimate R by the invariance property of MLE, as follows:
ˆR=H(1,ˆγ2ˆσ1ˆσ2ˆγ1,−ˆγ1,1+ˆγ2ˆσ2(ˆμ1−ˆμ2−ˆσ1ˆγ1),−1ˆγ2). | (4.2) |
Alternatively, whenever γj>0, j=1,2 and μ1−σ1/γ1≤μ2−σ2/γ2 (or γj<0, j=1,2 and μ1−σ1/γ1≥μ2−σ2/γ2), the same invariance property can be applied, leading to:
ˆR=1−H(1,ˆγ1ˆσ2ˆσ1ˆγ2,−ˆγ2,1+ˆγ1ˆσ1(ˆμ2−ˆμ1−ˆσ2ˆγ2),−1ˆγ1). | (4.3) |
This is due to the Theorems 3.1 and 3.4 that describe R in terms of the function H (which is an integral, hence a continuous and measurable function).
Whenever a single set of realizations of the random variables involved is available, the MLE approach above is of utmost importance. This is the case, for example, of asset selection, when a single time series of observed returns is available for each asset.
On the other hand, to illustrate the suitability of the analytical closed-form expressions hereby derived, a direct simulation approach can be carried out. In such case, several samples of size n can be drawn from each random variable, which are then used to estimate the empirical value of R and can be repeated several times. Both approaches will be explored in the next section.
To evaluate the correctness of the closed-form expression for R given in Theorem 3.1, we generate N Monte-Carlo samples, each of which is size n, of the random variables GEV(μ2,σ2,γ2) and GEV(μ1,σ1,γ1). In these cases, the values of μ2,σ2,γ2,μ1,σ1,γ1 are pre-specified.
The GEV distribution with negative shape parameter is treated in Tables 2 and 3 where we analyze the estimates ˆR, bias and root mean squared error (RMSE). Table 4 deals with positive shape parameter.
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1151 | -0.0004 | 0.0330 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4280 | 0.0070 | 0.0502 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4591 | 0.0059 | 0.0525 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0796 | 0.0002 | 0.0241 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5277 | 0.0021 | 0.0466 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5669 | 0.0017 | 0.0492 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1387 | 0.0027 | 0.0389 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5129 | -0.0037 | 0.0462 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5362 | 0.0009 | 0.0461 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1025 | -0.0010 | 0.0313 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3873 | -0.0022 | 0.0463 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4159 | -0.0043 | 0.0511 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0573 | 0.0017 | 0.0270 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4275 | 0.0083 | 0.0530 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4807 | 0.0009 | 0.0474 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1120 | -0.0016 | 0.0277 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4279 | -0.0089 | 0.0474 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4500 | -0.0022 | 0.0506 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1234 | 0.0003 | 0.0302 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3701 | 0.0014 | 0.0488 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.4034 | -0.0045 | 0.0472 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0797 | 0.0025 | 0.0276 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4134 | 0.0026 | 0.0403 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4794 | -0.0055 | 0.0548 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1303 | -0.0032 | 0.0325 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.4068 | -0.0069 | 0.0457 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4344 | -0.0015 | 0.0452 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1147 | -0.0000 | 0.0106 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4355 | -0.0005 | 0.0151 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4645 | 0.0005 | 0.0156 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0798 | 0.0001 | 0.0084 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5299 | -0.0001 | 0.0159 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5693 | -0.0007 | 0.0151 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1414 | -0.0000 | 0.0112 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5096 | -0.0004 | 0.0162 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5374 | -0.0003 | 0.0158 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1018 | -0.0003 | 0.0098 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3852 | -0.0001 | 0.0152 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4119 | -0.0003 | 0.0152 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0587 | 0.0003 | 0.0074 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4366 | -0.0008 | 0.0154 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4814 | 0.0002 | 0.0150 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1106 | -0.0002 | 0.0101 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4196 | -0.0006 | 0.0157 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4471 | 0.0008 | 0.0150 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1243 | -0.0006 | 0.0103 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3722 | -0.0007 | 0.0158 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.3974 | 0.0014 | 0.0154 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0822 | 0.0000 | 0.0086 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4166 | -0.0005 | 0.0160 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4736 | 0.0004 | 0.0162 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1265 | 0.0007 | 0.0105 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.3998 | 0.0001 | 0.0155 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4323 | 0.0006 | 0.0151 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.0617 | 0.0624 | -0.0007 | 0.0244 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.4288 | 0.4306 | -0.0018 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 0.3 | 0.4491 | 0.4478 | 0.0012 | 0.0514 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.0906 | 0.0891 | 0.0015 | 0.0284 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.4443 | 0.4463 | -0.0020 | 0.0507 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 0.3 | 0.4717 | 0.4727 | -0.0010 | 0.0499 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.1299 | 0.1300 | -0.0001 | 0.0332 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.4419 | 0.4423 | -0.0004 | 0.0509 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 0.3 | 0.4611 | 0.4602 | 0.0008 | 0.0497 |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.1469 | 0.1474 | -0.0004 | 0.0359 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.4764 | 0.4774 | -0.0010 | 0.0486 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 1.0 | 0.4947 | 0.4940 | 0.0007 | 0.0504 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.1846 | 0.1847 | -0.0001 | 0.0382 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.4980 | 0.4985 | -0.0006 | 0.0520 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 1.0 | 0.5240 | 0.5235 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.2144 | 0.2164 | -0.0019 | 0.0431 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.4951 | 0.4925 | 0.0026 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 1.0 | 0.5128 | 0.5124 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.1868 | 0.1883 | -0.0015 | 0.0393 |
0.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.4906 | 0.4894 | 0.0013 | 0.0523 |
0.4 | 1.5 | 1.5 | 0.5 | 0.5 | 1.5 | 0.5076 | 0.5072 | 0.0004 | 0.0495 |
2.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.2237 | 0.2243 | -0.0006 | 0.0414 |
0.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.5056 | 0.5052 | 0.0004 | 0.0501 |
0.4 | 1.0 | 1.5 | 0.5 | 0.7 | 1.5 | 0.5318 | 0.5310 | 0.0008 | 0.0496 |
2.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.2461 | 0.2451 | 0.0011 | 0.0444 |
0.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.5046 | 0.5042 | 0.0004 | 0.0491 |
0.4 | 1.5 | 1.5 | 0.5 | 0.9 | 1.5 | 0.5220 | 0.5226 | -0.0006 | 0.0490 |
For the simulation, for each line in the Tables 2–4, the following procedure was carried out:
(1) for each Monte-Carlo sample, the estimate ˆR is computed empirically, simply as n−1∑iI(xi<yi), where I(.) is an indicator function, which is 1 for true arguments and 0 otherwise;
(2) ˆRMC is evaluated by taking the sample mean of the Monte-Carlo samples ˆR;
(3) the bias is computed as the difference between the value obtained by Theorem 3.1 and ˆRMC. The same follows for the RMSE, which also considers the true value as the analytically obtained one.
As expected, Tables 2 and 3 illustrate that the analytical results obtained match the empirical ones. This is a clear evidence of the correctness of the new expressions hereby derived. Besides, it is clear that increasing the sample size n leads to more precise estimations of R, indicating reduced bias and greater consistency. In Table 4 we observe the same good behavior of the estimator characterized by low bias and RMSE.
In order to evaluate the proposed framework, we modeled stock prices log-returns as GEV and compare them in a reliability sense. For this, we assume that the returns are independent. To meet independence requirements, we take stock log-returns that are not correlated and correlations are measured using Pearson, Kendal and Spearman methods.
Denote X1,X2,X3 and X4 the stock price log-returns of BBAS3, ITUB4, VALE3 and VIIA3, respectively. The data sets are retrieved directly through the software R [34] by the command
quantmod::getSymbols("BBAS3.SA", src = "yahoo", auto.assign = FALSE, from = '2022-01-01', to = '2023-04-30', return.class = 'xts').
The data sets have information in Brazilian currency (R$, BRL).
Summary statistics for the data sets X1,X2,X3 and X4 are presented in Table 5. A boxplot is presented in Figure 2 showing the symmetry around zero of the log-returns and that X4 has more dispersion than the others.
RV | Data Set | min | 1st Qu. | Median | Mean | 3rd Qu. | Max | n |
X1 | BBAS3 | -0.1057 | -0.0097 | 0.0019 | 0.0012 | 0.0136 | 0.0736 | 330 |
X2 | ITUB4 | -0.0492 | -0.0105 | 0.0004 | 0.0006 | 0.0109 | 0.0794 | 330 |
X3 | VALE3 | -0.0689 | -0.0140 | 0.0001 | -0.0002 | 0.0128 | 0.0989 | 330 |
X4 | VIIA3 | -0.1075 | -0.0344 | -0.006 | -0.0030 | 0.0231 | 0.1504 | 330 |
Maximum likelihood (ML) estimates and Kolmogorov-Smirnov (KS) p-values are given in Table 6. Figure 3 shows histograms and the fit of the GEV model to X1,X2,X3 and X4. For each data set, the empirical CDF (ECDF) is compared to the theoretical model in Figure 4.
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0063 | 0.0219 | -0.2535 | -803.9168 | 0.0147 |
X2 | ITUB4 | -0.0064 | 0.0165 | -0.1545 | -870.9392 | 0.4299 |
X3 | VALE3 | -0.0095 | 0.0222 | -0.1631 | -774.0895 | 0.2331 |
X4 | VIIA3 | -0.0217 | 0.0396 | -0.1170 | -567.6979 | 0.6996 |
Although the p-value of the Kolmogorov-Smirnov test is small for the BBAS3 (X1), the graphical analysis does not invalidate the good fit of the distribution to the data.
Aiming at estimating probabilities of the type R=P(X<Y) via Theorems 3.1–3.4, we need X and Y to be independent RVs. In this sense, we analyzed the dependency structures of X1,⋯,X4 using the Pearson, Kendal and Spearman correlation matrices, and the results are presented in the Appendix. The pairs chosen to be compared are X3−X2, X3−X4 and X3−X1.
Reliability measures of the type R=P(X<Y) appear in a decision process of an investor. In summary, when X and Y represent profit RVs and R<1/2, it is advisable that the investor chooses the variable X. If R>1/2, the opposite occurs. The case R=1/2 is inconclusive. Thus, knowing how to evaluate R accurately is important to support the decision process. In this sense, Table 7 presents the estimates of P(X3<X1), P(X3<X2) and P(X3<X4) and the 95% bootstrap confidence intervals (CI). For all subsequent Tables, ˆR is the estimate obtained by first fitting the dataset to a GEV distribution and using the results from Remark 4.1 (invariance property of MLE of the parameters and the analytical expression obtained for R). Also, ˆRNP is a non-parametric estimation of R, which considers a similar approach as in the simulation study (uses an indicator function to compare samples from the two distributions). Finally, ˆRBoot is the mean value of the bootstrap estimation of R using the non-parametric approach and the CI reported is for such mean.
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5283 | 0.5242 | 0.5277 | (0.4883; 0.5676) |
P(X3<X2) | 0.5174 | 0.5242 | 0.5181 | (0.4766; 0.5569) |
P(X3<X4) | 0.4506 | 0.4364 | 0.4500 | (0.4055; 0.4944) |
The estimates of R indicate that, within the analyzed time period, VALE3 would be preferred over BBAS3 and ITUB4, while VIIA3 would be preferred over VALE3. On the other hand, the bootstrap CI estimates indicate that only the case P(X3<X4) was conclusive since 0.5 does not belong to the estimated confidence interval.
It is important to compare different distributions as candidate models for the log-returns modelling. Considering previous results in the literature [11], we compared the performance of the GEV and generalized logistic distributions as models for daily returns, as presented in Table 8.
RV | Data set | GEV | Generalized Logistic |
X1 | BBAS3 | -803.92 | - 833.43 |
X2 | ITUB4 | -870.94 | - 882.01 |
X3 | VALE3 | -774.09 | - 787.81 |
X4 | VIIA3 | -567.70 | - 569.60 |
It is possible to see that both GEV and generalized logistic provided quite similar modelling capabilities (about the same log-likelihood values). Since the two distributions considered have three parameters, there is no need to consider information criteria.
Instead of considering the log-returns for the closing prices, as in the previous analyses, one may study how the weekly maximums (or minimums) behave. This has a direct economic interpretation: a proxy for greater profits (or greater losses, i.e., shortfalls and values-at-risk) and has been explored in the literature [11,12,18].
Following previous studies [12], it is of interest to explore the modelling scenario where the maximum returns are considered GEV random variables. Table 9 show the GEV ML estimates for such case.
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0128 | 0.0393 | -0.1157 | -116.3016 | 0.8076 |
X2 | ITUB4 | -0.091 | 0.0397 | -0.4319 | -126.5530 | 0.8292 |
X3 | VALE3 | -0.0197 | 0.0426 | -0.2003 | -114.0183 | 0.9365 |
X4 | VIIA3 | -0.0464 | 0.0750 | -0.1339 | -72.9854 | 0.5378 |
Table 9 indicates that the GEV random variable is adequate for every case considered, which is expected as the EVT predicts such convergence. Now that the parameters have been obtained, Table 10 presents the reliability calculations.
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5453 | 0.5522 | 0.5413 | (0.4483; 0.6334) |
P(X3<X2) | 0.5409 | 0.5522 | 0.5483 | (0.4421; 0.6358) |
P(X3<X4) | 0.4364 | 0.4328 | 0.4350 | (0.3379; 0.5350) |
Since R=0.5 is within the confidence intervals, the metric becomes inconclusive. It is possible to notice, on the other hand, that if ˆR is considered as a metric by itself, it would precisely reflect the high volatility of VIIA3, which suffered severe instability and losses during the time window analyzed.
Besides the direct application to financial assets selection, some engineering applications can also benefit from the new relations hereby defined. One application is illustrated in the next subsection.
Stress-strength reliability can be also applied to the modelling and comparison of carbon fibers of lengths 10 and 20 mm. The data represent the strength data measured in GPa (gigapascal), for single carbon fibers tested under tension. The data are frequently used in the literature and are also presented below (e.g., [28]).
Carbon fibers of length 20 mm:
X=(1.312,1.314,1.479,1.552,1.700,1.803,1.861,1.865,1.944,1.958,1.966,1.977,2.006,2.021,2.027,2.055,2.063,2.098,2.140,2.179,2.224,2.240,2.253,2.270,2.272,2.274,2.301,2.301,2.359,2.382,2.382,2.426,2.434,2.435,2.478,2.490,2.511,2.514,2.535,2.554,2.566,2.570,2.586,2.629,2.633,2.642,2.648,2.684,2.697,2.726,2.770,2.773,2.800,2.809,2.818,2.821,2.848,2.880,2.954,3.012,3.067,3.084,3.090,3.096,3.128,3.233,3.433,3.585,3.585). |
Carbon fibers of length 10 mm:
Y=(1.901,2.132,2.203,2.228,2.257,2.350,2.361,2.396,2.397,2.445,2.454,2.474,2.518,2.522,2.525,2.532,2.575,2.614,2.616,2.618,2.624,2.659,2.675,2.738,2.740,2.856,2.917,2.928,2.937,2.937,2.977,2.996,3.030,3.125,3.139,3.145,3.220,3.223,3.235,3.243,3.264,3.272,3.294,3.332,3.346,3.377,3.408,3.435,3.493,3.501,3.537,3.554,3.562,3.628,3.852,3.871,3.886,3.971,4.024,4.027,4.225,4.395,5.020). |
Table 11 and Figure 5 show the summary statistics of X and Y in which it is possible to observe that Y (carbon fibers of length 10 mm) tend to have greater strength values than X (carbon fibers of length 20 mm).
RV | Data set | Min. | 1st Qu. | Median | Mean | 3rd Qu. | Max. | n |
X | Carbon fibers of 20 mm | 1.312 | 2.098 | 2.478 | 2.451 | 2.773 | 3.585 | 69 |
Y | Carbon fibers of 10 mm | 1.901 | 2.554 | 2.996 | 3.059 | 3.421 | 5.020 | 63 |
ML estimates and KS p-values for the GEV model are presented in Table 12 and the good fit can be observed in Figure 6.
Data set | ˆμ | ˆσ | ˆγ | KS p-value |
X | 2.2781 | 0.4956 | -0.2851 | 0.9978 |
Y | 2.7904 | 0.5245 | -0.0747 | 0.8216 |
The estimates of stress-strength reliability and bootstrap 95% CI are ˆR=0.774 and (0.699,0.851), respectively. It is easy to conclude that since P(X<Y)>1/2 and 0.5 is not within the CI, that X<Y statistically, i.e., the fibers with length of 10 mm have statistically greater strength values than the ones of length of 20 mm.
Despite its significant limitations, Markowitz's modern portfolio theory is still relied upon by many practitioners because of its user-friendly simplicity. This way, studying alternative approaches that are also of straightforward comprehension is of utmost importance.
In this paper, we studied the stress-strength reliability R=P(X<Y) when both X and Y follow three-parameter GEV distributions. In summary, when X and Y represent return RVs and R<1/2, it is advisable that the investor chooses the variable X. If R>1/2, the opposite occurs. The case R=1/2 is inconclusive. Thus, exact expressions for R have been obtained in terms of the extreme-value H-function with minimal parameter restrictions. With additional restrictions, it was shown that R can be calculated in terms of H-functions and even in terms of standard functions (such as exponential functions).
Monte-Carlo simulations attested to the performance of the analytical closed-form expressions hereby derived. By applying our methodology to real-world financial data, we could orient a stock selection procedure by calculating P(X<Y) when both X and Y represent stock returns. Besides, an engineering application was also described, where carbon fibers tested under tension were modelled in a stress-strength reliability sense.
In this appendix, we present the correlation matrices of the data sets (log-returns) modeled in Section 5. Thus, Tables 13–15 give the correlations of Pearson, Spearman and Kendal, respectively.
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.61 | 0.10 | 0.29 |
X2 | 1.00 | 0.15 | 0.31 | |
X3 | 1.00 | 0.03 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.63 | 0.15 | 0.31 |
X2 | 1.00 | 0.22 | 0.34 | |
X3 | 1.00 | 0.06 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.45 | 0.10 | 0.21 |
X2 | 1.00 | 0.15 | 0.24 | |
X3 | 1.00 | 0.04 | ||
X4 | 1.00 |
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
The authors acknowledge the support provided by the University of Brasilia (UnB). Additionally, M.O. acknowledges the Coordination for the Improvement of Higher Education Personnel (CAPES) for awarding him a scholarship for a Master of Science program.
The authors declare no conflicts of interest.
[1] |
K. C. Chou, H. B. Shen, Recent progress in protein subcellular location prediction, Anal. Biochem., 370 (2007), 1–16. https://doi.org/10.1016/j.ab.2007.07.006 doi: 10.1016/j.ab.2007.07.006
![]() |
[2] | R. F. Murphy, M. V. Boland, M. Velliste, Towards a systematics for protein subcellular location: quantitative description of protein localization patterns and automated analysis of fluorescence microscope images, in Proceedings International Conference on Intelligent System Molecular Biology, 8 (2000), 251–259. |
[3] |
J. Cao, W. Liu, J. He, H. Gu, Mining proteins with non-experimental annotations based on an active sample selection strategy for predicting protein subcellular localization, PLoS One, 8 (2013), e67343. https://doi.org/10.1371/journal.pone.0067343 doi: 10.1371/journal.pone.0067343
![]() |
[4] |
H. B. Shen, J. Yang, K. C. Chou, Methodology development for predicting subcellular localization and other attributes of proteins, Expert Rev. Proteomics, 4 (2007), 453–463. https://doi.org/10.1586/14789450.4.4.453 doi: 10.1586/14789450.4.4.453
![]() |
[5] |
A. Reinhardt, T. Hubbard, Using neural networks for prediction of the subcellular location of proteins, Nucleic Acids Res., 26 (1998), 2230–2236. https://doi.org/10.1093/nar/26.9.2230 doi: 10.1093/nar/26.9.2230
![]() |
[6] |
J. Cedano, P. Aloy, J. A. Perez-Pons, E. Querol, Relation between amino acid composition and cellular location of proteins, J. Mol. Biol., 266 (1997), 594–600. https://doi.org/10.1006/jmbi.1996.0804 doi: 10.1006/jmbi.1996.0804
![]() |
[7] |
Y. X. Pan, Z. Z. Zhang, Z. M. Guo, G. Y. Feng, Z. D. Huang, L. He, Application of pseudo amino acid composition for predicting protein subcellular location: stochastic signal processing approach, J. Protein Chem., 22 (2003), 395–402. https://doi.org/10.1023/a:1025350409648 doi: 10.1023/a:1025350409648
![]() |
[8] |
J. Y. Shi, S. Zhang, Q. Pan, G. Zhou, Using pseudo amino acid composition to predict protein subcellular location: approached with amino acid composition distribution, Amino Acids, 35 (2008), 321–327. https://doi.org/10.1007/s00726-007-0623-z doi: 10.1007/s00726-007-0623-z
![]() |
[9] |
H. Lin, H. Ding, F. Guo, A. Zhang, J. Huang, Predicting subcellular localization of mycobacterial proteins by using Chou's pseudo amino acid composition, Protein Pept. Lett., 15 (2008), 739–744. https://doi.org/10.2174/092986608785133681 doi: 10.2174/092986608785133681
![]() |
[10] |
K. Chou, Prediction of protein cellular attributes using pseudo-amino acid composition, Proteins, 43 (2001), 246–255. https://doi.org/10.1002/prot.1035 doi: 10.1002/prot.1035
![]() |
[11] |
T. Liu, X. Zheng, C. Wang, J. Wang, Prediction of subcellular location of apoptosis proteins using pseudo amino acid composition: an approach from auto covariance transformation, Protein Pept. Lett., 17 (2010), 1263–1269. https://doi.org/10.2174/092986610792231528 doi: 10.2174/092986610792231528
![]() |
[12] |
Y. Shen, J. Tang, F. Guo, Identification of protein subcellular localization via integrating evolutionary and physicochemical information into Chou's general PseAAC, J. Theor. Biol., 462 (2019), 230–239. https://doi.org/10.1016/j.jtbi.2018.11.012 doi: 10.1016/j.jtbi.2018.11.012
![]() |
[13] |
Y. H. Yao, Z. X. Shi, Q. Dai, Apoptosis protein subcellular location prediction based on position-specific scoring matrix, J. Comput. Theor. Nanos., 11 (2014), 2073–2078. https://doi.org/10.1166/jctn.2014.3607 doi: 10.1166/jctn.2014.3607
![]() |
[14] |
T. Liu, P. Tao, X. Li, Y. Qin, C. Wang, Prediction of subcellular location of apoptosis proteins combining tri-gram encoding based on PSSM and recursive feature elimination, J. Theor. Biol., 366 (2015), 8–12. https://doi.org/10.1016/j.jtbi.2014.11.010 doi: 10.1016/j.jtbi.2014.11.010
![]() |
[15] |
S. Wang, W. Li, Y. Fei, An improved process for generating uniform PSSMs and its application in protein subcellular localization via various global dimension reduction techniques, IEEE Access, 7 (2019), 42384–42395. https://doi.org/10.1109/ACCESS.2019.2907642 doi: 10.1109/ACCESS.2019.2907642
![]() |
[16] |
X. Cheng, X. Xiao, K. C. Chou, pLoc-mHum: predict subcellular localization of multi-location human proteins via general PseAAC to winnow out the crucial GO information. Bioinformatics, 34 (2018), 1448–1456. https://doi.org/10.1093/bioinformatics/btx711 doi: 10.1093/bioinformatics/btx711
![]() |
[17] |
X. Cheng, S. Zhao, W. Lin, X. Xiao, K. Chou, pLoc-mAnimal: predict subcellular localization of animal proteins with both single and multiple sites, Bioinformatics, 33 (2017), 3524–3531. https://doi.org/10.1093/bioinformatics/btx476 doi: 10.1093/bioinformatics/btx476
![]() |
[18] |
X. Cheng, X. Xiao, K.C. Chou, pLoc-mGneg: Predict subcellular localization of Gram-negative bacterial proteins by deep gene ontology learning via general PseAAC, Genomics, 110 (2017), 231–239. https://doi.org/10.1016/j.ygeno.2017.10.002 doi: 10.1016/j.ygeno.2017.10.002
![]() |
[19] |
X. Cheng, X. Xiao, K. C. Chou, pLoc-mEuk: Predict subcellular localization of multi-label eukaryotic proteins by extracting the key GO information into general PseAAC, Genomics, 110 (2018), 50–58. https://doi.org/10.1016/j.ygeno.2017.08.005 doi: 10.1016/j.ygeno.2017.08.005
![]() |
[20] |
K. Chou, Y. Cai, A new hybrid approach to predict subcellular localization of proteins by incorporating gene ontology, Biochem. Biophys. Res. Commun., 311 (2003), 743–747. https://doi.org/10.1016/j.bbrc.2003.10.062 doi: 10.1016/j.bbrc.2003.10.062
![]() |
[21] |
S. Wan, M. Mak, S. Kung, GOASVM: A subcellular location predictor by incorporating term-frequency gene ontology into the general form of Chou's pseudo-amino acid composition, J. Theor. Biol., 323 (2013), 40–48. https://doi.org/10.1016/j.jtbi.2013.01.012 doi: 10.1016/j.jtbi.2013.01.012
![]() |
[22] |
S. Wan, M. Mak, S. Kung, mGOASVM: Multi-label protein subcellular localization based on gene ontology and support vector machines, BMC Bioinf., 13 (2012), 290. https://doi.org/10.1186/1471-2105-13-290 doi: 10.1186/1471-2105-13-290
![]() |
[23] |
K. C. Chou, Y. D. Cai, Using functional domain composition and support vector machines for prediction of protein subcellular location, J. Biol. Chem., 277 (2002), 45765–45769. https://doi.org/10.1074/jbc.M204161200 doi: 10.1074/jbc.M204161200
![]() |
[24] |
K. Chou, H. Shen, A new method for predicting the subcellular localization of eukaryotic proteins with both single and multiple sites: Euk-mPLoc 2.0, PLoS One, 5 (2010), e9931. https://doi.org/10.1371/journal.pone.0009931 doi: 10.1371/journal.pone.0009931
![]() |
[25] |
Y. Cai, K. Chou, Nearest neighbour algorithm for predicting protein subcellular location by combining functional domain composition and pseudo-amino acid composition, Biochem. Biophys. Res. Commun., 305 (2003), 407–411. https://doi.org/10.1016/s0006-291x(03)00775-7 doi: 10.1016/s0006-291x(03)00775-7
![]() |
[26] |
K. Chou, Y. Cai, Predicting subcellular localization of proteins by hybridizing functional domain composition and pseudo-amino acid composition, J. Cell. Biochem., 91 (2004), 1197–1203. https://doi.org/10.1002/jcb.10790 doi: 10.1002/jcb.10790
![]() |
[27] |
X. Pan, L. Chen, M. Liu, Z. Niu, T. Huang, Y. Cai, Identifying protein subcellular locations with embeddings-based node2loc, IEEE/ACM Trans. Comput. Biol. Bioinf., 19 (2022), 666–675. https://doi.org/10.1109/TCBB.2021.3080386 doi: 10.1109/TCBB.2021.3080386
![]() |
[28] |
X. Pan, H. Li, T. Zeng, Z. Li, L. Chen, T. Huang, et al., Identification of protein subcellular localization with network and functional embeddings, Front. Genet., 11 (2021), 626500. https://doi.org/10.3389/fgene.2020.626500 doi: 10.3389/fgene.2020.626500
![]() |
[29] |
H. Liu, B. Hu, L. Chen, Identifying protein subcellular location with embedding features learned from networks, Curr. Proteomics, 18 (2021), 646–660. https://doi.org/10.2174/1570164617999201124142950 doi: 10.2174/1570164617999201124142950
![]() |
[30] | R. Wang, L. Chen, Identification of human protein subcellular location with multiple networks, Curr. Proteomics, 19 (2022), 344–356. |
[31] |
R. Su, L. He, T. Liu, X. Liu, L. Wei, Protein subcellular localization based on deep image features and criterion learning strategy, Briefings Bioinf., 22 (2020), bbaa313. https://doi.org/10.1093/bib/bbaa313 doi: 10.1093/bib/bbaa313
![]() |
[32] |
M. Ullah, F. Hadi, J. Song, D. Yu, PScL-DDCFPred: an ensemble deep learning-based approach for characterizing multiclass subcellular localization of human proteins from bioimage data, Bioinformatics, 38 (2022), 4019–4026. https://doi.org/10.1093/bioinformatics/btac432 doi: 10.1093/bioinformatics/btac432
![]() |
[33] |
M. Ullah, K. Han, F. Hadi, J. Xu, J. Song, D. Yu, PScL-HDeep: image-based prediction of protein subcellular location in human tissue using ensemble learning of handcrafted and deep learned features with two-layer feature selection, Briefings Bioinf., 22 (2021), bbab278. https://doi.org/10.1093/bib/bbab278 doi: 10.1093/bib/bbab278
![]() |
[34] | G. Tsoumakas, I. Vlahavas, Random k-Labelsets: An ensemble method for multilabel classification, in Machine Learning: ECML 2007, (2007), 406–417. https://doi.org/10.1007/978-3-540-74958-5_38 |
[35] |
L. Breiman, Random forests, Mach. Learn., 45 (2001), 5–32. https://doi.org/10.1023/A:1010933404324 doi: 10.1023/A:1010933404324
![]() |
[36] |
K. C. Chou, Z. C. Wu, X. Xiao, iLoc-Hum: using the accumulation-label scale to predict subcellular locations of human proteins with both single and multiple sites, Mol. Biosyst., 8 (2012), 629–641. https://doi.org/10.1039/c1mb05420a doi: 10.1039/c1mb05420a
![]() |
[37] |
H. B. Shen, K. C. Chou, A top-down approach to enhance the power of predicting human protein subcellular localization: Hum-mPLoc 2.0, Anal. Biochem., 394 (2009), 269–274. https://doi.org/10.1016/j.ab.2009.07.046 doi: 10.1016/j.ab.2009.07.046
![]() |
[38] |
W. Z. Lin, J. Fang, X. Xiao, K. Chou, iLoc-Animal: a multi-label learning classifier for predicting subcellular localization of animal proteins, Mol. Biosyst., 9 (2013), 634–644. https://doi.org/10.1039/c3mb25466f doi: 10.1039/c3mb25466f
![]() |
[39] |
H. B. Shen, K. C. Chou, Gneg-mPLoc: a top-down strategy to enhance the quality of predicting subcellular localization of Gram-negative bacterial proteins, J. Theor. Biol., 264 (2010), 326–333. https://doi.org/10.1016/j.jtbi.2010.01.018 doi: 10.1016/j.jtbi.2010.01.018
![]() |
[40] |
X. Xiao, Z. C. Wu, K. C. Chou, A multi-label classifier for predicting the subcellular localization of gram-negative bacterial proteins with both single and multiple sites, PLoS One, 6 (2011), e20592. https://doi.org/10.1371/journal.pone.0020592 doi: 10.1371/journal.pone.0020592
![]() |
[41] |
G. Tsoumakas, I. Katakis, Multi-label classification: An overview, Int. J. Data Warehouse. Min., 3 (2007), 1–13. https://doi.org/10.4018/jdwm.2007070101 doi: 10.4018/jdwm.2007070101
![]() |
[42] | S. Al-Maadeed, Kernel collaborative label power set system for multi-label classification, in Qatar Foundation Annual Research Forum Volume 2013 Issue 1, Hamad bin Khalifa University Press, 2013 (2013). https://doi.org/10.5339/qfarf.2013.ICTP-028 |
[43] |
J. P. Zhou, L. Chen, Z. H. Guo, iATC-NRAKEL: An efficient multi-label classifier for recognizing anatomical therapeutic chemical classes of drugs, Bioinformatics, 36 (2020), 1391–1396. https://doi.org/10.1093/bioinformatics/btz757 doi: 10.1093/bioinformatics/btz757
![]() |
[44] |
J. P. Zhou, L. Chen, T. Wang, M. Liu, iATC-FRAKEL: A simple multi-label web-server for recognizing anatomical therapeutic chemical classes of drugs with their fingerprints only, Bioinformatics, 36 (2020), 3568–3569. https://doi.org/10.1093/bioinformatics/btaa166 doi: 10.1093/bioinformatics/btaa166
![]() |
[45] |
X. Li, L. Lu, L. Chen, Identification of protein functions in mouse with a label space partition method, Math. Biosci. Eng., 19 (2022), 3820–3842. https://doi.org/10.3934/mbe.2022176 doi: 10.3934/mbe.2022176
![]() |
[46] |
H. Li, S. Zhang, L. Chen, X. Pan, Z. Li, T. Huang, et al., Identifying functions of proteins in mice with functional embedding features, Front. Genet., 13 (2022), 909040. https://doi.org/10.3389/fgene.2022.909040 doi: 10.3389/fgene.2022.909040
![]() |
[47] |
L. Chen, Z. Li, T. Zeng, Y. Zhang, H. Li, T. Huang, et al., Predicting gene phenotype by multi-label multi-class model based on essential functional features, Mol. Genet. Genomics, 296 (2021), 905–918. https://doi.org/10.1007/s00438-021-01789-8 doi: 10.1007/s00438-021-01789-8
![]() |
[48] |
Y. Zhu, B. Hu, L. Chen, Q. Dai, iMPTCE-Hnetwork: a multi-label classifier for identifying metabolic pathway types of chemicals and enzymes with a heterogeneous network, Comput. Math. Methods Med., 2021 (2021), 6683051. https://doi.org/10.1155/2021/6683051 doi: 10.1155/2021/6683051
![]() |
[49] |
J. Che, L. Chen, Z. Guo, S. Wang, Aorigele, Drug target group prediction with multiple drug networks, Comb. Chem. High Throughput Screen., 23 (2020), 274–284. https://doi.org/10.2174/1386207322666190702103927 doi: 10.2174/1386207322666190702103927
![]() |
[50] |
H. Wang, L. Chen, PMPTCE-HNEA: Predicting metabolic pathway types of chemicals and enzymes with a heterogeneous network embedding algorithm, Curr. Bioinf., 18 (2023), 748–759. https://doi.org/10.2174/1574893618666230224121633 doi: 10.2174/1574893618666230224121633
![]() |
[51] | J. Read, P. Reutemann, B. Pfahringer, MEKA: A multi-label/multi-target extension to WEKA, J. Mach. Learn. Res., 17 (2016), 1–5. |
[52] |
B. Ran, L. Chen, M. Li, Y. Han, Q. Dai, Drug-Drug interactions prediction using fingerprint only, Comput. Math. Methods Med., 2022 (2022), 7818480. https://doi.org/10.1155/2022/7818480 doi: 10.1155/2022/7818480
![]() |
[53] |
M. Onesime, Z. Yang, Q. Dai, Genomic island prediction via chi-square test and random forest algorithm, Comput. Math. Methods Med., 2021 (2021), 9969751. https://doi.org/10.1155/2021/9969751 doi: 10.1155/2021/9969751
![]() |
[54] |
L. Chen, K. Chen, B. Zhou, Inferring drug-disease associations by a deep analysis on drug and disease networks, Math. Biosci. Eng., 20 (2023), 14136–14157. https://doi.org/10.3934/mbe.2023632 doi: 10.3934/mbe.2023632
![]() |
[55] |
P. Chen, T. Shen, Y. Zhang, B. Wang, A sequence-segment neighbor encoding schema for protein hotspot residue prediction, Curr. Bioinf., 15 (2020), 445–454. https://doi.org/10.2174/1574893615666200106115421 doi: 10.2174/1574893615666200106115421
![]() |
[56] |
Z. B. Lv, J. Zhang, H. Ding, Q. Zou, RF-PseU: A random forest predictor for rna pseudouridine sites, Front. Bioeng. Biotechnol., 8 (2020), 134. https://doi.org/10.3389/fbioe.2020.00134 doi: 10.3389/fbioe.2020.00134
![]() |
[57] |
F. Huang, Q. Ma, J. Ren, J. Li, F. Wang, T. Huang, et al., Identification of smoking associated transcriptome aberration in blood with machine learning methods, Biomed. Res. Int., 2023 (2023), 445–454. https://doi.org/10.1155/2023/5333361 doi: 10.1155/2023/5333361
![]() |
[58] |
F. Huang, M. Fu, J. Li, L. Chen, K. Feng, T. Huang, et al., Analysis and prediction of protein stability based on interaction network, gene ontology, and kegg pathway enrichment scores, Biochim. Biophys. Acta. Proteins Proteom., 1871 (2023), 140889. https://doi.org/10.1016/j.bbapap.2023.140889 doi: 10.1016/j.bbapap.2023.140889
![]() |
[59] |
J. Ren, Y. Zhang, W. Guo, K. Feng, Y. Yuan, T. Huang, et al., Identification of genes associated with the impairment of olfactory and gustatory functions in COVID-19 via machine-learning methods, Life (Basel), 13 (2023), 798. https://doi.org/10.3390/life13030798 doi: 10.3390/life13030798
![]() |
[60] |
K. C. Chou, C. T. Zhang, Prediction of protein structural classes, Crit. Rev. Biochem. Mol. Biol., 30 (1995), 275–349. https://doi.org/10.3109/10409239509083488 doi: 10.3109/10409239509083488
![]() |
[61] |
K. C. Chou, Z. C. Wu, X. Xiao, iLoc-Euk: A multi-label classifier for predicting the subcellular localization of singleplex and multiplex eukaryotic proteins, PLoS One, 6 (2011), e18258. https://doi.org/10.1371/journal.pone.0018258 doi: 10.1371/journal.pone.0018258
![]() |
[62] | S. Tang, L. Chen, iATC-NFMLP: Identifying classes of anatomical therapeutic chemicals based on drug networks, fingerprints and multilayer perceptron. Curr. Bioinf., 17 (2022), 814–824. |
[63] |
H. Zhao, Y. Li, J. Wang, A convolutional neural network and graph convolutional network-based method for predicting the classification of anatomical therapeutic chemicals, Bioinformatics, 37 (2021), 2841–2847. https://doi.org/10.1093/bioinformatics/btab204 doi: 10.1093/bioinformatics/btab204
![]() |
[64] |
W. Chen, H. Yang, P. Feng, H. Ding, H. Lin, iDNA4mC: identifying DNA N4-methylcytosine sites based on nucleotide chemical properties, Bioinformatics, 33 (2017), 3518–3523. https://doi.org/10.1093/bioinformatics/btx479 doi: 10.1093/bioinformatics/btx479
![]() |
[65] |
L. Wei, P. Xing, R. Su, G. Shi, Z. S. Ma, Q. Zou, CPPred-RF: A sequence-based predictor for identifying cell-penetrating peptides and their uptake efficiency, J. Proteome Res., 16 (2017), 2044–2053. https://doi.org/10.1021/acs.jproteome.7b00019 doi: 10.1021/acs.jproteome.7b00019
![]() |
[66] |
S. R. Safavian, D. Landgrebe, A survey of decision tree classifier methodology, T-SMCA, 21 (1991), 660–674. https://doi.org/10.1109/21.97458 doi: 10.1109/21.97458
![]() |
[67] |
C. Cortes, V. Vapnik, Support-vector networks, Mach. Learn., 20 (1995), 273–297. https://doi.org/10.1007/BF00994018 doi: 10.1007/BF00994018
![]() |
![]() |
![]() |
1. | Rebeca Klamerick Lima, Felipe Sousa Quintino, Tiago A. da Fonseca, Luan Carlos de Sena Monteiro Ozelim, Pushpa Narayan Rathie, Helton Saulo, Assessing the Impact of Copula Selection on Reliability Measures of Type P(X < Y) with Generalized Extreme Value Marginals, 2024, 5, 2673-3951, 180, 10.3390/modelling5010010 | |
2. | Melquisadec Oliveira, Felipe S. Quintino, Dióscoros Aguiar, Pushpa N. Rathie, Helton Saulo, Tiago A. da Fonseca, Luan Carlos de Sena Monteiro Ozelim, On the Stress–Strength Reliability of Transmuted GEV Random Variables with Applications to Financial Assets Selection, 2024, 26, 1099-4300, 441, 10.3390/e26060441 | |
3. | Felipe Sousa Quintino, Pushpa Narayan Rathie, Luan Carlos de Sena Monteiro Ozelim, Tiago Alves da Fonseca, Estimation of P(X < Y) Stress–Strength Reliability Measures for a Class of Asymmetric Distributions: The Case of Three-Parameter p-Max Stable Laws, 2024, 16, 2073-8994, 837, 10.3390/sym16070837 | |
4. | Rebeca Klamerick Lima, Felipe Sousa Quintino, Melquisadec Oliveira, Luan Carlos de Sena Monteiro Ozelim, Tiago A. da Fonseca, Pushpa Narayan Rathie, Multicomponent Stress–Strength Reliability with Extreme Value Distribution Margins: Its Theory and Application to Hydrological Data, 2024, 7, 2571-8800, 529, 10.3390/j7040032 | |
5. | Refah Alotaibi, Mazen Nassar, Zareen A. Khan, Ahmed Elshahhat, Statistical analysis of stress–strength in a newly inverted Chen model from adaptive progressive type-Ⅱ censoring and modelling on light-emitting diodes and pump motors, 2024, 9, 2473-6988, 34311, 10.3934/math.20241635 |
sign(γ) | Distribution | Reference | Parameter restriction |
0 | Gumbel | [25] | σ1=σ2 or σ1=2σ2 or σ2/σ1>1 |
1 | Fréchet | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[30] | μ1=μ2=0 | ||
[29] | μ1=μ2=0 and γ1=γ2 | ||
[33] | μ1=μ2=0 | ||
−1 | Weibull (min) | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[26] | μ1=μ2=0 and γ1=γ2=γ | ||
[27] | μ1=μ2=μ and γ1=γ2=γ | ||
[28] | μ1=μ2=0 | ||
*p and q are coprime integers. |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1151 | -0.0004 | 0.0330 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4280 | 0.0070 | 0.0502 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4591 | 0.0059 | 0.0525 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0796 | 0.0002 | 0.0241 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5277 | 0.0021 | 0.0466 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5669 | 0.0017 | 0.0492 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1387 | 0.0027 | 0.0389 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5129 | -0.0037 | 0.0462 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5362 | 0.0009 | 0.0461 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1025 | -0.0010 | 0.0313 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3873 | -0.0022 | 0.0463 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4159 | -0.0043 | 0.0511 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0573 | 0.0017 | 0.0270 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4275 | 0.0083 | 0.0530 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4807 | 0.0009 | 0.0474 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1120 | -0.0016 | 0.0277 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4279 | -0.0089 | 0.0474 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4500 | -0.0022 | 0.0506 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1234 | 0.0003 | 0.0302 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3701 | 0.0014 | 0.0488 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.4034 | -0.0045 | 0.0472 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0797 | 0.0025 | 0.0276 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4134 | 0.0026 | 0.0403 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4794 | -0.0055 | 0.0548 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1303 | -0.0032 | 0.0325 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.4068 | -0.0069 | 0.0457 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4344 | -0.0015 | 0.0452 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1147 | -0.0000 | 0.0106 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4355 | -0.0005 | 0.0151 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4645 | 0.0005 | 0.0156 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0798 | 0.0001 | 0.0084 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5299 | -0.0001 | 0.0159 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5693 | -0.0007 | 0.0151 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1414 | -0.0000 | 0.0112 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5096 | -0.0004 | 0.0162 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5374 | -0.0003 | 0.0158 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1018 | -0.0003 | 0.0098 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3852 | -0.0001 | 0.0152 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4119 | -0.0003 | 0.0152 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0587 | 0.0003 | 0.0074 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4366 | -0.0008 | 0.0154 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4814 | 0.0002 | 0.0150 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1106 | -0.0002 | 0.0101 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4196 | -0.0006 | 0.0157 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4471 | 0.0008 | 0.0150 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1243 | -0.0006 | 0.0103 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3722 | -0.0007 | 0.0158 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.3974 | 0.0014 | 0.0154 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0822 | 0.0000 | 0.0086 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4166 | -0.0005 | 0.0160 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4736 | 0.0004 | 0.0162 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1265 | 0.0007 | 0.0105 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.3998 | 0.0001 | 0.0155 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4323 | 0.0006 | 0.0151 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.0617 | 0.0624 | -0.0007 | 0.0244 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.4288 | 0.4306 | -0.0018 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 0.3 | 0.4491 | 0.4478 | 0.0012 | 0.0514 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.0906 | 0.0891 | 0.0015 | 0.0284 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.4443 | 0.4463 | -0.0020 | 0.0507 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 0.3 | 0.4717 | 0.4727 | -0.0010 | 0.0499 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.1299 | 0.1300 | -0.0001 | 0.0332 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.4419 | 0.4423 | -0.0004 | 0.0509 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 0.3 | 0.4611 | 0.4602 | 0.0008 | 0.0497 |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.1469 | 0.1474 | -0.0004 | 0.0359 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.4764 | 0.4774 | -0.0010 | 0.0486 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 1.0 | 0.4947 | 0.4940 | 0.0007 | 0.0504 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.1846 | 0.1847 | -0.0001 | 0.0382 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.4980 | 0.4985 | -0.0006 | 0.0520 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 1.0 | 0.5240 | 0.5235 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.2144 | 0.2164 | -0.0019 | 0.0431 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.4951 | 0.4925 | 0.0026 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 1.0 | 0.5128 | 0.5124 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.1868 | 0.1883 | -0.0015 | 0.0393 |
0.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.4906 | 0.4894 | 0.0013 | 0.0523 |
0.4 | 1.5 | 1.5 | 0.5 | 0.5 | 1.5 | 0.5076 | 0.5072 | 0.0004 | 0.0495 |
2.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.2237 | 0.2243 | -0.0006 | 0.0414 |
0.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.5056 | 0.5052 | 0.0004 | 0.0501 |
0.4 | 1.0 | 1.5 | 0.5 | 0.7 | 1.5 | 0.5318 | 0.5310 | 0.0008 | 0.0496 |
2.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.2461 | 0.2451 | 0.0011 | 0.0444 |
0.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.5046 | 0.5042 | 0.0004 | 0.0491 |
0.4 | 1.5 | 1.5 | 0.5 | 0.9 | 1.5 | 0.5220 | 0.5226 | -0.0006 | 0.0490 |
RV | Data Set | min | 1st Qu. | Median | Mean | 3rd Qu. | Max | n |
X1 | BBAS3 | -0.1057 | -0.0097 | 0.0019 | 0.0012 | 0.0136 | 0.0736 | 330 |
X2 | ITUB4 | -0.0492 | -0.0105 | 0.0004 | 0.0006 | 0.0109 | 0.0794 | 330 |
X3 | VALE3 | -0.0689 | -0.0140 | 0.0001 | -0.0002 | 0.0128 | 0.0989 | 330 |
X4 | VIIA3 | -0.1075 | -0.0344 | -0.006 | -0.0030 | 0.0231 | 0.1504 | 330 |
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0063 | 0.0219 | -0.2535 | -803.9168 | 0.0147 |
X2 | ITUB4 | -0.0064 | 0.0165 | -0.1545 | -870.9392 | 0.4299 |
X3 | VALE3 | -0.0095 | 0.0222 | -0.1631 | -774.0895 | 0.2331 |
X4 | VIIA3 | -0.0217 | 0.0396 | -0.1170 | -567.6979 | 0.6996 |
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5283 | 0.5242 | 0.5277 | (0.4883; 0.5676) |
P(X3<X2) | 0.5174 | 0.5242 | 0.5181 | (0.4766; 0.5569) |
P(X3<X4) | 0.4506 | 0.4364 | 0.4500 | (0.4055; 0.4944) |
RV | Data set | GEV | Generalized Logistic |
X1 | BBAS3 | -803.92 | - 833.43 |
X2 | ITUB4 | -870.94 | - 882.01 |
X3 | VALE3 | -774.09 | - 787.81 |
X4 | VIIA3 | -567.70 | - 569.60 |
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0128 | 0.0393 | -0.1157 | -116.3016 | 0.8076 |
X2 | ITUB4 | -0.091 | 0.0397 | -0.4319 | -126.5530 | 0.8292 |
X3 | VALE3 | -0.0197 | 0.0426 | -0.2003 | -114.0183 | 0.9365 |
X4 | VIIA3 | -0.0464 | 0.0750 | -0.1339 | -72.9854 | 0.5378 |
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5453 | 0.5522 | 0.5413 | (0.4483; 0.6334) |
P(X3<X2) | 0.5409 | 0.5522 | 0.5483 | (0.4421; 0.6358) |
P(X3<X4) | 0.4364 | 0.4328 | 0.4350 | (0.3379; 0.5350) |
RV | Data set | Min. | 1st Qu. | Median | Mean | 3rd Qu. | Max. | n |
X | Carbon fibers of 20 mm | 1.312 | 2.098 | 2.478 | 2.451 | 2.773 | 3.585 | 69 |
Y | Carbon fibers of 10 mm | 1.901 | 2.554 | 2.996 | 3.059 | 3.421 | 5.020 | 63 |
Data set | ˆμ | ˆσ | ˆγ | KS p-value |
X | 2.2781 | 0.4956 | -0.2851 | 0.9978 |
Y | 2.7904 | 0.5245 | -0.0747 | 0.8216 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.61 | 0.10 | 0.29 |
X2 | 1.00 | 0.15 | 0.31 | |
X3 | 1.00 | 0.03 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.63 | 0.15 | 0.31 |
X2 | 1.00 | 0.22 | 0.34 | |
X3 | 1.00 | 0.06 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.45 | 0.10 | 0.21 |
X2 | 1.00 | 0.15 | 0.24 | |
X3 | 1.00 | 0.04 | ||
X4 | 1.00 |
sign(γ) | Distribution | Reference | Parameter restriction |
0 | Gumbel | [25] | σ1=σ2 or σ1=2σ2 or σ2/σ1>1 |
1 | Fréchet | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[30] | μ1=μ2=0 | ||
[29] | μ1=μ2=0 and γ1=γ2 | ||
[33] | μ1=μ2=0 | ||
−1 | Weibull (min) | [25] | μ1=μ2 and (γ1=γ2 or γ2=2γ1 |
or γ2/γ1=p/q)* | |||
[26] | μ1=μ2=0 and γ1=γ2=γ | ||
[27] | μ1=μ2=μ and γ1=γ2=γ | ||
[28] | μ1=μ2=0 | ||
*p and q are coprime integers. |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1151 | -0.0004 | 0.0330 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4280 | 0.0070 | 0.0502 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4591 | 0.0059 | 0.0525 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0796 | 0.0002 | 0.0241 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5277 | 0.0021 | 0.0466 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5669 | 0.0017 | 0.0492 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1387 | 0.0027 | 0.0389 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5129 | -0.0037 | 0.0462 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5362 | 0.0009 | 0.0461 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1025 | -0.0010 | 0.0313 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3873 | -0.0022 | 0.0463 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4159 | -0.0043 | 0.0511 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0573 | 0.0017 | 0.0270 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4275 | 0.0083 | 0.0530 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4807 | 0.0009 | 0.0474 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1120 | -0.0016 | 0.0277 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4279 | -0.0089 | 0.0474 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4500 | -0.0022 | 0.0506 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1234 | 0.0003 | 0.0302 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3701 | 0.0014 | 0.0488 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.4034 | -0.0045 | 0.0472 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0797 | 0.0025 | 0.0276 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4134 | 0.0026 | 0.0403 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4794 | -0.0055 | 0.0548 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1303 | -0.0032 | 0.0325 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.4068 | -0.0069 | 0.0457 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4344 | -0.0015 | 0.0452 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.1147 | 0.1147 | -0.0000 | 0.0106 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -0.3 | 0.4350 | 0.4355 | -0.0005 | 0.0151 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -0.3 | 0.4650 | 0.4645 | 0.0005 | 0.0156 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.0798 | 0.0798 | 0.0001 | 0.0084 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -0.3 | 0.5298 | 0.5299 | -0.0001 | 0.0159 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -0.3 | 0.5686 | 0.5693 | -0.0007 | 0.0151 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.1414 | 0.1414 | -0.0000 | 0.0112 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -0.3 | 0.5092 | 0.5096 | -0.0004 | 0.0162 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -0.3 | 0.5371 | 0.5374 | -0.0003 | 0.0158 |
2.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.1015 | 0.1018 | -0.0003 | 0.0098 |
0.0 | 1.5 | -1.0 | 0.0 | 0.5 | -1.0 | 0.3851 | 0.3852 | -0.0001 | 0.0152 |
0.4 | 1.5 | -1.0 | 0.5 | 0.5 | -1.0 | 0.4116 | 0.4119 | -0.0003 | 0.0152 |
2.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.0590 | 0.0587 | 0.0003 | 0.0074 |
0.0 | 1.0 | -1.0 | 0.0 | 0.7 | -1.0 | 0.4358 | 0.4366 | -0.0008 | 0.0154 |
0.4 | 1.0 | -1.0 | 0.5 | 0.7 | -1.0 | 0.4816 | 0.4814 | 0.0002 | 0.0150 |
2.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.1104 | 0.1106 | -0.0002 | 0.0101 |
0.0 | 1.5 | -1.0 | 0.0 | 0.9 | -1.0 | 0.4190 | 0.4196 | -0.0006 | 0.0157 |
0.4 | 1.5 | -1.0 | 0.5 | 0.9 | -1.0 | 0.4478 | 0.4471 | 0.0008 | 0.0150 |
2.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.1237 | 0.1243 | -0.0006 | 0.0103 |
0.0 | 1.5 | -1.5 | 0.0 | 0.5 | -1.5 | 0.3715 | 0.3722 | -0.0007 | 0.0158 |
0.4 | 1.5 | -1.5 | 0.5 | 0.5 | -1.5 | 0.3989 | 0.3974 | 0.0014 | 0.0154 |
2.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.0822 | 0.0822 | 0.0000 | 0.0086 |
0.0 | 1.0 | -1.5 | 0.0 | 0.7 | -1.5 | 0.4160 | 0.4166 | -0.0005 | 0.0160 |
0.4 | 1.0 | -1.5 | 0.5 | 0.7 | -1.5 | 0.4739 | 0.4736 | 0.0004 | 0.0162 |
2.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.1271 | 0.1265 | 0.0007 | 0.0105 |
0.0 | 1.5 | -1.5 | 0.0 | 0.9 | -1.5 | 0.3999 | 0.3998 | 0.0001 | 0.0155 |
0.4 | 1.5 | -1.5 | 0.5 | 0.9 | -1.5 | 0.4329 | 0.4323 | 0.0006 | 0.0151 |
μ2 | σ2 | γ2 | μ1 | σ1 | γ1 | R | ˆRMC | Bias | RMSE |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.0617 | 0.0624 | -0.0007 | 0.0244 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 0.3 | 0.4288 | 0.4306 | -0.0018 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 0.3 | 0.4491 | 0.4478 | 0.0012 | 0.0514 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.0906 | 0.0891 | 0.0015 | 0.0284 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 0.3 | 0.4443 | 0.4463 | -0.0020 | 0.0507 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 0.3 | 0.4717 | 0.4727 | -0.0010 | 0.0499 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.1299 | 0.1300 | -0.0001 | 0.0332 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 0.3 | 0.4419 | 0.4423 | -0.0004 | 0.0509 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 0.3 | 0.4611 | 0.4602 | 0.0008 | 0.0497 |
2.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.1469 | 0.1474 | -0.0004 | 0.0359 |
0.0 | 1.5 | 1.0 | 0.0 | 0.5 | 1.0 | 0.4764 | 0.4774 | -0.0010 | 0.0486 |
0.4 | 1.5 | 1.0 | 0.5 | 0.5 | 1.0 | 0.4947 | 0.4940 | 0.0007 | 0.0504 |
2.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.1846 | 0.1847 | -0.0001 | 0.0382 |
0.0 | 1.0 | 1.0 | 0.0 | 0.7 | 1.0 | 0.4980 | 0.4985 | -0.0006 | 0.0520 |
0.4 | 1.0 | 1.0 | 0.5 | 0.7 | 1.0 | 0.5240 | 0.5235 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.2144 | 0.2164 | -0.0019 | 0.0431 |
0.0 | 1.5 | 1.0 | 0.0 | 0.9 | 1.0 | 0.4951 | 0.4925 | 0.0026 | 0.0497 |
0.4 | 1.5 | 1.0 | 0.5 | 0.9 | 1.0 | 0.5128 | 0.5124 | 0.0004 | 0.0480 |
2.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.1868 | 0.1883 | -0.0015 | 0.0393 |
0.0 | 1.5 | 1.5 | 0.0 | 0.5 | 1.5 | 0.4906 | 0.4894 | 0.0013 | 0.0523 |
0.4 | 1.5 | 1.5 | 0.5 | 0.5 | 1.5 | 0.5076 | 0.5072 | 0.0004 | 0.0495 |
2.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.2237 | 0.2243 | -0.0006 | 0.0414 |
0.0 | 1.0 | 1.5 | 0.0 | 0.7 | 1.5 | 0.5056 | 0.5052 | 0.0004 | 0.0501 |
0.4 | 1.0 | 1.5 | 0.5 | 0.7 | 1.5 | 0.5318 | 0.5310 | 0.0008 | 0.0496 |
2.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.2461 | 0.2451 | 0.0011 | 0.0444 |
0.0 | 1.5 | 1.5 | 0.0 | 0.9 | 1.5 | 0.5046 | 0.5042 | 0.0004 | 0.0491 |
0.4 | 1.5 | 1.5 | 0.5 | 0.9 | 1.5 | 0.5220 | 0.5226 | -0.0006 | 0.0490 |
RV | Data Set | min | 1st Qu. | Median | Mean | 3rd Qu. | Max | n |
X1 | BBAS3 | -0.1057 | -0.0097 | 0.0019 | 0.0012 | 0.0136 | 0.0736 | 330 |
X2 | ITUB4 | -0.0492 | -0.0105 | 0.0004 | 0.0006 | 0.0109 | 0.0794 | 330 |
X3 | VALE3 | -0.0689 | -0.0140 | 0.0001 | -0.0002 | 0.0128 | 0.0989 | 330 |
X4 | VIIA3 | -0.1075 | -0.0344 | -0.006 | -0.0030 | 0.0231 | 0.1504 | 330 |
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0063 | 0.0219 | -0.2535 | -803.9168 | 0.0147 |
X2 | ITUB4 | -0.0064 | 0.0165 | -0.1545 | -870.9392 | 0.4299 |
X3 | VALE3 | -0.0095 | 0.0222 | -0.1631 | -774.0895 | 0.2331 |
X4 | VIIA3 | -0.0217 | 0.0396 | -0.1170 | -567.6979 | 0.6996 |
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5283 | 0.5242 | 0.5277 | (0.4883; 0.5676) |
P(X3<X2) | 0.5174 | 0.5242 | 0.5181 | (0.4766; 0.5569) |
P(X3<X4) | 0.4506 | 0.4364 | 0.4500 | (0.4055; 0.4944) |
RV | Data set | GEV | Generalized Logistic |
X1 | BBAS3 | -803.92 | - 833.43 |
X2 | ITUB4 | -870.94 | - 882.01 |
X3 | VALE3 | -774.09 | - 787.81 |
X4 | VIIA3 | -567.70 | - 569.60 |
RV | Data set | ˆμ | ˆσ | ˆγ | −llmax | KS p-value |
X1 | BBAS3 | -0.0128 | 0.0393 | -0.1157 | -116.3016 | 0.8076 |
X2 | ITUB4 | -0.091 | 0.0397 | -0.4319 | -126.5530 | 0.8292 |
X3 | VALE3 | -0.0197 | 0.0426 | -0.2003 | -114.0183 | 0.9365 |
X4 | VIIA3 | -0.0464 | 0.0750 | -0.1339 | -72.9854 | 0.5378 |
R=P(X<Y) | ˆR | ˆRNP | ˆRBoot | 95% CI |
P(X3<X1) | 0.5453 | 0.5522 | 0.5413 | (0.4483; 0.6334) |
P(X3<X2) | 0.5409 | 0.5522 | 0.5483 | (0.4421; 0.6358) |
P(X3<X4) | 0.4364 | 0.4328 | 0.4350 | (0.3379; 0.5350) |
RV | Data set | Min. | 1st Qu. | Median | Mean | 3rd Qu. | Max. | n |
X | Carbon fibers of 20 mm | 1.312 | 2.098 | 2.478 | 2.451 | 2.773 | 3.585 | 69 |
Y | Carbon fibers of 10 mm | 1.901 | 2.554 | 2.996 | 3.059 | 3.421 | 5.020 | 63 |
Data set | ˆμ | ˆσ | ˆγ | KS p-value |
X | 2.2781 | 0.4956 | -0.2851 | 0.9978 |
Y | 2.7904 | 0.5245 | -0.0747 | 0.8216 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.61 | 0.10 | 0.29 |
X2 | 1.00 | 0.15 | 0.31 | |
X3 | 1.00 | 0.03 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.63 | 0.15 | 0.31 |
X2 | 1.00 | 0.22 | 0.34 | |
X3 | 1.00 | 0.06 | ||
X4 | 1.00 |
X1 | X2 | X3 | X4 | |
X1 | 1.00 | 0.45 | 0.10 | 0.21 |
X2 | 1.00 | 0.15 | 0.24 | |
X3 | 1.00 | 0.04 | ||
X4 | 1.00 |