Processing math: 75%
Research article

WG-ICRN: Protein 8-state secondary structure prediction based on Wasserstein generative adversarial networks and residual networks with Inception modules


  • Received: 06 December 2022 Revised: 18 January 2023 Accepted: 11 February 2023 Published: 20 February 2023
  • Protein secondary structure is the basis of studying the tertiary structure of proteins, drug design and development, and the 8-state protein secondary structure can provide more adequate protein information than the 3-state structure. Therefore, this paper proposes a novel method WG-ICRN for predicting protein 8-state secondary structures. First, we use the Wasserstein generative adversarial network (WGAN) to extract protein features in the position-specific scoring matrix (PSSM). The extracted features are combined with PSSM into a new feature set of WG-data, which contains richer feature information. Then, we use the residual network (ICRN) with Inception to further extract the features in WG-data and complete the prediction. Compared with the residual network, ICRN can reduce parameter calculations and increase the width of feature extraction to obtain more feature information. We evaluated the prediction performance of the model using six datasets. The experimental results show that the WGAN has excellent feature extraction capabilities, and ICRN can further improve network performance and improve prediction accuracy. Compared with four popular models, WG-ICRN achieves better prediction performance.

    Citation: Shun Li, Lu Yuan, Yuming Ma, Yihui Liu. WG-ICRN: Protein 8-state secondary structure prediction based on Wasserstein generative adversarial networks and residual networks with Inception modules[J]. Mathematical Biosciences and Engineering, 2023, 20(5): 7721-7737. doi: 10.3934/mbe.2023333

    Related Papers:

    [1] Hammad Siddiqi . Financial market disruption and investor awareness: the case of implied volatility skew. Quantitative Finance and Economics, 2022, 6(3): 505-517. doi: 10.3934/QFE.2022021
    [2] Takashi Kanamura . Diversification effect of commodity futures on financial markets. Quantitative Finance and Economics, 2018, 2(4): 821-836. doi: 10.3934/QFE.2018.4.821
    [3] Oleg S. Sukharev . The restructuring of the investment portfolio: the risk and effect of the emergence of new combinations. Quantitative Finance and Economics, 2019, 3(2): 390-411. doi: 10.3934/QFE.2019.2.390
    [4] Zhongxian Men, Tony S. Wirjanto . A new variant of estimation approach to asymmetric stochastic volatilitymodel
    . Quantitative Finance and Economics, 2018, 2(2): 325-347. doi: 10.3934/QFE.2018.2.325
    [5] Sang Phu Nguyen, Toan Luu Duc Huynh . Portfolio optimization from a Copulas-GJR-GARCH-EVT-CVAR model: Empirical evidence from ASEAN stock indexes. Quantitative Finance and Economics, 2019, 3(3): 562-585. doi: 10.3934/QFE.2019.3.562
    [6] Samuel Asante Gyamerah . Modelling the volatility of Bitcoin returns using GARCH models. Quantitative Finance and Economics, 2019, 3(4): 739-753. doi: 10.3934/QFE.2019.4.739
    [7] Cemile Özgür, Vedat Sarıkovanlık . An application of Regular Vine copula in portfolio risk forecasting: evidence from Istanbul stock exchange. Quantitative Finance and Economics, 2021, 5(3): 452-470. doi: 10.3934/QFE.2021020
    [8] Hongxuan Huang, Zhengjun Zhang . An intrinsic robust rank-one-approximation approach for currency portfolio optimization. Quantitative Finance and Economics, 2018, 2(1): 160-189. doi: 10.3934/QFE.2018.1.160
    [9] Ngo Thai Hung . Equity market integration of China and Southeast Asian countries: further evidence from MGARCH-ADCC and wavelet coherence analysis. Quantitative Finance and Economics, 2019, 3(2): 201-220. doi: 10.3934/QFE.2019.2.201
    [10] Ick Jin . Systematic ESG risk and hedge fund. Quantitative Finance and Economics, 2024, 8(2): 387-409. doi: 10.3934/QFE.2024015
  • Protein secondary structure is the basis of studying the tertiary structure of proteins, drug design and development, and the 8-state protein secondary structure can provide more adequate protein information than the 3-state structure. Therefore, this paper proposes a novel method WG-ICRN for predicting protein 8-state secondary structures. First, we use the Wasserstein generative adversarial network (WGAN) to extract protein features in the position-specific scoring matrix (PSSM). The extracted features are combined with PSSM into a new feature set of WG-data, which contains richer feature information. Then, we use the residual network (ICRN) with Inception to further extract the features in WG-data and complete the prediction. Compared with the residual network, ICRN can reduce parameter calculations and increase the width of feature extraction to obtain more feature information. We evaluated the prediction performance of the model using six datasets. The experimental results show that the WGAN has excellent feature extraction capabilities, and ICRN can further improve network performance and improve prediction accuracy. Compared with four popular models, WG-ICRN achieves better prediction performance.



    1. Introduction

    Portfolio selection aims at either maximizing the return or minimizing the risk. In 1952, Markowitz (1952) suggests to select the portfolio by minimizing the standard deviation at a given expected return under the assumption that asset returns are normally distributed. This means that standard deviation is chosen as the risk measure. Markowitz's work laid down the cornerstone for modern portfolio selection theory framework.

    Risk measures and probability distributions are two important constituents of the portfolio selection theory. Traditional Markowitz's model (Markowitz, 1952) is established based on normality assumption and standard deviation is chosen as the risk measure.

    One disadvantage of taking standard deviation (StD) as a risk measure is that the loss in the extreme cases tends to be underestimated. To overcome such a difficulty, the idea of Value at Risk (VaR) is also widely used in practice. Artzner et al. (1999) suggests that desirable risk measure should be "coherent". However, VaR does not fulfill the subadditivity condition as required by the definition of "coherence". Yiu (2004) proposed an optimal portfolio selection under Value-at-Risk. On the other hand, Expected Shortfall (ES) is coherent as a popular risk measure for portfolio selection that aims at averaging the tail uncertainties.

    It is well-known that financial data cannot be described satisfactorily by normal distribution. The normality assumption is restrictive and is generally violated due to financial market uncertainties and managers' risk aversion. As Behr and Ptter (2009) pointed out, alternatives for multivariate normal distribution are necessary for portfolio selection. A desirable alternative model should be able to explain tail heaviness, skewness, and excess kurtosis. Various heavy tailed distributions have been applied to portfolio selection problems. Among these, Mandelbrot (1997) concluded that the daily rate of return of stock price data exhibit heavy tailed distributions; Hu and Kercheval (2010) apply multivariate skewed t and student t distribution for efficient frontier analysis; Generalized hyperbolic distribution is extensively studied in (Behr and Ptter, 2009; Eberlein, 2001; Hellmich and Kassberger, 2011; Hu and Kercheval, 2007; Surya and Kurniawan, 2014; Socgnia and Wilcox, 2014), with special cases including hyperbolic distribution (Bingham and Kiesel, 2001; Eberlein and Keller, 1995), Variance Gamma distribution (Seneta, 2004), Normal Inverse Gaussian distribution (Barndor-Nielsen, 1995), etc.

    Recently, Asymmetric Laplace distribution has received various attention in the literature, to name a few, (Ayebo and Kozubowski, 2003; Kollo and Srivastava, 2005; Kozubowski and Podgrski, 1999; Kozubowski and Podgrski, 2001; Punathumparambath, 2012). Compared to Normal distribution, the Asymmetric Laplace distribution describes asymmetry, steep peak, and tail heaviness better. Portfolio selection models are extensively studied under Asymmetric Laplace framework. Zhu (2007), Kozubowski and Podgrski (2001) apply Asymmetric Laplace distribution to financial data. By assuming that the asset data is generated from autoregressive moving average (ARMA) time series models with Asymmetric Laplace noise, Zhu (2007) establish the asymptotic inference theory under very mild conditions and present methods of computing conditional Value at Risk (CVaR). Zhao et al. (2015) further propose a so-called mean-CVaR-skewness portfolio selection strategy under Asymmetric Laplace distribution, this model can be further transformed to quadratic programming problem with explicit solutions.

    In this paper, we extended Hu's work (Hu, 2010) to Asymmetric Laplace framework. We first derived the equivalence of mean-VaR/ES/Std-skewness-kurtosis models, and show that these models can be reduced to quadratic programming problem. Since Zhao et al. (2015) utilized moment estimation for parameter estimation of Asymmetric Laplace distribution which less efficient compare to maximum likelihood estimation. Taken into consideration of the normal mean-variance mixture of Asymmetric Laplace distribution, followed by Expectation-Maximization algorithm for multivariate Laplace distribution in Arslan (2010), we derived the EM algorithm for Asymmetric Laplace Distributions that outperforms moment estimation in Zhao et al. (2015). The advantage of the proposed EM algorithm is to alleviate the complicated calculation of Bessel function. This improves many existing methods of estimating Asymmetric Laplace distributions, for example, Hrlimann (2013), Kollo and Srivastava (2005) and Visk (2009). Extensive simulation studies and efficient frontier analysis are complemented to confirm that our algorithm performs better than moment estimation for parameter estimation.

    The rest of the article is organized as follows. In Section 2, properties of Asymmetric Laplace distributions and coherent risk measures are summarized. In Section 3, portfolio selection under Asymmetric Laplace framework are derived, complement with Expectation-Maximization (EM) algorithm for parameter estimation of Asymmetric Laplace distributions. In Section 4, simulation studies are provided to show the efficiency of the Expectation-Maximization procedure. Section 5 presents real data analysis of Asymmetric Laplace Distributions based portfolio selection models, followed by conclusive remarks in Section 6.


    2. Preliminary knowledge


    2.1. Asymmetric Laplace distribution

    Kotz et al. (2001) proposed the Asymmetric Laplace Distribution with density function

    f(x)=2exΣ1μ(2π)n/2|Σ|1/2(xΣ1x2+μΣ1μ)v/2Kv((2+μΣ1μ)(xΣ1x)), (2.1)

    denoted as XALn(μ,Σ). Here, n is the dimension of random vector X, v=(2n)/2 and Kv(u) is the modified Bessel function of the third kind with the following two popular representations:

    Kv(u)=12(u2)v0tv1exp{tu24t}dt,u>0, (2.2)
    Kv(u)=(u/2)vΓ(1/2)Γ(v+1/2)1eut(t21)v1/2dt,u>0,v1/2. (2.3)

    When μ=0n, we can obtain Symmetric Laplace distribution SL(Σ) with density

    f(x)=2(2π)n/2|Σ|1/2(xΣ1x/2)ν/2Kν(2xΣ1x).

    When n=1, we have Σ=σ11=σ. In such cases, (2.1) becomes the univariate Laplace distribution AL1(μ,σ) distribution with parameters μ and σ. The corresponding density function is

    f(x)=1γexp{|x|σ2[γμsign(x)]} with γ=μ2+2σ2. (2.4)

    The symmetric case (μ=0) leads to the univariate Laplace distribution SL1(0,σ).

    Figure 1 displays plot of symmetric densities and AL densities. Symmetric densities including standard normal distribution, student t distribution with 2 degrees of freedom, and univariate symmetric Laplace distribution, denoted as N(0,1),t(2),SL1(0,1). The student t distribution possesses heavier tail than normal distribution, whereas SL1(0,1) distribution imposes greater peakedness and heavier tail than normal case. As for plots of AL densities, when μ>0, the density skews to the right. On the other hand, when μ<0, the density skews to the left.

    Figure 1. Univariate densities.

    Important results of univariate and multivariate asymptotic Laplace distributions that will be used later on are presented below.

    Proposition 2.1. (See Kotz, 2001)

    (1). If X=(X1,,Xn) follows multivariate Asymmetric Laplace distribution, i.e., XALn(μ,Σ), n is the number of securities. The linear combination wX=w1X1++wnXn follows univariate Asymmetric Laplace distribution, i.e. wXAL1(μ,σ), with μ=wμ,σ=wΣw,w=(w1,,wn).

    (2). Assume that univariate random variable YAL1(μ,σ). To measure the asymmetry and peakedness of the distribution, define the skewness (Skew[Y]) and kurtosis (Kurt[Y]) as the third and fourth standardized moment of a random variable Y. Then,

    Skew[Y]=E(YEY)3[E(YEY)2]3/2=2μ3+3μσ2(μ2+σ2)3/2,Kurt[Y]=E(YEY)4[Var(Y)]2=9μ4+6σ4+18μ2σ2(μ2+σ2)2.

    (3). Let X=(X1,X2,,Xn)ALn(μ,Σ). Then the first and second order moments of X are

    E(X)=μandCov(X)=Σ+μμ.

    (4). The Asymmetric Laplace distribution can be represented as a mixture of normal vector and a standard exponential variable, i.e., XALn(μ,Σ) can be represented as

    X=μZ+Z1/2Y,

    where YNn(0,Σ),ZExp(1). This indicate that we can simulate multivariate Asymmetric Laplace random vector XALn(μ,Σ) as follows:

    1. Generate a multivariate normal variable YNn(0,Σ);

    2. Generate a standard exponential variable ZExp(1);

    3. Construct Asymmetric Laplace random vector as X=μZ+Z1/2Y.

    Figure 2 displays several realizations of bivariate Asymmetric Laplace distribution with different levels of asymmetry and peakedness.

    Figure 2. Bivariate Asymmetric Laplace data with μ cases: (a1, a2, a3): μ=(0,0); (b1, b2, b3): μ=(1,1); (c1, c2, c3): μ=(1,1). Covariance matrix Σ cases. (a1, b1, c1): σ11=σ22=1,σ12=σ21=0; (a2, b2, c2): σ11=σ22=1,σ12=σ21=0.8; (a3, b3, c3): σ11=σ22=1,σ12=σ21=0.8.

    2.2. Risk measures

    Since mean and covariance matrix cannot be used to characterize non-Gaussian distribution, alternative risk measures are necessary for portfolio selection problems. Artzner et al. (1999) suggests that a desirable risk measure should be defined fulfilling certain properties and such a risk measure is said to be coherent.

    A risk measure ϕ that maps a random variable to a real number is coherent if it satisfies the following conditions:

    1). Translation invariance: ϕ(l+h)=ϕ(l)+h, for all random losses l and all hR;

    2). Subadditivity: ϕ(l+h)ϕ(l)+ϕ(h), for all random losses l,h;

    3). Positive homogeneity: ϕ(λl)=λϕ(l) for all random losses l and all λ>0;

    4). Monotonicity: ϕ(l1)ϕ(l2) for all random losses l1,l2 with l1l2 almost surely.

    Standard deviation is not coherent in general excepting the Gaussian cases. VaR is coherent when the underlying distribution is elliptically distributed. Expected Shortfall, or the so-called conditional value at risk (CVaR) is a coherent risk measure since it always satisfies subadditivity, monotonicity, positive homogeneity, and convexity. For any fixed α(0,1), α-VaR is the α-quantile loss while α-ES is the average of all β-VaR for β(α,1). Both VaR and CVaR measure the potential maximal loss. VaR and ES can be written as

    VaRα=F1(α)andESα=E[L|L-VaRα]=1αVaRαVaRβdβ,

    where F() is the cumulative distribution function of loss L and ESα is the expected loss above VaRα. Thus, the estimation process are

    VaRαfX(x)dx=αandESα=1αVaRαxf(x)dx. (2.5)

    Under normality assumption, VaRα and ESα are

    VaRα=μ+σΦ1(1α),ESα=μ+σψ(Φ1(1α))α.

    where ψ() as the normal density distribution, and Φ1() is the quantile distribution.

    As shown in Hu et al. (2010), portfolio selected by minimizing standard deviation, VaRα, and ESα are the equivalent under elliptical distribution assumption.

    It is well-documented that asset securities are not normally distributed. As an alternative to Gaussian distribution, Asymmetric Laplace distribution exhibits tail-heaviness, skewness, and peakedness.


    3. Portfolio selection under ALD framework

    Let X=(X1,X2,,Xn)ALn(μP,ΣP) be the return vectors of n securities, and w=(w1,w2,,wn) is the allocation weight vector. Then, the portfolio is defined as

    P(w)=wX=ni=1wiXi.

    According to Proposition 2.1 (2), P(w)AL1(μ,σ) with μ=wμ,σ=wΣw.

    From Theorem 3.1–3.2 below, in order to select a portfolio under Asymmetric Laplace distribution, it suffices to obtain the unknown parameters μP and ΣP. Thus portfolio selection models under Asymmetric Laplace distribution lead to parameter estimation for ALn(μP,ΣP). Zhao et al. (2015) proposed the multi-objective portfolio selection model under Asymmetric Laplace framework and derived the simplified model that can be reformulated as quadratic programming problem. However, to estimate the unknown parameters, the authors adopt a moment estimation method that is less efficient compared to maximum likelihood method. Since Asymmetric Laplace distribution can be represented as a mixture of exponential distribution and multivariate normal distribution, we derived the Expectation-Maximization algorithm for parameter estimation of Asymmetric Laplace distribution. The algorithm for estimating these unknown parameters is discussed in Section 3.2.


    3.1. Portfolio selection theorems

    Theorem 3.1. Let X=(X1,,Xn)ALn(μP,ΣP) be a n-dimensional random vector that follow multivariate Asymmetric Laplace distribution, each element (Xi,i=1,2,,n) represent a stock. Let w be the weight vector and P(w)=wX=ni=1wiXi be the portfolio. Then, under Asymmetric Laplace framework, risk measures of StD, VaRα, and ESα at α(0,1) level formulated as

    Standard Deviation: StD(P(w))=σ2;Value at Risk: VaRα(P(w))=σ2γ+μlnαγ(γ+μ)σ2;Expected Shortfall: ESα(P(w))=σ2γ+μσ2γ+μlnαγ(γ+μ)σ2.

    Here, μ=wμP=mean(P(w)),σ=wΣPw=std(P(w)) and γ=μ2+2σ2.

    Proof. Let μP=(μ1,,μn) be the mean return vector of the securities (X1,,Xn) and ΣP=(σP)pi,j=1 be the scale matrix of (X1,,Xn). Denote the allocation vector by w=(w1,,wn). Then, the portfolio P(w)=ni=1wiXi follows univariate Asymmetric Laplace distribution with

    P(w)=ni=1wiXiAL1(μ,σ) with μ=ni=1μiwi,σ=(ni=1nj=1σPijwiwj)1/2.

    If μ=0n, the univariate symmetric Asymmetric Laplace distribution becomes AL1(0,σ) with density

    g(x)=1γexp{|x|σ2γ} with γ=2σ.

    Thus, standard deviation (StD) of portfolio P(w)=wX is

    StD(P(w))=+1γ|x|exp{γσ2|x|}dx=2+0xγexp{γσ2x}dx=2σ4γ3=σ2.

    According to the definition of VaRα and ESα as defined in (2.5) and univariate Asymmetric Laplace density (2.4), we have

    VaRα1γexp{|x|σ2[γμsgn(x)]}dx=α,σ2γ(γ+μ)exp{γ+μσ2VaRα}=α.

    Thus, VaRα and ESα are

    VaRα(P(w))=σ2μ2+2σ2+μlnα(μ2+2σ2+μμ2+2σ2)σ2=σ2γ+μlnαγ(γ+μ)σ2;ESα(P(w))=1αVaRαxfX(x)dx=1αVaRαx1γexp{|x|σ2[γμsgn(x)]}dx=σ2μ+μ2+2σ2σ2μ+μ2+2σ2ln{2α+α(μ2+μμ2+2σ2)σ2}=σ2γ+μσ2γ+μlnαγ(γ+μ)σ2.

    Then we have the following theorem.

    Theorem 3.2. Let XALn(μP,ΣP). Then, portfolio P(w)=wX with following models based on ESα, VaRα, and StD (as defined in Theorem 3.1)

    minwESα(P(w)) or minwVaRα(P(w)) or minwStD(P(w))maxwSkew[P(w)]=2μ3+3μσ2(μ2+σ2)3/2maxwKurt[P(w)]=9μ4+6σ4+18μ2σ2(μ2+σ2)2s.t.wμ=r0,w1=1.

    are equivalent. Here, μ=wμP=mean[P(w)],σ=wΣPw=std[P(w)],w=(w1,w2,,wn).

    Proof. Let g(μ,σ)=σ2μ+μ2+2σ2. Then, ESα[P(w)] and VaRα[P(w)] are

    VaRα[P(w)]=g(μ,σ)ln(2α+αμg(μ,σ))=g(μ,σ)[lnα+ln(2+μg(μ,σ))],ESα[P(w)]=g(μ,σ)g(μ,σ)ln(2α+αμg(μ,σ))=(1lnα)g(μ,σ)g(μ,σ)ln(2+μg(μ,σ)).

    Differentiating the above expressions with respect to σ, we have

    VaRα[P(w)]σ=g(μ,σ)σ[lnαln(2+μg(μ,σ))+μg(μ,σ)2+μg(μ,σ)]>0,ESα[P(w)]σ=g(μ,σ)σ[1lnαln(2+μg(μ,σ))+μg(μ,σ)2+μ2+g(μ,σ)]>0,

    where

    g(μ,σ)σ=[σ2μ+μ2+2σ2]σ=2σμ+μ2+σ2μ2+2σ2(μ2+μ2+2σ2)2>0.

    The derivative of skewness measure with respect to σ is

    Skew[P(w)]σ=[2μ3+3μσ2(μ2+σ2)3/2]σ=3μσ3(μ2+σ2)5/2<0.

    The derivative of kurtosis measure with respect to σ is

    Kurt[P(w)]σ=9μ4+6σ4+18μ2σ2(μ2+σ2)2σ=12μ4σ312μ2σ5(μ2+σ2)4<0.

    The monotonicity of VaRα[P(w)], ESα[P(w)], Skew[P(w)], and Kurt[P(w)] with respect to σ indicate that the portfolio selection problems based on these risk measures are equivalent. This means that minimizing VaRα[P(w)], ESα[P(w)], StD[P(w)] are equivalent to minimizing wΣPw.


    3.2. Parameter estimation of Asymmetric Laplace distribution

    Assume X=(X1,X2,,Xn)ALn(μ,Σ). Let x1,x2,,xTRn be the T observations. We aim at fitting a multivariate Asymmetric Laplace distribution ALn(μ,Σ) with unknown parameters μ,Σ.

    Hrlimann (2013), Kollo and Srivastava (2005), Visk (2009) consider moment matching methods that is less efficient than maximum likelihood estimation. Kotz et al. (2002) and Kotz et al. (2001) presented the maximum likelihood estimators for parameter estimation of Asymmetric Laplace distributions. However, maximum likelihood estimation require computation of complicated Bessel function. Thus we derived the expectation-maximization algorithm for parameter estimation of Asymmetric Laplace distribution.


    3.2.1. Moment estimation

    As Zhao et al. (2015) pointed out, according to Proposition 2.1 (3), Asymmetric Laplace distribution can be estimated via moment method (Moment-AL) with

    ˆμ=ˉxandˆΣ=cov(X)ˆμˆμ,

    where ˉx=1nni=1xi,Cov(X)=ni=1(xiˉx)T(xiˉx).


    3.2.2. Maximum likelihood estimation

    Consider sample points x1,x2,,xn and density function of Asymmetric Laplace distribution as defined in (2.1). Taken logarithm with respect to likelihood function, the log-likelihood is

    (μ,Σ)=lnL(μ,Σ)=Tt=1lnf(xt;μ,Σ)=Tt=1xtΣ1μ+Tln2Tn2ln(2π)T2ln(|Σ|)+ν2Tt=1ln(xtΣ1xt)νT2ln(2+μΣ1μ)+Tt=1lnKv{(2+μΣ1μ)(xtΣ1xt)}.

    Generally, we can directly maximize the log-likelihood function (μ,Σ) with respect to parameters μ,Σ and thus obtain the maximum likelihood estimator. Unfortunately, the density function involves modified Bessel function of the third kind with density (2.2), (2.3) that are too complex and complicated for numerical maximization. However, Gaussian-Exponential mixture representation of the Asymmetric Laplace distribution allows us to employ the expectation-maximization algorithm without involving modified Bessel functions.


    3.2.3. Expectation-maximization algorithm

    Then we derive the Expectation-Maximization algorithm for parameter estimation of multivariate Asymmetric Laplace Distribution (mALD), we follow the EM derivation for Multivariate Skew Laplace distribution in Arslan (2010).

    Let X=(X1,X2,,Xn) be Asymmetric Laplace distributed random vector. Proposition 2.1 suggests that X can be generated from a latent random variable Z=z through multivariate Gaussian distribution with zμ,zΣ, i.e. X|Z=zNn(zμ,zΣ),ZExp(1) with density

    fX|Z(x,z)=1(2π)n/2|zΣ|1/2exp{12(xzμ)(zΣ)1(xzμ)},fZ(z)=ez1{z0}.

    Thus the joint density function of X and Z is

    fX,Z(x,z)=fX|Z(x,z)fZ(z)=1(2π)n2zn2|Σ|12exp{12zxΣ1x+xΣ1μz2μΣ1μz1{z0}}.

    Suppose that there are T observations X1,,XT generated from the latent random variables z1,z2,,zT respectively. The complete data is defined as {(xt,zt)},t=1,2,,T. In the EM algorithm, xt and zt are the observed and missing data respectively. The log-likelihood up to an additive constant can be written as

    ˜L(μ,Σ)=Tt=1lnfX,Z(xt,zt)=T2ln|Σ|12Tt=11ztxtΣ1xt+Tt=1xtΣ1μ12μΣ1μTt=1ztTt=1zt1{zt0}.

    Note that the last term of the above equation does not contain any unknown parameters and thus is negligible. Then, the E-step becomes

    E(˜L(μ,Σ)|xt,ˆμ,ˆΣ)T2ln|Σ|+Tt=1xtΣ1μ12Tt=1E(z1t|xt,ˆμ,ˆΣ)xtΣ1xt12μΣ1μTt=1E(zt|xt,ˆμ,ˆΣ).

    where E(zt|x,ˆμ,ˆΣ) and E(z1t|x,ˆμ,ˆΣ) are the conditional expectations of zt and z1t given xt and the current estimates ˆμ,ˆΣ.

    To evaluate conditional expectations E(z1t|xt,ˆμ,ˆΣ) and E(zt|xt,ˆμ,ˆΣ), we need the conditional density of Z given X, fZ|X. After some straightforward algebra, the conditional distribution of Z given X is an inverse Gaussian distribution with density function

    fZ|X(z|x,μ,Σ)=fX,Z(x,z)fX(x)=1(2π)n2zn2|Σ|12exp{12zxΣ1x+xΣ1μz2μΣ1μz1{z0}}2exΣ1μ(2π)n/2|Σ|1/2(xΣ1x2+μΣ1μ)v/2Kv((2+μΣ1μ)(xΣ1x))=(2+μΣ1μxΣ1x)v/2zn/2exp{12[z1xΣ1x+z(μΣ1μ+z1{z0})]}2Kv((2+μΣ1μ)(xΣ1x)). (3.1)

    Lemma 3.1. (GIG (Stacy, 1962)) A random variable X follows Generalized Inverse Gaussian distribution(denoted as XN(λ,χ,ψ)) if its density function could be represented as

    f(x)=χλ(χψ)λ2Kλ(χψ)xλ1exp{12(χx1+ψx)},x>0.

    where Kλ denotes the third kind modified Bessel function, and the parameters satisfy

    {χ>0,ψ0,ifλ<0;χ>0,ψ>0,ifλ=0;χ0,ψ>0,ifλ>0.

    After some algebraic manipulations, it is easy to show that Z|X follows Generalized Inverse Gaussian distribution:

    Z|XN(2n2,xΣ1x,2+μΣ1μ).

    If χ>0,ψ>0, the moments could be calculated through the following formulas:

    E(Xα)=(χψ)α/2Kλ+α(χψ)Kλ(χψ),αR,E(lnX)=dE(Xα)dα|α=0.

    Denote χ=xΣ1x,ψ=2+μΣ1μ. Then, Z|XN(2n2,χ,ψ). From the conditional density function of (3.1), we can obtain the conditional expectations with the following moment properties:

    at=E(zt|xt,ˆμ,ˆΣ)=χtψKn22(χtψ)Kn21(χtψ),t=1,2,,T;bt=E(z1t|xt,ˆμ,ˆΣ)=ψχtKn2(χtψ)Kn21(χtψ),t=1,2,,T.

    where χt=xtΣ1xt, R(λ)=Kλ+1(x)Kλ(x) is strictly decreasing in x with limxRλ(x)=1 and limx0+Rλ(x)=. Thus, at>0,bt>0,t=1,2,,T.

    Finally, if the conditional expectation E(zt|xt,ˆμ,ˆΣ) and E(z1t|xt,ˆμ,ˆΣ) are replaced by at and bt respectively, the objective function becomes

    Q(μ,Σ|xt,ˆμ,ˆΣ)=T2ln|Σ|+Tt=1xtΣ1μ12Tt=1btxtΣ1xt12μΣ1μTt=1at. (3.2)

    Denote S=Σ1. The objective function (3.2) becomes

    Q(μ,Σ|xt,ˆμ,ˆS)=T2ln|S|+Tt=1xtSμ12Tt=1btxtSxt12μSμTt=1at. (3.3)

    Taking derivative of objective function (3.3) with respect to μ,S, we obtain

    Qμ=Tt=1xtSTt=1atμS=0,QS=T2S112Tt=1btxtxt+Tt=1xtμ12Tt=1atμμ=0.

    Substituting S by Σ and setting these derivatives to zero yield

    Tt=1xtΣ1Tt=1atμΣ1=0,T2Σ12Tt=1btxtxt+Tt=1xtμ12Tt=1atμμ=0.

    Thus, maximization of the objective function Q(μ,Σ|xt,ˆμ,ˆΣ) can be achieved by the following iterative updating formulas:

    ˆμ=ˉxˉa;ˆΣ=¯btxtxtˉxˉxˉa.

    where ˉa,ˉb stand for the average of {at}Tt=1 and {bt}Tt=1 respectively and ˉx is the average of {xt}Tt=1. In what follows, we present the iterative reweighted Expectation-Maximization algorithm for parameter estimation of Asymmetric Laplace distribution.

    Algorithm 1 Iterative reweighting algorithm
    1. Set iteration number k=1 and select initial estimates for parameters μ(0),Σ(0).
    2. (E-Step) At k-th iteration with current estimates μ(k),Σ(k), define the corresponding log-likelihood as
                            l(k)=logTt=1f(xt|μ(k),Σ(k)),k=1,2,.
    With notations χt=xtΣ1xt,ψ=2+μΣ1μ, we can obtain iterative weights
                            at=E(zt|xt,ˆμ,ˆΣ)=χtψK2n2(χtψ)K1n2(χtψ),t=1,2,,T;
                            bt=E(z1t|xt,ˆμ,ˆΣ)=ψχtKn2(χtψ)K1n2(χtψ),t=1,2,,T.
    3. (M-Step) Employ the following iteration formulas to calculate the new estimates μ(k+1),Σ(k+1) at (k+1)-th iteration:
                            μ(k+1)=ˉxˉa,Σ(k+1)=¯btxtxtˉxˉxˉa.(3.4)
    The log-likelihood at (k+1)-th iteration becomes
                            l(k+1)=logTt=1f(xt|μ(k+1),Σ(k+1)).
    4. Repeat these iteration steps until convergence with criterion l(k+1)l(k)<ε, where ε>0 is a small number that control the convergence precision, for convenience, we take ε=1e16.

    4. Simulation studies

    To evaluate the performance of portfolio selection models and parameter estimation methods in Section 3, we generate 100 datasets from Gaussian distribution and Asymmetric Laplace distribution respectively. Each dataset consists of T=200 observations with the following parameter settings: Case (1): n=3,μ=(0.03,0.06,0.09); Case (2): n=5,μ=(0.01,0.02,0.06,0.08,0.09); Case (3): n=10,μ=(0.01,0.02,0.03,,0.10). For each case, we set Σ=diag(μ/10). All the simulation studies are carried out on a PC with Intel Core i7 3.6 GHz processor under R platform.

    Each dataset are estimated under both multivariate Gaussian and Asymmetric Laplace distribution. ALD (EM-AL) is estimated using the EM algorithm described in Section 3.2. We evaluate the estimation performance using Bias measure, defined as Bias=ˆμμ1+ˆΣΣ1. The mean log-likelihood and mean bias of the simulated 200 datasets are reported in Table 1.

    Table 1. Model fitting results of Gaussian data and Asymmetric Laplace data using Gauss Model and EM-AL Model.
    Gaussian Data
    Log-Likelihood Bias
    Gauss EM-AL Gauss Moment-AL EM-AL
    Case (1) 709.6159 641.2472 0.0153 0.0451 0.0183
    Case (2) 1363.0231 1260.3676 0.0280 0.0886 0.0319
    Case (3) 2593.1151 2432.2192 0.0672 0.3309 0.0766
    Asymmetric Laplace Data
    Log-Likelihood Bias
    Gauss EM-AL Gauss Moment-AL EM-AL
    Case (1) 619.1798 735.0847 0.0542 0.0252 0.0226
    Case (2) 1250.7110 1463.9149 0.0870 0.0376 0.0302
    Case (3) 2432.3401 2919.9878 0.3750 0.1304 0.0832
     | Show Table
    DownLoad: CSV

    Table 1 indicate that if the model is correctly specified, the estimation performance is always the best in terms of bias. If the data is generated from Gaussian distribution, the estimation from Gaussian model is the best, so does Asymmetric Laplace distribution. If data is generated from Gaussian distribution, then the estimation log-likelihood of Gaussian model is larger than Asymmetric Laplace distribution, this is true for Asymmetric Laplace data as well.

    Figure 3 show that for Gaussian data, since Gaussian data fit the model better, efficient frontiers under Gaussian data are more close to Gaussian models; Figure 4 indicate that for generated Asymmetric Laplace data, efficient frontiers nearly equivalent to true Asymmetric Laplace data. Figure 34 suggest that we can first modeling data using Gaussian and Asymmetric Laplace distribution, and use the fitted log-likelihood to determine the distribution, then we evaluate the performance with the corresponding efficient frontier analysis.

    Figure 3. Efficient frontiers of simulated Gaussian data of case (1)–(3) using Gaussian and EM-AL model.
    Figure 4. Efficient frontiers of simulated Asymmetric Laplace data of case (1)–(3) using Gaussian and EM-AL model.

    5. Real data analysis

    In this section, we apply our proposed methodology to two real financial datasets, Hang Seng Index and Nasdaq Index, both datasets are downloaded from Bloomberg, with daily data range from January 4, 2011, to December 29, 2017. The variable of interest is the rate of returns multiplied by the annualized ratio \sqrt{252}, formulated as

    \text{LogRet}\, (t) = \sqrt{252} \Big\{ \log\big(\text{price}[t+1]\big)-\log\big(\text{price}[t]\big) \Big\}, \quad t = 1, 2, \cdots, 1721.

    These two datasets are analyzed through efficient frontier analysis under ALD framework on \texttt{R} platform. Theorem 3.1–3.2 indicate that portfolio selection models under ALD framework can be reduced to the following quadratic programming problem:

    \min\limits_{{\boldsymbol{w}}} \sigma^2 = {\boldsymbol{w}}' {\bf{\Sigma}} {\boldsymbol{w}} \quad \text{ s.t. } \quad {\boldsymbol{w}}'{\boldsymbol{\mu}} = r_0\, , {\boldsymbol{w}}'{\boldsymbol{1}} = 1.

    with explicit solution (see Lai and Xing, 2008) as follows:

    \hat{{\boldsymbol{w}}} = \frac{D-r_0B}{AD-B^2}{\bf{\Sigma}}^{-1}{\boldsymbol{1}} + \frac{r_0A-B}{AD-B^2}{\bf{\Sigma}}^{-1}{\boldsymbol{\mu}}. (5.1)

    Here, A = {\bf{1}}'{\bf{\Sigma}}^{-1}{\bf{1}}, B = {\bf{1}}'{\bf{\Sigma}}^{-1}{\boldsymbol{\mu}} and D = {\boldsymbol{\mu}}'{\bf{\Sigma}}^{-1}{\boldsymbol{\mu}}.


    5.1. Example 1: Hang seng index

    In the first example, we construct a portfolio consisting of 8 Hang Seng indexes: HK1, HK175, HK2007, HK2318, HK4, HK6, HK66. The summary descriptive statistics are reported in Table 2. It is clear that the all stock returns exhibit larger skewness and kurtosis. The median of these stocks are close to zero, the log-likelihood of Asymmetric Laplace distribution is larger than gaussian distribution, indicating that Asymmetric Laplace distribution would be a good fit than gaussian distribution.

    Table 2. Hang Seng data statistics.
    Descriptive Statistics
    StD Mean Median Skewness Kurtosis Jarq.Test Jarq.Prob
    HK1 19.7819 0.2370 0.0000 3.9601 71.8163 375244.7434 0.0000
    HK175 3.7101 0.2183 0.0000 1.1510 28.6192 59264.8528 0.0000
    HK2007 2.1968 0.1115 0.0000 0.5928 16.5623 19825.4392 0.0000
    HK2318 12.6007 0.3431 0.0000 0.6174 6.2371 2908.6947 0.0000
    HK4 5.8690 0.0409 0.0000 0.4894 7.6362 4263.7750 0.0000
    HK6 13.0712 0.1517 0.0000 -1.1430 20.3112 30037.3385 0.0000
    HK66 5.8708 0.1545 0.0000 -1.7941 19.9392 29510.3684 0.0000
    Gaussian EM-AL
    Log-likelihood -39707.45 -37865.83
    Parameter Estimation
    HK1 HK175 HK2007 HK2318 HK4 HK6 HK66
    \mu 0.2370 0.2183 0.1115 0.3431 0.0409 0.1517 0.1545
    \Sigma HK1 HK175 HK2007 HK2318 HK4 HK6 HK66
    HK1 406.4426 12.8089 11.4547 130.0934 68.3857 94.4069 60.3087
    HK175 12.8089 7.3656 1.6330 12.0356 4.7467 4.8725 3.6070
    HK2007 11.4547 1.6330 3.8506 9.9261 4.6830 3.9055 2.4395
    HK2318 130.0934 12.0356 9.9261 174.0423 42.1879 48.2877 35.0499
    HK4 68.3857 4.7467 4.6830 42.1879 44.6336 26.5314 17.6232
    HK6 94.4069 4.8725 3.9055 48.2877 26.5314 205.1471 32.5570
    HK66 60.3087 3.6070 2.4395 35.0499 17.6232 32.5570 41.6153
     | Show Table
    DownLoad: CSV

    Then we fit the data to Asymmetric Laplace distribution through EM algorithm as described in Section 3.2. Parameter estimation results are displayed in Table 2, we construct portfolios under Asymmetric Laplace framework at different levels of expected return. Consider increasing target expected return values

    r_0 = 0.040\, , 0.0745\, , 0.1081\, , 0.1417\, , 0.1753\, , 0.2088\, , 0.2424\, , 0.2760\, , 0.3096\, , 0.3431\, .

    Portfolio selection results are summarized in Table 3, with kurtosis, skewness, sharpe ratio, VaR and ES results at \alpha = 0.01, 0.05, 0.10 levels.

    Table 3. Efficient frontier results of Hang Seng data.
    r \mu \sigma Skew Kurt Sharpe VaR_{0.01} ES_{0.01} VaR_{0.05} ES_{0.05} VaR_{0.10} ES_{0.10}
    1 0.0409 0.0409 2.5461 0.0482 6.0016 0.0161 6.9429 8.7229 4.0782 5.8582 2.8444 4.6244
    2 0.0745 0.0745 2.0565 0.1086 6.0079 0.0362 5.5080 6.9254 3.2268 4.6442 2.2444 3.6618
    3 0.1081 0.1081 1.7299 0.1869 6.0233 0.0625 4.5256 5.6960 2.6420 3.8124 1.8308 3.0011
    4 0.1417 0.1417 1.6650 0.2537 6.0430 0.0851 4.2684 5.3771 2.4841 3.5928 1.7156 2.8243
    5 0.1753 0.1753 1.8891 0.2763 6.0510 0.0928 4.8095 6.0606 2.7960 4.0471 1.9288 3.1799
    6 0.2088 0.2088 2.3199 0.2682 6.0480 0.0900 5.9208 7.4601 3.4434 4.9827 2.3764 3.9157
    7 0.2424 0.2424 2.8656 0.2523 6.0425 0.0846 7.3493 9.2579 4.2774 6.1861 2.9544 4.8631
    8 0.2760 0.2760 3.4724 0.2372 6.0375 0.0795 8.9467 11.2679 5.2108 7.5321 3.6018 5.9231
    9 0.3096 0.3096 4.1134 0.2247 6.0337 0.0753 10.6386 13.3966 6.1999 8.9578 4.2882 7.0462
    10 0.3431 0.3431 4.7749 0.2147 6.0307 0.0719 12.3871 15.5962 7.2222 10.4313 4.9978 8.2069
     | Show Table
    DownLoad: CSV

    The efficient frontier tendencies are displayed in Figure 5. It is suggested that aggressive investors should impose higher confidence levels and conservative investors may choose smaller confidence levels. Figure 6 depicts the kurtosis, skewness, and sharpe ratio tendency of portfolio selection models. Results show that Sharpe Ratio, Skewness and Kurtosis increase fast and decreases slowly down as the target expected returns increases.

    Figure 5. ALD Efficient Frontier of Hang Seng Data.
    Figure 6. Skewness, Kurtosis and Sharpe Ratio Tendency of Hang Seng Data.

    5.2. Example 2: Nasdaq index

    In the second example, we consider Nasdaq index, including CTRP, MNST, NFLX, NTES, NVDA, TTWO, and report the descriptive statistics in Table 4. All the indexes exhibit significant skewness and kurtosis. Jarq.Test results indicate that this dataset deviates from normality significantly. We fit the log returns data to Gaussian and Asymmetric Laplace distributions. Since Asymmetric Laplace model achieve higher log-likelihood results compared to Gaussian model, we choose EM-AL model for data fitting. Parameter estimation results are displayed in Table 4.

    Table 4. Nasdaq data statistics.
    Descriptive Statistics
    StD Mean Median Skewness Kurtosis Jarq.Test Jarq.Prob
    CTRP 12.5853 0.2100 0.0000 1.5907 16.5100 20786.4126 0.0000
    MNST 10.0679 0.4904 0.1587 1.9632 25.7988 50065.6119 0.0000
    NFLX 32.9311 1.5015 -0.0079 1.0314 20.0251 29796.6074 0.0000
    NTES 58.0910 2.7817 1.2700 1.2457 26.0418 50315.0307 0.0000
    NVDA 25.2360 1.6024 0.3175 1.7413 40.3981 120864.0386 0.0000
    TTWO 12.2826 0.8786 0.1587 2.1392 52.3926 203127.9368 0.0000
    Gaussian EM-AL
    Log-Likelihood -46400.25 -42798.03
    Parameter Estimation
    CTRP MNST NFLX NTES NVDA TTWO
    \mu 0.2100 0.4904 1.5015 2.7817 1.6024 0.8786
    \Sigma CTRP MNST NFLX NTES NVDA TTWO
    CTRP 187.1236 25.4281 129.9104 246.6268 46.1539 40.0369
    MNST 25.4281 116.7107 51.2572 82.0634 26.2438 23.5405
    NFLX 129.9104 51.2572 900.1261 327.6980 122.5361 83.8650
    NTES 246.6268 82.0634 327.6980 2380.8894 213.1414 121.1493
    NVDA 46.1539 26.2438 122.5361 213.1414 259.7019 53.8530
    TTWO 40.0369 23.5405 83.8650 121.1493 53.8530 111.0020
     | Show Table
    DownLoad: CSV

    Then we consider increasing target expected returns

    r_0 = 0.2100, 0.4958, 0.7815, 1.0673, 1.3530, 1.6388, 1.9245, 2.2102, 2.4960, 2.7817.

    Results of skewness, kurtosis, sharpe ratio and VaR, ES results are summarized in Table 5. Figure 7 displays the efficient frontiers at confidence level \alpha = 0.01, 0.05, 0.10. These results show that the portfolio capture higher risk at higher \alpha levels. Figure 8 displays the skewness, kurtosis, and Sharpe Ratio Tendency. The optimal portfolios can be obtained from (5.1) with the corresponding VaR, ES, skewness, kurtosis and Sharpe Ratio.

    Table 5. Efficient frontier analysis of Nasdaq data.
    r \mu \sigma Skew Kurt Sharpe VaR_{0.01} ES_{0.01} VaR_{0.05} ES_{0.05} VaR_{0.10} ES_{0.10}
    1 0.2100 0.2100 8.5237 0.0739 6.0036 0.0246 23.0672 28.9903 13.5344 19.4575 9.4288 15.3519
    2 0.4958 0.4958 7.5957 0.1951 6.0254 0.0653 19.8220 24.9509 11.5675 16.6963 8.0125 13.1413
    3 0.7815 0.7815 7.8219 0.2973 6.0590 0.0999 19.7857 24.9397 11.4907 16.6447 7.9183 13.0723
    4 1.0673 1.0673 9.1167 0.3472 6.0806 0.1171 22.7065 28.6414 13.1547 19.0896 9.0409 14.9758
    5 1.3530 1.3530 11.1127 0.3608 6.0870 0.1218 27.5608 34.7713 15.9560 23.1665 10.9581 18.1686
    6 1.6388 1.6388 13.5025 0.3597 6.0865 0.1214 33.4993 42.2627 19.3952 28.1586 13.3208 22.0843
    7 1.9245 1.9245 16.1117 0.3541 6.0838 0.1194 40.0422 50.5132 23.1898 33.6608 15.9318 26.4028
    8 2.2102 2.2102 18.8494 0.3478 6.0808 0.1173 46.9392 59.2083 27.1927 39.4619 18.6883 30.9575
    9 2.4960 2.4960 21.6671 0.3418 6.0781 0.1152 54.0563 68.1800 31.3251 45.4488 21.5353 35.6590
    10 2.7817 2.7817 24.5371 0.3365 6.0757 0.1134 61.3178 77.3329 35.5424 51.5576 24.4416 40.4567
     | Show Table
    DownLoad: CSV
    Figure 7. ALD efficient frontier of Hang Seng data.
    Figure 8. Skewness, kurtosis and Sharpe Ratio tendency of Nasdaq data.

    Table 5 and Figure 7 suggests that as r_0 increases, all ES (ES_{0.01}, ES_{0.05}, ES_{0.10}) increases, indicating that higher return is derived from higher risk. It is interesting that under the ALD assumption, as r_0 increases, Sharpe ratio and skewness first decreases then increases accordingly. As \alpha increases, VaR and ES measures decreases. Thus, conservative investors can choose larger \alpha levels and aggressive investors would select smaller \alpha levels.


    6. Conclusion and prospects

    In this paper, we derive several equivalent portfolio selection models under ALD framework, these models can be transformed to quadratic programming problem with explicit solutions. The Expectation-Maximization algorithm for parameter estimation of Asymmetric Laplace distribution is obtained and outperforms moment estimation.

    There are several advantages of Asymmetric Laplace distribution based models. First, the equivalence of risk measures such as VaR, ES and StD at maximization of skewness and minimization of kurtosis faciliate portfolio selection models significantly. Second, confidence levels of these models offer investors various portfolio selection choices. Conservative investors can choose larger \alpha levels and aggressive investors can select smaller \alpha levels. Finally, ALD model is able to explain skewness and kurtosis in financial data. Therefore, the Asymmetric Laplace distribution can be widely applied to handle real financial datasets.

    We may further extend the Asymmetric Laplace based portfolio selection model to cases of mixture Asymmetric Laplace distributions. Another direction is to combine clustering techniques (see Dias et al., 2015; Iorio et al., 2018) with Asymmetric Laplace distribution for portfolio selection of time series models.


    Acknowledgments

    Chi Tim, Ng's work is supported by the 2016 Chonnam National University Research Program grant (No. 2016–2762).


    Conflict of interest

    The authors declare no conflict of interest.




    [1] A. W. Senior, R. Evans, J. Jumper, J. Kirkpatrick, L. Sifre, T. Green, et al., Improved protein structure prediction using potentials from deep learning, Nature, 577 (2020), 706–710. https://doi.org/10.1038/s41586-019-1923-7 doi: 10.1038/s41586-019-1923-7
    [2] J. Jumper, R. Evans, A. Pritzel, T. Green, M. Figurnov, O. Ronneberger, et al., Highly accurate protein structure prediction with AlphaFold, Nature, 596 (2021), 583–589. https://doi.org/10.1038/s41586-021-03819-2 doi: 10.1038/s41586-021-03819-2
    [3] J. Zhou, O. Troyanskaya, Deep supervised and convolutional generative stochastic network for protein secondary structure prediction, in Proceedings of the 31st International Conference on Machine Learning (ICML-14), 32 (2014), 745–753.
    [4] A. Yaseen, Y. H. Li, Template-based c8-scorpion: A protein 8-state secondary structure prediction method using structural information and context-based features, BMC Bioinformatics, 15 (2014). https://doi.org/10.1186/1471-2105-15-S8-S3 doi: 10.1186/1471-2105-15-S8-S3
    [5] W. Kabsch, C. Sander, Dictionary of protein secondary structure, Biopolymers, 22 (1983), 2577–2637.
    [6] B. Rost, C. Sander, Combining evolutionary information and neural networks to predict protein secondary structure, Proteins., 19 (1994), 55–72. https://doi.org/10.1002/prot.340190108 doi: 10.1002/prot.340190108
    [7] Y. Yang, J. Gao, J. Wang, R. Heffernan, J. Hanson, K. Paliwal, et al., Sixty-five years of the long march in protein secondary structure prediction: The final stretch?, Brief. Bioinform., 19 (2018), 482–494. https://doi.org/10.1093/bib/bbw129 doi: 10.1093/bib/bbw129
    [8] Y. Ma, Y. Liu, J. Cheng, Protein secondary structure prediction based on data partition and semi-random subspace method, Sci. Rep., 8 (2018), 1–10. https://doi.org/10.1038/s41598-018-28084-8 doi: 10.1038/s41598-018-28084-8
    [9] M. Lasfar, H. Bouden, A method of data mining using hidden markov models (HMMs) for protein secondary structure prediction, Procedia Comput. Sci., 127 (2018), 42–51. https://doi.org/10.1016/j.procs.2018.01.096 doi: 10.1016/j.procs.2018.01.096
    [10] A. Drozdetskiy, C. Cole, J. Procter, et al. JPred4: A protein secondary structure prediction server, Nucleic Acids Res., 43 (2015), 389–394. https://doi.org/10.1093/nar/gkv332 doi: 10.1093/nar/gkv332
    [11] D. T. Jones, Protein secondary structure prediction based on position-specific scoring matrices, J. Mol. Biol., 292 (1999), 195–202. https://doi.org/10.1006/jmbi.1999.3091 doi: 10.1006/jmbi.1999.3091
    [12] A. Busia, N. Jaitly, Next-step conditioned deep convolutional neural networks improve protein secondary structure prediction, preprint, arXiv: 2017: 1702.0386.
    [13] B. Z. Zhang, J. Y. Li, Q. Lü, Prediction of 8-state protein secondary structures by a novel deep learning architecture, BMC Bioinformatics, 19 (2018), 1–13. https://doi.org/10.1186/s12859-018-2280-5 doi: 10.1186/s12859-018-2280-5
    [14] S. Krieger, J. Kececioglu, Boosting the accuracy of protein secondary structure prediction through nearest neighbor search and method hybridization, Bioinformatics, 36 (2020). https://doi.org/10.1093/bioinformatics/btaa336 doi: 10.1093/bioinformatics/btaa336
    [15] M. R. Uddin, S. Mahbub, Saifur Rahman, M., Bayzid, M.S. SAINT: Self-attention augmented inception-inside-inception network improves protein secondary structure prediction, Bioinformatics, 36 (2020), 4599–4608. https://doi.org/10.1093/bioinformatics/btaa531 doi: 10.1093/bioinformatics/btaa531
    [16] K. Kotowski, T. Smolarczyk, I. Roterman-Konieczna, K. Stapor, ProteinUnet-An efficient alternative to SPIDER3-single for sequence-based prediction of protein secondary structures, J. Comput. Chem., 42 (2021), 50–59. https://doi.org/10.1002/jcc.26432 doi: 10.1002/jcc.26432
    [17] P. M. Sonsare, C. Gunavathi, Cascading 1D-convnet bidirectional long short term memory network with modified COCOB optimizer: A novel approach for protein secondary structure prediction, Chaos Soliton. Fract., 153 (2021), 111446. https://doi.org/10.1016/j.chaos.2021.111446 doi: 10.1016/j.chaos.2021.111446
    [18] M. J. Zvelebil, J. O. Baum, Understanding Bioinformatics, Garland Science, New York, 2007.
    [19] I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., Generative adversarial networks, Commun. ACM, 63 (2020), 139–144. https://doi.org/10.1145/3422622 doi: 10.1145/3422622
    [20] R. Wang, X. Xiao, B. Guo, Q. Qin, R.Chen, An effective image denoising method for UAV images via improved generative adversarial networks, Sensors, 18 (2018), 1985. https://doi.org/10.3390/s18071985 doi: 10.3390/s18071985
    [21] S. Yu, H. Chen, E. B. Garcia Reyes, N. Poh, Gaitgan: Invariant gait feature extraction using generative adversarial networks, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017 (2017), 30–37.
    [22] Y. Zhao, H. Zhang, Y. Liu, Protein secondary structure prediction based on generative confrontation and convolutional neural network, IEEE Access, 8 (2020), 199171–199178. https://doi.org/10.1109/ACCESS.2020.3035208 doi: 10.1109/ACCESS.2020.3035208
    [23] L. Abualigah, A. Diabat, S. Mirjalili, M. Abd Elaziz, A. H. Gandomi, The arithmetic optimization algorithm, Comput. Method. Appl. M., 376 (2021), 113609. https://doi.org/10.1016/j.cma.2020.113609 doi: 10.1016/j.cma.2020.113609
    [24] L. Abualigah, A. Diabat, P. Sumari, A. H. Gandomi, Applications, deployments, and integration of internet of drones (iod): A review, IEEE Sens. J., 21 (2021), 25532–25546. https://doi.org/10.1109/JSEN.2021.3114266 doi: 10.1109/JSEN.2021.3114266
    [25] L. Abualigah, M. Abd Elaziz, P. Sumari, Z. W. Geem, A. H. Gandomi, Reptile search algorithm (rsa): A nature-inspired meta-heuristic optimizer, Expert Syst. Appl., 191 (2022), 116158. https://doi.org/10.1016/j.eswa.2021.116158 doi: 10.1016/j.eswa.2021.116158
    [26] A. E. Ezugwu, J. O. Agushaka, L. Abualigah, S. Mirjalili, A. H. Gandomi, Prairie dog optimization algorithm, Neural Comput. Appl., 34 (2022), 20017–20065. https://doi.org/10.1007/s00521-022-07530-9 doi: 10.1007/s00521-022-07530-9
    [27] J. O. Agushaka, A. E. Ezugwu, L. Abualigah, Gazelle optimization algorithm: A novel nature-inspired metaheuristic optimizer, Neural Comput. Appl., 35 (2023), 4099–4131. https://doi.org/10.1007/s00521-022-07854-6 doi: 10.1007/s00521-022-07854-6
    [28] L. Abualigah, D. Yousri, M. Abd Elaziz, A. A. Ewees, M. A. Al-Qaness, A. H. Gandomi, Aquila optimizer: A novel meta-heuristic optimization algorithm, Comput. Ind. Eng., 157 (2021), 107250. https://doi.org/10.1016/j.cie.2021.107250 doi: 10.1016/j.cie.2021.107250
    [29] M. Arjovsky, S. Chintala, L. Bottou, Wasserstein generative adversarial networks, in International Conference on Machine Learning, 70 (2017), 214–223.
    [30] K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (2016), 770–778.
    [31] M. Farooq, A. Hafeez, Covid-resnet: A deep learning framework for screening of covid19 from radiographs, preprint, arXiv: 2003.14395.
    [32] Z. Wu, C. Shen, A. Van Den Hengel, Wider or deeper: Revisiting the resnet model for visual recognition, Pattern Recogn., 90 (2019), 119–133. https://doi.org/10.1016/j.patcog.2019.01.006 doi: 10.1016/j.patcog.2019.01.006
    [33] K. He, X. Zhang, S. Ren, J. Sun, Identity mappings in deep residual networks, in European Conference on Computer Vision, Springer, Cham, (2016), 630–645.
    [34] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, et al., Going deeper with convolutions, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, (2015), 1–9.
    [35] G. Wang, R. L. Dunbrack, Pisces: Recent improvements to a PDB sequence culling server, Nucleic Acids Res., 33 (2005), W94–W98. https://doi.org/10.1093/nar/gki402 doi: 10.1093/nar/gki402
    [36] J. Moult, K. Fidelis, A. Kryshtafovych, T. Schwede, A. Tramontano, Critical assessment of methods of protein structure prediction (casp)-round x, Proteins., 82 (2014), 1–6. https://doi.org/10.1002/prot.24452 doi: 10.1002/prot.24452
    [37] J. Moult, K. Fidelis, A. Kryshtafovych, T. Schwede, A. Tramontano, Critical assessment of methods of protein structure prediction: Progress and new directions in round xi, Proteins., 84 (2016), 4–14. https://doi.org/10.1002/prot.25064 doi: 10.1002/prot.25064
    [38] J. Moult, K. Fidelis, A. Kryshtafovych, T. Schwede, A. Tramontano, Critical assessment of methods of protein structure prediction (casp)-round xii, Proteins., 86 (2018), 7–15. https://doi.org/10.1002/prot.25415 doi: 10.1002/prot.25415
    [39] A. Kryshtafovych, T. Schwede, M. Topf, K. Fidelis, J. Moult, Critical assessment of methods of protein structure prediction (casp)-round xiii, Proteins., 87 (2019), 1011–1020. https://doi.org/10.1002/prot.25823 doi: 10.1002/prot.25823
    [40] A. Kryshtafovych, T. Schwede, M. Topf, K. Fidelis, J. Moult, Critical assessment of methods of protein structure prediction (casp)-round xiv, Proteins., 89 (2021), 1607–1617. https://doi.org/10.1002/prot.26237 doi: 10.1002/prot.26237
    [41] J. A. Cuff, G. J. Barton, Evaluation and improvement of multiple sequence methods for protein secondary structure prediction, Proteins., 34 (1999), 508–519.
    [42] S. F. Altschul, T. L. Madden, A. A. Schäffer, J. Zhang, Z. Zhang, W. Miller et al., Gapped blast and psi-blast: a new generation of protein database search programs, Nucleic Acids Res., 25 (1997), 3389–3402. https://doi.org/10.1093/nar/25.17.3389 doi: 10.1093/nar/25.17.3389
    [43] B. Rost, C. Sander, R. Schneider, Redefining the goals of protein secondary structure prediction, J. Mol. Biol., 235 (1994), 13–26. https://doi.org/10.1016/S0022-2836(05)80007-5 doi: 10.1016/S0022-2836(05)80007-5
    [44] Y. Guo, W. Li, B. Wang, H. Liu, D. Zhou, Deepaclstm: Deep asymmetric convolutional long short-term memory neural models for protein secondary structure prediction, BMC bioinformatics, 20 (2019), 1–12. https://doi.org/10.1186/s12859-019-2940-0 doi: 10.1186/s12859-019-2940-0
    [45] A. R. Ratul, M. Turcotte, M. H. Mozaffari, W. S. Lee, Prediction of 8-state protein secondary structures by 1D-Inception and BD-LSTM, BioRxiv, 2019 (2019), 871921. https://doi.org/10.1101/871921 doi: 10.1101/871921
    [46] Z. Li, Y. Yu, Protein secondary structure prediction using cascaded convolutional and recurrent neural networks, preprint, arXiv: 1604.07176.
    [47] I. Drori, I. Dwivedi, P. Shrestha, J. Wan, Y. Wang, Y. He, et al., High quality prediction of protein q8 secondary structure by diverse neural network architectures, preprint, arXiv: 1811.07143.
  • This article has been cited by:

    1. Gaoke Liao, Zhenghui Li, Ziqing Du, Yue Liu, The Heterogeneous Interconnections between Supply or Demand Side and Oil Risks, 2019, 12, 1996-1073, 2226, 10.3390/en12112226
    2. Masoud Rahiminezhad Galankashi, Farimah Mokhatab Rafiei, Maryam Ghezelbash, Portfolio selection: a fuzzy-ANP approach, 2020, 6, 2199-4730, 10.1186/s40854-020-00175-4
    3. Luca Merlo, Lea Petrella, Valentina Raponi, Forecasting VaR and ES using a joint quantile regression and its implications in portfolio allocation, 2021, 133, 03784266, 106248, 10.1016/j.jbankfin.2021.106248
    4. Hongyue Guo, Mengjun Wan, Lidong Wang, Xiaodong Liu, Witold Pedrycz, Weighted Fuzzy Clustering for Time Series With Trend-Based Information Granulation, 2024, 54, 2168-2267, 903, 10.1109/TCYB.2022.3190705
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2345) PDF downloads(72) Cited by(2)

Article outline

Figures and Tables

Figures(9)  /  Tables(8)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog