
The emergence of COVID-19 has broken the silence of humanity and people are gradually becoming concerned about pneumonia-related diseases; thus, improving the recognition rate of pneumonia-related diseases is an important task. Neural networks have a remarkable effectiveness in medical diagnoses, though the internal parameters need to be set in accordance to different data sets; therefore, an important challenge is how to further improve the efficiency of neural network models. In this paper, we proposed a learning exponential distribution optimizer based on chaotic evolution, and we optimized Resnet50 for COVID classification, in which the model is abbreviated as IEDO-net. The algorithm introduces a criterion for judging the distance of the signal-to-noise ratio, a chaotic evolution mechanism is designed according to this criterion to effectively improve the search efficiency of the algorithm, and a rotating flight mechanism is introduced to improve the search capability of the algorithm. In the computed tomography (CT) image data of COVID-19, the accuracy, sensitivity, specificity, precision, and F1 score of the optimized Resnet50 were 94.42%, 93.40%, 94.92%, 94.29% and 93.84%, respectively. The proposed network model is compared with other algorithms and models, and ablation experiments and convergence and statistical analyses are performed. The results show that the diagnostic performance of IEDO-net is competitive, which validates the feasibility and effectiveness of the proposed network.
Citation: Chengtian Ouyang, Huichuang Wu, Jiaying Shen, Yangyang Zheng, Rui Li, Yilin Yao, Lin Zhang. IEDO-net: Optimized Resnet50 for the classification of COVID-19[J]. Electronic Research Archive, 2023, 31(12): 7578-7601. doi: 10.3934/era.2023383
[1] | Saranya Muniyappan, Arockia Xavier Annie Rayan, Geetha Thekkumpurath Varrieth . DTiGNN: Learning drug-target embedding from a heterogeneous biological network based on a two-level attention-based graph neural network. Mathematical Biosciences and Engineering, 2023, 20(5): 9530-9571. doi: 10.3934/mbe.2023419 |
[2] | Jiahui Wen, Haitao Gan, Zhi Yang, Ran Zhou, Jing Zhao, Zhiwei Ye . Mutual-DTI: A mutual interaction feature-based neural network for drug-target protein interaction prediction. Mathematical Biosciences and Engineering, 2023, 20(6): 10610-10625. doi: 10.3934/mbe.2023469 |
[3] | Peter Hinow, Philip Gerlee, Lisa J. McCawley, Vito Quaranta, Madalina Ciobanu, Shizhen Wang, Jason M. Graham, Bruce P. Ayati, Jonathan Claridge, Kristin R. Swanson, Mary Loveless, Alexander R. A. Anderson . A spatial model of tumor-host interaction: Application of chemotherapy. Mathematical Biosciences and Engineering, 2009, 6(3): 521-546. doi: 10.3934/mbe.2009.6.521 |
[4] | Wen Zhu, Yuxin Guo, Quan Zou . Prediction of presynaptic and postsynaptic neurotoxins based on feature extraction. Mathematical Biosciences and Engineering, 2021, 18(5): 5943-5958. doi: 10.3934/mbe.2021297 |
[5] | Bo Zhou, Bing Ran, Lei Chen . A GraphSAGE-based model with fingerprints only to predict drug-drug interactions. Mathematical Biosciences and Engineering, 2024, 21(2): 2922-2942. doi: 10.3934/mbe.2024130 |
[6] | Xinglong Yin, Lei Liu, Huaxiao Liu, Qi Wu . Heterogeneous cross-project defect prediction with multiple source projects based on transfer learning. Mathematical Biosciences and Engineering, 2020, 17(2): 1020-1040. doi: 10.3934/mbe.2020054 |
[7] | Huiqing Wang, Sen Zhao, Jing Zhao, Zhipeng Feng . A model for predicting drug-disease associations based on dense convolutional attention network. Mathematical Biosciences and Engineering, 2021, 18(6): 7419-7439. doi: 10.3934/mbe.2021367 |
[8] | Rachael C. Adams, Behnam Rashidieh . Can computers conceive the complexity of cancer to cure it? Using artificial intelligence technology in cancer modelling and drug discovery. Mathematical Biosciences and Engineering, 2020, 17(6): 6515-6530. doi: 10.3934/mbe.2020340 |
[9] | Dong Ma, Shuang Li, Zhihua Chen . Drug-target binding affinity prediction method based on a deep graph neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 269-282. doi: 10.3934/mbe.2023012 |
[10] | Xianfang Wang, Qimeng Li, Yifeng Liu, Zhiyong Du, Ruixia Jin . Drug repositioning of COVID-19 based on mixed graph network and ion channel. Mathematical Biosciences and Engineering, 2022, 19(4): 3269-3284. doi: 10.3934/mbe.2022151 |
The emergence of COVID-19 has broken the silence of humanity and people are gradually becoming concerned about pneumonia-related diseases; thus, improving the recognition rate of pneumonia-related diseases is an important task. Neural networks have a remarkable effectiveness in medical diagnoses, though the internal parameters need to be set in accordance to different data sets; therefore, an important challenge is how to further improve the efficiency of neural network models. In this paper, we proposed a learning exponential distribution optimizer based on chaotic evolution, and we optimized Resnet50 for COVID classification, in which the model is abbreviated as IEDO-net. The algorithm introduces a criterion for judging the distance of the signal-to-noise ratio, a chaotic evolution mechanism is designed according to this criterion to effectively improve the search efficiency of the algorithm, and a rotating flight mechanism is introduced to improve the search capability of the algorithm. In the computed tomography (CT) image data of COVID-19, the accuracy, sensitivity, specificity, precision, and F1 score of the optimized Resnet50 were 94.42%, 93.40%, 94.92%, 94.29% and 93.84%, respectively. The proposed network model is compared with other algorithms and models, and ablation experiments and convergence and statistical analyses are performed. The results show that the diagnostic performance of IEDO-net is competitive, which validates the feasibility and effectiveness of the proposed network.
Drug-target interactions (DTIs) involve the binding of a drug to the relevant site of a target protein to trigger a biochemical reaction [1]. The efficacy is related to the biological activity of the protein. However, it is complicated for experiments to predict a drug's success and drug discovery is time-consuming and expensive [2,3], which is estimated to typically take 12–15 years and cost over $100 million [4]. For these reasons, in the past decades, computer-aided drug design (CADD) has been proposed to discriminate new drugs and consists of processes such as virtual screening, molecular docking, and QSAR methods [5]. Currently, due to limited ligand data and the limited information on the structure of novel target proteins [6], these approaches are inappropriate and inefficient given the growth of available biological and chemical data [2]. Recently, with the advent of various deep learning methods, a significant future trend in AI-based drug discovery has been identified [7]. It is essential for drug discovery to accurately predict the number of DTIs [8]. Therefore, it is urgent to devise richer and more compatible computational methods to differentiate between potential DTIs.
The concept of "guilt-by-association" [9] has been described in DTIs prediction. It is defined that if drug A has target proteins, and the action event between drug B is similar to drug A, targets interactions are likely to appear, and the reverse is also true. Machine learning methods are used for DTIs prediction and can successfully solve the assumption. For instance, Mei et al. [10] proposed bipartite local models (BLMs) that considered neighbors' interaction profiles where neighbor-based interaction-profile inferring (NII) can be effective in defining a new candidate problem. Luo et al. [11] used an inductive matrix completion method, in which seven kinds of drug/target-related similarities were included in an integrated network (e.g., drugs, proteins, diseases, and side-effects). Ezzat et al. [12] proposed graph regularized matrix factorization (GRMF) and weighted graph regularized matrix factorization (WGRMF) methods that introduced graph regularization into the matrix factorization in order to learn manifolds. Moreover, a preprocessing step (WKNKN) has been developed to rescore unknown drug-target pairs that were previously regarded as null values. Although these methods have been proven to be effective, there are challenges to overcome complex data structures such as interaction networks of drugs or targets. Furthermore, the rapid growth of drug/target-related data has outpaced their ability to process and analyze information. With the emergence of diverse and enriched feature representations, the efficacy of the above methods may limit the exploration of more comprehensive topological information and node characteristics between drugs and target proteins.
Network-based algorithms and feature-based algorithms become famous in the field. Generally, identifying DTIs is considered as a binary classification task by extracting features vectors of drugs as well as targets. Several number of heterogeneous data have been integrated into a heterogeneous network to boost the accuracy of DTIs prediction tasks [13]. The deep belief network (DBN) [14] has been proposed to build an end-to-end method for abstracting raw input samples. Moreover, sequence-based approaches are universal. Different architectures [15,16,17,18] have been developed for feature extraction of sequence information. DrugVQA [19] employs a bidirectional long-short time memory network to tackle the prediction problem. Furthermore, graph-based methods are suitable for the two-dimensional representation of structural information. Zhao et al. [20] utilized a combination of graph convolutional network (GCN) and deep neural network (DNN) to enhance the identification of DTIs. GNN was coupled with CNN, which was designed as drug feature and target feature extraction method [21]. LASSO has been employed by You et al. [22] as a feature procession. Thafar et al. [23] constructed the DTi2Vec model including graph embedding which capture relationships between drugs and targets and then these features are fed into the ensemble classifier for prediction analyses. Huang et al. designed a molecular sub-structure representation and used massive unlabeled biomedical data through an augment transformer [24]. Peng et al. [25] introduced CNN to identify DTIs and trained the denoising autoencoder (DAE) as a feature selector. Although these methods can effectively predict DTIs, the problem of parameter count and computation amount need to be given more attention.
The broad learning system (BLS) [26] is characterized by a relatively simple neural network architecture comprising only three layers of neurons. Inspired by the concept of the random vector functional-link neural network (RVFLNN) [27,28], its training procedure is facilitated through pseudo-inverse calculations. Due to its training procedure and flat structure, BLS has the advantages of fast computing speed and few training parameters. Therefore, BLS has been widely applied to various disciplines including medicine [29]. For instance, Fan et al. [30] proposed a stacked ensemble classifier build by BLS for the prediction of interactions between lncRNA and proteins. Zheng et al. [31] designed a modified BLS-based model to predict miRNA-disease associations using sequence similarities of microRNA (miRNAs). The above applications of BLS in this area have been proven to be useful. However, there is a lack of related research for DTIs based on BLS. Additionally, since labeled data volumes are always sparse and insufficient, prediction modeling is to performance is inadequate. By fusing information from multiple aspects to overcome the limitations, the above methods can improve the performance, indicating that these combined models could solve the challenge of interaction matrix sparsity.
In this study, we developed a novel model called ConvBLS-DTI to predict DTIs. Compared with the previous DTI predictive methods, ConvBLS-DTI integrates matrix factorization with the broad learning system, yielding reliable DTI prediction results. The task of DTI prediction is formulated as a binary classification problem to determine whether a drug-target pair is a DTI. The major contributions of this paper are as follows:
1) We address the challenges of data sparsity and incompleteness by employing a WKNKN algorithm as a pre-processing step, which help to mitigate the adverse effects of a large number missing interaction value.
2) We propose a matrix factorization technique used on the interaction matrix to generate two latent feature matrices for drugs and targets, thereby enabling the learning of low-dimensional vector representations of features.
3) Based on the CNN algorithm, ConvBLS-DTI can handle the DTIs prediction, taking the extracted drug-target pairs feature vectors as inputs.
The architecture of the proposed ConvBLS-DTI method is depicted in Figure 1. It is primarily composed of three sessions: First, we utilized the WKNKN algorithm to alleviate the sparsity of the DTI matrix, thus enhancing the input information complement of the model and improving its predictive performance. After construction of the DTI matrix, matrix factorization is used to decompose DTI matrix into two feature matrices of low ranks which obtains vector representation of the drug features and target features. Then, the drug feature vectors combine with the target feature vectors together to get the final feature vectors. Finally, ConvBLS is built for classification. A CNN is leveraged to enhance the nodes' representation, followed by a broad learning module, which further enable satisfactory results in effective identification of DTIs.
We initially reconstruct the interaction matrix using the computational preprocessing technique, which can effectively complement the interaction matrix for the identification of DTIs and improve the known DTI samples. As shown on the left side of step A in Figure 1, the green circles, the red triangles, and the blue lines separately denote drugs, targets and the known interaction. D={di}ndi=1 and T={tj}ntj=1 are separately described as each node for drugs and targets, where nd is the number of drugs and nt is the number of targets. The associations between nd drugs and nt targets are represented by an interaction matrix Y∈{0,1}, in which Yij=1 indicates a known interaction between drug di and target tj, and Yij=0 otherwise. In addition, the similarity matrix of both drugs and targets is represented as SD∈Rnd×nd and ST∈Rnt×nt.
Numerous unknown interactions can significantly impact the evaluation outcome of the model and introduce prediction bias. In DTIs prediction, weighted k-nearest neighbors (k-NN) has been employed to leverage similarity measures to promote further prediction performance. Weighted k-NN considers both neighbor similarity and distances, incorporating distance weights to calculate likelihood values of unconfirmed drug-target interactions. Specifically, given a drug-target pair, the algorithm first identifies the k-NN and assigns weights to each neighbor based on their similarity and distance. Weight involved in WKNKN [12] is computed by Gaussian weighting method. The calculated weighted likelihood values can be used to predict the likelihood of unknown DTIs within the matrix. Here, the specific operation is achieved through the following three steps:
Yd(d)=1Md∑Ki=1ωdiY(di) | (2.1) |
where Yd(d) denotes the likelihood score of interaction for drug di. Md is defined as a normalization term, ω coefficient represents the weights of the K nearest known neighbors of drug di. Similarly, the same terms are computed to estimate the interaction likelihood score of the target tj:
Yt(t)=1Mt∑Ki=1ωtjY(tj) | (2.2) |
where Yt(t) denotes the likelihood score of interaction for target tj. Mt is the normalization term and ω coefficient represents the weights of the K nearest known neighbors of target tj. Finally, the derived formula is as follows:
YWKNKN=max(Yd+Yt2,Y) | (2.3) |
Therefore, if Yij is 0, YWKNKN replaces it with an average of the weighted interaction likelihood value. For the matrix representation, 0 and 1 denote the absence and presence of interactions between drugs and targets, respectively. Likelihood serves as a measure of the possibility of interaction between a drug and a target, typically ranging from 0 to 1. Higher likelihood values indicate a higher likelihood of interaction, while lower likelihood values suggest a lower likelihood.
Considering that most studies mainly concentrate on extracting features from drugs and targets individually and less on the relationships of the DTI, neighbor regularization logistic matrix factorization (NRLMF) [32] is used to represent drugs and targets in the right part of step B. NRLMF is an unsupervised learning strategy that mainly infers unknowns through known interactions and their similarities, so no negative samples are required. The valid connections are denoted as the modified interaction matrix made of known and unknown interactions. As shown in Figure 1, the DTI probabilities can be defined as a logistic function:
Pi,j=exp(uivTj)1+exp(uivTj) | (2.4) |
where each term ui∈R1×r is denoted as the r-dimensional potential representation of each drug di. Similarly, each term vj∈R1×r represents the r-dimensional potential representation of each target tj. In this way, the potential feature vectors for all drugs and all targets can be summarized as U=(uT1,…,uTnd) and V=(vT1,…,vTnt), where T refers to the transpose of the matrix.
The neighborhood regularization method proposes to add the nearest neighbors of drugs and targets to further increase information diversity and enable higher accuracy without overfitting. The neighborhood regularization is achieved by:
α2∑ndi∑ndjSPiμ||ui−uj||2f | (2.5) |
α2∑nti∑ntjSQϑj||vi−vj||2f | (2.6) |
where α is the Laplace regularization parameter, and ||⋅||f is the Frobenius norm of the matrix, and the parameters SP and SQ represent neighbors similarity measure matrix are given by:
SPiμ=SDiμ if dμ∈Wd(di) else SPiμ=0 | (2.7) |
SQϑj=STϑj if tϑ∈Wt(tj) else SQϑj=0 | (2.8) |
where SD and ST denote as similarity matrix, and Wd(di) is defined as the nearest neighbors of a node di, and Wt(tj) is defined as the nearest neighbors of a node tj.
The matrix factorization (MF) method decomposes the interaction matrix into two low-rank matrices. The MF is formulated as a feature extraction task to obtain the description of the drugs and their targets as features. The feature matrix is obtained by maximizing the objective function via the posterior probability distribution:
maxU,VP(U,V|Y,σ2d,σ2t) | (2.9) |
where Y denotes the interaction matrix, σ2d and σ2t are parameters that control the variance of Gaussian distribution of drug set and target set.
Thus, drugs and targets can be denoted as two r-dimensional feature representations. As illustrated in Figure 2, the drug feature is U−D=[DF1,DF2,...,DFr], and the target feature is V−T=[TF1,TF2,...,TFr]. Then, the drug feature vectors and target feature vectors are merged and assigned the label based on the interaction matrix Y. To ensure the quality of the model, the number of negative samples is equal to positive number in each dataset. Negative samples are randomly generated based on the interaction matrix Y. The pairwise drug–target feature vector is used as input to the neural network, which can be expressed as FV = [DF1,DF2,...,DFr,TF1,TF2,...,TFr].
After concatenating the features of drugs and targets, in order to achieve better performance and train more effectively, the ConvBLS model is used as a classification method to determine the predictions of the DTIs. As shown in Figure 3, we developed a broad learning system that combined convolutional neural network to extract high-quality drug and target representations for better prediction.
ConvBLS mostly includes two parts: The 1D-CNN module and enhancement nodes. CNN block is used to learn a representative features of targets and drugs. The input data of 1D-CNN is a one-dimensional feature vectors, and the convolution kernel is also in one-dimensional form. The enhancement layer is responsible for further feature extracting. The detailed network structure is shown in Figure 3. This section completes the classification task with ConvBLS.
Given the lack of learning ability of the original feature mapping, the CNN block is selected for sequence data of drug-target features. It contains multiple groups of feature mapping nodes composed of a 1D-CNN layer and a max pooling layer. To solve complex tasks, learning models increasingly go deeply. The multiscale random convolution feature is expanded to improve robustness. The detailed computational procedure is as follows:
First, the drug–target predictive model ConvBLS is constructed based on previous obtained feature data FV. The input is connected to the mapping matrix by applying 1-D convolution kernels to generate the corresponding feature representation. All the random items can act as the convolution kernel so that we can achieve the output:
FC=φ(Conv(X,KC)) | (2.10) |
where X is the input feature vectors, KC is the convolution kernel, Conv(.) is denoted as the convolution function, and φ(.) is the activation function. The descriptor of the mapping feature is called FC. Then, the down-sampling method is used to allow the feature to be robust:
FP=max_pool(FC) | (2.11) |
where FP is the result after max pooling function. Next, an enhancement layer is built. Using random weights and nonlinear transformation, the enhancement nodes are obtained:
Ej=ψ(FPWej+bej)j=1,2,…,n | (2.12) |
where ψ(.) is the activation function, weights and bias represented as Wej and bej, which are randomly initialized, and n is the group number of enhanced nodes. All of the enhancement nodes can be represented as En≡[E1,E2,…En]. Finally, the improved feature layer and enhancement layer are concatenated into one matrix as a single neural network. Hence, featurization constitutes the outputs of weight of the BLS, based on Y=HWdt.
Wdt=H+Y=[FP|En]+Y | (2.13) |
The ridge regression approximation algorithm [33] is utilized to determine the [FP|En]+:
[FP|En]+=limλ→0(λI+[FP|En][FP|En]T)−1[FP|En]T | (2.14) |
Here, we describe the dataset used in this paper and provide the experiment setup and evaluation metrics for comparing model performance in subsequent experiments.
In this study, two benchmark datasets are used for evaluating our proposed model: the Yamanishi's dataset and Luo's dataset. The first one is the gold benchmark dataset created by Yamanishi et al. [34]. It is classified into four categories based on the target protein class, namely: (ⅰ) enzyme (E), (ⅱ) ion channel (IC), (ⅲ) G protein-coupled receptor (GPCR), and (ⅳ) nuclear receptor (NR). Since the discovery of the interactions in these datasets 14 years ago, we implemented the completed version of the original golden standard datasets collected by Liu et al. [35]. The new datasets added information on the KEGG pathways [36], DrugBank [37], and ChEMBL [38] databases. The second one was developed by Luo et al. [11], consisting of four categories of nodes (drugs, proteins, diseases, and side-effects) and six types of connections (drug-target interaction, drug-drug interactions, protein-protein interactions, drug-disease associations, protein-disease associations, and drug-side-effect associations). Table 1 lists the detailed statistical entries of the complete datasets included in our analysis. Sparsity represents the proportion of known DTI numbers in all possible DTI combinations.
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Table 2 lists the parameter settings in the experiments depending on datasets. The best parameters of ConvBLS-DTI were selected by performing a grid search. Some key parameters were set as follows: The number of the nearest known neighbors K is set to 5 for NR and 7 for others; the feature dimension r is set to 50 for a relatively small dataset NR, and 100 was an appropriate setting for GPCR, IC, and E datasets. The convolution kernel size is taken from {3–9}. The Tanh function was chosen as the activation function for every layer. A number of experiments are performed to determine the optimal classification parameters of BLS. Specifically, the shrinkage scale (sc) of the enhancement nodes plays a central role in this experiment. The parameters of all baseline methods were set based on the suggestions from the respective studies available in the literature.
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
For the cross-validation experiments, there are three different experimental settings for comparison, depending on whether the drug and target involved in the test pair are training entities:
1) CVd: Predicts the interactions between testing drugs and training targets;
2) CVt: Predicts the interactions between training drugs and testing targets;
3) CVdt: Predicts the interactions between testing drugs and testing targets.
The 10-fold cross-validation is one of the most widely available methods. All models were trained and tested using 10-fold cross-validation. In this study, the final results are given with the area under the receiver operating characteristic curve (AUC) and the area under the precision-recall curve (AUPR) to judge the prediction performance. They are widely used in this field [39,40]. Since there are few true DTIs, AUPR is a more precise quality indicator than AUC because it punished those in which lots of false positive examples were found from the top-ranked prediction score [41], so we consider it as an evaluation sign. In addition, the Sen score was another metric used in this study. The average values are used for the results of each dataset.
Experiments were run under the environment of Windows 10 Professional Edition and i5-7200H CPU. Our aim of this study was to construct an efficient computation method with excellent performance for DTI prediction. Therefore, we first observe the performance of two BLS-based models from different perspectives on the four datasets. Then, we compared the prediction results of our model with representative methods under three settings: NRLMF [32], DTINet [11], WKNNIR [42], DTi2Vec [43], ADA-GRMFC [44], and BLS-DTI. Finally, the optimal of core parameter in the experiment was reported.
We first compared our model with the BLS-DTI. Tables 3 and 4 list the AUC and AUPR results on the prediction tasks. As shown in Table 3, our model is found to outperform BLS-DTI in the AUC and AUPR. It highlights the importance of the feature extraction ability in the BLS. The enhancement layer is included in the two networks. We considered the prediction performance of BLS is insufficient for DTI prediction tasks due to the lack of the ability to obtain deep features. ConvBLS-DTI provides the better performance in terms of CNN method. Specifically, ConvBLS-DTI exhibits higher results than BLS-DTI for the E dataset, providing 0.22 higher AUC score, an improvement of 28%, with a 0.152 greater AUPR value, an improvement of 19%. The same positive results are also found in the other three datasets, which indicates that the performance of the model ConvBLS-DTI is improved when the CNN method is added to the BLS network.
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
For a more comprehensive evaluation, the following Table 4 shows the addition AUC and AUPR scores of each experimental setting on four datasets. Similarly, ConvBLS-DTI obtain higher performance in all scenarios, outperforming the other method BLS-DTI. Compared with NR and GPCR datasets, IC and E datasets contribute to higher AUC and AUPR scores, with AUPR values of 0.947 and 0.961, respectively. The possible reason is that the number of DTIs in the NR and GPCR categories is smaller than the other categories, especially for NR with only 166 drug-target pairs.
In this section, under the same datasets, evaluation metrics and experimental scenarios (CVd, CVt, and CVdt), six advanced methods, including NRLMF, DTINet, WKNNIR, DTi2Vec, ADA-GRMFC, and BLS-DTI, are involved into the performance comparison. Tables 5 and 6 show the AUC and AUPR of the methods participating in the CVd and CVt settings. In general, based on the main evaluation metrics, our method has overall better performance than the other methods under different scenarios. For CVd, ConvBLS-DTI shows a high performance in all datasets. For CVt, minimal difference is found in the AUC score obtained using IC and GPCR datasets, but the AUPR result achieved by our method increases by 2.05%, 4.41%, 3.6%, 5.54%, and 6.12% on NR, GPCR, IC, E, and Luo datasets, respectively, compared with that of the second-best model. In particular, ConvBLS-DTI performs better than BLS-DTI. In the results of predicting novel drugs and known targets, ConvBLS-DTI is better than other methods. In the experiment scenario of CVd, ConvBLS-DTI achieves AUPR values of 0.917, 0.968, 0.972, 0.958, and 0.972 on NR, IC, GPCR, E, and Luo datasets, respectively. In the experiment scenario of CVt, the AUPR values of ConvBLS-DTI are 0.846, 0.950, 0.946, 0.952, and 0.954 on NR, IC, GPCR, E, and Luo datasets, respectively. Overall, it can be concluded that the proposed ConvBLS-DTI is superior to all the compared methods and proves that broad learning system can also be a rational tool to help for predicting DTIs.
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
In particular, the ConvBLS-DTI has satisfactory performance under the CVdt setting. The AUC and AUPR histograms for the different algorithms are shown in Figures 4 and 5, respectively. The results are entirely consistent across all datasets given in the CVdt (specifically in terms of AUPR metric). For CVdt, the AUC and AUPR values of ConvBLS-DTI are higher than other methods on all datasets, although DTi2Vec is very competitive compared with ConvBLS-DTI. Overall, our method improves the AUPR more than the AUC.
To measure the impact of the WKNKN method on ConvBLS-DTI, ablation experiments were conducted by removing the WKNKN method under three different CV strategies using the Luo et al. dataset. The variant of ConvBLS-DTI without the WKNKN method is denoted as ConvBLS-DTI (without WKNKN). Performance comparisons between ConvBLS-DTI and the variant in terms of AUC and AUPR are presented in Tables 7 and 8, respectively. The findings in Tables 7 and 8 suggest that the utilization of the WKNKN method contributes to improve the performance of ConvBLS-DTI.
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |
In this study, the datasets IC and GPCR were applied to test the influence of the convolution kernel size. As illustrated in Figure 6, by varying the size of the convolutional kernel (3, 4, 5, 6, 7, 8, and 9), the AUPR value of the ConvBLS-DTI method progressively improved with an increase in kernel size and reached its optimal performance when kernel size was set to 5. Subsequently, the performance showed a decline. A kernel size set to 5 achieved good results.
In this paper, we aimed to solve the problem of sparsity and incompletion of the drug interaction data. A new framework called ConvBLS-DTI was proposed to predict DTIs by applying an advanced fusion of BLS approach. Our method integrates WKNKN, MF, and BLS to improve the DTI prediction results. The method takes advantage of matrix factorization for the latent low-dimensional feature representation and predicts DTIs based on the broad learning architecture. Moreover, the WKNKN algorithm was used as a preprocessing step to increase the availability of relevant information for a large number of missing correlations. Compared with the BLS-DTI, our model achieved AUC and AUPR values of 0.971 and 0.967, respectively, for the IC dataset under tenfold-cross-validation experiments. These findings illustrate that the combination of CNN and BLS could improve the prediction performance for DTIs. Additionally, compared with other previous methods, the best AUC and AUPR values of the proposed method were 0.9643 and 0.9681 for the E dataset and CVdt setting, respectively. The results show that our model acquires improved prediction effect on AUC and AUPR using extensive experimental verification.
In future studies, greater emphasis will be placed on optimizing the BLS structure to enhance the feature extraction ability. In fact, the results of the present prediction model can be heavily influenced by the mapping algorithms and the effectiveness of the dataset. Therefore, our models might be further developed using other deep-learning models to increase the identifying power. Overall, with the availability of more data and the development of new approaches, it is expected that more applications of our model can be achieved.
The authors that declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work was supported by the project of the Natural Science Foundation of Shandong Province, China (Natural Science Foundation of Shandong Province, No. ZR2019PEE018), Shandong Province Science and Technology SMES Innovation Ability Enhancement Project (Natural Science Foundation of Shandong Province, No. 2021TSGC1063), Major Scientific and Technological Innovation Projects of Shandong Province (Natural Science Foundation of Shandong Province, No. 2019JZZY020101), and the project of the Natural Science Foundation of Qingdao (No. 23-2-1-216-zyyd-jch).
The authors declare that there are no conflicts of interest.
[1] |
Q. Yuan, K. Chen, Y. Yu, N. Q. K. Le, M. C. H. Chua, Prediction of anticancer peptides based on an ensemble model of deep learning and machine learning using ordinal positional encoding, Briefings Bioinf., 24 (2023), bbac630. https://doi.org/10.1093/bib/bbac630 doi: 10.1093/bib/bbac630
![]() |
[2] |
N. Q. K. Le, Potential of deep representative learning features to interpret the sequence information in proteomics, Proteomics, 22 (2022), 2100232. https://doi.org/10.1002/pmic.202100232 doi: 10.1002/pmic.202100232
![]() |
[3] |
P. Aggarwal, N. K. Mishra, B. Fatimah, P. Singh, A. Gupta, S. D. Joshi, COVID-19 image classification using deep learning: Advances, challenges and opportunities, Comput. Biol. Med., 144 (2022), 105350. https://doi.org/10.1016/j.compbiomed.2022.105350 doi: 10.1016/j.compbiomed.2022.105350
![]() |
[4] |
O. S. Albahri, A. A. Zaidan, A. S. Albahri, B. B. Zaidan, K. H. Abdulkareem, Z. T. Al-qaysi, et al., Systematic review of artificial intelligence techniques in the detection and classification of COVID-19 medical images in terms of evaluation and benchmarking: Taxonomy analysis, challenges, future solutions and methodological aspects, J. Infect. Public Health, 13 (2020), 1381–1396. https://doi.org/10.1016/j.jiph.2020.06.028 doi: 10.1016/j.jiph.2020.06.028
![]() |
[5] |
T. W. Cenggoro, B. Pardamean, A systematic literature review of machine learning application in COVID-19 medical image classification, Procedia Comput. Sci., 216 (2023), 749–756. https://doi.org/10.1016/j.procs.2022.12.192 doi: 10.1016/j.procs.2022.12.192
![]() |
[6] |
Y. Hu, K. Liu, K. Ho, D. Riviello, J. Brown, A. R. Chang, et al., A simpler machine learning model for acute kidney injury risk stratification in hospitalized patients, J. Clin. Med., 11 (2022), 5688. https://doi.org/10.3390/jcm11195688 doi: 10.3390/jcm11195688
![]() |
[7] |
E. Hussain, M. Hasan, M. A. Rahman, I. Lee, T. Tamanna, M. Z. Parvez, CoroDet: A deep learning based classification for COVID-19 detection using chest X-ray images, Chaos, Solitons Fractals, 142 (2021), 110495. https://doi.org/10.1016/j.chaos.2020.110495 doi: 10.1016/j.chaos.2020.110495
![]() |
[8] |
M. A. Ozdemir, G. D. Ozdemir, O. Guren, Classification of COVID-19 electrocardiograms by using hexaxial feature mapping and deep learning, BMC Med. Inf. Decis. Making, 21 (2021), 170. https://doi.org/10.1186/s12911-021-01521-x doi: 10.1186/s12911-021-01521-x
![]() |
[9] |
A. M. Ismael, A. Şengür, Deep learning approaches for COVID-19 detection based on chest X-ray images, Expert Syst. Appl., 164 (2021), 114054. https://doi.org/10.1016/j.eswa.2020.114054 doi: 10.1016/j.eswa.2020.114054
![]() |
[10] |
K. Kc, Z. Yin, M. Wu, Z. Wu, Evaluation of deep learning-based approaches for COVID-19 classification based on chest X-ray images, Signal, Image Video Process., 15 (2021), 959–966. https://doi.org/10.1007/s11760-020-01820-2 doi: 10.1007/s11760-020-01820-2
![]() |
[11] |
G. Muhammad, M. S. Hossain, COVID-19 and non-COVID-19 classification using multi-layers fusion from lung ultrasound images, Inf. Fusion, 72 (2021), 80–88. https://doi.org/10.1016/j.inffus.2021.02.013 doi: 10.1016/j.inffus.2021.02.013
![]() |
[12] |
X. Wang, X. Deng, Q. Fu, Q. Zhou, J. Feng, H. Ma, et al., A weakly-supervised framework for COVID-19 classification and lesion localization from chest CT, IEEE Trans. Med. Imaging, 39 (2020), 2615–2625. https://doi.org/10.1109/TMI.2020.2995965 doi: 10.1109/TMI.2020.2995965
![]() |
[13] |
M. Riaz, M. Bashir, I. Younas, Metaheuristics based COVID-19 detection using medical images: A review, Comput. Biol. Med., 2022 (2022), 105344. https://doi.org/10.1016/j.compbiomed.2022.105344 doi: 10.1016/j.compbiomed.2022.105344
![]() |
[14] |
D. Zhu, S. Wang, C. Zhou, S. Yan, J. Xue, Human memory optimization algorithm: A memory-inspired optimizer for global optimization problems, Expert Syst. Appl., 237 (2024), 121597. https://doi.org/10.1016/j.eswa.2023.121597 doi: 10.1016/j.eswa.2023.121597
![]() |
[15] |
D. Zhu, S. Wang, J. Shen, C. Zhou, T. Li, S. Yan, A multi-strategy particle swarm algorithm with exponential noise and fitness-distance balance method for low-altitude penetration in secure space, J. Comput. Sci., 74 (2023), 102149. https://doi.org/10.1016/j.jocs.2023.102149 doi: 10.1016/j.jocs.2023.102149
![]() |
[16] | J. Kennedy, R. Eberhart, Particle swarm optimization, in Proceedings of ICNN'95-International Conference on Neural Networks, IEEE, (1995), 1942–1948. https://doi.org/10.1109/ICNN.1995.488968 |
[17] | K. V. Price, Differential evolution, in Handbook of Optimization: From Classical to Modern Approach, Springer, (2013), 187–214. https://doi.org/10.1007/978-3-642-30504-7_8 |
[18] | M. Dorigo, M. Birattari, T. Stutzle, Ant colony optimization, IEEE Comput. Intell. Mag., 1 (2006), 28–39. https://doi.org/10.1109/MCI.2006.329691 |
[19] | S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Adv. Eng. Software, 69 (2014), 46–61. https://doi.org/10.1016/j.advengsoft.2013.12.007 |
[20] |
S. Mirjalili, A. Lewis, The whale optimization algorithm, Adv. Eng. Software, 95 (2016), 51–67. https://doi.org/10.1016/j.advengsoft.2016.01.008 doi: 10.1016/j.advengsoft.2016.01.008
![]() |
[21] |
J. Xue, B. Shen, A novel swarm intelligence optimization approach: Sparrow search algorithm, Syst. Sci. Control Eng., 8 (2020), 22–34. https://doi.org/10.1080/21642583.2019.1708830 doi: 10.1080/21642583.2019.1708830
![]() |
[22] |
A. A. Heidari, S. Mirjalili, H. Faris, I. Aljarah, M. Mafarja, H. Chen, Harris hawks optimization: Algorithm and applications, Future Gener. Comput. Syst., 97 (2019), 849–872. https://doi.org/10.1016/j.future.2019.02.028 doi: 10.1016/j.future.2019.02.028
![]() |
[23] |
W. Zhao, Z. Zhang, L. Wang, Manta ray foraging optimization: An effective bio-inspired optimizer for engineering applications, Eng. Appl. Artif. Intell., 87 (2020), 103300. https://doi.org/10.1016/j.engappai.2019.103300 doi: 10.1016/j.engappai.2019.103300
![]() |
[24] |
M. Abdel-Basset, D. El-Shahat, M. Jameel, M. Abouhawwash, Exponential distribution optimizer (EDO): A novel math-inspired algorithm for global optimization and engineering problems, Artif. Intell. Rev., 56 (2023), 9329–9400. https://doi.org/10.1007/s10462-023-10403-9 doi: 10.1007/s10462-023-10403-9
![]() |
[25] |
A. Dixit, A. Mani, R. Bansal, CoV2-Detect-Net: Design of COVID-19 prediction model based on hybrid DE-PSO with SVM using chest X-ray images, Inf. Sci., 571 (2021), 676–692. https://doi.org/10.1016/j.ins.2021.03.062 doi: 10.1016/j.ins.2021.03.062
![]() |
[26] |
D. A. D. Júnior, L. B. da Cruz, J. O. B. Diniz, G. L. F. da Silva, G. B. Junior, A. C. Silva, et al., Automatic method for classifying COVID-19 patients based on chest X-ray images, using deep features and PSO-optimized XGBoost, Expert Syst. Appl., 183 (2021), 115452. https://doi.org/10.1016/j.eswa.2021.115452 doi: 10.1016/j.eswa.2021.115452
![]() |
[27] |
M. A. A. Albadr, S. Tiun, M. Ayob, F. T. AL-Dhief, Particle swarm optimization-based extreme learning machine for COVID-19 detection, Cognit. Comput., 2022 (2022), 1–16. https://doi.org/10.1007/s12559-022-10063-x doi: 10.1007/s12559-022-10063-x
![]() |
[28] |
M. A. Elaziz, K. M. Hosny, A. Salah, M. M. Darwish, S. Lu, A. T. Sahlol, New machine learning method for image-based diagnosis of COVID-19, PLoS One, 15 (2020), e0235187. https://doi.org/10.1371/journal.pone.0235187 \newpage doi: 10.1371/journal.pone.0235187
![]() |
[29] |
E. S. M. El-Kenawy, S. Mirjalili, A. Ibrahim, M. Alrahmawy, M. El-Said, R. M. Zaki, et al., Advanced meta-heuristics, convolutional neural networks, and feature selectors for efficient COVID-19 X-ray chest image classification, IEEE Access, 9 (2021), 36019–36037. https://doi.org/10.1109/ACCESS.2021.3061058 doi: 10.1109/ACCESS.2021.3061058
![]() |
[30] |
S. Pathan, P. C. Siddalingaswamy, P. Kumar, M. M. M. Pai, T. Ali, U. R. Acharya, Novel ensemble of optimized CNN and dynamic selection techniques for accurate COVID-19 screening using chest CT images, Comput. Biol. Med., 137 (2021), 104835. https://doi.org/10.1016/j.compbiomed.2021.104835 doi: 10.1016/j.compbiomed.2021.104835
![]() |
[31] |
A. Basu, K. H. Sheikh, E. Cuevas, R. Sarkar, COVID-19 detection from CT scans using a two-stage framework, Expert Syst. Appl., 193 (2022), 116377. https://doi.org/10.1016/j.eswa.2021.116377 doi: 10.1016/j.eswa.2021.116377
![]() |
[32] |
M. H. Nadimi-Shahraki, H. Zamani, S. Mirjalili, Enhanced whale optimization algorithm for medical feature selection: A COVID-19 case study, Comput. Biol. Med., 148 (2022), 105858. https://doi.org/10.1016/j.compbiomed.2022.105858 doi: 10.1016/j.compbiomed.2022.105858
![]() |
[33] | S. Elghamrawy, A. E. Hassanien, Diagnosis and prediction model for COVID-19 patient's response to treatment based on convolutional neural networks and whale optimization algorithm using CT images, preprint, MedRxiv, 2020. https://doi.org/10.1101/2020.04.16.20063990 |
[34] |
T. Goel, R. Murugan, S. Mirjalili, D. K. Chakrabartty, OptCoNet: An optimized convolutional neural network for an automatic diagnosis of COVID-19, Appl. Intell., 51 (2021), 1351–1366. https://doi.org/10.1007/s10489-020-01904-z doi: 10.1007/s10489-020-01904-z
![]() |
[35] |
T. Hu, M. Khishe, M. Mohammadi, G. Parvizi, S. H. T. Karim, T. A. Rashid, Real-time COVID-19 diagnosis from X-Ray images using deep CNN and extreme learning machines stabilized by chimp optimization algorithm, Biomed. Signal Process. Control, 68 (2021), 102764. https://doi.org/10.1016/j.bspc.2021.102764 doi: 10.1016/j.bspc.2021.102764
![]() |
[36] |
D. Singh, V. Kumar, M. Kaur, Classification of COVID-19 patients from chest CT images using multi-objective differential evolution–based convolutional neural networks, Eur. J. Clin. Microbiol. Infect. Dis., 39 (2020), 1379–1389. https://doi.org/10.1007/s10096-020-03901-z doi: 10.1007/s10096-020-03901-z
![]() |
[37] | M. S. Iraji, M. Feizi-Derakhshi, J. Tanha, Deep learning for COVID-19 diagnosis based feature selection using binary differential evolution algorithm, preprint, 2021, arXiv: PPR343118. |
[38] |
A. M. Sahan, A. S. Al-Itbi, J. S. Hameed, COVID-19 detection based on deep learning and artificial bee colony, Periodicals Eng. Nat. Sci., 9 (2021), 29–36. http://doi.org/10.21533/pen.v9i1.1774 doi: 10.21533/pen.v9i1.1774
![]() |
[39] | F. Sadeghi, O. Rostami, M. K. Yi, A deep learning approach for detecting COVID-19 using the chest X-ray images, CMC-Comput. Mater. Continua, 75 (2023), 751–768. |
[40] |
H. M. Balaha, E. M. El-Gendy, M. M. Saafan, CovH2SD: A COVID-19 detection approach based on Harris Hawks Optimization and stacked deep learning, Expert Syst. Appl., 186 (2021), 115805. https://doi.org/10.1016/j.eswa.2021.115805 doi: 10.1016/j.eswa.2021.115805
![]() |
[41] |
W. M. Bahgat, H. M. Balaha, Y. AbdulAzeem, M. M. Badawy, An optimized transfer learning-based approach for automatic diagnosis of COVID-19 from chest x-ray images, PeerJ Comput. Sci., 7 (2021), e555. https://doi.org/10.7717/peerj-cs.555 doi: 10.7717/peerj-cs.555
![]() |
[42] | K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, (2016), 770–778. |
[43] |
L. Wen, X. Li, L. Gao, A transfer convolutional neural network for fault diagnosis based on ResNet-50, Neural Comput. Appl., 32 (2020), 6111–6124. https://doi.org/10.1007/s00521-019-04097-w doi: 10.1007/s00521-019-04097-w
![]() |
[44] |
J. Yang, J. Yu, C. Huang, Adaptive multistrategy ensemble particle swarm optimization with Signal-to-Noise ratio distance metric, Inf. Sci., 612 (2022), 1066–1094. https://doi.org/10.1016/j.ins.2022.07.165 doi: 10.1016/j.ins.2022.07.165
![]() |
[45] |
D. Zhu, Z. Huang, S. Liao, C. Zhou, S. Yan, G. Chen, Improved bare bones particle swarm optimization for DNA sequence design, IEEE Trans. Nanobiosci., 22 (2022), 603–613. https://doi.org/10.1109/TNB.2022.3220795 doi: 10.1109/TNB.2022.3220795
![]() |
[46] |
H. T. Kahraman, S. Aras, E. Gedikli, Fitness-distance balance (FDB): A new selection method for meta-heuristic search algorithms, Knowledge-Based Syst., 190 (2020), 105169. https://doi.org/10.1016/j.knosys.2019.105169 doi: 10.1016/j.knosys.2019.105169
![]() |
[47] |
R. Zheng, A. G. Hussien, R. Qaddoura, H. Jia, L. Abualigah, S. Wang, A multi-strategy enhanced African vultures optimization algorithm for global optimization problems, J. Comput. Des. Eng., 10 (2023), 329–356. https://doi.org/10.1093/jcde/qwac135 doi: 10.1093/jcde/qwac135
![]() |
[48] |
D. Zhu, S. Wang, C. Zhou, S. Yan, Manta ray foraging optimization based on mechanics game and progressive learning for multiple optimization problems, Appl. Soft Comput., 145 (2023), 110561. https://doi.org/10.1016/j.asoc.2023.110561 doi: 10.1016/j.asoc.2023.110561
![]() |
[49] |
R. Murugan, T. Goel, S. Mirjalili, D. K. Chakrabartty, WOANet: Whale optimized deep neural network for the classification of COVID-19 from radiography images, Biocybern. Biomed. Eng., 41 (2021), 1702–1718. https://doi.org/10.1016/j.bbe.2021.10.004 doi: 10.1016/j.bbe.2021.10.004
![]() |
[50] | C. Szegedy, W. Liu, Y. Jia, Going deeper with convolutions, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, IEEE, (2015), 1–9. |
[51] | K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, preprint, arXiv: 1409.1556. |
[52] |
L. Kong, J. Cheng, Classification and detection of COVID-19 X-Ray images based on DenseNet and VGG16 feature fusion, Biomed. Signal Process. Control, 77 (2022), 103772. https://doi.org/10.1016/j.bspc.2022.103772 doi: 10.1016/j.bspc.2022.103772
![]() |
[53] |
A. Karacı, VGGCOV19-NET: Automatic detection of COVID-19 cases from X-ray images using modified VGG19 CNN architecture and YOLO algorithm, Neural Comput. Appl., 34 (2022), 8253–8274. https://doi.org/10.1007/s00521-022-06918-x doi: 10.1007/s00521-022-06918-x
![]() |
[54] |
A. Krizhevsky, I. Sutskever, G. E. Hinton, Imagenet classification with deep convolutional neural networks, Commun. ACM, 60 (2017), 84–90. https://doi.org/10.1145/3065386 doi: 10.1145/3065386
![]() |
[55] |
G. Muhammad, M. S. Hossain, COVID-19 and non-COVID-19 classification using multi-layers fusion from lung ultrasound images, Inf. Fusion, 72 (2021), 80–88. https://doi.org/10.1016/j.inffus.2021.02.013 doi: 10.1016/j.inffus.2021.02.013
![]() |
[56] | I. Bankman, Handbook of Medical Image Processing and Analysis, Elsevier, 2008. |
[57] |
Y. Song, S. Zheng, L. Li, X. Zhang, X. Zhang, Z. Huang, et al., Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images, IEEE/ACM Trans. Comput. Biol. Bioinf., 18 (2021), 2775–2780. https://doi.org/10.1109/TCBB.2021.3065361 doi: 10.1109/TCBB.2021.3065361
![]() |
[58] |
Y. Pathak, P. K. Shukla, K. V. Arya, Deep bidirectional classification model for COVID-19 disease infected patients, IEEE/ACM Trans. Comput. Biol. Bioinf., 18 (2020), 1234–1241. https://doi.org/10.1109/TCBB.2020.3009859 doi: 10.1109/TCBB.2020.3009859
![]() |
[59] |
Y. Pathak, P. K. Shukla, A. Tiwari, S. Stalin, S. Singh, P. K. Shukla, Deep transfer learning based classification model for COVID-19 disease, IRBM, 43 (2022), 87–92. https://doi.org/10.1016/j.irbm.2020.05.003 doi: 10.1016/j.irbm.2020.05.003
![]() |
[60] |
X. Fan, X. Feng, Y. Dong, H. Hou, COVID-19 CT image recognition algorithm based on transformer and CNN, Displays, 72 (2022), 102150. https://doi.org/10.1016/j.displa.2022.102150 doi: 10.1016/j.displa.2022.102150
![]() |
[61] |
A. S. Ebenezer, S. D. Kanmani, M. Sivakumar, S. J. Priya, Effect of image transformation on EfficientNet model for COVID-19 CT image classification, Mater. Today Proc., 51 (2022), 2512–2519. https://doi.org/10.1016/j.matpr.2021.12.121 doi: 10.1016/j.matpr.2021.12.121
![]() |
[62] |
N. S. Shaik, T. K. Cherukuri, Transfer learning based novel ensemble classifier for COVID-19 detection from chest CT-scans, Comput. Biol. Med., 141 (2022), 105127. https://doi.org/10.1016/j.compbiomed.2021.105127 doi: 10.1016/j.compbiomed.2021.105127
![]() |
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |