
As the focus of the new round of technological revolution, it is crucial to explore the role of artificial intelligence (AI) technology innovation in improving total factor productivity (TFP). Based on the data from 30 Chinese provinces from 2003 to 2021, this article measured AI innovation using the number of patent applications and empirically investigated the effects of AI technology innovation on TFP. The results demonstrated that AI technology innovation exerts significantly positive influences on the TFP. The mechanism analyses revealed that AI technology innovation improves TFP by upgrading industrial structures and promoting human capital. The subsample results indicated that the promotion effect of AI technology innovation on TFP is significant only in areas with high levels of marketization, financial development, and digital infrastructure. The panel quantile regression results indicated that as the TFP increases, the promoting effect of AI technology innovation on TFP gradually strengthens. This study offers comprehensive empirical evidence for understanding the impacts of AI technology innovation on TFP, giving a reference for further enhancing the level of AI development and promoting a sustainable economic development.
Citation: Shuang Luo, Wenting Lei, Peng Hou. Impact of artificial intelligence technology innovation on total factor productivity: an empirical study based on provincial panel data in China[J]. National Accounting Review, 2024, 6(2): 172-194. doi: 10.3934/NAR.2024008
[1] | Saranya Muniyappan, Arockia Xavier Annie Rayan, Geetha Thekkumpurath Varrieth . DTiGNN: Learning drug-target embedding from a heterogeneous biological network based on a two-level attention-based graph neural network. Mathematical Biosciences and Engineering, 2023, 20(5): 9530-9571. doi: 10.3934/mbe.2023419 |
[2] | Jiahui Wen, Haitao Gan, Zhi Yang, Ran Zhou, Jing Zhao, Zhiwei Ye . Mutual-DTI: A mutual interaction feature-based neural network for drug-target protein interaction prediction. Mathematical Biosciences and Engineering, 2023, 20(6): 10610-10625. doi: 10.3934/mbe.2023469 |
[3] | Peter Hinow, Philip Gerlee, Lisa J. McCawley, Vito Quaranta, Madalina Ciobanu, Shizhen Wang, Jason M. Graham, Bruce P. Ayati, Jonathan Claridge, Kristin R. Swanson, Mary Loveless, Alexander R. A. Anderson . A spatial model of tumor-host interaction: Application of chemotherapy. Mathematical Biosciences and Engineering, 2009, 6(3): 521-546. doi: 10.3934/mbe.2009.6.521 |
[4] | Wen Zhu, Yuxin Guo, Quan Zou . Prediction of presynaptic and postsynaptic neurotoxins based on feature extraction. Mathematical Biosciences and Engineering, 2021, 18(5): 5943-5958. doi: 10.3934/mbe.2021297 |
[5] | Bo Zhou, Bing Ran, Lei Chen . A GraphSAGE-based model with fingerprints only to predict drug-drug interactions. Mathematical Biosciences and Engineering, 2024, 21(2): 2922-2942. doi: 10.3934/mbe.2024130 |
[6] | Xinglong Yin, Lei Liu, Huaxiao Liu, Qi Wu . Heterogeneous cross-project defect prediction with multiple source projects based on transfer learning. Mathematical Biosciences and Engineering, 2020, 17(2): 1020-1040. doi: 10.3934/mbe.2020054 |
[7] | Huiqing Wang, Sen Zhao, Jing Zhao, Zhipeng Feng . A model for predicting drug-disease associations based on dense convolutional attention network. Mathematical Biosciences and Engineering, 2021, 18(6): 7419-7439. doi: 10.3934/mbe.2021367 |
[8] | Rachael C. Adams, Behnam Rashidieh . Can computers conceive the complexity of cancer to cure it? Using artificial intelligence technology in cancer modelling and drug discovery. Mathematical Biosciences and Engineering, 2020, 17(6): 6515-6530. doi: 10.3934/mbe.2020340 |
[9] | Dong Ma, Shuang Li, Zhihua Chen . Drug-target binding affinity prediction method based on a deep graph neural network. Mathematical Biosciences and Engineering, 2023, 20(1): 269-282. doi: 10.3934/mbe.2023012 |
[10] | Xianfang Wang, Qimeng Li, Yifeng Liu, Zhiyong Du, Ruixia Jin . Drug repositioning of COVID-19 based on mixed graph network and ion channel. Mathematical Biosciences and Engineering, 2022, 19(4): 3269-3284. doi: 10.3934/mbe.2022151 |
As the focus of the new round of technological revolution, it is crucial to explore the role of artificial intelligence (AI) technology innovation in improving total factor productivity (TFP). Based on the data from 30 Chinese provinces from 2003 to 2021, this article measured AI innovation using the number of patent applications and empirically investigated the effects of AI technology innovation on TFP. The results demonstrated that AI technology innovation exerts significantly positive influences on the TFP. The mechanism analyses revealed that AI technology innovation improves TFP by upgrading industrial structures and promoting human capital. The subsample results indicated that the promotion effect of AI technology innovation on TFP is significant only in areas with high levels of marketization, financial development, and digital infrastructure. The panel quantile regression results indicated that as the TFP increases, the promoting effect of AI technology innovation on TFP gradually strengthens. This study offers comprehensive empirical evidence for understanding the impacts of AI technology innovation on TFP, giving a reference for further enhancing the level of AI development and promoting a sustainable economic development.
Drug-target interactions (DTIs) involve the binding of a drug to the relevant site of a target protein to trigger a biochemical reaction [1]. The efficacy is related to the biological activity of the protein. However, it is complicated for experiments to predict a drug's success and drug discovery is time-consuming and expensive [2,3], which is estimated to typically take 12–15 years and cost over $100 million [4]. For these reasons, in the past decades, computer-aided drug design (CADD) has been proposed to discriminate new drugs and consists of processes such as virtual screening, molecular docking, and QSAR methods [5]. Currently, due to limited ligand data and the limited information on the structure of novel target proteins [6], these approaches are inappropriate and inefficient given the growth of available biological and chemical data [2]. Recently, with the advent of various deep learning methods, a significant future trend in AI-based drug discovery has been identified [7]. It is essential for drug discovery to accurately predict the number of DTIs [8]. Therefore, it is urgent to devise richer and more compatible computational methods to differentiate between potential DTIs.
The concept of "guilt-by-association" [9] has been described in DTIs prediction. It is defined that if drug A has target proteins, and the action event between drug B is similar to drug A, targets interactions are likely to appear, and the reverse is also true. Machine learning methods are used for DTIs prediction and can successfully solve the assumption. For instance, Mei et al. [10] proposed bipartite local models (BLMs) that considered neighbors' interaction profiles where neighbor-based interaction-profile inferring (NII) can be effective in defining a new candidate problem. Luo et al. [11] used an inductive matrix completion method, in which seven kinds of drug/target-related similarities were included in an integrated network (e.g., drugs, proteins, diseases, and side-effects). Ezzat et al. [12] proposed graph regularized matrix factorization (GRMF) and weighted graph regularized matrix factorization (WGRMF) methods that introduced graph regularization into the matrix factorization in order to learn manifolds. Moreover, a preprocessing step (WKNKN) has been developed to rescore unknown drug-target pairs that were previously regarded as null values. Although these methods have been proven to be effective, there are challenges to overcome complex data structures such as interaction networks of drugs or targets. Furthermore, the rapid growth of drug/target-related data has outpaced their ability to process and analyze information. With the emergence of diverse and enriched feature representations, the efficacy of the above methods may limit the exploration of more comprehensive topological information and node characteristics between drugs and target proteins.
Network-based algorithms and feature-based algorithms become famous in the field. Generally, identifying DTIs is considered as a binary classification task by extracting features vectors of drugs as well as targets. Several number of heterogeneous data have been integrated into a heterogeneous network to boost the accuracy of DTIs prediction tasks [13]. The deep belief network (DBN) [14] has been proposed to build an end-to-end method for abstracting raw input samples. Moreover, sequence-based approaches are universal. Different architectures [15,16,17,18] have been developed for feature extraction of sequence information. DrugVQA [19] employs a bidirectional long-short time memory network to tackle the prediction problem. Furthermore, graph-based methods are suitable for the two-dimensional representation of structural information. Zhao et al. [20] utilized a combination of graph convolutional network (GCN) and deep neural network (DNN) to enhance the identification of DTIs. GNN was coupled with CNN, which was designed as drug feature and target feature extraction method [21]. LASSO has been employed by You et al. [22] as a feature procession. Thafar et al. [23] constructed the DTi2Vec model including graph embedding which capture relationships between drugs and targets and then these features are fed into the ensemble classifier for prediction analyses. Huang et al. designed a molecular sub-structure representation and used massive unlabeled biomedical data through an augment transformer [24]. Peng et al. [25] introduced CNN to identify DTIs and trained the denoising autoencoder (DAE) as a feature selector. Although these methods can effectively predict DTIs, the problem of parameter count and computation amount need to be given more attention.
The broad learning system (BLS) [26] is characterized by a relatively simple neural network architecture comprising only three layers of neurons. Inspired by the concept of the random vector functional-link neural network (RVFLNN) [27,28], its training procedure is facilitated through pseudo-inverse calculations. Due to its training procedure and flat structure, BLS has the advantages of fast computing speed and few training parameters. Therefore, BLS has been widely applied to various disciplines including medicine [29]. For instance, Fan et al. [30] proposed a stacked ensemble classifier build by BLS for the prediction of interactions between lncRNA and proteins. Zheng et al. [31] designed a modified BLS-based model to predict miRNA-disease associations using sequence similarities of microRNA (miRNAs). The above applications of BLS in this area have been proven to be useful. However, there is a lack of related research for DTIs based on BLS. Additionally, since labeled data volumes are always sparse and insufficient, prediction modeling is to performance is inadequate. By fusing information from multiple aspects to overcome the limitations, the above methods can improve the performance, indicating that these combined models could solve the challenge of interaction matrix sparsity.
In this study, we developed a novel model called ConvBLS-DTI to predict DTIs. Compared with the previous DTI predictive methods, ConvBLS-DTI integrates matrix factorization with the broad learning system, yielding reliable DTI prediction results. The task of DTI prediction is formulated as a binary classification problem to determine whether a drug-target pair is a DTI. The major contributions of this paper are as follows:
1) We address the challenges of data sparsity and incompleteness by employing a WKNKN algorithm as a pre-processing step, which help to mitigate the adverse effects of a large number missing interaction value.
2) We propose a matrix factorization technique used on the interaction matrix to generate two latent feature matrices for drugs and targets, thereby enabling the learning of low-dimensional vector representations of features.
3) Based on the CNN algorithm, ConvBLS-DTI can handle the DTIs prediction, taking the extracted drug-target pairs feature vectors as inputs.
The architecture of the proposed ConvBLS-DTI method is depicted in Figure 1. It is primarily composed of three sessions: First, we utilized the WKNKN algorithm to alleviate the sparsity of the DTI matrix, thus enhancing the input information complement of the model and improving its predictive performance. After construction of the DTI matrix, matrix factorization is used to decompose DTI matrix into two feature matrices of low ranks which obtains vector representation of the drug features and target features. Then, the drug feature vectors combine with the target feature vectors together to get the final feature vectors. Finally, ConvBLS is built for classification. A CNN is leveraged to enhance the nodes' representation, followed by a broad learning module, which further enable satisfactory results in effective identification of DTIs.
We initially reconstruct the interaction matrix using the computational preprocessing technique, which can effectively complement the interaction matrix for the identification of DTIs and improve the known DTI samples. As shown on the left side of step A in Figure 1, the green circles, the red triangles, and the blue lines separately denote drugs, targets and the known interaction. D={di}ndi=1 and T={tj}ntj=1 are separately described as each node for drugs and targets, where nd is the number of drugs and nt is the number of targets. The associations between nd drugs and nt targets are represented by an interaction matrix Y∈{0,1}, in which Yij=1 indicates a known interaction between drug di and target tj, and Yij=0 otherwise. In addition, the similarity matrix of both drugs and targets is represented as SD∈Rnd×nd and ST∈Rnt×nt.
Numerous unknown interactions can significantly impact the evaluation outcome of the model and introduce prediction bias. In DTIs prediction, weighted k-nearest neighbors (k-NN) has been employed to leverage similarity measures to promote further prediction performance. Weighted k-NN considers both neighbor similarity and distances, incorporating distance weights to calculate likelihood values of unconfirmed drug-target interactions. Specifically, given a drug-target pair, the algorithm first identifies the k-NN and assigns weights to each neighbor based on their similarity and distance. Weight involved in WKNKN [12] is computed by Gaussian weighting method. The calculated weighted likelihood values can be used to predict the likelihood of unknown DTIs within the matrix. Here, the specific operation is achieved through the following three steps:
Yd(d)=1Md∑Ki=1ωdiY(di) | (2.1) |
where Yd(d) denotes the likelihood score of interaction for drug di. Md is defined as a normalization term, ω coefficient represents the weights of the K nearest known neighbors of drug di. Similarly, the same terms are computed to estimate the interaction likelihood score of the target tj:
Yt(t)=1Mt∑Ki=1ωtjY(tj) | (2.2) |
where Yt(t) denotes the likelihood score of interaction for target tj. Mt is the normalization term and ω coefficient represents the weights of the K nearest known neighbors of target tj. Finally, the derived formula is as follows:
YWKNKN=max(Yd+Yt2,Y) | (2.3) |
Therefore, if Yij is 0, YWKNKN replaces it with an average of the weighted interaction likelihood value. For the matrix representation, 0 and 1 denote the absence and presence of interactions between drugs and targets, respectively. Likelihood serves as a measure of the possibility of interaction between a drug and a target, typically ranging from 0 to 1. Higher likelihood values indicate a higher likelihood of interaction, while lower likelihood values suggest a lower likelihood.
Considering that most studies mainly concentrate on extracting features from drugs and targets individually and less on the relationships of the DTI, neighbor regularization logistic matrix factorization (NRLMF) [32] is used to represent drugs and targets in the right part of step B. NRLMF is an unsupervised learning strategy that mainly infers unknowns through known interactions and their similarities, so no negative samples are required. The valid connections are denoted as the modified interaction matrix made of known and unknown interactions. As shown in Figure 1, the DTI probabilities can be defined as a logistic function:
Pi,j=exp(uivTj)1+exp(uivTj) | (2.4) |
where each term ui∈R1×r is denoted as the r-dimensional potential representation of each drug di. Similarly, each term vj∈R1×r represents the r-dimensional potential representation of each target tj. In this way, the potential feature vectors for all drugs and all targets can be summarized as U=(uT1,…,uTnd) and V=(vT1,…,vTnt), where T refers to the transpose of the matrix.
The neighborhood regularization method proposes to add the nearest neighbors of drugs and targets to further increase information diversity and enable higher accuracy without overfitting. The neighborhood regularization is achieved by:
α2∑ndi∑ndjSPiμ||ui−uj||2f | (2.5) |
α2∑nti∑ntjSQϑj||vi−vj||2f | (2.6) |
where α is the Laplace regularization parameter, and ||⋅||f is the Frobenius norm of the matrix, and the parameters SP and SQ represent neighbors similarity measure matrix are given by:
SPiμ=SDiμ if dμ∈Wd(di) else SPiμ=0 | (2.7) |
SQϑj=STϑj if tϑ∈Wt(tj) else SQϑj=0 | (2.8) |
where SD and ST denote as similarity matrix, and Wd(di) is defined as the nearest neighbors of a node di, and Wt(tj) is defined as the nearest neighbors of a node tj.
The matrix factorization (MF) method decomposes the interaction matrix into two low-rank matrices. The MF is formulated as a feature extraction task to obtain the description of the drugs and their targets as features. The feature matrix is obtained by maximizing the objective function via the posterior probability distribution:
maxU,VP(U,V|Y,σ2d,σ2t) | (2.9) |
where Y denotes the interaction matrix, σ2d and σ2t are parameters that control the variance of Gaussian distribution of drug set and target set.
Thus, drugs and targets can be denoted as two r-dimensional feature representations. As illustrated in Figure 2, the drug feature is U−D=[DF1,DF2,...,DFr], and the target feature is V−T=[TF1,TF2,...,TFr]. Then, the drug feature vectors and target feature vectors are merged and assigned the label based on the interaction matrix Y. To ensure the quality of the model, the number of negative samples is equal to positive number in each dataset. Negative samples are randomly generated based on the interaction matrix Y. The pairwise drug–target feature vector is used as input to the neural network, which can be expressed as FV = [DF1,DF2,...,DFr,TF1,TF2,...,TFr].
After concatenating the features of drugs and targets, in order to achieve better performance and train more effectively, the ConvBLS model is used as a classification method to determine the predictions of the DTIs. As shown in Figure 3, we developed a broad learning system that combined convolutional neural network to extract high-quality drug and target representations for better prediction.
ConvBLS mostly includes two parts: The 1D-CNN module and enhancement nodes. CNN block is used to learn a representative features of targets and drugs. The input data of 1D-CNN is a one-dimensional feature vectors, and the convolution kernel is also in one-dimensional form. The enhancement layer is responsible for further feature extracting. The detailed network structure is shown in Figure 3. This section completes the classification task with ConvBLS.
Given the lack of learning ability of the original feature mapping, the CNN block is selected for sequence data of drug-target features. It contains multiple groups of feature mapping nodes composed of a 1D-CNN layer and a max pooling layer. To solve complex tasks, learning models increasingly go deeply. The multiscale random convolution feature is expanded to improve robustness. The detailed computational procedure is as follows:
First, the drug–target predictive model ConvBLS is constructed based on previous obtained feature data FV. The input is connected to the mapping matrix by applying 1-D convolution kernels to generate the corresponding feature representation. All the random items can act as the convolution kernel so that we can achieve the output:
FC=φ(Conv(X,KC)) | (2.10) |
where X is the input feature vectors, KC is the convolution kernel, Conv(.) is denoted as the convolution function, and φ(.) is the activation function. The descriptor of the mapping feature is called FC. Then, the down-sampling method is used to allow the feature to be robust:
FP=max_pool(FC) | (2.11) |
where FP is the result after max pooling function. Next, an enhancement layer is built. Using random weights and nonlinear transformation, the enhancement nodes are obtained:
Ej=ψ(FPWej+bej)j=1,2,…,n | (2.12) |
where ψ(.) is the activation function, weights and bias represented as Wej and bej, which are randomly initialized, and n is the group number of enhanced nodes. All of the enhancement nodes can be represented as En≡[E1,E2,…En]. Finally, the improved feature layer and enhancement layer are concatenated into one matrix as a single neural network. Hence, featurization constitutes the outputs of weight of the BLS, based on Y=HWdt.
Wdt=H+Y=[FP|En]+Y | (2.13) |
The ridge regression approximation algorithm [33] is utilized to determine the [FP|En]+:
[FP|En]+=limλ→0(λI+[FP|En][FP|En]T)−1[FP|En]T | (2.14) |
Here, we describe the dataset used in this paper and provide the experiment setup and evaluation metrics for comparing model performance in subsequent experiments.
In this study, two benchmark datasets are used for evaluating our proposed model: the Yamanishi's dataset and Luo's dataset. The first one is the gold benchmark dataset created by Yamanishi et al. [34]. It is classified into four categories based on the target protein class, namely: (ⅰ) enzyme (E), (ⅱ) ion channel (IC), (ⅲ) G protein-coupled receptor (GPCR), and (ⅳ) nuclear receptor (NR). Since the discovery of the interactions in these datasets 14 years ago, we implemented the completed version of the original golden standard datasets collected by Liu et al. [35]. The new datasets added information on the KEGG pathways [36], DrugBank [37], and ChEMBL [38] databases. The second one was developed by Luo et al. [11], consisting of four categories of nodes (drugs, proteins, diseases, and side-effects) and six types of connections (drug-target interaction, drug-drug interactions, protein-protein interactions, drug-disease associations, protein-disease associations, and drug-side-effect associations). Table 1 lists the detailed statistical entries of the complete datasets included in our analysis. Sparsity represents the proportion of known DTI numbers in all possible DTI combinations.
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Table 2 lists the parameter settings in the experiments depending on datasets. The best parameters of ConvBLS-DTI were selected by performing a grid search. Some key parameters were set as follows: The number of the nearest known neighbors K is set to 5 for NR and 7 for others; the feature dimension r is set to 50 for a relatively small dataset NR, and 100 was an appropriate setting for GPCR, IC, and E datasets. The convolution kernel size is taken from {3–9}. The Tanh function was chosen as the activation function for every layer. A number of experiments are performed to determine the optimal classification parameters of BLS. Specifically, the shrinkage scale (sc) of the enhancement nodes plays a central role in this experiment. The parameters of all baseline methods were set based on the suggestions from the respective studies available in the literature.
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
For the cross-validation experiments, there are three different experimental settings for comparison, depending on whether the drug and target involved in the test pair are training entities:
1) CVd: Predicts the interactions between testing drugs and training targets;
2) CVt: Predicts the interactions between training drugs and testing targets;
3) CVdt: Predicts the interactions between testing drugs and testing targets.
The 10-fold cross-validation is one of the most widely available methods. All models were trained and tested using 10-fold cross-validation. In this study, the final results are given with the area under the receiver operating characteristic curve (AUC) and the area under the precision-recall curve (AUPR) to judge the prediction performance. They are widely used in this field [39,40]. Since there are few true DTIs, AUPR is a more precise quality indicator than AUC because it punished those in which lots of false positive examples were found from the top-ranked prediction score [41], so we consider it as an evaluation sign. In addition, the Sen score was another metric used in this study. The average values are used for the results of each dataset.
Experiments were run under the environment of Windows 10 Professional Edition and i5-7200H CPU. Our aim of this study was to construct an efficient computation method with excellent performance for DTI prediction. Therefore, we first observe the performance of two BLS-based models from different perspectives on the four datasets. Then, we compared the prediction results of our model with representative methods under three settings: NRLMF [32], DTINet [11], WKNNIR [42], DTi2Vec [43], ADA-GRMFC [44], and BLS-DTI. Finally, the optimal of core parameter in the experiment was reported.
We first compared our model with the BLS-DTI. Tables 3 and 4 list the AUC and AUPR results on the prediction tasks. As shown in Table 3, our model is found to outperform BLS-DTI in the AUC and AUPR. It highlights the importance of the feature extraction ability in the BLS. The enhancement layer is included in the two networks. We considered the prediction performance of BLS is insufficient for DTI prediction tasks due to the lack of the ability to obtain deep features. ConvBLS-DTI provides the better performance in terms of CNN method. Specifically, ConvBLS-DTI exhibits higher results than BLS-DTI for the E dataset, providing 0.22 higher AUC score, an improvement of 28%, with a 0.152 greater AUPR value, an improvement of 19%. The same positive results are also found in the other three datasets, which indicates that the performance of the model ConvBLS-DTI is improved when the CNN method is added to the BLS network.
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
For a more comprehensive evaluation, the following Table 4 shows the addition AUC and AUPR scores of each experimental setting on four datasets. Similarly, ConvBLS-DTI obtain higher performance in all scenarios, outperforming the other method BLS-DTI. Compared with NR and GPCR datasets, IC and E datasets contribute to higher AUC and AUPR scores, with AUPR values of 0.947 and 0.961, respectively. The possible reason is that the number of DTIs in the NR and GPCR categories is smaller than the other categories, especially for NR with only 166 drug-target pairs.
In this section, under the same datasets, evaluation metrics and experimental scenarios (CVd, CVt, and CVdt), six advanced methods, including NRLMF, DTINet, WKNNIR, DTi2Vec, ADA-GRMFC, and BLS-DTI, are involved into the performance comparison. Tables 5 and 6 show the AUC and AUPR of the methods participating in the CVd and CVt settings. In general, based on the main evaluation metrics, our method has overall better performance than the other methods under different scenarios. For CVd, ConvBLS-DTI shows a high performance in all datasets. For CVt, minimal difference is found in the AUC score obtained using IC and GPCR datasets, but the AUPR result achieved by our method increases by 2.05%, 4.41%, 3.6%, 5.54%, and 6.12% on NR, GPCR, IC, E, and Luo datasets, respectively, compared with that of the second-best model. In particular, ConvBLS-DTI performs better than BLS-DTI. In the results of predicting novel drugs and known targets, ConvBLS-DTI is better than other methods. In the experiment scenario of CVd, ConvBLS-DTI achieves AUPR values of 0.917, 0.968, 0.972, 0.958, and 0.972 on NR, IC, GPCR, E, and Luo datasets, respectively. In the experiment scenario of CVt, the AUPR values of ConvBLS-DTI are 0.846, 0.950, 0.946, 0.952, and 0.954 on NR, IC, GPCR, E, and Luo datasets, respectively. Overall, it can be concluded that the proposed ConvBLS-DTI is superior to all the compared methods and proves that broad learning system can also be a rational tool to help for predicting DTIs.
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
In particular, the ConvBLS-DTI has satisfactory performance under the CVdt setting. The AUC and AUPR histograms for the different algorithms are shown in Figures 4 and 5, respectively. The results are entirely consistent across all datasets given in the CVdt (specifically in terms of AUPR metric). For CVdt, the AUC and AUPR values of ConvBLS-DTI are higher than other methods on all datasets, although DTi2Vec is very competitive compared with ConvBLS-DTI. Overall, our method improves the AUPR more than the AUC.
To measure the impact of the WKNKN method on ConvBLS-DTI, ablation experiments were conducted by removing the WKNKN method under three different CV strategies using the Luo et al. dataset. The variant of ConvBLS-DTI without the WKNKN method is denoted as ConvBLS-DTI (without WKNKN). Performance comparisons between ConvBLS-DTI and the variant in terms of AUC and AUPR are presented in Tables 7 and 8, respectively. The findings in Tables 7 and 8 suggest that the utilization of the WKNKN method contributes to improve the performance of ConvBLS-DTI.
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |
In this study, the datasets IC and GPCR were applied to test the influence of the convolution kernel size. As illustrated in Figure 6, by varying the size of the convolutional kernel (3, 4, 5, 6, 7, 8, and 9), the AUPR value of the ConvBLS-DTI method progressively improved with an increase in kernel size and reached its optimal performance when kernel size was set to 5. Subsequently, the performance showed a decline. A kernel size set to 5 achieved good results.
In this paper, we aimed to solve the problem of sparsity and incompletion of the drug interaction data. A new framework called ConvBLS-DTI was proposed to predict DTIs by applying an advanced fusion of BLS approach. Our method integrates WKNKN, MF, and BLS to improve the DTI prediction results. The method takes advantage of matrix factorization for the latent low-dimensional feature representation and predicts DTIs based on the broad learning architecture. Moreover, the WKNKN algorithm was used as a preprocessing step to increase the availability of relevant information for a large number of missing correlations. Compared with the BLS-DTI, our model achieved AUC and AUPR values of 0.971 and 0.967, respectively, for the IC dataset under tenfold-cross-validation experiments. These findings illustrate that the combination of CNN and BLS could improve the prediction performance for DTIs. Additionally, compared with other previous methods, the best AUC and AUPR values of the proposed method were 0.9643 and 0.9681 for the E dataset and CVdt setting, respectively. The results show that our model acquires improved prediction effect on AUC and AUPR using extensive experimental verification.
In future studies, greater emphasis will be placed on optimizing the BLS structure to enhance the feature extraction ability. In fact, the results of the present prediction model can be heavily influenced by the mapping algorithms and the effectiveness of the dataset. Therefore, our models might be further developed using other deep-learning models to increase the identifying power. Overall, with the availability of more data and the development of new approaches, it is expected that more applications of our model can be achieved.
The authors that declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work was supported by the project of the Natural Science Foundation of Shandong Province, China (Natural Science Foundation of Shandong Province, No. ZR2019PEE018), Shandong Province Science and Technology SMES Innovation Ability Enhancement Project (Natural Science Foundation of Shandong Province, No. 2021TSGC1063), Major Scientific and Technological Innovation Projects of Shandong Province (Natural Science Foundation of Shandong Province, No. 2019JZZY020101), and the project of the Natural Science Foundation of Qingdao (No. 23-2-1-216-zyyd-jch).
The authors declare that there are no conflicts of interest.
[1] |
Acemoglu D, Autor D, Dorn D, et al. (2014) Return of the solow paradox? It, productivity, and employment in US Manufacturing. Am Econ Rev 104: 394–399. https://doi.org/10.1257/aer.104.5.394 doi: 10.1257/aer.104.5.394
![]() |
[2] | Acemoglu D, Restrepo P (2019) Artificial intelligence, automation and work. In: Agrawal, A., Gans, J., Goldfarb, A., (eds), The Economics of Artificial Intelligence: An Agenda, University of Chicago Press, 197–236. https://doi.org/10.7208/chicago/9780226613475.003.0008 |
[3] |
Acemoglu D, Restrepo P (2018) The race between man and machine: implications of technology for growth, factor shares, and employment. Am Econ Rev 108: 1488–1542. https://doi.org/10.1257/aer.20160696 doi: 10.1257/aer.20160696
![]() |
[4] |
Aghion P, Blundell RW, Griffith R, et al. (2009) The effects of entry on incumbent innovation and productivity. Rev Econ Stat 91: 20–32. https://doi.org/10.1162/rest.91.1.20 doi: 10.1162/rest.91.1.20
![]() |
[5] | Aghion P, Howitt P (1992) A model of growth through creative destruction. Econometrica 60: 323–351. |
[6] |
Alrowwad A, Abualooush SH, Masa'Deh R (2020) Innovation and intellectual capital as intermediary variables among transformational leadership, transactional leadership, and organizational performance. J Manag Dev 39: 196–222. https://doi.org/10.1108/JMD-02-2019-0062 doi: 10.1108/JMD-02-2019-0062
![]() |
[7] | Brynjolfsson E, Rock D, Syverson C (2019) Artificial intelligence and the modern productivity paradox: a clash of expectations and statistics. In: Agrawal, A., Gans, J., Goldfarb, A., (eds), The Economics of Artificial Intelligence: An Agenda, University of Chicago Press, 23–60. https://doi.org/10.7208/chicago/9780226613475.003.0001 |
[8] |
Cao J, Law SH, Samad ARBA, et al. (2022) Effect of financial development and technological innovation on green growth-analysis based on spatial durbin model. J Clean Prod 365. https://doi.org/10.1016/j.jclepro.2022.132865 doi: 10.1016/j.jclepro.2022.132865
![]() |
[9] |
Chang L, Taghizadeh-Hesary F, Mohsin M (2023) Role of artificial intelligence on green economic development: joint determinates of natural resources and green total factor productivity. Resour Policy 82. https://doi.org/10.1016/j.resourpol.2023.103508 doi: 10.1016/j.resourpol.2023.103508
![]() |
[10] |
Dong F, Hu MY, Gao YJ, et al. (2022) How does digital economy affect carbon emissions? Evidence from global 60 countries. Sci Total Environ 852. https://doi.org/10.1016/j.scitotenv.2022.158401 doi: 10.1016/j.scitotenv.2022.158401
![]() |
[11] |
Galor O, Moav O (2002) Natural selection and the origin of economic growth. Q J Econ 117: 1133–1191. https://doi.org/10.1162/003355302320935007 doi: 10.1162/003355302320935007
![]() |
[12] |
Ge P, Liu T, Huang X (2023) The effects and drivers of green financial reform in promoting environmentally-biased technological progress. J Environ Manage 339. https://doi.org/10.1016/j.jenvman.2023.117915. doi: 10.1016/j.jenvman.2023.117915
![]() |
[13] | Graetz G, Michaels G (2015) Robots at work: the impact on productivity and jobs. Centre for Economic Performance, LSE. |
[14] |
Hopenhayn HA (2014) Firms, misallocation, and aggregate productivity: a review. Annu Rev Econom 6: 735–770. https://doi.org/10.1146/annurev-economics-082912-110223 doi: 10.1146/annurev-economics-082912-110223
![]() |
[15] |
Jiang J, Su P, Ge Z (2021) The high-and new-technology enterprise identification, marketization process and the total factor productivity of enterprise. Kybernetes 50: 528–549. https://doi.org/10.1108/K-11-2019-0743 doi: 10.1108/K-11-2019-0743
![]() |
[16] | Jiang W, Li P (2022) Ai and tfp: "technology dividend" or "technology gap". J Stat Inform 37:26–35. |
[17] |
Kijek A, Kijek T (2020) Nonlinear effects of human capital and r & d on tfp: evidence from european regions. Sustainability 12. https://doi.org/10.3390/su12051808 doi: 10.3390/su12051808
![]() |
[18] |
Lee H, Yang SA, Kim K (2019) The role of fintech in mitigating information friction in supply chain finance. Asian Development Bank Economics Working Paper Series. http://dx.doi.org/10.22617/WPS190574-2 doi: 10.22617/WPS190574-2
![]() |
[19] |
Lei Z, Wang D (2023) Digital transformation and total factor productivity: empirical evidence from china. Plos One 18. https://doi.org/10.1371/journal.pone.0292972 doi: 10.1371/journal.pone.0292972
![]() |
[20] | Lewbel A (1997) Constructing instruments for regressions with measurement error when no additional data are available, with an application to patents and r & d. Econometrica, 1201–1213. |
[21] | Liang S, Dong Q (2023) Management's macroeconomic cognition and corporate default risk. J Quant Technol Econ 40:200–220. |
[22] |
Lin B, Zhu J (2019) The role of renewable energy technological innovation on climate change: empirical evidence from china. Sci Total Environ 659: 1505–1512. https://doi.org/10.1016/j.scitotenv.2018.12.449 doi: 10.1016/j.scitotenv.2018.12.449
![]() |
[23] |
Liu J, Chang H, Forrest JY, et al. (2020) Influence of artificial intelligence on technological innovation: evidence from the panel data of china's manufacturing sectors. Technol Forecast Soc 158. https://doi.org/10.1016/j.techfore.2020.120142 doi: 10.1016/j.techfore.2020.120142
![]() |
[24] |
Meng T, Yu D, Ye L, et al. (2023) Impact of digital city competitiveness on total factor productivity in the commercial circulation industry: evidence from china's emerging first-tier cities. Humanit Soc Sci Commun 10. https://doi.org/10.1057/s41599-023-02390-7 doi: 10.1057/s41599-023-02390-7
![]() |
[25] |
Nordhaus WD (2021) Are we approaching an economic singularity? Information technology and the future of economic growth. Am Econ J Macroecon 13: 299–332. https://doi.org/10.1257/mac.20170105 doi: 10.1257/mac.20170105
![]() |
[26] | Pan W, He Z, Pan H (2021) Research on spatiotemporal evolution and distribution dynamics of digital economy development in china. China Soft Sci 10: 137–147. |
[27] |
Pan X, Chu J, Tian M, et al. (2022) Non-linear effects of outward foreign direct investment on total factor energy efficiency in china. Energy 239. https://doi.org/10.1016/j.energy.2021.122293 doi: 10.1016/j.energy.2021.122293
![]() |
[28] |
Ren XH, Zeng GD, Gozgor G (2023) How does digital finance affect industrial structure upgrading? Evidence from chinese prefecture-level cities. J Environ Manage 330. https://doi.org/10.1016/j.jenvman.2022.117125 doi: 10.1016/j.jenvman.2022.117125
![]() |
[29] | Ren Y, Liu Y, Li H (2023) Artificial intelligence technology innovationand enterprise total factor productivity. Bus Manag J 45: 50–60. |
[30] |
Song W, Mao H, Han X (2021) The two-sided effects of foreign direct investment on carbon emissions performance in china. Sci Total Environ 791. https://doi.org/10.1016/j.scitotenv.2021.148331 doi: 10.1016/j.scitotenv.2021.148331
![]() |
[31] |
Tang C, Xu YY, Hao Y, et al. (2021) What is the role of telecommunications infrastructure construction in green technology innovation? A firm-level analysis for china. Energ Econ 103. https://doi.org/10.1016/j.eneco.2021.105576 doi: 10.1016/j.eneco.2021.105576
![]() |
[32] | Tang S, Lai X, Huang R (2019) How can fintech innovation affect tfp: facilitating or inhibiting? theoretical analysis framework and regional practice. China Soft Sci 7: 134–144. |
[33] | Tang S, Wu X, Zhu J (2020) Digital Finance and Enterprise Technology Innovation: Structural Feature, Mechanism Identification and Effect Difference under Financial Supervision. Manag World 36: 52–66. |
[34] | Valli V, Saccone D (2009) Structural change and economic development in china and india. Eur J Comp Econ 6 |
[35] |
Wang CG, Liu TS, Zhu Y, et al. (2022) Digital economy, environmental regulation and corporate green technology innovation: evidence from china. Int J Env Res Pub He 19. https://doi.org/10.3390/ijerph192114084. doi: 10.3390/ijerph192114084
![]() |
[36] |
Wang KL, Sun TT, Xu RY, et al. (2023) The impact of artificial intelligence on total factor productivity: empirical evidence from china's manufacturing enterprises. Econ Change Restruct 56: 1113–1146. https://doi.org/10.1007/s10644-022-09467-4 doi: 10.1007/s10644-022-09467-4
![]() |
[37] | Wang X, Fan G (2000) Sustainability of china's economic growth. Economic Science Press, Shanghai. |
[38] | Wang X, Hu L, Fan G (2021) Marketization index of china's provinces:neri report 2021. Social Sciences Academic Press (China). |
[39] | Wang Z, Han C, Zhu W (2022) Research on the impact of digital finance development on complexity of export technology. World Econ Stud 8: 26–42. |
[40] |
Xiong J, Chen L (2022) Dialect diversity and total factor productivity: evidence from chinese listed companies. Front Psychol 13. https://doi.org/10.3389/fpsyg.2022.1017397 doi: 10.3389/fpsyg.2022.1017397
![]() |
[41] |
Yan Z, Zou B, Du K, Li K (2020) Do renewable energy technology innovations promote china's green productivity growth? Fresh evidence from partially linear functional-coefficient models. Energ Econ 90. https://doi.org/10.1016/j.eneco.2020.104842 doi: 10.1016/j.eneco.2020.104842
![]() |
[42] |
Yao S, Zhang S, Zhang X (2019) Renewable energy, carbon emission and economic growth: a revised environmental kuznets curve perspective. J Clean Prod 235: 1338–1352. https://doi.org/10.1016/j.jclepro.2019.07.069 doi: 10.1016/j.jclepro.2019.07.069
![]() |
[43] |
You J, Xiao H (2022) Can fdi facilitate green total factor productivity in china? Evidence from regional diversity. Environ Sci Pollut R 29: 49309–49321. https://doi.org/10.1007/s11356-021-18059-0 doi: 10.1007/s11356-021-18059-0
![]() |
[44] |
Zeng S, Shu X, Ye W (2022) Total factor productivity and high-quality economic development: a theoretical and empirical analysis of the yangtze river economic belt, china. Int J Env Res Pub He 19. https://doi.org/10.3390/ijerph19052783 doi: 10.3390/ijerph19052783
![]() |
[45] |
Zhai S, Liu Z (2023) Artificial intelligence technology innovation and firm productivity: evidence from china. Financ Res Lett 58. https://doi.org/10.1016/j.frl.2023.104437 doi: 10.1016/j.frl.2023.104437
![]() |
[46] | Zhang B, Sun X (2015) Total factor productivity of economic growth. Journal of Ocean University of China (Social Sciences), 73–78. |
[47] | Zhang J, Wu G, Peng J (2004) The estimation of china's provincial capital stock: 1952–2000. Economic Research Journal, 33–44. |
[48] |
Zheng W, Walsh PP (2019) Economic growth, urbanization and energy consumption - a provincial level analysis of china. Energ Econ 80: 153–162. https://doi.org/10.1016/j.eneco.2019.01.004 doi: 10.1016/j.eneco.2019.01.004
![]() |
[49] |
Zhou C, Sun Z, Qi S, et al. (2023) Green credit guideline and enterprise export green-sophistication. J Environ Manage 336. https://doi.org/10.1016/j.jenvman.2023.117648 doi: 10.1016/j.jenvman.2023.117648
![]() |
[50] |
Zou S, Liao Z, Fan X (2024) The impact of the digital economy on urban total factor productivity: mechanisms and spatial spillover effects. Sci Rep-Uk 14: 396. https://doi.org/10.1038/s41598-023-49915-3 doi: 10.1038/s41598-023-49915-3
![]() |
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |
Dataset | Drugs | Targets | Interactions | Sparsity |
NR | 54 | 26 | 166 | 0.118 |
GPCR | 223 | 95 | 1096 | 0.052 |
IC | 210 | 204 | 2331 | 0.054 |
E | 445 | 664 | 4256 | 0.014 |
Luo | 708 | 1512 | 1923 | 0.002 |
Parameter | Value |
K | K ∈ {1, 2, 3, 5, 7, 9} |
r | r ∈ {50,100} |
sc | 2 |
filter size | 4 |
number of filters | 5 |
enhancement nodes | n ∈ [100,1000] |
Method | E | IC | GPCR | NR | ||||
AUC | AUPR | AUC | AUPR | AUC | AUPR | AUC | AUPR | |
BLS-DTI | 0.789 | 0.814 | 0.907 | 0.916 | 0.854 | 0.863 | 0.854 | 0.863 |
ConvBLS-DTI | 0.969 | 0.962 | 0.971 | 0.967 | 0.968 | 0.961 | 0.968 | 0.961 |
Dataset | Method | CVd | CVt | CVdt | ||||||
AUC | AUPR | SEN | AUC | AUPR | SEN | AUC | AUPR | SEN | ||
NR | BLS-DTI | 0.8882 | 0.8753 | 0.8510 | 0.8048 | 0.7942 | 0.8088 | 0.9274 | 0.9182 | 0.9403 |
ConvBLS-DTI | 0.9509 | 0.9531 | 0.9186 | 0.9693 | 0.9688 | 0.9313 | 0.9545 | 0.95099 | 0.9153 | |
IC | BLS-DTI | 0.8830 | 0.8723 | 0.6141 | 0.9572 | 0.9547 | 0.7505 | 0.6981 | 0.6461 | 0.9160 |
ConvBLS-DTI | 0.9747 | 0.9770 | 0.9496 | 0.9719 | 0.9732 | 0.9390 | 0.9541 | 0.9654 | 0.9377 | |
GPCR | BLS-DTI | 0.9077 | 0.8950 | 0.6758 | 0.9134 | 0.8980 | 0.6570 | 0.7262 | 0.6837 | 0.8750 |
ConvBLS-DTI | 0.9557 | 0.9632 | 0.9317 | 0.9242 | 0.9460 | 0.9277 | 0.8926 | 0.9170 | 0.9000 | |
E | BLS-DTI | 0.8553 | 0.8182 | 0.7191 | 0.8689 | 0.8527 | 0.6807 | 0.8874 | 0.8675 | 0.6444 |
ConvBLS-DTI | 0.9614 | 0.9677 | 0.9314 | 0.9588 | 0.9691 | 0.9413 | 0.9643 | 0.9681 | 0.9330 | |
The green part is the best performance in comparison models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.842 | 0.701 | 0.817 | 0.917 | 0.866 | 0.856 | 0.937 |
IC | 0.904 | 0.842 | 0.929 | 0.897 | 0.802 | 0.861 | 0.968 | |
GPCR | 0.831 | 0.752 | 0.834 | 0.955 | 0.827 | 0.886 | 0.973 | |
E | 0.857 | 0.769 | 0.86 | 0.846 | 0.841 | 0.891 | 0.958 | |
Luo | 0.92 | 0.881 | 0.902 | 0.861 | 0.859 | 0.901 | 0.979 | |
CVt | NR | 0.813 | 0.756 | 0.82 | 0.654 | 0.814 | 0.833 | 0.865 |
IC | 0.938 | 0.879 | 0.949 | 0.908 | 0.938 | 0.802 | 0.947 | |
GPCR | 0.958 | 0.907 | 0.956 | 0.866 | 0.896 | 0.897 | 0.951 | |
E | 0.943 | 0.841 | 0.927 | 0.853 | 0.939 | 0.862 | 0.960 | |
Luo | 0.835 | 0.838 | 0.851 | 0.911 | 0.952 | 0.753 | 0.969 | |
Indicated in blue is the best result in each category compared with other models. |
Setting | Dataset | NRLMF | DTINet | WKNNIR | DTi2Vec | ADA-GRMFC | BLS-DTI | ConvBLS-DTI |
CVd | NR | 0.532 | 0.346 | 0.571 | 0.912 | 0.607 | 0.857 | 0.917 |
IC | 0.514 | 0.47 | 0.529 | 0.911 | 0.39 | 0.882 | 0.968 | |
GPCR | 0.486 | 0.373 | 0.502 | 0.953 | 0.384 | 0.885 | 0.972 | |
E | 0.371 | 0.215 | 0.423 | 0.863 | 0.426 | 0.834 | 0.958 | |
Luo | 0.476 | 0.299 | 0.492 | 0.945 | 0.721 | 0.906 | 0.972 | |
CVt | NR | 0.522 | 0.435 | 0.63 | 0.639 | 0.466 | 0.829 | 0.846 |
IC | 0.735 | 0.526 | 0.781 | 0.917 | 0.824 | 0.842 | 0.950 | |
GPCR | 0.803 | 0.574 | 0.858 | 0.875 | 0.631 | 0.906 | 0.946 | |
E | 0.724 | 0.379 | 0.719 | 0.876 | 0.825 | 0.902 | 0.952 | |
Luo | 0.303 | 0.138 | 0.571 | 0.899 | 0.878 | 0.804 | 0.954 | |
Indicated in blue is the best result for each category comparing all other models. |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9785 | 0.9691 | 0.9590 |
ConvBLS-DTI (without WKNKN) | 0.9758 | 0.9613 | 0.9516 |
Model | CVd | CVt | CVdt |
ConvBLS-DTI | 0.9718 | 0.9535 | 0.9522 |
ConvBLS-DTI (without WKNKN) | 0.9636 | 0.9463 | 0.9478 |