Loading [MathJax]/jax/element/mml/optable/GeneralPunctuation.js
Research article Special Issues

An adaptive feature selection algorithm based on MDS with uncorrelated constraints for tumor gene data classification


  • The developing of DNA microarray technology has made it possible to study the cancer in view of the genes. Since the correlation between the genes is unconsidered, current unsupervised feature selection models may select lots of the redundant genes during the feature selecting due to the over focusing on genes with similar attribute. which may deteriorate the clustering performance of the model. To tackle this problem, we propose an adaptive feature selection model here in which reconstructed coefficient matrix with additional constraint is introduced to transform original data of high dimensional space into a low-dimensional space meanwhile to prevent over focusing on genes with similar attribute. Moreover, Alternative Optimization (AO) is also proposed to handle the nonconvex optimization induced by solving the proposed model. The experimental results on four different cancer datasets show that the proposed model is superior to existing models in the aspects such as clustering accuracy and sparsity of selected genes.

    Citation: Wenkui Zheng, Guangyao Zhang, Chunling Fu, Bo Jin. An adaptive feature selection algorithm based on MDS with uncorrelated constraints for tumor gene data classification[J]. Mathematical Biosciences and Engineering, 2023, 20(4): 6652-6665. doi: 10.3934/mbe.2023286

    Related Papers:

    [1] Xiwen Qin, Shuang Zhang, Dongmei Yin, Dongxue Chen, Xiaogang Dong . Two-stage feature selection for classification of gene expression data based on an improved Salp Swarm Algorithm. Mathematical Biosciences and Engineering, 2022, 19(12): 13747-13781. doi: 10.3934/mbe.2022641
    [2] Kunpeng Li, Zepeng Wang, Yu Zhou, Sihai Li . Lung adenocarcinoma identification based on hybrid feature selections and attentional convolutional neural networks. Mathematical Biosciences and Engineering, 2024, 21(2): 2991-3015. doi: 10.3934/mbe.2024133
    [3] Ji-Ming Wu, Wang-Ren Qiu, Zi Liu, Zhao-Chun Xu, Shou-Hua Zhang . Integrative approach for classifying male tumors based on DNA methylation 450K data. Mathematical Biosciences and Engineering, 2023, 20(11): 19133-19151. doi: 10.3934/mbe.2023845
    [4] Depei Wang, Lianglun Cheng, Tao Wang . Fairness-aware genetic-algorithm-based few-shot classification. Mathematical Biosciences and Engineering, 2023, 20(2): 3624-3637. doi: 10.3934/mbe.2023169
    [5] Kun Yu, Mingxu Huang, Shuaizheng Chen, Chaolu Feng, Wei Li . GSEnet: feature extraction of gene expression data and its application to Leukemia classification. Mathematical Biosciences and Engineering, 2022, 19(5): 4881-4891. doi: 10.3934/mbe.2022228
    [6] Chen Ma, Zhihao Yao, Qinran Zhang, Xiufen Zou . Quantitative integration of radiomic and genomic data improves survival prediction of low-grade glioma patients. Mathematical Biosciences and Engineering, 2021, 18(1): 727-744. doi: 10.3934/mbe.2021039
    [7] Xu Zhang, Dongdong Chen, Wenmin Yang, JianhongWu . Identifying candidate diagnostic markers for tuberculosis: A critical role of co-expression and pathway analysis. Mathematical Biosciences and Engineering, 2019, 16(2): 541-552. doi: 10.3934/mbe.2019026
    [8] Huiqing Wang, Xiao Han, Jianxue Ren, Hao Cheng, Haolin Li, Ying Li, Xue Li . A prognostic prediction model for ovarian cancer using a cross-modal view correlation discovery network. Mathematical Biosciences and Engineering, 2024, 21(1): 736-764. doi: 10.3934/mbe.2024031
    [9] Huili Yang, Wangren Qiu, Zi Liu . Anoikis-related mRNA-lncRNA and DNA methylation profiles for overall survival prediction in breast cancer patients. Mathematical Biosciences and Engineering, 2024, 21(1): 1590-1609. doi: 10.3934/mbe.2024069
    [10] Chaofan Song, Tongqiang Liu, Huan Wang, Haifeng Shi, Zhuqing Jiao . Multi-modal feature selection with self-expression topological manifold for end-stage renal disease associated with mild cognitive impairment. Mathematical Biosciences and Engineering, 2023, 20(8): 14827-14845. doi: 10.3934/mbe.2023664
  • The developing of DNA microarray technology has made it possible to study the cancer in view of the genes. Since the correlation between the genes is unconsidered, current unsupervised feature selection models may select lots of the redundant genes during the feature selecting due to the over focusing on genes with similar attribute. which may deteriorate the clustering performance of the model. To tackle this problem, we propose an adaptive feature selection model here in which reconstructed coefficient matrix with additional constraint is introduced to transform original data of high dimensional space into a low-dimensional space meanwhile to prevent over focusing on genes with similar attribute. Moreover, Alternative Optimization (AO) is also proposed to handle the nonconvex optimization induced by solving the proposed model. The experimental results on four different cancer datasets show that the proposed model is superior to existing models in the aspects such as clustering accuracy and sparsity of selected genes.



    Cancer is considered to be the greatest threat against human health and a significant barrier for life expectancy increasing in the world. An early diagnosis is significant for differentiation of the characteristics of the tumor and enhancing the quality of life of the patients. At present, there is much data that can be used to assist cancer early diagnosis, in which gene expression data obtained by DNA microarray technology is considered to be the most important data. However, directly using raw gene expression data to classify cancer not only induces high storage costs and huge computational burden but also leads to poor clustering or classification accuracy due to the high dimensionality and redundancy of the raw gene expression data. Therefore, extracting the key genes related to tumors from the large number of genes in the raw data, i.e., feature selection, is one of the noteworthy research areas for cancer classification.

    Using the feature selection method can extract key genes strongly relevant to the certain cancer. Key genes extracted from the raw gene data should have properties such as low redundancy, low dimensionality and high classification power. Feature selection methods can be divided into three categories: filtering method, wrapper method and embedded method. Filtering methods score genes according to rules; for example: divergence or correlation, to select genes first, and then employ the selected genes to train filter [1,2]. While easy to be fulfilled, filtering methods may lose the structural information concealed in the raw data since correlation between genes are unconsidered. Wrapper methods (e.g., RFE-SVM [3]) use the classifier to evaluate and update the selected genes iteratively. Although it can achieve higher accuracy rates comparing with filtering methods, wrapper methods suffer heavy computational burdens.

    The embedded method selects the genes by following the objective function of a certain model, thus can retain the structural information concealed in the raw gene data. In addition, it's computational complexity is lower than that of the wrapper method. At present, the embedded methods can be divided into two categories: supervised feature selection (SFS) and unsupervised feature selection (UFS), according to whether labels are required. The SFS methods (e.g., Max-Relevance and Min-Redundancy (mRMR) [4], Regularized logistic regression (RLR) [5], etc.) employ labels to select genes with discrimination capability. Obviously, SFS methods is unsuitable in the scene which data label is unavailable. Although labelling data can handle this problem, it will consume a lot of time and labors. To extract key genes from the unlabeled data, the UFS models including Spectral Feature Selection (SPEC), Unsupervised Discriminative Feature Selection (UDFS), non-negative discriminant feature selection algorithm (NDFS) and Joint Embedding Learning and Sparse Regression (JELSR) [6,7,8,9] are proposed. Although the above methods can reveal the local flow structure of the gene data, clustering performance of them is unsatisfied since overall structural information concealed in gene data is unconsidered by them [10]. To handle this problem, Jin et al. proposed the adaptive feature selection method (MDS-AUFS) [11], which first employs multidimensional deflation to transform high dimensional original data into a low dimensional space while preserving the global structure of the original gene data. Then, the sparse regression item and probabilistic neighborhood graph are introduced in the proposed model to sparse select genes and to preserve the local structure of the low dimensional data, respectively. However, since Jin's method tends to extract the highly correlated genes, the genes extracted by the method loses the diversity of the genes, which may deteriorate to the clustering ability of the model [12]. Although generalized irrelevance constraints is introduced by Li to preserve the diversity of the selected genes [13], how to preserve the global structure of the original data is still unconsidered in this work. Namely, Li's method is also faced the problem that the inherent structure of the original data is changed.

    To address the above problem, inspired by works of Jin and Li, we propose a new unsupervised feature selection model here which is called adaptive unsupervised feature selection for multidimensional deformation based on uncorrelated constraints (UCMAFS). Specifically, the objective function of the proposed model consists of three items. The first item employs reconstructed coefficient matrix to map high-dimensional original genetic data into a low-dimensional space which is obtained by using multidimensional deflation method (MDS). The second item encourages reconstructed coefficient matrix to be row sparsity, which may be suitable for selecting sparse features related to the tumor. The third item employs probability neighborhood matrix to encourage reconstructed coefficient matrix preserving local flow structure of the original data. Moreover, an uncorrelated constraint is imposed on the reconstructed coefficient matrix to preserve the diversity of the selected genes. The contributions of this paper are summarized as follows.

    1) A new unsupervised feature selection model is proposed by integrating structure learning and sparse feature selection together.

    2) The reconstructed coefficient matrix and its 2,1-norm constraint are introduced in proposed model to map high dimension original data into a low dimensional space meanwhile to guarantee sparsity of the selected features.

    3) Global and local structures of original data are preserved by using MDS method and probabilistic neighborhood matrices respectively in the proposed model.

    4) An uncorrelated constraint is imposed on the reconstructed coefficient matrix of the proposed model to enhance the diversity of the selected genes.

    The subsequent part of the paper is organized as follows, in Section 2, we present the UCMAUFS model in detail and give its optimization algorithm. Comparative experiment with some benchmark models on real cancer datasets are designed to demonstrate the effectiveness of the UCMAUFS in the Section 3. Conclusions are given in Section 4.

    In this paper, we use X=[x1,x2,...xn]Rd×n to represent the gene expression data, where xiRd denotes the i-th sample and n is the number of samples. InRn×n is an identity matrix and 1n is an n-dimensional vector with all elements of 1. The trace of a matrix A=(ai,j)Rn×n is written as Tr(A), and ||A||2,1 represents the 2,1-norm of the matrix A:

    ||A||2,1=ni=1(nj=1a2i,j)2 (1)

    Although UFS algorithm can select genes by revealing the inherent structural information of unlabelled data, it ignores the correlation information concealed in the genes, so highly correlated genes may coexist in the selected genes, which reduces the diversity of selected genes, thus deteriorate the clustering performance of the model. To preserve the inherent structure of the data and the diversity of the genes simultaneously, we propose the UCMAUFS model here, whose objective function is shown below.

    minW,P{||WTXY||2F+α||W||2,1+βni,j(||WTxiWTxj||2pi,j+λp2i,j)}s.t.WTUpW=I,0pi,j1,nj=1pi,j=1 (2)

    where α and β are regularization parameters to balance the fitting accuracy of adaptive structure learning and the sparsity of the feature selection coefficient matrix. WRd×p is the reconstructed coefficient matrix which transforms high dimensional original data X to the low-dimensional representation Y, where Y is obtained by the MDS algorithm, thus the first item of the model can preserve the global structure of the feature space by employing the spatial structure invariant property of the MDS algorithm, while the 2,1-norm term in the second item forces row of the W matrix to be sparse, giving the W matrix feature selection capability. And pi,j is a element of the probabilistic neighborhood matrix PRn×n, whose values is the probabilitie that xi and xj are neighbors of each other, which can be obtained by solving for the following equation:

    minW,Pni,j{(||WTxiWTxj||2pi,j+λp2i,j)}s.t.0pi,j1,nj=1pi,j=1 (3)

    where λ is a regularization parameter, and the regularization term λp2i,j can effectively avoid the trivial solution.

    In addition, the probabilistic neighborhood matrix is utilized to preserve the local flow structure concealed in the original data, therefore, a new uncorrelated constraint is proposed here according to the original uncorrelated constraint [13]:

    WT(X(In+βLp)XT+αM)W=WTUpW=I (4)

    where Lp=Dp(P+PT)/2, is the Laplace matrix constructed by the probabilistic neighborhood matrix P, Dp is a diagonal matrix with the i-th diagonal element which can be described to nj=1(pi,j+pj,i)/2. The diagonal matrix MRd×d with i-th diagonal element described in the (5) is proposed to relax the uncorrelated constraint with respect to W:

    mi,i=12||wi||22+ε (5)

    where wi is the i-th row of W. We can find that the (4) allows the model to effectively reduce the redundancy gene selected, thus to find more discriminative genes, leads the classification ability of the model to improving.

    Since the non-convex optimization problem (2) is to minimize an objective function which is composed of two regularization terms subject to three constraints, it is hard to solve directly. Inspired by the optimization methods which are described in the literature [14,15], we handle the (2) by using the Alternative Optimization (AO), i.e., by given one optimized matrix to update the other, thus an original coupled matrix optimization problem can be transformed into multiple sub-problems to handle iteratively. Details of the UCMAUFS model are written in the Algorithm 1 described as follow.

    Algorithm 1 UCMAUFS
    Input: gene expression data matrix X, regularization parameter α and β, the clustering number q, number of selected genes s.
    Output: The top s ranked genes.
    Initialization:
    1: Identity matrix to be set as the initialization matrix of the W and M; λ is initialized to be 1;
    2: A low-dimensional representation matrix Y of X are obtained by using the MDS algorithm;
    Repeat:
    1: Given variables W update P and λ (see algorithm2);
    2: Calculate Lp=Dp(P+PT)/2;
    3: Given variables P update W and M (see algorithm3);
    Until Convergence
    Sort all genes based on ||wi||2 in descending order. The top s ranked genes are selected.

    Given the W, the optimization problem (2) can be described as follow.

    min0pi,j1nj=1pi,j=1i,j{||WTxiWTxj||22pi,j+λp2i,j} (6)

    Let bi,j=12λ||WTxiWTxj||2, Define the square matrix B=[bT1,bT2,...,bTn]TRn×n, then (6) can be expressed as:

    min0pi,j11npi=112 (7)

    The Lagrange function of problem (7) is described as:

    \begin{equation} L({\mathit{\boldsymbol{p}}_i}, \mu , {\mathit{\boldsymbol{v}}_i}) = \frac{1}{2}||{\mathit{\boldsymbol{p}}_i} + {\mathit{\boldsymbol{b}}_i}|{|^2} - \mu (\mathit{\boldsymbol{1}}_n{\mathit{\boldsymbol{p}}_i} - 1) - \mathit{\boldsymbol{v}}_i^T{\mathit{\boldsymbol{p}}_i} \end{equation} (8)

    where \mu and \mathit{\boldsymbol{v}}_i are Lagrangian multipliers. According to the KTT condition, the optimal solution of problem (8) is

    \begin{equation} {p_{i, j}} = \max ( - {b_{i, j}} + \mu , 0) \end{equation} (9)

    Sort the elements of each row of matrix \mathit{\boldsymbol{B}} in descending order to form matrix {\mathit{\boldsymbol{B}}^ * } \in {R^{n \times n}} . The value of k can be determined by the following inequality:

    \begin{equation} \left\{ \begin{array}{l} B_{i, \bar k}^* + \mu > 0{\rm{\qquad for }}{\kern 1pt}{\kern 1pt}{\kern 1pt}{\kern 1pt}\bar k{\rm{{\kern 1pt}{\kern 1pt} = {\kern 1pt}{\kern 1pt}1, {\kern 1pt}{\kern 1pt}}}...{\rm{, {\kern 1pt}{\kern 1pt}}}k\\ B_{i, \bar k}^* + \mu \le 0{\rm{\qquad for }}{\kern 1pt}{\kern 1pt}{\kern 1pt}{\kern 1pt}\bar k{\rm{{\kern 1pt}{\kern 1pt} = {\kern 1pt}{\kern 1pt}}}k{\rm{{\kern 1pt}{\kern 1pt}+{\kern 1pt}{\kern 1pt}1, {\kern 1pt}{\kern 1pt}}}...{\rm{, {\kern 1pt}{\kern 1pt}}}n \end{array} \right. \end{equation} (10)

    Since the \mathit{\boldsymbol{p}}_i should follow the constraint \mathit{\boldsymbol{1}}_n \mathit{\boldsymbol{p}}_i = 1 , thus \mu is further obtained by (11).

    \begin{equation} \mu = \frac{1}{k}(1 - \sum\limits_{d = 1}^k {b_{i, d}^*} ) \end{equation} (11)

    Substitute the (11) into the (9) to obtain the optimal value of \mathit{\boldsymbol{P}} :

    \begin{equation} {p_{i, j}} = \max ( - {b_{i, j}} - \frac{1}{k}(1 - \sum\limits_{d = 1}^k {b_{i, d}^*} ), 0) \end{equation} (12)

    Similar to the approach proposed in the literature [14], we set the regularization parameter \lambda according to the number of nearest neighbors k :

    \begin{equation} \lambda = \frac{1}{{2n}}\sum\limits_{i = 1}^n {(kb_{i, k + 1}^*} - \sum\limits_{j = 1} {{b_{i, j}})} \end{equation} (13)

    Algorithm 2 Given variables \mathit{\boldsymbol{W}} to update \mathit{\boldsymbol{P}}
    Input: gene expression data matrix \mathit{\boldsymbol{X}}, low-dimensional representation matrix \mathit{\boldsymbol{Y}}, Reconstructed coefficient matrix \mathit{\boldsymbol{W}}, regularization parameter \lambda.
    Repeat:
    1: Obtain the matrix \mathit{\boldsymbol{B}};
    2: Determine the value of k by using the inequality (10);
    3: Update \mathit{\boldsymbol{P}} according to equation (12);
    4: Update \lambda according to equation (13);
    Until Convergence
    Output: Probability matrix \mathit{\boldsymbol{P}} and regularization parameter \lambda.

    Given \mathit{\boldsymbol{P}} , the optimization problem is described as

    \begin{equation} \begin{aligned} &\mathop {\min }\limits_{{\mathit{\boldsymbol{W}}^T}{\mathit{\boldsymbol{U}}_P}\mathit{\boldsymbol{W}} = \mathit{\boldsymbol{I}}} \{ Tr({\mathit{\boldsymbol{W}}^T}\mathit{\boldsymbol{X}}{\mathit{\boldsymbol{X}}^T}\mathit{\boldsymbol{W}} - 2{\mathit{\boldsymbol{W}}^T}\mathit{\boldsymbol{X}}\mathit{\boldsymbol{Y}}) + Tr({\mathit{\boldsymbol{W}}^T}(\alpha \mathit{\boldsymbol{M}})\mathit{\boldsymbol{W}}) + Tr({\mathit{\boldsymbol{W}}^T}\mathit{\boldsymbol{X}}(\beta {\mathit{\boldsymbol{L}}_p}){\mathit{\boldsymbol{X}}^T}\mathit{\boldsymbol{W}}) \end{aligned} \end{equation} (14)

    Given \mathit{\boldsymbol{W}}^T\mathit{\boldsymbol{U}}_p\mathit{\boldsymbol{W}} = \mathit{\boldsymbol{I}} and \mathit{\boldsymbol{P}} , solving problem (14) is equivalent to solving the following problem.

    \begin{equation} \begin{array}{l} \mathop {\max }\limits_{\mathit{\boldsymbol{W}}} {Tr({\mathit{\boldsymbol{W}}^T}\mathit{\boldsymbol{X}}\mathit{\boldsymbol{Y}})}\\ \begin{array}{*{20}{c}} {s.t.}&{{\mathit{\boldsymbol{W}}^T}{\mathit{\boldsymbol{U}}_P}\mathit{\boldsymbol{W}} = \mathit{\boldsymbol{I}}} \end{array} \end{array} \end{equation} (15)

    Transform the (15) to the (16)

    \begin{equation} \mathop {\max }\limits_{{\mathit{\boldsymbol{Q}}^T}\mathit{\boldsymbol{Q}} = \mathit{\boldsymbol{I}}} Tr({\mathit{\boldsymbol{Q}}^T}\mathit{\boldsymbol{C}}) \end{equation} (16)

    Where:

    \begin{equation} \begin{array}{*{20}{c}} {\mathit{\boldsymbol{Q}} = {{({\mathit{\boldsymbol{U}}_p})}^{\frac{1}{2}}}\mathit{\boldsymbol{W}}, }&{\mathit{\boldsymbol{C}} = {{({\mathit{\boldsymbol{U}}_p})}^{ - \frac{1}{2}}}\mathit{\boldsymbol{X}}\mathit{\boldsymbol{Y}}} \end{array} \end{equation} (17)

    By solving (16) and (17), we can find that \mathit{\boldsymbol{W}} can be described as

    \begin{equation} {\mathit{\boldsymbol{W}} = {{({\mathit{\boldsymbol{U}}_p})}^{ - \frac{1}{2}}}\mathit{\boldsymbol{Q}}} \end{equation} (18)

    In summary, Optimization problem (14) can be solved efficiently [16]. The details are described in Algorithm 3.

    Algorithm 3 Given variables \mathit{\boldsymbol{P}} to update \mathit{\boldsymbol{W}}
    Input: gene expression data matrix \mathit{\boldsymbol{X}}, low-dimensional representation matrix \mathit{\boldsymbol{Y}}, regularization parameter \alpha and \beta, Diagonal matrix \mathit{\boldsymbol{M}}; Laplace matrix \mathit{\boldsymbol{L}}_p.
    Repeat:
    1: Obtain {\mathit{\boldsymbol{U}}_p} = \mathit{\boldsymbol{X}}({\mathit{\boldsymbol{I}}_n} + \beta {\mathit{\boldsymbol{L}}_p}){\mathit{\boldsymbol{X}}^T} + \alpha \mathit{\boldsymbol{M}};
    2: Obtain the matrix \mathit{\boldsymbol{C}} according to equation (17);
    3: Obtain \mathit{\boldsymbol{U}} and \mathit{\boldsymbol{V}} by performing a tight singular value decomposition of \mathit{\boldsymbol{C}};
    4: Obtain \mathit{\boldsymbol{Q}}=\mathit{\boldsymbol{U}}\mathit{\boldsymbol{V}}^T;
    5: Update \mathit{\boldsymbol{W}}=(\mathit{\boldsymbol{U}}_p)^{- \frac{1}{2}} \mathit{\boldsymbol{Q}};
    6: Update \mathit{\boldsymbol{M}};
    Until Convergence
    Output: Reconstructed coefficient matrix \mathit{\boldsymbol{W}} and Diagonal matrix \mathit{\boldsymbol{M}}.

    Comparative experiment on the four publicly available cancer gene datasets, including a lung cancer dataset, a colon cancer dataset, a lymphoma dataset and a glioma dataset are employed to evaluate the performance of the UCMSUFS. All dataset were downloaded from https://jundongl.github.io/scikit-feature/datasets.html (accessed on 1 May 2021), and details of the datasets are shown in Table 1.

    Table 1.  Details of the datasets.
    Dataset No. of samples No. of genes No. of Classes
    Lung Cancer 203 3312 5
    Colon Cancer 62 2000 2
    Lymphoma 96 4026 9
    Glioma 50 4434 4

     | Show Table
    DownLoad: CSV

    We use the K-means clustering algorithm [17] to evaluate the clustering accuracy (ACC) of the new unsupervised feature selection model (2) proposed in this paper. Specifically, the key genes are extracted by using UCMAUFS model and other comparison models, respectively, and then k-means algorithm are employed to cluster those extracted genes. The accuracy of the clustering is obtained by using (19) finally.

    \begin{equation} ACC = 1/n\sum\nolimits_{i = 1}^n {\delta (map({c_i}), {C_i})} \end{equation} (19)

    Where c_i is the cluster label of the i -th sample \mathit{\boldsymbol{x}}_i and C_i is the actual label of \mathit{\boldsymbol{x}}_i . \delta (.) is the \delta -function and function map(.) is the optimal mapping function [18], which projects each cluster label into the actual label by using the Kuhn-Munkres algorithm. Obviously, larger ACC values indicate higher clustering performance.

    Comparative experiment with some classical feature selection models including MDS-AUFS [11], URAFS [13], UDFS [7], NDFS [8] and JELSR [9] are employed to evaluate the performance of proposed UCMSUFS model. To smooth ACC fluctuating induced by initial points choosing in the clustering algorithm, we repeat the clustering algorithm 20 times in each experiment and take the average of ACC. In different dataset, the average of ACC versus number of genes selected of different models are plotted in the Figure 1.

    Figure 1.  ACC performance of the six models on the four gene expression datasets. (a) Colon Cancer; (b) Glioma; (c) Lung Cancer; (d) Lymphoma.

    It can be seen that the beginning, the clustering accuracy of all models on the four data sets almost trend to increasing with the number of selected genes when the number of selected genes is small. However, clustering accuracy of all models on the four data sets show decreasing with the number of selected genes after the clustering accuracy achieves its maximum. It can be inferred that all models perform well due to the reasonable number of selected genes, while further increases in the selected genes may induce redundant genes into models which may weaken the performance of the models. Moreover, we can find that proposed model shows the almost best clustering performance on all datasets with no more than 40 genes selected. For example, UCMAUFS achieves a maximum clustering accuracy of 83.37% on the colon cancer dataset, much higher than that of MDS-AUFS with 80.65% clustering accuracy and so on. Similarly, on the lymphoma dataset in Figure 1(d), UCMAUFS also shows the best clustering accuracy in all models. Although UCMAUFS, MDS-AUFS and JELSR show the similar clustering accuracy on the glioma dataset, UCMAUFS achieves the highest clustering accuracy with only 20 genes selected (Figure 1(b)) which indicates that UCMAUFS has the best ability of sparse feature selecting. Noteworthy, in Figure 1(c), the MDS-AUFS model shows the consistent increases of clustering accuracy with number of selected genes increasing, even close to the maximum clustering accuracy of UCMAUFS at the 50 genes selected.

    Average and maximum clustering accuracy of each model are respectively described in Figure 2 and Table 2 to clearly exhibit the performance of the proposed model. In addition, the best and second best results are written in bold and underlined text respectively in the Table 2. We can find that proposed UCMAUFS model exhibits apparently best performance in all models.

    Figure 2.  The average ACC of all models on four gene expression datasets.
    Table 2.  Maximum clusering accuracy of all modles on four gene expression datasets.
    URAFS UDFS NDFS JELSR MDS-AUFS UCMAUFS
    Colon Cancer 0.7097 0.7581 0.6774 0.7581 0.8065 0.8387
    Glioma 0.5400 0.6200 0.5600 0.6400 0.6400 0.6400
    Lung Cancer 0.5172 0.7044 0.5369 0.5665 0.7192 0.7438
    Lymphoma 0.4792 0.4375 0.5833 0.6146 0.6563 0.6979

     | Show Table
    DownLoad: CSV

    Computational complexity of UCMAUFS and others comparison models are analyzed and written in Table 3. According to Algorithm 1, the computational complexity of UCMAUFS mainly determined by the reconstructed coefficient matrix \mathit{\boldsymbol{W}} and the probability neighborhood matrix \mathit{\boldsymbol{P}} updating. Since the computation complexity of updating coefficient matrix \mathit{\boldsymbol{W}} and the probability neighborhood matrix \mathit{\boldsymbol{P}} are described as O(n^3) and O(d^3) , respectively. Thus, we can conclude that the computational complexity of UCMAUFS is O(n^3+d^3) .

    Table 3.  Comparison of the computational complexity.
    Method Computational Complexity
    URAFS O(d^3+n^2d+n^2)
    UDFS O(d^3+n^2c)
    NDFS O(d^3+n^2c)
    JELSR O(d^3+n^2q)
    MDS-AUFS O(min{(n, d)}^3+n^2q)
    UCMAUFS O(n^3+d^3)

     | Show Table
    DownLoad: CSV

    Convergences versus number of iterations of UCMAFUS model on four datasets are plotted in the Figure 3. It is clearly that UCMAFUS converges in all gene datasets quickly, almost achieving steady-state within 15 iterations.

    Figure 3.  The convergence of UCMAUFS on four gene expression datasets.

    The heatmap is employed to further demonstrate the clustering performance of the UCMAUFS model. Specifically, the heatmap of the top ten genes selected by using UCMAUFS model on the four gene datasets are shown in Figure 4, where redder color block indicates higher expression level of the corresponding gene while bluer color block indicates lower expression level of the corresponding gene. It can be seen that in the binary dataset Colon, the selected genes of different categories have different gene expression profiles. In other multi-categorical datasets, using a single gene may hard to distinguish different category, however, we can find that, overall, the gene expression profiles of different categories appear apparent distinguish.

    Figure 4.  Heatmap of the 10 top ranked genes of 4 gene expression data. (a) Colon Cancer; (b) Glioma; (c) Lung Cancer; (d) Lymphoma.

    To deeply discuss the clustering performance of the proposed model, the T-score and its P-value of the selected top ten genes in the Colon Cancer dataset and the F-score and its P-value of the selected top ten genes in the others datasets are presented in Table 4. It is clearly that seven of the top ten genes in the Colon Cancer dataset have T-score values greater than 2, meanwhile their P-value values are less than 0.05. While the top ten genes in the other three datasets all have large F-scores and its P-values are less than 0.05. This indicates that most of the selected genes of proposed model have good clustering performance.

    Table 4.  T-score or F-score of the top ten genes in the four datasets.
    Colon Cancer Glioma Lung Cancer Lymphoma
    T-score P-value F-score P-value F-score P-value F-score P-value
    2.637 1.06 \times 10^{-2} 4.114 1.10 \times 10^{-2} 18.47 6.49 \times 10^{-13} 6.687 8.55 \times 10^{-7}
    2.905 5.16 \times 10^{-3} 5.308 3.00 \times 10^{-3} 29.86 1.94 \times 10^{-19} 2.750 9.00 \times 10^{-3}
    2.475 1.62 \times 10^{-2} 8.718 1.09 \times 10^{-4} 24.92 1.01 \times 10^{-16} 6.732 7.73 \times 10^{-7}
    4.370 7.39 \times 10^{-4} 9.323 6.30 \times 10^{-5} 56.37 1.11 \times 10^{-31} 4.695 8.40 \times 10^{-5}
    2.366 2.13 \times 10^{-2} 5.284 3.00 \times 10^{-3} 10.55 9.02 \times 10^{-8} 3.145 4.00 \times 10^{-3}
    1.602 1.14 \times 10^{-1} 9.042 8.10 \times 10^{-5} 17.64 2.14 \times 10^{-12} 4.285 2.23 \times 10^{-4}
    4.556 6.50 \times 10^{-7} 3.674 1.90 \times 10^{-2} 6.649 4.90 \times 10^{-5} 3.562 1.00 \times 10^{-3}
    1.917 6.00 \times 10^{-2} 4.928 5.00 \times 10^{-3} 23.09 1.13 \times 10^{-15} 9.744 1.46 \times 10^{-9}
    1.484 1.43 \times 10^{-1} 3.093 3.60 \times 10^{-2} 3.761 6.00 \times 10^{-3} 3.815 6.98 \times 10^{-4}
    3.886 8.70 \times 10^{-5} 13.46 2.00 \times 10^{-6} 10.45 1.07 \times 10^{-7} 2.757 9.00 \times 10^{-3}

     | Show Table
    DownLoad: CSV

    In this paper, a new sparse embedded model is proposed to sparse selects genes with high relevant to cancer meanwhile to preserve the diversity of selected genes. By handling this new model, we propose an AO algorithm called UCMAUFS here. The comparative experiments with others models on four different cancer datasets shows that UCMAUFS can achieve the maximum clustering accuracy with the least selected genes which indicates UCMAUFS can obtain higher clustering accuracy by filtering out redundant genes.

    The authors declare there is no conflict of interest.



    [1] S. M. Kopka, A. D. Long, E. T. Ito, L. Tolleri, M. M. Riehle, E. S. Paegle, et al., Global gene expression profiling in Escherichia coli K12: The effects of integration host factor, J. Biol. Chem., 275 (2000), 29672–29684. https://doi.org/10.1074/jbc.M213060200 doi: 10.1074/jbc.M213060200
    [2] M. Berta, J. M. Renes, M. M. Wilde, Identifying the information gain of a quantum measurement, IEEE Trans. Inform. Theory, 60 (2014), 7987–8006. https://doi.org/10.1109/TIT.2014.2365207 doi: 10.1109/TIT.2014.2365207
    [3] T. R. Golub, D. K. Slonim, P. Tamayo, C. Huard, M. Gaasenbeek, J. P. Mesiroy, et al., Molecular classification of cancer: class discovery and class prediction by gene expression monitoring, Science, 286 (1999), 531–537. https://doi.org/10.1126/science.286.5439.531 doi: 10.1126/science.286.5439.531
    [4] H. Peng, F. Long, C. Ding, Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy, IEEE Trans. Pattern Anal. Mach. Intell, , 27 (2005), 1226–1238. https://doi.org/10.1109/TPAMI.2005.159 doi: 10.1109/TPAMI.2005.159
    [5] L. Y. Li, Z. P. Liu, Biomarker discovery for predicting spontaneous preterm birth from gene expression data by regularized logistic regression, Comput. Struct. Biotechnol. J., 18 (2020), 3434–3446. https://doi.org/10.1016/j.csbj.2020.10.028 doi: 10.1016/j.csbj.2020.10.028
    [6] Z. Zhao, H. Liu, Spectral feature selection for supervised and un-supervised Learning, in Proceedings of the 24th international conference on Machine learning, 227 (2007), 1151–1157. https://doi.org/10.1145/1273496.1273641
    [7] Y. Yang, H. T. Shen, Z. Ma, Z. Huang, X. Zhou, L_2, 1-Norm regularized discrimiNative feature selection for unsupervised learning, in Proceedings of the 22nd International joint Conference on Artificial Intelligence, (2011), 1589–1594. https://doi.org/10.5591/978-1-57735-516-8/IJCAI11-267
    [8] Z. Li, Y. Yang, J. Liu, X. Zhou, H. Lu, Unsupervised feature selection using nonnegative spectral analysis, in Proceedings of the Twenty-Sixth AAAI Conference on Artificial Interlligence, 26 (2012), 1026–1032. https://doi.org/10.1609/aaai.v26i1.8289
    [9] C. P. Hou, F. P. Nie, D. Y. Yi, Y. Wu, Feature selection via joint embedding Learning and sparse regression, in Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, (2011), 1324–1229. https://doi.org/10.5591/978-1-57735-516-8/IJCAI11-224
    [10] L. Du, Y. D. Shen, Unsupervised feature selection with adaptive structure learning, in Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, (2015), 209–218. https://doi.org/10.1145/2783258.2783345
    [11] B. Jin, C. L. Fu, Y. Jin, W. Yang, S. B. Li, G. Y. Zhang, et al., An adaptive unsupervised feature selection algorithm based on MDS for tumor gene data classification, Sensors, 21 (2021), 3627. https://doi.org/10.3390/s21113627 doi: 10.3390/s21113627
    [12] X. Y. Xu, X. Wu, F. L. Wei, W. Zhong, F. P. Nie, A general framework for feature selection under orthogonal regression with global redundancy minimization, IEEE Trans. Knowl. Data Eng., 34 (2021), 5056–5069. https://doi.org/10.1109/TKDE.2021.3059523 doi: 10.1109/TKDE.2021.3059523
    [13] L. X. Li, H. Zhang, R. Zhang, Y. Liu, Generalized uncorrelated regression with adaptive graph for unsupervised feature selection, IEEE Trans. Neural Networks Learn. Syst., 30 (2019), 1587–1595. https://doi.org/10.1109/TNNLS.2018.2868847 doi: 10.1109/TNNLS.2018.2868847
    [14] M. Yang, L. Zhang, X. C. Feng, D. Zhang, Sparse representation based fisher discrimination dictionary learning for image classification, International Journal of Computer Vision, 109 (2014), 209–232. https://doi.org/10.1007/s11263-014-0722-8 doi: 10.1007/s11263-014-0722-8
    [15] S. L. Peng, Y. Yang, W. Liu, F. Li, X. K. Liao, Discriminant projection shared dictionary learning for classification of tumors using gene expression data, IEEE/ACM Trans. Comput. Biol. Bioinforma., 18 (2021), 1464–1473. https://doi.org/10.1109/TCBB.2019.2950209 doi: 10.1109/TCBB.2019.2950209
    [16] J. Huang, F. P. Nie, H. Huang, C. Ding, Robust manifold nonnegative matrix factorization, ACM Trans. Knowl. Discovery Data, 8 (2014), 1–21. https://doi.org/10.1145/2601434 doi: 10.1145/2601434
    [17] R. Zhang, X. L. Li, Unsupervised feature selection via data reconstruction and side information, IEEE Trans. Image Process., 29 (2020), 8097–8106. https://doi.org/10.1109/TIP.2020.3011253 doi: 10.1109/TIP.2020.3011253
    [18] A. Strehl, J. Ghosh, Cluster ensembles-A knowledge reuse framework for combining multiple partitions, J. Mach. Learn. Res., 3 (2020), 583–617. https://doi.org/10.1162/153244303321897735 doi: 10.1162/153244303321897735
  • Reader Comments
  • © 2023 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(2347) PDF downloads(71) Cited by(0)

Figures and Tables

Figures(4)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog