Research article Special Issues

FCM-DNN: diagnosing coronary artery disease by deep accuracy fuzzy C-means clustering model


  • Cardiovascular disease is one of the most challenging diseases in middle-aged and older people, which causes high mortality. Coronary artery disease (CAD) is known as a common cardiovascular disease. A standard clinical tool for diagnosing CAD is angiography. The main challenges are dangerous side effects and high angiography costs. Today, the development of artificial intelligence-based methods is a valuable achievement for diagnosing disease. Hence, in this paper, artificial intelligence methods such as neural network (NN), deep neural network (DNN), and fuzzy C-means clustering combined with deep neural network (FCM-DNN) are developed for diagnosing CAD on a cardiac magnetic resonance imaging (CMRI) dataset. The original dataset is used in two different approaches. First, the labeled dataset is applied to the NN and DNN to create the NN and DNN models. Second, the labels are removed, and the unlabeled dataset is clustered via the FCM method, and then, the clustered dataset is fed to the DNN to create the FCM-DNN model. By utilizing the second clustering and modeling, the training process is improved, and consequently, the accuracy is increased. As a result, the proposed FCM-DNN model achieves the best performance with a 99.91% accuracy specifying 10 clusters, i.e., 5 clusters for healthy subjects and 5 clusters for sick subjects, through the 10-fold cross-validation technique compared to the NN and DNN models reaching the accuracies of 92.18% and 99.63%, respectively. To the best of our knowledge, no study has been conducted for CAD diagnosis on the CMRI dataset using artificial intelligence methods. The results confirm that the proposed FCM-DNN model can be helpful for scientific and research centers.

    Citation: Javad Hassannataj Joloudari, Hamid Saadatfar, Mohammad GhasemiGol, Roohallah Alizadehsani, Zahra Alizadeh Sani, Fereshteh Hasanzadeh, Edris Hassannataj, Danial Sharifrazi, Zulkefli Mansor. FCM-DNN: diagnosing coronary artery disease by deep accuracy fuzzy C-means clustering model[J]. Mathematical Biosciences and Engineering, 2022, 19(4): 3609-3635. doi: 10.3934/mbe.2022167

    Related Papers:

    [1] Biao Cai, Qing Xu, Cheng Yang, Yi Lu, Cheng Ge, Zhichao Wang, Kai Liu, Xubin Qiu, Shan Chang . Spine MRI image segmentation method based on ASPP and U-Net network. Mathematical Biosciences and Engineering, 2023, 20(9): 15999-16014. doi: 10.3934/mbe.2023713
    [2] Ke Bi, Yue Tan, Ke Cheng, Qingfang Chen, Yuanquan Wang . Sequential shape similarity for active contour based left ventricle segmentation in cardiac cine MR image. Mathematical Biosciences and Engineering, 2022, 19(2): 1591-1608. doi: 10.3934/mbe.2022074
    [3] Jianhua Song, Lei Yuan . Brain tissue segmentation via non-local fuzzy c-means clustering combined with Markov random field. Mathematical Biosciences and Engineering, 2022, 19(2): 1891-1908. doi: 10.3934/mbe.2022089
    [4] Zijian Wang, Yaqin Zhu, Haibo Shi, Yanting Zhang, Cairong Yan . A 3D multiscale view convolutional neural network with attention for mental disease diagnosis on MRI images. Mathematical Biosciences and Engineering, 2021, 18(5): 6978-6994. doi: 10.3934/mbe.2021347
    [5] Qing Zou, Zachary Miller, Sanja Dzelebdzic, Maher Abadeer, Kevin M. Johnson, Tarique Hussain . Time-Resolved 3D cardiopulmonary MRI reconstruction using spatial transformer network. Mathematical Biosciences and Engineering, 2023, 20(9): 15982-15998. doi: 10.3934/mbe.2023712
    [6] Yuqing Zhang, Yutong Han, Jianxin Zhang . MAU-Net: Mixed attention U-Net for MRI brain tumor segmentation. Mathematical Biosciences and Engineering, 2023, 20(12): 20510-20527. doi: 10.3934/mbe.2023907
    [7] Qian Wu, Yuyao Pei, Zihao Cheng, Xiaopeng Hu, Changqing Wang . SDS-Net: A lightweight 3D convolutional neural network with multi-branch attention for multimodal brain tumor accurate segmentation. Mathematical Biosciences and Engineering, 2023, 20(9): 17384-17406. doi: 10.3934/mbe.2023773
    [8] R. Bhavani, K. Vasanth . Brain image fusion-based tumour detection using grey level co-occurrence matrix Tamura feature extraction with backpropagation network classification. Mathematical Biosciences and Engineering, 2023, 20(5): 8727-8744. doi: 10.3934/mbe.2023383
    [9] Jiajun Zhu, Rui Zhang, Haifei Zhang . An MRI brain tumor segmentation method based on improved U-Net. Mathematical Biosciences and Engineering, 2024, 21(1): 778-791. doi: 10.3934/mbe.2024033
    [10] Yu Li, Meilong Zhu, Guangmin Sun, Jiayang Chen, Xiaorong Zhu, Jinkui Yang . Weakly supervised training for eye fundus lesion segmentation in patients with diabetic retinopathy. Mathematical Biosciences and Engineering, 2022, 19(5): 5293-5311. doi: 10.3934/mbe.2022248
  • Cardiovascular disease is one of the most challenging diseases in middle-aged and older people, which causes high mortality. Coronary artery disease (CAD) is known as a common cardiovascular disease. A standard clinical tool for diagnosing CAD is angiography. The main challenges are dangerous side effects and high angiography costs. Today, the development of artificial intelligence-based methods is a valuable achievement for diagnosing disease. Hence, in this paper, artificial intelligence methods such as neural network (NN), deep neural network (DNN), and fuzzy C-means clustering combined with deep neural network (FCM-DNN) are developed for diagnosing CAD on a cardiac magnetic resonance imaging (CMRI) dataset. The original dataset is used in two different approaches. First, the labeled dataset is applied to the NN and DNN to create the NN and DNN models. Second, the labels are removed, and the unlabeled dataset is clustered via the FCM method, and then, the clustered dataset is fed to the DNN to create the FCM-DNN model. By utilizing the second clustering and modeling, the training process is improved, and consequently, the accuracy is increased. As a result, the proposed FCM-DNN model achieves the best performance with a 99.91% accuracy specifying 10 clusters, i.e., 5 clusters for healthy subjects and 5 clusters for sick subjects, through the 10-fold cross-validation technique compared to the NN and DNN models reaching the accuracies of 92.18% and 99.63%, respectively. To the best of our knowledge, no study has been conducted for CAD diagnosis on the CMRI dataset using artificial intelligence methods. The results confirm that the proposed FCM-DNN model can be helpful for scientific and research centers.



    Cardiac personalized modelling has been used for the non-invasive diagnosis and treatment of heart rhythm disorders, including risk classification of patients with heart attacks [1,2,3], prediction of the location of re-entry [4], and guidance for clinical ablation [5]. The key to the clinical application of heart models is the accurate creation of the personalized model, which is currently mostly segmented by experienced experts. Manual segmentation is subjective, irreproducible and time consuming, while model simulation takes a great deal of time and is time-critical in clinical practice, so how to minimize heart modeling time makes an automated segmentation method extremely important for the clinical application of personalized heart modelling.

    Before the rise of deep learning, classical medical image segmentation algorithms such as region-based, grayscale-based, and edge-based algorithms were well established for medical images [6,7,8,9], while traditional machine learning techniques such as model-based methods (e.g., active contour and deformable models) and atlas-based methods (e.g., single-atlas and multi-atlas) have achieved good performance [10,11,12,13]. Nevertheless, both classical image segmentation algorithms and machine learning techniques usually require some prior knowledge or feature annotation processing to achieve better results. In contrast, deep learning-based algorithms do not rely on these operations; it automatically discovers and learns complex features from the data for target segmentation and detection. These features are often learned directly from the data through generic learning procedures and end-to-end methods. This allows deep learning-based algorithms to be easily applied to other domains. Deep learning-based segmentation algorithms are gradually surpassing previously advanced traditional methods and are gaining popularity in research, not only because of developments and advances in computer hardware such as graphics processing units (GPU) and tensor processing units (TPU), but also because of the increase in publicly available datasets and open source code. This trend can be observed in Figure 1, where the number of deep learning-based cardiac image segmentation papers has grown considerably in the last few years, especially after year 2018. We searched the web of science database for the keywords cardiac image segmentation and deep learning, and all types of articles were counted.

    Figure 1.  Overview of the number of papers published between 01 January 2016 and 01 December 2021 on deep learning-based methods for cardiac image segmentation.

    Accurate localization and segmentation of medical images is a necessary prerequisite for diagnosis and treatment planning of cardiac diseases [14]. With the development of deep learning technology, various deep learning algorithms have been introduced into medical image processing and analysis with good results [15,16]. Convolutional neural networks (CNN) are one of the most common neural networks in medical image analysis, which are computationally fast and simple, and requiring no major adjustments to the network architecture [17]. CNN have been used with great success for medical image classification and segmentation, but a major drawback of this patch-based approach is that a separate network must be deployed for each patch at the time of inference, due to multiple overlapping in the image patches, which results in a large amount of redundancy and wasted resources. To solve this problem, fully convolutional networks [18] were created, which were designed to have an encoding-decoding structure that allows them to receive inputs of arbitrary size and produce outputs with the same size. However, this encoder-decoder structure also poses some limitations, such as the loss of some features, so there are many variants based on FCNs, but the most famous one is U-Net [19], which uses hopping connections to recover feature information in the down-sample paths to reduce the loss of spatial context information and thus obtain more accurate results. Subsequently U-Net gradually dominated in medical image processing, but it and its variants [20,21,22] also faced the lack of ability to build remotely correlated models. This is mainly due to the inherent limitations of the convolution operation [23].

    On the other hand, the success of Transformer, which captures remote dependencies, has made possible the solution of the above problem in recent years. Transformer was designed for sequence modeling and transformation tasks, and it is known for its focus on modeling remote dependencies in data. Its great success in the language domain has motivated researchers to investigate its adaptability to computer vision, especially since it has achieved good results on some recent image classification and segmentation tasks [24,25,26]. ViT [25] first introduced transformer to computer vision tasks by segmenting an image into 16 non-overlapping patches, feeding them into standard transformer with positional embedding and comparing it with the CNN-based approach, ViT achieved a fairly good performance, which broke the monopoly of U-Net in computer vision. With the advent of VIT, more and more transformer-based image processing became popular, such as Swin-transformer [24] proposed a hierarchical transformer and a sliding window attention-based transformer, while the pyramidal visual transformer (PVT) [26] proposed a gradual shrinkage strategy to control the scale of feature maps and proposed a spatially reduced attention (SRA) layer to replace the traditional multiple head attention (MHA) layer in encoders, which were designed mainly to reduce the computational complexity. But these transformer-based networks have a limitation of unable extracting low-level features like convolutional operations [27], so some detailed features will be ignored.

    To solve the above problem, we propose TF-Unet, a medical image segmentation framework that combines Transformer and U-Net. To fully utilize the advantages of both, we use two convolutional layers to learn high-resolution features and spatial location information in the learning feature phase, and use Transformer blocks to establish remote dependencies in the decoding phase. In terms of structure, inspired by the U-Net network structure, we divide the network into encoder-decoder blocks, and the self-attentive features of the coding blocks are combined with different high-resolution decoding features through hopping connections to reduce information loss. The results show that such a design allows our framework to maintain the advantages of both Convolution and Transformer, while facilitating the segmentation of medical images. Experimental results show that our proposed hybrid network has better performance and robustness compared to previous methods based on pure convolution and pure transformer.

    The ACDC dataset: (1) Raw Nifti images of 100 patients were used as the training set, and clinical experts used the corresponding manual reference analysis of ED and ES time phases as segmentation criteria, where trabecular and papillary muscles were included in the ventricular blood pool; (2) raw Nifti images of another 50 patients were used as the test set, providing only basic patient information: height and weight, and ED and ES time phases. ACDC data were acquired using 1.5 T and 3.0 T MRI scanners with retrospective or prospective balanced steady-state free-feed sequences. The scan parameters were as follows: layer thickness of 5–8 mm, layer spacing of 5 mm, layer thickness and layer spacing combined were typically 5–10 mm, matrix size was 256 × 256, FOV was 300 × 330 mm2, and one complete cardiac cycle consisted of 28–40 time phases.

    The Synapse dataset: (1) Raw Nifti images of 30 patients were used as training set; (2) Raw Nifti images of another 20 patients were used as test set, these 50 scans were taken at the portal venography stage with different volumes (512 × 512 × 85–512 × 512 × 198) and fields of view (approximately 280 × 280 × 280–500 × 500 × 650 mm3). The planar resolution varied from 0.54 × 0.54 to 0.98 × 0.98 mm2, while the slice thickness ranged from 2.5 to 5.0 mm.

    The general architecture of TF-Unet is shown in Figure 2, which maintains a U-shape similar to that of U-net [11] and consists of two main branches, i.e., encoder and decoder. Specifically, the encoder includes the feature extraction block, the transformer block, and the down-sampling block. The decoder branch includes the transformer block, the up-sampling block and the deconvolution block that finally maps the output. And, to recover the image details in the prediction, we add residual connections [28] between the corresponding feature pyramids of the encoder and decoder in a symmetric manner.

    Figure 2.  The overall architecture of TF-Unet, which is composed of encoder, decoder and skip connections. H, W, C represent the height, width and sequence length of each input patch, respectively.

    The feature extraction block is mainly responsible for converting each input image I into a high-dimensional tensor IRH4×W4×C, where H, W, C record the height, width, and sequence length of each input patch, respectively. Unlike Jieneng Chen et al. [24], who first flattened the input image directly and then preprocessed it in one dimension, we use a feature extraction layer which extracts low-level but high-resolution 3D features directly from the image and has more accurate spatial information at the pixel level.

    We use two consecutive convolutional layers with a kernel size of 3 and step sizes of 2 and 1 and use LeakyReLU nonlinear activation functions and LayerNorm for each layer, which not only allows us to encode spatial information more accurately than the faceted position encoding used in the transformer, but also helps to reduce computational complexity while providing equally sized perceptual fields.

    After the feature extraction block, we pass the high-dimensional tensor I to the Transformer block in two consecutive layers. The ability of the transformer to establish remote dependencies is fully exploited to establish the connection between the high-resolution features extracted in the upper layer and the multi-scale features obtained by convolutional downsampling in the next layer. Unlike the traditional multi-headed self-attentive module, this paper uses the Swin-Transformer module [24], who is constructed based on a sliding window. Since the window-based self-attentive module lacks cross-window connections, this limits its modeling capabilities. In order to introduce cross-window connectivity while maintaining efficient computation of non-overlapping windows, Ze Liu et al. [24] proposed a sliding window partitioning approach. In Figure 3, two consecutive transformer modules are given. Each Swin-Transformer block consists of a LayerNorm (LN) layer, a multi-headed self-attentive module, a skip connection, and an MLP (Multilayer Perceptron) with a LeakyReLU nonlinearity. The window-based multi-headed self-attention (W-MSA) module and the sliding window-based multi-headed self-attention (SW-MSA) module are applied in the two consecutive Transformer blocks, respectively. Based on this window division mechanism, the consecutive sliding Transformer blocks can be represented as Eqs (1)–(4).

    41yl=WMSA(LN(yl1))+yl1 (1)
    42yl=MLP(LN(yl))+yl (2)
    43yl+1=SWMSA(LN(yl))+yl (3)
    44yl+1=MLP(LN(yl+1))+yl+1 (4)
    Figure 3.  Two consecutive transformer blocks, the left is a fixed window, the right is a sliding window. Each transformer block is composed of LayerNorm layer, multi-head self-attention module and 2-layer MLP with LeakyReLU non-linearity.

    where l is the index of the layer. W-MSA and SW-MSA denote the volume-based multi-headed self-attentive and its transfer version. where yl and yl denotes the output of the W-MSA module and the MLP module of layer l, respectively. The computational complexity of SW-MSA on a volume of H × W × D patches is 4HWDC2 + 2SHSWSDHWDC, however, the computational complexity of naïve multi-headed self-attention (MSA) is 4HWDC2 + 2(HWD)2C. SH, SW, SD represent the height, width and depth of the sliding window respectively. SW-MSA greatly reduces the computational complexity of MSA, so our proposed algorithm is more efficient. The sliding window segmentation approach introduces connections between adjacent non-overlapping windows in the previous layer and has been found to be effective in image classification, object detection and semantic segmentation [23].

    In calculating the self-attention, we refer to Han Hu et al. [29,30] and add the relative position bias, and the specific formula for calculating the self-attention is as follows:

    47Attention(Q,K,V)=SoftMax(QKTd+B)V (5)

    where Q, K, V represent the query matrix, key matrix and value matrix, respectively. d is generally taken as the dimension of Q or K. BR(2H1)×(2W1) is the relative position encoding.

    Instead of completing the cascaded feature operations by using linear layers as in Swin-Unet [23], we directly use the convolution operation with stride size of 2. The reason for this is that the layered features generated by convolutional down-sampling help to model the target object at multiple scales. After such processing, the feature resolution is down-sampled by a factor of 2 and the feature dimension is increased to twice the original dimension.

    Corresponding to the Convolutional down-sampling, we also make changes in the up-sampling layer. We use stepwise deconvolution to up-sample the low-resolution feature map into a high-resolution feature map, i.e., by reconstructing the adjacent dimensional feature map into a higher-resolution feature map (2x up-sampling) and correspondingly reducing the feature dimension to half of the original dimension, and then by skip connecting, the features extracted from the encoder's down-sampling are combined with the decoder up-sampled features are merged. A deconvolution operation is also performed in the last patch extension block to produce the final result.

    Similar to U-Net [19], skip connections are used to fuse multiscale features from the encoder with up-sample features from the decoder. We splice shallow and deep features together to reduce the loss of spatial information due to down-sample.

    To fairly compare the experimental results, we test three times on the ACDC dataset to take the average, and to verify the robustness of our algorithm, we do the same test on the Synapse dataset.

    We ran all experiments based on Python 3.6, pytorch 1.8.1 and Ubutun 20.04. All training programs were executed on an NVIDIA 2080 GPU with 11 GB of RAM. The initial learning rate was set to 0.01, and we used the "poly" decay strategy [31] by default. As described in Eq (6):

    59lr=initial_lr×(1tem_epochmax_epoch)γ (6)

    where max_epoch represents the total number of training generations, default 1000, tem_epoch represents the current training generations, γ is the hyperparameter, default take 0.9.

    The default optimizer is stochastic gradient descent (SGD) and we set the momentum to 0.99. The weight decay is set to 3e-5. We use the weighted sum of the cross-entropy loss and the dice loss as the loss function. The training epochs is 1000 and each epoch contains 250 iterations.

    All images in the same dataset are firstly resampled to the same target spacing and then cropped to the same size. Since there are not enough training samples, some data enhancement operations, such as rotation, scaling, Gaussian blur, Gaussian noise, brightness and contrast adjustment, are performed during the training process

    In conducting experiments on the ACDC dataset, we designed two experimental scenarios; one is to make full use of the dataset, we use all 100 training data as the training set and 50 test data as the test set. The other is to quantitatively evaluate our results, we divide the 100 labeled training data into 70 training sets, 10 validation sets and 20 test sets. The real labels of the 20 cases used for testing were not put into the training.

    Figures 4 and 5 show the results of the first and second scenario, respectively. We randomly selected several patients' results for visualization [32].

    Figure 4.  The patient's consecutive nine-layer cardiac MRI results, the first and third columns are the original images, sequentially from the base to the apex of the heart, and the second and fourth columns correspond to the segmentation results on the left, respectively.
    Figure 5.  Cardiac MRI results in a patient with ground truth, the first columns are the original images, sequentially from the base to the apex of the heart, the second columns are the ground truth, and the third column are the segmentation results.

    The results of the second scenario are as follows, Table 1 shows the quantitative calculations and comparisons of RV, MYO and LV using the dice coefficients, and Figure 5 shows the raw plots of several randomly selected patient data, ground truth and predicted results. Due to the random nature of data partitioning, the results of the other methods in Table Ⅰ are taken from the results in the corresponding papers.

    Table 1.  Comparison on the ACDC MRI dataset (average dice score in %, dice score in % for each class).
    Methods DSC(avg) RV MYO LV
    R50-Unet [27] 87.55 87.10 80.36 94.92
    R50-Attn Unet [33] 86.75 87.58 79.20 93.47
    VIT [25] 81.45 81.46 70.71 92.18
    R50-VIT [25] 87.57 86.07 81.88 94.75
    TransUNet [27] 89.71 88.86 84.54 95.73
    Swin-Unet [23] 90.00 88.55 85.62 95.83
    Ours 91.72 90.16 89.40 95.60

     | Show Table
    DownLoad: CSV

    For Synapse data, we only use the second scenario in ACDC, we chose a part of labeled training set for testing, with training sample: validation sample: test sample = 14:7:9. We used the mean dice similarity coefficient (DSC) for eight abdominal organs, namely the aorta, gall bladder, spleen, left kidney, right kidney, liver, pancreas and stomach, to evaluate the model performance. Figure 6 shows the results of the different layers of patients from the Synapse dataset, with different colors representing different organs, as shown in the legend in Figure 6.

    Figure 6.  Results for one patient from the Synapse dataset, the first column is the original image, the second column is the ground truth, and the third column is the segmentation result. Due to the excessive number of scanned layers, we choose the intermediate layers to display, starting from layer 85 and displaying every ten layers at intervals until layer 125.

    In this section, we introduce the importance of learning rate strategies. In order to verify the effect of different learning rate strategies on the results, we did controlled experiments with four functions, inv, multistep, poly, step, and the results of the four methods are shown in Figure 7.

    Figure 7.  Results of four different learning rate strategies.

    In this section, we discuss in detail the experimental results obtained by our algorithm and explore the impact of different factors on the model performance, which we have compared on the ACDC and Synapse datasets, respectively. Specifically, we discuss the effects of different learning rate strategies on network performance.

    Analysis from a quantitative perspective. From Table 1, the best transformer-based model is Swin-Unet, which has an average dice coefficient of 90%. The best convolution-based model is R50-U-Net whose average dice coefficient is 87.55%, while our proposed TF-Unet is 1.72% higher than that of Swin-Unet and 4.17% higher than that of R50-U-Net. Considering that the current accuracy of these networks themselves is already very high, our proposed network improvement is still very effective, suggesting that our method can achieve better edge prediction. Analysis from a qualitative perspective. As can be seen in Figure 5, the middle represents the patient's true value and the rightmost represents our predicted value. By comparing layer by layer, the results obtained by our method are very close to the true value, and very good results are achieved even for the right ventricle, which is difficult to segment. In this work, we demonstrate that by combining Transformer with convolutional operations, better global and remote semantic information interactions can be learned, resulting in better segmentation results.

    It is well known that most of the deep learning networks cannot predict the results well for the test set without labeled values, but the network model based on TF-Unet can get good results. Observing Figure 4, we can conclude that the results obtained with our method are generally quite accurate for the layers other than the root tip layer. However, in the lower right corner of Figure 4, i.e., the apical layer, our method does not segment it. On the one hand, the apical layer has less segmentation in the training set, which makes it difficult for the network to learn features in this region; on the other hand, the true areas of both RV and LV in the apical layer are small and easily confused with surrounding vessels or tissues, leading to difficulties in segmentation.

    Figure 8 summarizes the learning process of our proposed network, it was observed that the training loss and validation loss decrease with the increase of iterations and reach a stable state at about 200 generations without overfitting. And the dice coefficient of the validation set increases with the number of iterations and reaches a steady state at 800 generations.

    Figure 8.  (a) The blue solid line represents training loss curves, (b) The red solid line represents validation loss curves, (c) The green dotted line represents dice score curves during validation.

    Quantitatively, as shown in Table 2, we performed experiments on Synapse and compared our TF-Unet with various transformer-based and Unet-based baselines. The main evaluation metric is the dice factor. the best performing transformer -based approach is Swin-Unet, which achieves an average score of 79.13. In contrast, DualNorm-UNet reports the best CNN-based results with an average of 80.37, slightly higher than Swin-Unet. our TF-Unet is able to outperform both Swin-Unet and DualNorm-UNet average performance by 6.33% and 5.09%, respectively, which is a considerable improvement on Synapse. Qualitatively, as can be seen in Figure 6, the middle column indicates the true value, and the rightmost column indicates the prediction result. For the segmentation of multiple organs, our proposed TF-Unet network still performs well, but there are some shortcomings for the stomach, as shown by the red boxes in the four lower right panels in Figure 6, one is that the prediction result is not smooth enough and there are many bursts, and the other is that it is difficult to segment to complex boundaries

    Table 2.  Comparison on the Synapse multi-organ CT dataset (average dice score in %, dice score in % for each class).
    Methods DSC(avg) Aorta Gallbladder Kidney(L) Kidney(R) Liver Pancreas Spleen Stomach
    R50-Unet [27] 74.68 87.74 63.66 80.60 78.19 93.74 56.90 85.87 74.16
    DualNorm-UNet [34] 80.37 86.52 55.51 88.64 86.29 95.64 55.91 94.62 79.80
    VIT [25] 67.86 70.19 45.10 74.70 67.40 91.32 42.00 81.75 70.44
    R50-VIT [25] 71.29 73.73 55.13 75.80 72.20 91.51 45.99 81.99 73.95
    SQNet [35] 73.76 83.55 61.17 76.87 69.40 91.53 56.55 85.82 65.24
    TransUNet [27] 77.48 87.23 63.16 81.87 77.02 94.08 55.86 85.08 75.62
    Swin-Unet [23] 79.13 85.47 66.53 83.28 79.61 94.29 56.58 90.66 76.60
    Ours 85.46 87.45 63.10 92.44 93.05 96.21 79.06 88.80 83.57

     | Show Table
    DownLoad: CSV

    Observing Figure 7 we can easily see that the results of all the functions are close except for the inv function. Through Figure 9 we speculate that this is because the learning rate of the inv function decreases too fast at the beginning of the iteration, and although it can speed up the search for the optimal solution, it is also easy to ignore the optimal solution and fall into the local optimal solution, leading to relatively poor results. The other three learning rates are all gradually decreasing, and although there is a big difference in the intermediate stages, the results do not differ much. These experiments show that the learning rate strategy has some influence on the experimental results, but it is generally enough to find the learning rate with the appropriate decreasing speed, and the different learning rate functions do not differ greatly.

    Figure 9.  Four different learning rate functions.

    In this paper, we propose a new medical image segmentation network TF-Unet. TF-Unet is built on the intertwined backbone of convolution and self-attention, which makes good use of the underlying features of CNN to build hierarchical object concepts at multiple scales through U-shaped hybrid architectural design. In addition play Transformer's powerful self-attention mechanism that entangles long-term dependencies with convolutionally extracted features to capture the global context. Based on this hybrid structure, TF-Unet has made a great progress in previous Transformer-based segmentation methods. In the future, we hope that TF-Unet can replace manual segmentation operations for cardiac modeling, effectively improve the efficiency of personalized modeling, and accelerate the development of personalized cardiac models in clinical applications.

    This study was supported by the Natural Science Foundation of China (NSFC) under grant number 62171408 and 81901841, the Key Research and Development Program of Zhejiang Province under grant number 2020C03016, the Major Scientific Project of Zhejiang Lab under grant number 2020ND8AD01, and Fundamental Research Funds for the Central Universities under grant number DUT21YG102.

    The authors declare that there are no conflicts of interest.



    [1] F. Jiang, Y. Jiang, H. Zhi, Y. Dong, H. Li, S. Ma, et al., Artificial intelligence in healthcare: past, present and future, Stroke Vasc. Neurol., 2 (2017). https://doi.org/10.1136/svn-2017-000101 doi: 10.1136/svn-2017-000101
    [2] M. D. McCradden, E. A. Stephenson, J. A. Anderson, Clinical research underlies ethical integration of healthcare artificial intelligence, Nat. Med., 26 (2020), 1325-1326. https://doi.org/10.1038/s41591-020-1035-9 doi: 10.1038/s41591-020-1035-9
    [3] K. H. Yu, A. L. Beam, I. S. Kohane, Artificial intelligence in healthcare, Nat. Biomed. Eng., 2 (2018), 719-731. https://doi.org/10.1038/s41551-018-0305-z doi: 10.1038/s41551-018-0305-z
    [4] O. Asan, A. E. Bayrak, A. Choudhury, Artificial intelligence and human trust in healthcare: focus on clinicians, J. Med. Internet Res., 22 (2020), e15154. https://doi.org/10.2196/15154 doi: 10.2196/15154
    [5] D. Shen, G. Wu, H. I. Suk, Deep learning in medical image analysis, Annu. Rev. Biomed. Eng., 19 (2017), 221-248. https://doi.org/10.1146/annurev-bioeng-071516-044442 doi: 10.1146/annurev-bioeng-071516-044442
    [6] G. Litjens, T. Kooi, B. E. Bejnordi, A. A. A. Setio, F. Ciompi, M. Ghafoorian, et al., A survey on deep learning in medical image analysis, Med. Image Anal., 42 (2017), 60-88. https://doi.org/10.1016/j.media.2017.07.005 doi: 10.1016/j.media.2017.07.005
    [7] M. I. Razzak, S. Naz, A. Zaib, Deep learning for medical image processing: Overview, challenges and the future, Classif. BioApps, (2018), 323-350.
    [8] J. H. Thrall, D. Fessell, P. V. Pandharipande, Rethinking the approach to artificial intelligence for medical image analysis: the case for precision diagnosis, J. Am. Coll. Radiol., 18 (2021), 174-179. https://doi.org/10.1016/j.jacr.2020.07.010 doi: 10.1016/j.jacr.2020.07.010
    [9] Y. Zhang, Z. Wang, J. Zhang, C. Wang, Y. Wang, H. Chen, et al., Deep learning model for classifying endometrial lesions, J. Transl. Med., 19 (2021), 1-13. https://doi.org/10.1186/s12967-020-02660-x doi: 10.1186/s12967-020-02660-x
    [10] C. Zheng, L. Chen, J. Jian, J. Li, Z. Gao, Efficacy evaluation of interventional therapy for primary liver cancer using magnetic resonance imaging and CT scanning under deep learning and treatment of vasovagal reflex, J. Supercomput., 77 (2021), 7535-7548. https://doi.org/10.1007/s11227-020-03539-w doi: 10.1007/s11227-020-03539-w
    [11] G. A. Roth, C. Johnson, A. Abajobir, F. Abd-Allah, S. F. Abera, G. Abyu, et al., Global, regional, and national burden of cardiovascular diseases for 10 causes, 1990 to 2015, J. Am. Coll. Cardiol., 70 (2017), 1-25.
    [12] K. H. Miao, J. H. Miao, Coronary heart disease diagnosis using deep neural networks, Int. J. Adv. Comput. Sci. Appl., 9 (2018), 1-8. https://doi.org/10.14569/IJACSA.2018.091001 doi: 10.14569/IJACSA.2018.091001
    [13] A. Gupta, H. S. Arora, R. Kumar, B. Raman, DMHZ: a decision support system based on machine computational design for heart disease diagnosis using z-alizadeh sani dataset, in 2021 International Conference on Information Networking (ICOIN), (2021), 818-823. https://doi.org/10.1109/ICOIN50884.2021.9333884
    [14] A. D. Villa, E. Sammut, A. Nair, R. Rajani, R. Bonamini, A. Chiribiri, Coronary artery anomalies overview: the normal and the abnormal, World J. Radiol., 8 (2016), 537. https://doi.org/10.4329/wjr.v8.i6.537 doi: 10.4329/wjr.v8.i6.537
    [15] R. Alizadehsani, M. Abdar, M. Roshanzamir, A. Khosravi, P. M. Kebria, F. Khozeimeh, et al., Machine learning-based coronary artery disease diagnosis: A comprehensive review, Comput. Biol. Med., 111 (2019), 103346. https://doi.org/10.1016/j.compbiomed.2019.103346 doi: 10.1016/j.compbiomed.2019.103346
    [16] T. M. Williamson, C. Moran, A. McLennan, S. Seidel, P. P. Ma, M. L. Koerner, T. S. Campbell, Promoting adherence to physical activity among individuals with cardiovascular disease using behavioral counseling: A theory and research-based primer for health care professionals, Prog. Cardiovasc. Dis., (2020). https://doi.org/10.1016/j.pcad.2020.12.007
    [17] J. H. Joloudari, E. H. Joloudari, H. Saadatfar, M. Ghasemigol, S. M. Razavi, A. Mosavi, et al., Coronary artery disease diagnosis; ranking the significant features using a random trees model, Int. J. Environ. Res. Public Health, 17 (2020), 731. https://doi.org/10.3390/ijerph17030731 doi: 10.3390/ijerph17030731
    [18] M. V. Dyke, S. Greer, E. Odom, L. Schieb, A. Vaughan, M. Kramer, et al., Heart disease death rates among blacks and whites aged ≥ 35 years-United States, 1968–2015, MMWR Surveillance Summaries, 67 (2018), 1. https://doi.org/10.15585/mmwr.ss6705a1
    [19] D. Mozaffarian, E. J. Benjamin, A. S. Go, D. K. Arnett, M. J. Blaha, M. Cushman, et al., Heart disease and stroke statistics—2015 update: a report from the American heart association, Circulation, 131 (2015), e29-e322.
    [20] E. J. Benjamin, S. S. Virani, C. W. Callaway, A. M. Chamberlain, A. R. Chang, S. Cheng, et al., Heart disease and stroke statistics—2018 update: a report from the American heart association, Circulation, 137 (2018), e67-e492. https://doi.org/10.1161/CIR.0000000000000573 doi: 10.1161/CIR.0000000000000573
    [21] H. Larochelle, Y. Bengio, J. Louradour, P. Lamblin, Exploring strategies for training deep neural networks, J. Mach. Learn. Res., 10 (2009).
    [22] R. O. Bonow, D. L. Mann, D. P. Zipes, P. Libby, Braunwald's Heart Disease E-Book: A Textbook of Cardiovascular Medicine, Elsevier Health Sciences, 2011.
    [23] E.G. Nabel, E. Braunwald, A tale of coronary artery disease and myocardial infarction, New Engl. J. Med., 366 (2012), 54-63. https://doi.org/10.1056/NEJMra1112570 doi: 10.1056/NEJMra1112570
    [24] İ. Babaoglu, O. Findik, E. Ülker, A comparison of feature selection models utilizing binary particle swarm optimization and genetic algorithm in determining coronary artery disease using support vector machine, Expert Syst. Appl., 37 (2010), 3177-3183. https://doi.org/10.1016/j.eswa.2009.09.064 doi: 10.1016/j.eswa.2009.09.064
    [25] M. Kumar, R. B. Pachori, U. R. Acharya, Characterization of coronary artery disease using flexible analytic wavelet transform applied on ECG signals, Biomed. Signal Proces. Control, 31 (2017), 301-308. https://doi.org/10.1016/j.bspc.2016.08.018 doi: 10.1016/j.bspc.2016.08.018
    [26] R. Alizadehsani, J. Habibi, M. J. Hosseini, R. Boghrati, A. Ghandeharioun, B. Bahadorian, et al., Diagnosis of coronary artery disease using data mining techniques based on symptoms and ecg features, Eur. J. Scientific Res., 82 (2012), 542-553.
    [27] R. Alizadehsani, J. Habibi, M. J. Hosseini, H. Mashayekhi, R. Boghrati, A. Ghandeharioun, et al., A data mining approach for diagnosis of coronary artery disease, Comput. Meth. Prog. Bio., 111(2013), 52-61. https://doi.org/10.1016/j.cmpb.2013.03.004 doi: 10.1016/j.cmpb.2013.03.004
    [28] R. Alizadehsani, J. Habibi, Z. A. Sani, H. Mashayekhi, R. Boghrati, A. Ghandeharioun, et al., Diagnosing coronary artery disease via data mining algorithms by considering laboratory and echocardiography features, Res. Cardiov. Med., 2 (2013), 133. https://doi.org/10.5812/cardiovascmed.10888 doi: 10.5812/cardiovascmed.10888
    [29] R. Alizadehsani, M. H. Zangooei, M. J. Hosseini, J. Habibi, A. Khosravi, M. Roshanzamir, et al., Coronary artery disease detection using computational intelligence methods, Knowledge-Based Syst., 109 (2016), 187-197. https://doi.org/10.1016/j.knosys.2016.07.004 doi: 10.1016/j.knosys.2016.07.004
    [30] A. D. Dolatabadi, S. E. Z. Khadem, B. M. Asl, Automated diagnosis of coronary artery disease (CAD) patients using optimized SVM, Comput. Meth. Prog. Bio., 138 (2017), 117-126. https://doi.org/10.1016/j.cmpb.2016.10.011 doi: 10.1016/j.cmpb.2016.10.011
    [31] Z. Arabasadi, R. Alizadehsani, M. Roshanzamir, H. Moosaei, A. A. Yarifard, et al., Computer aided decision making for heart disease detection using hybrid neural network-Genetic algorithm, Comput. Meth. Prog. Bio., 141 (2017), 19-26. https://doi.org/10.1016/j.cmpb.2017.01.004 doi: 10.1016/j.cmpb.2017.01.004
    [32] R. Alizadehsani, M. J. Hosseini, A. Khosravi, F. Khozeimeh, M. Roshanzamir, N. Sarrafzadegan, et al., Non-invasive detection of coronary artery disease in high-risk patients based on the stenosis prediction of separate coronary arteries, Comput. Meth. Prog. Bio., 162 (2018), 119-127. https://doi.org/10.1016/j.cmpb.2018.05.009 doi: 10.1016/j.cmpb.2018.05.009
    [33] M. Abdar, W. Książek, U. R. Acharya, R. S. Tan, V. Makarenkov, P. Pławiak, A new machine learning technique for an accurate diagnosis of coronary artery disease, Comput. Meth. Prog. Bio., 179 (2019), 104992. https://doi.org/10.1016/j.cmpb.2019.104992 doi: 10.1016/j.cmpb.2019.104992
    [34] C. Blake, UCI Repository of Machine Learning Databases, 1998. Available from: http://www.ics.uci.edu/~mlearn/MLRepository.html.
    [35] R. W. Hamersvelt, M. Zreik, M. Voskuil, M. A. Viergever, I. Išgum, T. Leiner, et al., Deep learning analysis of left ventricular myocardium in CT angiographic intermediate-degree coronary stenosis improves the diagnostic accuracy for identification of functionally significant stenosis, Eur. Rad., 29 (2019), 2350-2359. https://doi.org/10.1007/s00330-018-5822-3 doi: 10.1007/s00330-018-5822-3
    [36] U. R. Acharya, H. Fujita, O. S. Lih, M. Adam, J. H. Tan, C. K. Chua, Automated detection of coronary artery disease using different durations of ECG segments with convolutional neural network, Knowledge-Based Syst., 132 (2017), 62-71. https://doi.org/10.1016/j.knosys.2017.06.003 doi: 10.1016/j.knosys.2017.06.003
    [37] A. L. Goldberger, L. A. Amaral, L. Glass, J. M. Hausdorff, P. C. Ivanov, R. G. Mark, et al., PhysioBank, PhysioToolkit, and PhysioNet: components of a new research resource for complex physiologic signals, Circulation, 101 (2000), e215-e220. https://doi.org/10.1161/01.CIR.101.23.e215 doi: 10.1161/01.CIR.101.23.e215
    [38] J. H. Tan, Y. Hagiwara, W. Pang, I. Lim, S. L. Oh, M. Adam, et al., Application of stacked convolutional and long short-term memory network for accurate identification of CAD ECG signals, Comput. Biol. Med., 94 (2018), 19-26. https://doi.org/10.1016/j.compbiomed.2017.12.023 doi: 10.1016/j.compbiomed.2017.12.023
    [39] U. R. Acharya, H. Fujita, M. Adam, O. S. Lih, V. K. Sudarshan, T. J. Hong, et al., Automated characterization and classification of coronary artery disease and myocardial infarction by decomposition of ECG signals: A comparative study, Inform. Sci., 377 (2017), 17-29. https://doi.org/10.1016/j.ins.2016.10.013 doi: 10.1016/j.ins.2016.10.013
    [40] U. R. Acharya, H. Fujita, S. L. Oh, Y. Hagiwara, J. H. Tan, M. Adam, et al., Deep convolutional neural network for the automated diagnosis of congestive heart failure using ECG signals, Appl. Intell., 49 (2019), 16-27. https://doi.org/10.1007/s10489-018-1179-1 doi: 10.1007/s10489-018-1179-1
    [41] M. M. Ghiasi, S. Zendehboudi, A. A. Mohsenipour, Decision tree-based diagnosis of coronary artery disease: CART model, Comput. Meth. Prog. Bio., 192 (2020), 105400. https://doi.org/10.1016/j.cmpb.2020.105400 doi: 10.1016/j.cmpb.2020.105400
    [42] L. Verma, S. Srivastava, P. Negi, A hybrid data mining model to predict coronary artery disease cases using non-invasive clinical data, J. Med. Syst., 40 (2016), 178. https://doi.org/10.1007/s10916-016-0536-z doi: 10.1007/s10916-016-0536-z
    [43] N. M. Idris, Y. K. Chiam, K. D. Varathan, W. A. W. Ahmad, K. H. Chee, Y. M. Liew, Feature selection and risk prediction for patients with coronary artery disease using data mining, Med. Biol. Eng. Comput., 58 (2020), 3123-3140. https://doi.org/10.1007/s11517-020-02268-9 doi: 10.1007/s11517-020-02268-9
    [44] D, . Velusamy, K. Ramasamy, Ensemble of heterogeneous classifiers for diagnosis and prediction of coronary artery disease with reduced feature subset, Comput. Meth. Prog. Bio., 198 (2020), 105770. https://doi.org/10.1016/j.cmpb.2020.105770 doi: 10.1016/j.cmpb.2020.105770
    [45] I. Goodfellow, D. Warde-Farley, M. Mirza, A. Courville, Y. Bengio, Maxout networks, in International Conference On Machine Learning, PMLR, (2013), 1319-1327.
    [46] J. H. Joloudari, M. Haderbadi, A. Mashmool, M. GhasemiGol, S. S. Band, A. Mosavi, Early detection of the advanced persistent threat attack using performance analysis of deep learning, IEEE Access, 8 (2020), 186125-186137. https://doi.org/10.1109/ACCESS.2020.3029202 doi: 10.1109/ACCESS.2020.3029202
    [47] Y. Ito, Approximation of functions on a compact set by finite sums of a sigmoid function without scaling, Neural Networks, 4 (1991), 817-826. https://doi.org/10.1016/0893-6080(91)90060-I doi: 10.1016/0893-6080(91)90060-I
    [48] N. Hassan, N. Akamatsu, A new approach for contrast enhancement using sigmoid function, Inter Arab J. Inf. Techn., 1 (2004).
    [49] X. Li, X. Zhang, W. Huang, Q. Wang, Truncation cross entropy loss for remote sensing image captioning, IEEE Transactions Geosci. Remote Sens., 59 (2020), 5246-5257. https://doi.org/10.1109/TGRS.2020.3010106 doi: 10.1109/TGRS.2020.3010106
    [50] C. Otto, D. Wang, A. K. Jain, Clustering millions of faces by identity, IEEE Transactions Pattern Anal. Mach. Intell., 40 (2017), 289-303. https://doi.org/10.1109/TPAMI.2017.2679100 doi: 10.1109/TPAMI.2017.2679100
    [51] E. H. Ruspini, A new approach to clustering, Inform. Control, 15 (1969), 22-32. https://doi.org/10.1016/S0019-9958(69)90591-9 doi: 10.1016/S0019-9958(69)90591-9
    [52] R. O. Duda, P. E. Hart, Hart PE Pattern Classification And Scene Analysis, New York: Wiley, 1973.
    [53] R. Veloso, F. Portela, M. F. Santos, A. Silva, F. Rua, A. Abelha, et al., A clustering approach for predicting readmissions in intensive medicine, Procedia Technol., 16 (2014), 1307-1316. https://doi.org/10.1016/j.protcy.2014.10.147 doi: 10.1016/j.protcy.2014.10.147
    [54] H. S. Park, C. H. Jun, A simple and fast algorithm for K-medoids clustering, Expert Syst. Appl., 36 (2009), 3336-3341. https://doi.org/10.1016/j.eswa.2008.01.039 doi: 10.1016/j.eswa.2008.01.039
    [55] R. O. Duda, P. E. Hart, Pattern Classification And Scene Analysis, Wiley New York, 1973.
    [56] J. C. Dunn, Well-separated clusters and optimal fuzzy partitions, J. Cybern., 4 (1974), 95-104. https://doi.org/10.1080/01969727408546059 doi: 10.1080/01969727408546059
    [57] J. C. Bezdek, Objective function clustering, in Pattern Recognition With Fuzzy Objective Function Algorithms, Springer, (1981), 43-93. https://doi.org/10.1007/978-1-4757-0450-1_3
    [58] M. S. Yang, A survey of fuzzy clustering, Math. Comput. Model., 18 (1993), 1-16. https://doi.org/10.1016/0895-7177(93)90202-A doi: 10.1016/0895-7177(93)90202-A
    [59] G. Govaert, M. Nadif, Clustering with block mixture models, Pattern Recogn., 36 (2003), 463-473. https://doi.org/10.1016/S0031-3203(02)00074-2 doi: 10.1016/S0031-3203(02)00074-2
    [60] S. Bandyopadhyay, U. Maulik, A. Mukhopadhyay, Multiobjective genetic clustering for pixel classification in remote sensing imagery, IEEE Trans. Geosci. Remote Sens., 45 (2007), 1506-1511. https://doi.org/10.1109/TGRS.2007.892604 doi: 10.1109/TGRS.2007.892604
    [61] R. Xu, D. Wunsch, Survey of clustering algorithms, IEEE Trans. Neural Networks, 16 (2005), 645-678. https://doi.org/10.1109/TNN.2005.845141 doi: 10.1109/TNN.2005.845141
    [62] J. H. Joloudari, H. Saadatfar, A. Dehzangi, S. Shamshirband, Computer-aided decision-making for predicting liver disease using PSO-based optimized SVM with feature selection, Inform. Medicine Unlocked, 17 (2019), 100255. https://doi.org/10.1016/j.imu.2019.100255 doi: 10.1016/j.imu.2019.100255
    [63] M. Abdar, M. Zomorodi-Moghadam, R. Das, I. H. Ting, Performance analysis of classification algorithms on early detection of liver disease, Expert Syst. Appl., 67 (2017), 239-251. https://doi.org/10.1016/j.eswa.2016.08.065 doi: 10.1016/j.eswa.2016.08.065
    [64] C. H. Weng, C. K. Huang, R. P. Han, Disease prediction with different types of neural network classifiers, Telemat. Inform., 33 (2016), 277-292. https://doi.org/10.1016/j.tele.2015.08.006 doi: 10.1016/j.tele.2015.08.006
    [65] M. Diwakar, A. Tripathi, K. Joshi, M. Memoria, P. Singh, Latest trends on heart disease prediction using machine learning and image fusion, Materials Today: Proceed., 37 (2021), 3213-3218. https://doi.org/10.1016/j.matpr.2020.09.078 doi: 10.1016/j.matpr.2020.09.078
    [66] J. H. Moon, W. C. Cha, M. J. Chung, K. S. Lee, B. H. Cho, J. H. Choi, Automatic stenosis recognition from coronary angiography using convolutional neural networks, Comput. Meth. Prog. Bio., 198 (2021), 105819. https://doi.org/10.1016/j.cmpb.2020.105819 doi: 10.1016/j.cmpb.2020.105819
  • This article has been cited by:

    1. Kelei He, Chen Gan, Zhuoyuan Li, Islem Rekik, Zihao Yin, Wen Ji, Yang Gao, Qian Wang, Junfeng Zhang, Dinggang Shen, Transformers in medical image analysis, 2023, 3, 26671026, 59, 10.1016/j.imed.2022.07.002
    2. Xiang Li, Minglei Li, Pengfei Yan, Guanyi Li, Yuchen Jiang, Hao Luo, Shen Yin, Deep Learning Attention Mechanism in Medical Image Analysis: Basics and Beyonds, 2023, 2653-6226, 10.53941/ijndi0201006
    3. Jasmine El-Taraboulsi, Claudia P. Cabrera, Caroline Roney, Nay Aung, Deep neural network architectures for cardiac image segmentation, 2023, 4, 26673185, 100083, 10.1016/j.ailsci.2023.100083
    4. Md Rabiul Islam, Marwa Qaraqe, Erchin Serpedin, CoST-UNet: Convolution and swin transformer based deep learning architecture for cardiac segmentation, 2024, 96, 17468094, 106633, 10.1016/j.bspc.2024.106633
    5. Hamed Aghapanah, Reza Rasti, Saeed Kermani, Faezeh Tabesh, Hossein Yousefi Banaem, Hamidreza Pour Aliakbar, Hamid Sanei, William Paul Segars, CardSegNet: An adaptive hybrid CNN-vision transformer model for heart region segmentation in cardiac MRI, 2024, 115, 08956111, 102382, 10.1016/j.compmedimag.2024.102382
    6. Fan Yang, Fan Wang, Pengwei Dong, Bo Wang, HCA-former: Hybrid Convolution Attention Transformer for 3D Medical Image Segmentation, 2024, 90, 17468094, 105834, 10.1016/j.bspc.2023.105834
    7. Jiaji Wang, Shuihua Wang, Yudong Zhang, Deep learning on medical image analysis, 2024, 2468-2322, 10.1049/cit2.12356
    8. Swetha S, Aasiya Rafee, Manjula S H, Venugopal K R, 2023, Optimizing Left Atrium Segmentation: A Modified U-NET Architecture with MRI Image Slicing, 979-8-3503-1416-8, 1, 10.1109/ICDDS59137.2023.10434364
    9. Haiyan Song, Cuihong Liu, Shengnan Li, Peixiao Zhang, TS-GCN: A novel tumor segmentation method integrating transformer and GCN, 2023, 20, 1551-0018, 18173, 10.3934/mbe.2023807
    10. Bettina Baeßler, Sandy Engelhardt, Amar Hekalo, Anja Hennemuth, Markus Hüllebrand, Ann Laube, Clemens Scherer, Malte Tölle, Tobias Wech, Perfect Match: Radiomics and Artificial Intelligence in Cardiac Imaging, 2024, 17, 1941-9651, 10.1161/CIRCIMAGING.123.015490
    11. Sarv Priya, Durjoy D. Dhruba, Sarah S. Perry, Pritish Y. Aher, Amit Gupta, Prashant Nagpal, Mathews Jacob, Optimizing Deep Learning for Cardiac MRI Segmentation: The Impact of Automated Slice Range Classification, 2024, 31, 10766332, 503, 10.1016/j.acra.2023.07.008
    12. Narjes Benameur, Ramzi Mahmoudi, Mohamed Deriche, Amira Fayouka, Imene Masmoudi, Mohamed Hedi Jemaa, Nessrine Zoghlami, 2024, Automatic Deep Learning-based Myocardial Contours Segmentation from Cine MRI Images, 979-8-3503-7413-1, 76, 10.1109/SSD61670.2024.10548280
    13. Qile Zhang, Xiaoliang Jiang, Xiuqing Huang, Chun Zhou, MSR-Net: Multi-Scale Residual Network Based on Attention Mechanism for Pituitary Adenoma MRI Image Segmentation, 2024, 12, 2169-3536, 119371, 10.1109/ACCESS.2024.3449925
    14. Shuai Zhang, Yanmin Niu, LcmUNet: A Lightweight Network Combining CNN and MLP for Real-Time Medical Image Segmentation, 2023, 10, 2306-5354, 712, 10.3390/bioengineering10060712
    15. Xiaoke Lan, Honghuan Chen, Wenbing Jin, DRI-Net: segmentation of polyp in colonoscopy images using dense residual-inception network, 2023, 14, 1664-042X, 10.3389/fphys.2023.1290820
    16. Giorgos Papanastasiou, Nikolaos Dikaios, Jiahao Huang, Chengjia Wang, Guang Yang, Is Attention all You Need in Medical Image Analysis? A Review, 2024, 28, 2168-2194, 1398, 10.1109/JBHI.2023.3348436
    17. Wenli Cheng, Jiajia Jiao, An adversarially consensus model of augmented unlabeled data for cardiac image segmentation (CAU+), 2023, 20, 1551-0018, 13521, 10.3934/mbe.2023603
    18. Gang Lin, 2023, Segmentation of Cardiac MRI Images Based on nnU-Net, 978-1-6654-5703-3, 103, 10.1109/IPEC57296.2023.00026
    19. David Pospíšil, Monika Míková, Jan Řehoř, Kristýna Hochmanová, Veronika Bulková, Nela Hejtmánková, Biomedical engineering in contemporary Czech cardiology, 2024, 23, 1213807X, 110, 10.36290/kar.2024.023
    20. Tian Ma, JiaLi Qian, Xue Qin, DanDan Yu, 2024, Lightweight Dental Lesion Segmentation Algorithm Based on Transformer and CNN, 979-8-3503-8599-1, 183, 10.1109/ICIVC61627.2024.10837496
    21. Yaliang Tong, Kun Liu, Yuquan He, Ping Yang, CSS‐UNet: Convolution‐State Space Enhanced UNet for Cardiac MRI Segmentation, 2025, 61, 0013-5194, 10.1049/ell2.70175
    22. Ovais Iqbal Shah, Danish Raza Rizvi, Aqib Nazir Mir, Transformer-Based Innovations in Medical Image Segmentation: A Mini Review, 2025, 6, 2661-8907, 10.1007/s42979-025-03929-y
  • Reader Comments
  • © 2022 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(4068) PDF downloads(134) Cited by(8)

Figures and Tables

Figures(13)  /  Tables(8)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog