
Hepatocellular carcinoma is a highly malignant tumor and patients yield limited benefits from the existing treatments. The application of immune checkpoint inhibitors is promising but the results described in the literature are not favorable. It is therefore urgent to systematically analyze the immune microenvironment of HCC and screen the population best suited for the application of immune checkpoint inhibitors to provide a basis for clinical treatment. In this study, we collected The Cancer Genome Atlas Liver Hepatocellular Carcinoma (TCGA-LIHC)-related data sets to evaluate the immune microenvironment and immune cell infiltration (ICI) in HCC. Three independent ICI subtypes showing significant differences in survival were identified. Further, TCGA-LIHC immunophenoscore (IPS) was used to identify the differentially expressed genes between high- and low-IPS in HCC, so as to identify the immune gene subtypes in HCC tumors. The ICI score model for HCC was constructed, whereby we divided HCC samples into high- and low-score groups based on the median ICI score. The differences between these groups in genomic mutation load and immunotherapy benefit in HCC were examined in detail to provide theoretical support for accurate immunotherapy strategy in HCC. Finally, four genes were screened, which could accurately predict the subtype based on the tumor immune infiltration score. The findings may provide a basis and simplify the process for screening clinical drugs suitable for relevant subgroups.
Citation: Yu Jiang, Lijuan Lin, Huiming Lv, He Zhang, Lili Jiang, Fenfen Ma, Qiuyue Wang, Xue Ma, Shengjin Yu. Immune cell infiltration and immunotherapy in hepatocellular carcinoma[J]. Mathematical Biosciences and Engineering, 2022, 19(7): 7178-7200. doi: 10.3934/mbe.2022339
[1] | Jinke Wang, Lubiao Zhou, Zhongzheng Yuan, Haiying Wang, Changfa Shi . MIC-Net: multi-scale integrated context network for automatic retinal vessel segmentation in fundus image. Mathematical Biosciences and Engineering, 2023, 20(4): 6912-6931. doi: 10.3934/mbe.2023298 |
[2] | Chen Yue, Mingquan Ye, Peipei Wang, Daobin Huang, Xiaojie Lu . SRV-GAN: A generative adversarial network for segmenting retinal vessels. Mathematical Biosciences and Engineering, 2022, 19(10): 9948-9965. doi: 10.3934/mbe.2022464 |
[3] | G. Prethija, Jeevaa Katiravan . EAMR-Net: A multiscale effective spatial and cross-channel attention network for retinal vessel segmentation. Mathematical Biosciences and Engineering, 2024, 21(3): 4742-4761. doi: 10.3934/mbe.2024208 |
[4] | Caixia Zheng, Huican Li, Yingying Ge, Yanlin He, Yugen Yi, Meili Zhu, Hui Sun, Jun Kong . Retinal vessel segmentation based on multi-scale feature and style transfer. Mathematical Biosciences and Engineering, 2024, 21(1): 49-74. doi: 10.3934/mbe.2024003 |
[5] | Yinlin Cheng, Mengnan Ma, Liangjun Zhang, ChenJin Jin, Li Ma, Yi Zhou . Retinal blood vessel segmentation based on Densely Connected U-Net. Mathematical Biosciences and Engineering, 2020, 17(4): 3088-3108. doi: 10.3934/mbe.2020175 |
[6] | Yanxia Sun, Xiang Li, Yuechang Liu, Zhongzheng Yuan, Jinke Wang, Changfa Shi . A lightweight dual-path cascaded network for vessel segmentation in fundus image. Mathematical Biosciences and Engineering, 2023, 20(6): 10790-10814. doi: 10.3934/mbe.2023479 |
[7] | Jinzhu Yang, Meihan Fu, Ying Hu . Liver vessel segmentation based on inter-scale V-Net. Mathematical Biosciences and Engineering, 2021, 18(4): 4327-4340. doi: 10.3934/mbe.2021217 |
[8] | Tong Shan, Jiayong Yan, Xiaoyao Cui, Lijian Xie . DSCA-Net: A depthwise separable convolutional neural network with attention mechanism for medical image segmentation. Mathematical Biosciences and Engineering, 2023, 20(1): 365-382. doi: 10.3934/mbe.2023017 |
[9] | Rafsanjany Kushol, Md. Hasanul Kabir, M. Abdullah-Al-Wadud, Md Saiful Islam . Retinal blood vessel segmentation from fundus image using an efficient multiscale directional representation technique Bendlets. Mathematical Biosciences and Engineering, 2020, 17(6): 7751-7771. doi: 10.3934/mbe.2020394 |
[10] | Jun Liu, Zhenhua Yan, Chaochao Zhou, Liren Shao, Yuanyuan Han, Yusheng Song . mfeeU-Net: A multi-scale feature extraction and enhancement U-Net for automatic liver segmentation from CT Images. Mathematical Biosciences and Engineering, 2023, 20(5): 7784-7801. doi: 10.3934/mbe.2023336 |
Hepatocellular carcinoma is a highly malignant tumor and patients yield limited benefits from the existing treatments. The application of immune checkpoint inhibitors is promising but the results described in the literature are not favorable. It is therefore urgent to systematically analyze the immune microenvironment of HCC and screen the population best suited for the application of immune checkpoint inhibitors to provide a basis for clinical treatment. In this study, we collected The Cancer Genome Atlas Liver Hepatocellular Carcinoma (TCGA-LIHC)-related data sets to evaluate the immune microenvironment and immune cell infiltration (ICI) in HCC. Three independent ICI subtypes showing significant differences in survival were identified. Further, TCGA-LIHC immunophenoscore (IPS) was used to identify the differentially expressed genes between high- and low-IPS in HCC, so as to identify the immune gene subtypes in HCC tumors. The ICI score model for HCC was constructed, whereby we divided HCC samples into high- and low-score groups based on the median ICI score. The differences between these groups in genomic mutation load and immunotherapy benefit in HCC were examined in detail to provide theoretical support for accurate immunotherapy strategy in HCC. Finally, four genes were screened, which could accurately predict the subtype based on the tumor immune infiltration score. The findings may provide a basis and simplify the process for screening clinical drugs suitable for relevant subgroups.
Globally, hundreds of millions of people suffer from eye diseases. Conditions such as diabetic retinopathy, cataracts, age-related macular degeneration, uncorrected refractive errors and glaucoma contribute significantly to visual impairment [1]. Among various fundus pathologies, changes in the morphology and structure of retinal vessels are frequently observed, carrying vital information about the overall health of the eye and the body. Hence, segmentation of the retinal vessels holds significant importance for the early screening, diagnosis and treatment of related diseases. Initially, specialists relied on the manual segmentation of retinal vessels, which is a laborious, monotonous and time-consuming task [2], particularly when dealing with a large number of segments. Therefore, the development of computer-aided diagnostic systems for the rapid and efficient automated segmentation of fundus vessels has become immensely important for ocular medical research and applications.
However, due to its own characteristics, there are a number of challenges in the retinal vessel segmentation task. First, retinal images often exhibit uneven brightness [3], resulting in low contrast between the blood vessels and the background, particularly in capillaries, which are difficult to discern visually. Second, the presence of hemorrhages, hard exudates and other interferences in the fundus retinal image, which are affected by the lesion, can cause significant noise in vessel segmentation [4]. Lastly, the retinal vasculature is highly structured. Unlike polyp and skin datasets, retinal vessels start from the fovea and gradually spread throughout the eye varying in thickness and overlapping with each other, which makes segmentation more difficult.
In the past, numerous researchers have dedicated their efforts to extracting the structure and morphology of the retina from fundus images. Currently, the primary segmentation methods can be categorized into two groups: traditional methods using hand-designed features and methods employing deep learning. Traditional segmentation methods, which extract the vascular structure directly by building algorithmic models, can be broadly classified into matched filtering methods [5], morphological [6] and mathematical modeling methods [7], vascular tracking methods [8] and variable model methods [9]. Although traditional methods have shown progress in retinal vessel segmentation, they heavily depend on manual feature extraction and parameter selection, which can be labor-intensive. Moreover, these methods yield poor segmentation accuracy for processes of capturing vessel endings and cannot meet the requirements of clinical practice.
With the continuous development of deep learning, the proposal of convolutional neural networks has led to impressive achievements in medical image segmentation. Fully convolutional networks (FCNs) [10] constitute the basis of the first method to use convolutional neural networks for semantic segmentation. Fu et al. [11] introduced a network model called DeepVessel which utilizes an FCN as a framework and incorporates multi-scale features and conditional random fields. At that time DeepVessel achieved remarkable results on tasks of retinal vessel segmentation. Following this, Ronneberger et al. [12] proposed a classical network known as U-Net, which demonstrated excellent feature extraction capability and outstanding performance. Since then, U-Net and its variants [13,14,15,16] are still the most popular segmentation models in medical image processing and are widely employed for medical image segmentation tasks, such as retinal vessel segmentation. Nevertheless, these methods exhibit certain limitations in attempts to achieve high-precision retina segmentation. These include a restricted receptive domain, the loss of detailed information caused by changes in feature map size and a decline in segmentation accuracy, as attributed to the semantic gap.
To solve the above problems, we propose an efficient and flexible multi-fusion network with grouped attention (MAG-Net) for accurate retinal vessel segmentation. First, we use the classical data preprocessing scheme proposed by Jiang et al. [17] to initially alleviate the problems of uneven brightness in fundus images. Second, we adopt a U-shaped encoder-decoder structure [18], which performs well on denoising tasks, and introduce depthwise separable convolution and atrous convolution into the encoder. These operations enable enlargement of receptive field without inflation of the parameters and ensure that the feature map can learn more long-range contextual information. Additionally, we introduce a grouped attention enhancement (GAE) module for skip connections to further enhance the traditional channel attention method. By using high-level features to complement low-level features, GAE aims to reduce the semantic gap between the encoder and decoder. Finally, to further improve the model's learning capability, we propose a multi-scale feature fusion (MFF) module. This module integrates features from different decoder levels to obtain a more comprehensive and precise feature representation, helping the image restoration.
Our primary contributions can be divided into the following points:
1) We propose a new multi-fusion network with grouped attention that enhances feature extraction capability and improves model accuracy and robustness.
2) We introduce a grouped attention fusion module that facilitates information interaction between the encoder and decoder, preserving valuable information across different levels of features.
3) We have evaluated the proposed model on publicly available datasets (DRIVE, CHASE and STARE), and the results demonstrate the model's strong segmentation performance and overall effectiveness.
In recent years, U-Net and its variant models have demonstrated excellent segmentation capabilities on tasks of retinal vessel segmentation. The incorporation of innovative optimization mechanisms, including convolutional strategies [19], feature fusion techniques [20] and attention mechanisms [21], has further enhanced the segmentation performance of these models.
In traditional convolutional neural networks, the application of the receptive field plays a crucial role in segmentation [22]. How to improve the receptive field and enhance the feature extraction ability has been a popular research topic. DUNet [23] introduces deformable convolution, which adaptively adjusts the receptive field based on the scale and shape of the vessels. This adjustment allows the model to capture retinal vessels with diverse shapes and scales. SCS-Net [24] is a novel scale and context sensitive network for retinal vessel segmentation, and it includes a scale-aware feature aggregation module to dynamically adjust the receptive field to obtain multi-scale information. Zhang et al. [25] proposed a semi-isotropic model that maintains a relatively large receptive field at different stages of the convolutional network, demonstrating excellent segmentation performance. To enhance the feature extraction capability, Liu et al. proposed a ResDO-UNet [26] and applied it to retinal blood vessel segmentation by combining it with the DO-conv [27] layer. MCDAU-Net [28] was designed with a cascading dilated spatial pyramid pool module that enhances the receptive field and generates feature maps that are rich in contextual information.
In recent years, attention mechanisms have transcended their initial application of natural language processing and gained significant traction in medical image segmentation, especially on tasks of retinal vessel segmentation. For example, Guo et al. [21] proposed a lightweight network called SA-UNet, which incorporates channel attention to generate a spatial attention map of features along the spatial dimension for adaptive feature refinement. FANet [29] introduces a novel approach by integrating the current training features with the mask generated in the previous iteration to generate hard attention, which aims to suppress unwanted feature clutter. WA-Net [30] has been proposed as a width attention-based convolutional neural network that weighs the channel relevance via a width attention module, which considers the location and position correlations of feature maps. To overcome the difficulty of segmenting small blood vessels, a generator with attention augmented convolution has been proposed in [31] to highlight the region of interest in the whole image. DOLG-NeXt [32] integrates SE-Net and ConvNeXt, which overcomes the problem that the large transformer-based model performs poorly in small biomedical image environments, as well as realizes effective feature extraction and fine target segmentation.
While these methods have excelled on tasks of retinal vessel segmentation, they are not without limitations. Challenges such as the loss of fine-grained details during upsampling, significant computational overhead and the inclusion of excessive irrelevant information in attention mechanisms still persist.
The complex structure of blood vessels in fundus retinal images, which are susceptible to uneven illumination and noise, poses challenges for retinal vessel segmentation. To improve the effectiveness of retinal vessel segmentation, we propose the MAG-Net, a multi-fusion network with grouped attention.
Figure 1 illustrates the overall network structure of the MAG-Net, which consists of three components: the encoder-decoder structure, GAE module and MFF module. The backbone of our model is based on the three-layer U-Net structure, renowned for its exceptional performance and versatility on diverse medical image segmentation tasks. Initially, the input image is preprocessed and divided into greyscale images of size 48 × 48, which are then fed into the encoder. To downsample the feature maps, a maximum pooling operation is applied between the encoder blocks, reducing the dimensions by half. Similarly, a transposed convolution operation is employed between the decoder blocks to double the size of the feature maps. The bottleneck is composed of two convolution operations followed by batch normalization and ReLU activation. The letter C illustrates the changes in the feature maps for channels at each stage in Figure 1. In the MAG-Net, the hybrid convolutional fusion (HCF) module was first designed to obtain a larger receptive field and better feature extraction capability. Second, the GAE module is utilized to establish connections between corresponding encoder and decoder layers, facilitating the utilization of multi-scale features and enhancing the integration of detailed information from different locations. Finally, the MFF module combines feature maps generated at different scales in the decoder to better capture information at different levels and scales in the image. The three important modules are described in detail below.
To enhance the extraction capability of the encoder, the HCF module has been incorporated, and each original encoder module has been replaced by an HCF module. The HCF module combines depthwise separable convolutions and a spatial attention module, as illustrated in Figure 2. By employing depthwise convolutions with varying dilation rates, the HCF module encodes the fused features through element-level summation. Feature maps are then fed into the spatial attention module, making the network prioritize the vasculature. Next, a pointwise convolution is applied to perform linear transformations and nonlinear activations on individual pixels in each channel. Finally, we use a standard convolution operation, followed by DropBlock to prevent overfitting and accelerate the network convergence.
Spatial attention is a machine learning technique that is commonly used in image segmentation tasks; it emphasizes the interaction between features at different locations in spatial dimensions to help extract local features. The spatial attention module first performs global average pooling and global maximum pooling operations on the input features F∈RH×W×C along the channel dimension to obtain global information in the form of Fsavg∈RH×W×1 and Fsmax∈RH×W×1, and it subsequently generates spatial attention weights Ms∈RH×W×1 applying a 7 × 7 convolution operation and a sigmoid activation function. Finally, the spatial attention weights are multiplied element-wise with the input features and aggregated to obtain the output features Fs, thus allowing the network to focus on features at different locations for enhancement. In summary, the output features Fs can be expressed as
Fs=F⋅Ms=F⋅δ(f7×7[Fsavg;Fsmax])=F⋅δ(f7×7[avgpooling(F);maxpooling(F)]) | (3.1) |
where δ denotes a sigmoid function, f7×7 denotes a 7 × 7 convolution operation and [;] represents channel concatenation.
The integration of skip connections in U-Net facilitates information exchange between the encoder and the decoder, preventing information loss and enhancing training efficiency. However, the presence of skip connections can lead to a semantic gap problem, as caused by information distortion and loss within the encoder and the decoder. To mitigate this issue, the GAE module has been introduced into the skip connection. This module incorporates feature maps from different levels and utilizes high-level features, which contain rich category information, to guide the low-level features. The goal is to achieve more precise segmentation of intricate details.
The channel attention module, which is often used to selectively generate importance weights for each channel, is an integral component of the GAE module. The module details are depicted in Figure 3. Assuming that Xi denotes the input feature and H, W and C denote its height, width and number of input channels respectively. The channels are first compressed by using global average pooling Pavg and global maximum pooling Pmax to obtain the outputs Pavg(Xi)∈R1×1×C and Pmax(Xi)∈R1×1×C. This approach allows the pooling pixels to be taken into account and the important features to be extracted. A multi-layer perceptron (MLP) is used to stimulate the channels and adaptively recalibrate the channel relationships. The MLP consists of two fully connected layers with the first fully connected layer having an output channel number of C/r, where r=4. Subsequently, a ReLU activation function is introduced, and the output channel number of the second fully connected layer is restored to C. Subsequently, the stimulated results are fed into a sigmoid function and summed to obtain the channel attention weights βi. The exact computational procedure can be expressed as follows:
βi=δ(Ms(Pavg(Xi)))+δ(Ms(Pmax(Xi))) | (3.2) |
The specific structure of the GAE module is shown in Figure 4.
First, we use standard and transposed convolution operations to resize the low-level and high-level features in the encoder, respectively, to the same size and concatenate them into multi-scale features along the channel dimension. The multi-scale features are given as follows:
MF=[f3×3(Flow);up(Fhigh)] | (3.3) |
where Flow indicates the low-level features and Fhigh indicates the high-level features. Additionally, MF denotes the multi-scale features, f3×3 denotes a 3 × 3 standard convolution and up() means a transposed convolution operation.
Second, we employ a channel attention module to extract the channel attention weights of the feature maps at different scales, denoted as β1 and β2, respectively. The detailed calculation will be explained in the next paragraph. The multi-scale channel attention weights are obtained by cascading, and they can be expressed as follows:
β=[β1;β2] | (3.4) |
where β is the multi-scale channel attention weight. Next, a softmax function is applied to β to obtain the recalibrated soft attention atti. atti contains the complete spatial information and the attention weights on the channel dimensions that facilitate the interaction between a particular feature channel attention and another feature channel attention. This operation is defined as follows:
atti=Softmax(βi)=exp(βi)1∑i=0exp(βi),i=0,1 | (3.5) |
att is obtained by concatenating atti to denote the multi-scale channel attention weights after channel attention interactions, and it is expressed as follows:
att=[att1;att2] | (3.6) |
Subsequently, the recalibrated multi-scale channel attention weight att is multiplied by the multi-scale features to obtain the feature Fc. Finally, Fc is the output after applying a 1 × 1 convolution operation, which rescales channels and sums with adjusted low-level features.
Fc=f3×3(Flow)+f1×1(MF×att) | (3.7) |
where f1×1 denotes a 1 × 1 convolution operation.
Due to the multivariate shape and structure of fundus vessels, existing vessel segmentation methods still exhibit limitations. To overcome the problem of feature adaptation for retinal vessels at different scales, we introduced the MFF module. This module leverages bilinear interpolation upsampling to complete the semantic interaction between neighboring low-level features and high-level features as well as the aggregation of features at different scales. It merges spatial and channel information, obtaining detailed local and global information. The structure of the MFF module is shown in Figure 5.
Specifically, the MFF module initiates at the bottleneck which is positioned between the encoder and decoder. It performs a level-by-level decoding process while employing a 1 × 1 convolution operation at different scales, reducing the number of feature channels to 16 to decrease the computational effort. Next, the MFF module conducts an operation by applying bottom-up bilinear interpolation along the decoder, thus merging with the adjacent upper-level feature maps via element-by-element summation. Thereafter, a nonlinear activation function, ReLU, is applied to improve the fitting of the nonlinear model and the nonlinear transformation. After upsampling and feature merging, the feature map achieves a flexible and adaptive feature transformation and incorporates differences in information between scales, complementing the spatial and channel information. In addition, the MFF module employs upsampling and cascades the activated features from each layer through bilinear interpolation to obtain multi-scale cascaded features (MCFs). Finally, through a series of convolution operations, batch normalization and nonlinear activation, the MCFs recombine the spatial and channel information to generate multi-scale aggregated features, which are cascaded with an output feature map along the decoder to obtain the final output.
We validated our method on three standard public datasets (DRIVE, CHASE, STARE). These datasets are described in detail as follows:
1) DRIVE [33]. This dataset contains 40 digital retinal images, namely, the corresponding groundtruth images and the corresponding masked images, 20 of which are used for training and the other 20 for testing. The size of each image is 565 × 584 pixels, and each image contains annotations of retinal regions and vascular regions.
2) CHASE [34]. This dataset contains 28 digital retinal images, namely, the corresponding groundtruth images and the corresponding masked images, 14 of which are normal retinal images, and the other 14 are retinal images with lesions. The size of each image is 999 × 960 pixels. Twenty images were set as the training set, and the other eight images were set as the test set.
3) STARE [35]. This dataset contains 20 fundus images, namely, the corresponding groundtruth images and the corresponding masked images with an image size of 605 × 700 pixels. We divided the dataset with the first 10 images used for training and the last 10 images used for testing.
The segmentation performance of retinal vessels was quantitatively evaluated in this study by using several metrics: Dice coefficient, accuracy, sensitivity and specificity. These metrics were assessed by using a confusion matrix. The Dice coefficient measures the similarity between the predicted and labeled results, while the accuracy represents the ratio of correctly segmented pixels to the total pixels providing an overall assessment of segmentation accuracy. Sensitivity evaluates the model's ability to segment the vascular region by measuring the proportion of correctly segmented positive samples. Specificity quantifies the model's capability to segment non-vascular regions by measuring the proportion of correctly segmented negative samples. Accuracy reflects the accuracy of the model's prediction as a positive sample. Each metric is defined as follows:
Dice=2×TP2×TP+FN+FP | (4.1) |
Accuracy=TP+TNTP+FN+TN+FP | (4.2) |
Sensitivity=TPTP+FN | (4.3) |
Specificity=TNTN+FP | (4.4) |
Precision=TPTP+FP | (4.5) |
where true positive (TP) represents correctly segmented vessel pixels, true negative (TN) represents correctly segmented background pixels, false positive (FP) denotes the background pixels incorrectly segmented as vessel pixels and false negative (FN) denotes the vessel pixels incorrectly segmented as background pixels.
Our deep learning method, implemented through the use of a PyTorch framework, was evaluated on an Ubuntu 64-bit operating system using a QuadroRTX 6000 server. During training, we employed a random patch approach with a patch size set to 48 × 48 pixels and a total number of 104800 patches. The model underwent 100 iterations with an initial learning rate of 0.001. For the DRIVE dataset and CHASE dataset, the training batch size was 128 and the threshold was 0.48. For the STARE dataset, we set the batch size to 64 and the threshold to 0.48. The model was optimized by using the Adam optimizer, with the exponential decay rate β1 = 0.9, β2 = 0.999 and the momentum ϵ = 1 × 10−8.
The loss function we use is the cross-entropy loss function, defined as follows:
Lossce(y,ˆy)=−∑^yilog^yi+(1−yi)log(1−^yi) | (5.1) |
where yi denotes the true label and ^yi represents the predicted label.
The problem of uneven brightness in the original retinal image significantly interferes with automatic segmentation. Preprocessing of the fundus images can effectively alleviate this problem. We have processed the original images by employing the following four steps and the effect of each preprocessing step was as shown in Figure 6.
1) Three-channel fusion of the color images to convert the corresponding greyscale images;
2) Normalization of greyscale images to mean and standard deviation to improve the convergence speed of the model;
3) Adaptive histogram equalization (CLAHE) of greyscale images to increase the contrast between the blood vessels and the background and suppress noise;
4) Gamma correction for vessel images to suppress illumination unevenness and centerline reflections.
To verify the validity of each component in the MAG-Net, we conducted ablation experiments on the DRIVE, CHASE and STARE datasets. Specifically, we evaluated the impact of HCF, GAE and MFF, individually. Our baseline model was U-Net, while the standard model encompassed U-Net augmented with HCF, GAE and MFF. In the experimental setup, we also examined models without specific components, denoted as "w/o HCF", "w/o GAE" and "w/o MFF", respectively. The experimental results, displayed in Tables 1–3, highlight the most favorable outcomes for better visibility and analysis.
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9891 | 0.9856 | 0.8375 | 0.9704 | 0.8529 |
w/o HCF | 0.989 | 0.9846 | 0.8474 | 0.9705 | 0.8549 |
w/o GAE | 0.9892 | 0.9837 | 0.8544 | 0.9705 | 0.8556 |
w/o MFF | 0.9891 | 0.9842 | 0.8473 | 0.9701 | 0.8532 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9867 | 0.9883 | 0.7876 | 0.9765 | 0.7970 |
w/o HCF | 0.9878 | 0.9885 | 0.7957 | 0.9772 | 0.8031 |
w/o GAE | 0.9878 | 0.9872 | 0.8104 | 0.9769 | 0.8040 |
w/o MFF | 0.9880 | 0.9876 | 0.8014 | 0.9767 | 0.8009 |
ours | 0.9884 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | AUC_ROC | Pr | Sp | Se | Acc | Dice |
baseline | 0.9886 | 0.9166 | 0.9932 | 0.7461 | 0.9742 | 0.8161 |
w/o HCF | 0.9895 | 0.9195 | 0.9929 | 0.7578 | 0.9748 | 0.8219 |
w/o GAE | 0.9890 | 0.9138 | 0.9915 | 0.7604 | 0.9738 | 0.8167 |
w/o MFF | 0.9882 | 0.9141 | 0.9915 | 0.7676 | 0.9743 | 0.8208 |
ours | 0.9895 | 0.9158 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
As can be seen in Tables 1–3, the MAG-Net achieved better overall performance compared than the baseline model. The area under the receiver operating characteristic curve (AUC_ROC), accuracy, sensitivity and Dice coefficient for the MAG-Net respectively reached 0.9895, 0.9708, 0.8588 and 0.8576 on the DRIVE dataset, 0.9884, 0.9773, 0.8123 and 0.8069 on the CHASE dataset and 0.9895, 0.7781, 0.9743 and 0.8228 on the STARE dataset. When we removed the HCF module from the standard model and replaced it with the standard convolution operation in U-Net, the model showed a slight decrease of 0.04 and 1.14% for the AUC_ROC and 0.06 and 1.66% for sensitivity on the DRIVE dataset and CHASE dataset, respectively, and a decrease of 2.01% for sensitivity on the STARE dataset. It demonstrates that increasing the perceptual field by using depthwise atrous convolutions with different dilation rates is effective in segmenting vascular pixels accurately. When we removed the GAE, the data in Tables 1–3 revealed that the absence of the GAE module has a small effect, as the effects were more pronounced for the accuracy and Dice coefficient, which decreased by 0.03 and 0.29%, 0.04 and 0.29% and 0.05 and 0.61%, respectively. Results emphasize the importance of the GAE module in the MAG-Net. The effectiveness of the MFF module is also well demonstrated by the fact that the three metrics, i.e., sensitivity, accuracy and the Dice coefficient, decreased by 1.15, 0.07 and 0.42% on the DRIVE dataset and 1.09, 0.06 and 0.60% on the CHASE dataset without the MFF.
In Figure 7, we show the segmentation results of various models in the ablation experiments on three datasets, aiming to visually validate the effectiveness of our module in segmenting retinal vessel details. The figure showcases six columns, (a) to (f), representing the preprocessed fundus images, ground truth annotations and the segmentation results obtained by using "w/o HCF", "w/o GAE", "w/o MFF" and the MAG-Net, respectively. The results on the DRIVE, CHASE and STARE datasets are shown from top to bottom in the figure above. Upon examination of the images, it is evident that all four models possess the ability to accurately segment the main trunk of the vessel and finer vessels across different datasets. Using capillary segmentation as a significant metric, the MAG-Net demonstrates its exceptional performance. We have highlighted the segmentation of some capillaries by marking and enlarging the same areas of each image with red boxes.
The GAE module is a very important part of the MAG-Net, and to validate the reliability of our proposed grouped attention, we conducted a series of ablation experiments for the GAE module. We selected the widely used SE module [36], CBAM [37] module and GC block [38] to substitute the attention computation method in the GAE module and conducted experiments on the DRIVE dataset. Table 4 shows the results of our experiments.
Model | AUC_ROC | Sp | Se | Acc | Dice |
MAG-Net+SE | 0.9892 | 0.9837 | 0.8517 | 0.9702 | 0.8542 |
MAG-Net+CBAM | 0.9891 | 0.9850 | 0.844 | 0.9706 | 0.8547 |
MAG-Net+GC | 0.9894 | 0.9849 | 0.8471 | 0.9708 | 0.8559 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Although the addition of the SE, CBAM and GC block improves the model's performance to a certain extent, these three methods are slightly inferior compared to the grouped attention approach. This is because the GAE module not only learns the channel weights of features at different scales, it also retains the details of low-level features and high-level features to the greatest extent possible. At the same time, the GAE module is simple in structure and achieves efficient segmentation with only a small amount of computation.
Figure 8 presents a visual comparison of our model, MAG-Net, with other competing models. We selected U-Net [12], SA-UNet [21] and Attention-UNet [39] as the competition, and we applied a test image from both the DRIVE and CHASE datasets for visual analysis, arranged in left-to-right order. Columns (a)–(f) display the preprocessed original image, the groundtruth image, U-Net segmented results, Attention-UNet segmented results, SA-UNet segmented results and MAG-Net segmented results, respectively. Under each image, we included an enlarged local area to promote the comparison of the results.
By comparing the detailed regions, it can be seen that U-Net, Attention U-Net, SA-UNet and MAG-Net already have the ability to extract the vascular backbone from the original image and achieves satisfactory segmentation for most vascular regions. However, when comparing the groudtruth images, disconnections and mis-segmentations in the capillary region are easy to observe. Comparing the details, it can be seen that the MAG-Net does well at denoising, minimizes mis-segmentation and has excellent vessel connectivity.
To further demonstrate the superiority of our method, we compared the MAG-Net with some other deep learning-based retinal vessel segmentation methods. We evaluated them by using four metrics: specificity, sensitivity, accuracy and the Dice coefficient, respectively. Tables 5–7 indicate their segmentation results on the DRIVE, CHASE, and STARE datasets, respectively. When compared to existing segmentation methods, our proposed method showed significant improvements for most metrics, especially accuracy, outperforming the other methods on all three datasets. As can be seen in Table 5, the MAG-Net exhibits competitive ability that is not exhibited by other methods, with the best accuracy and Dice coefficient, which improved by 0.04 and 2.73%, respectively. As shown in Table 6, on the CHASE dataset, the MAG-Net excelled in the cases of the specificity and accuracy metrics, achieving 98.75 and 97.73% respectively. As can be seen in Table 7, similar to the CHASE dataset, the MAG-Net performed best for specificity and accuracy metrics, with a respective improvement of 0.28 and 0.18%.
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9813 | 0.7792 | 0.9556 | 0.8171 |
DUNet [23] | 2019 | 0.9800 | 0.7963 | 0.9566 | 0.8237 |
NFN+ [41] | 2020 | 0.9813 | 0.7996 | 0.9582 | 0.8295 |
SA-UNet [21] | 2020 | 0.9840 | 0.8212 | 0.9698 | 0.8263 |
MPS-Net [42] | 2021 | 0.9740 | 0.8361 | 0.9563 | 0.8278 |
SCS-Net [24] | 2021 | 0.9838 | 0.8289 | 0.9697 | 0.8189 |
AACA-MLA-D-UNet [43] | 2022 | 0.9805 | 0.8046 | 0.9581 | 0.8303 |
Bridge-Net [44] | 2022 | 0.9818 | 0.7853 | 0.9565 | 0.8203 |
CRAUNet [45] | 2022 | - | 0.7954 | 0.9586 | 0.8302 |
SDDC-Net [46] | 2023 | 0.9808 | 0.8603 | 0.9704 | 0.8289 |
ours | 2023 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9862 | 0.8298 | 0.9712 | 0.8475 |
DUNet [23] | 2019 | 0.9752 | 0.8155 | 0.9610 | 0.7883 |
NFN+ [41] | 2020 | 0.9880 | 0.8003 | 0.9688 | 0.8369 |
SA-UNet [21] | 2020 | 0.9835 | 0.8573 | 0.9755 | 0.8153 |
SCS-Net [24] | 2021 | 0.9839 | 0.8365 | 0.9744 | - |
MPS-Net [42] | 2021 | 0.9795 | 0.8488 | 0.9668 | 0.8332 |
AACA-MLA-D-UNet [43] | 2022 | 0.9801 | 0.8402 | 0.9673 | 0.8246 |
Bridge-Net [44] | 2022 | 0.9840 | 0.8132 | 0.9667 | 0.8293 |
CRAUNet [45] | 2022 | - | 0.8259 | 0.9659 | 0.8156 |
SDDC-Net [46] | 2023 | 0.9789 | 0.8268 | 0.9669 | 0.7965 |
ours | 2023 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9820 | 0.7756 | 0.9634 | 0.7928 |
DUNet [23] | 2019 | 0.9878 | 0.7595 | 0.9641 | 0.8143 |
NFN+ [41] | 2020 | 0.9863 | 0.7963 | 0.9672 | 0.8298 |
SCS-Net [24] | 2021 | 0.9839 | 0.8207 | 0.9736 | - |
MPS-Net [42] | 2021 | 0.9819 | 0.8566 | 0.9689 | 0.8491 |
AACA-MLA-D-UNet [43] | 2022 | 0.9870 | 0.7914 | 0.9665 | 0.8276 |
Bridge-Net [44] | 2022 | 0.9864 | 0.8002 | 0.9668 | 0.8289 |
SDDC-Net [46] | 2023 | 0.9784 | 0.7988 | 0.9642 | 0.7776 |
ours | 2023 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
Generalization ability is a very important performance evaluation metric for retinal vessel segmentation tasks. We evaluated the generalization ability of the model by performing cross-validation on different training and test sets. Table 8 gives a comparison of the generalization ability of the proposed method with three other methods, under the condition of the training being performed on the DRIVE dataset and testing performed on the STARE dataset, as well as the opposite experiment. Experiments show that our method has good overall performance and excellent performance on the Dice coefficient and sensitivity metrics, which shows that our method has excellent generalization ability in the task of vessel segmentation.
Database | Methods | Sp | Se | Acc | Dice |
STARE (trained on DRIVE) | MS-CANet [47] | - | - | 0.9673 | 0.7826 |
GDF-Net [48] | 0.9795 | 0.7089 | 0.9588 | - | |
AA-WGAN [31] | 0.9883 | 0.7839 | 0.9647 | - | |
ours | 0.9745 | 0.8471 | 0.9648 | 0.7855 | |
DRIVE (trained on STARE) | MS-CANet [47] | - | - | 0.9701 | 0.8035 |
GDF-Net [48] | 0.9902 | 0.7289 | 0.9593 | - | |
AA-WGAN [31] | 0.9928 | 0.8250 | 0.9635 | - | |
ours | 0.9828 | 0.8255 | 0.9690 | 0.8236 |
In addition, we present the Receiver operating characteristic (ROC) curves for the different models in the ablation experiment for the three datasets. As can be seen in Figure 9, the values of the area under the curve for all four models were close to 1, indicating the models' superior performance for retinal vessel segmentation. Of these, the MAG-Net had the largest area of the ROC curve among all four models for all three datasets, further emphasizing the superiority of the MAG-Net over the other models in the three ablation experiments.
Furthermore, we compared our work with other state of the art models in terms of the number of model parameters on the DRIVE dataset. The comparison results are shown in Table 9. It can be seen that our model demonstrated lower complexity than the other methods. In other words, our model performs well in terms of efficiency.
U-Net | ResUnet | R2U-Net | AACA-MLA-D-UNet | MRC-Net [49] | ours | |
# Param. | 4.32M | 32.61M | 39.09M | 2.03M | 0.9M | 0.6M |
This paper presents a new retinal vessel segmentation framework, the MAG-Net, which combines a multi-scale technique and an attention mechanism to realize improvements. The HCF module was introduced, and it utilizes spatial attention and multiple convolutions to expand the receptive field of the encoder, minimize noise and optimize the performance of the feature extractor at the source. The GAE module uses high-level features to guide low-level features on skip connections, optimizing segmentation details and alleviating the semantic gap between the encoder and decoder. The MFF module addresses the information loss problem during upsampling in the decoder by integrating features of multiple scales, which serves to supplement the detailed information. We validated our method on three datasets, i.e., the DRIVE, the CHASE and the STARE datasets. The experimental results show that our method achieves superior retinal vessel segmentation performance, as compared to U-Net, Attention-UNet and SA-UNet.
However, there are certain limitations of our work. First, the current implementation relies on annotated datasets, which can be expensive and restricts the availability of a larger number of annotated datasets. Second, our network was specifically designed for retinal vessel segmentation, while fundus images contain a wealth of information related to various fundus diseases. A single-target segmentation architecture is temporarily unable to meet the diagnostic needs of multiple fundus diseases.
Regarding future research, our research goals can be divided into two main points. First, we will focus on the improvement and optimization of the model to make it more adaptable to the specificity of fundus images and to increase its robustness for applicability to different pathological situations. Second, transfer learning and self-supervised learning are effective means to address data scarcity and labeling difficulties. In the absence of large-scale labeled data, we will consider exploring how to learn knowledge from other medical image tasks through transfer learning and use self-supervised learning to improve the performance of the model.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
Our work was supported in part by the National Natural Science Foundation of China (No. 61962054), the National Natural Science Foundation of China (No. 61163036), the Cultivation Plan of Major Scientific Research Projects of Northwest Normal University (No. NWNU-LKZD2021-06).
We declare that there is no conflict of interest.
[1] |
S. F. Altekruse, S. J. Henley, J. E. Cucinelli, K. A. McGlynn, Changing hepatocellular carcinoma incidence and liver cancer mortality rates in the United States, Am. J. Gastroenterol., 4 (2014), 542-553. https://doi.org/10.1038/ajg.2014.11 doi: 10.1038/ajg.2014.11
![]() |
[2] |
L. A. Torre, F. Bray, R. L. Siegel, J. Ferlay, J. Lortet-Tieulent, A. Jemal, Global cancer statistics, CA Cancer J. Clin., 2 (2015), 87-108. https://doi.org/10.3322/caac.21262 doi: 10.3322/caac.21262
![]() |
[3] |
J. M. Llovet, S. Ricci, V. Mazzaferro, P. Hilgard, E. Gane, J. F. Blanc, et al., Sorafenib in advanced hepatocellular carcinoma, N. Engl. J. Med., 4 (2008), 378-390. https://doi.org/10.1056/NEJMoa0708857 doi: 10.1056/NEJMoa0708857
![]() |
[4] |
M. Kudo, R. S. Finn, S. Qin, K. H. Han, K. Ikeda, F. Piscaglia, et al., Lenvatinib versus sorafenib in first-line treatment of patients with unresectable hepatocellular carcinoma: a randomised phase 3 non-inferiority trial, Lancet, 10126 (2018), 1163-1173. https://doi.org/10.1016/S0140-6736(18)30207-1 doi: 10.1016/S0140-6736(18)30207-1
![]() |
[5] |
N. H. Bhayani, Y. Jiang, O. Hamed, E. T. Kimchi, K. F. Staveley-Ocarroll, N. J. Gusani, Advances in the pharmacologic treatment of hepatocellular carcinoma, Curr. Clin. Pharmacol., 4 (2015), 299-304. https://doi.org/10.2174/1574884710666151020100059 doi: 10.2174/1574884710666151020100059
![]() |
[6] |
D. Neureiter, S. Stintzing, T. Kiesslich, M. Ocker, Hepatocellular carcinoma: Therapeutic advances in signaling, epigenetic and immune targets, World J. Gastroenterol., 25 (2019), 3136-3150. https://doi.org/10.3748/wjg.v25.i25.3136 doi: 10.3748/wjg.v25.i25.3136
![]() |
[7] |
A. B. El-Khoueiry, B. Sangro, T. Yau, T. S. Crocenzi, M. Kudo, C. Hsu, et al., Nivolumab in patients with advanced hepatocellular carcinoma (CheckMate 040): an open-label, non-comparative, phase 1/2 dose escalation and expansion trial, Lancet, 10088 (2017), 2492-2502. https://doi.org/10.1016/S0140-6736(17)31046-2 doi: 10.1016/S0140-6736(17)31046-2
![]() |
[8] |
A. X. Zhu, R. S. Finn, J. Edeline, S. Cattan, S. Ogasawara, D. Palmer, et al., Pembrolizumab in patients with advanced hepatocellular carcinoma previously treated with sorafenib (KEYNOTE-224): a non-randomised, open-label phase 2 trial, Lancet Oncol., 7 (2018), 940-952. https://doi.org/10.1016/S1470-2045(18)30351-6 doi: 10.1016/S1470-2045(18)30351-6
![]() |
[9] |
P. Federico, A. Petrillo, P. Giordano, D. Bosso, A. Fabbrocini, M. Ottaviano, et al., Immune checkpoint inhibitors in hepatocellular carcinoma, current status and novel perspectives, Cancers, 10 (2020), 3025. https://doi.org/10.3390/cancers12103025 doi: 10.3390/cancers12103025
![]() |
[10] |
M. Binnewies, E. W. Roberts, K. Kersten, V. Chan, D. F. Fearon, M. Merad, et al., Understanding the tumor immune microenvironment (TIME) for effective therapy, Nat. Med., 5 (2018), 541-550. https://doi.org/10.1038/s41591-018-0014-x doi: 10.1038/s41591-018-0014-x
![]() |
[11] |
X. Tekpli, T. Lien, A. H. Røssevold, D. Nebdal, E. Borgen, H. O. Ohnstad, et al., An independent poor-prognosis subtype of breast cancer defined by a distinct tumor immune microenvironment, Nat. Commun., 1 (2019), 5499. https://doi.org/10.1038/s41467-019-13329-5 doi: 10.1038/s41467-019-13329-5
![]() |
[12] |
D. C. Lazăr, M. F. Avram, I. Romoșan, M. Cornianu, S. Tăban, A. Goldiș, Prognostic significance of tumor immune microenvironment and immunotherapy: Novel insights and future perspectives in gastric cancer, World J. Gastroenterol., 32 (2018), 3583-3616. https://doi.org/10.3748/wjg.v24.i32.3583 doi: 10.3748/wjg.v24.i32.3583
![]() |
[13] |
C. M. Balch, L. B. Riley, Y. J. Bae, M. A. Salmeron, C. D. Platsoucas, A. V. Eschenbach, et al., Patterns of human tumor-infiltrating lymphocytes in 120 human cancers, Arc. Surg., 2 (1990), 200-205. https://doi.org/10.1001/archsurg.1990.01410140078012 doi: 10.1001/archsurg.1990.01410140078012
![]() |
[14] |
R. A. Caruso, R. Bellocco, M. Pagano, G. Bertoli, L. Rigoli, C. Inferrera, Prognostic value of intratumoral neutrophils in advanced gastric carcinoma in a high-risk area in northern Italy, Mod. Pathol., 8 (2002), 831-837. https://doi.org/10.1097/01.MP.0000020391.98998.6B doi: 10.1097/01.MP.0000020391.98998.6B
![]() |
[15] |
R. D. Bense, C. Sotiriou, M. J. Piccart-Gebhart, J. Haanen, M. van Vugt, E. G. E. de Vries, et al., Relevance of tumor-infiltrating immune cell composition and functionality for disease outcome in breast cancer, J. Natl. Cancer Inst., 1 (2017), 192. https://doi.org/10.1093/jnci/djw192 doi: 10.1093/jnci/djw192
![]() |
[16] | S. C. Zhang, Z. Q. Hu, J. H. Long, G. M. Zhu, Y. Wang, Y. Jia, et al., Clinical implications of tumor-infiltrating immune cells in breast cancer, J. Cancer, 10 (2019), 6175-6184. https://doi.org/10.7150/jca.35901. eCollection 2019 |
[17] |
M. Mroweh, T. Decaens, P. N. Marche, Z. M. Jilkova, F. Clément, Modulating the crosstalk between the tumor and its microenvironment using RNA interference: A treatment strategy for hepatocellular carcinoma, Int. J. Mol. Sci., 21 (2020), 5250. https://doi.org/10.3390/ijms21155250 doi: 10.3390/ijms21155250
![]() |
[18] |
T. Shiraki, E. Takayama, H. Magari, T. Nakata, T. Maekita, S. Enomoto, et al., Altered cytokine levels and increased CD4+CD57+ T cells in the peripheral blood of hepatitis C virus-related hepatocellular carcinoma patients, Oncol. Rep., 26 (2011), 201-208. https://doi.org/10.3892/or.2011.1258 doi: 10.3892/or.2011.1258
![]() |
[19] |
A. Mantovani, M. Locati, Tumor-associated macrophages as a paradigm of macrophage plasticity, diversity, and polarization: lessons and open questions, Arterioscler. Thromb. Vasc. Biol., 33 (2013), 1478-1483. https://doi.org/10.1161/ATVBAHA.113.300168 doi: 10.1161/ATVBAHA.113.300168
![]() |
[20] |
M. R. de Galarreta, E. Bresnahan, P. Molina-Sánchez, K. E. Lindblad, B. Maier, D. Sia, et al., β-catenin activation promotes immune escape and resistance to Anti-PD-1 therapy in hepatocellular carcinoma, Cancer Discov., 9 (2019), 1124-1141. https://doi.org/10.1158/2159-8290 doi: 10.1158/2159-8290.CD-19-0074
![]() |
[21] |
J. J. Harding, S. Nandakumar, J. Armenia, D. N. Khalil, M. Albano, M. Ly, et al., Prospective genotyping of hepatocellular carcinoma: Clinical implications of next-generation sequencing for matching patients to targeted and immune therapies, Clin. Cancer Res., 25 (2019), 2116-2126. https://doi.org/10.1158/1078-0432 doi: 10.1158/1078-0432.CCR-18-2293
![]() |
[22] |
R. S. Finn, B. Y. Ryoo, P. Merle, M. Kudo, M. Bouattour, H. Y. Lim, et al., Pembrolizumab as second-line therapy in patients with advanced hepatocellular carcinoma in KEYNOTE-240: A randomized, double-blind, phase Ⅲ trial, J. Clin. Oncol., 38 (2020), 193-202. https://doi.org/10.1200/JCO.19.01307 doi: 10.1200/JCO.19.01307
![]() |
[23] |
A. Dalessio, A. Cammarota, M. G. Prete, T. Pressiani, L. Rimassa, The evolving treatment paradigm of advanced hepatocellular carcinoma: putting all the pieces back together, Curr. Opin. Oncol., 33 (2021), 386-394. https://doi.org/10.1097/CCO.0000000000000744 doi: 10.1097/CCO.0000000000000744
![]() |
[24] |
R. C. Miksch, M. B. Schoenberg, M. Weniger, F. Bösch, S. Ormanns, B. Mayer, et al., Prognostic impact of tumor-infiltrating lymphocytes and neutrophils on survival of patients with upfront resection of pancreatic cancer, Cancers, 11 (2019), 39. https://doi.org/:10.3390/cancers11010039 doi: 10.3390/cancers11010039
![]() |
[25] |
S. Sousa, R. Brion, M. Lintunen, P. Kronqvist, J. Sandholm, J. Mönkkönen, et al., Human breast cancer cells educate macrophages toward the M2 activation status, Breast Cancer Res., 17 (2015), 101. https://doi.org/10.1186/s13058-015-0621-0 doi: 10.1186/s13058-015-0621-0
![]() |
[26] |
J. Lan, L. Sun, F. Xu, L. Liu, F. Hu, D. Song, et al., M2 macrophage-derived exosomes promote cell migration and invasion in colon cancer, Cancer Res., 79 (2019), 146-158. https://doi.org/10.1158/0008-5472 doi: 10.1158/0008-5472.CAN-18-0014
![]() |
[27] |
Y. Kurebayashi, H. Ojima, H. Tsujikawa, N. Kubota, J. Maehara, Y. Abe, et al., Landscape of immune microenvironment in hepatocellular carcinoma and its additional impact on histological and molecular classification, Hepatology, 68 (2018), 1025-1041. https://doi.org/10.1002/hep.29904 doi: 10.1002/hep.29904
![]() |
[28] |
Q. F. Chen, W. Li, P. H. Wu, L. J. Shen, Z. L. Huang, Significance of tumor-infiltrating immunocytes for predicting prognosis of hepatitis B virus-related hepatocellular carcinoma, World J. Gastroenterol., 25 (2019), 5266-5282. https://doi.org/10.3748/wjg.v25.i35.5266 doi: 10.3748/wjg.v25.i35.5266
![]() |
[29] |
N. Rohr-Udilova, F. Klinglmüller, R. Schulte-Hermann, J. Stift, M. Herac, M. Salzmann, et al., Deviations of the immune cell landscape between healthy liver and hepatocellular carcinoma, Sci. Rep., 8 (2018), 6220. https://doi.org/10.1038/s41598-018-24437-5 doi: 10.1038/s41598-018-24437-5
![]() |
[30] |
Y. Peng, C. Liu, M. Li, W. Li, M. Zhang, X. Jiang, et al., Identification of a prognostic and therapeutic immune signature associated with hepatocellular carcinoma, Cancer Cell. Int., 21 (2021), 98. https://doi.org/10.1186/s12935-021-01792-4 doi: 10.1186/s12935-021-01792-4
![]() |
[31] |
M. Morita, N. Nishida, K. Sakai, T. Aoki, H. Chishina, M. Takita, et al., Immunological microenvironment predicts the survival of the patients with hepatocellular carcinoma treated with anti-PD-1 antibody, Liver Cancer, 10 (2021), 380-393. https://doi.org/10.1159/000516899 doi: 10.1159/000516899
![]() |
[32] |
W. Ding, Y. Tan, Y. Qian, W. Xue, Y. Wang, P. Jiang, et al., Clinicopathologic and prognostic significance of tumor-associated macrophages in patients with hepatocellular carcinoma: A meta-analysis, Plos One, 14 (2019), e0223971. https://doi.org/10.1371/journal.pone.0223971 doi: 10.1371/journal.pone.0223971
![]() |
[33] |
O. W. Yeung, C. M. Lo, C. C. Ling, X. Qi, W. Geng, C. X. Li, et al., Alternatively activated (M2) macrophages promote tumour growth and invasiveness in hepatocellular carcinoma, J. Hepatol., 62 (2015), 607-616. https://doi.org/10.1016/j.jhep.2014.10.029 doi: 10.1016/j.jhep.2014.10.029
![]() |
[34] |
T. Flecken, N. Schmidt, S. Hild, E. Gostick, O. Drognitz, R. Zeiser, et al., Immunodominance and functional alterations of tumor-associated antigen-specific CD8+ T-cell responses in hepatocellular carcinoma, Hepatology, 59 (2014), 1415-1426. https://doi.org/10.1002/hep.26731 doi: 10.1002/hep.26731
![]() |
[35] |
A. M. van der Leun, D. S. Thommen, T. N. Schumacher, CD8(+) T cell states in human cancer: insights from single-cell analysis, Nat. Rev. Cancer, 20 (2020), 218-232. https://doi.org/10.1038/s41568-019-0235-4 doi: 10.1038/s41568-019-0235-4
![]() |
[36] |
H. Liao, W. Chen, Y. Dai, J. J. Richardson, J. Guo, K. Yuan, et al., Expression of programmed cell death-ligands in hepatocellular carcinoma: Correlation with immune microenvironment and survival outcomes, Front. Oncol., 9 (2019), 883. https://doi.org/10.3389/fonc.2019.00883 doi: 10.3389/fonc.2019.00883
![]() |
[37] |
Y. Zhou, Y. Zhang, G. Guo, X. Cai, H. Yu, Y. Cai, et al., Nivolumab plus ipilimumab versus pembrolizumab as chemotherapy-free, first-line treatment for PD-L1-positive non-small cell lung cancer, Clin. Transl. Med., 10 (2020), 107-115. https://doi.org/10.1002/ctm2.14 doi: 10.1002/ctm2.14
![]() |
[38] |
S. P. Patel, R. Kurzrock, PD-L1 expression as a predictive biomarker in cancer immunotherapy, Mol. Cancer Ther., 14 (2015), 847-856. https://doi.org/10.1158/1535-7163 doi: 10.1158/1535-7163.MCT-14-0983
![]() |
[39] |
C. Sun, R. Mezzadra, T. N. Schumacher, Regulation and function of the PD-L1 checkpoint, Immunity, 48 (2018), 434-452. https://doi.org/10.1016/j.immuni.2018.03.014 doi: 10.1016/j.immuni.2018.03.014
![]() |
[40] |
Z. Wang, Z. Lv, Q. Xu, L. Sun, Y. Yuan, Identification of differential proteomics in Epstein-Barr virus-associated gastric cancer and related functional analysis, Cancer Cell. Int., 21 (2021), 368. https://doi.org/10.1186/s12935-021-02077-6 doi: 10.1186/s12935-021-02077-6
![]() |
[41] |
S. W. Cheng, P. C. Chen, M. H. Lin, T. R. Ger, H. W. Chiu, Y. F. Lin, GBP5 repression suppresses the metastatic potential and PD-L1 expression in triple-negative breast cancer, Biomedicines, 9 (2021), 371. https://doi.org/10.3390/biomedicines9040371 doi: 10.3390/biomedicines9040371
![]() |
[42] |
M. Watanabe, M. Shibata, T. Inaishi, T. Ichikawa, I. Soeda, N. Miyajima, et al., MZB1 expression indicates poor prognosis in estrogen receptor-positive breast cancer, Oncol. Lett., 5 (2020), 198. https://doi.org/10.3892/ol.2020.12059 doi: 10.3892/ol.2020.12059
![]() |
[43] |
W. Wu, Z. Yang, F. Long, L. Luo, Q. Deng, J. Wu, et al., COL1A1 and MZB1 as the hub genes influenced the proliferation, invasion, migration and apoptosis of rectum adenocarcinoma cells by weighted correlation network analysis, Bioorg. Chem., 95 (2020), 103457. https://doi.org/10.1016/j.bioorg.2019 doi: 10.1016/j.bioorg.2019.103457
![]() |
[44] |
W. A. Freed-Pastor, L. J. Lambert, Z. A. Ely, N. B. Pattada, A. Bhutkar, G. Eng, et al., The CD155/TIGIT axis promotes and maintains immune evasion in neoantigen-expressing pancreatic cancer, Cancer Cell, 39 (2021), 1342-1360. https://doi.org/10.1016/j.ccell. 2021.07.007 doi: 10.1016/j.ccell.2021.07.007
![]() |
[45] |
Z. Ge, M. P. Peppelenbosch, D. Sprengers, J. Kwekkeboom, TIGIT, the next step towards successful combination immune checkpoint therapy in cancer, Front. Immunol., 12 (2021), 699895. https://doi.org/10.3389/fimmu.2021.699895 doi: 10.3389/fimmu.2021.699895
![]() |
[46] |
L. Mao, Y. Xiao, Q. C. Yang, S. C. Yang, L. L. Yang, Z. J. Sun, TIGIT/CD155 blockade enhances anti-PD-L1 therapy in head and neck squamous cell carcinoma by targeting myeloid-derived suppressor cells, Oral. Oncol., 121 (2021), 105472. https://doi.org/10.1016/j. oraloncology.2021.105472 doi: 10.1016/j.oraloncology.2021.105472
![]() |
[47] |
N. Li, B. Li, X. Zhan, Comprehensive analysis of tumor microenvironment identified prognostic immune-related gene signature in ovarian cancer, Front. Genet., 12 (2021), 616073. https://doi.org/10.3389/fgene.2021.616073 doi: 10.3389/fgene.2021.616073
![]() |
[48] |
S. A. Bhat, S. F. Walton, T. Ventura, X. Liu, J. S. McCarthy, S. T. G. Burgess, et al., Early immune suppression leads to uncontrolled mite proliferation and potent host inflammatory responses in a porcine model of crusted versus ordinary scabies, Plos Negl. Trop. Dis., 9 (2020), e0008601. https://doi.org/10.1371/journal.pntd.0008601 doi: 10.1371/journal.pntd.0008601
![]() |
[49] |
N. E. Ready, P. A. Ott, M. D. Hellmann, J. Zugazagoitia, C. L. Hann, F. de Braud, et al., Nivolumab monotherapy and nivolumab plus ipilimumab in recurrent small cell lung cancer: Results from the checkMate 032 randomized cohort, J. Thorac. Oncol., 3 (2020), 426-435. https://doi.org/10.1016/j.jtho.2019.10.004 doi: 10.1016/j.jtho.2019.10.004
![]() |
[50] |
R. M. Samstein, C. H. Lee, A. N. Shoushtari, M. D. Hellmann, R. Shen, Y. Y. Janjigian, et al., Tumor mutational load predicts survival after immunotherapy across multiple cancer types. Nat. Genet., 2 (2019), 202-206. https://doi.org/10.1038/s41588-018-0312-8 doi: 10.1038/s41588-018-0312-8
![]() |
[51] |
L. Liu, X. Bai, J. Wang, X. R. Tang, D. H. Wu, S. S. Du, et al., Combination of TMB and CNA stratifies prognostic and predictive responses to immunotherapy across metastatic cancer, Clin. Cancer Res., 24 (2019), 7413-7423. https://doi.org/10.1158/1078-0432 doi: 10.1158/1078-0432
![]() |
[52] |
R. Cristescu, R. Mogg, M. Ayers, A. Albright, E. Murphy, J. Yearley, et al., Pan-tumor genomic biomarkers for PD-1 checkpoint blockade-based immunotherapy, Science, 6411 (2018), 3593. https://doi.org/10.1126/science.aar3593 doi: 10.1126/science.aar3593
![]() |
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9891 | 0.9856 | 0.8375 | 0.9704 | 0.8529 |
w/o HCF | 0.989 | 0.9846 | 0.8474 | 0.9705 | 0.8549 |
w/o GAE | 0.9892 | 0.9837 | 0.8544 | 0.9705 | 0.8556 |
w/o MFF | 0.9891 | 0.9842 | 0.8473 | 0.9701 | 0.8532 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9867 | 0.9883 | 0.7876 | 0.9765 | 0.7970 |
w/o HCF | 0.9878 | 0.9885 | 0.7957 | 0.9772 | 0.8031 |
w/o GAE | 0.9878 | 0.9872 | 0.8104 | 0.9769 | 0.8040 |
w/o MFF | 0.9880 | 0.9876 | 0.8014 | 0.9767 | 0.8009 |
ours | 0.9884 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | AUC_ROC | Pr | Sp | Se | Acc | Dice |
baseline | 0.9886 | 0.9166 | 0.9932 | 0.7461 | 0.9742 | 0.8161 |
w/o HCF | 0.9895 | 0.9195 | 0.9929 | 0.7578 | 0.9748 | 0.8219 |
w/o GAE | 0.9890 | 0.9138 | 0.9915 | 0.7604 | 0.9738 | 0.8167 |
w/o MFF | 0.9882 | 0.9141 | 0.9915 | 0.7676 | 0.9743 | 0.8208 |
ours | 0.9895 | 0.9158 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
Model | AUC_ROC | Sp | Se | Acc | Dice |
MAG-Net+SE | 0.9892 | 0.9837 | 0.8517 | 0.9702 | 0.8542 |
MAG-Net+CBAM | 0.9891 | 0.9850 | 0.844 | 0.9706 | 0.8547 |
MAG-Net+GC | 0.9894 | 0.9849 | 0.8471 | 0.9708 | 0.8559 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9813 | 0.7792 | 0.9556 | 0.8171 |
DUNet [23] | 2019 | 0.9800 | 0.7963 | 0.9566 | 0.8237 |
NFN+ [41] | 2020 | 0.9813 | 0.7996 | 0.9582 | 0.8295 |
SA-UNet [21] | 2020 | 0.9840 | 0.8212 | 0.9698 | 0.8263 |
MPS-Net [42] | 2021 | 0.9740 | 0.8361 | 0.9563 | 0.8278 |
SCS-Net [24] | 2021 | 0.9838 | 0.8289 | 0.9697 | 0.8189 |
AACA-MLA-D-UNet [43] | 2022 | 0.9805 | 0.8046 | 0.9581 | 0.8303 |
Bridge-Net [44] | 2022 | 0.9818 | 0.7853 | 0.9565 | 0.8203 |
CRAUNet [45] | 2022 | - | 0.7954 | 0.9586 | 0.8302 |
SDDC-Net [46] | 2023 | 0.9808 | 0.8603 | 0.9704 | 0.8289 |
ours | 2023 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9862 | 0.8298 | 0.9712 | 0.8475 |
DUNet [23] | 2019 | 0.9752 | 0.8155 | 0.9610 | 0.7883 |
NFN+ [41] | 2020 | 0.9880 | 0.8003 | 0.9688 | 0.8369 |
SA-UNet [21] | 2020 | 0.9835 | 0.8573 | 0.9755 | 0.8153 |
SCS-Net [24] | 2021 | 0.9839 | 0.8365 | 0.9744 | - |
MPS-Net [42] | 2021 | 0.9795 | 0.8488 | 0.9668 | 0.8332 |
AACA-MLA-D-UNet [43] | 2022 | 0.9801 | 0.8402 | 0.9673 | 0.8246 |
Bridge-Net [44] | 2022 | 0.9840 | 0.8132 | 0.9667 | 0.8293 |
CRAUNet [45] | 2022 | - | 0.8259 | 0.9659 | 0.8156 |
SDDC-Net [46] | 2023 | 0.9789 | 0.8268 | 0.9669 | 0.7965 |
ours | 2023 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9820 | 0.7756 | 0.9634 | 0.7928 |
DUNet [23] | 2019 | 0.9878 | 0.7595 | 0.9641 | 0.8143 |
NFN+ [41] | 2020 | 0.9863 | 0.7963 | 0.9672 | 0.8298 |
SCS-Net [24] | 2021 | 0.9839 | 0.8207 | 0.9736 | - |
MPS-Net [42] | 2021 | 0.9819 | 0.8566 | 0.9689 | 0.8491 |
AACA-MLA-D-UNet [43] | 2022 | 0.9870 | 0.7914 | 0.9665 | 0.8276 |
Bridge-Net [44] | 2022 | 0.9864 | 0.8002 | 0.9668 | 0.8289 |
SDDC-Net [46] | 2023 | 0.9784 | 0.7988 | 0.9642 | 0.7776 |
ours | 2023 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
Database | Methods | Sp | Se | Acc | Dice |
STARE (trained on DRIVE) | MS-CANet [47] | - | - | 0.9673 | 0.7826 |
GDF-Net [48] | 0.9795 | 0.7089 | 0.9588 | - | |
AA-WGAN [31] | 0.9883 | 0.7839 | 0.9647 | - | |
ours | 0.9745 | 0.8471 | 0.9648 | 0.7855 | |
DRIVE (trained on STARE) | MS-CANet [47] | - | - | 0.9701 | 0.8035 |
GDF-Net [48] | 0.9902 | 0.7289 | 0.9593 | - | |
AA-WGAN [31] | 0.9928 | 0.8250 | 0.9635 | - | |
ours | 0.9828 | 0.8255 | 0.9690 | 0.8236 |
U-Net | ResUnet | R2U-Net | AACA-MLA-D-UNet | MRC-Net [49] | ours | |
# Param. | 4.32M | 32.61M | 39.09M | 2.03M | 0.9M | 0.6M |
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9891 | 0.9856 | 0.8375 | 0.9704 | 0.8529 |
w/o HCF | 0.989 | 0.9846 | 0.8474 | 0.9705 | 0.8549 |
w/o GAE | 0.9892 | 0.9837 | 0.8544 | 0.9705 | 0.8556 |
w/o MFF | 0.9891 | 0.9842 | 0.8473 | 0.9701 | 0.8532 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | AUC_ROC | Sp | Se | Acc | Dice |
baseline | 0.9867 | 0.9883 | 0.7876 | 0.9765 | 0.7970 |
w/o HCF | 0.9878 | 0.9885 | 0.7957 | 0.9772 | 0.8031 |
w/o GAE | 0.9878 | 0.9872 | 0.8104 | 0.9769 | 0.8040 |
w/o MFF | 0.9880 | 0.9876 | 0.8014 | 0.9767 | 0.8009 |
ours | 0.9884 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | AUC_ROC | Pr | Sp | Se | Acc | Dice |
baseline | 0.9886 | 0.9166 | 0.9932 | 0.7461 | 0.9742 | 0.8161 |
w/o HCF | 0.9895 | 0.9195 | 0.9929 | 0.7578 | 0.9748 | 0.8219 |
w/o GAE | 0.9890 | 0.9138 | 0.9915 | 0.7604 | 0.9738 | 0.8167 |
w/o MFF | 0.9882 | 0.9141 | 0.9915 | 0.7676 | 0.9743 | 0.8208 |
ours | 0.9895 | 0.9158 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
Model | AUC_ROC | Sp | Se | Acc | Dice |
MAG-Net+SE | 0.9892 | 0.9837 | 0.8517 | 0.9702 | 0.8542 |
MAG-Net+CBAM | 0.9891 | 0.9850 | 0.844 | 0.9706 | 0.8547 |
MAG-Net+GC | 0.9894 | 0.9849 | 0.8471 | 0.9708 | 0.8559 |
ours | 0.9895 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9813 | 0.7792 | 0.9556 | 0.8171 |
DUNet [23] | 2019 | 0.9800 | 0.7963 | 0.9566 | 0.8237 |
NFN+ [41] | 2020 | 0.9813 | 0.7996 | 0.9582 | 0.8295 |
SA-UNet [21] | 2020 | 0.9840 | 0.8212 | 0.9698 | 0.8263 |
MPS-Net [42] | 2021 | 0.9740 | 0.8361 | 0.9563 | 0.8278 |
SCS-Net [24] | 2021 | 0.9838 | 0.8289 | 0.9697 | 0.8189 |
AACA-MLA-D-UNet [43] | 2022 | 0.9805 | 0.8046 | 0.9581 | 0.8303 |
Bridge-Net [44] | 2022 | 0.9818 | 0.7853 | 0.9565 | 0.8203 |
CRAUNet [45] | 2022 | - | 0.7954 | 0.9586 | 0.8302 |
SDDC-Net [46] | 2023 | 0.9808 | 0.8603 | 0.9704 | 0.8289 |
ours | 2023 | 0.9836 | 0.8588 | 0.9708 | 0.8576 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9862 | 0.8298 | 0.9712 | 0.8475 |
DUNet [23] | 2019 | 0.9752 | 0.8155 | 0.9610 | 0.7883 |
NFN+ [41] | 2020 | 0.9880 | 0.8003 | 0.9688 | 0.8369 |
SA-UNet [21] | 2020 | 0.9835 | 0.8573 | 0.9755 | 0.8153 |
SCS-Net [24] | 2021 | 0.9839 | 0.8365 | 0.9744 | - |
MPS-Net [42] | 2021 | 0.9795 | 0.8488 | 0.9668 | 0.8332 |
AACA-MLA-D-UNet [43] | 2022 | 0.9801 | 0.8402 | 0.9673 | 0.8246 |
Bridge-Net [44] | 2022 | 0.9840 | 0.8132 | 0.9667 | 0.8293 |
CRAUNet [45] | 2022 | - | 0.8259 | 0.9659 | 0.8156 |
SDDC-Net [46] | 2023 | 0.9789 | 0.8268 | 0.9669 | 0.7965 |
ours | 2023 | 0.9875 | 0.8123 | 0.9773 | 0.8069 |
Model | Year | Sp | Se | Acc | Dice |
R2U-Net [40] | 2018 | 0.9820 | 0.7756 | 0.9634 | 0.7928 |
DUNet [23] | 2019 | 0.9878 | 0.7595 | 0.9641 | 0.8143 |
NFN+ [41] | 2020 | 0.9863 | 0.7963 | 0.9672 | 0.8298 |
SCS-Net [24] | 2021 | 0.9839 | 0.8207 | 0.9736 | - |
MPS-Net [42] | 2021 | 0.9819 | 0.8566 | 0.9689 | 0.8491 |
AACA-MLA-D-UNet [43] | 2022 | 0.9870 | 0.7914 | 0.9665 | 0.8276 |
Bridge-Net [44] | 2022 | 0.9864 | 0.8002 | 0.9668 | 0.8289 |
SDDC-Net [46] | 2023 | 0.9784 | 0.7988 | 0.9642 | 0.7776 |
ours | 2023 | 0.9906 | 0.7781 | 0.9743 | 0.8228 |
Database | Methods | Sp | Se | Acc | Dice |
STARE (trained on DRIVE) | MS-CANet [47] | - | - | 0.9673 | 0.7826 |
GDF-Net [48] | 0.9795 | 0.7089 | 0.9588 | - | |
AA-WGAN [31] | 0.9883 | 0.7839 | 0.9647 | - | |
ours | 0.9745 | 0.8471 | 0.9648 | 0.7855 | |
DRIVE (trained on STARE) | MS-CANet [47] | - | - | 0.9701 | 0.8035 |
GDF-Net [48] | 0.9902 | 0.7289 | 0.9593 | - | |
AA-WGAN [31] | 0.9928 | 0.8250 | 0.9635 | - | |
ours | 0.9828 | 0.8255 | 0.9690 | 0.8236 |
U-Net | ResUnet | R2U-Net | AACA-MLA-D-UNet | MRC-Net [49] | ours | |
# Param. | 4.32M | 32.61M | 39.09M | 2.03M | 0.9M | 0.6M |