
We examined whether mechanical stretch affected expression of muscle-specific microRNAs (miRNAs) that regulate proliferation (miR-133a) or differentiation (miR-1, -206). Real-time quantitative RT-PCR was used to assess miRNA regulation in the murine myoblast cell line C2C12 exposed to stretch regimens that promote either proliferation (high stretch: 17%, 1 Hz) or differentiation (moderate stretch: 10%, 0.5 Hz) after adding media that promotes differentiation. Controls consisted of myoblasts cultured under static conditions. While miRNA expression was not affected by high stretch, a significant effect of stretch (P < 0.05) was seen after 4 days with the moderate stretch regimen. All three microRNAs were upregulated by stretch, with the most significant increase for miR-1. Myoblast maturation was enhanced with a moderate stretch regimen, as assessed by a higher percentage of nuclei in straited fibers and an increase in Mef2c gene expression. Correspondingly, HDAC4 protein expression, a direct target of miR-1 and repressor of Mef2, was decreased with the moderate stretch regimen. Over-expression of miR-1 abrogated the effect of stretch on miR-1, miR-133a and miR-206 levels compared to its negative control but did not alter miR-133a or miR-206 levels. Treatment with an antisense mRNA to miR-1 similarly diminished the stretch-mediated response. Results indicate that the differential response of skeletal myoblasts to moderate and high stretch cyclic stretch regimens is due, in part, to muscle specific miRNA expression.
Citation: Caroline Rhim, William E. Kraus, George A. Truskey. Biomechanical effects on microRNA expression in skeletal muscle differentiation[J]. AIMS Bioengineering, 2020, 7(3): 147-164. doi: 10.3934/bioeng.2020014
[1] | Tong Shan, Jiayong Yan, Xiaoyao Cui, Lijian Xie . DSCA-Net: A depthwise separable convolutional neural network with attention mechanism for medical image segmentation. Mathematical Biosciences and Engineering, 2023, 20(1): 365-382. doi: 10.3934/mbe.2023017 |
[2] | Yuqing Zhang, Yutong Han, Jianxin Zhang . MAU-Net: Mixed attention U-Net for MRI brain tumor segmentation. Mathematical Biosciences and Engineering, 2023, 20(12): 20510-20527. doi: 10.3934/mbe.2023907 |
[3] | Zhigao Zeng, Cheng Huang, Wenqiu Zhu, Zhiqiang Wen, Xinpan Yuan . Flower image classification based on an improved lightweight neural network with multi-scale feature fusion and attention mechanism. Mathematical Biosciences and Engineering, 2023, 20(8): 13900-13920. doi: 10.3934/mbe.2023619 |
[4] | Jiajun Zhu, Rui Zhang, Haifei Zhang . An MRI brain tumor segmentation method based on improved U-Net. Mathematical Biosciences and Engineering, 2024, 21(1): 778-791. doi: 10.3934/mbe.2024033 |
[5] | Yanxia Sun, Xiang Li, Yuechang Liu, Zhongzheng Yuan, Jinke Wang, Changfa Shi . A lightweight dual-path cascaded network for vessel segmentation in fundus image. Mathematical Biosciences and Engineering, 2023, 20(6): 10790-10814. doi: 10.3934/mbe.2023479 |
[6] | Hong Yu, Wenhuan Lu, Qilong Sun, Haiqiang Shi, Jianguo Wei, Zhe Wang, Xiaoman Wang, Naixue Xiong . Design and analysis of a robust breast cancer diagnostic system based on multimode MR images. Mathematical Biosciences and Engineering, 2021, 18(4): 3578-3597. doi: 10.3934/mbe.2021180 |
[7] | Zijian Wang, Yaqin Zhu, Haibo Shi, Yanting Zhang, Cairong Yan . A 3D multiscale view convolutional neural network with attention for mental disease diagnosis on MRI images. Mathematical Biosciences and Engineering, 2021, 18(5): 6978-6994. doi: 10.3934/mbe.2021347 |
[8] | Ning Sheng, Dongwei Liu, Jianxia Zhang, Chao Che, Jianxin Zhang . Second-order ResU-Net for automatic MRI brain tumor segmentation. Mathematical Biosciences and Engineering, 2021, 18(5): 4943-4960. doi: 10.3934/mbe.2021251 |
[9] | Dongwei Liu, Ning Sheng, Tao He, Wei Wang, Jianxia Zhang, Jianxin Zhang . SGEResU-Net for brain tumor segmentation. Mathematical Biosciences and Engineering, 2022, 19(6): 5576-5590. doi: 10.3934/mbe.2022261 |
[10] | Xiaoguang Liu, Meng Chen, Tie Liang, Cunguang Lou, Hongrui Wang, Xiuling Liu . A lightweight double-channel depthwise separable convolutional neural network for multimodal fusion gait recognition. Mathematical Biosciences and Engineering, 2022, 19(2): 1195-1212. doi: 10.3934/mbe.2022055 |
We examined whether mechanical stretch affected expression of muscle-specific microRNAs (miRNAs) that regulate proliferation (miR-133a) or differentiation (miR-1, -206). Real-time quantitative RT-PCR was used to assess miRNA regulation in the murine myoblast cell line C2C12 exposed to stretch regimens that promote either proliferation (high stretch: 17%, 1 Hz) or differentiation (moderate stretch: 10%, 0.5 Hz) after adding media that promotes differentiation. Controls consisted of myoblasts cultured under static conditions. While miRNA expression was not affected by high stretch, a significant effect of stretch (P < 0.05) was seen after 4 days with the moderate stretch regimen. All three microRNAs were upregulated by stretch, with the most significant increase for miR-1. Myoblast maturation was enhanced with a moderate stretch regimen, as assessed by a higher percentage of nuclei in straited fibers and an increase in Mef2c gene expression. Correspondingly, HDAC4 protein expression, a direct target of miR-1 and repressor of Mef2, was decreased with the moderate stretch regimen. Over-expression of miR-1 abrogated the effect of stretch on miR-1, miR-133a and miR-206 levels compared to its negative control but did not alter miR-133a or miR-206 levels. Treatment with an antisense mRNA to miR-1 similarly diminished the stretch-mediated response. Results indicate that the differential response of skeletal myoblasts to moderate and high stretch cyclic stretch regimens is due, in part, to muscle specific miRNA expression.
Glioma is one of the most aggressive types of brain tumors, accounting for approximately 80% of all cranial malignancies, with a high recurrence, mortality rate, low cure rate and slightly predominant in males. The common clinical treatments of glioma include radiotherapy, chemotherapy and surgery, which all depend on medical imaging, processing and analysis. Among all medical imaging, Magnetic Resonance Imaging (MRI) has been a crucial technique in the diagnosis and treatment of glioma, since it cannot only clearly display the soft tissue structures based on multimodal images, but provide non-invasive and radiation-free imaging for the human body [1]. Multi-modalities show the different anatomical and functional information of glioma, which are beneficial for the depiction of the type, shape, size and boundary of the tumors, as shown in Figure 1. Accurate delineation MRI of glioma requires true labels of data, including background (label 0), necrotic and non-enhancing tumor (label 1), edema (label 2) and enhancing tumor (label 4). Physicians often delineate tumor regions manually in a slice-by-slice manner from a large amount of MRI images, which is tedious and time-consuming, and the accuracy depends highly on the experience of the experts and threshold setting. Therefore, automatic and accurate segmentation of brain tumors plays a vital role in clinical diagnosis and treatment [2].
Convolutional neural networks (CNNs) have achieved impressive achievements in brain tumor segmentation because of their powerful feature extraction capabilities. Fully Convolutional Networks (FCNs) [3] abandon the full connection layer to reduce network parameters and achieve pixel-level segmentation. Especially, since Ronneberger introduced the U-Net [4] based on FCN in 2015, encoder-decoder symmetric networks with skip connection have demonstrated exceptional performance in medical image semantic segmentation tasks [5,6,7,8]. In order to meet demands of clinical applications in terms of accuracy and efficiency, a large number of networks based on U-Net have appeared to optimize the performance by improving network structure, adding new modules, expanding to 3D networks and so on. For example, Jungo et al. enhanced the U-Net architecture by introducing residual flow and improving the pooling flow to combine global and local feature information [9]. Cinar et al. [10] proposed a hybrid architecture that combined the strengths of DenseNet121 and U-Net to achieve more accurate and precise segmentation results for small tumor regions. Wu et al. [11] introduced DE-ResUNet, which connected T1-weighted image and texture features to the dual encoder and used a channel attention mechanism to extract more useful informative features. Jiang et al. [12] proposed a dual-branch decoder network DDU-net. Çiçek et al. proposed the 3D U-Net [13], Mehta et al. [14] proposed a 3D U-Net brain tumor segmentation network and Abdollahi et al. [15] proposed the network V-Net to expand 2D into 3D network, which can extract more spatial contextual information and achieve remarkable results in the segmentation of volumetric medical image data. Zhu et al. [16] proposed a segmentation method that outperformed several state-of-the-art methods on the BraTS benchmarks by fusing deep semantic and edge information from multimodal MRI. Li et al. [17] introduced X-Net, a dual decoder structure that combined CNN and Transformer features, which showed outstanding performance in medical image segmentation tasks. Similarly, Xu et al. [18] also integrated CNN and Transformer, adopted a multi-dimensional statistical feature extraction module and obtained excellent results in brain tumor and heart segmentation tasks.
Attention mechanisms are widely regarded as crucial components in medical image processing. In MRI scans of the brain, where the tumor area is relatively small, these mechanisms play a vital role in focusing on the tumor region and extracting relevant and informative features. By leveraging various attention mechanisms, the accuracy of small tumor segmentation can be significantly improved. For example, Kong et al. [19] presented a 3D FCN with a dual-attention mechanism, incorporating both global attention and local attention, to enhance significantly the segmentation performance for glioma segmentation based on MRI. Attention U-Net [20] added attention gates (AGs) in encoder and decoder to readjust the output characteristics of encoder, a gating signal can be generated by the module to remove irrelevant and noise blurring responses and highlight the significant features transmitted through skipping connections. He et al. [21] proposed a cloud-based medical image segmentation method, which improved segmentation accuracy by introducing an interactive fusion attention module. Squeeze-and-excitation (SE*) was proposed to automatically acquire the importance of each feature channel, and then increased the weight of the relevant features in the 2018 CVPR. Roy et al. [22] introduced three SE* module extended structures in the fully convolutional neural network. Shuffle attention (SA) is a multi-branch attention mechanism [23] that combines spatial attention and channel attention simultaneously. Unlike SE*, which mainly focuses on channel attention, SA overcomes the limitations of using a single attention mechanism and incorporates global attention. Furthermore, SA offers the advantage of reducing computational complexity compared to SE*.
In the field of brain tumor segmentation, the currently existing 3D encoder-decoder networks have obtained excellent results. However, these networks often have a large number of parameters, which results in high computational cost. To maintain good segmentation efficiency while reducing computation and parameters, 3D lightweight networks are proposed for precisely segmenting brain glioma, such as MobileNets [24], ShuffleNet [25] and GhostNet [26], etc. Zhou et al. [27] presented an effective 3D residual neural network (ERV-Net) with lower computational memory consumption. Reddy et al. [28] proposed a lightweight CNN framework that extracted features from brain MRI images and then classified them as normal or abnormal. Luo et al. [29] used a lightweight and efficient hierarchical decoupled convolutional network (HDCNet), which replaced 3D convolution with a new hierarchical decoupled convolutional module, to achieve efficient exploration of spatial contextual information of images at multiple scales and perspectives. Zhang et al. [30] introduced a hierarchical multi-scale segmentation network (HMNet) with lightweight components, which further reduced the complexity and computational overhead of the network.
In the current field of brain tumor segmentation, there are still some issues. First, a large number of parameters and computational cost are required by traditional CNN networks to directly extract feature information from high-resolution brain MRI volumetric data, which can lead to difficulties in running on resource-limited devices. Second, due to the small size and diverse shapes of glioma regions, as well as the blurred boundary between them and healthy tissue, tumor area segmentation is extremely challenging. Therefore, we are committed to developing a lightweight network that can maintain high accuracy while achieving real-time performance in the brain tumor segmentation task. In this paper, a 3D lightweight network SDS-Net is proposed, which cleverly simplifies parameters and computational resources to obtain high accuracy in glioma segmentation while maintaining low computational complexity. SDS-Net demonstrates the ability to rapidly and accurately segment brain tumors, and can be easily embedded in resource-limited devices to provide a more convenient and efficient solution for clinical diagnosis and treatment. Additionally, we have developed a brain tumor segmentation and visualization software, further enhancing the convenience and efficiency of brain tumor segmentation in clinical applications. Our major contributions are given as follows:
● A novel 3D lightweight encoder-decoder network, named SDS-Net, is introduced for multimodal brain tumor MRI image segmentation. SDS-Net adopts a lightweight backbone structure and extensively explores the rich multimodal and boundary features of brain tumor images, significantly reducing network computation and parameters while maintaining excellent segmentation performance.
● The effective 3D lightweight backbone blocks, known as lightweight feature extraction (LFE) and lightweight feature fusion (LFF) modules, have been developed. The LFE module can effectively extract the local feature information of multimodal MRI images, and the LFF module combines high-level abstract features from the encoder with features from the decoder to achieve a more comprehensive feature representation. The two lightweight modules together make our proposed SDS-Net more lightweight, and enable it to exhibit efficient performance and benefit on resource-limited devices.
● To achieve more precise segmentation results, 3D SA and 3D SE modules are separately incorporated into the encoder and decoder of SDS-Net. By aggregating essential spatial and channel features from different modalities and refining the learning of edge features, the expression capability of crucial features and edges is significantly enhanced, leading to more accurate results in brain tumor segmentation.
● A segmentation and visualization system of brain MRI images is designed to assist doctors in diagnosis and subsequent treatment, which can be used in medical institutions such as hospitals and clinics to provide more accurate diagnosis and treatment plans.
The rest of the paper is structured as follows. Section 2 describes the overall architecture of the proposed network and provides detailed information about the experimental setup, including data preprocessing and evaluation indicators. Section 3 is the quantitative and qualitative analysis of the experimental results. Section 4 draws the main conclusions of this paper.
In this work, several experiments are conducted on MICCAI Brain Tumor Segmentation (BRATS) Challenge datasets [31,32,33] and the results of the proposed network are qualitatively evaluated on the segmentation and visualization system. The BRATS 2020 contains MRI images of 369 patients with different severity of gliomas and BRATS 2021 contains 1251 patients. Each patient dataset contains four modalities T1, T2, Flair and T1C, in NIFTI format, with a size of 155 × 240 × 240 and the same voxel size of 1.0 × 1.0 × 1.0 mm3. The training dataset consists of three different types of brain tumor labels, which were hand drawn by medical experts, i.e., the whole tumor (WT) (label 1, 2 and 4), the enhancing tumor (ET) (label 4) and the tumor core (TC) (labels 1 and 4).
Since there are differences in the intensity values from different patients, the z-score standardization operation is required to achieve a common image intensity level. The normalization equation (Eq (1)) can be used to calculate the normalized intensity value X∗ of the sample data, where X represents the original intensity value of the data, while μ and σ denote the mean value and standard deviation of the data, respectively.
X∗=X−μσ | (1) |
Cropping is performed at the center of the image with a size of 128 × 128 × 128 to remove irrelevant regions and address the problem of class imbalance. Four MRI modalities are fused into 4 × 128 × 128 × 128 block as the input of the network, 128 × 128 × 128 corresponds to the input volume size and the number 4 represents four MRI modalities of Flair, T1, T1ce and T2. Preprocessing steps not only eliminate the class imbalance problem but also reduce training time and memory requirements, the results of preprocessing are shown in Figure 2.
The proposed 3D lightweight network SDS-Net for brain tumor segmentation is structured as shown in Figure 3, mainly includes four key components of lightweight feature extraction (LFE) module, lightweight feature fusion (LFF) module, 3D SA module and 3D SE module. In Figure 3, the gray and gray-green blocks are 3D feature extraction blocks that consist of LFE modules with residual connection and 3D SA modules. The yellow-green and blue blocks are respectively LFF modules composed of depthwise separable convolution (DSC) merging blocks and 3D SE module. GN is used for feature normalization in the encoder and decoder, which is superior to Batch Normalization when dealing with small batch size, effectively reducing the risk of overfitting. SDS-Net combines the low-resolution feature map in the encoder with the high-resolution feature map in the decoder through skip connections to capture features at different scales, while can also help to preserve local details. The number of channels is set to 16, 32, 64,128 and 256, and the kernel size of the convolution block is 3 × 3 × 3. The SA module is employed to enhance the ability of SDS-Net to recognize the critical features, and the SE module is used to improve the fine feature expression capability of SDS-Net; we will describe the two modules in detail below. Furthermore, in order to avoid the issues of gradient vanishing and gradient explosion, the convolution operation incorporates GN and ReLU functions to accelerate the convergence speed of the training process [34]. Table 1 shows the detailed description of the SDS-Net structure. The encoder of SDS-Net extracts features by expanding the channel dimension of the feature map and decreasing the resolution. Its decoder gradually reduces the channel size and restores resolution. To effectively combine low-level detail features with high-level semantic features, a skip connection is used to establish corresponding connections and expand path feature maps.
Layer | Input | Type | Stride | Output | Params |
LFE1 | 128 × 128 × 128 × 4 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 9,456 |
Maxpool1 | 128 × 128 × 128 × 16 | 2 × 2 × 2 | 2 | 64 × 64 × 64 × 16 | 0 |
LFE 2 | 64 × 64 × 64 × 16 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 43,616 |
Maxpool2 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 2 | 32 × 32 × 32 × 32 | 0 |
SA1 | 32 × 32 × 32 × 32 | 32 × 32 × 32 × 64 | 55,456 | ||
Maxpool3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 2 | 16 × 16 × 16 × 64 | 0 |
SA2 | 16 × 16 × 16 × 64 | 16 × 16 × 16 × 128 | 221,504 | ||
Maxpool4 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 2 | 8 × 8 × 8 × 128 | 0 |
SA3 | 8 × 8 × 8 × 128 | 8 × 8 × 8 × 256 | 885,376 | ||
UpSampling1 | 8 × 8 × 8 × 256 | 2 × 2 × 2 | 16 × 16 × 16 × 256 | 0 | |
LFF1 | 16 × 16 × 16 × 256 | 3 × 3 × 3 | 1 | 16 × 16 × 16 × 128 | 106,496 |
UpSampling2 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 32 × 32 × 32 × 128 | 0 | |
LFF2 | 32 × 32 × 32 × 128 | 3 × 3 × 3 | 1 | 32 × 32 × 32 × 64 | 913,664 |
UpSampling3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 64 × 64 × 64 × 64 | 0 | |
LFF3 | 64 × 64 × 64 × 64 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 229,504 |
UpSampling4 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 128 × 128 × 128 × 32 | 0 | |
LFF4 | 128 × 128 × 128 × 32 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 57,920 |
SE | 128 × 128 × 128 × 16 | 128 × 128 × 128 × 3 | 681 | ||
Total Params | 2,523,673 |
Class imbalance is a challenging issue, especially for smaller tumors, as verified by the BRATS datasets, where the tumor area is only 1.5% of MRI images and the ET represents only 11% of the WT [35]. In order to eliminate the impact of background on segmentation results in brain MRI images and better concentrate on tumor regions, attention mechanism is introduced into medical image segmentation. The multi-branch architecture and group features have been used to build CNN networks, such as SKNet [36], ResNet [37] and AlexNet [38]. SA, a multi-branch attention mechanism, is employed in this study to enhance the performance of the SDS-Net network, specifically in the form of a 3D SA module. The feature map is divided into smaller sub-features by grouping along the channel dimension. Then, each of these sub-features is then processed separately to construct channel attention and spatial attention.
As shown in Figure 4, the 3D SA module groups the features X∈Rc×h×w×d, where c, h, w and d represent the number of channels, height, width and depth of the feature map, respectively. The feature map is first divided into several sub-features along the channel dimension, and then the spatial and channel attention mechanisms are constructed using shuffle units for each sub-feature to concatenate the spatial and channel dependencies of the features. In the channel attention mechanism, the global average pooling is used to obtain the global information of the channel. Subsequently, a linear transformation (using weights and biases) is applied, and the output is passed through the Sigmoid activation function to achieve more complex feature learning. Finally, the sub-features x'c and x's are aggregated to better extract complementary information between feature maps, and the "Channel Shuffle" [25] operation is used to communicate feature information to solve the problem of information disfluency between groups. The result of experiments shows SDS-Net can further improve the recognition ability of glioma key features by using the 3D SA module, and achieve better segmentation results of tumor sub-regions while reducing computation parameters slightly.
x'c=xc×f[FC(Fgp(xc))] | (2) |
x's=xs×f[FC(GN(xs))] | (3) |
where, xc and xs represent the channel path feature map and spatial path feature map, respectively. x'c and x's are the corresponding outputs of the two paths. Fgp denotes the average pooling operation and f is the sigmoid activation function. FC represents the linear transformation function and GN refers to the Group Normalization operation.
In order to alleviate the problem of edge detail loss by convolution, and further improve the network's fine feature expression capability, the 3D SE module [39,40] is added to the decoder. The structure is illustrated in Figure 5 and described by the following equation, primarily consisting of 1 × 1 × 1 convolution and up-sampling operation. First, the 3D SE module converts the features of different layers in the decoder into 3 channels through 1 × 1 × 1 convolution, and then the output feature map is subjected to an up-sampling operation to recover the image size. This approach helps to fuse the feature information from different layers and reduce the loss of partial information, thereby further enhancing the segmentation performance.
ˇys=Cn(ys)+Up(ˇys−1),s=1,2,3⋯X | (4) |
where ˇys represents the output of each layer in the SE module, ys denotes the output of each layer in the decoder, Cn refers to the 1 × 1 × 1 convolution with a channel size of 3, Up is the up-sampling operation and X is set to 3 in this experiment.
As shown in Figure 6(A), to reduce the parameters and computation of the network, DSC [24] and traditional convolution are combined to construct the lightweight backbone blocks, namely LFE and LFF modules. LFE module first performs two 3 × 3 × 3 convolution operations, comparing with a 5 × 5 × 5 convolution, using two stacked smaller convolutional kernels can result in less computational consumption, while keeping the same receptive field and perceptual background. DSCs with different convolution kernels are used for each channel's feature map and decompose the traditional convolution into two operations of deep convolution and point-by-point convolution. Deep convolution (DC) only applies one convolution in each channel of the input data to obtain a set of convoluted output feature maps, thus significantly reducing the number of parameters. Pointwise convolution followed DC applies a 1 × 1 × 1 convolution in the output feature graph and combines all the feature graphs of the deep convolution. The computation of each pointwise convolution with 1 × 1 × 1 kernel size is very small, resulting in a significant reduction in the complexity of the SDS-Net network. Finally, the output feature maps are fused by residual connection to enhance the reusability of feature information in the network.
The structure of the LFF module is depicted in Figure 6(B), where a 3 × 3 × 3 convolution operation is first performed, the size of the feature map remains unchanged and the number of channels is reduced by half, then the features are connected to the corresponding layer of the encoder by skip connection. The fusion of the features transmitted from the encoder effectively enhances and optimizes the local feature representation, thereby improving the ability to perceive details. The feature map is processed by multi-channels with the same DSC as the LFE module. The major advantage of the three-dimensional lightweight backbone blocks is the use of fewer parameters and computation, while introducing more local features and enhancing the network's feature representation capacity. The experimental results indicate that by incorporating the LFE and LFF modules, SDS-Net achieves precise segmentation of gliomas with limited computational memory. The specific details are shown in the following equation:
x'1=f[C2n(C2n(x1))] | (5) |
y1=f[P2n(D2n(x'1))+x'1] | (6) |
x2=Cat[f[Cn/2(x2)],y] | (7) |
y2=f[Pn(Dn(x2))] | (8) |
where, x1 and x2 are the inputs of the LFE and LFF modules, respectively, while y1 and y2 are their corresponding outputs. C, D and P represent traditional convolution, deep convolution and pointwise convolution, respectively. Additionally, n denotes the number of channels, f is the ReLU activation function and Cat indicates the concatenation of two feature maps.
The segmentation performance of SDS-Net is evaluated quantitatively by Dice coefficient and Hausdorff distance. The use of these metrics can help researchers more comprehensively and accurately evaluate the performance of segmentation networks, while providing more accurate auxiliary tools for tumor diagnosis and treatment in clinical medicine.
Dice coefficient (Eq (9)) is one of the commonly used metrics to evaluate segmentation accuracy. We compare the predicted region with the actual region, and a higher value indicates that the prediction result is more similar to the ground truth. Here, True Positive (TP), False Positive (FP) and False Negative (FN) represent the number of correctly predicted tumor pixels, non-tumor pixels in tumor tissue, and pixels incorrectly predicted as healthy tissue, respectively. Dice coefficient ranges from 0 to 1, and the value 1 indicates that the prediction is consistent with the ground truth.
Dice=2TP2TP+FP+FN | (9) |
Hausdorff distance (Eq (10)) is the similarity of the boundary between the predicted region and the ground truth, where X and Y represent the predicted tumor region and ground truth, respectively, and d(X,Y) is the Euclidean distance between x and y points on the predicted and ground truth.
d(X,Y)=max{maxx∈Xminy∈Yd(x,y),maxy∈Yminx∈Xd(x,y)} | (10) |
The proposed network SDS-Net is built in the PyTorch framework with the Adam optimizer. The learning rate is set to 0.0003 and the batch size is fixed at 1. The network is trained using an NVIDIA GeForce RTX 3060 (12G) with a training period of 200 epochs, and the early stop is set to 20 to prevent overfitting. The network is optimized using a weighted mixed loss of Binary cross-entropy loss and Dice loss.
We compare the proposed SDS-Net with the various advanced networks on the BRATS 2020 dataset, including CNN, variants of CNN combined with Transformer and other lightweight methods. The quantitative comparison experimental results are shown in Table 2. Through this comparative analysis, the lightweight networks SDS-Net in this paper has demonstrated superior performance in the task of brain tumor segmentation. Specifically, compared to traditional CNN networks [13,41], our proposed SDS-Net achieves high accuracy while having lower computational complexity, using only 2.52 M parameters and 68.18 G FLOPs. This makes SDS-Net more efficient to run on resource-limited devices. Compared to the variant of CNN combined with Transformer [42,43], SDS-Net can fully leverage the local features from multimodal and the high-quality spatial and channel information, thereby better capturing image features during the segmentation process. For TC regions located between other sub-tumor regions and the difficulty to distinguish the correct boundary, our proposed method achieves the best Dice coefficient of 88.9% in the challenging region. Compared to other lightweight methods [30,44,45], Zhang et al. [30] introduced a hierarchical multi-scale network, achieving an excellent parameter count of 0.80 M. Chen et al. [45] proposed a novel 3D dilated multi-fiber network that demonstrates remarkable computational efficiency with 27.04 G FLOPs. Although SDS-Net has slightly higher parameter and computational complexity than these two networks, our method still achieves satisfactory results in the brain tumor segmentation task. With the flexible design of the network structure, SDS-Net achieves higher Dice coefficient in the WT, ET and TC regions, with values of 92.7, 80.0 and 88.9%, respectively.
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Nuechterlein et al. [44] | 76.51 G | 3.63 M | 0.883 | 0.737 | 0.814 | 7.10 | 31.30 | 14.62 |
Zhang et al. [30] | 129.40 G | 0.80 M | 0.901 | 0.781 | 0.823 | 5.95 | 21.34 | 7.06 |
Isensee et al. [41] | 1902.15 G | 16.32 M | 0.907 | 0.814 | 0.848 | 6.94 | 5.85 | 5.07 |
Wang et al. [42] | 333.00 G | 32.99 M | 0.901 | 0.787 | 0.855 | 4.96 | 17.95 | 9.77 |
Xing et al. [43] | 71.77 G | 32.99 M | 0.920 | 0.800 | 0.864 | 4.57 | 5.27 | 5.32 |
Chen et al. [45] | 27.04 G | 3.88 M | 0.901 | 0.754 | 0.824 | − | − | − |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.927 | 0.800 | 0.889 | 17.86 | 12.93 | 11.35 |
We also compared the proposed SDS-Net with other networks on the BRATS 2021 dataset, and the detailed results of these comparative experiments are shown in Table 3. Peiris et al. [46] enhanced the generalization and robustness of the segmentation network by employing dual adversarial learning and introducing uncertain information. In comparison, the proposed SDS-Net exhibits superior performance in segmentation accuracy and uses 1/13 of its parameters. Liu et al. [40] combined residual blocks and spatial group-wise attention blocks, achieving the best Dice coefficients in ET and TC segmentation tasks, with 83.3 and 86.8%, respectively. Jiang et al. [47] proposed a novel medical image segmentation network SwinBTS combined with Transformer, which showed excellent performance in the brain tumor segmentation task, achieving 91.8% Dice score on the WT region segmentation. In our experiments, we carefully considered the balance between network performance and resource consumption. SDS-Net achieved Dice scores of 91.8, 82.5 and 86.8% in WT, ET and TC, respectively. In addition, the proposed network uses 2.52 M parameters and 68.18 G FLOPs. Therefore, the lightweight network proposed in this paper has more prominent advantages overall, achieving better segmentation results with fewer parameters and computational complexity.
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Çiçek et al. [13] | 1902.15 G | 16.32 M | 0.880 | 0.757 | 0.782 | 25.03 | 14.11 | 13.76 |
Liu et al. [40] | − | − | 0.916 | 0.833 | 0.868 | 5.94 | 19.28 | 7.57 |
Peiris et al. [46] | 776.11 G | 19.17 M | 0.908 | 0.814 | 0.854 | 5.37 | 21.83 | 8.56 |
Jiang et al. [47] | − | − | 0.918 | 0.832 | 0.848 | 3.65 | 16.03 | 14.51 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
We perform ablation experiments to evaluate the results of different SDS-Net modules on the BRATS 2021 dataset. The U-Net network structure has shown promising results in brain tumor segmentation. First, we use the modified U-Net as the benchmark network, where the encoder includes 5 convolution modules and 4 max pooling layers for feature extraction and resolution reduction. The decoder consists of 4 de-convolution modules and 4 tri-linear interpolation up-sampling operations to restore the edge information of the image. To fairly show the efficiency of each SDS-Net module, the SDS-Net has the same input size, network layers number and training parameters as the benchmark network. We gradually add feature extraction module SE, attention mechanism SA module and lightweight backbone blocks to the benchmark network, and then compare the computational complexity and segmentation precision through ablation experiments.
The results are illustrated in Table 4. First, to improve the feature fusion in the decoder, we add the 3D SE module to the benchmark. As can be seen from the Table 4, the Dice coefficient of WT, ET and TC is increased by 0.4, 0.3 and 0.9%, respectively, while the Hausdorff distance is decreased by 0.47, 1.65 and 1.13 mm, respectively. The evaluation metrics in the three tumor regions are all improved, but the computational cost of the network is slightly increased. Second, in order to reduce the computation burden, the lightweight 3D SA module is also added to the benchmark network, the last three convolutional modules in the encoder are replaced with triple 3D SA modules, which not only reduce the number of parameters, but also enhance the network's ability to extract key features, resulting in the optimal Dice coefficient and the lowest Hausdorff distance for the WT and ET regions. The experiments show that replacing by the 3D SA modules can light the FLOPs to 190.27 G and greatly reduce computation parameters to 4.80 M, while maintaining the segmentation precision. Finally, the lightweight blocks with the combination of DSCs and traditional convolution, two LFE modules are adopted to the encoder and four LFF modules are used to the decoder, respectively, which further achieve the more lightweight network with parameter amount 2.52 M and computational cost 68.18 G FLOPs. Compared with the benchmark network, our SDS-Net has significant improvements in all metrics of the tumor regions, with Dice coefficient enhancing by 0.6, 0.7 and 1.2% for WT, ET and TC, respectively, while the Hausdorff distance is decreased by 0.92, 1.74 and 2.47 mm for WT, ET and TC, respectively.
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Baseline | 202.89 G | 9.45 M | 0.912 | 0.818 | 0.856 | 21.99 | 14.87 | 14.46 |
Baseline + SE | 202.95 G | 9.45 M | 0.916 | 0.821 | 0.865 | 21.52 | 13.22 | 13.33 |
Baseline + SE+ SA | 190.27 G | 4.80 M | 0.918 | 0.830 | 0.861 | 20.02 | 15.13 | 14.74 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Figure 7 shows the results of ablation experiments for different modules, and the Dice coefficients are shown using boxplots and bar graph. From the figure, it can be observed that the proposed SDS-Net presents fewer instances of abnormal and scattered data, indicating that it has high stability and consistency. Although the performance of SDS-Net in the ET region is slightly lower than baseline + SE + SA method, overall, it shows significant effectiveness in improving segmentation accuracy. To show the segmentation effect more clearly, Figure 8 shows the visual segmentation results of the ablation experiment. In the second row, we can observe that the baseline network exhibits limited performance in accurately segmenting the individual edema region in the upper right corner. However, with the gradual introduction of functional modules into the baseline network, resulting in SDS-Net, the segmentation performance for the individual region gradually approaches the ground truth. Therefore, this further verifies the effectiveness of each model and demonstrates the performance of our SDS-Net.
To evaluate the effectiveness of different attention mechanisms in SDS-Net, ablation experiments are performed on the BRATS 2021 dataset. Three different attention mechanisms, including the CBAM module, the 3D SE* module and the 3D SA module, are explored and the performances are compared in the segmentation tasks. Table 5 illustrates the results of the ablation experiments. From Table 5, 3D CBAM obtained the second place in Hausdorff distance of the WT region, but it does not perform well in other metrics. In the ET region, the 3D SE* achieved the optimal results, with a highest Dice coefficient of 82.7% and a lowest Hausdorff distance of 13.00. 3D SA almost outperformed the other two different attention mechanisms in all metrics, except for a slightly lower metric of the ET region than 3D SE*, the Dice coefficients of WT, ET and TC brain tumor regions are 91.8, 82.5 and 86.8% respectively, and the optimal Hausdorff distance is achieved in the WT and TC regions. Thus, we choose 3D SA as the attention module for SDS-Net encoder. Figure 9 visualizes the segmentation results obtained by different attention mechanisms of the network SDS-Net, the visualization shows the results obtained from 3D SA are closer to the ground truth than CBAM and 3D SE*. The 3D SA module incorporates an ultra-lightweight attention mechanism that combines spatial attention and channel attention; unfortunately, there is no notable change of network parameters and computational cost comparing to the other two attention mechanisms.
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
SDS-Net + CBAM | 68.18 G | 2.53 M | 0.901 | 0.798 | 0.845 | 22.03 | 13.52 | 13.43 |
SDS-Net + SE* | 68.18 G | 2.53 M | 0.915 | 0.827 | 0.859 | 23.95 | 13.00 | 13.62 |
SDS-Net + SA | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
The Binary cross-entropy (BCE) loss (Eq (11)) is commonly used in binary classification tasks. This loss function is calculated on a pixel-by-pixel basis, comparing the probability of the actual voxel value (Pi) with the predicted value (Gi). When Pi is equal to 1, the loss function is log (Gi); when Pi is equal to 0, the loss function is log (1−Gi), where N is the number of voxels.
BCE=−[∑Ni=1PilogGi+(1−Pi)log(1−Gi)] | (11) |
Another frequently used loss function (Eq (12)) is the Dice loss function, which aims to optimize the similarity between the actual voxel value (Pi) and the predicted mask value (Gi).
Dice=1−2∑Ni=1PiGi∑Ni=1Pi+∑Ni=1Gi | (12) |
In this experiment, the BCE-Dice loss is taken to be the loss function (Eq (13)), which is a weighted hybrid of the Binary cross-entropy and Dice loss function. The convergence of the loss function is an important indicator for measuring the performance of CNNs. Faster convergence speed cannot only significantly reduce the training complexity but also enhance the segmentation performance. Figure 10 shows the convergence trends of the loss functions for 3D U-Net, Benchmark network and our SDS-Net trained on BRATS 2021 dataset. We can observe that SDS-Net converges faster comparing to other networks.
BCEDice=0.5BCE+Dice | (13) |
The common evaluation metrics to measure network performance and complexity are FLOPs and the number of parameters, respectively. Table 6 presents the number of parameters and FLOPs for the 3D U-Net, Benchmark network and SDS-Net with the same brain image size of 128 × 128 × 128. SDS-Net achieves 2.52 M parameters and 68.18 G FLOPs, while, 3D U-Net and Benchmark network get 16.32, 9.45 M parameters and 1902.15,202.89 G FLOPs, respectively. In comparison, our SDS-Net uses fewer parameters and lower computational cost. The number of parameters and FLOPs of the 3D conventional convolution can be calculated as Eqs (14) and (15).
Parameters=Cin×Cout×kH×kw×kD+Cout | (14) |
FLOPs=2×Oh×Ow×Od×Parameters | (15) |
The number of parameters and FLOPs for the 3D DSC can be calculated as Eqs (16) and (17).
Parameters∗=(Cin×kH×kw×kD+1)+(Cin+1)×Cout | (16) |
FLOPs∗=2×Oh×Ow×Od×Parameters∗ | (17) |
where Cin and Cout represent the number of input and output channels of the feature map respectively. The kernel size is kH×kw×kD. The depth, height and width of the output feature map are Oh, Ow and Od, respectively.
Network | Parameters(M) | FLOPs(G) |
3D U-Net | 16.32 M | 1902.15 G |
Baseline | 9.45 M | 202.89 G |
SDS-Net (Ours) | 2.52 M | 68.18 G |
CNN has become one of the most popular researches in data processing and analysis, such as image registration, segmentation and fusion. Especially in the medical image processing, deep learning technology has been widely used in the automatic detection and diagnosis of diseases, among which brain tumor segmentation is a crucial task. CNN typically requires large amounts of parameters and computational resources, which may pose challenges in medical devices with limited computing resources. Therefore, lightweight CNN has become increasingly vital to medical data application. In this study, a lightweight segmentation network is proposed that not only achieves better segmentation results with fewer parameters and computational complexity, but can also be applied to medical devices to assist doctors in tumor diagnosis and treatment plans.
The segmentation and visualization system consists of the image acquisition module, image pre-processing module, 3D image segmentation module and image visualization module. Figure 11 shows the interface of the brain tumor segmentation and visualization system, which includes information of the patient, 2D slices of all MRI modalities and the segmentation results of the tumor area. Three feature types are drawn with different colors to differentiate various labels, while the visualization module can calculate the area size of the segmentation result and fuse the tumor area with the original image. The system is developed by PyCharm 2021.2.2 and PyQt5 with Windows 10, 16 GB memory and RTX3060 GPU, which can be used in medical institutions such as hospitals and clinics. 3D image segmentation module loads the lightweight segmentation network SDS-Net, so it can work stably, fast, accurately and conveniently display the segmentation results of brain MRI in the same group.
In this paper, we introduce SDS-Net, a lightweight and efficient 3D network designed for brain tumor MRI segmentation, which consists of four key components of LFE module, LFF module, 3D SA module and 3D SE module. Encoder block consists of LFE module with residual connection and 3D SA module, decoder block with DSC is composed of LFF module and 3D SE module. Specially, SA is placed in the encoder and SE fine feature expression is added to the decoder of SDS-Net; by the operations, SDS-Net can efficiently concentrate on the region of interest in the images and extract the information rich image features. SDS-Net combines DSC and traditional convolution in the LFE and LFF modules, which further helps to decrease the network parameters and computer cost, making the network more lightweight and efficient. GN for feature normalization and skip connection is also used in the encoder and decoder. In addition, BCE-Dice loss function helps to classify pixels more accurately and makes the network converge faster. Finally, ablation experiments explore the effect of the addition of different modules to SDS-Net, including using SE module, lightweight backbone blocks and replacing the last three feature extraction modules by 3D SA modules. Moreover, we compare SDS-Net with the advanced approaches on the BRATS datasets, and through quantitative analysis and visual comparison of SDS-Net segmentation results with the labels, it can be found that SDS-Net has better computational efficiency, lower parameter amount and the best segmentation performance.
However, this study also has certain limitations: First, while multi-modal image data is currently processed using early fusion methods, further research is needed to determine whether the later fusion of multi-modal images is superior. Second, in the future, we plan to migrate the proposed method to the field of ischemic stroke lesion segmentation.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This research was funded by the Anhui University Scientific Research Project under Grant (2022AH050660).
The authors declare there is no conflict of interest.
[1] | Burkholder TJ, Lieber RL (2001) Sarcomere length operating range of vertebrate muscles during movement. J Exp Biol 204: 1529-1536. |
[2] |
Zhang SJ, Truskey GA, Kraus WE (2007) Effect of cyclic stretch on β1D-integrin expression and activation of FAK and RhoA. Am J Physiol Cell Physiol 292: C2057-2069. doi: 10.1152/ajpcell.00493.2006
![]() |
[3] | Vandenburgh HH (1992) Mechanical forces and their second messengers in stimulating cell growth in vitro. Am J Physiol 262: R350-R355. |
[4] |
Vandenburgh HH, Karlisch P (1989) Longitudinal growth of skeletal myotubes in vitro in a new horizontal mechanical cell stimulator. In Vitro Cell Dev Biol 25: 607-616. doi: 10.1007/BF02623630
![]() |
[5] |
Kumar A, Murphy R, Robinson P, et al. (2004) Cyclic mechanical strain inhibits skeletal myogenesis through activation of focal adhesion kinase, Rac-1 GTPase, and NF-κB transcription factor. FASEB J 18: 1524-1535. doi: 10.1096/fj.04-2414com
![]() |
[6] |
Rauch C, Loughna PT (2005) Static stretch promotes MEF2A nuclear translocation and expression of neonatal myosin heavy chain in C2C12 myocytes in a calcineurin- and p38-dependent manner. Am J Physiol Cell Physiol 288: C593-C605. doi: 10.1152/ajpcell.00346.2004
![]() |
[7] | Kook SH, Lee HJ, Chung WT, et al. (2008) Cyclic mechanical stretch stimulates the proliferation of C2C12 myoblasts and inhibits their differentiation via prolonged activation of p38 MAPK. Mol Cells 25: 479-486. |
[8] |
Zhao X, Ito A, Kane CD, et al. (2001) The modular nature of histone deacetylase HDAC4 confers phosphorylation-dependent intracellular trafficking. J Biol Chem 276: 35042-35048. doi: 10.1074/jbc.M105086200
![]() |
[9] |
Zhou X, Richon VM, Wang AH, et al. (2000) Histone deacetylase 4 associates with extracellular signal-regulated kinases 1 and 2, and its cellular localization is regulated by oncogenic Ras. Proc Natl Acad Sci USA 97: 14329-14333. doi: 10.1073/pnas.250494697
![]() |
[10] |
McKinsey TA, Zhang CL, Lu J, et al. (2000) Signal-dependent nuclear export of a histone deacetylase regulates muscle differentiation. Nature 408: 106-111. doi: 10.1038/35040593
![]() |
[11] |
Miska EA, Langley E, Wolf D, et al. (2001) Differential localization of HDAC4 orchestrates muscle differentiation. Nucleic Acids Res 29: 3439-3447. doi: 10.1093/nar/29.16.3439
![]() |
[12] |
Chen JF, Mandel EM, Thomson JM, et al. (2006) The role of microRNA-1 and microRNA-133 in skeletal muscle proliferation and differentiation. Nat Genet 38: 228-233. doi: 10.1038/ng1725
![]() |
[13] |
Rao PK, Kumar RM, Farkhondeh M, et al. (2006) Myogenic factors that regulate expression of muscle-specific microRNAs. Proc Natl Acad Sci USA 103: 8721-8726. doi: 10.1073/pnas.0602831103
![]() |
[14] |
Kim HK, Lee YS, Sivaprasad U, et al. (2006) Muscle-specific microRNA miR-206 promotes muscle differentiation. J Cell Biol 174: 677-687. doi: 10.1083/jcb.200603008
![]() |
[15] |
Baskerville S, Bartel DP (2005) Microarray profiling of microRNAs reveals frequent coexpression with neighboring miRNAs and host genes. RNA 11: 241-247. doi: 10.1261/rna.7240905
![]() |
[16] |
Wienholds E, Plasterk RHA (2005) MicroRNA function in animal development. FEBS Lett 579: 5911-5922. doi: 10.1016/j.febslet.2005.07.070
![]() |
[17] |
Lagos-Quintana M, Rauhut R, Yalcin A, et al. (2002) Identification of tissue-specific microRNAs from mouse. Curr Biol 12: 735-739. doi: 10.1016/S0960-9822(02)00809-6
![]() |
[18] |
Sempere LF, Freemantle S, Pitha-Rowe I, et al. (2004) Expression profiling of mammalian microRNAs uncovers a subset of brain-expressed microRNAs with possible roles in murine and human neuronal differentiation. Genome Biol 5: R13. doi: 10.1186/gb-2004-5-3-r13
![]() |
[19] |
Ma K, Chan JKL, Zhu G, et al. (2005) Myocyte enhancer factor 2 acetylation by p300 enhances its DNA binding activity, transcriptional activity, and myogenic differentiation. Mol Cell Biol 25: 3575-3582. doi: 10.1128/MCB.25.9.3575-3582.2005
![]() |
[20] |
Wang DZ (2006) Micro or mega: how important are MicroRNAs in muscle? Cell Cycle 5: 1015-1016. doi: 10.4161/cc.5.10.2742
![]() |
[21] |
McCarthy JJ (2008) MicroRNA-206: the skeletal muscle-specific myomiR. Biochim Biophys Acta 1779: 682-691. doi: 10.1016/j.bbagrm.2008.03.001
![]() |
[22] |
Yuasa K, Hagiwara Y, Ando M, et al. (2008) MicroRNA-206 is highly expressed in newly formed muscle fibers: implications regarding potential for muscle regeneration and maturation in muscular dystrophy. Cell Struct Funct 33: 163-169. doi: 10.1247/csf.08022
![]() |
[23] |
Cheng CS, El-Abd Y, Bui K, et al. (2014) Conditions that promote primary human skeletal myoblast culture and muscle differentiation in vitro. Am J Physiol Cell Physiol 306: C385-C395. doi: 10.1152/ajpcell.00179.2013
![]() |
[24] |
Kuang W, Tan J, Duan Y, et al. (2009) Cyclic stretch induced miR-146a upregulation delays C2C12 myogenic differentiation through inhibition of Numb. Biochem Bioph Res Commun 378: 259-263. doi: 10.1016/j.bbrc.2008.11.041
![]() |
[25] |
McCarthy JJ, Esser KA (2007) MicroRNA-1 and microRNA-133a expression are decreased during skeletal muscle hypertrophy. J Appl Physiol 102: 306-313. doi: 10.1152/japplphysiol.00932.2006
![]() |
[26] |
Care A, Catalucci D, Felicetti F, et al. (2007) MicroRNA-133 controls cardiac hypertrophy. Nat Med 13: 613-618. doi: 10.1038/nm1582
![]() |
[27] |
Carson JA, Wei L (2000) Integrin signaling's potential for mediating gene expression in hypertrophying skeletal muscle. J Appl Physiol 88: 337-343. doi: 10.1152/jappl.2000.88.1.337
![]() |
[28] |
Blau HM, Pavlath GK, Hardeman EC, et al. (1985) Plasticity of the differentiated state. Science 230: 758-766. doi: 10.1126/science.2414846
![]() |
[29] |
Livak KJ, Schmittgen TD (2001) Analysis of relative gene expression data using real-time quantitative PCR and the 2−ΔΔCT method. Methods 25: 402-408. doi: 10.1006/meth.2001.1262
![]() |
[30] | Zar JH (1999) Biostatistical Analysis. Upper Saddle River NJ: Prentice Hall. |
[31] |
Feng Y, Cao JH, Li XY, et al. (2011) Inhibition of miR-214 expression represses proliferation and differentiation of C2C12 myoblasts. Cell Biochem Funct 29: 378-383. doi: 10.1002/cbf.1760
![]() |
[32] |
Wei X, Li H, Zhang B, et al. (2016) miR-378a-3p promotes differentiation and inhibits proliferation of myoblasts by targeting HDAC4 in skeletal muscle development. RNA Biol 13: 1300-1309. doi: 10.1080/15476286.2016.1239008
![]() |
[33] |
Liu L, Li TM, Liu XR, et al. (2019) MicroRNA-140 inhibits skeletal muscle glycolysis and atrophy in endotoxin-induced sepsis in mice via the WNT signaling pathway. Am J Physiol Cell Physiol 317: C189-C199. doi: 10.1152/ajpcell.00419.2018
![]() |
[34] |
Su Y, Yu Y, Liu C, et al. (2020) Fate decision of satellite cell differentiation and self-renewal by miR-31-IL34 axis. Cell Death Differ 27: 949-965. doi: 10.1038/s41418-019-0390-x
![]() |
[35] |
Ge Y, Sun Y, Chen J (2011) IGF-II is regulated by microRNA-125b in skeletal myogenesis. J Cell Biol 192: 69-81. doi: 10.1083/jcb.201007165
![]() |
[36] |
McCarthy JJ, Esser KA, Andrade FH (2007) MicroRNA-206 is overexpressed in the diaphragm but not the hindlimb muscle of mdx mouse. Am J Physiol Cell Physiol 293: C451-C457. doi: 10.1152/ajpcell.00077.2007
![]() |
[37] |
Farh KKH, Grimson A, Jan C, et al. (2005) The widespread impact of mammalian MicroRNAs on mRNA repression and evolution. Science 310: 1817-1821. doi: 10.1126/science.1121158
![]() |
[38] |
Sayed D, Hong C, Chen IY, et al. (2007) MicroRNAs play an essential role in the development of cardiac hypertrophy. Circ Res 100: 416-424. doi: 10.1161/01.RES.0000257913.42552.23
![]() |
[39] |
Van Rooij E, Olson EN (2007) MicroRNAs: powerful new regualtors of heart disease and provocative therapeutic targets. J Clin Invest 117: 2369-2376. doi: 10.1172/JCI33099
![]() |
[40] |
Van Rooij E, Sutherland LB, Liu N, et al. (2006) A signature pattern of stress-responsive microRNAs that can evoke cardiac hypertrophy and heart failure. Proc Natl Acad Sci USA 103: 18255-18260. doi: 10.1073/pnas.0608791103
![]() |
[41] |
Yang B, Lin H, Xiao J, et al. (2007) The muscle-specific microRNA miR-1 regulates cardiac arrhythmogenic potential by targeting GJA1 and KCNJ2. Nat Med 13: 486-491. doi: 10.1038/nm1569
![]() |
[42] |
Potthoff MJ, Olson EN (2007) MEF2: a central regulator of diverse developmental programs. Development 134: 4131-4140. doi: 10.1242/dev.008367
![]() |
[43] |
Chan JKL, Sun L, Yang XJ, et al. (2003) Functional characterization of an amino-terminal region of HDAC4 that possesses MEF2 binding and transcriptional repressive activity. J Biol Chem 278: 23515-234521. doi: 10.1074/jbc.M301922200
![]() |
[44] |
Liu N, Williams AH, Kim Y, et al. (2007) An intragenic MEF2-dependent enhancer directs muscle-specific expression of microRNAs 1 and 133. Proc Natl Acad Sci USA 104: 20844-20849. doi: 10.1073/pnas.0710558105
![]() |
[45] |
Zhao Y, Samal E, Srivastava D (2005) Serum response factor regulates a muscle-specific microRNA that targets Hand2 during cardiogenesis. Nature 436: 214-220. doi: 10.1038/nature03817
![]() |
[46] |
Van Rooij E, Liu N, Olson EN (2008) MicroRNAs flex their muscles. Trends Genet 24: 159-166. doi: 10.1016/j.tig.2008.01.007
![]() |
![]() |
![]() |
1. | Bin Guo, Ning Cao, Ruihao Zhang, Peng Yang, GETNet: Group Normalization Shuffle and Enhanced Channel Self-Attention Network Based on VT-UNet for Brain Tumor Segmentation, 2024, 14, 2075-4418, 1257, 10.3390/diagnostics14121257 | |
2. | Novsheena Rasool, Javaid Iqbal Bhat, A Critical Review on Segmentation of Glioma Brain Tumor and Prediction of Overall Survival, 2024, 1134-3060, 10.1007/s11831-024-10188-2 | |
3. | Hao Luo, Dongmei Zhou, Yongjian Cheng, Siqi Wang, MPEDA-Net: A lightweight brain tumor segmentation network using multi-perspective extraction and dense attention, 2024, 91, 17468094, 106054, 10.1016/j.bspc.2024.106054 | |
4. | Bin Guo, Ning Cao, Ruihao Zhang, Peng Yang, SCENet: Small Kernel Convolution with Effective Receptive Field Network for Brain Tumor Segmentation, 2024, 14, 2076-3417, 11365, 10.3390/app142311365 | |
5. | Ruihao Zhang, Peng Yang, Can Hu, Bin Guo, MEASegNet: 3D U-Net with Multiple Efficient Attention for Segmentation of Brain Tumor Images, 2025, 15, 2076-3417, 3791, 10.3390/app15073791 | |
6. | Peng Yang, Ruihao Zhang, Can Hu, Bin Guo, GMDNet: Grouped Encoder-Mixer-Decoder Architecture Based on the Role of Modalities for Brain Tumor MRI Image Segmentation, 2025, 14, 2079-9292, 1658, 10.3390/electronics14081658 | |
7. | Junjie Li, Hong Peng, Bing Li, Zhicai Liu, Rikong Lugu, Bingyan He, Global–Local Feature Fusion Network Based on Nonlinear Spiking Neural Convolutional Model for MRI Brain Tumor Segmentation, 2025, 0129-0657, 10.1142/S0129065725500364 | |
8. | Lesia Mochurad, Comparative analysis of audio-MAE and MAE-AST models for real-time audio classification, 2025, 66, 0005-1144, 503, 10.1080/00051144.2025.2504749 |
Layer | Input | Type | Stride | Output | Params |
LFE1 | 128 × 128 × 128 × 4 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 9,456 |
Maxpool1 | 128 × 128 × 128 × 16 | 2 × 2 × 2 | 2 | 64 × 64 × 64 × 16 | 0 |
LFE 2 | 64 × 64 × 64 × 16 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 43,616 |
Maxpool2 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 2 | 32 × 32 × 32 × 32 | 0 |
SA1 | 32 × 32 × 32 × 32 | 32 × 32 × 32 × 64 | 55,456 | ||
Maxpool3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 2 | 16 × 16 × 16 × 64 | 0 |
SA2 | 16 × 16 × 16 × 64 | 16 × 16 × 16 × 128 | 221,504 | ||
Maxpool4 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 2 | 8 × 8 × 8 × 128 | 0 |
SA3 | 8 × 8 × 8 × 128 | 8 × 8 × 8 × 256 | 885,376 | ||
UpSampling1 | 8 × 8 × 8 × 256 | 2 × 2 × 2 | 16 × 16 × 16 × 256 | 0 | |
LFF1 | 16 × 16 × 16 × 256 | 3 × 3 × 3 | 1 | 16 × 16 × 16 × 128 | 106,496 |
UpSampling2 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 32 × 32 × 32 × 128 | 0 | |
LFF2 | 32 × 32 × 32 × 128 | 3 × 3 × 3 | 1 | 32 × 32 × 32 × 64 | 913,664 |
UpSampling3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 64 × 64 × 64 × 64 | 0 | |
LFF3 | 64 × 64 × 64 × 64 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 229,504 |
UpSampling4 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 128 × 128 × 128 × 32 | 0 | |
LFF4 | 128 × 128 × 128 × 32 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 57,920 |
SE | 128 × 128 × 128 × 16 | 128 × 128 × 128 × 3 | 681 | ||
Total Params | 2,523,673 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Nuechterlein et al. [44] | 76.51 G | 3.63 M | 0.883 | 0.737 | 0.814 | 7.10 | 31.30 | 14.62 |
Zhang et al. [30] | 129.40 G | 0.80 M | 0.901 | 0.781 | 0.823 | 5.95 | 21.34 | 7.06 |
Isensee et al. [41] | 1902.15 G | 16.32 M | 0.907 | 0.814 | 0.848 | 6.94 | 5.85 | 5.07 |
Wang et al. [42] | 333.00 G | 32.99 M | 0.901 | 0.787 | 0.855 | 4.96 | 17.95 | 9.77 |
Xing et al. [43] | 71.77 G | 32.99 M | 0.920 | 0.800 | 0.864 | 4.57 | 5.27 | 5.32 |
Chen et al. [45] | 27.04 G | 3.88 M | 0.901 | 0.754 | 0.824 | − | − | − |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.927 | 0.800 | 0.889 | 17.86 | 12.93 | 11.35 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Çiçek et al. [13] | 1902.15 G | 16.32 M | 0.880 | 0.757 | 0.782 | 25.03 | 14.11 | 13.76 |
Liu et al. [40] | − | − | 0.916 | 0.833 | 0.868 | 5.94 | 19.28 | 7.57 |
Peiris et al. [46] | 776.11 G | 19.17 M | 0.908 | 0.814 | 0.854 | 5.37 | 21.83 | 8.56 |
Jiang et al. [47] | − | − | 0.918 | 0.832 | 0.848 | 3.65 | 16.03 | 14.51 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Baseline | 202.89 G | 9.45 M | 0.912 | 0.818 | 0.856 | 21.99 | 14.87 | 14.46 |
Baseline + SE | 202.95 G | 9.45 M | 0.916 | 0.821 | 0.865 | 21.52 | 13.22 | 13.33 |
Baseline + SE+ SA | 190.27 G | 4.80 M | 0.918 | 0.830 | 0.861 | 20.02 | 15.13 | 14.74 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
SDS-Net + CBAM | 68.18 G | 2.53 M | 0.901 | 0.798 | 0.845 | 22.03 | 13.52 | 13.43 |
SDS-Net + SE* | 68.18 G | 2.53 M | 0.915 | 0.827 | 0.859 | 23.95 | 13.00 | 13.62 |
SDS-Net + SA | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Network | Parameters(M) | FLOPs(G) |
3D U-Net | 16.32 M | 1902.15 G |
Baseline | 9.45 M | 202.89 G |
SDS-Net (Ours) | 2.52 M | 68.18 G |
Layer | Input | Type | Stride | Output | Params |
LFE1 | 128 × 128 × 128 × 4 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 9,456 |
Maxpool1 | 128 × 128 × 128 × 16 | 2 × 2 × 2 | 2 | 64 × 64 × 64 × 16 | 0 |
LFE 2 | 64 × 64 × 64 × 16 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 43,616 |
Maxpool2 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 2 | 32 × 32 × 32 × 32 | 0 |
SA1 | 32 × 32 × 32 × 32 | 32 × 32 × 32 × 64 | 55,456 | ||
Maxpool3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 2 | 16 × 16 × 16 × 64 | 0 |
SA2 | 16 × 16 × 16 × 64 | 16 × 16 × 16 × 128 | 221,504 | ||
Maxpool4 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 2 | 8 × 8 × 8 × 128 | 0 |
SA3 | 8 × 8 × 8 × 128 | 8 × 8 × 8 × 256 | 885,376 | ||
UpSampling1 | 8 × 8 × 8 × 256 | 2 × 2 × 2 | 16 × 16 × 16 × 256 | 0 | |
LFF1 | 16 × 16 × 16 × 256 | 3 × 3 × 3 | 1 | 16 × 16 × 16 × 128 | 106,496 |
UpSampling2 | 16 × 16 × 16 × 128 | 2 × 2 × 2 | 32 × 32 × 32 × 128 | 0 | |
LFF2 | 32 × 32 × 32 × 128 | 3 × 3 × 3 | 1 | 32 × 32 × 32 × 64 | 913,664 |
UpSampling3 | 32 × 32 × 32 × 64 | 2 × 2 × 2 | 64 × 64 × 64 × 64 | 0 | |
LFF3 | 64 × 64 × 64 × 64 | 3 × 3 × 3 | 1 | 64 × 64 × 64 × 32 | 229,504 |
UpSampling4 | 64 × 64 × 64 × 32 | 2 × 2 × 2 | 128 × 128 × 128 × 32 | 0 | |
LFF4 | 128 × 128 × 128 × 32 | 3 × 3 × 3 | 1 | 128 × 128 × 128 × 16 | 57,920 |
SE | 128 × 128 × 128 × 16 | 128 × 128 × 128 × 3 | 681 | ||
Total Params | 2,523,673 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Nuechterlein et al. [44] | 76.51 G | 3.63 M | 0.883 | 0.737 | 0.814 | 7.10 | 31.30 | 14.62 |
Zhang et al. [30] | 129.40 G | 0.80 M | 0.901 | 0.781 | 0.823 | 5.95 | 21.34 | 7.06 |
Isensee et al. [41] | 1902.15 G | 16.32 M | 0.907 | 0.814 | 0.848 | 6.94 | 5.85 | 5.07 |
Wang et al. [42] | 333.00 G | 32.99 M | 0.901 | 0.787 | 0.855 | 4.96 | 17.95 | 9.77 |
Xing et al. [43] | 71.77 G | 32.99 M | 0.920 | 0.800 | 0.864 | 4.57 | 5.27 | 5.32 |
Chen et al. [45] | 27.04 G | 3.88 M | 0.901 | 0.754 | 0.824 | − | − | − |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.927 | 0.800 | 0.889 | 17.86 | 12.93 | 11.35 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Çiçek et al. [13] | 1902.15 G | 16.32 M | 0.880 | 0.757 | 0.782 | 25.03 | 14.11 | 13.76 |
Liu et al. [40] | − | − | 0.916 | 0.833 | 0.868 | 5.94 | 19.28 | 7.57 |
Peiris et al. [46] | 776.11 G | 19.17 M | 0.908 | 0.814 | 0.854 | 5.37 | 21.83 | 8.56 |
Jiang et al. [47] | − | − | 0.918 | 0.832 | 0.848 | 3.65 | 16.03 | 14.51 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
Baseline | 202.89 G | 9.45 M | 0.912 | 0.818 | 0.856 | 21.99 | 14.87 | 14.46 |
Baseline + SE | 202.95 G | 9.45 M | 0.916 | 0.821 | 0.865 | 21.52 | 13.22 | 13.33 |
Baseline + SE+ SA | 190.27 G | 4.80 M | 0.918 | 0.830 | 0.861 | 20.02 | 15.13 | 14.74 |
SDS-Net (Ours) | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Methods | FLOPs | Parameters | Dice | Hausdorff | ||||
WT | ET | TC | WT | ET | TC | |||
SDS-Net + CBAM | 68.18 G | 2.53 M | 0.901 | 0.798 | 0.845 | 22.03 | 13.52 | 13.43 |
SDS-Net + SE* | 68.18 G | 2.53 M | 0.915 | 0.827 | 0.859 | 23.95 | 13.00 | 13.62 |
SDS-Net + SA | 68.18 G | 2.52 M | 0.918 | 0.825 | 0.868 | 21.07 | 13.13 | 11.99 |
Network | Parameters(M) | FLOPs(G) |
3D U-Net | 16.32 M | 1902.15 G |
Baseline | 9.45 M | 202.89 G |
SDS-Net (Ours) | 2.52 M | 68.18 G |