
St. Martin's Island was declared an ecologically critical area of Bangladesh in 1999, but this has had limited effect on the conservation of the island's natural coral resources, on which a thriving tourism industry and the local inhabitants depend. The introduction of a tourism entrance fee can benefit conservation management on the island, but research on the amount that tourists are willing to pay is absent. The objective of this paper is to determine an appropriate entrance fee amount tourists would be willing to pay (WTP) for visiting St. Martin's Island using contingent valuation method questionnaire surveys and interviews of tourists on the island (n = 327) and the factors that influence their decision. Significance testing and regression analysis were used to assess survey data. A large majority of respondents suggested that they would be willing to pay between 0.78 and 7.8 USD; however, 24.5% said that they would pay nothing and indicated that such reluctance to pay was based on a belief that the responsibility should not fall on themselves as individuals, rather than a lack of financial capacity. Evidence suggests that even greater tourism entrance fees would still be accepted and amenable to tourists. If a fee of 4.29 USD was introduced, between 350,000 and 3.51 million USD, or 1.93 million USD, could be generated annually. The level of education, income, and a general concern for the environment significantly influenced WTP amounts. This study is aimed at assisting policy decision-makers and conservation managers of St. Martin's Island; required policy actions are briefly discussed.
Citation: Seema Rani, Michael Bennett, Md. Kawser Ahmed, Xiongzhi Xue, Keliang Chen, Mohammad Shamsul Alam, Antaya March, Pierre Failler. Potential of establishing a tourism entrance fee for the conservation management of St. Martin's Island, Bangladesh[J]. Green Finance, 2025, 7(1): 1-23. doi: 10.3934/GF.2025001
[1] | Bandar Abdullah Aloyaydi, Subbarayan Sivasankaran, Hany Rizk Ammar . Influence of infill density on microstructure and flexural behavior of 3D printed PLA thermoplastic parts processed by fusion deposition modeling. AIMS Materials Science, 2019, 6(6): 1033-1048. doi: 10.3934/matersci.2019.6.1033 |
[2] | K.A. Lazopoulos, A.K. Lazopoulos . Beam bending and Λ-fractional analysis. AIMS Materials Science, 2023, 10(4): 604-617. doi: 10.3934/matersci.2023034 |
[3] | Guadalupe Sanchez-Olivares, Fausto Calderas, Antonio Sanchez-Solis, Luis Medina-Torres, Leonardo R. Moreno, Octavio Manero . Assessment of extrusion-sonication process on flame retardant polypropylene by rheological characterization. AIMS Materials Science, 2016, 3(2): 620-633. doi: 10.3934/matersci.2016.2.620 |
[4] | Marek Konieczny . Mechanical properties and failure analysis of laminated magnesium-intermetallic composites. AIMS Materials Science, 2022, 9(4): 572-583. doi: 10.3934/matersci.2022034 |
[5] | Kiyotaka Obunai, Daisuke Mikami, Tadao Fukuta, Koichi Ozaki . Microstructure and mechanical properties of newly developed SiC-C/C composites under atmospheric conditions. AIMS Materials Science, 2018, 5(3): 494-507. doi: 10.3934/matersci.2018.3.494 |
[6] | R. Daulath Banu, R. Karunanithi, S. Sivasankaran, B. Subramanian, Abdullah A. Alhomidan . Influence of graphene nanoplatelets (GNPs) and aluminum-carbon layered double hydroxides (Al-C LDH) in polypropylene matrix of hybrid composite structures on the microstructure and mechanical performances. AIMS Materials Science, 2024, 11(5): 882-917. doi: 10.3934/matersci.2024043 |
[7] | Mohammed O. Atteaa Al-hassany, Ali Al-Dulaimy, Amir Al-Sammarraie, Abed Fares Ali . Effect of fiberglass form on the tensile and bending characteristic of epoxy composite material. AIMS Materials Science, 2020, 7(5): 583-595. doi: 10.3934/matersci.2020.5.583 |
[8] | Hanh C. Nguyen, Shigeru Nagasawa, Kensei Kaneko . Strength estimation of silicon nitride ceramics using a round-notched specimen subjected to shearing-tool indentation. AIMS Materials Science, 2020, 7(5): 518-533. doi: 10.3934/matersci.2020.5.518 |
[9] | Bhavik Ardeshana, Umang Jani, Ajay Patel, Anand Joshi . An approach to modelling and simulation of single-walled carbon nanocones for sensing applications. AIMS Materials Science, 2017, 4(4): 1010-1028. doi: 10.3934/matersci.2017.4.1010 |
[10] | Zulzamri Salleh, Md Mainul Islam, Jayantha Ananda Epaarachchi, Haibin Su . Mechanical properties of sandwich composite made of syntactic foam core and GFRP skins. AIMS Materials Science, 2016, 3(4): 1704-1727. doi: 10.3934/matersci.2016.4.1704 |
St. Martin's Island was declared an ecologically critical area of Bangladesh in 1999, but this has had limited effect on the conservation of the island's natural coral resources, on which a thriving tourism industry and the local inhabitants depend. The introduction of a tourism entrance fee can benefit conservation management on the island, but research on the amount that tourists are willing to pay is absent. The objective of this paper is to determine an appropriate entrance fee amount tourists would be willing to pay (WTP) for visiting St. Martin's Island using contingent valuation method questionnaire surveys and interviews of tourists on the island (n = 327) and the factors that influence their decision. Significance testing and regression analysis were used to assess survey data. A large majority of respondents suggested that they would be willing to pay between 0.78 and 7.8 USD; however, 24.5% said that they would pay nothing and indicated that such reluctance to pay was based on a belief that the responsibility should not fall on themselves as individuals, rather than a lack of financial capacity. Evidence suggests that even greater tourism entrance fees would still be accepted and amenable to tourists. If a fee of 4.29 USD was introduced, between 350,000 and 3.51 million USD, or 1.93 million USD, could be generated annually. The level of education, income, and a general concern for the environment significantly influenced WTP amounts. This study is aimed at assisting policy decision-makers and conservation managers of St. Martin's Island; required policy actions are briefly discussed.
Brain tumors are abnormal cell growths located in or near brain tissue that damage the nervous system, causing symptoms such as headaches, dizziness, dementia, seizures, and other neurological signs [1]. Magnetic resonance imaging (MRI)—including T1-weighted (T1), post-contrast T1-weighted (T1CE), T2-weighted (T2), and fluid-attenuated inversion recovery (FLAIR) sequences—is a prevalent diagnostic tool for brain tumors due to its sensitivity to soft tissue and high image contrast, as shown in Figure 1. Physicians utilize MRI for lesion diagnosis, but accuracy can be hindered by factors such as fatigue and emotional state. Automated methods have garnered extensive attention in the medical field due to their capability to objectively and accurately analyze imaging information.
Most multimodal approaches assume complete data availability; however, in reality, missing modalities are common. As illustrated in Figure 2, various missing scenarios can occur during both training and inference stages. The absence of certain MRI sequences may fail to capture tumor characteristics, thereby limiting a comprehensive understanding of the tumor [2]. Therefore, it is crucial for multimodal learning methods to maintain robustness in the presence of missing modalities during inference.
Currently, a prevalent approach to tackle segmentation arising from missing modality is knowledge distillation [3,4], where information is transferred from a teacher-student network to recover missing data, but this can be computationally intensive. Another method is image synthesis [5], leveraging generative models to reconstruct the missing data. However, synthetic images may introduce noise to the task. Additionally, mapping available modalities into a common latent subspace aims to compensate for or recover the missing information [6,7,8]. However, existing approaches often require training multiple sets of parameters to address various missing modality scenarios, thereby escalating the model's complexity and computational overhead.
With the expansion of data scale and enhancement of computational resources, researchers favor general neural networks for diverse tasks, minimizing the need for task-specific model design and training. Recently, transformer [9] has shown great potential in natural language processing, visual recognition, intensive prediction. However, its complex architecture and high computational demands limit comprehensive fine-tuning for downstream tasks, especially accurate segmentation, potentially leading to overfitting and reduced generalization ability.
Inspired by recent advancements in prompt learning [10,11,12] and efficient fine-tuning techniques [13,14,15], we introduce a novel brain tumor segmentation framework, called DPONet. This framework employs an encoder-decoder structure for the segmentation network, enhancing performance in both incomplete and complete modality scenarios. Specifically, we leverage image frequency information as frequency filtering prompt (FFP) to facilitate the pre-trained model in extracting discriminative features. Furthermore, by learning a series of spatial perturbation prompt (SPP), we map these discriminative features into a common latent space, mitigating the challenges of modality fusion in the decoder. Finally, we validate the robustness of our approach on two commonly used public datasets. To sum up, our main contributions are as follows:
● We propose a new framework for incomplete-modal image segmentation that effectively handles common cases of missing modalities. This approach requires only 7% of the trainable parameters to adjust the pre-trained model, thereby avoiding the heavy fine-tuning typically necessary for transformers.
● We introduce a frequency filtering prompt to extract spatial frequency components from images. This method addresses the model's oversight of target domain features and enhances its adaptation to brain tumor datasets.
● We propose a spatial perturbation prompt that incorporates learnable parameters into a spatial modulation model. This aims to achieve consistent multimodal feature embeddings even in the presence of missing partial modalities.
Incomplete multimodal learning refers to scenarios in multimodal learning tasks where partial modality information is missing or incomplete. This issue becomes particularly prominent in brain tumor segmentation tasks, where medical imaging data is typically composed of multiple MRI sequences. The absence of one modality results in the challenge of incomplete modality information learning. Many studies [16,17,18] are devoted to solving this problem, demonstrating impressive performance in various incomplete multimodal learning tasks. Zhou et al. [16] showed that there exists a certain correlation within the latent representations of modalities, which can be utilized to describe missing modalities by calculating the correlation between modalities in a latent space. Ting et al. [17] combines available modality information to estimate the latent features of missing modalities. Liu et al. [18] explicitly considers the relationship between modalities and regions, giving different attention to different modalities for each region. However, these models require full fine-tuning of the pre-trained model, which increases the computational burden and leads to a decrease in generalization ability.
The task of most neural networks is to learn the optimal points in functions. Fourier Transform establishes the transformation relationship between the function in the spatial domain and the frequency domain, so that we can analyze a function by the frequency component to approximate the objective function more effectively [19]. The frequency of an image represents the intensity of gray change in the image. Fourier transform analyzes the features by analyzing the coefficients of each frequency component [20]. The performance of computer vision models is significantly affected by the Fourier statistical properties of the training data and show a certain sensitivity to the Fourier basis direction, and their robustness can be improved by learning this sensitivity [21]. For example, Fang et al. [22] and Xu et al. [23] argued that different parts of the same organ in MRI images exhibit regularity and that high-frequency structural information can more effectively capture these similarities and regularities.
Prompt learning is an effective transfer learning approach in natural language processing [10,24,25], which fine-tunes pre-trained models on source tasks by embedding contextual prompts. Recently, prompts have also been employed in computer vision tasks [26,27,28] and multimodal learning tasks [11,29,30], introducing self-adaptation in the input space to optimize the target task. For instance, Jia et al. [26] proposed the Pyramid Vision Transformer model (PVT), achieving downstream performance comparable to full fine-tuning by adding a small number of learnable prompt embeddings on the patch embedding. Different from the PVT model, Bahng et al. [27] further proposed a method to learn a single disturbance to adjust the pixel space and affect the model output. These studies suggest that continuously adjusting and optimizing prompts can enhance the adaptability of model. Lee et al. [29] treats different scenarios of missing modalities as different types of inputs and employs learnable prompts to guide the predictions of model under various missing conditions. Qiu et al. [30] utilizes an intermediate classifier to generate a prompt for each missing scenario based on intermediate features for segmentation prediction. The difference is that our work does not require learning a set of prompts for each missing scenario but aims to learn generic visual prompts and generalize them to modulate feature space in missing scenes.
In this paper, we focus on brain tumor segmentation under common missing modality scenarios. We simulate real-world data incompleteness by assuming absences of one or multiple modalities (Figure 2). Additionally, due to the difficulty of fully training a pre-trained transformer with limited computational resources, we design a discriminative prompt optimization network that avoids fine-tuning the entire pre-trained model. In this section, we will elaborate on the framework and its components.
The pyramid vision transformer (PVT) [31] introduces a progressive shrinking strategy within the transformer block to control the scale of feature maps for dense prediction tasks. We chose the backbone is initialized with the weights pre-trained on ImageNet. PVT comprises four stages, each consisting of a patch embedding layer and l transformer encoder layers, which generate feature maps of different scales. Given an input image X∈RH×W×C, the patch embedding layer divides the sample X into HWpi non-overlapping patches, where pi represents the patch size of the i-th layer. As the stage progresses, the patch size decreases accordingly. The flattened patches are then fed into a linear projection to obtain embedded patches. The embedded patches, along with positional embedding information, are subsequently input into the transformer encoder to produce a feature map x of size Hpi×Wpi×C.This process can be described as follows:
xl=MLP(LN(SRA(xl−1))), | (3.1) |
where xl−1 represents the feature map output from the previous layer, SRA(⋅) denotes the spatial reduction attention proposed in PVT, and LN(⋅) and MLP(⋅) refer to normalization and multi-layer perceptron operations, respectively. SRA is like multi-head attention. The formula is as follows:
SQA=Attention(QWQ,SRA(K)WK,SR(V)WV), | (3.2) |
where WQ, WK, and WV are the parameters of the linear projections. SRA(⋅) is used to reduce the spatial dimension. This can be expressed as:
SRA(x)=LN(Reshape(xi,ri)WS), | (3.3) |
The ri represents the feature map reduction rate for stage i.
The Reshape(⋅) operation reshapes the input x∈Rhi×wi×ci to hiwir2i×(r2ici)). The WS is a linear projection that reduces the dimensionality of the input. The attention calculation is as follows:
Attention(q,k,v)=Softmax(qkT√d)v, | (3.4) |
where q, k and v are the query, key, and value transform matrices, and d is the dimension.
We consider a multimodal dataset consisting of N(N=4) modalities, M= FLAIR, T1CE, T1 and T2. The dataset is denoted as D=D14,D13,…,Di,…,D0, where D14 represents the complete set of modalities, and other sets represent missing modalities subsets, such as D0=XF0,XT1c0,XT10,XT21 indicating only T2 mode is available. Xmk represents the input sample, where m denotes the modality type, and k represents the modal state. For the model, it is unaware of which specific modality is missing, therefore, we introduce placeholder values (set to 0) to assign to the missing modality data XF0,XT1c0,XT10,XT20 to ensure the format of the multimodal input.
We propose a novel discriminative prompt optimization network, as shown in Figure 3, which aims to provide natural insertion points for intermediate features of the network while preserving the integrity of the pre-trained model and enabling fine-tuning for downstream tasks. We adapt a pre-trained transformer as feature extractor and keep it frozen during training. Multimodal images D={Xkm}k=[0,1] are fed into four extractors, and task-relevant information is aggregated through discriminative prompts to fully exploiting the discriminative features. Next, a spatial perturbation prompt module is introduced, which hierarchically fuses the discriminative features of available modalities and maps them to a shared feature representation space to learn cross-modal shared information. Furthermore, the fused features are mapped back to the original input size through up-sampling in the decoder, and the resulting segmentation masks are obtained from these feature maps. Notably, during training, the trainable parameters are confined to the prompt components and the decoder.
The frequency filtering prompt method, as illustrated in Figure 4, utilizes Fourier transform to extract frequency features and jointly modulates the intermediate features with image embeddings. The frequency processing method decomposes images into different frequency components, which are distributed across different spatial locations of the image, encouraging the model to focus on critical information of the image [21]. The core idea is to remodulate the intermediate features using frequency domain information, shifting the distribution from the pre-trained dataset to the target dataset. Furthermore, since there may be commonalities between features of different modalities, even if image data from a particular modality is missing, the remaining modalities still contain corresponding frequency information, which enhances the robustness of the information to a certain extent. Taking a single branch as an example, for a given image, we apply the fast Fourier transform (FFT) along the spatial dimension to obtain frequency components corresponding to different spatial locations. FFT is applied to each channel to convert the spatial domain representation into a frequency representation in the frequency domain, and filtering operations are performed in the frequency domain. Then, an attention mask is learned in the frequency domain to analyze the dominant frequency components in the feature map. Finally, the feature representation is transformed back to the spatial domain using inverse FFT (iFFT). The transformation from the spatial domain to the frequency domain is expressed as follows:
F(x)(μ,υ)=H−1∑h=0W−1∑w=0x(h,w)e−i2π(hμH+wυW), | (3.5) |
After obtaining the frequency representation, different frequency components are modulated by filtering through the attention mechanism. Specifically, the attention mechanism compresses information across channels through convolution and a sigmoid function. The expression of the frequency filtering mechanism is as follows:
F′(x)=Fx⊗σ(conv([AvgPool(Fx),Maxpool(Fx)])), | (3.6) |
where, σdenotes the Sigmoid function, AvgPool(⋅) and MaxPool(⋅) represent the average pooling and max pooling operations respectively.
Finally, the inverse FFT is used to transform back to the spatial domain features:
x′(h,w)=1H⋅WH−1∑h=0W−1∑w=0F′(x)ei2π(hμH+wυW), | (3.7) |
Inspired by AdaptFormer [32], we employ a frequency enhancement adaptor, a bottleneck structure that limits the number of parameters. It takes the combination of filtered frequency features and image features as input and generates relevant frequency prompts through a down-projection layer, a lightweight multi-layer perceptron, and an up-projection layer. Formally, this process can be expressed as:
pif=MLPup(GELU(MLPidown(x′+x))), | (3.8) |
Thirdly, the generated prompts are appended to the transformer layers to facilitate the model in learning more representative and discriminative image features.
To enable the model to handle missing modalities, we employ null values for filling, however, such null values are likely to disturb the feature space and result in failure of modal feature fusion. Therefore, we propose learnable spatial perturbation prompts, as show in Figure 5, aiming to learn a task-specific visual prompt (P) within a latent space that encourages the sharing of cross-modal information. Prompts interact dynamically with input features, facilitating adaptive modal fusion rather than simply injecting fixed information using learning prompts.
First, the extracted discriminative features are concatenated through element-wise addition fic=[fif,fit1c,fit1,fit2] and then passed through a 3 × 3 convolutional layer followed by a Sigmoid activation function to generate prompt weights ωi∈[0,1]. These weights describe the importance of each spatial data point in the input. Inspired by EVP [27], we add random visual embeddings of the same size as the transformer tokens, train only these random embeddings in the training phase, and the trained visual prompts as the guidance for the model, denoted as Fi=(Fitoken,pim). The process can be described as:
ωi=σ(conv([fif,fit1c,fit1,fit2])), | (3.9) |
pim=conv(N∑c=1ωipic), | (3.10) |
Fi=transformer(fic+pim), | (3.11) |
where, σ is the Sigmoid function. Finally, the cross-modal information features (F) are fed into Transformer encoder block to establish cross-modal long-range dependencies.
We introduce a consistency loss to optimize the prompts to capture task-shared knowledge and transform it into representations that are beneficial for the task. Specifically, we map the feature maps obtained from the transformer encoder stages to the same size as the input image and use mean squared error ensuring that the model learns coherent and consistent information at each stage. Note that, since shallower layers may lack sufficient semantic information, we apply the consistency loss only in the last two stages of the transformer encoder.
Lm=1NN∑i=1M∑m=1(ˆfi−fmi)2, | (3.12) |
where, N is the number of samples, M is the number of decoder layers, and the rescaled features of images in transformer layer m, and their average is denoted as ˆfi=1m∑mk=1fki.
In addition, we mapped the feature map into a segmentation map, and calculated Dice loss from the ground truth to prompt the model capture consistent feature representations.
Ld=1NN∑i=1M∑m=1Dice(yi−f(xim)), | (3.13) |
where, yi denotes the ground-truth labels of the image xi, and f(xim) denotes the prediction corresponding to the m-th layer features of the image.
The feature consistency loss and prediction consistency loss are combined to supervise prompt generation.
Lc=γLm+(1−γ)Ld, | (3.14) |
where, γ is the weight parameter used to balance the two losses. We experiment with different values of γ and found that γ=0.3 gives the best result.
The convolutional decoder gradually restores the spatial resolution from the fused features to the original segmentation space. The convolutional decoder employs skip connections to merge features from different modalities at specific hierarchical levels into the encoder, to preserve more low-level details. Therefore, the overall processing steps are as follows:
Di=conv(upsample(conv(fic,Di−1))), | (3.15) |
where Di is the feature map from the i-th layer of the convolutional decoder, and fic is the combined feature from multiple encoder layers.
We employ a hybrid loss to measure the difference between the predictions and the ground truth. Dice Loss is used to calculate the similarity between the predicted segmentation result and the true segmentation result. Cross-entropy loss measures the prediction performance by quantifying the difference between the predicted probability distribution and the true probability distribution. Gradients are calculated based on the feedback of the sum of the two losses to update the parameters. The definition is as follows:
LDice=−2∑Niyif(xi)∑Niyi+∑Nif(xi), | (3.16) |
LCE=−∑Niyilogp(f(xi)), | (3.17) |
where f(xi) and yi represent the prediction and ground-truth labels, respectively. Besides, N is the number of pixels, p(⋅) is the SoftMax of the prediction. Last, our hybid loss function Lseg can be given by
Lseg=Lc+LDice+LCE, | (3.18) |
We use two public datasets from the Multimodal Brain Tumor Segmentation Challenge (BraTS) to demonstrate the effectiveness of the proposed method, BraTS 2018 and BraTS 2020 [33,34,35]. BraTS 2018 contains 285 cases of patients for training, while BraTS 2020 includes 369 cases for training and 125 for validation. In these datasets, each case comprises four MRI modalities: Flair, T1ce, T1, and T2. The volume of each modality is 240 × 240 × 155, aligned within the same spatial space. Medical experts provide manual pixel-level annotations of three mutually inclusive tumor regions in each image, namely, whole tumor (WT), tumor core (TC), and enhancing tumor (ET). WT encompasses all tumor tissues, while TC comprises ET, necrosis, and non-enhancing tumor core.
Data preprocessing is performed on the two datasets before training. For each dataset, we slice along the axial plane of the 3D medical images. To eliminate non-informative slices and irrelevant background regions, thereby saving training efficiency and time, we use central slices as the training data and reshape each 2D slice to 224 × 224. We design a simulation method for missing modalities. The MRI modalities are randomly removed from the input. The missing modality can be any one or multiple modalities, and the missing rate for each modality is random. The purpose of this is to simulate the scenario where missing modalities may occur in real-world situations.
In this study, our method is implemented in Pytorch utilizing a single NVIDIA Tesla V100 32 GB GPU. We adopt the U-Net architecture composed of transformer blocks as the benchmark, and the transformer is pre-trained on ImageNet-1K. We utilize the SGD optimizer with an initial learning rate of 0.01. After many experiments and parameter tuning, we set our model to train 100 epochs with an initial learning rate of 1e−2 and a batch size of 12. For the segmentation task, we use the Dice coefficient (which computes the similarity of two sets), the Hausdorff distance (HD95, which measures the distance between two sets), and the sensitivity (the ratio of the number of positive samples correctly identified by the model to the number of all true positive samples) as performance metrics to evaluate various methods.
We focus on exploring the robustness of discriminative optimization networks to general incompleteness in multimodal image without fine-tuning the entire pretraind model. In this chapter, we first introduce the excellent results obtained by our method. Subsequently, a series of ablation experiments on the proposed components. Considering that the BraTS 2020 dataset contains many patient cases and is representative, we experimented with it in the ablation study.
As shown in Table 1, our method achieves remarkable performance in Dice score on both the modality-complete and modality-missing scenarios. For example, our proposed approach has significantly better mean Dice scores for whole tumors, tumor cores, and enhanced tumors than suboptimal approaches. From the experimental results in Table 2, we observed that the baseline model generally exhibited unsatisfactory performance on the T1 modality. However, our model achieved significant improvements in this aspect, effectively enhancing the performance under the T1 modality. In Figures 6 and 7, we present the visualization of segmentation results. Furthermore, Table 3 clearly exhibits that our method outperforms other approaches in terms of HD95 and sensitivity under complete modality testing, further validating the superior performance of our approach.
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | D | Z | T | Q | Our | D | Z | T | Q | Our | D | Z | T | Q | Our |
✓ | 86.1 | 86.1 | 86.5 | 86.7 | 93.9 | 71.0 | 70.9 | 71.5 | 71.0 | 93.3 | 46.3 | 46.3 | 45.6 | 47.2 | 76.1 | |||
✓ | 76.8 | 78.5 | 77.4 | 79.5 | 91.6 | 81.5 | 84.0 | 83.4 | 84.3 | 95.3 | 74.9 | 80.1 | 78.9 | 81.4 | 88.4 | |||
✓ | 77.2 | 78.0 | 78.1 | 79.5 | 89.1 | 66.0 | 65.9 | 66.8 | 67.7 | 91.9 | 37.3 | 38.0 | 41.3 | 39.1 | 71.6 | |||
✓ | 87.3 | 87.4 | 89.1 | 86.9 | 95.2 | 69.2 | 68.8 | 69.3 | 69.9 | 93.5 | 38.2 | 42.4 | 43.6 | 42.8 | 74.6 | |||
✓ | ✓ | 87.7 | 87.8 | 88.4 | 88.4 | 94.5 | 83.5 | 84.8 | 86.4 | 86.3 | 95.8 | 75.9 | 79.4 | 81.7 | 80.1 | 88.9 | ||
✓ | ✓ | 81.1 | 81.8 | 81.2 | 83.1 | 92.1 | 83.4 | 83.6 | 85.2 | 85.8 | 95.4 | 78.0 | 80.1 | 79.2 | 81.7 | 88.3 | ||
✓ | ✓ | 89.7 | 89.8 | 89.9 | 89.8 | 95.5 | 73.1 | 73.8 | 73.9 | 74.4 | 94.3 | 41.0 | 45.9 | 48.2 | 46.8 | 77.3 | ||
✓ | ✓ | 87.7 | 87.8 | 88.0 | 87.9 | 94.4 | 73.1 | 73.4 | 73.3 | 72.9 | 94.1 | 45.7 | 46.8 | 50.1 | 47.3 | 77.5 | ||
✓ | ✓ | 89.9 | 89.9 | 90.5 | 90.1 | 95.5 | 74.1 | 74.6 | 75.5 | 74.5 | 94.1 | 49.3 | 48.6 | 48.6 | 49.5 | 76.6 | ||
✓ | ✓ | 89.9 | 89.3 | 90.0 | 90.0 | 95.6 | 84.7 | 84.8 | 85.5 | 86.6 | 95.9 | 76.7 | 81.9 | 81.8 | 81.2 | 88.9 | ||
✓ | ✓ | ✓ | 90.7 | 90.1 | 90.7 | 90.6 | 95.6 | 85.1 | 85.2 | 86.5 | 86.7 | 95.8 | 76.8 | 82.1 | 81.8 | 81.8 | 88.8 | |
✓ | ✓ | ✓ | 90.6 | 90.6 | 90.3 | 90.6 | 95.7 | 75.2 | 75.6 | 75.9 | 75.8 | 94.7 | 49.9 | 50.3 | 52.5 | 51.1 | 78.0 | |
✓ | ✓ | ✓ | 90.7 | 90.4 | 90.6 | 90.8 | 95.8 | 85.0 | 85.3 | 86.4 | 86.4 | 96.0 | 77.1 | 78.7 | 81.0 | 80.0 | 88.9 | |
✓ | ✓ | ✓ | 88.3 | 88.2 | 88.7 | 88.9 | 94.6 | 83.5 | 84.2 | 86.5 | 86.5 | 95.8 | 77.0 | 79.3 | 78.5 | 82.1 | 88.9 | |
✓ | ✓ | ✓ | ✓ | 91.1 | 90.6 | 90.6 | 91.0 | 95.9 | 85.2 | 84.6 | 87.4 | 86.4 | 95.9 | 78.0 | 79.9 | 81.6 | 81.0 | 88.9 |
Average | 87.0 | 87.1 | 87.3 | 87.6 | 94.3 | 78.2 | 78.6 | 79.6 | 79.7 | 94.8 | 61.5 | 64.0 | 64.9 | 64.9 | 82.8 |
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | Z | Y | T | L | Our | Z | Y | T | L | Our | Z | Y | T | L | Our |
✓ | 81.2 | 76.3 | 86.6 | 84.8 | 94.3 | 64.2 | 56.7 | 68.8 | 69.4 | 94.4 | 43.1 | 16.0 | 41.4 | 47.6 | 76.2 | |||
✓ | 72.2 | 42.8 | 77.8 | 75.8 | 92.6 | 75.4 | 65.1 | 81.5 | 82.9 | 95.4 | 72.6 | 66.3 | 75.7 | 73.7 | 89.2 | |||
✓ | 67.5 | 15.5 | 78.7 | 74.4 | 90.9 | 56.6 | 16.8 | 65.6 | 66.1 | 93.2 | 32.5 | 8.1 | 44.5 | 37.1 | 74.7 | |||
✓ | 86.1 | 84.2 | 88.4 | 88.7 | 95.2 | 61.2 | 47.3 | 66.7 | 66.4 | 94.2 | 39.3 | 8.1 | 40.5 | 35.6 | 74.8 | |||
✓ | ✓ | 83.0 | 84.1 | 88.2 | 86.3 | 95.0 | 78.6 | 80.3 | 84.8 | 84.2 | 96.1 | 74.5 | 68.7 | 77.7 | 75.3 | 90.0 | ||
✓ | ✓ | 74.4 | 62.1 | 81.8 | 77.2 | 93.1 | 78.6 | 78.2 | 83.5 | 83.4 | 95.7 | 74.0 | 70.7 | 77.1 | 74.7 | 89.5 | ||
✓ | ✓ | 87.1 | 87.3 | 89.7 | 89.0 | 95.6 | 65.9 | 61.6 | 72.0 | 70.8 | 95.2 | 43.0 | 9.5 | 44.4 | 41.2 | 77.9 | ||
✓ | ✓ | 82.2 | 84.2 | 88.4 | 88.7 | 94.9 | 61.2 | 47.3 | 66.7 | 66.4 | 95.1 | 45.0 | 16.5 | 47.7 | 48.7 | 77.7 | ||
✓ | ✓ | 87.6 | 87.9 | 90.3 | 89.9 | 95.9 | 69.8 | 62.6 | 71.8 | 70.9 | 95.1 | 47.5 | 17.4 | 48.3 | 45.4 | 78.1 | ||
✓ | ✓ | 87.1 | 87.5 | 89.5 | 89.7 | 95.6 | 77.9 | 80.8 | 84.8 | 84.4 | 96.1 | 75.1 | 64.8 | 76.8 | 75.0 | 90.0 | ||
✓ | ✓ | ✓ | 87.3 | 87.7 | 90.4 | 88.9 | 95.7 | 79.8 | 80.9 | 85.2 | 84.1 | 96.2 | 75.5 | 65.7 | 77.4 | 74.0 | 90.0 | |
✓ | ✓ | ✓ | 87.8 | 88.4 | 89.7 | 89.9 | 96.0 | 71.5 | 63.7 | 74.1 | 72.7 | 95.5 | 47.7 | 19.4 | 50.0 | 44.8 | 78.7 | |
✓ | ✓ | ✓ | 88.1 | 88.8 | 90.6 | 90.4 | 96.0 | 79.6 | 80.7 | 85.8 | 84.6 | 96.3 | 75.7 | 66.4 | 76.6 | 73.8 | 90.1 | |
✓ | ✓ | ✓ | 82.7 | 80.9 | 88.4 | 86.1 | 95.1 | 80.4 | 79.0 | 85.8 | 84.4 | 96.2 | 74.8 | 68.3 | 78.5 | 75.4 | 90.1 | |
✓ | ✓ | ✓ | ✓ | 89.6 | 88.8 | 90.6 | 90.1 | 96.1 | 85.8 | 80.1 | 85.9 | 84.5 | 96.3 | 77.6 | 68.4 | 80.4 | 75.5 | 90.0 |
Average | 82.9 | 76.4 | 87.3 | 86.0 | 94.8 | 72.4 | 65.4 | 77.5 | 77.0 | 95.4 | 59.9 | 42.3 | 62.5 | 59.9 | 83.8 |
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
Ding et al. | 86.13 | 71.93 | 58.98 | 72.35 | - | - | - | - | - | - | - | - |
Zhang et al. | 87.08 | 78.69 | 64.08 | 76.62 | 2.90 | 6.21 | 44.64 | 17.92 | 99.60 | 99.81 | 99.82 | 99.74 |
Ting et al. | 90.71 | 84.60 | 79.07 | 84.79 | 4.05 | 5.78 | 33.77 | 14.53 | 90.98 | 83.90 | 77.68 | 84.18 |
Qiu et al. | 87.58 | 79.67 | 64.87 | 77.37 | 2.82 | 5.71 | 43.92 | 17.48 | 99.66 | 99.83 | 99.81 | 99.77 |
baseline(fine-tune) | 77.63 | 78.94 | 70.85 | 93.56 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline(frozen) | 58.11 | 61.09 | 40.88 | 89.16 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
our | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
We further conducted experiments to analyze the robustness of our proposed method to varying missing modality rates between the training and testing phases. As shown in Figure 8(a), we trained the model with a 70% missing rate and randomly removed multiple modalities to simulate modality missing scenarios for testing. We found that, compared to the baseline, our DPONet method was robust to different missing rates during testing. Moreover, in Figure 8(b), where we used 10%, 70%, and 90% to represent the degree of missingness during training (through many experiments, we found that these missing rates are representative), we observed that when training with more complete modality data, the performance was significantly higher when testing with low missing rates. In this paper, the experiments based on the general reality that collecting complete modality data cannot be guaranteed. However, there are still some publicly available datasets with complete modalities. Therefore, we trained the models using complete data, as shown in Figure 8(c), where the baseline model could not handle data missing, our method consistently improved upon the baseline.
We explored the effects of frequency filtering prompts and spatial perturbation prompts, the results showing in Table 4, our method achieved a higher Dice score of 93.02. The term baseline (fine-tune) refers to a pre-trained transformer that is comprehensively fine-tuned on the BraTS dataset. The term baseline (frozen) refers to a baseline model where the pre-trained backbone parameters are frozen.
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
baseline (fine-tune) | 77.63 | 78.94 | 70.85 | 75.81 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline (frozen) | 58.11 | 61.09 | 40.88 | 53.36 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
baseline + FFP | 93.65 | 92.40 | 85.08 | 90.38 | 2.45 | 2.04 | 2.16 | 2.22 | 96.54 | 96.11 | 91.26 | 94.64 |
baseline + SPP | 94.56 | 94.40 | 87.37 | 92.11 | 2.47 | 2.05 | 2.22 | 2.25 | 96.59 | 96.07 | 90.53 | 94.40 |
baseline + FFP + SPP | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
We introduced frequency filtering prompts into the baseline model, the model achieved comparable performance to fine-tuned model, demonstrating the efficiency of proposed component. Furthermore, as shown in Figure 9, during training with complete modalities, when a significant portion of modalities were absent during inference (i.e., retaining only one modality), the baseline model suffered a severe performance degradation. Excitingly, when prompts were introduced, the model was able to perform image segmentation normally even with a single modality input, indicating that the proposed visual prompts facilitated the encoder to learn discriminative features across modalities.
We introduced the spatial perturbation prompts module into the baseline, the overall robustness of the model was improved. As shown in Table 4, our method achieved a higher Dice score of 93.02, exceeding the baseline model by 17.21. Furthermore, the Dice score for the ET region saw a significant increase, indicating that the spatial perturbation prompt facilitated the fusion of inter-modal information and preserved more edge details and small-scale information. Figure 10 visualizes the segmentation results before and after using the spatial perturbation prompt, clearly demonstrating that more small-scale lesion areas are preserved.
Additionally, in Table 5, we described the parameter information before and after adding the module. It indicates that our method only introduced approximately 7% of the total trainable parameters but achieved excellent segmentation performance. Once extended to large models with billions of parameters, our proposed method will be more favorable and suitable for multimodal downstream tasks with missing modalities, achieving a favorable trade-off between computational cost and performance.
Method | Param (M) | Tunable Param (M) |
baseline (fine-tune) | 194.82 | 194.82 |
baseline (frozen) | 194.82 | 49.30 |
baseline + FFP | 160.42 | 58.97 |
baseline + SPP | 173.93 | 48.69 |
baseline + FFP + SPP | 153.43 | 10.58 |
In this paper, we introduce a parameter-efficient and discriminatively optimized segmentation network that exhibits robust adaptability to generalized missing modality inputs. Our model filters frequency features to generate discriminative visual cues and introduces learnable spatial perturbation prompts into shared feature representations, effectively addressing the challenge of incomplete multimodal brain tumor segmentation. Compared to fine-tuning the entire transformer model, our approach requires only 7% of the trainable parameters while demonstrating superior performance in handling real-world scenarios with missing modality data. Extensive experiments and ablation studies on the publicly available BraTS2018 and BraTS2020 datasets validate the effectiveness of our proposed method.
In this work, we investigate a parametrically efficient incomplete modal image segmentation method for brain tumors. Although our model successfully captures consistent features by mapping robust multimodal features to the same potential space, we must point out that our model cannot recover information about missing modalites from available multimodal inputs. Therefore, our next plan will study how to use the available multimodal image to estimate the missing modal information to obtain rich image information.
This work is supported by the National Nature Science Foundation of China (No.U24A20231, No.62272283), New Twentieth Items of Universities in Jinan (No.2021GXRC049).
All authors declare no conflicts of interest in this paper.
[1] |
Ahmad SA, Hanley N (2009) Willingness to pay for reducing crowding effect damages in marine parks in Malaysia. Singap Econ Rev 54: 21–39. https://doi.org/10.1142/S0217590809003124 doi: 10.1142/S0217590809003124
![]() |
[2] |
Arin T, Kramer RA (2002) Divers' willingness to pay to visit marine sanctuaries: an exploratory study. Ocean Coastal Manage 45: 171–183. https://doi.org/10.1016/S0964-5691(02)00049-2 doi: 10.1016/S0964-5691(02)00049-2
![]() |
[3] | Arrow K, Solow R, Portney PR, et al. (1993) Report of the NOAA panel on contingent valuation. Federal Register 58: 4601–4614. |
[4] |
Asafu-Adjaye J, Tapsuwan S (2008) A contingent valuation study of scuba diving benefits: Case study in Mu Ko Similan Marine National Park, Thailand. Tourism Manage 29: 1122–1130. https://doi.org/10.1016/j.tourman.2008.02.005 doi: 10.1016/j.tourman.2008.02.005
![]() |
[5] | Ashik TM (2023) All that ails St Martin's Island. The Daily Star. News Article published on 22 May 2023. Accessed on: 6 April 2024. Available from: https://www.thedailystar.net/opinion/views/news/all-ails-st-martins-island-3326651. |
[6] | Asih ENN, Nugraha WA (2020) Marine tourism in Gili Labak Island: Willingness to pay method as an effort to preserve coral reef in Gili Labak Island, Madura, Indonesia. Aquaculture, Aquarium, Conserv Legis 13: 3789–3797. |
[7] |
Baral N, Stern MJ, Bhattarai R (2008) Contingent valuation of ecotourism in Annapurna conservation area, Nepal: Implications for sustainable park finance and local development. Ecol Econ 66: 218–227. https://doi.org/10.1016/j.ecolecon.2008.02.004 doi: 10.1016/j.ecolecon.2008.02.004
![]() |
[8] | Barker NH (2003) Ecological and socio-economic impacts of dive and snorkel tourism in St. Lucia, West Indies (Doctoral dissertation, University of York). Available from: https://etheses.whiterose.ac.uk/9867/1/423741.pdf. |
[9] | Baskara KA, Hendarto RM, Susilowati I (2017) Economic's valuation of marine protected area (MPA) of Karimunjawa, Jepara-Indonesia. Aquaculture, Aquarium, Conserv Legis 10: 1554–1568. |
[10] | Billah M (2022) Why govt plans to save St Martins is falling by the wayside. The Business Standard. Online Article published on 4 January 2022. Accessed on: 6 April 2024. Available from: https://www.tbsnews.net/features/panorama/why-govt-plans-save-st-martins-falling-wayside-352819. |
[11] | Blackwell BD, Asafu-Adjaye J (2020) Adding jewels to the crown: The marginal recreational value of Noosa National Park and implications for user fees. Discussion Paper Series No. 622, University of Queensland, School of Economics. Available from: https://www.researchgate.net/profile/Boyd-Blackwell-2/publication/342623509_Adding_jewels_to_the_crown_The_marginal_recreational_value_of_Noosa_National_Park_and_implications_for_user_fees/links/5efd7159299bf18816fa4ada/Adding-jewels-to-the-crown-The-marginal-recreational-value-of-Noosa-National-Park-and-implications-for-user-fees.pdf. |
[12] | Burke L, Maidens J (2004) Reefs at Risk in the Caribbean. World Resources Institute. ISBN 1-56973-567-0. Available from: https://www.wri.org/research/reefs-risk-caribbean. |
[13] | Burke L, Selig L (2002) Reefs at risk in Southeast Asia. World Resources Institute. ISBN 1-56973-490-9. Available from: https://www.wri.org/research/reefs-risk-southeast-asia. |
[14] | Burke L, Reytar K, Spalding M, et al. (2011) Reefs at risk revisited. Washington, DC: World Resources Institute (WRI). ISBN 978-1-56973-762-0. Available from: https://bvearmb.do/handle/123456789/1787 |
[15] |
Carson RT, Groves T (2007) Incentive and informational properties of preference questions. Environ Resour Econ37: 181–210. https://doi.org/10.1007/s10640-007-9124-5 doi: 10.1007/s10640-007-9124-5
![]() |
[16] |
Casey JF, Schuhmann PW (2019) PACT or no PACT are tourists willing to contribute to the Protected Areas Conservation Trust in order to enhance marine resource conservation in Belize? Mar Policy 101: 8–14. https://doi.org/10.1016/j.marpol.2018.12.002 doi: 10.1016/j.marpol.2018.12.002
![]() |
[17] |
Casey JF, Brown C, Schuhmann P (2010) Are tourists willing to pay additional fees to protect corals in Mexico? J Sustain Tour 18: 557–573. https://doi.org/10.1080/09669580903513079 doi: 10.1080/09669580903513079
![]() |
[18] |
Cerrano C, Milanese M, Ponti M (2017) Diving for science‐science for diving: volunteer scuba divers support science and conservation in the Mediterranean Sea. Aquatic Conserv: Marine Freshwater Ecosyst 27: 303–323. https://doi.org/10.1002/aqc.2663 doi: 10.1002/aqc.2663
![]() |
[19] | Cesar H, Burke L, Pet-Soede L (2003) The economics of worldwide coral reef degradation. International Coral Reef Action Network. Available from: https://agris.fao.org/search/en/providers/122412/records/6473698a08fd68d546062c52 |
[20] |
Chaudhry P, Tewari VP (2016) Estimating recreational value of Mahatama Gandhi Marine National Park, Andaman and Nicobar Islands, India. Interdiscipl Environ Rev 17: 47–59. https://doi.org/10.1504/IER.2016.074877 doi: 10.1504/IER.2016.074877
![]() |
[21] | Clery E, Rhead R (2013) Education and attitudes towards the environment. Background paper prepared for the Education for all global monitoring report 2013/4, Teaching and learning: achieving quality for all, UNESCO. Available from: https://policycommons.net/artifacts/8226149/education-and-attitudes-towards-the-environment/9141679/. |
[22] | Cummings RG, Brookshire DS, Shultz WD (1986) Valuing environmental goods: A state of the arts assessment of the contingent valuation method. Totowa NJ: Rowan and Allenheld. Available from: https://www.semanticscholar.org/paper/VALUING-ENVIRONMENTAL-GOODS%3A-A-STATE-OF-THE-ARTS-OF-Cummings-Brookshire/dc7577ca2bdea29c00911e440067763547534f0b. |
[23] |
Depondt F, Green E (2006) Diving user fees and the financial sustainability of marine protected areas: Opportunities and impediments. Ocean Coastal Manage 49: 188–202. https://doi.org/10.1016/j.ocecoaman.2006.02.003 doi: 10.1016/j.ocecoaman.2006.02.003
![]() |
[24] |
Dharmaratne GS, Sang FY, Walling LJ (2000) Tourism potentials for financing protected areas. Annals Tourism Res 27: 590–610. https://doi.org/10.1016/S0160-7383(99)00109-7 doi: 10.1016/S0160-7383(99)00109-7
![]() |
[25] | Dixit IAM, Kumar L, Kumar P (2010) Valuing the services of coral reef systems for sustainable coastal management: A case study of the Gulf of Kachchh, India. In Valuation of Regulating Services of Ecosystems (pp. 187–210) Routledge. ISBN 9780203847602. Available from: https://www.taylorfrancis.com/chapters/edit/10.4324/9780203847602-17/valuing-services-coral-reef-systems-sustainable-coastal-management-case-study-gulf-kachchh-india-india-arun-dixit-lalit-kumar-pushpam-kumar. |
[26] |
Failler P, Montocchio C, de Battisti AB, et al. (2019) Sustainable financing of marine protected areas: the case of the Martinique regional marine reserve of "Le Prêcheur". Green Financ 1: 110–129. https://doi.org/10.3934/GF.2019.2.110 doi: 10.3934/GF.2019.2.110
![]() |
[27] |
Faizan M, Sasekumar A, Chenayah, S (2016) Estimation of local tourists willingness to pay. Reg Stud Mar Sci 7: 142–149. https://doi.org/10.1016/j.rsma.2016.06.005 doi: 10.1016/j.rsma.2016.06.005
![]() |
[28] |
Franzen A, Meyer R (2010) Environmental attitudes in cross-national perspective: A multilevel analysis of the ISSP 1993 and 2000. Eur Sociol Rev 26: 219–234. https://doi.org/10.1093/esr/jcp018 doi: 10.1093/esr/jcp018
![]() |
[29] | Frey UJ, Pirscher F (2019) Distinguishing protest responses in contingent valuation: A conceptualization of motivations and attitudes behind them. PloS One 14: e0209872. |
[30] | Frost J (2017) Nonparametric tests vs. parametric tests. Statistics by Jim Blog. Available from: https://statisticsbyjim.com/hypothesis-testing/nonparametric-parametric-tests/ |
[31] |
Gazi MY, Mowsumi TJ, Ahmed MK (2020) Detection of coral reefs degradation using geospatial techniques around Saint Martin's Island, Bay of Bengal. Ocean Sci J 55: 419–431. https://doi.org/10.1007/s12601-020-0029-3 doi: 10.1007/s12601-020-0029-3
![]() |
[32] | Government of India (2016) Lakshadweep tourism policy – 2016. Department of Tourism, Lakshadweep Administration. Available from: https://www.lakshadweeptourism.com/documents/Tourism%20Policy%202016.pdf. |
[33] |
Grafeld S, Oleson K, Barnes M, et al. (2016) Divers' willingness to pay for improved coral reef conditions in Guam: An untapped source of funding for management and conservation? Ecol Econ 128: 202–213. https://doi.org/10.1016/j.ecolecon.2016.05.005 doi: 10.1016/j.ecolecon.2016.05.005
![]() |
[34] |
Habib MHR, Rahman M, Uddin MM, et al. (2024) Application of AHP and geospatial technologies to assess ecotourism suitability: A case study of Saint Martin's Island in Bangladesh. Reg Stud Mar Sci 70: 103357. https://doi.org/10.1016/j.rsma.2023.103357 doi: 10.1016/j.rsma.2023.103357
![]() |
[35] |
Han F, Yang Z, Wang H, et al. (2011) Estimating willingness to pay for environment conservation: a contingent valuation study of Kanas Nature Reserve, Xinjiang, China. Environ Monit Assess 180: 451–459. https://doi.org/10.1007/s10661-010-1798-4 doi: 10.1007/s10661-010-1798-4
![]() |
[36] | Hasan R (2022) Saint Martin's Island: 2022 Travel Guide. BTA Holidays. Online Article published on 20 August 2022. Accessed on: 6 April 2024. Available from: https://bangladesh-travel-assistance.com/saint-martins-island/. |
[37] | Islam MN, van Amstel A, Ahmed MK (2021) Climate change-induced livelihood vulnerability and adaptation of St. Martin's Island's community, Bangladesh. Bangladesh II: Climate Change Impacts, Mitigation and Adaptation in Developing Countries, 267–282. Available from: https://research.wur.nl/en/publications/bangladesh-ii-climate-change-impacts-mitigation-and-adaptation-in-2. |
[38] | Khan H, Giurca Vasilescu L (2008) The willingness to pay for recreational services: An empirical investigation with the application of multivariate analysis of two public parks in Northern Pakistan. SSRN 1279466. Available from: https://papers.ssrn.com/sol3/papers.cfm?abstract_id = 1279466. |
[39] |
Mathieu LF, Langford IH, Kenyon W (2003) Valuing marine parks in a developing country: a case study of the Seychelles. Environ Dev Econ 8: 373–390. https://doi.org/10.1017/S1355770X0300196 doi: 10.1017/S1355770X0300196
![]() |
[40] | Mitchell RC, Carson RT (1989) Using surveys to value public goods: The Contingent Valuation Method. Washington D.C.: Resources for The Future. |
[41] |
Murphy SE, Campbell I, Drew JA (2018) Examination of tourists' willingness to pay under different conservation scenarios; Evidence from reef manta ray snorkeling in Fiji. PloS One 13: e0198279. https://doi.org/10.1371/journal.pone.0198279 doi: 10.1371/journal.pone.0198279
![]() |
[42] |
Perera NM (2016) Co-existence of coral reef conservation and tourism at Pigeon Island National Park. J Tropical Forestry Environ 6: 20–35. https://doi.org/10.31357/jtfe.v6i1.2614 doi: 10.31357/jtfe.v6i1.2614
![]() |
[43] |
Peters H, Hawkins JP (2009) Access to marine parks: A comparative study in willingness to pay. Ocean Coastal Manage 52: 219–228. https://doi.org/10.1016/j.ocecoaman.2008.12.001 doi: 10.1016/j.ocecoaman.2008.12.001
![]() |
[44] |
Strieder Philippssen J, Soares Angeoletto FH, Santana RG (2017) Education level and income are important for good environmental awareness: a case study from south Brazil. Ecología Austral 27: 39–44. https://doi.org/10.25260/EA.17.27.1.0.300 doi: 10.25260/EA.17.27.1.0.300
![]() |
[45] | Rajasuriya A, Zahir H, Muley EV, et al. (2002) Status of coral reefs in South Asia: Bangladesh, India, Maldives, Sri Lanka. In Proceedings of the Ninth International Coral Reef Symposium, Bali, 23–27 October 2000, 2: 841–845. Available from: https://www.researchgate.net/publication/27668118_Status_of_coral_reefs_in_South_Asia_Bangladesh_India_Maldives_Sri_Lanka. |
[46] | Rajasuriya A, Zahir H, Venkataraman K, et al. (2004) Status of coral reefs in South Asia: Bangladesh, Chagos, India, Maldives and Sri Lanka. Available from: https://www.researchgate.net/publication/283437263_Status_of_coral_reefs_in_South_Asia_Bangladesh_Chagos_India_Maldives_and_Sri_Lanka. |
[47] |
Rani S, Ahmed MK, Xiongzhi X, et al. (2020) Economic valuation and conservation, restoration & management strategies of Saint Martin's coral island, Bangladesh. Ocean Coastal Manage 183: 105024. https://doi.org/10.1016/j.ocecoaman.2019.105024 doi: 10.1016/j.ocecoaman.2019.105024
![]() |
[48] | Rankin J, Robinson A (2018) Accounting for protest zeros in contingent valuation studies: A review of literature, HEG Working Paper, No. 18–01, University of East Anglia, Health Economics Group (HEG), Norwich. Available from: https://www.econstor.eu/bitstream/10419/197777/1/1027441459.pdf. |
[49] |
Roberts RM, Jones KW, Seidl A, et al. (2017) Conservation finance and sustainable tourism: the acceptability of conservation fees to support the Tambopata National Reserve, Peru. J Sustain Tour 25: 1353–1366. https://doi.org/10.1080/09669582.2016.1257630 doi: 10.1080/09669582.2016.1257630
![]() |
[50] |
Rudd MA, Tupper MH (2002) The impact of Nassau grouper size and abundance on scuba diver site selection and MPA economics. Coastal Manage 30: 133–151. https://doi.org/10.1080/089207502753504670 doi: 10.1080/089207502753504670
![]() |
[51] |
Schuhmann PW, Skeete R, Waite R, et al. (2019) Visitors' willingness to pay marine conservation fees in Barbados. Tourism Manage 71: 315–326. https://doi.org/10.1016/j.tourman.2018.10.011 doi: 10.1016/j.tourman.2018.10.011
![]() |
[52] |
Seenprachawong U (2003) Economic valuation of coral reefs at Phi Phi Islands, Thailand. Int J Global Environ Issues 3: 104–114. https://doi.org/10.1504/IJGENVI.2003.002413 doi: 10.1504/IJGENVI.2003.002413
![]() |
[53] | Shams J (2023) Strict steps to save Saint Martin's. The Business Post. News Article published 10 June 2023. Accessed on: 5 April 2024. Available from: https://businesspostbd.com/front/strict-steps-to-save-saint-martins-2023-06-10. |
[54] |
Souter DW, Linden O (2000) The health and future of coral reef systems. Ocean Coastal Manage 43: 657–688. https://doi.org/10.1016/S0964-5691(00)00053-3 doi: 10.1016/S0964-5691(00)00053-3
![]() |
[55] | Staff Correspondent - NewAge Bangladesh (2022) Govt plan to limit tourists to St Martins draws flak. NewAge Bangladesh. News Article published on: 11 June 2022. Accessed on: 6 April 2024. Available from: https://www.newagebd.net/article/172989/govt-plan-to-limit-tourists-to-st-martins-draws-flak. |
[56] |
Sultana I, Alam SMI, Das DK (2018) An Annotated Avifaunal Checklist of the Saint Martin's Island of Bangladesh. J Asiatic Society Bangladesh Sci 44: 149–158. https://doi.org/10.3329/jasbs.v44i2.46557 doi: 10.3329/jasbs.v44i2.46557
![]() |
[57] | Tapsuwan S (2005) Valuing the willingness to pay for environmental conservation and management: A case study of scuba diving levies in Moo Koh Similan Islands Marine National Park, Thailand. ACE05: 34th Australian Conference of Economists, Melbourne, Australia, 26-28 September 2005. Sydney, Australia: Economic Society of Australia. Available from: https://espace.library.uq.edu.au/view/UQ: 102859. |
[58] |
Terk E, Knowlton N (2010) The role of SCUBA diver user fees as a source of sustainable funding for coral reef marine protected areas. Biodiversity 11: 78–84. https://doi.org/10.1080/14888386.2010.9712651 doi: 10.1080/14888386.2010.9712651
![]() |
[59] | Thai National Parks (2021) Similan Islands. Online web article. Accessed on: 7 April 2024. Available from: https://www.thainationalparks.com/mu-ko-similan-national-park. |
[60] |
Thur SM (2010) User fees as sustainable financing mechanisms for marine protected areas: An application to the Bonaire National Marine Park. Mar Policy 34: 63–69. https://doi.org/10.1016/j.marpol.2009.04.008 doi: 10.1016/j.marpol.2009.04.008
![]() |
[61] |
Tisdell CA, Wilson C, Swarna Nantha H (2008) Contingent valuation as a dynamic process. J Socio-Econ 37: 1443–1458. https://doi.org/10.1016/j.socec.2007.04.005 doi: 10.1016/j.socec.2007.04.005
![]() |
[62] |
Togridou A, Hovardas T, Pantis JD (2006) Determinants of visitors' willingness to pay for the National Marine Park of Zakynthos, Greece. Ecol Econ 60: 308–319. https://doi.org/10.1016/j.ecolecon.2005.12.006 doi: 10.1016/j.ecolecon.2005.12.006
![]() |
[63] | Tomascik T (1997) Management Plan for Coral Resources of Narikel Jinjira (St. Martin's Island) Final Report. National Conservation Strategy Implementation Project – 1, Dhaka, Bangladesh. Available from: https://www.researchgate.net/profile/Tomas-tom-Tomascik/publication/304077895_MANAGEMENT_PLAN_FOR_CORAL_RESOURCES_OF_NARIKEL_JINJIRA_St_Martin's_Island/links/5765925208aedbc345f3820c/MANAGEMENT-PLAN-FOR-CORAL-RESOURCES-OF-NARIKEL-JINJIRA-St-Martins-Island.pdf. |
[64] |
Uddin MM, Schneider P, Asif MRI, et al. (2021) Fishery-based ecotourism in developing countries can enhance the social-ecological resilience of coastal fishers—a case study of Bangladesh. Water 13: 292. https://doi.org/10.3390/w13030292 doi: 10.3390/w13030292
![]() |
[65] |
Uyarra MC, Gill JA, Côté IM (2010) Charging for nature: marine park fees and management from a user perspective. Ambio 39: 515–523. https://doi.org/10.1007/s13280-010-0078-4 doi: 10.1007/s13280-010-0078-4
![]() |
[66] |
Wielgus J, Balmford A, Lewis TB, et al. (2010) Coral reef quality and recreation fees in marine protected areas. Conserv Lett 3: 38–44. https://doi.org/10.1111/j.1755-263X.2009.00084.x doi: 10.1111/j.1755-263X.2009.00084.x
![]() |
[67] |
Wielgus J, Chadwick-Furman NE, Zeitouni N, et al. (2003) Effects of coral reef attribute damage on recreational welfare. Marine Resour Econ 18: 225–237. https://doi.org/10.1086/mre.18.3.42629397 doi: 10.1086/mre.18.3.42629397
![]() |
[68] |
Zakai D, Chadwick-Furman NE (2002) Impacts of intensive recreational diving on reef corals at Eilat, northern Red Sea. Biol Conserv 105: 179–187. https://doi.org/10.1016/S0006-3207(01)00181-1 doi: 10.1016/S0006-3207(01)00181-1
![]() |
[69] | Zinat MA, Roy P (2015) Biodiversity of St Martin's under threat. The Daily Star. Online news article published on 16 October 2015. Accessed on: 7 April 2024. Available from: https://www.thedailystar.net/backpage/biodiversity-st-martins-under-threat-157891. |
![]() |
![]() |
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | D | Z | T | Q | Our | D | Z | T | Q | Our | D | Z | T | Q | Our |
✓ | 86.1 | 86.1 | 86.5 | 86.7 | 93.9 | 71.0 | 70.9 | 71.5 | 71.0 | 93.3 | 46.3 | 46.3 | 45.6 | 47.2 | 76.1 | |||
✓ | 76.8 | 78.5 | 77.4 | 79.5 | 91.6 | 81.5 | 84.0 | 83.4 | 84.3 | 95.3 | 74.9 | 80.1 | 78.9 | 81.4 | 88.4 | |||
✓ | 77.2 | 78.0 | 78.1 | 79.5 | 89.1 | 66.0 | 65.9 | 66.8 | 67.7 | 91.9 | 37.3 | 38.0 | 41.3 | 39.1 | 71.6 | |||
✓ | 87.3 | 87.4 | 89.1 | 86.9 | 95.2 | 69.2 | 68.8 | 69.3 | 69.9 | 93.5 | 38.2 | 42.4 | 43.6 | 42.8 | 74.6 | |||
✓ | ✓ | 87.7 | 87.8 | 88.4 | 88.4 | 94.5 | 83.5 | 84.8 | 86.4 | 86.3 | 95.8 | 75.9 | 79.4 | 81.7 | 80.1 | 88.9 | ||
✓ | ✓ | 81.1 | 81.8 | 81.2 | 83.1 | 92.1 | 83.4 | 83.6 | 85.2 | 85.8 | 95.4 | 78.0 | 80.1 | 79.2 | 81.7 | 88.3 | ||
✓ | ✓ | 89.7 | 89.8 | 89.9 | 89.8 | 95.5 | 73.1 | 73.8 | 73.9 | 74.4 | 94.3 | 41.0 | 45.9 | 48.2 | 46.8 | 77.3 | ||
✓ | ✓ | 87.7 | 87.8 | 88.0 | 87.9 | 94.4 | 73.1 | 73.4 | 73.3 | 72.9 | 94.1 | 45.7 | 46.8 | 50.1 | 47.3 | 77.5 | ||
✓ | ✓ | 89.9 | 89.9 | 90.5 | 90.1 | 95.5 | 74.1 | 74.6 | 75.5 | 74.5 | 94.1 | 49.3 | 48.6 | 48.6 | 49.5 | 76.6 | ||
✓ | ✓ | 89.9 | 89.3 | 90.0 | 90.0 | 95.6 | 84.7 | 84.8 | 85.5 | 86.6 | 95.9 | 76.7 | 81.9 | 81.8 | 81.2 | 88.9 | ||
✓ | ✓ | ✓ | 90.7 | 90.1 | 90.7 | 90.6 | 95.6 | 85.1 | 85.2 | 86.5 | 86.7 | 95.8 | 76.8 | 82.1 | 81.8 | 81.8 | 88.8 | |
✓ | ✓ | ✓ | 90.6 | 90.6 | 90.3 | 90.6 | 95.7 | 75.2 | 75.6 | 75.9 | 75.8 | 94.7 | 49.9 | 50.3 | 52.5 | 51.1 | 78.0 | |
✓ | ✓ | ✓ | 90.7 | 90.4 | 90.6 | 90.8 | 95.8 | 85.0 | 85.3 | 86.4 | 86.4 | 96.0 | 77.1 | 78.7 | 81.0 | 80.0 | 88.9 | |
✓ | ✓ | ✓ | 88.3 | 88.2 | 88.7 | 88.9 | 94.6 | 83.5 | 84.2 | 86.5 | 86.5 | 95.8 | 77.0 | 79.3 | 78.5 | 82.1 | 88.9 | |
✓ | ✓ | ✓ | ✓ | 91.1 | 90.6 | 90.6 | 91.0 | 95.9 | 85.2 | 84.6 | 87.4 | 86.4 | 95.9 | 78.0 | 79.9 | 81.6 | 81.0 | 88.9 |
Average | 87.0 | 87.1 | 87.3 | 87.6 | 94.3 | 78.2 | 78.6 | 79.6 | 79.7 | 94.8 | 61.5 | 64.0 | 64.9 | 64.9 | 82.8 |
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | Z | Y | T | L | Our | Z | Y | T | L | Our | Z | Y | T | L | Our |
✓ | 81.2 | 76.3 | 86.6 | 84.8 | 94.3 | 64.2 | 56.7 | 68.8 | 69.4 | 94.4 | 43.1 | 16.0 | 41.4 | 47.6 | 76.2 | |||
✓ | 72.2 | 42.8 | 77.8 | 75.8 | 92.6 | 75.4 | 65.1 | 81.5 | 82.9 | 95.4 | 72.6 | 66.3 | 75.7 | 73.7 | 89.2 | |||
✓ | 67.5 | 15.5 | 78.7 | 74.4 | 90.9 | 56.6 | 16.8 | 65.6 | 66.1 | 93.2 | 32.5 | 8.1 | 44.5 | 37.1 | 74.7 | |||
✓ | 86.1 | 84.2 | 88.4 | 88.7 | 95.2 | 61.2 | 47.3 | 66.7 | 66.4 | 94.2 | 39.3 | 8.1 | 40.5 | 35.6 | 74.8 | |||
✓ | ✓ | 83.0 | 84.1 | 88.2 | 86.3 | 95.0 | 78.6 | 80.3 | 84.8 | 84.2 | 96.1 | 74.5 | 68.7 | 77.7 | 75.3 | 90.0 | ||
✓ | ✓ | 74.4 | 62.1 | 81.8 | 77.2 | 93.1 | 78.6 | 78.2 | 83.5 | 83.4 | 95.7 | 74.0 | 70.7 | 77.1 | 74.7 | 89.5 | ||
✓ | ✓ | 87.1 | 87.3 | 89.7 | 89.0 | 95.6 | 65.9 | 61.6 | 72.0 | 70.8 | 95.2 | 43.0 | 9.5 | 44.4 | 41.2 | 77.9 | ||
✓ | ✓ | 82.2 | 84.2 | 88.4 | 88.7 | 94.9 | 61.2 | 47.3 | 66.7 | 66.4 | 95.1 | 45.0 | 16.5 | 47.7 | 48.7 | 77.7 | ||
✓ | ✓ | 87.6 | 87.9 | 90.3 | 89.9 | 95.9 | 69.8 | 62.6 | 71.8 | 70.9 | 95.1 | 47.5 | 17.4 | 48.3 | 45.4 | 78.1 | ||
✓ | ✓ | 87.1 | 87.5 | 89.5 | 89.7 | 95.6 | 77.9 | 80.8 | 84.8 | 84.4 | 96.1 | 75.1 | 64.8 | 76.8 | 75.0 | 90.0 | ||
✓ | ✓ | ✓ | 87.3 | 87.7 | 90.4 | 88.9 | 95.7 | 79.8 | 80.9 | 85.2 | 84.1 | 96.2 | 75.5 | 65.7 | 77.4 | 74.0 | 90.0 | |
✓ | ✓ | ✓ | 87.8 | 88.4 | 89.7 | 89.9 | 96.0 | 71.5 | 63.7 | 74.1 | 72.7 | 95.5 | 47.7 | 19.4 | 50.0 | 44.8 | 78.7 | |
✓ | ✓ | ✓ | 88.1 | 88.8 | 90.6 | 90.4 | 96.0 | 79.6 | 80.7 | 85.8 | 84.6 | 96.3 | 75.7 | 66.4 | 76.6 | 73.8 | 90.1 | |
✓ | ✓ | ✓ | 82.7 | 80.9 | 88.4 | 86.1 | 95.1 | 80.4 | 79.0 | 85.8 | 84.4 | 96.2 | 74.8 | 68.3 | 78.5 | 75.4 | 90.1 | |
✓ | ✓ | ✓ | ✓ | 89.6 | 88.8 | 90.6 | 90.1 | 96.1 | 85.8 | 80.1 | 85.9 | 84.5 | 96.3 | 77.6 | 68.4 | 80.4 | 75.5 | 90.0 |
Average | 82.9 | 76.4 | 87.3 | 86.0 | 94.8 | 72.4 | 65.4 | 77.5 | 77.0 | 95.4 | 59.9 | 42.3 | 62.5 | 59.9 | 83.8 |
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
Ding et al. | 86.13 | 71.93 | 58.98 | 72.35 | - | - | - | - | - | - | - | - |
Zhang et al. | 87.08 | 78.69 | 64.08 | 76.62 | 2.90 | 6.21 | 44.64 | 17.92 | 99.60 | 99.81 | 99.82 | 99.74 |
Ting et al. | 90.71 | 84.60 | 79.07 | 84.79 | 4.05 | 5.78 | 33.77 | 14.53 | 90.98 | 83.90 | 77.68 | 84.18 |
Qiu et al. | 87.58 | 79.67 | 64.87 | 77.37 | 2.82 | 5.71 | 43.92 | 17.48 | 99.66 | 99.83 | 99.81 | 99.77 |
baseline(fine-tune) | 77.63 | 78.94 | 70.85 | 93.56 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline(frozen) | 58.11 | 61.09 | 40.88 | 89.16 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
our | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
baseline (fine-tune) | 77.63 | 78.94 | 70.85 | 75.81 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline (frozen) | 58.11 | 61.09 | 40.88 | 53.36 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
baseline + FFP | 93.65 | 92.40 | 85.08 | 90.38 | 2.45 | 2.04 | 2.16 | 2.22 | 96.54 | 96.11 | 91.26 | 94.64 |
baseline + SPP | 94.56 | 94.40 | 87.37 | 92.11 | 2.47 | 2.05 | 2.22 | 2.25 | 96.59 | 96.07 | 90.53 | 94.40 |
baseline + FFP + SPP | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
Method | Param (M) | Tunable Param (M) |
baseline (fine-tune) | 194.82 | 194.82 |
baseline (frozen) | 194.82 | 49.30 |
baseline + FFP | 160.42 | 58.97 |
baseline + SPP | 173.93 | 48.69 |
baseline + FFP + SPP | 153.43 | 10.58 |
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | D | Z | T | Q | Our | D | Z | T | Q | Our | D | Z | T | Q | Our |
✓ | 86.1 | 86.1 | 86.5 | 86.7 | 93.9 | 71.0 | 70.9 | 71.5 | 71.0 | 93.3 | 46.3 | 46.3 | 45.6 | 47.2 | 76.1 | |||
✓ | 76.8 | 78.5 | 77.4 | 79.5 | 91.6 | 81.5 | 84.0 | 83.4 | 84.3 | 95.3 | 74.9 | 80.1 | 78.9 | 81.4 | 88.4 | |||
✓ | 77.2 | 78.0 | 78.1 | 79.5 | 89.1 | 66.0 | 65.9 | 66.8 | 67.7 | 91.9 | 37.3 | 38.0 | 41.3 | 39.1 | 71.6 | |||
✓ | 87.3 | 87.4 | 89.1 | 86.9 | 95.2 | 69.2 | 68.8 | 69.3 | 69.9 | 93.5 | 38.2 | 42.4 | 43.6 | 42.8 | 74.6 | |||
✓ | ✓ | 87.7 | 87.8 | 88.4 | 88.4 | 94.5 | 83.5 | 84.8 | 86.4 | 86.3 | 95.8 | 75.9 | 79.4 | 81.7 | 80.1 | 88.9 | ||
✓ | ✓ | 81.1 | 81.8 | 81.2 | 83.1 | 92.1 | 83.4 | 83.6 | 85.2 | 85.8 | 95.4 | 78.0 | 80.1 | 79.2 | 81.7 | 88.3 | ||
✓ | ✓ | 89.7 | 89.8 | 89.9 | 89.8 | 95.5 | 73.1 | 73.8 | 73.9 | 74.4 | 94.3 | 41.0 | 45.9 | 48.2 | 46.8 | 77.3 | ||
✓ | ✓ | 87.7 | 87.8 | 88.0 | 87.9 | 94.4 | 73.1 | 73.4 | 73.3 | 72.9 | 94.1 | 45.7 | 46.8 | 50.1 | 47.3 | 77.5 | ||
✓ | ✓ | 89.9 | 89.9 | 90.5 | 90.1 | 95.5 | 74.1 | 74.6 | 75.5 | 74.5 | 94.1 | 49.3 | 48.6 | 48.6 | 49.5 | 76.6 | ||
✓ | ✓ | 89.9 | 89.3 | 90.0 | 90.0 | 95.6 | 84.7 | 84.8 | 85.5 | 86.6 | 95.9 | 76.7 | 81.9 | 81.8 | 81.2 | 88.9 | ||
✓ | ✓ | ✓ | 90.7 | 90.1 | 90.7 | 90.6 | 95.6 | 85.1 | 85.2 | 86.5 | 86.7 | 95.8 | 76.8 | 82.1 | 81.8 | 81.8 | 88.8 | |
✓ | ✓ | ✓ | 90.6 | 90.6 | 90.3 | 90.6 | 95.7 | 75.2 | 75.6 | 75.9 | 75.8 | 94.7 | 49.9 | 50.3 | 52.5 | 51.1 | 78.0 | |
✓ | ✓ | ✓ | 90.7 | 90.4 | 90.6 | 90.8 | 95.8 | 85.0 | 85.3 | 86.4 | 86.4 | 96.0 | 77.1 | 78.7 | 81.0 | 80.0 | 88.9 | |
✓ | ✓ | ✓ | 88.3 | 88.2 | 88.7 | 88.9 | 94.6 | 83.5 | 84.2 | 86.5 | 86.5 | 95.8 | 77.0 | 79.3 | 78.5 | 82.1 | 88.9 | |
✓ | ✓ | ✓ | ✓ | 91.1 | 90.6 | 90.6 | 91.0 | 95.9 | 85.2 | 84.6 | 87.4 | 86.4 | 95.9 | 78.0 | 79.9 | 81.6 | 81.0 | 88.9 |
Average | 87.0 | 87.1 | 87.3 | 87.6 | 94.3 | 78.2 | 78.6 | 79.6 | 79.7 | 94.8 | 61.5 | 64.0 | 64.9 | 64.9 | 82.8 |
Modalities | Dice (%) ↑ | |||||||||||||||||
Complete | Core | Enhancing | ||||||||||||||||
F | T1 | T1c | T2 | Z | Y | T | L | Our | Z | Y | T | L | Our | Z | Y | T | L | Our |
✓ | 81.2 | 76.3 | 86.6 | 84.8 | 94.3 | 64.2 | 56.7 | 68.8 | 69.4 | 94.4 | 43.1 | 16.0 | 41.4 | 47.6 | 76.2 | |||
✓ | 72.2 | 42.8 | 77.8 | 75.8 | 92.6 | 75.4 | 65.1 | 81.5 | 82.9 | 95.4 | 72.6 | 66.3 | 75.7 | 73.7 | 89.2 | |||
✓ | 67.5 | 15.5 | 78.7 | 74.4 | 90.9 | 56.6 | 16.8 | 65.6 | 66.1 | 93.2 | 32.5 | 8.1 | 44.5 | 37.1 | 74.7 | |||
✓ | 86.1 | 84.2 | 88.4 | 88.7 | 95.2 | 61.2 | 47.3 | 66.7 | 66.4 | 94.2 | 39.3 | 8.1 | 40.5 | 35.6 | 74.8 | |||
✓ | ✓ | 83.0 | 84.1 | 88.2 | 86.3 | 95.0 | 78.6 | 80.3 | 84.8 | 84.2 | 96.1 | 74.5 | 68.7 | 77.7 | 75.3 | 90.0 | ||
✓ | ✓ | 74.4 | 62.1 | 81.8 | 77.2 | 93.1 | 78.6 | 78.2 | 83.5 | 83.4 | 95.7 | 74.0 | 70.7 | 77.1 | 74.7 | 89.5 | ||
✓ | ✓ | 87.1 | 87.3 | 89.7 | 89.0 | 95.6 | 65.9 | 61.6 | 72.0 | 70.8 | 95.2 | 43.0 | 9.5 | 44.4 | 41.2 | 77.9 | ||
✓ | ✓ | 82.2 | 84.2 | 88.4 | 88.7 | 94.9 | 61.2 | 47.3 | 66.7 | 66.4 | 95.1 | 45.0 | 16.5 | 47.7 | 48.7 | 77.7 | ||
✓ | ✓ | 87.6 | 87.9 | 90.3 | 89.9 | 95.9 | 69.8 | 62.6 | 71.8 | 70.9 | 95.1 | 47.5 | 17.4 | 48.3 | 45.4 | 78.1 | ||
✓ | ✓ | 87.1 | 87.5 | 89.5 | 89.7 | 95.6 | 77.9 | 80.8 | 84.8 | 84.4 | 96.1 | 75.1 | 64.8 | 76.8 | 75.0 | 90.0 | ||
✓ | ✓ | ✓ | 87.3 | 87.7 | 90.4 | 88.9 | 95.7 | 79.8 | 80.9 | 85.2 | 84.1 | 96.2 | 75.5 | 65.7 | 77.4 | 74.0 | 90.0 | |
✓ | ✓ | ✓ | 87.8 | 88.4 | 89.7 | 89.9 | 96.0 | 71.5 | 63.7 | 74.1 | 72.7 | 95.5 | 47.7 | 19.4 | 50.0 | 44.8 | 78.7 | |
✓ | ✓ | ✓ | 88.1 | 88.8 | 90.6 | 90.4 | 96.0 | 79.6 | 80.7 | 85.8 | 84.6 | 96.3 | 75.7 | 66.4 | 76.6 | 73.8 | 90.1 | |
✓ | ✓ | ✓ | 82.7 | 80.9 | 88.4 | 86.1 | 95.1 | 80.4 | 79.0 | 85.8 | 84.4 | 96.2 | 74.8 | 68.3 | 78.5 | 75.4 | 90.1 | |
✓ | ✓ | ✓ | ✓ | 89.6 | 88.8 | 90.6 | 90.1 | 96.1 | 85.8 | 80.1 | 85.9 | 84.5 | 96.3 | 77.6 | 68.4 | 80.4 | 75.5 | 90.0 |
Average | 82.9 | 76.4 | 87.3 | 86.0 | 94.8 | 72.4 | 65.4 | 77.5 | 77.0 | 95.4 | 59.9 | 42.3 | 62.5 | 59.9 | 83.8 |
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
Ding et al. | 86.13 | 71.93 | 58.98 | 72.35 | - | - | - | - | - | - | - | - |
Zhang et al. | 87.08 | 78.69 | 64.08 | 76.62 | 2.90 | 6.21 | 44.64 | 17.92 | 99.60 | 99.81 | 99.82 | 99.74 |
Ting et al. | 90.71 | 84.60 | 79.07 | 84.79 | 4.05 | 5.78 | 33.77 | 14.53 | 90.98 | 83.90 | 77.68 | 84.18 |
Qiu et al. | 87.58 | 79.67 | 64.87 | 77.37 | 2.82 | 5.71 | 43.92 | 17.48 | 99.66 | 99.83 | 99.81 | 99.77 |
baseline(fine-tune) | 77.63 | 78.94 | 70.85 | 93.56 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline(frozen) | 58.11 | 61.09 | 40.88 | 89.16 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
our | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
Method | Dice ↑ | HD95 ↓ | Sensitivity ↑ | |||||||||
WT | TC | ET | Avg | WT | TC | ET | Avg | WT | TC | ET | Avg | |
baseline (fine-tune) | 77.63 | 78.94 | 70.85 | 75.81 | 2.61 | 2.09 | 2.39 | 2.36 | 86.28 | 86.50 | 82.74 | 85.17 |
baseline (frozen) | 58.11 | 61.09 | 40.88 | 53.36 | 2.83 | 2.29 | 2.97 | 2.70 | 81.41 | 84.68 | 85.90 | 84.00 |
baseline + FFP | 93.65 | 92.40 | 85.08 | 90.38 | 2.45 | 2.04 | 2.16 | 2.22 | 96.54 | 96.11 | 91.26 | 94.64 |
baseline + SPP | 94.56 | 94.40 | 87.37 | 92.11 | 2.47 | 2.05 | 2.22 | 2.25 | 96.59 | 96.07 | 90.53 | 94.40 |
baseline + FFP + SPP | 94.96 | 94.12 | 89.98 | 93.02 | 2.58 | 2.09 | 2.21 | 2.29 | 96.81 | 96.32 | 93.01 | 95.38 |
Method | Param (M) | Tunable Param (M) |
baseline (fine-tune) | 194.82 | 194.82 |
baseline (frozen) | 194.82 | 49.30 |
baseline + FFP | 160.42 | 58.97 |
baseline + SPP | 173.93 | 48.69 |
baseline + FFP + SPP | 153.43 | 10.58 |