
Citation: Zhi-xin Zheng, Jun-qing Li, Hong-yan Sang. A hybrid invasive weed optimization algorithm for the economic load dispatch problem in power systems[J]. Mathematical Biosciences and Engineering, 2019, 16(4): 2775-2794. doi: 10.3934/mbe.2019138
[1] | Hongan Li, Qiaoxue Zheng, Wenjing Yan, Ruolin Tao, Xin Qi, Zheng Wen . Image super-resolution reconstruction for secure data transmission in Internet of Things environment. Mathematical Biosciences and Engineering, 2021, 18(5): 6652-6671. doi: 10.3934/mbe.2021330 |
[2] | Zhijing Xu, Jingjing Su, Kan Huang . A-RetinaNet: A novel RetinaNet with an asymmetric attention fusion mechanism for dim and small drone detection in infrared images. Mathematical Biosciences and Engineering, 2023, 20(4): 6630-6651. doi: 10.3934/mbe.2023285 |
[3] | Yingying Xu, Songsong Dai, Haifeng Song, Lei Du, Ying Chen . Multi-modal brain MRI images enhancement based on framelet and local weights super-resolution. Mathematical Biosciences and Engineering, 2023, 20(2): 4258-4273. doi: 10.3934/mbe.2023199 |
[4] | Jimin Yu, Jiajun Yin, Shangbo Zhou, Saiao Huang, Xianzhong Xie . An image super-resolution reconstruction model based on fractional-order anisotropic diffusion equation. Mathematical Biosciences and Engineering, 2021, 18(5): 6581-6607. doi: 10.3934/mbe.2021326 |
[5] | Qing Zou, Zachary Miller, Sanja Dzelebdzic, Maher Abadeer, Kevin M. Johnson, Tarique Hussain . Time-Resolved 3D cardiopulmonary MRI reconstruction using spatial transformer network. Mathematical Biosciences and Engineering, 2023, 20(9): 15982-15998. doi: 10.3934/mbe.2023712 |
[6] | Linglei Meng, XinFang Shang, FengXiao Gao, DeMao Li . Comparative study of imaging staging and postoperative pathological staging of esophageal cancer based on smart medical big data. Mathematical Biosciences and Engineering, 2023, 20(6): 10514-10529. doi: 10.3934/mbe.2023464 |
[7] | Zhuang Zhang, Wenjie Luo . Hierarchical volumetric transformer with comprehensive attention for medical image segmentation. Mathematical Biosciences and Engineering, 2023, 20(2): 3177-3190. doi: 10.3934/mbe.2023149 |
[8] | Liwei Deng, Yuanzhi Zhang, Jingjing Qi, Sijuan Huang, Xin Yang, Jing Wang . Enhancement of cone beam CT image registration by super-resolution pre-processing algorithm. Mathematical Biosciences and Engineering, 2023, 20(3): 4403-4420. doi: 10.3934/mbe.2023204 |
[9] | Qiming Li, Chengcheng Chen . A robust and high-precision edge segmentation and refinement method for high-resolution images. Mathematical Biosciences and Engineering, 2023, 20(1): 1058-1082. doi: 10.3934/mbe.2023049 |
[10] | Shuaiyu Bu, Yuanyuan Li, Guoqiang Liu, Yifan Li . MAET-SAM: Magneto-Acousto-Electrical Tomography segmentation network based on the segment anything model. Mathematical Biosciences and Engineering, 2025, 22(3): 585-603. doi: 10.3934/mbe.2025022 |
High-resolution (HR) magnetic resonance imaging (MRI) unveils enhanced structural details and textures, essential for accurate diagnosis and pathological analysis of bodily organs. However, the resolution of the medical image is often constrained by factors like imaging hardware limitations, prolonged scanning durations, and lower signal-to-noise ratios (SNR) [1]. Improving spatial resolution usually involves the sacrifice of decreased SNR and increased scanning time [2].
Recently, super-resolution (SR) has emerged as a post-processing technique for upscaling the resolution of MRI images [2,3,4]. Existing SR methods include interpolation-based, regularization-based, and learning-based methods [5,6]. Interpolation methods usually blur sharp edges and can hardly recover fine details or handle complex textures [7]. Using deep convolutional neural networks (CNN) in the SR image has shown notable success in high-quality reconstruction performance [8]. After the pioneering work of SRCNN [9], a multitude of CNN-based SR models have been proposed, such as EDSR [10], RCAN [11], and SwinIR [12], significantly improving SR performance. The superior reconstruction performance of CNN-based methods, such as SAN [13] and HAN [14], primarily stems from their deep architecture, residual learning, and diverse attention mechanisms [7,15]. Deepening the network's layers can enlarge receptive fields and facilitate its ability to comprehend the intricate mapping between the low-resolution (LR) inputs and HR counterparts. The adoption of residual learning facilitates deeper SR networks, as it effectively mitigates issues associated with gradient vanishing and explosion. Since CNN-based SR methods develop rapidly, transformer-based SR methods emerged to further improve SR performance [12,16,17]. As an alternative to CNN, transformer-based methods make full use of long-range dependency information rather than local features, greatly improving SR performance. However, the transformer-based SR model usually has large model parameters and is difficult to train.
Although previous work has made significant progress, the deep SR model is still challenging to train because of its expensive GPU computation and time costs, leading to decreased performance of the state-of-the-art methods [18]. Therefore, the SR methods proposed ahead are not suitable for limited computation resources and limited diagnosis time in medical applications.
To tackle the aforementioned issues and challenges, we propose the multi-distillation residual network (MDRN), which has a superior trade-off between reconstruction quality and computation consumption. Specifically, we propose a feature multi-distillation residual block (FMDRB), used in MDRN, which selectively retains certain features and sends others to the subsequent steps. To maximize the feature distillation capability, we incorporate a contrast-aware channel attention layer (CCA) to enhance the aggregation of diverse refined information. Our approach focuses on leveraging more informative features such as edges, textures, and small vessels for MRI image reconstruction.
In general, our main contributions can be summarized as follows:
1) We propose a multi-distillation residual network (MDRN) applied to efficient and fast super-resolution MRI that learns extra discriminative feature representations and is lightweight enough for limited computation costs. Our MDRN is suitable for super-resolution MRI and clinical applications.
2) We introduce a CCA block to our FMDRB that can guide the model to focus on recovering high-frequency information. Based on that, CCA maximizes the power of the MDRN network. Besides, it is suitable for low-level vision and has better performance than the plain channel attention block.
3) Thanks to the unique design of MDRN, it outperforms previous CNN-based SR models even under smaller GPU conditions. The proposed method obtains the best trade-off between inference time and reconstruction quality, showing the competitive advantage of our MDRN over state-of-the-art (SOTA) methods, as supported by quantitative and qualitative evidence.
We propose a multi-distillation residual network (MDRN) for efficient and fast super-resolution MRI, whose architecture is shown in Figure 1. In Section 2.1, we provide an overview of the MDRN structure. In Section 2.2, we introduce the core module: feature multi-distillation residual block (FMDRB). Drawing inspiration from the common residual block (RB) [10] and information multi-distillation block (IMDB) [19], our network comprises a series of stacked FMDRBs forming the main chain, as demonstrated in Figure 1.
Given ILR as the LR input of MDRN, the network reconstructs the SR output ISR from the LR input. As in previous works, we adopt a shallow feature extraction, deep feature extraction, and post-upsample structure. The process of shallow feature F0 extracted from the input ILR is as follows:
F0=DSF(ILR), | (1) |
where HSF(⋅) demonstrates the function of shallow feature extractor, specifically one convolution operation.
The subsequent part of MDRN involves the integration of multiple FMDRBs, which are put in a chain manner with feature distillation connections. This design facilitates the gradual refinement of the initial extracted features, culminating in the generation of deep features. The deep feature extraction part can be described as follows:
Fk=DDFk(Fk−1),k=1,…,n, | (2) |
where DDFk(⋅) stands for the function of k -th FMDRB, and Fk−1 and Fk represent the input and output features of the k-th FMDRB, respectively. After the iterative refinement process by the FMDRBs, one 1×1 convolution layer is put at the end of a feature extraction part to assemble the fused distilled features. Following the fusion operation, a 3×3 convolution layer is put here to smooth the inductive bias of the aggregated features as follows:
Ffusion=Daggregated(Concat(F1,⋯,Fn)), | (3) |
where Concat denotes the fusion operation through channel concatenation of all the distillation features, Daggregated denotes the operation, which is one 3×3 convolution following one 1×1 convolution, and Ffusion is the fused and aggregated features. Finally, the SR output ISR is generated by the reconstruction module as follows:
ISR=DREC(Ffusion+F0), | (4) |
where DREC(⋅) denotes the function of the upscale reconstruction part. The initial extracted feature F0 is added to the assembled features Ffusion through skip connection, and ISR is the output of the network. The upsample reconstruction works through a convolution layer, whose output channels are quadratic in relation to the upscale factor with a 3×3 kernel size and a sub-pixel shuffle operation that is non-parametric.
The shallow extracted features predominantly contain low-frequency information, whereas deep extracted features focus more on restoring fading high-frequency information. The skip connection path enables MDRN to directly transmit low frequencies to the reconstruction process, which can help combine information and achieve more stable training.
Inspired by the concept of feature distillation and residual learning, we designed the core module--feature multi-distillation residual block (FMDRB), which is more efficient and lightweight than the traditional residual modules. Different from the common residual block (two convolutions and one activation with identity connection), FMDRB uses an additional path with convolution for feature distillation and improved residual blocks stacked in the main chain as refinement layers that process coarse features gradually. We describe the complete structure as follows:
Fdistilled_1=D1(Fin),Fremain_1=R1(Fin),Fdistilled_2=D2(Fremain_1),Fremain_2=R2(Fremain_1),Fdistilled_3=D2(Fremain_2),Fremain_3=R2(Fremain_2),Fremain_4=R4(Fremain_3),Fout=Concat(Fdistilled_1,Fdistilled_2,Fdistilled_3,Fremain_4), | (5) |
where D denotes the distillation operation, R denotes the layer for remaining features, and the subscript number represents the number of layers. The output feature Fout fuses the right-most features processed in the main chain and distilled features in the distillation paths. As described in the above equations, the distillation operation works concurrently with the residual learning; this structure shows more efficiency and flexibility than the original residual block commonly used. As such, this block is called feature multi-distillation residual block.
As shown in Figure 1 below, the feature distillation path in each level is performed by one 1×1 convolution layer that effectively compresses feature channels at a fixed ratio; for example, we use input channels divided by 2. Although most convolutions in the SR model use 3×3 kernel size, we note that employing the 1×1 convolution for channel reduction, as done in numerous other CNN models, is more efficient. As we replace the convolution in the distillation path, the parameter amount is significantly reduced. The convolutions located in the main body of MDRN still use a 3×3 kernel size, which better refines the features in the main path and more effectively utilizes spatial information in context.
As shown in Figure 1, despite the improvements mentioned above, we also introduce the base unit of FMDRB, named BSRB [20], which allows more flexible residual learning than a common residual block. Specifically, it uses a 3×3 Blueprint Separable Convolution (BSConv) [21], an identity connection, and the ReLU activation layer. BSConv is a 1×1 pointwise convolution followed by a 3×3 depthwise convolution, which differs from the standard convolution.
The initial concept of channel attention, widely recognized as the squeeze-and-excitation (SE) module, has been extensively used in image processing tasks. The significance of a feature map is predominantly determined by the activation of high-value regions, as these areas are critical for classification or detection. Consequently, global average and maximum pooling are commonly utilized to capture global information in these high- or mid-level visions. While average pooling can indeed enhance the PSNR value, it lacks the capability to retain structural, textural, and edge information, which are crucial for improving image detail (as related to SSIM) [19]. As illustrated in Figure 1, the contrast-aware channel attention module is specific to low-level vision. Specifically, we replace global average pooling with the summation of standard deviation and mean (evaluating the contrast degree of a feature map). Let us denote X=[x1,x2,…,xc,…,xC] as the input, which has C feature maps with spatial size of H×W. Therefore, the contrast information value can be calculated by
zc=HGC(xc)=√1HW∑(i,j)∈xc(xi,jc−1HW∑(i,j)∈xcxi,jc)2+1HW∑(i,j)∈xcxi,jc, | (6) |
where zc is the c-th element of output. HGC indicates the global contrast information evaluation function. With the assistance of the CCA module, our network can steadily improve the accuracy of super-resolution.
We used the public clinical dataset from The Cancer Imaging Archive [22], which is available at https://www.cancerimagingarchive.net/collection/vestibular-schwannoma-seg/, named MRI-brain below. The dataset contains labeled MRI images obtained from 242 patients who received Gamma Knife radiation treatment and have been diagnosed with vestibular schwannoma. The images were acquired on a 32-channel Siemens Avanto 1.5T scanner. We used 5000 slices in the MRI-brain dataset for the training set. For testing the performance of our method, we used the remaining 1000 slices as the testing set. The dataset is enough for training and testing since one patient has approximately 140-160 slices.
In data preprocessing, first, we converted the DICOM raw files to NumPy files with voxels. Then, the image pixel data was clipped to range below 2000 and normalized to range [0, 1]. Third, we used bicubic interpolation as the degradation function of the original HR image to the LR image. The preprocessing workflow is shown in Figure 2.
We trained our model with 5×10−4 learning rate updated by StepLR scheduler and minimizing the L1 loss function. For the purpose of reducing the training burden, we got patches 192×192 from whole HR images as the input to the network. We used the ADAM optimizer with β1 = 0.9, β2 = 0.99. The entire MDRN procedure took approximately 48 h (20, 000 iterations per epoch, 200 epochs) for training and evaluation on the MRI dataset on a single GeForce RTX 3090 GPU with 24 GB of memory.
Following previous works, peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) were used to assess the model's performance. The calculation of these evaluation metrics is written below:
PSNR=10log10(MAX2MSE),MSE=1mn∑m−1i=0∑n−1j=0[Ix(i,j)−Iy(i,j)]2, | (7) |
SSIM=(2μxμy+c1)(σxy+c2)(μ2x+μ2y+c1)(σ2x+σ2y+c2). | (8) |
We verified the effectiveness of each proposed component in our MDRN introduced before in detail on the same dataset under the same experiment setting. As shown in Table 1, we itemized the performance of specific methods.
Base | R1 | R2 | R3 | R4 | R5 | R6 | R7 | Ours | |
Multi-distillation (inside block) | √ | √ | √ | √ | √ | ||||
BSRB | √ | √ | √ | √ | √ | ||||
Using CCA | √ | √ | √ | √ | √ | ||||
Multi-distillation (outside block) | √ | √ | √ | √ | |||||
PSNR | 31.07 | 31.08 | 31.26 | 31.54 | 31.89 | 31.89 | 31.97 | 31.53 | 32.46 |
The Base refers to the model EDSR, which is a common residual block stacked in one path with one long skip connection, keeping the basic style of the mostly used SR SOTA model. The result of R1 shows the effectiveness of the distillation path outside the FMDRB. The result of R2 verifies the effectiveness of the basic unit (BSRB); as we can see, the block used alone enhances the performance, overtaking the model constructed from common residual blocks. The result of R3 shows the role of CCA in this proposed method. Results from R4 to R7 with/without the feature distillation operation outside/inside the proposed FMDRB, BSRB, and CCA obtain different SR results and outperform the before model, which further verifies the effectiveness of each proposed method. When the basic residual units (FMDRBs) are stacked in a chain manner, which is the common structure in the popular SR models, the model gets lower performance. However, when adding the feature distillation connections to the main chain of the residual blocks, which is the so-called FMDRB, the enhanced distillation block gets better performance.
The distillation structure is useful not only inside the enhanced distillation block but also outside the basic block. The result R6 is without/with the CCA layer; the result using CCA outperforms the result not using CCA, which verifies that the CCA layer maximizes the performance of FDRB.
We put the contrast-aware channel attention block in the tail position of the proposed FMDRB, which maximizes the capability of the proposed module. To prove the effectiveness of the attention module, we used other attention blocks for comparison, such as CA and IIA. As shown in Table 2, the results of the ablation study aiming at attention block show that the CCA is effective and has the best ability for immediate features.
Attention block | w/o | CA | IIA | CCA |
PSNR | 31.97 | 31.98 | 32.12 | 32.46 |
SSIM | 0.8767 | 0.8771 | 0.8778 | 0.8761 |
The proposed MDRN has inherited the advantages of the residual network and combines the advantages of the feature distillation network. To prove the excellent performance of MDRN, we compared our model with popular state-of-the-art SR models, including NTIRE2017 winner EDSR [10], RCAN [11], large-scale SAN [13], HAN [14], novel IGAN [15], RFDN [23], and the recent DINet [24]. Since most SR SOTA models are tested on DIV2K, which are 3-channel natural images, the performance comparison between different methods cannot be directly done from cited papers; they were re-tested on the MRI-brain dataset, composed of single-channel clinical images.
Table 3 demonstrates the comparison of quantitative results for 2×, 4×, and 8× SR. Our MDRN outperforms existing methods on MR-brain test datasets of all scales. Without using tricks like self-ensemble, the proposed MDRN network still achieves significant improvements compared to recent advanced methods. It is notably worth noticing that our model is much better than the EDSR, which shares a similar basic architecture with MDRN and shows some superiority over RFDN, which also uses the feature distillation strategy as MDRN. MDRN outperforms methods such as SAN and IGAN, which have more computationally intensive attention modules. Specifically, MDRN obtains superior results by 1.82 dB improvement in PSNR compared to the base EDSR in 4× scale, and its SSIM wins over previous methods. MDRN gains better results by up to 0.44 dB in terms of PSNR than DIPNet.
Memory | Time | Scale 2 | Scale 4 | Scale 8 | |
[M] | (ms) | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Bicubic | -- | -- | 33.66/0.9299 | 28.44/0.8159 | 24.40/0.6580 |
EDSR [10] | 2192.74 | 72.36 | 34.98*/0.9025* | 30.64*/0.8697* | 26.17*/0.7513* |
RCAN [11] | 2355.20 | 498.26 | 38.27*/0.9614* | 31.65**/0.9019* | 26.21*/0.7778* |
SAN [13] | 5017.60 | 805.23 | 34.85*/0.9318* | 31.09*/0.8432* | 25.39*/0.7359* |
IGAN [15] | 2099.20 | 335.77 | 33.91*/0.9173* | 31.73*/0.8744* | 26.32*/0.7804* |
HAN [14] | 5038.98 | 719.07 | 34.97*/0.9576* | 31.03*/0.8424* | 25.66*/0.7612* |
RFDN [23] | 813.06 | 49.51 | 38.31**/0.9620* | 31.98*/0.8795* | 26.28*/0.7794* |
DIPNet [24] | 521.02 | 28.79 | 38.27**/0.9614* | 32.02**/0.8712* | 26.33*/0.7884* |
Ours | 325.21 | 27.88 | 39.19/0.9686 | 32.46/0.8761 | 26.47/0.8696 |
*p < 0.05, **p < 0.001 |
The efficiency of a SR model can be assessed through various metrics, such as the number of parameters, runtime, computational complexity (FLOPs), and GPU memory consumption. These metrics play pivotal roles in the deployment of models in different aspects. Among these evaluation metrics, the runtime is the most direct indicator of a network's efficiency and is used as the primary evaluation metric. Memory consumption is also an important metric because it determines whether the model can be deployed to the edge device. In a clinical setting, the SR MRI model will be put into a small GPU, and models needing large-memory GPU will not work as intended. Our MDRN model gets the best PSNR, which is over 32 dB, only using 325.21 M GPU memory and 27.88 ms valid runtime, as shown in Table 3, showing a competitive advantage over other methods. To test the validation of experiment results, we analyzed the statistical significance of the results. As shown in Table 3, we calculated the P value of the results using the data of every epoch as a collection of random variables.
BraTS-Gli | BraTS-Meni | |
PSNR/SSIM | PSNR/SSIM | |
Bicubic | 32.94/0.9099 | 30.25/0.8689 |
EDSR [10] | 36.35*/0.9610* | 33.33*/0.9196* |
RCAN [11] | 36.94**/0.9513* | 33.86*/0.9160* |
SAN [13] | 37.06*/0.9514* | 34.02*/0.9191* |
IGAN [15] | 37.09*/0.9620* | 34.13*/0.9217* |
HAN [14] | 37.33*/0.9521* | 33.83*/0.9197* |
RFDN [23] | 38.17**/0.9600** | 34.08**/0.9214* |
DIPNet [24] | 38.38**/0.9623* | 34.17*/0.9218* |
Ours | 38.92/0.9635 | 34.25/0.9225 |
*p < 0.05, **p < 0.001 |
For a more intuitive demonstration of the gap between these methods, we show the comparison of zoomed results of various methods. As shown in Figure 3, we randomly select some results from the test set for evaluation. Taking "img_050112" as an example, most SR methods can reconstruct the general composition, but only IGAN and MDRN recover the more detailed textures and sharper edges. In zoomed details of "img_05011", we can see that IGAN, SAN, and RFDN do not clearly restore the small vessels, while our MDRN obviously does (shown in red arrows). Additionally, as seen in "img_05024", MDRN is closer to the ground truth, recovering the cerebrospinal fluid and not generating blurring artifacts (shown in yellow arrows). Our MDRN can output more high-frequency information, like enhanced contrast edges, than other methods. Through the observations of visual results, it is verified that MDRN has superiority in complex feature representations and recovery ability over previous works.
Deep learning-based methods have been proven to work effectively in the domain of medical image processing, including SR reconstruction for MR images. Based on the bottleneck of the SR task, we propose a novel lightweight and fast SR model named MDRN using multi-distillation residual learning.
Figure 4 provides an overview of the comparison of the performance and computation efficiency of the proposed method and other methods. It is evident that MDRN achieves the best execution time. Except for SAN and HAN using transformer structure, the computation complexity of SAN and HAN is O(n2) and of other models is O(n). The quadratic computation complexity O(n) in relation to the query/key/value sequence length n leads to high computation costs when using self-attention with a global receptive field. For a precise assessment of the computation complexity of our method, we compare it using quantitative metrics with several representative open-source models, as shown in Table 3. Quantitative results show that our MDRN consumes lower computation resources while maintaining 32+ PSNR. MDRN has a better trade-off between performance and cost.
We conducted generalization experiments by applying the super-resolution model trained on head and neck magnetic resonance imaging (MRI) images to pelvic CT images, aiming to validate the model's generalization performance on different datasets (Table 5). The results demonstrate that our model achieves a PSNR of 31.4 dB on the pelvic dataset at a 4× magnification factor. This outcome indicates that our MDRN exhibits favorable generalization performance and is capable of completing super-resolution tasks on new datasets. Visual quality is shown in Figure 5.
Scale | 2× | 4× | 8× |
PSNR | 36.55 | 32.35 | 27.79 |
SSIM | 0.8882 | 0.8938 | 0.8928 |
In this paper, we propose the MDRN, a lightweight CNN model, for efficient and fast super-resolution MRI tasks using the innovative multi-distillation strategy. Our findings show remarkable superiority of MDRN over current SR methods, supported by both quantitative metrics and visual evidence. Notably, MDRN excels at learning discriminative features and striking a better balance between computational efficiency and reconstruction performance by integrating the feature distillation mechanism into the network architecture. Extensive evaluations conducted on an MRI-brain dataset underline the favorable performance of MDRN over existing methods in both computational cost and accuracy for medical scenarios.
We declare that we have not used generative AI tools to generate the scientific writing of this paper.
We declare that we have no known financial interests or personal relationships that could have appeared to influence the work reported in this paper. There is no professional or other personal interest of any kind in any product, service or company that could influence the work reported in this paper.
[1] | Z. X. Liang and J. D. Glover, A zoom feature for a dynamic programming solution to economic dispatch including transmission losses, IEEE T. Power Syst., 7 (1992), 544–550. |
[2] | G. Xiong and D. Shi, Orthogonal learning competitive swarm optimizer for economic dispatch problems, Appl. Soft. Comput. J., 66 (2018), 134–148. |
[3] | R. A. Jabr, A. H. Coonick and B. J. Cory, A homogeneous linear programming algorithm for the security constrained economic dispatch problem, IEEE T. Power Syst., 15 (2000), 930–936. |
[4] | S. Muralidharan, K. Srikrishna and S. Subramanian, Self-adaptive dynamic programming technique for economic power dispatch, Int. J. Power Energy Syst., 27 (2007), 340–345. |
[5] | N. Sinha, R. Chakrabarti and P. K. Chattopadhyay, evolutionary programming techniques for economic load dispatch, IEEE T. Evol. Comput., 7 (2003), 83–94. |
[6] | J. Q. Li, H. Y. Sang, Y. Y. Han, et al., Efficient multi-objective optimization algorithm for hybrid flow shop scheduling problems with setup energy consumptions, J. Clean. Prod., 181 (2018), 584–598. |
[7] | J. Q. Li, S. C. Bai, P. Y. Duan, et al., An improved artificial bee colony algorithm for addressing distributed flow shop with distance coefficient in a prefabricated system, Int. J. Prod. Res., (2019). |
[8] | J. Q. Li, Q. K. Pan and K. Mao, A discrete teaching-learning-based optimisation algorithm for realistic flowshop rescheduling problems, Eng. Appl. Artif. Intell., 37 (2015), 279–292. |
[9] | J. Q. Li, P. Y. Duan, H. Y. Sang, et al., An efficient optimization algorithm for resource-constrained steelmaking scheduling problems, IEEE Access, 6 (2018), 33883–33894. |
[10] | J. Q. Li, P. Y. Duan, J. D. Cao, et al., A hybrid Pareto-based tabu search for the distributed flexible job shop scheduling problem with E/T criteria, IEEE Access, 6 (2018), 58883–58897. |
[11] | J. Q. Li, Q. K. Pan and M. F. Tasgetiren, A discrete artificial bee colony algorithm for the multi-objective flexible job-shop scheduling problem with maintenance activities, Appl. Math. Model., 38 (2014), 1111–1132. |
[12] | J. Q. Li, Q. K. Pan and S. X. Xie, An effective shuffled frog-leaping algorithm for multi-objective flexible job shop scheduling problems, Appl. Math. Comput., 218 (2012), 9353–9371. |
[13] | J. Q. Li, Q. K. Pan and K. Z. Gao, Pareto-based discrete artificial bee colony algorithm for multi-objective flexible job shop scheduling problems, Int. J. Adv. Manuf. Technol., 55 (2011), 1159–1169. |
[14] | J. Q. Li, A hybrid multi-objective artificial bee colony algorithm for flexible task scheduling problems in Cloud computing system, Cluster Comput., (2019), 1–24. |
[15] | P. Y. Duan, J. Q. Li, Y. Wang, et al., Solving chiller loading optimization problems using an improved teaching‐learning‐based optimization algorithm, Optim, Control Appl. Met., 39 (2018), 65–77. |
[16] | Z. X. Zheng, J. Q. Li and P.Y. Duan, Optimal chiller loading by improved artificial fish swarm algorithm for energy saving, Math. Comput. Simul., 155 (2019), 227–243. |
[17] | Z. X. Zheng and J. Q. Li, Optimal chiller loading by improved invasive weed optimization algorithm for reducing energy consumption, Energy Build, 161(2018), 80–88. |
[18] | S. Kumar and R. Naresh, Nonconvex economic load dispatch using an efficient real-coded genetic algorithm, Appl. Soft Comput. J., 9 (2009), 321–329. |
[19] | P. Subbaraj, R. Rengaraj and S. Salivahanan, Enhancement of Self-adaptive real-coded genetic algorithm using Taguchi method for Economic dispatch problem, Appl. Soft Comput. J., 11 (2011), 83–92. |
[20] | S. O. Orero and M. R. Irving, Large scale unit commitment using a hybrid genetic algorithm, Int. J. Electr. Power Energy Syst., 19 (1997), 45–55. |
[21] | P. Chen and H. Chang, Large-scale economic dispatch by genetic algorithm, IEEE T. Power Syst., 10 (1995), 1919–1926. |
[22] | A. I. Selvakumar, A new particle swarm optimization solution to nonconvex economic dispatch problems, IEEE T. Power Syst., 22 (2007), 42–51. |
[23] | J. B. Park, Y. W. Jeong, J. R. Shin, et al., An improved particle swarm optimization for nonconvex economic dispatch problems, IEEE T. Power Syst., 25 (2010), 156–166. |
[24] | Z. Gaing, Particle swarm optimization to solving the economic dispatch considering the generator constraints, IEEE T. Power Syst., 18 (2003), 1187–1195. |
[25] | K. T. Chaturvedi, M. Pandit and L. Srivastava, Self-organizing hierarchical particle swarm optimization for nonconvex economic dispatch, IEEE T. Power Syst., 23 (2008), 1079–1087. |
[26] | Q. Qin, S. Cheng, X. Chu, et al., Solving non-convex/non-smooth economic load dispatch problems via an enhanced particle swarm optimization, Appl. Soft Comput. J., 59 (2017), 229–242. |
[27] | X. S. Yang, S. S. S. Hosseini and A. H. Gandomi, Firefly Algorithm for solving non-convex economic dispatch problems with valve loading effect, Appl. Soft Comput. J., 12 (2012), 1180–1186. |
[28] | K. Bhattacharjee, A. Bhattacharya and S. H. N. Dey, Oppositional real coded chemical reaction optimization for different economic dispatch problems, Int. J. Electr. Power Energy Syst., 55 (2014), 378–391. |
[29] | A. S. Reddy and K. Vaisakh, Shuffled differential evolution for large scale economic dispatch, Electr. Power Syst. Res., 96 (2013), 237–245. |
[30] | D. Zou, S. Li, G. G. Wang, et al., An improved differential evolution algorithm for the economic load dispatch problems with or without valve-point effects, Appl. Energy, 181 (2016), 375–390. |
[31] | B. R. Adarsh, T. Raghunathan, T. Jayabarathi, et al., Economic dispatch using chaotic bat algorithm, Energy, 96 (2016), 666–675. |
[32] | A. K. Barisal and R. C. Prusty, Large scale economic dispatch of power systems using oppositional invasive weed optimization, Appl. Soft Comput. J., 29 (2015), 122–137. |
[33] | S. Banerjee, D. Maity and C. K. Chanda, Teaching learning based optimization for economic load dispatch problem considering valve point loading effect, Int. J. Electr. Power Energy Syst., 73 (2015), 456–464. |
[34] | M. A. Al-Betar, M. A. Awadallah, A. T. Khader, et al., Tournament-based harmony search algorithm for non-convex economic load dispatch problem, Appl. Soft Comput. J., 47 (2016), 449–459. |
[35] | M. Pradhan, P. K. Roy and T. Pal, Grey wolf optimization applied to economic load dispatch problems, Int. J. Electr. Power Energy Syst., 83 (2016), 325–334. |
[36] | T. Jayabarathi, T. Raghunathan, B. R. Adarsh, et al., Economic dispatch using hybrid grey wolf optimizer, Energy, 111 (2016), 630–641. |
[37] | M. Kumar and J. S. Dhillon, Hybrid artificial algae algorithm for economic load dispatch, Appl. Soft Comput., 71 (2018), 89–109. |
[38] | M. Modiri-Delshad and N. A. Rahim, Solving non-convex economic dispatch problem via backtracking search algorithm, Energy, 77 (2014), 372–381. |
[39] | J. J. Q. Yu and V. O. K. Li, A social spider algorithm for solving the non-convex economic load dispatch problem, Neurocomputing, 171 (2016), 955–965. |
[40] | A. I. Selvakumar and K. Thanushkodi, Optimization using civilized swarm: Solution to economic dispatch with multiple minima, Electr. Power Syst. Res., 79 (2009), 8–16. |
[41] | M. Basu, Kinetic gas molecule optimization for nonconvex economic dispatch problem, Int. J. Electr. Power Energy Syst., 80 (2016), 325–332. |
[42] | J. Cai, Q. Li, L. Li, et al., A hybrid FCASO-SQP method for solving the economic dispatch problems with valve-point effects, Energy, 38 (2012), 346–353. |
[43] | J. S. Alsumait, J. K. Sykulski and A. K. Al-Othman, A hybrid GA-PS-SQP method to solve power system valve-point economic dispatch problems, Appl. Energy, 87 (2010), 1773–1781. |
[44] | J. Cai, Q. Li, L. Li, et al., A hybrid CPSO-SQP method for economic dispatch considering the valve-point effects, Energy Convers. Manag., 53 (2012), 175–181. |
[45] | S. Sayah and A. Hamouda, A hybrid differential evolution algorithm based on particle swarm optimization for nonconvex economic dispatch problems, Appl. Soft Comput. J., 13 (2013), 1608–1619. |
[46] | A. R. Mehrabian and C. Lucas, A novel numerical optimization algorithm inspired from weed colonization, Ecol. Inform., 1 (2006), 355–366. |
[47] | T. A. A. Victoire and A. E. Jeyakumar, Reserve constrained dynamic dispatch of units with valve-point effects, IEEE T. Power Syst., 20 (2005), 1273–1282. |
[48] | T. Niknam and F. Golestaneh, Enhanced adaptive particle swarm optimisation algorithm for dynamic economic dispatch of units considering valve-point effects and ramp rates, IET Gener. Transm. Distrib., 6 (2012), 424–435. |
[49] | M. A. Elhameed and A. A. El-Fergany, Water cycle algorithm-based economic dispatcher for sequential and simultaneous objectives including practical constraints, Appl. Soft Comput. J., 58 (2017), 145–154. |
[50] | E. Afzalan and M. Joorabian, An improved cuckoo search algorithm for power economic load dispatch, Int. Trans. Electr. Energy Syst., 25 (2015), 958–975. |
[51] | Y. Labbi, D. B. Attous, H. A. Gabbar, et al., A new rooted tree optimization algorithm for economic dispatch with valve-point effect, Int. J. Electr. Power Energy Syst., 79 (2016), 298–311. |
[52] | N. Ghorbani and E. Babaei, Exchange market algorithm for economic load dispatch, Int. J. Electr. Power Energy Syst., 75 (2016), 19–27. |
Base | R1 | R2 | R3 | R4 | R5 | R6 | R7 | Ours | |
Multi-distillation (inside block) | √ | √ | √ | √ | √ | ||||
BSRB | √ | √ | √ | √ | √ | ||||
Using CCA | √ | √ | √ | √ | √ | ||||
Multi-distillation (outside block) | √ | √ | √ | √ | |||||
PSNR | 31.07 | 31.08 | 31.26 | 31.54 | 31.89 | 31.89 | 31.97 | 31.53 | 32.46 |
Attention block | w/o | CA | IIA | CCA |
PSNR | 31.97 | 31.98 | 32.12 | 32.46 |
SSIM | 0.8767 | 0.8771 | 0.8778 | 0.8761 |
Memory | Time | Scale 2 | Scale 4 | Scale 8 | |
[M] | (ms) | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Bicubic | -- | -- | 33.66/0.9299 | 28.44/0.8159 | 24.40/0.6580 |
EDSR [10] | 2192.74 | 72.36 | 34.98*/0.9025* | 30.64*/0.8697* | 26.17*/0.7513* |
RCAN [11] | 2355.20 | 498.26 | 38.27*/0.9614* | 31.65**/0.9019* | 26.21*/0.7778* |
SAN [13] | 5017.60 | 805.23 | 34.85*/0.9318* | 31.09*/0.8432* | 25.39*/0.7359* |
IGAN [15] | 2099.20 | 335.77 | 33.91*/0.9173* | 31.73*/0.8744* | 26.32*/0.7804* |
HAN [14] | 5038.98 | 719.07 | 34.97*/0.9576* | 31.03*/0.8424* | 25.66*/0.7612* |
RFDN [23] | 813.06 | 49.51 | 38.31**/0.9620* | 31.98*/0.8795* | 26.28*/0.7794* |
DIPNet [24] | 521.02 | 28.79 | 38.27**/0.9614* | 32.02**/0.8712* | 26.33*/0.7884* |
Ours | 325.21 | 27.88 | 39.19/0.9686 | 32.46/0.8761 | 26.47/0.8696 |
*p < 0.05, **p < 0.001 |
BraTS-Gli | BraTS-Meni | |
PSNR/SSIM | PSNR/SSIM | |
Bicubic | 32.94/0.9099 | 30.25/0.8689 |
EDSR [10] | 36.35*/0.9610* | 33.33*/0.9196* |
RCAN [11] | 36.94**/0.9513* | 33.86*/0.9160* |
SAN [13] | 37.06*/0.9514* | 34.02*/0.9191* |
IGAN [15] | 37.09*/0.9620* | 34.13*/0.9217* |
HAN [14] | 37.33*/0.9521* | 33.83*/0.9197* |
RFDN [23] | 38.17**/0.9600** | 34.08**/0.9214* |
DIPNet [24] | 38.38**/0.9623* | 34.17*/0.9218* |
Ours | 38.92/0.9635 | 34.25/0.9225 |
*p < 0.05, **p < 0.001 |
Scale | 2× | 4× | 8× |
PSNR | 36.55 | 32.35 | 27.79 |
SSIM | 0.8882 | 0.8938 | 0.8928 |
Base | R1 | R2 | R3 | R4 | R5 | R6 | R7 | Ours | |
Multi-distillation (inside block) | √ | √ | √ | √ | √ | ||||
BSRB | √ | √ | √ | √ | √ | ||||
Using CCA | √ | √ | √ | √ | √ | ||||
Multi-distillation (outside block) | √ | √ | √ | √ | |||||
PSNR | 31.07 | 31.08 | 31.26 | 31.54 | 31.89 | 31.89 | 31.97 | 31.53 | 32.46 |
Attention block | w/o | CA | IIA | CCA |
PSNR | 31.97 | 31.98 | 32.12 | 32.46 |
SSIM | 0.8767 | 0.8771 | 0.8778 | 0.8761 |
Memory | Time | Scale 2 | Scale 4 | Scale 8 | |
[M] | (ms) | PSNR/SSIM | PSNR/SSIM | PSNR/SSIM | |
Bicubic | -- | -- | 33.66/0.9299 | 28.44/0.8159 | 24.40/0.6580 |
EDSR [10] | 2192.74 | 72.36 | 34.98*/0.9025* | 30.64*/0.8697* | 26.17*/0.7513* |
RCAN [11] | 2355.20 | 498.26 | 38.27*/0.9614* | 31.65**/0.9019* | 26.21*/0.7778* |
SAN [13] | 5017.60 | 805.23 | 34.85*/0.9318* | 31.09*/0.8432* | 25.39*/0.7359* |
IGAN [15] | 2099.20 | 335.77 | 33.91*/0.9173* | 31.73*/0.8744* | 26.32*/0.7804* |
HAN [14] | 5038.98 | 719.07 | 34.97*/0.9576* | 31.03*/0.8424* | 25.66*/0.7612* |
RFDN [23] | 813.06 | 49.51 | 38.31**/0.9620* | 31.98*/0.8795* | 26.28*/0.7794* |
DIPNet [24] | 521.02 | 28.79 | 38.27**/0.9614* | 32.02**/0.8712* | 26.33*/0.7884* |
Ours | 325.21 | 27.88 | 39.19/0.9686 | 32.46/0.8761 | 26.47/0.8696 |
*p < 0.05, **p < 0.001 |
BraTS-Gli | BraTS-Meni | |
PSNR/SSIM | PSNR/SSIM | |
Bicubic | 32.94/0.9099 | 30.25/0.8689 |
EDSR [10] | 36.35*/0.9610* | 33.33*/0.9196* |
RCAN [11] | 36.94**/0.9513* | 33.86*/0.9160* |
SAN [13] | 37.06*/0.9514* | 34.02*/0.9191* |
IGAN [15] | 37.09*/0.9620* | 34.13*/0.9217* |
HAN [14] | 37.33*/0.9521* | 33.83*/0.9197* |
RFDN [23] | 38.17**/0.9600** | 34.08**/0.9214* |
DIPNet [24] | 38.38**/0.9623* | 34.17*/0.9218* |
Ours | 38.92/0.9635 | 34.25/0.9225 |
*p < 0.05, **p < 0.001 |
Scale | 2× | 4× | 8× |
PSNR | 36.55 | 32.35 | 27.79 |
SSIM | 0.8882 | 0.8938 | 0.8928 |