
Citation: David W. Schwartzman. Life’s Critical Role in the Long-term Carbon Cycle: the Biotic Enhancement of Weathering[J]. AIMS Geosciences, 2017, 3(2): 216-238. doi: 10.3934/geosci.2017.2.216
[1] | Keying Du, Liuyang Fang, Jie Chen, Dongdong Chen, Hua Lai . CTFusion: CNN-transformer-based self-supervised learning for infrared and visible image fusion. Mathematical Biosciences and Engineering, 2024, 21(7): 6710-6730. doi: 10.3934/mbe.2024294 |
[2] | Yongmei Ren, Xiaohu Wang, Jie Yang . Maritime ship recognition based on convolutional neural network and linear weighted decision fusion for multimodal images. Mathematical Biosciences and Engineering, 2023, 20(10): 18545-18565. doi: 10.3934/mbe.2023823 |
[3] | Yong Tian, Tian Zhang, Qingchao Zhang, Yong Li, Zhaodong Wang . Feature fusion–based preprocessing for steel plate surface defect recognition. Mathematical Biosciences and Engineering, 2020, 17(5): 5672-5685. doi: 10.3934/mbe.2020305 |
[4] | Zhijing Xu, Jingjing Su, Kan Huang . A-RetinaNet: A novel RetinaNet with an asymmetric attention fusion mechanism for dim and small drone detection in infrared images. Mathematical Biosciences and Engineering, 2023, 20(4): 6630-6651. doi: 10.3934/mbe.2023285 |
[5] | Dawei Li, Suzhen Lin, Xiaofei Lu, Xingwang Zhang, Chenhui Cui, Boran Yang . IMD-Net: Interpretable multi-scale detection network for infrared dim and small objects. Mathematical Biosciences and Engineering, 2024, 21(1): 1712-1737. doi: 10.3934/mbe.2024074 |
[6] | Shuai Cao, Biao Song . Visual attentional-driven deep learning method for flower recognition. Mathematical Biosciences and Engineering, 2021, 18(3): 1981-1991. doi: 10.3934/mbe.2021103 |
[7] | Jianguo Xu, Cheng Wan, Weihua Yang, Bo Zheng, Zhipeng Yan, Jianxin Shen . A novel multi-modal fundus image fusion method for guiding the laser surgery of central serous chorioretinopathy. Mathematical Biosciences and Engineering, 2021, 18(4): 4797-4816. doi: 10.3934/mbe.2021244 |
[8] | Xiaoli Zhang, Kunmeng Liu, Kuixing Zhang, Xiang Li, Zhaocai Sun, Benzheng Wei . SAMS-Net: Fusion of attention mechanism and multi-scale features network for tumor infiltrating lymphocytes segmentation. Mathematical Biosciences and Engineering, 2023, 20(2): 2964-2979. doi: 10.3934/mbe.2023140 |
[9] | Zhigao Zeng, Cheng Huang, Wenqiu Zhu, Zhiqiang Wen, Xinpan Yuan . Flower image classification based on an improved lightweight neural network with multi-scale feature fusion and attention mechanism. Mathematical Biosciences and Engineering, 2023, 20(8): 13900-13920. doi: 10.3934/mbe.2023619 |
[10] | Haifeng Song, Weiwei Yang, Songsong Dai, Haiyan Yuan . Multi-source remote sensing image classification based on two-channel densely connected convolutional networks. Mathematical Biosciences and Engineering, 2020, 17(6): 7353-7377. doi: 10.3934/mbe.2020376 |
As an image enhancement technology, image fusion is utilized to combine images captured by different types of sensors or under distinct shooting settings, aiming to obtain images with more comprehensive scene representation information [1]. Among various applications of image fusion, infrared and visible image fusion serves as a typical example. Infrared images are captured using infrared sensors, relying on thermal radiation. They are characterized by prominent targets, minimal environmental influence, high imaging noise and blurred details [2]. On the other hand, visible images exhibit rich texture details, high resolution and sensitivity to lighting conditions [3]. Image fusion enables the integration of unique and shared information from both modalities to generate fused images with enhanced texture and salient targets. These fused images play a crucial role in subsequent high-level vision tasks such as semantic segmentation [4] and nighttime vehicle target detection [5].
The research on infrared and visible image fusion has led to the development of various traditional methods that have been proposed [6,7,8,9,10]. These methods are often highly interpretable but rely on hand-designed fusion rules, which can limit their performance when dealing with more complex scene fusion tasks. However, with the advancements in deep learning, an increasing number of deep learning methods are being applied to image fusion tasks [11,12,13,14,15]. Deep networks exhibit strong capabilities in characterizing image features, surpassing the limitations of traditional feature extraction methods. They adopt a data-driven approach, enabling end-to-end generation of fused images.
In order to enhance the performance metrics of fusion, many existing deep learning-based fusion methods incorporate complex network modules that require more storage and computational resources to achieve improved performance metrics. For instance, Long et al. [16] proposed a network that aggregates residual dense blocks, combining dense connected blocks with residual connected blocks. Pu et al. [17] introduced a complex contextual information perceptual module for image reconstruction. Xu et al. [18] employed dissociative representation learning in an auto-encoder-based approach. These methods have demonstrated performance improvements in fusion results; however, they also introduce greater computational complexity due to the inclusion of complex modules in the network.
Furthermore, existing fusion algorithms often employ fusion layers that incorporate intricate fusion modules or fusion rules, with the primary aim of improving evaluation metrics. However, these algorithms often overlook the characteristics of different modalities. Notably, auto-encoder-based methods [18,19,20,21] utilize hand-designed fusion strategies for combining depth features. The use of such hand-designed fusion strategies may not assign proper weights to the depth features, leading to limitations in the performance of the fusion methods.
Currently, there is a lack of research on lightweight fusion models, which aim to reduce model parameters and convolutional depth channels. One example is PMGI [22], which performs information extraction through gradient and intensity scale preservation. It achieves this by reusing and fusing features extracted with fewer convolutional layers. Another lightweight model, FLFuse [23], generates fused images using a weight sharing encoder and feature swapping training strategy to ensure efficiency. However, FLFuse fails to fully extract and fuse image features due to its shallow network channel dimension and simplistic implicit fusion strategy, resulting in subpar visual effects and performance metrics.
We focus on exploring lightweight fusion methods based on structural re-parameterization. Existing structural re-parameterization methods have demonstrated high performance in training and fast inference speeds, making them effective for advanced vision tasks [24,25,26,27]. They are likely to be crucial in addressing the imbalance between fusion performance and computational resource consumption. However, directly applying these structural re-parameterization blocks designed for high-level vision tasks provides limited improvement for infrared and visible image fusion. Specific structural re-parameterization blocks tailored for fusion tasks are required to efficiently extract richer information from different modal features.
To address the limitations of existing image fusion methods, this paper proposes a novel approach that combines edge operators with structural re-parameterization. This approach enables the rapid generation of fused images with enhanced edge texture information and prominent targets, effectively addressing the imbalance between fusion performance and computational resource consumption. The major contributions of this paper are outlined as follows:
● A fast edge convolution fusion network (FECFusion) for infrared and visible images is proposed, which combines edge operations with structural re-parameterization for the first time to rapidly generate fused images with rich edge texture information and salient target, solving the problem of imbalance between fusion performance and computational resource consumption.
● A structural re-parameterization edge convolution block (RECB) is proposed, which can deeply mine the edge information in the source images and improve the performance of the fusion model without introducing additional inference burden.
● An attention fusion module (AFM) is designed to sufficiently fuze the unique and common information of different modal features to effectively integrate the feature information of the source images with less computational effort.
In the current literature, there are numerous works that primarily focus on preserving texture details in images [28,29,30]. In contrast, our work aims to address the challenge of balancing lightweight design and performance in image fusion networks. One approach is IFCNN [31], where both the encoder and decoder components employ only two convolutional layers for feature extraction and image reconstruction. Additionally, the fusion rules are adjusted based on the source image type, resulting in a unified network capable of handling various fusion tasks. Another method, SDNet [32], tackles the fusion task by incorporating the generated fused image reconstruction into a squeezed network structure of the source image. This forces the fused image to contain more information from the source images. SeAFusion [33] utilizes dense blocks with gradient residuals for feature extraction and employs the semantic segmentation task loss to guide the training of the fusion network. Recently, FLFuse [23] achieves feature extraction implicitly through a weight sharing encoder and feature swapping training strategy, enabling the generation of fused images in a lightweight and fast manner.
However, existing methods for infrared and visible image fusion only reduce the network model's parameters through conventional lightweight network design approaches, which can lead to a degradation in fusion performance.
Many existing methods for infrared and visible image fusion rely on attention mechanisms or multi-scale feature extraction to enhance network performance, but these approaches often come at the cost of increased computational complexity. Finding networks that effectively extract image features while maintaining high computational efficiency is challenging. In ACNet [34], Ding et al. proposed a method to convert multi-branch structures into a single-branch structure, thereby improving the performance of convolutional networks. In another work by Ding et al. [35], a concise VGG-like backbone network called RepVGG was introduced. RepVGG utilizes structural re-parameterization as its core technique, enabling efficient feature extraction while reducing network computation. RepVGG has demonstrated excellent performance in target detection tasks. Building upon this work, Ding et al. [36] proposed six network structures that can be structurally re-parameterized. The authors also explained the underlying reasons why the structural re-parameterization method is effective. Given the success of structural re-parameterization in various vision tasks, this approach holds promise for addressing the challenge of balancing fusion performance and computational resource consumption in infrared and visible image fusion tasks.
Unfortunately, the direct use of those structural re-parameterization blocks designed for advanced vision tasks provides little improvement for infrared and visible image fusion tasks. They still require the structural re-parameterization block specifically designed for the image fusion task to quickly extract the full wealth of information from the different modal features.
Therefore, we propose a new image fusion method, FECFusion, which substantially reduces computational resource consumption while maintaining high fusion performance through a well-designed structural re-parameterization technique.
In this paper, FECFusion utilizes end-to-end convolutional neural networks to perform feature extraction, feature fusion and image reconstruction, enabling efficient and straightforward fusion tasks. The network architecture, as depicted in Figure 1, consists of three main components: an encoder, a fusion layer and a decoder. In the encoder, a two-branch structure is employed, comprising one convolutional layer and two structural re-parameterization edge convolution blocks. This setup allows for the extraction of depth features from both the infrared and visible images. The fusion layer combines these extracted features, leveraging the complementary information present in the two modalities. Subsequently, the decoder, consisting of three structural re-parameterization edge convolution blocks, reconstructs the hybrid features obtained from the fusion layer to generate the final fused image. Overall, FECFusion offers a simple and efficient solution for infrared and visible image fusion, utilizing end-to-end convolutional neural networks for image fusion of the two modalities.
To ensure that the fused images better retain the edge feature information of the source images, FECFusion has designed a structural re-parameterization edge convolution block(RECB) for improving the performance in infrared and visible image fusion tasks. In addition, we use an attention fusion module (AFM) to better fuse the feature information of different modal images extracted from different branches.
Although it has some effects to use standard convolution to extract infrared and visible image feature information for fusion networks, it is inferior to complex models in terms of fusion performance. However, replacing standard convolution with complex blocks would make the network consume more computational resources. Therefore, the structural reparameterization technique is introduced in this paper to enrich the characterization capability of the network without increasing the computational resource consumption of the network in the inference stage.
To ensure that the fused images are better able to retain the edge feature information of the source images, we design a structural re-parameterization edge convolution block(RECB) for improving the performance in infrared and visible image fusion task. The specific structure of RECB is shown in Figure 2.
In particular, the RECB consists of four elaborated operators as follows.
1) The branch of standard 3 × 3 convolution
To guarantee the basic performance of the module, we use a standard 3 × 3 convolution. This convolution is represented as:
Fn=Kn∗X+Bn, | (3.1) |
where Fn represents the output feature of 3 × 3 standard convolution. Kn represents the convolution kernel weight of 3 × 3 standard convolution. X represents the input feature. Bn represents the offset of 3 × 3 standard convolution.
2) The branch of feature expansion convolution
The representational power of the fusion task is improved by expanding the channels of the features, which helps to improve the extraction of more feature information. Specifically, the branch uses 1 × 1 convolution to expand the channel dimension of the features and 3 × 3 convolution to extract the feature information, which is expressed as:
Fe=Kn∗(Ke∗X+Be)+Bn, | (3.2) |
where Fe represents the output feature of the feature expansion convolution branch. Ke represents the convolution kernel of 1 × 1 convolution. Be represents the offset of 1 × 1 convolution.
3) The branch of Sobel filter
Edge information is tremendously helpful for the performance improvement of the fusion task. Since it is usually difficult for the network model to learn the weights of the edge detection filters through training, a pre-defined Sobel edge filter is embedded in this branch for extracting the first-order spatial derivatives and learning the scaling factors of the filters. Specifically, the input features are firstly scaled by 1 × 1 convolution, then the edge information is extracted by horizontal and vertical Sobel filters, which are processed as follows:
Dx=[+10−1+20−2+10−1]andDy=[+1+2+1000−1−2−1], | (3.3) |
FDx=(SDx⋅Dx)⊗(Kx∗X+Bx)+BDx,FDy=(SDy⋅Dy)⊗(Ky∗X+By)+BDy, | (3.4) |
Fsobel=FDx+FDy. | (3.5) |
where Kx, Bx and Ky, By are the weights, bias of the 1 × 1 convolution in the horizontal and vertical directions. SDx, BDx and SDy, BDy are the scaling parameters and bias with the shape of C× 1 × 1 × 1. ⊗ and ∗ represent DWConv and normal convolution. (SDx⋅Dx), (SDy⋅Dy) are in the shape of C× 1 × 3 × 3.
4) The branch of Laplacian filter
In addition to the Sobel operator for extracting the first-order spatial derivatives, this branch employs a more stable Laplacian edge filter that is more robust to noise to extract the second-order spatial derivatives of the image edge information. Similarly, this branch also uses 1 × 1 convolution for scaling and then uses the Laplacian operator to extract the edge information, processed as:
Dlap=[0+10+1−4+10+10], | (3.6) |
Flap=(Slap⋅Dlap)⊗(Klap∗K+Blap)+Blap. | (3.7) |
where Klap, Blap are the weights, bias of the 1 × 1 convolution. Slap, Blap are scaling factors and bias of DWConv, respectively.
In addition, the BN layer is not used in the RECB, unlike the structural re-parameterization block designed for advanced vision tasks, because the BN layer would hinder the performance of the fusion network. Finally, the output features of these four branches are summed and mapped to the nonlinear activation layer:
F=Fn+Fe+Fsobel+Flap. | (3.8) |
where F is the output characteristic of RECB. The nonlinear activation layer used in this experiment is LeakyRelu.
The above RECB is the structure of the training phase. After the training is completed, the parameters of the four branch structures are equivalent to a 3 × 3 convolution parameter through the structure re-parameterization technique, so that the same effect can be obtained only through the 3 × 3 convolution processing after the structure re-parameterization in the inference phase.
Since these two features come from source images of different modalities, they have object focus of different scenes, with complementary information and public information of each other. Thus, the fusion module should have to focus on the fusion of complementary information and public information of different modalities.
In Figure 3, it is evident that detecting pedestrians in visible images at night can be challenging due to inadequate lighting conditions. However, in infrared thermal images, pedestrians are clearly highlighted. Therefore, the key challenge lies in fusing these two features by leveraging the complementary information that exists in only one of the modalities. In the case of a well-illuminated thermal target, such as the vehicle in Figure 3, both cameras are capable of sensing it. During the fusion process, it is important to enhance both features simultaneously. If a method is used that focuses solely on processing the complementary information, there is a risk of weakening one of the features. In order to effectively fuse the infrared and visible features, it is crucial to devise a fusion approach that preserves and enhances the salient information from both modalities. This will ensure that both the infrared highlights and the visible details are effectively integrated, leading to improved detection results.
In order to better solve the problem of fusion of different modal information, the element-by-element addition method for extracting complementary information of heteromodal images and the element-by-element multiplication method for extracting common information of heteromodal images are employed here. The element-by-element addition and element-by-element multiplication methods are expressed as:
Xadd=Xvi+Xir,Xmul=Xvi∗Xir. | (3.9) |
where Xvi and Xirrepresent the depth features of infrared and visible images extracted by the encoder, respectively. Element-by-element addition Xadd represents the addition of elements to visible image features and thermal target features to accumulate complementary information of different modes, while element-by-element multiplication Xmul represents the multiplication of elements to visible image features and thermal target features to enhance common information of different modes.
As in the simple example in Figure 4, 0 represents that the target is not sensed by the sensor and 1 represents that the target is sensed by the sensor. Suppose there are two cases of sensed and un-sensed target for infrared and visible sensors, respectively, which are illustrated by 0 and 1. Therefore, four cases are generated, and the goal is to retain all the information as long as possible. Element-by-element addition preserves the target information to the maximum extent possible, the left in Figure 4, with the target being sensed by at least one of the sensors. Element-by-element multiplication allows filtering out the common information, the right image in Figure 4, with the target needing to be sensed by two sensors at the same time. Once the two types of information are obtained separately, they are preserved by feature concatenation, combining the unique information from both modalities. This approach allows for the retention of crucial information while effectively combining the features extracted from each sensor.
Therefore, the attention fusion module (AFM) based on element-by-element addition and element-by-element multiplication is designed to better fuse the information of these two depth features. The specific structure of the AFM is shown in Figure 5.
The upper branch in AFM is used to enhance the common information of different modal features, while the bottom branch enhances the feature information of different modal features through the attention module and aggregates the complementary information of different modal features by feature summation, and then cascades the common and complementary information to allow both feature information to be retained as much as possible.AFM is processed as follows:
Y=Cat(Xvi∗Xir,A(Xvi)+A(Xir)). | (3.10) |
where Cat represents concatenation; A represents the attention module, and CBAM is used in this article.
The loss function is the key to guide the training of deep neural networks to achieve the desired results. The loss of structural similarity is often used to maintain a clear intensity distribution of the fused image. However, in the fusion task, the fused image needs to be similar to the two source images at the same time, and there is more complementary information between the two source images, which will weaken the complementary information region and lead to a decrease in fusion performance.
In order to better promote the recovery of texture details, this paper uses content loss Lcontent and traditional loss Ltradition to jointly constrain the training of the network. The total loss Lfusion formula is as follows:
Lfusion=λLtradition+Lcontent. | (3.11) |
where λ is the weight coefficient to balance these two losses.
The design of traditional loss function can enhance the similarity between the fused image and the two types of source images, guide the network to generate the fusion result with complete information faster, and avoid the single and incomplete information in the fusion result. The traditional loss Ltradition calculation formula is :
Ltradition=1HW‖If−0.5∗(Iir+Ivi)‖1. | (3.12) |
where If represents the fused image. Iir and Ivi represent the infrared and visible images. ‖⋅‖ refers to the l1-norm.
In order to promote the model to fuse more meaningful information, retain the saliency in the infrared image and the edge texture information of the source image, the content loss Lcontent with bilateral filtering is designed in this paper. The content loss consists of two parts: the intensity loss Lin and the edge gradient loss with bilateral filtering Lgrad. The formula is as follows:
Lcontent=μ1Lin+μ2Lgrad. | (3.13) |
where μ1, μ2 are the weighting coefficients for balancing these two losses.
Among them, the intensity loss Lin constrains the overall apparent intensity of the fused image. In order to better retain the salient target, the pixel intensity of the fused image should be biased towards the maximum intensity of the infrared and visible images. The formula of strength loss is as follows:
Lin=1HW‖If−Max(Iir,Ivi)‖1. | (3.14) |
where Max(⋅) stands of the element-wise maximum calculation.
In addition, in order to make the network model better preserve the edge texture details of the fused image, the existing methods use the maximum edge gradient of the source image to constrain the training of the network, but this loss is easily affected by noise in the infrared image. To this end, this paper uses a bilateral filter that preserves the edge gradient to denoise the infrared image, thereby reducing the noise of the fused image. The following is the calculation formula of edge gradient loss of bilateral filtering:
Lgrad=1HW‖|∇If|−Max(|∇Bila(Iir)|,|∇Ivi|)‖1. | (3.15) |
where ∇ is the gradient operator for measuring image texture information, and the Sobel operator is used to calculate the gradient in this paper. |⋅| indicates the absolute operation. Bila represents a bilateral filter.
In this paper, FECFusion is trained with the MSRS dataset [37]. Since the existing infrared and visible image fusion dataset is small, the MSRS training set part of the infrared and visible images common dataset is expanded from 1083 to 26,112 pairs of images, and the size of the training set image pairs after data enhancement is 64 pixel × 64 pixel, which can basically meet the training requirements. In order to evaluate the effectiveness of FECFusion, it is tested on the test set of the MSRS dataset and picks 361 pairs of images as the test subject. In addition, to more comprehensively evaluate the generalization performance of FECFusion, it selected 42 and 300 pairs of images on the TNO[38] and M3FD[39] datasets, respectively, for generalization comparison experiments.
Since none of the current public datasets for infrared and visible image fusion have reference images, the quality of the fusion result images cannot be directly evaluated by ground truth, therefore, we evaluate the visualization image effects of different algorithms by human subjective visual perception as a qualitative assessment, and by objective generic image quality evaluation index results as a quantitative assessment.
In this paper, standard deviation (SD) [40], mutual information (MI) [41], visual information fidelity (VIF) [42], sum of correlation differences (SCD) [43], entropy (EN) [44] and Qabf [45] are used. SD evaluates the contrast and distribution of the fused images from a statistical point of view. MI measures the amount of information from the source image to the fused image. VIF reflects the fidelity of the fused information from a human visual point of view. SCD measures the difference between the source image and the fused image. EN measures the amount of information contained in the image. Qabf evaluates the amount of fused edge information from the source image. All the above metrics are positive metrics, and higher values mean better fusion results.
FECFusion is compared with seven fusion algorithms, including DenseFuse[46], FusionGAN[47], IFCNN[31], SDNet[32], U2Fusion[48], FLFuse[23] and PIAFusion[37]. All the compared algorithms are experimented in public code, where the relevant settings of the experiments are kept constant. In the superparameter settings of the proposed network, the network optimizer uses Adam, epoch = 10, batch size = 64, learning rate is 1 ×10−4, loss function parameters are λ=10, μ1=12, μ2=45. The parameters of bilateral filtering are σd=0.05, σr=8.0, and window size is 11 × 11. The training process of FECFusion is summarized in Algorithm 1.
Besides the comparative and generalization experiments, the effectiveness of RECB and AFM is verified by ablation experiments in this paper. In addition, FECFusion is verified to be helpful for the advanced vision task through segmentation experiments. Finally, we have compared the operational efficiency of FECFusion with other methods and compare the computational resource consumption with and without structural re-parameterization. Our experiments are all conducted on a GeForce RTX 2080Ti 11GB and an Intel Core i5-12600KF, with PyTorch of a deep learning framework.
Algorithm 1: Training procedure |
Input: Infrared images Iir and visible images Ivi Output: Fused images If ![]() |
It is an important challenge for the image fusion algorithm to generalize the performance of different scenes. In the MSRS dataset, we have chosen two daytime and two nighttime images to evaluate the subjective visualization performance, and the comparison results are shown in Figures 6 and 7. We mark the texture detail information with green boxes and the highlighted target information with red boxes.
In the daytime scene depicted in Figure 6, we can observe the performance of different fusion methods. DenseFuse, SDNet and U2Fusion fail to effectively highlight the infrared target and do not fully utilize the background information present in the visible image. FusionGAN manages to highlight the salient target to some extent, but it can be seen from the green box that it blurs the background. In contrast, IFCNN and FLFuse weaken the texture details of the background, as evident from the green box. Only PIAFusion and the method proposed in this paper successfully integrate the relevant information, effectively preserving both the infrared target and the background texture details. Therefore, it is evident that the method proposed in this paper exhibits superior performance in the daytime scene, achieving a balanced fusion result that highlights the target while retaining the background information.
In the night scene depicted in Figure 7, the visible image contains limited texture information, while the infrared image contains both background texture details and a salient target. Many existing fusion methods tend to overemphasize the information from one modality, making it challenging to achieve satisfactory results across different scenes.
Among the fusion methods examined, DenseFuse, SDNet and U2Fusion exhibit a bias towards infrared images, and weaken infrared targets. FusionGAN introduces artifacts into the fused images. Only IFCNN, PIAFusion, and the method proposed in this paper are capable of generating fused images with higher contrast in black night scenes. FLFuse, which is also a lightweight method, performs poorly in this scenario. It fails to fully leverage the characteristics of both modal images, leading to a degradation in both the background and the infrared target. Therefore, the method proposed in this paper demonstrates good performance in night scenes as well. It effectively captures the characteristics of both modal images and achieves better contrast, thereby preserving the infrared target while retaining background details.
In this section, we perform quantitative evaluation on the MARS dataset and select six metrics for evaluation. The comparison of the metrics of different methods is shown in Table 1, where red represents the best result and blue represents the second one.
Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | |
DenseFuse | 7.0692 | 2.5409 | 0.6752 | 1.3296 | 5.8397 | 0.3552 |
FusionGAN | 5.4694 | 1.9155 | 0.4253 | 0.8015 | 5.2260 | 0.1208 |
IFCNN | 7.5947 | 2.7399 | 0.8283 | 1.6658 | 6.3109 | 0.5540 |
SDNet | 5.3258 | 1.7398 | 0.3758 | 0.8364 | 4.8891 | 0.2944 |
U2Fusion | 5.6231 | 1.8953 | 0.3967 | 1.0034 | 4.7525 | 0.2908 |
FLFuse | 6.4790 | 2.0697 | 0.4860 | 1.1189 | 5.5157 | 0.3198 |
PIAFusion | 7.9268_ | 4.1774 | 0.9072_ | 1.7395_ | 6.4304_ | 0.6324 |
Ours | 8.1413 | 3.6805_ | 0.9282 | 1.8153 | 6.5104 | 0.5619_ |
From Table 1, it is clear that our method shows significant advantages in four metrics, SD, VIF, SCD and EN, while its performance in MI and Qabf is second only to PIAFusion. The value of SD is the best indicating that the fusion result of this paper method Shencheng achieves high contrast between infrared target and background; the highest value of VIF indicates that the fused image generated by this paper method is more in line with The highest value of VIF indicates that the fused images generated by this method are more consistent with the human visual system; the highest values of SCD and EN indicate that this method can generate fused images with more edge details and contain more realistic results than other methods. In conclusion, the quantitative experimental results show that this method can generate fused images with more information while reducing the computational effort.
In the fusion task, it is required that the fusion model has a stronger generalization capability, which is applicable in different scenes. Therefore, we selected 20 and 300 pairs of images in the TNO and M3FD datasets, respectively, to evaluate the generalization ability of FECFusion. A qualitative comparison of the different algorithms on the TNO and M3FD datasets is presented in Figures 8 and 9.
From the figures, it is evident that DenseFuse, SDNet, U2Fusion and FLFuse tend to blend the background and the target together, making it difficult to distinguish the salient infrared target. FusionGAN, on the other hand, exhibits high overlap with the infrared image and lacks the inclusion of background information. In comparison, IFCNN and PIAFusion show performance similar to the proposed method in this paper. However, it is important to note that these methods may not match the inference speed of the proposed method, which offers faster processing capabilities. Therefore, based on objective evaluation and considering the faster inference speed, the proposed method in this paper demonstrates competitive performance and provides a promising solution for infrared and visible image fusion tasks.
The results of quantitative metrics for the generalization experiments are shown in Table 2. The metrics performance of our method is the best or the second best on both datasets, which indicates that our method can both preserve the texture details of the source image and improve the contrast of the target. In conclusion, the qualitative and quantitative results show that FECFusion performs excellently in generalization. In addition, the method in this paper effectively maintains the intensity distribution of the target region and preserves the texture details of the background region, benefiting from the proposed RECB and AFM.
Dataset | Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
TNO | DenseFuse | 8.5765 | 2.1987 | 0.6704 | 1.5916 | 6.3422 | 0.3427 |
FusionGAN | 8.6703 | 2.3353 | 0.6541 | 1.3788 | 6.5578 | 0.2339 | |
IFCNN | 9.0058 | 2.4154 | 0.7996 | 1.6850 | 6.7413 | 0.5066 | |
SDNet | 9.0679 | 2.2606 | 0.7592 | 1.5587 | 6.6947 | 0.4290 | |
U2Fusion | 8.8553 | 1.8730 | 0.6787 | 1.5862 | 6.4230 | 0.4245 | |
FLFuse | 9.2628 | 2.1925 | 0.8084 | 1.7308 | 6.3658 | 0.4177 | |
PIAFusion | 9.1093 | 3.2464 | 0.8835 | 1.6540 | 6.8937 | 0.5556 | |
Ours | 9.2721 | 3.7136 | 0.9496 | 1.7312 | 6.9856 | 0.5311 | |
M3FD | DenseFuse | 8.6130 | 2.8911 | 0.6694 | 1.5051 | 6.4264 | 0.3709 |
FusionGAN | 8.8571 | 2.9921 | 0.5176 | 1.1292 | 6.4750 | 0.2530 | |
IFCNN | 9.2815 | 2.9560 | 0.7738 | 1.5353 | 6.6966 | 0.6053 | |
SDNet | 8.8855 | 3.1798 | 0.6329 | 1.3914 | 6.6102 | 0.5005 | |
U2Fusion | 9.0141 | 2.7531 | 0.7061 | 1.5488 | 6.6285 | 0.5303 | |
FLFuse | 8.7580 | 3.2425 | 0.6986 | 1.4975 | 6.5744 | 0.2640 | |
PIAFusion | 10.1639 | 4.6942 | 0.9300 | 1.3363 | 6.8036 | 0.6348 | |
Ours | 9.9899 | 4.3123 | 0.9350 | 1.5502 | 6.7685 | 0.6440 |
In order to verify the effectiveness of adding RECB and AFM to our FECFusion, ablation experiments are designed in this section to further analyze the role of these two proposed modules in the network model. First, RECB is a structure that equates multiple branch structures into a single branch structure by structural re-parameterization; therefore, in the ablation experiments, the RECB part is directly replaced with the structure after structural re-parameterization for training, i.e., the structure of a single ordinary convolution. For the ablation experiments of AFM, the network is trained with a direct feature cascade instead of AFM. This experiment is performed on the MSRS dataset, and the experimental results are shown in the Figure 10, where the background texture is marked with a green solid box and the infrared salient targets are marked with a red solid box.
From the experimental results, it can be seen that without RECB, the fused images are blurred at the edges to some extent, which proves that the module contributes to maintaining the edge information of the fusion results. Without AFM, the saliency of the fusion results decreases. If both RECB and AFM are not available, the fusion results have a decreased target significance and blurred edge texture. The evaluation metrics for this ablation experiment are shown in Table 3. We observe that the absence of both RECB and AFM leads to a decrease in the evaluation metric values to different degrees, proving the effectiveness of each part of our FECFusion.
RECB | AFM | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
✓ | ✓ | 8.1413 | 3.6805 | 0.9282 | 1.8153 | 6.5104 | 0.5619 |
✓ | ✘ | 7.4551 | 3.0648 | 0.7234 | 1.6117 | 6.0100 | 0.4488 |
✘ | ✓ | 7.7954 | 3.0922 | 0.8355 | 1.8060 | 6.2925 | 0.5098 |
✘ | ✘ | 6.6285 | 2.6117 | 0.5500 | 1.2793 | 5.6158 | 0.4362 |
To verify the execution efficiency of the proposed algorithm, the average processing time of forward propagation of each fusion method is tested on the MSRS dataset in this paper, and the comparison results are shown in Table 4 where red represents the best and blue represents the second best. It can be seen that our method is more efficient than most methods, while FLFuse is faster than our method, the method in this paper works better, so these differences in running efficiency are acceptable.
Algorithm | DenseFuse | FusionGAN | IFCNN | SDNet | U2Fusion | FLFuse | PIAFusion | Ours |
Running time | 0.374 | 0.082 | 0.019 | 0.014 | 0.155 | 0.001 | 0.081 | 0.002_ |
In addition, with reference to the image size used in the MSRS data set, the data of 640 × 480 × 1 is used as the input of the network forward propagation, and the parameters and weights of the network model are calculated by the TorchSummary library. The forward propagation time, running storage space, parameter quantity, weight size and the cumulative deviation of the fusion results pixel by pixel before and after the structural re-parameterization are compared experimentally. The comparison results are shown in Table 5.
Structural re-parameterisation | Forward time | Forward pass size | Params | Params size | Cumulative pixel deviation of results |
W/O | 0.0299 s | 5451.57 MB | 146,389 | 0.56 MB | / |
W | 0.0020 s | 2601.57 MB | 145,477 | 0.55 MB | 1 × 10-4 |
By comparing the results, it can be seen that there is almost no difference in the fusion results of the network before and after the structural re-parameterization, indicating that the structural re-parameterization can effectively reduce the running time, running storage space, parameter quantity and weight size in the case of very low deviation.
Semantic segmentation algorithms are an important general-purpose computer vision method whose performance reflects well on the semantic information of the fused resultant image. To verify that the fused images can be helpful for subsequent vision tasks, DeepLabV3+[49], a semantic segmentation model pre-trained on the Cityscapes dataset[50], is also used in this section to evaluate the performance of the fused images, and the semantic segmentation results are shown in the Figure 11.
From the experimental results, the semantic segmentation results of the fused result images are all a little better than the infrared and visible images, especially at night when the lighting conditions are poorer, the visible sensors have difficulty capturing enough information, and the semantic segmentation models often have difficulty detecting hot targets such as travelers, so to some extent, it can be shown that the image fusion has an enhanced effect on subsequent vision tasks.
In this paper, we propose FECFusion, an infrared and visible image fusion network based on fast edge convolution. The network consists of several key components. First, the main part of the network employs the RECB to extract features, including detailed texture features and salient image features. These extracted features are then fused using the AFM, and the fused image is reconstructed. After the completion of training, the network undergoes a structural re-parameterization operation to optimize the inference speed and storage space required while preserving the original training effectiveness. Through subjective and objective experimental results, we demonstrate that FECFusion achieves superior fusion results compared to other algorithms. It offers better real-time performance and requires less inference memory footprint, making it more suitable for practical engineering applications that involve the design of custom hardware accelerated circuits. In future research, we will explore specific applications of FECFusion on mobile devices and further optimize its performance. This includes enhancing the network's ability to learn multi-scale image features and achieve better fusion results with lower computational resource consumption.
The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work is supported by National Natural Science Foundation of China(NO.62266025).
The authors declare there is no conflict of interest.
[1] | Brantley SL, Megonigal JP, Scatena FN, et al. (2011) Twelve testable hypotheses on the geobiology of weathering. Geobiology 9: 140-165. |
[2] |
Akob DM, and Küsel K (2011) Where microorganisms meet rocks in the Earth's Critical Zone. Biogeosciences 8: 3531-3543. doi: 10.5194/bg-8-3531-2011
![]() |
[3] | Brecheisen ZS, Richter D deB (2014) Ordering interfluves: a simple proposal for understanding critical zone evolution. Procedia Earth Planetary Sci 10: 77–81. |
[4] | Field JP, Breshears DD, Law DJ, et al. (2015) Critical Zone Services: Expanding Context, Constraints, and Currency beyond Ecosystem Services. Vadose Zone J 14. |
[5] | Summers S, Thomson BC, Whiteley A, et al. (2016). Mesophilic mineral weathering bacteria inhabit the critical-zone of a perennially cold basaltic environment. Geomicrobiology J 33: 52-62. |
[6] |
Schwartzman DW, Volk T (1989) Biotic enhancement of weathering and the habitability of Earth. Nat 340: 457-460. doi: 10.1038/340457a0
![]() |
[7] | Schwartzman D (1999 2002) Life, Temperature, and the Earth: The Self-Organizing Biosphere. New York: Columbia University Press. |
[8] | Schwartzman DW (2008) Coevolution of the Biosphere and Climate, In: S.E. Jorgensen SE, Fath B (eds.), Encyclopedia of Ecology, 1st Edition, Oxford: Elsevier B.V., 648-658. |
[9] | Brantley SL, Lebedeva M, Hausrath EM (2012) A geobiological view of weathering and erosion. In Knoll AH, Canfield DE, Konhauser KO (eds.) Fundamentals of Geobiology, 1st Edition., Oxford: Blackwell Publishing Ltd., 205-227. |
[10] | Lovelock JE, Watson AJ (1982) The regulation of carbon dioxide and climate: Gaia or geochemistry. Planet. Space Sci. 30: 795-802. |
[11] | Richter D deB, Billings SA (2015) "One physical system": Tansley's ecosystem as Earth's critical zone. New Phytol 206: 900-912. |
[12] | Bray AW, Oelkers EH, Bonneville S, et al. (2013) How bugs get their food: Linking mineral surface chemistry to nutrient availability. Mineral Mag 77: 765. |
[13] | Quirk J, Beerling DJ, Banwart SA, et al. (2012) Evolution of trees and mycorrhizal fungi intensifies silicate mineral weathering. Biol Lett 8: 1006-1011. |
[14] | Quirk J, Andrews MY, Leake JR, et al. (2014) Ectomycorrhizal fungi and past high CO2 atmospheres enhance mineral weathering through increased below-ground carbon-energy fluxes. Biol Lett 10: 387-393. |
[15] |
Quirk J, Leake JR, Banwart SA, et al. (2014) Weathering by tree-root-associating fungi diminishes under simulated Cenozoic atmospheric CO2 decline. Biogeosciences 11: 321-331. doi: 10.5194/bg-11-321-2014
![]() |
[16] |
Burghelea C, Zaharescu DG, Dontsova K, et al. (2015) Mineral nutrient mobilization by plants from rock: influence of rock type and arbuscular mycorrhiza. Biogeochemistry 124:187-203. doi: 10.1007/s10533-015-0092-5
![]() |
[17] |
Callesen I, Harrison R, Stupak I, et al. (2016) Carbon storage and nutrient mobilization from soil minerals by deep roots and rhizospheres. For Ecol Manage 359: 322-331. doi: 10.1016/j.foreco.2015.08.019
![]() |
[18] | Quirk J, Leake JR, Johnson DA, et al. (2015) Constraining the role of early land plants in Palaeozoic weathering and global cooling. Proc R Soc B 282: 20151115. |
[19] |
Taylor LL, Leake JR, Quirk J, et al. (2009) Biological weathering and the long-term carbon cycle: integrating mycorrhizal evolution and function into the current paradigm. Geobiology 7: 171-191. doi: 10.1111/j.1472-4669.2009.00194.x
![]() |
[20] | Lucas Y (2001) The role of plants in controlling rates and products of weathering: Importance of biological pumping. Annu Rev Earth Planet Sci 29: 135-163. |
[21] | Buss HL, Sak PB, Webb SM, et al. (2008) Weathering of the Rio Blanco quartz diorite, Luquillo Mountains, Puerto Rico: Coupling oxidation, dissolution, and fracturing. Geochim Cosmochim Acta 72: 4488-4507. |
[22] |
Navarre-Sitchler AK, Cole DR, Rother G, et al. (2013) Porosity and surface area evolution during weathering of two igneous rocks. Geochim Cosmochim Acta 109: 400-413. doi: 10.1016/j.gca.2013.02.012
![]() |
[23] | Dorn RI (2014) Ants as a powerful biotic agent of olivine and plagioclase dissolution. Geol 42: 771-774. |
[24] |
Chadwick OA, Derry LA, Vitousek PM, et al. (1999) Changing sources of nutrients during four million years of ecosystem development. Nat 397: 491-497. doi: 10.1038/17276
![]() |
[25] | Chadwick KD, Asner GP (2016) Tropical soil nutrient distributions determined by biotic and hillslope processes. Biogeochemistry 127: 273-289. |
[26] |
Porder S, Clark DA, Vitousek PM (2006) Persistence of rock-derived nutrients in the wet tropical forests of La Selva, Costa Rica. Ecology 87: 594-602. doi: 10.1890/05-0394
![]() |
[27] |
Callesen I, Harrison R, Stupak I, et al. (2016) Carbon storage and nutrient mobilization from soil minerals by deep roots and rhizospheres. For Ecol Manage 359: 322-331. doi: 10.1016/j.foreco.2015.08.019
![]() |
[28] |
Calmels D, Gaillardet J, Brenot A, et al. (2007) Sustained sulfide oxidation by physical erosion processes in the Mackenzie River basin: Climatic perspectives. Geol 35: 1003-1006. doi: 10.1130/G24132A.1
![]() |
[29] |
Torres MA, West AJ, Li G (2014) Sulphide oxidation and carbonate dissolution as a source of CO2 over geological timescales. Nat 507: 346-349. doi: 10.1038/nature13030
![]() |
[30] | Torres MA, West AJ, Clark KE (2015) Geomorphic regime modulates hydrologic control of chemical weathering in the Andes–Amazon. Geochim Cosmochim Acta 166: 105–128. |
[31] |
Caves JK, Jost AB, Laub KV, et al. (2016) Cenozoic carbon cycle imbalances and a variable weathering feedback. Earth Planet Sci Lett 450: 152-163. doi: 10.1016/j.epsl.2016.06.035
![]() |
[32] | West AJ, Torres M, Moosdorf N, et al. (2016) Glacial weathering, sulfide oxidation, and the geologic evolution of CO2. Goldschmidt Conf Abstr 3402. |
[33] | Larson DW (1987) The absorption and release of water by lichens. In: Peveling E (ed.), Progress and Problems in Lichenology in the Eighties, Bibliothca lichenological 25, Berlin:J. Cramer, 351-360. |
[34] |
Li Z, Liu L, Chen J, Teng HH (2016) Cellular dissolution at `hypha-and spore-mineral interfaces revealing unrecognized mechanisms and scales of fungal weathering. Geol 44: 319-322. doi: 10.1130/G37561.1
![]() |
[35] | Jones D, Wilson MJ (1985) Chemical activity of lichens on mineral surfaces; a review. Int Biodeterior 21: 99-104. |
[36] |
Chen J, Blume H-P, Beyer L (2000) Weathering of rocks induced by lichen colonization–a review. Catena 39: 121-146. doi: 10.1016/S0341-8162(99)00085-5
![]() |
[37] |
Amundson R, Heimsath A, Owen, et al. (2015) Hillslope soils and vegetation. Geomorphol 234: 122-13. doi: 10.1016/j.geomorph.2014.12.031
![]() |
[38] |
Hahm WJ, Riebe CS, Lukens CE, et al. (2014) Bedrock composition regulates mountain ecosystems and landscape evolution. Proc Natl Acad Sci USA. 111: 3338-3343. doi: 10.1073/pnas.1315667111
![]() |
[39] | Schwartzman DW (2015) The case for a hot Archean climate and its implications to the history of the biosphere. Arxiv org April 1. |
[40] |
Sheldon ND (2006) Precambrian paleosols and atmospheric CO2 levels. Precambrian Res 147: 148-155. doi: 10.1016/j.precamres.2006.02.004
![]() |
[41] | Hren MT, Tice MM, Chamberlain CP (2009) Oxygen and hydrogen isotope evidence for a temperate climate 3.42 billion years ago. Nat 462: 205-208. |
[42] |
Blake RE, Chang SJ, Lepland A (2010) Phosphate oxygen isotopic evidence for a temperate and biologically active Archaean ocean. Nat 464: 1029-1032. doi: 10.1038/nature08952
![]() |
[43] | Rosing MT, Bird DK, Sleep NH, et al. (2010) No climate paradox under the faint early Sun. Nat 464: 744-747. |
[44] | Driese SG, Jirsa MA, Ren M, et al. (2011) Neoarchean paleoweathering of tonalite and metabasalt: Implications for reconstructions of 2.69 Ga early terrestrial ecosystems and paleoatmospheric chemistry. Precambrian Res 189: 1-17. |
[45] |
Sheldon, ND, Tabor, NJ (2009) Quantitative paleoenvironmental and paleoclimatic reconstruction using paleosols. Earth-Sci Rev 95: 1-52. doi: 10.1016/j.earscirev.2009.03.004
![]() |
[46] | de Wit MJ, Furnes H (2016) 3.5-Ga hydrothermal fields and diamictites in the Barberton Greenstone Belt-Paleoarchean crust in cold environments. Sci Adv 2: e1500368. |
[47] | Airapetian VS , Glocer A, Grono G, et al. (2016) Prebiotic chemistry and atmospheric warming of early Earth by an active young Sun. Nat Geosci 9: 452-455. |
[48] | Som, SM, Buick R, Hagadorn JW, et al. (2016) Earth's air pressure 2.7 billion years ago constrained to less than half of modern levels. Nat Geosci 9: 448-451. |
[49] | Som SM, Catling DC, Harnmeijer JP, et al. (2012) Air density 2.7 billion years ago limited to less than twice modern levels by fossil raindrop imprints. Nat 484: 359-362. |
[50] |
Kavanagh L, Goldblatt C (2015) Using raindrops to constrain past atmospheric density. Earth Planet Sci Lett 413: 51-58. doi: 10.1016/j.epsl.2014.12.032
![]() |
[51] |
Sahagian D, Proussevitch A (2007) Paleoelevation measurement on the basis of vesicular basalts. Rev Mineral Geochem 66: 195-213. doi: 10.2138/rmg.2007.66.8
![]() |
[52] |
Fournier, GP, Alm, E J (2015) Ancestral Reconstruction of a Pre-LUCA Aminoacyl-tRNA Synthetase Ancestor Supports the Late Addition of Trp to the Genetic Code. J Mol Evol 80: 171-185. doi: 10.1007/s00239-015-9672-1
![]() |
[53] |
Romero-Romero ML, Risso VA, Martinez-Rodriguez S, et al. (2016) Selection for Protein Kinetic Stability Connects Denaturation Temperatures to Organismal Temperatures and Provides Clues to Archaean Life. PLoS ONE, 11: e0156657. doi: 10.1371/journal.pone.0156657
![]() |
[54] | Tartèse, R, Chaussidon M, Gurenko A, et al. (2017) Warm Archean oceans reconstructed from oxygen isotope composition of early-life remnants. Geochem Perspect Lett 3: 55-65. |
[55] | Schwartzman DW, Volk T (1991) Biotic enhancement of weathering and surface temperatures on Earth since the origin of life. Glob Planet Change Sect 4: 357-371. |
[56] |
Kasting JF, Ackerman TP (1986) Climatic consequences of a very high CO2 level in Earth's early atmosphere. Sci 234: 1383-1385. doi: 10.1126/science.11539665
![]() |
[57] | Charnay B, Forget F, Wordsworth R, et al. (2013) Exploring the faint young Sun problem and the possible climates of the Archean Earth with a 3-D GCM. J Geophys Res: Atmos 118: 1-18. |
[58] |
Le Hir G, Teitler Y, Fluteau F, et al. (2014) The faint young Sun problem revisited with a 3-D climate–carbon model–Part 1. Clim Past 10: 697-713. doi: 10.5194/cp-10-697-2014
![]() |
[59] | Flament N, Coltice N, Rey PF (2013) The evolution of the 87Sr/86Sr of marine carbonates does not constrain continental growth. Precambrian Res 229: 177-188. |
[60] |
Jellinek AM, Jackson MG (2015) Connections between the bulk composition, geodynamics and habitability of Earth. Nat Geosci 8: 587-593. doi: 10.1038/ngeo2488
![]() |
[61] |
Dessert C, Dupre B, Gaillardet J, et al. (2003) Basalt weathering laws and the impact of basalt weathering on the global carbon cycle. Chem Geol 202: 257-273. doi: 10.1016/j.chemgeo.2002.10.001
![]() |
[62] |
Dupre B, Dessert C, Oliva P, et al. (2003) Rivers, chemical weathering and Earth's climate. C R Geosciences 335: 1141-1160. doi: 10.1016/j.crte.2003.09.015
![]() |
[63] |
Ibarra DE, Caves JK, Moon S, et al. (2016) Differential weathering of basaltic and granitic catchments from concentration discharge relationships. Geochim Cosmochim Acta 190: 265-293. doi: 10.1016/j.gca.2016.07.006
![]() |
[64] | Navarre-Sitchler A, Brantley S (2007) Basalt weathering across scales. Earth Planetary Sci Lett 261: 321-334. |
[65] | Eggleston CM, Hochella MF Jr, Parks GA (1989) Sample preparation and aging effects on dissolution rate and surface composition of diopside. Geochim Cosmochim Acta 53: 979-804. |
[66] | White AF, Brantley SL ( 2003) The effect of time on the weathering of silicate minerals: why do weathering rates differ in the laboratory and field? Chem Geol 202: 479-506. |
[67] | Grandstaff DE (1986) The dissolution rate of forsteritic olivine from Hawaiian beach sand. In: Coleman S, Dethier D (eds.) Rates of Chemical Weathering of Rocks and Minerals., New York: Academic Press, New York, 41-59. |
[68] |
Rimstidt JD, Brantley SL, Olsen AA (2012) Systematic review of forsterite dissolution rate data. Geochim Cosmochim Acta 99: 159-178. doi: 10.1016/j.gca.2012.09.019
![]() |
[69] | Wolff-Boenisch D, Gislason SR, Oelkers EH, et al. (2004) The dissolution rates of natural glasses as a function of their composition at pH 4 and 10.6, and temperatures from 25 to 74°C. Geochim Cosmochim Acta 68: 4843-4858. |
[70] | Shikazono N, Takino A, Ohtani, H (2005) An estimate of dissolution rate constant of volcanic glass in volcanic ash soil from the Mt. Fuji area, central Japan. Geochem J 39: 185-196. |
[71] |
Gao G-L, Ding G-E, Wu B, et al. (2014) Fractal Scaling of Particle Size Distribution and Relationships with Topsoil Properties Affected by Biological Soil Crusts. PLoS ONE 9: e88559. doi: 10.1371/journal.pone.0088559
![]() |
[72] | Schwartzman DW, Volk T (2015) Habitability for Complex Life and the Development and Self-Limitations of the Biotic Enhancement of Weathering, Puzzle of Earths Uninterrupted Habitability, Abstract Book, 12, London: Geological Society. |
[73] |
Rad S, Rivé K, Vittecoqa B, et al. (2013) Chemical weathering and erosion rates in the Lesser Antilles: An overview in Guadeloupe, Martinique and Dominica. J South Am Earth Sci 45: 331-344. doi: 10.1016/j.jsames.2013.03.004
![]() |
[74] | Hartmann J, Moosdorf N, Lauerwald R, et al. (2014) Global chemical weathering and associated P-release-The role of lithology, temperature and soil properties. Chem Geol 363: 145-163. |
[75] |
Rivé K, Gaillardet J, Agrinier P, et al. (2013) Carbon isotopes in the rivers from the Lesser Antilles: origin of the carbonic acid consumed by weathering reactions in the Lesser Antilles. Earth Surf Processes Landforms 38: 1020-1035. doi: 10.1002/esp.3385
![]() |
[76] | Maher K, Chamberlain CP (2014) Hydrologic regulation of chemical weathering and the geologic carbon cycle. Sci 343: 1502-1504. |
[77] |
Murphy BP, Johnson JPL, Gasparini NM, et al. (2016) Chemical weathering as a mechanism for the climatic control of bedrock river incision. Nat 532: 223-237. doi: 10.1038/nature17449
![]() |
[78] |
Anders AM (2016) How rain affects rock and rivers. Nat 532: 186-187. doi: 10.1038/532186a
![]() |
[79] | Van Cappellen P (2013) Where groundwater meets surface water. Mineral Mag 77: 2388. |
[80] |
Clair JSt, Moon S, Holbrook WS, et al. (2015) Geophysical imaging reveals topographic stress control of bedrock weathering. Sci 350: 534-538. doi: 10.1126/science.aab2210
![]() |
[81] | Anderson RS (2015) Pinched topography initiates the critical zone. Sci 350: 506-507. |
[82] | Bazilevskaya E, Lebedeva M, Pavich M, et al. (2013) Where fast weathering creates thin regolith and slow weathering creates thick regolith. Earth Surf Process Landform 38: 847-858. |
[83] |
West AJ (2012) Thickness of the chemical weathering zone and implications for erosional and climatic drivers of weathering and for carbon-cycle feedbacks. Geol 40: 811-814. doi: 10.1130/G33041.1
![]() |
[84] |
Herman F, Seward D, Valla PG, et al. (2013) Worldwide acceleration of mountain erosion under a cooling climate. Nat 504: 423-426. doi: 10.1038/nature12877
![]() |
[85] |
Galy V, Peucker-Ehrenbrink B, Eglinton T (2015) Global carbon export from the terrestrial biosphere controlled by erosion. Nat 521: 204-207. doi: 10.1038/nature14400
![]() |
[86] |
Li G, Hartmann J, Derry LA, et al. (2016) Temperature dependence of basalt weathering. Earth Planet Sci Lett 443: 59-69. doi: 10.1016/j.epsl.2016.03.015
![]() |
[87] | Torres MA, West AJ, Clark KE, et al. (2016) The acid and alkalinity budgets of weathering in the Andes–Amazon system: Insights into the erosional control of global biogeochemical cycles. Earth Planet Sci Lett 450: 381–391. |
[88] | Pelak NF, Parolari AJ , Porporato A (2016) Bistable plant–soil dynamics and biogenic controls on the soil production function. Earth Surf Process Landforms 41: 1011-1017. |
[89] |
Buendía C, Arens S, Hickler T, et al. (2014) On the potential vegetation feedbacks that enhance phosphorus availability - insights from a process-based model linking geological and ecological timescales. Biogeosciences 11: 3661-3683. doi: 10.5194/bg-11-3661-2014
![]() |
[90] |
Porada P, Weber B, Elbert W, et al. A (2014) Estimating impacts of lichens and bryophytes on global biogeochemical cycles. Global Biogeochem. Cycles 28: 71-85. doi: 10.1002/2013GB004705
![]() |
[91] | Kelemen PB, Matter J (2008) In situ carbonation of peridotite for CO2 storage. Proc Natl Acad Sci USA105: 17295-17300. |
[92] |
Schuiling RD, de Boer PL (2011) Rolling stones, fast weathering of olivine in shallow seas for cost-effective CO₂ capture and mitigation of global warming and ocean acidification. Earth Syst Dynam Discuss 2: 551-568. doi: 10.5194/esdd-2-551-2011
![]() |
[93] |
Ten Berge HFM., van der Meer HG, Steenhuizen JW, et al. (2012) Olivine weathering in soil, and its effects on growth and nutrient uptake in ryegrass (Lolium perenne L.). A Pot Experiment. PLOS One 7: e42098. doi: 10.1371/journal.pone.0042098
![]() |
[94] |
Hartmann J, West AJ, Renforth P, et al. (2013) Enhanced chemical weathering as a geoengineering strategy to reduce atmospheric carbon dioxide, supply nutrients, and mitigate ocean acidification. Rev Geophys 51: 113-149. doi: 10.1002/rog.20004
![]() |
[95] | Matter JM, Stute M, Snaebjörnsdottir SO, et al. (2016) Rapid carbon mineralization for permanent disposal of anthropogenic carbon dioxide emissions. Sci 352: 1312-1314. |
[96] | Taylor LL, Quirk J, Thorley RMS, et al. (2016) Enhanced weathering strategies for stabilizing climate and averting ocean acidification. Nat Clim Change 6: 402–406. |
[97] | Schuiling RD (2017) Olivine weathering against climate change. Natural Science 9: 21-26. |
[98] | Juarez S, Dontsova K, Le Galliard J-F , et al. (2016) Effect of elevated CO2 and temperature on abiotic and biologically-driven basalt weathering and C sequestration. Geophys Res Abstr 18. |
[99] | Navarre-Sitchler AK, Cole DR, Rother G, et al. (2013) Porosity and surface area evolution during weathering of two igneous rocks. Geochim Cosmochim Acta 109: 400-413. |
[100] | Navarre-Sitchler A, Brantley S, Rother G (2015) How Porosity Increases During Incipient Weathering of Crystalline Silicate Rocks. Rev Mineral Geochem 80: 331-354. |
[101] | Aghamiri R, Schwartzman DW (2002) Weathering rates of bedrock by lichens: a mini watershed study. Chem Geol 188: 249-259. |
[102] | Zambell CB, Adams JM, Goring ML, et al. (2012) Effect of lichen colonization on chemical weathering of hornblende granite as estimated by aqueous elemental flux. Chem Geol 291: 166-174. |
[103] |
Jackson TA, Keller WD (1970) A comparative study of the role of lichens and inorganic processes in the chemical weathering of recent Hawaiian lava flows. Am J Sci 269: 446-466. doi: 10.2475/ajs.269.5.446
![]() |
[104] | Brady PV, Dorn RI, Brazel AJ, et al. (1999) Direct measurement of the combined effects of lichen, rainfall, and temperature on silicate weathering. Geochim Cosmochim Acta 63: 3293-3300. |
[105] |
Stretch RC, Viles HA (2002) The nature and rate of weathering by lichens on lava flows on Lanzarote. Geomorphol 47: 87-94. doi: 10.1016/S0169-555X(02)00143-5
![]() |
[106] |
Lenton TM, Crouch M, Johnson M, et al. (2012) First plants cooled the Ordovician. Nat Geosci 5: 86-89. doi: 10.1038/ngeo1390
![]() |
[107] | Moulton KL, Berner RA (1998) Quantification of the effect of plants on weathering: studies in Iceland. Geol 26: 895-898. |
[108] |
Bormann BT, Wang D, Bormann FH, et al. (1998) Rapid, plant-induced weathering in an aggrading experimental ecosystem. Biogeochemistry 43: 129-155. doi: 10.1023/A:1006065620344
![]() |
[109] |
Berner RA, Kothaval Z (2001) Geocarb III: A revised model of atmospheric CO2 over Phanerozoic time. Am J Sci 301: 182-204. doi: 10.2475/ajs.301.2.182
![]() |
[110] |
Berner RA (2006) GEOCARBSULF: A combined model for Phanerozoic atmospheric O2 and CO2. Geochim Cosmochim Acta 70: 5653-5664. doi: 10.1016/j.gca.2005.11.032
![]() |
[111] | Berner RA, Kothaval Z (2001) Geocarb III: A revised model of atmospheric CO2 over Phanerozoic time. Am J Sci 301: 182-204. |
[112] | Royer DL, Berner RA, Montanez IP, et al. (2004) CO2 as a primary driver of Phanerozoic climate. GSA Today 14: 4-10. |
[113] |
Taylor LL, Banwart SA, Valdes PJ, et al. (2012) Evaluating the effects of terrestrial ecosystems, climate and carbon dioxide on weathering over geological time: a global-scale process-based approach. Phil. Trans. R. Soc. B, 367: 565-582. doi: 10.1098/rstb.2011.0251
![]() |
[114] |
Arens S, Kleidon A (2011) Eco-hydrological versus supply-limited weathering regimes and the potential for biotic enhancement of weathering at the global scale. Appl Geochem 26: S274-S278. doi: 10.1016/j.apgeochem.2011.03.079
![]() |
[115] | Von Bloh W, Bounama C, Eisenack K, et al. (2008) Estimating the biogenic enhancement factor of weathering using an inverse viability method. Ecol Modell 216: 245-251. |
[116] | Schwartzman DW, Brantley S (2013) The Geobiology of Weathering: the 13th Hypothesis. Mineral Mag 77: 2170. |
[117] | Beraldi-Campesi H (2013) Early life on land and the first terrestrial ecosystems. Available from: http://www.ecologicalprocesses.com/content/2/1/1. |
[118] |
Rastogi RP, Sinha RP, Moh SH, et al. (2014) Ultraviolet radiation and cyanobacteria. J Photochem Photobiol B 141: 154-169. doi: 10.1016/j.jphotobiol.2014.09.020
![]() |
[119] |
Lalonde SV, Konhauser KO (2015) Benthic perspective on Earth's oldest evidence for oxygenic photosynthesis. Proc Natl Acad Sci USA 112: 995-1000. doi: 10.1073/pnas.1415718112
![]() |
[120] | Gauger T, Konhauser K, Kappler A (2015) Protection of phototrophic iron(II)-oxidizing bacteria from UV irradiation by biogenic iron(III) minerals: Implications for early Archean banded iron formation. Geol 43 (12): 1067-1070. |
[121] | Lenton TM, Daines SJ (2016) Matworld-the biogeochemical effects of early life on land. New Phytol doi: 10.1111/nph.14338. |
[122] | Garcia AK, Schopf JW, Yokobori S-i, et al. (2017) Reconstructed ancestral enzymes suggest long-term cooling of Earth's photic zone since the Archean. Proc Natl Acad Sci USA 114: 4619-4624. |
[123] | Sleep NH, Zahnle K, Carbon dioxide cycling and implications for climate on ancient Earth. J Geophys Res 106: 1373-1399. |
[124] | Kanzaki Y, Murakami T (2015) Estimates of atmospheric CO2 in the Neoarchean–Paleoproterozoic from paleosols. Geochim Cosmochim Acta 159: 190-219. |
1. | Song Qian, Liwei Yang, Yan Xue, Ping Li, Xiaoyong Sun, SIFusion: Lightweight infrared and visible image fusion based on semantic injection, 2024, 19, 1932-6203, e0307236, 10.1371/journal.pone.0307236 | |
2. | 张鸿德 ZHANG Hongde, 冯鑫 FENG Xin, 杨杰铭 YANG Jieming, 邱国航 QIU Guohang, 基于双分支边缘卷积融合网络的红外与可见光图像融合方法, 2024, 53, 1004-4213, 0810004, 10.3788/gzxb20245308.0810004 | |
3. | Yongyu Luo, Zhongqiang Luo, Infrared and visible image fusion algorithm based on gradient attention residuals dense block, 2024, 10, 2376-5992, e2569, 10.7717/peerj-cs.2569 | |
4. | Zhen Pei, Jinbo Lu, Yongliang Qian, Lihua Fan, Hongyan Wang, Jinling Chen, A new method for fusing infrared and visible images in low-light environments based on visual perception and attention mechanism, 2025, 186, 01438166, 108800, 10.1016/j.optlaseng.2024.108800 | |
5. | Song Qian, Guzailinuer Yiming, Ping Li, Junfei Yang, Yan Xue, Shuping Zhang, LLE-Fuse: Lightweight Infrared and Visible Light Image Fusion Based on Low-Light Image Enhancement, 2025, 82, 1546-2226, 4069, 10.32604/cmc.2025.059931 | |
6. | 赵佳 Zhao Jia, 辛月兰 Xin Yuelan, 刘冀钊 Liu Jizhao, 王庆庆 Wang Qingqing, 脉冲耦合双对抗学习网络的红外与可见光图像融合, 2025, 62, 1006-4125, 1037008, 10.3788/LOP242143 |
Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | |
DenseFuse | 7.0692 | 2.5409 | 0.6752 | 1.3296 | 5.8397 | 0.3552 |
FusionGAN | 5.4694 | 1.9155 | 0.4253 | 0.8015 | 5.2260 | 0.1208 |
IFCNN | 7.5947 | 2.7399 | 0.8283 | 1.6658 | 6.3109 | 0.5540 |
SDNet | 5.3258 | 1.7398 | 0.3758 | 0.8364 | 4.8891 | 0.2944 |
U2Fusion | 5.6231 | 1.8953 | 0.3967 | 1.0034 | 4.7525 | 0.2908 |
FLFuse | 6.4790 | 2.0697 | 0.4860 | 1.1189 | 5.5157 | 0.3198 |
PIAFusion | 7.9268_ | 4.1774 | 0.9072_ | 1.7395_ | 6.4304_ | 0.6324 |
Ours | 8.1413 | 3.6805_ | 0.9282 | 1.8153 | 6.5104 | 0.5619_ |
Dataset | Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
TNO | DenseFuse | 8.5765 | 2.1987 | 0.6704 | 1.5916 | 6.3422 | 0.3427 |
FusionGAN | 8.6703 | 2.3353 | 0.6541 | 1.3788 | 6.5578 | 0.2339 | |
IFCNN | 9.0058 | 2.4154 | 0.7996 | 1.6850 | 6.7413 | 0.5066 | |
SDNet | 9.0679 | 2.2606 | 0.7592 | 1.5587 | 6.6947 | 0.4290 | |
U2Fusion | 8.8553 | 1.8730 | 0.6787 | 1.5862 | 6.4230 | 0.4245 | |
FLFuse | 9.2628 | 2.1925 | 0.8084 | 1.7308 | 6.3658 | 0.4177 | |
PIAFusion | 9.1093 | 3.2464 | 0.8835 | 1.6540 | 6.8937 | 0.5556 | |
Ours | 9.2721 | 3.7136 | 0.9496 | 1.7312 | 6.9856 | 0.5311 | |
M3FD | DenseFuse | 8.6130 | 2.8911 | 0.6694 | 1.5051 | 6.4264 | 0.3709 |
FusionGAN | 8.8571 | 2.9921 | 0.5176 | 1.1292 | 6.4750 | 0.2530 | |
IFCNN | 9.2815 | 2.9560 | 0.7738 | 1.5353 | 6.6966 | 0.6053 | |
SDNet | 8.8855 | 3.1798 | 0.6329 | 1.3914 | 6.6102 | 0.5005 | |
U2Fusion | 9.0141 | 2.7531 | 0.7061 | 1.5488 | 6.6285 | 0.5303 | |
FLFuse | 8.7580 | 3.2425 | 0.6986 | 1.4975 | 6.5744 | 0.2640 | |
PIAFusion | 10.1639 | 4.6942 | 0.9300 | 1.3363 | 6.8036 | 0.6348 | |
Ours | 9.9899 | 4.3123 | 0.9350 | 1.5502 | 6.7685 | 0.6440 |
RECB | AFM | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
✓ | ✓ | 8.1413 | 3.6805 | 0.9282 | 1.8153 | 6.5104 | 0.5619 |
✓ | ✘ | 7.4551 | 3.0648 | 0.7234 | 1.6117 | 6.0100 | 0.4488 |
✘ | ✓ | 7.7954 | 3.0922 | 0.8355 | 1.8060 | 6.2925 | 0.5098 |
✘ | ✘ | 6.6285 | 2.6117 | 0.5500 | 1.2793 | 5.6158 | 0.4362 |
Algorithm | DenseFuse | FusionGAN | IFCNN | SDNet | U2Fusion | FLFuse | PIAFusion | Ours |
Running time | 0.374 | 0.082 | 0.019 | 0.014 | 0.155 | 0.001 | 0.081 | 0.002_ |
Structural re-parameterisation | Forward time | Forward pass size | Params | Params size | Cumulative pixel deviation of results |
W/O | 0.0299 s | 5451.57 MB | 146,389 | 0.56 MB | / |
W | 0.0020 s | 2601.57 MB | 145,477 | 0.55 MB | 1 × 10-4 |
Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | |
DenseFuse | 7.0692 | 2.5409 | 0.6752 | 1.3296 | 5.8397 | 0.3552 |
FusionGAN | 5.4694 | 1.9155 | 0.4253 | 0.8015 | 5.2260 | 0.1208 |
IFCNN | 7.5947 | 2.7399 | 0.8283 | 1.6658 | 6.3109 | 0.5540 |
SDNet | 5.3258 | 1.7398 | 0.3758 | 0.8364 | 4.8891 | 0.2944 |
U2Fusion | 5.6231 | 1.8953 | 0.3967 | 1.0034 | 4.7525 | 0.2908 |
FLFuse | 6.4790 | 2.0697 | 0.4860 | 1.1189 | 5.5157 | 0.3198 |
PIAFusion | 7.9268_ | 4.1774 | 0.9072_ | 1.7395_ | 6.4304_ | 0.6324 |
Ours | 8.1413 | 3.6805_ | 0.9282 | 1.8153 | 6.5104 | 0.5619_ |
Dataset | Algorithm | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
TNO | DenseFuse | 8.5765 | 2.1987 | 0.6704 | 1.5916 | 6.3422 | 0.3427 |
FusionGAN | 8.6703 | 2.3353 | 0.6541 | 1.3788 | 6.5578 | 0.2339 | |
IFCNN | 9.0058 | 2.4154 | 0.7996 | 1.6850 | 6.7413 | 0.5066 | |
SDNet | 9.0679 | 2.2606 | 0.7592 | 1.5587 | 6.6947 | 0.4290 | |
U2Fusion | 8.8553 | 1.8730 | 0.6787 | 1.5862 | 6.4230 | 0.4245 | |
FLFuse | 9.2628 | 2.1925 | 0.8084 | 1.7308 | 6.3658 | 0.4177 | |
PIAFusion | 9.1093 | 3.2464 | 0.8835 | 1.6540 | 6.8937 | 0.5556 | |
Ours | 9.2721 | 3.7136 | 0.9496 | 1.7312 | 6.9856 | 0.5311 | |
M3FD | DenseFuse | 8.6130 | 2.8911 | 0.6694 | 1.5051 | 6.4264 | 0.3709 |
FusionGAN | 8.8571 | 2.9921 | 0.5176 | 1.1292 | 6.4750 | 0.2530 | |
IFCNN | 9.2815 | 2.9560 | 0.7738 | 1.5353 | 6.6966 | 0.6053 | |
SDNet | 8.8855 | 3.1798 | 0.6329 | 1.3914 | 6.6102 | 0.5005 | |
U2Fusion | 9.0141 | 2.7531 | 0.7061 | 1.5488 | 6.6285 | 0.5303 | |
FLFuse | 8.7580 | 3.2425 | 0.6986 | 1.4975 | 6.5744 | 0.2640 | |
PIAFusion | 10.1639 | 4.6942 | 0.9300 | 1.3363 | 6.8036 | 0.6348 | |
Ours | 9.9899 | 4.3123 | 0.9350 | 1.5502 | 6.7685 | 0.6440 |
RECB | AFM | Evaluation method | |||||
SD | MI | VIF | SCD | EN | Qabf | ||
✓ | ✓ | 8.1413 | 3.6805 | 0.9282 | 1.8153 | 6.5104 | 0.5619 |
✓ | ✘ | 7.4551 | 3.0648 | 0.7234 | 1.6117 | 6.0100 | 0.4488 |
✘ | ✓ | 7.7954 | 3.0922 | 0.8355 | 1.8060 | 6.2925 | 0.5098 |
✘ | ✘ | 6.6285 | 2.6117 | 0.5500 | 1.2793 | 5.6158 | 0.4362 |
Algorithm | DenseFuse | FusionGAN | IFCNN | SDNet | U2Fusion | FLFuse | PIAFusion | Ours |
Running time | 0.374 | 0.082 | 0.019 | 0.014 | 0.155 | 0.001 | 0.081 | 0.002_ |
Structural re-parameterisation | Forward time | Forward pass size | Params | Params size | Cumulative pixel deviation of results |
W/O | 0.0299 s | 5451.57 MB | 146,389 | 0.56 MB | / |
W | 0.0020 s | 2601.57 MB | 145,477 | 0.55 MB | 1 × 10-4 |