
Citation: Lin Kooi Ong, Frederick Rohan Walker, Michael Nilsson. Is Stroke a Neurodegenerative Condition? A Critical Review of Secondary Neurodegeneration and Amyloid-beta Accumulation after Stroke[J]. AIMS Medical Science, 2017, 4(1): 1-16. doi: 10.3934/medsci.2017.1.1
[1] | Jinyun Jiang, Jianchen Cai, Qile Zhang, Kun Lan, Xiaoliang Jiang, Jun Wu . Group theoretic particle swarm optimization for gray-level medical image enhancement. Mathematical Biosciences and Engineering, 2023, 20(6): 10479-10494. doi: 10.3934/mbe.2023462 |
[2] | Kunpeng Li, Zepeng Wang, Yu Zhou, Sihai Li . Lung adenocarcinoma identification based on hybrid feature selections and attentional convolutional neural networks. Mathematical Biosciences and Engineering, 2024, 21(2): 2991-3015. doi: 10.3934/mbe.2024133 |
[3] | Jun Gao, Qian Jiang, Bo Zhou, Daozheng Chen . Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Mathematical Biosciences and Engineering, 2019, 16(6): 6536-6561. doi: 10.3934/mbe.2019326 |
[4] | Yufeng Li, Chengcheng Liu, Weiping Zhao, Yufeng Huang . Multi-spectral remote sensing images feature coverage classification based on improved convolutional neural network. Mathematical Biosciences and Engineering, 2020, 17(5): 4443-4456. doi: 10.3934/mbe.2020245 |
[5] | Hong Yu, Wenhuan Lu, Qilong Sun, Haiqiang Shi, Jianguo Wei, Zhe Wang, Xiaoman Wang, Naixue Xiong . Design and analysis of a robust breast cancer diagnostic system based on multimode MR images. Mathematical Biosciences and Engineering, 2021, 18(4): 3578-3597. doi: 10.3934/mbe.2021180 |
[6] | Binju Saju, Neethu Tressa, Rajesh Kumar Dhanaraj, Sumegh Tharewal, Jincy Chundamannil Mathew, Danilo Pelusi . Effective multi-class lungdisease classification using the hybridfeature engineering mechanism. Mathematical Biosciences and Engineering, 2023, 20(11): 20245-20273. doi: 10.3934/mbe.2023896 |
[7] | Sushovan Chaudhury, Kartik Sau, Muhammad Attique Khan, Mohammad Shabaz . Deep transfer learning for IDC breast cancer detection using fast AI technique and Sqeezenet architecture. Mathematical Biosciences and Engineering, 2023, 20(6): 10404-10427. doi: 10.3934/mbe.2023457 |
[8] | S. Dinesh Krishnan, Danilo Pelusi, A. Daniel, V. Suresh, Balamurugan Balusamy . Improved graph neural network-based green anaconda optimization for segmenting and classifying the lung cancer. Mathematical Biosciences and Engineering, 2023, 20(9): 17138-17157. doi: 10.3934/mbe.2023764 |
[9] | Yingjian Yang, Wei Li, Yingwei Guo, Nanrong Zeng, Shicong Wang, Ziran Chen, Yang Liu, Huai Chen, Wenxin Duan, Xian Li, Wei Zhao, Rongchang Chen, Yan Kang . Lung radiomics features for characterizing and classifying COPD stage based on feature combination strategy and multi-layer perceptron classifier. Mathematical Biosciences and Engineering, 2022, 19(8): 7826-7855. doi: 10.3934/mbe.2022366 |
[10] | Haiyan Song, Cuihong Liu, Shengnan Li, Peixiao Zhang . TS-GCN: A novel tumor segmentation method integrating transformer and GCN. Mathematical Biosciences and Engineering, 2023, 20(10): 18173-18190. doi: 10.3934/mbe.2023807 |
In recent years, remote sensing (RS) imagery has been widely applied in various fields, with the continuous development of RS technology. The fields include environmental monitoring [1], natural disaster monitoring [2,3,4], mineral resource development [5] and cartography [6]. However, a large part of Earth's surface is covered by clouds [7], which results in a significant amount of ground information being inaccessible to satellites. Additionally, the texture and spectral information of images are inconstant. There are substantial challenges in subsequent object detection and analysis tasks. Therefore, accurate and effective cloud detection is crucial due to being a preprocessing step for getting over the challenges before the analysis of satellite images [8].
Currently, cloud detection methods can be categorized into multi-spectral threshold methods and machine learning-based methods. Multi-spectral threshold methods use spectral differences between clouds and the underlying surface in different wavelength bands to select thresholds for cloud detection. Rossow et al. used spectral thresholding based on the 11 μm brightness temperature data of cloud tops in the international satellite cloud climatology project [9]. Xiong et al. proposed a cloud detection approach based on hybrid multi-spectral features with dynamic thresholds [10]. The multi-spectral threshold methods have a simple structure and are easy to implement. However, it is challenging for them to distinguish between clouds and the underlying surface for them when the ground is covered with ice, snow, desert or a cloud cover that is thin.
Many cloud detection algorithms based on machine learning have emerged with the rapid development of computer science, including the support vector machine (SVM) [11,12,13], random forest (RF) [14,15], convolutional neural network (CNN) [16,17] and others. Li et al. [11] utilized brightness temperature and texture features, employing SVM as a classifier for cloud detection. Zhang et al. proposed a cloud detection method based on the multi-feature embedded learning SVM [12]. Fu et al. used the results of the ensemble threshold as training samples for the RF classifier [15]. Zhang et al. designed a cascaded feature attention module based on the encoder-decoder structure to obtain and enhance useful feature information [16]. Xie et al. proposed a CNN with two branches to detect different types of clouds [17]. However, all the machine learning methods used by them encounter the following challenges:
(1) Cloud edges are often semi-transparent, making it difficult for them to separate the underlying surface based solely on visual features. Therefore, it is essential to make full use of the multi-channel information from RS satellites.
(2) These machine learning methods require a sufficient amount of training data to achieve reliable performance [18]. However, fragmented cloud samples are often sparse and scattered, demanding a significant amount of effort and time for manual labeling.
The GMM, as an unsupervised learning method, does not require manual sample labeling. Therefore, it requires less preparation work before detection [19]. Furthermore, visual features can have difficulty in distinguishing between cloud edges and the underlying surface during manual labeling, and incorrect labeling may result in model inaccuracies. In contrast, the GMM algorithm use satellite multi-channel reflectance and brightness temperature as clustering features and can effectively leverage the spectral differences between clouds and the underlying surface for classification.
In summary, we propose a cloud detection algorithm based on a feature enhancement module and GMM clustering. The proposed method utilizes clustering-based feature analysis to select appropriate feature schemes for different time periods, enabling cloud detection throughout the day. Additionally, we employ the Laplacian operator to enhance the spectral characteristics of the cloud edges and broken cloud regions and improve the accuracy of the GMM classifier for cloud detection.
The paper is structured as follows: Section 2 presents the various applications of cloud detection. Section 3 discusses data preprocessing, with a focus on the selection of clustering features. Section 4 presents a modular description of the proposed algorithm, highlighting its key components and functionalities. Section 5 provides a comprehensive account of all experiments conducted and their corresponding observations, shedding light on the algorithm's performance and efficacy. Section 6 discusses the ethical considerations and potential societal impacts of enhanced cloud detection. Finally, Section 7 offers a summary of the paper, summarizing the major contributions.
(1) Wildfire detection
When utilizing satellite data for wildfire monitoring, clouds' mid-infrared reflectance might cause them to be mistakenly identified as fire. Hence, cloud detection is crucial. Zhang et al. set different thresholds during daytime and nighttime to detect clouds [20]. Kang et al. employed an algorithm utilizing Imager's reflectance and BT to eliminate clouds in aerosol optical depth retrieval [21].
(2) Cartography in agriculture
To reduce spectral redundancy in images and improve method efficiency, cloud detection is often a crucial step in data preprocessing during agricultural mapping. Xia et al. utilized cloud masks generated by the C Function of Mask algorithm [22]. Li et al. calculated cloud scores according to the spectral data from six bands to achieve cloud removal [23].
(3) Vegetation change analysis
Vegetation change analysis is particularly sensitive to cloud contamination because unmarked clouds above the vegetation can potentially be mistaken for change. Even minor errors in cloud detection can lead to significant inaccuracies in downstream vegetation change analysis. Huang et al. used clear view forest pixels as references to define cloud boundaries, separating clouds from the clear view surface in the spectral-temperature space [24]. Long et al. employed cloud identification algorithms provided by Google Earth Engine, utilizing the QA band to identify clouds and shadows in the images [25].
The cloud detection algorithm in this paper is based on multi-spectral data from the Himawari-8 satellite. As shown in Figure 1, the first step involves conducting a statistical analysis using probability density functions (PDFs) of pixel spectral data from clouds and underlying surfaces to select feature schemes for different time periods. The second step involves cloud detection based on the different feature schemes selected. During nighttime, the raw data of BT15 and BTD7−14 are directly subjected to GMM clustering to obtain the detection results. During the daytime, the detection process is divided into three parts. First, a feature enhancement module is applied to R3+4, and the enhanced data is clustered to detect cloud edges and broken clouds. Furthermore, the raw data is clustered to detect thick clouds. Finally, the results from the dual branches of clustering are combined to obtain the detection results.
The definition of clouds is crucial, and many scholars currently use a subjective approach to define different types of clouds. In this paper, cloud transparency and distribution are used as criteria for subjective definition.
(1) Thick clouds are opaque and typically appear over large areas.
(2) Cloud edges refer to clouds with transparency at the cloud boundaries. Cloud edges do not abruptly separate but gradually transition between the cloud layer and the non-cloud layer.
(3) Broken clouds exhibit varying degrees of transparency, with their most noticeable characteristic being their distribution in small patches across the sky.
Figure 2 illustrates various cloud types, with some cloud edges highlighted using red boxes.
The synchronous meteorological satellite Himawari-8 was officially launched in July 2015. The satellite is equipped with the advanced Himawari imager (AHI), which has 16 observation channels covering visible, near-infrared and thermal infrared ranges [26]. The central wavelengths of these channels for detecting various objects are shown in Table 1. Additionally, AHI can provide high-resolution observation data for the entire disk at intervals of 10 minutes, and its high temporal resolution data is advantageous for target detection and analysis [27]. The open accessibility of Himawari-8/AHI data, provided by the Japan Aerospace Exploration Agency, empowers researchers to extensively leverage this data for scientific inquiry and innovation. This ease of access and the widespread availability of this data have propelled its applications across the fields of Earth sciences, meteorology and environmental studies. For instance, it plays a pivotal role in monitoring natural disasters such as storm tracking, wildfire detection and climate change observations. Moreover, these datasets offer rich information for fields like agriculture, water resource management and urban planning.
Channel | Central wavelength/μm | Detection targets |
1 | 0.46 | aerosol, coastline |
2 | 0.51 | phytoplankton |
3 | 0.64 | vegetation, aerosol |
4 | 0.86 | stratus cloud |
5 | 1.6 | cloud top phase, snow |
6 | 2.3 | surface, cloud, snow |
7 | 3.9 | surface, wildfire |
8 | 6.2 | upper atmospheric water vapor, rainfall |
9 | 7.0 | mid-level atmospheric water vapor, rainfall |
10 | 7.3 | sulfur dioxide |
11 | 8.6 | water, sulfur dioxide, rainfall |
12 | 9.6 | ozone, airflow, wind |
13 | 10.4 | surface, cloud |
14 | 11.2 | surface, cloud |
15 | 12.4 | water, sea surface temperature |
16 | 13.3 | temperature, cloud altitude, cloudiness |
As shown in Figure 3, the data retrieved from the Himawari-8 satellite comes in the network common data form (NetCDF) format, requiring preprocessing before utilization. Initially, the NetCDF format will undergo conversion into physical variables. Channels 1 to 6 will denote reflectance data, while channels 7 to 16 will contain brightness temperature data. Following this, a sequence of geometric corrections will be applied, opting for the Geographic WGS84 Projection, succeeded by cropping procedures.
Moreover, different features may have varying measurement units or numerical ranges. Normalization can alleviate these effects, enabling the model to treat each feature more evenly. The normalization formula is as follows:
Y=X−XmeanXstd | (1) |
Where the mean and standard deviation of X are indicated by Xmean and Xstd, respectively. X and Y represent the initial and the normalized feature values, respectively.
The spectral data from the Himawari-8 satellite is analyzed using PDF to generate PDF distributions for reflectance, brightness temperature and brightness temperature difference of pixels corresponding to both clouds and the underlying surface. The selection of cluster features is completed based on the analysis of these PDF distributions. As shown in Figure 4, a random sample of cloud and underlying surface pixels from a series of RS images is used for statistical analysis.
Figure 5 shows the PDF distributions of different cluster features for cloud and underlying surface pixels. The PDF distributions for three cluster features, R3+4, BT15 and BTD7−14, exhibit some overlap for both cloud and underlying surface pixels. As shown in Figure 5(a), the PDF distribution of R3+4 for both cloud and underlying surface pixels roughly follows a normal distribution. Cloud PDF mainly ranges from 0.8 to 1.3, with a peak around 1.2, while the underlying surface PDF predominantly ranges from 0.2 to 0.35, with a peak around 0.3.
Figure 5(b) displays the PDF distribution of BT15 for clouds and underlying surface pixels. Cloud BT15 PDF is mainly distributed between 220 K and 285 K, displaying two peaks near 250 K and 270 K, respectively. In contrast, the underlying surface BT15 PDF mainly ranges from 280 K to 300 K, with a peak of around 290 K.
Figure 5(c) demonstrates that the PDF distributions of BTD7−14 for both cloud and underlying surface pixels also roughly follow a normal distribution. Cloud BTD7−14 PDF mainly ranges from 20 K to 30 K, with a peak around 23 K, while the underlying surface BTD7−14 PDF mainly ranges from 0 K to 7 K, with a peak around 2 K.
Although there are significant differences in the R3+4 PDF distributions between clouds and the underlying surface, during nighttime periods, the visible light band of AHI struggles to capture clear visible light images. Therefore, in this work, we use different cluster feature schemes for cloud detection during day and night periods: R3+4 during the day and BT15 and BTD7−14 at night.
During the process of cloud detection, cloud edges (the overlapping parts of PDFs for various cluster features in Figure 5) are prone to being misclassified as non-cloud pixels. This is primarily due to the minimal spectral feature differences between cloud edges and the underlying surface. Additionally, broken clouds are often sparse and scattered, causing their detection to be challenging. Therefore, as shown in Figure 6, to enhance the algorithm's efficiency in detecting cloud edges and broken clouds, convolution operations are employed to enhance the spectral features of cloud edges and broken clouds. The convolution operation involves sliding the sharpening convolution kernel over each pixel of the original image and computing the weighted sum of the convolution kernel with the pixel values. The formula for this enhancement is as follows:
g(x,y)=f(x,y)∗w(i,j) | (2) |
In the equation, f(x, y) represents the pixel value at (x, y), and w(i, j) represents the convolution kernel.
The Laplacian operator, often used in image processing, plays a crucial role in extracting gradient information from images [28,29]. It accentuates abrupt changes in pixel values, making it particularly useful in highlighting boundaries and fine details, such as edges of clouds or broken clouds. It is essentially a filter that amplifies high-frequency components, aiding in the detection and emphasis of subtle variations in the image. Compare the sharpening effects of the Laplacian operator with commonly used operators. As depicted in Figure 7, the sharpening results obtained using Roberts, Prewitt and Sobel operators exhibit missing cloud edges, whereas the sharpening results at the edges using the Laplacian operator demonstrate continuity. Therefore, the Laplacian operator is chosen to enhance the features of cloud edges and broken clouds. The definition of the Laplacian operator is as follows:
∇2f=∂2f∂2x+∂2f∂2y | (3) |
By separately taking the second derivatives of the Laplacian operator in the x and y directions through differencing, we obtain:
∇2f=f(x+1,y)+f(x−1,y)+f(x,y+1)+f(x,y−1)−4f(x,y) | (4) |
The above formulas can be transformed into the form of convolution kernels:
(0101−41010) | (5) |
After the Laplacian operator processing, the spectral features of cloud edges and broken clouds are enhanced. However, the extensive areas of thick clouds that manifest as low-frequency signals are removed. This can lead to sample imbalance and affect clustering results. Therefore, overlaying the image processed with the Laplacian operator onto the original image is crucial. The resulting feature-enhanced image can highlight cloud edges and broken cloud details while preserving the thick cloud portions from the original image. The overlay formula is as follows:
g(x,y)=f(x,y)∗w(i,j)=f(x,y)−[∇2f] | (6) |
Finally, based on Eq (6), we obtain the convolution kernel used in this paper:
w(i,j)=(0 - 10 - 15 - 10 - 10) | (7) |
Figure 8 provides a comparison of the results before and after feature enhancement. From Figure 8(a), (b), a noticeable improvement in the clarity of cloud edges and broken clouds that is within the RS image is observed after the feature enhancement. Additionally, the R3+4 values for cloud edges and broken clouds are extracted from the image for PDF comparison. Figure 8(c) shows that feature enhancement significantly increases the R3+4 values for cloud edges and broken clouds. This will effectively reduce the occurrence of false negatives in the algorithm for cloud edge and broken cloud pixels.
During detection, the diversity of cloud and the underlying surface results in complex distributions of reflectance and brightness temperature. GMM can represent the PDF of data by the way of using a convex combination of multiple Gaussian functions (a linear combination of multiple Gaussian functions). Given a sufficient number of samples, GMM can model distributions of any shape [30]. Therefore, GMM can accurately describe the probability distributions of reflectance and brightness temperature for both clouds and the underlying surface. Additionally, some spectral characteristics of clouds and underlying surfaces may overlap (as shown in Figure 5). The advantage of GMM compared to clustering algorithms like K-means and mean-shift lies in its more flexible modeling capability. Unlike K-means and other hard clustering methods, GMM has the ability to perform soft clustering which allows data points to belong to different clusters with certain probabilities. This flexibility helps in better understanding and representing relationships between data points, making the model more adaptable to clusters that exhibit overlap or have fuzzy boundaries. Additionally, GMM uses probability distributions to describe clusters, which makes it better suited for handling complex data distributions, especially when dealing with non-spherical distributions where GMM often outperforms other algorithms [31]. Its PDF formula is as follows:
P(x|θ)=K∑j=1βjf(x|μj,σj2) | (8) |
In the equation, x represents a vector X = {x1, x2, x3, ..., xn} consisting of n sample values, where X contains K categories. The weight of the jth class is denoted by βj, and the function f(x) is the PDF of a single Gaussian model with mean μj and variance σj for the jth class.
The set of data for the study area, R3+4, BT15 and BTD7-14, is denoted as X = {x1, x2, x3, ..., xn}, where xi is a sample drawn from the set X. The elements in the set are independent of each other and follow a GMM. As shown in Eq (7), the GMM density function contains a parameter set θ = {β, μ, σ2} and this parameter set is defined as θ. For a single Gaussian model, the parameters θ can be estimated using the Maximum Likelihood (ML) method, where the likelihood function is given by the PDF:
L(θ)=n∏i=1P(x|θ)=n∏i=11√2πσ2exp(−(xi−μ)22σ2) | (9) |
Similarly, the likelihood function can be derived from the GMM PDF as follows:
L(θ)=n∏i=1P(x|θ)=n∑i=1log[K∑j=1βjf(x|μj,σj2)] | (10) |
In the equation, there is a complex nonlinear relationship between the likelihood function L (θ) and the parameter set θ, making it impractical to directly maximize the likelihood function through differentiation. The Expectation-Maximization (EM) algorithm is a common method for maximum likelihood estimation and is suitable for data with hidden, unknown variables. It can appropriately estimate the optimal parameters of probability models [32,33]. The specific steps of the EM algorithm are as follows:
Step 1: Initialize the parameter set θ = {β, μ, σ2} for each category (hidden variables) based on prior knowledge.
Step 2: Calculate the probability that the data is generated by a Gaussian distribution as follows:
γ(i,j)=βjP(xi|μj,σj2)K∑j=1βjP(xi|μj,σj2) | (11) |
Step 3: Update the μ and σ2 for each Gaussian model as follows:
μj=n∑i=1γ(i,j)−xin∑i=1γ(i,j) | (12) |
σj2=n∑i=1γ(i,j)(xi−μj)(xi−μj)Tn∑i=1γ(i,j) | (13) |
Step 4: Update the β for each Gaussian model as follows:
βj=n∑n=1γ(i,j)n | (14) |
Repeat Steps 2 through Step 4 until the maximum likelihood value L (θ) approaches stability. Once this condition is met, the algorithm is complete.
Before applying the GMM algorithm for clustering, the value of K needs to be determined. In practical situations, the types of distributions contained in the data samples are unknown. As the number of sub-models in the observed samples increases, the mixture model better reflects the characteristics of the data samples. However, this also leads to a more complex model. A balance point should be established to maintain the simplicity of the mixture model and ensure the accuracy of the model fit. Therefore, how to determine the value of parameter K to ensure that the mixture model is both accurate and concise is a primary goal to optimize the model. We use the AIC and BIC criterions from information statistics to determine the parameter K and their expressions are as follows:
VAIC=2m−lnL(x|K,β,μ,σ2)VBIC=mlnn−lnL(x|K,β,μ,σ2) | (15) |
In the equations, m is the number of parameters in the mixture model, n is the number of samples, and lnL (x|K, β, μ, σ2) represents the maximum likelihood value. The minimum value of VAIC and VBIC is taken as the criteria for selecting the optimal model.
First, four sets of RS images of R3+4 are selected as sample data, and the range for selecting K is set between 1 and 14. VAIC and VBIC are calculated for different values of K. As shown in Figure 9, with the increase of K, both VAIC and VBIC show an overall decreasing trend, and the trend gradually stabilizes when K > 8. To further determine the specific value of K, values of K are set to 5, 6, 7 and 8, and the GMM cloud detection results for different K values are compared to determine the appropriate value of K. Figure 10 shows the comparison of cloud pixel omissions for K = 5, 6 and 7. As the increase in the value of K, the number of omitted cloud pixels gradually decreases. When K = 7, the algorithm produces detection results for various types of clouds that closely align with the actual cloud coverage. Furthermore, as shown in Figure 11, there is tiny little difference in the detection results of cloud edge details between K = 7 and K = 8. Therefore, K = 7 ensures that the mixture model is both accurate and concise.
The Himawari-8 data was provided by the Japan Aerospace Exploration Agency (https://www.eorc.jaxa.jp/ptree/). In this study, a dataset was constructed for AHI, encompassing partial data from 2021 and 2022. This dataset includes both full disk images and study area images, with over 1,400,000 pixels annotated.
The computer used for the experiments has an AMD R5 5600G processor with a base frequency of 2900 MHz, a 1 TB solid-state drive and 16 GB of RAM. The experiment uses the OpenCV and Scikit-Learn packages to implement the proposed algorithm.
The GMM algorithm selected in this study requires pre-determination of the number of clusters, similar to the K-means algorithm. Therefore, to demonstrate the advantage of GMM's soft clustering capability over K-means, a comparison was made. The selected images were taken during early morning or evening when the reflectance of clouds is lower compared to daytime, posing a significant challenge to the clustering algorithms. As shown in Figure 12 within the red box, K-means wrongly classified many low-reflectance clouds as underlying surfaces. This is because K-means tends to rigidly assign data points to the nearest cluster center when dealing with non-convex shapes or overlapping regions. Consequently, some low-reflectance clouds were erroneously categorized as land features by K-means. On the other hand, GMM is capable of better reflecting the probability of data points belonging to different clusters. Therefore, in such scenarios, GMM might more accurately identify areas with ambiguous boundaries. It excels in capturing complex data distributions and delineating regions with unclear boundaries, which might lead to a more precise identification of low-reflectance clouds compared to K-means.
The proposed algorithm uses a feature enhancement module to enhance the features of cloud edges and broken clouds, and it integrates the detection results of cloud edges and broken clouds in the final detection results, thereby improving the overall accuracy of the algorithm. Figure 13 shows the heatmap of the proposed algorithm, which reflects the aggregation of broken clouds and cloud edges with their respective categories. In Figure 13(a), (b), the model's attention to broken clouds is evident (highlighted in yellow boxes). Figure 13(b), (c) depict the attention to cloud edges, and it can be observed that the cloud edges of thin clouds also receive significant attention (highlighted in yellow boxes).
Traditional methods such as the multi-spectral threshold method are widely used, and machine learning methods like RF, KNN and GMM have efficient classification capabilities. Therefore, this paper implemented the proposed algorithm and the aforementioned algorithm and implemented cloud detection experiments on Himawari-8 images from different time phases and regions to validate the applicability of the proposed algorithm.
Figure 14 presents the comparison of cloud detection results using the multi-spectral threshold method [10], RF [15], KNN [34], GMM and the proposed algorithm. In the figure, the true labels obtained through visual interpretation are depicted in yellow, while the detection results are shown in white rendering. Professionally trained experts or analysts, following the cloud definition method outlined in Section 3.1, utilized Photoshop software to meticulously annotate the cloud's location pixel by pixel.
In Figure 14(a), the detection results of GMM and the proposed algorithm are the most accurate, while the multi-spectral threshold method and RF exhibit a slight underestimation of cloud pixels, primarily at the cloud edges. KNN shows the highest number of cloud pixel omissions (highlighted within the green bounding boxes in the images). In Figure 14(b), GMM and the proposed algorithm provide the best detection results, with the multi-spectral threshold method and RF exhibit the greatest number of cloud pixels omission. KNN exhibits a large area of omissions in the lower right corner of the image and it also misclassifies some non-cloud pixels as clouds at the cloud edges (highlighted within the purple bounding boxes for omissions and the yellow bounding boxes for misclassifications). In Figure 14(c), (d), the cloud edges in the upper right corners of the images are somewhat blurred. Upon comparison, it is evident that the proposed algorithm produces the most accurate detection results. Moreover, the multi-spectral threshold method, RF and GMM all exhibit varying degrees of omissions at the cloud edges. Multi-spectral threshold method has the most pronounced omission phenomenon, followed by RF (highlighted within the blue bounding boxes for omissions and the yellow bounding boxes for misclassifications).
For the detection of broken clouds, as depicted within the red box in Figure 14(d), the multi-spectral threshold method exhibits the highest number of missed broken cloud pixels, followed by RF and GMM. Additionally, KNN misclassifies numerous pixels as clouds. Analysis of the above comparison results reveals that in sufficient sunlight conditions during the day, the reflectance of clouds in the visible light spectrum is relatively high. During this time, the multi-spectral threshold method is more suitable. However, as it approaches evening, decreasing light levels can lead to higher threshold values, which is not conducive to cloud detection. RF and KNN, as supervised learning algorithms, have high requirements for sample size and the balance of samples among different categories. However, the small and dispersed nature of scattered clouds poses challenges in creating training samples for RF and KNN, making it difficult for these algorithms to learn the necessary features effectively.
In contrast, as an unsupervised learning algorithm, GMM can automatically learn the inherent relationships between different categories and perform classification without the need for an extensive set of training samples. Because the spectral characteristics of cloud edges are somewhat similar to those of underlying surfaces, the threshold set by the multi-spectral threshold method is relatively high for cloud edge detection. While GMM has several advantages in detecting large-area thick clouds, it does not consider the spectral feature similarity between cloud edges and underlying surfaces. Therefore, the proposed algorithm achieves the best detection results for various cloud types.
Additionally, as shown in Figure 15, full disk RS data from September 24, 2022, at UTC 03:00 are selected for cloud detection. Among the five algorithms, the multi-spectral threshold method and GMM achieve relatively good detection results, but there are instances of cloud underestimation in some regions (highlighted within the green boxes). RF and KNN show varying degrees of cloud false positives (highlighted within the red boxes). The proposed algorithm, an enhancement of GMM, reduces the underestimation of cloud pixels, resulting in the best detection performance among the five algorithms, demonstrating that this algorithm provides reliable results even for large-scale areas.
To quantitatively evaluate these methods, this study uses two evaluation metrics, the hit rate (HR) and the Hansen-Kuiper skill score (KSS), to analyze 15 RS images from different times and regions. The HR represents the proportion of correctly identified cloud and clear-sky pixels to the total number of pixels. The KSS score is used to provide a comprehensive overall measure. The formulas for calculating HR and KSS are as follows:
HR=TC+TUPA | (16) |
KSS=TC×TU−TF×FT(TC+FT)(TU+TF) | (17) |
In the formulas, TC represents the number of pixels correctly detected as clouds, TU is the number of pixels that are actually non-cloud but are detected as clouds, FT is the number of pixels that are actually clouds but are detected as non-cloud and PA is the total number of pixels in the image.
Figure 16 shows the evaluation results of different algorithms for detecting thick clouds, cloud edges and broken clouds. The values of HR and KSS approaching 1 indicate superior algorithm performance. For thick cloud detection, the HR values of the multi-spectral threshold method, RF, KNN, GMM and the proposed algorithm are 0.793, 0.832, 0.799, 0.974 and 0.976, respectively, and the KSS values are 0.692, 0.763, 0.642, 0.973 and 0.974, respectively. It is evident that GMM and the proposed algorithm outperform the other three algorithms in thick cloud detection. For cloud edge and broken cloud detection, the HR values of the five algorithms are 0.619, 0.687, 0.624, 0.963 and 0.974, respectively, and the KSS values are 0.458, 0.561, 0.563, 0.947 and 0.973, respectively. Compared to thick cloud detection, the HR and KSS values of the multi-spectral threshold method, RF and KNN significantly decrease in cloud edge and broken cloud detection, while those of GMM and the proposed algorithm remain above 0.96 and 0.94, respectively. Moreover, the proposed algorithm, which incorporates the feature enhancement module, shows an improvement of 0.011 in HR and 0.026 in KSS compared to GMM. This indicates that the feature enhancement module aids the classification model in extracting deeper cloud edge and broken cloud features. Additionally, as shown in Table 2, 20 images sized at 256 × 256 were selected for quantitative analysis to demonstrate the overall performance of the algorithm. Compared to GMM, the proposed algorithm shows an improvement of 0.002 in HR and 0.015 in KSS.
Algorithm | HR | KSS |
Multi-spectral threshold method | 0.763 | 0.572 |
RF | 0.811 | 0.65 |
KNN | 0.753 | 0.606 |
GMM | 0.973 | 0.958 |
Proposed algorithm | 0.975 | 0.973 |
When performing target detection and task analysis, the vast amount of RS data meant that the runtime of cloud detection algorithms is an important concern. To address this, RS images of sizes 1000 × 1100, 2401 × 2401 and 60016001 are selected to compare the runtime differences between different algorithms. As shown in Figure 17, when the image sizes are 1000 × 1100 and 2401 × 2401, the runtime differences among the four machine learning algorithms does not vary significantly, with GMM having the shortest runtime, followed by the proposed algorithm, RF and KNN. However, when the image size is 6001 × 6001, KNN's runtime is significantly longer than the other three algorithms (approximately 3.9 times longer than GMM and 3.65 times longer than RF), while the proposed algorithm with the feature enhancement module has a runtime only 7.45 seconds longer than GMM. Therefore, for cloud detection in large-sized RS images, the algorithm proposed in this study can maintain high accuracy while having a relatively short runtime.
While the proposed algorithm generally achieves satisfactory cloud detection results in most scenarios, it encounters errors, particularly when dealing with cloud overlap. This defect is especially obvious when dealing with cloud shadows. These shadows, appearing as dark regions in visible light images, share spectral characteristics similar to the underlying surfaces. The similarity poses a challenge in distinguishing clouds within these shadowed areas, leading to errors in identification. Figure 18 illustrates examples of cloud detection errors specifically within shadowed regions, highlighting numerous omitted pixels in the detection results. In addition, the proposed algorithm is tailored for satellites capable of providing 2 km resolution images. With higher image resolution, the diversity of surface types becomes more intricate, leading to fluctuations in the optimal K value. In such scenarios, extensive experimental analysis becomes crucial for re-evaluating the selection of the K value. Even then, pinpointing a stable K value can be challenging. For instance, in an image observed at 2 km resolution, a pixel m (x, y) might initially be identified as a cloud. However, as the resolution increases, the area represented by m (x, y) encompasses more pixels. Among these newly added pixels, some might reveal underlying surfaces that were previously unobservable at the 2 km resolution.
The future research goal is to combine thermal infrared spectroscopy with neural network algorithms to enhance the algorithm's transferability. Moreover, by exploring the imaging correlation between clouds and their shadows, there is potential to calculate the orientation and distance of shadows concerning cloud areas. This computational approach could be helpful in effectively filtering out these shadows from the detection results.
The improved cloud detection technology holds significant potential in disaster management. Accurate cloud coverage data is meaningful to monitoring and early warning systems for natural disasters such as storms and floods, aiding in the proactive implementation of response measures and minimizing potential losses. However, when applying these technologies, the critical importance of data accuracy and timeliness in decision-making must be considered to prevent the potential risks of misinformation or delays.
Furthermore, advancements in cloud detection technology may raise concerns regarding personal privacy, particularly in the monitoring and analysis of extensive areas. Therefore, safeguarding personal information and data privacy should be a crucial consideration when utilizing these algorithms, necessitating the establishment of stringent guidelines for data processing and sharing.
In this paper, we propose a multi-spectral data cloud detection algorithm based on a feature enhancement module and Gaussian mixture model (GMM) clustering. First, the spectral data of clouds and underlying surfaces are statistically analyzed. Based on the analysis results, an appropriate feature scheme is selected to achieve cloud detection throughout the day. Second, the Laplacian operator is used to enhance the spectral features of cloud edges and broken clouds. The enhanced data is then input into a GMM classifier for cloud edge and broken cloud detection. The original RS data is simultaneously subjected to GMM clustering to obtain thick cloud detection results. Finally, the results from the both branches are combined to obtain the final cloud detection result.
Comparing the results of the multi-spectral threshold methods, RF, KNN, GMM and the proposed algorithm, the following conclusions are drawn: The proposed algorithm achieves the highest accuracy with a hit rate (HR) of 0.976 for the case of thick cloud detection and a Hansen-Kuiper skill score (KSS) of 0.974 for the case of cloud edges and broken cloud detection. This means the proposed algorithm demonstrates significant advantages over other algorithms, with an HR and KSS improvement of 0.011 and 0.026, respectively, compared to GMM. Furthermore, the inclusion of the feature enhancement module in the proposed algorithm maintains a relatively short runtime. This further demonstrates that the proposed algorithm can achieve fast and high-precision cloud detection for different types of clouds. Therefore, the proposed algorithm can provide real-time cloud detection results for various tasks such as vegetation change analysis, disaster monitoring and cartography, offering a viable solution to alleviate the challenges posed by cloud interference.
The authors declare that they have not used Artificial Intelligence (AI) tools in the creation of this article.
This work is supported by the Major Scientific and Technological Projects of Yunnan Province: Research on Key Technologies of Ecological Environment Monitoring and Intelligent Management of Natural Resources in Yunnan (202202AD080010), the National Natural Science Foundation of China (52107017) and the Fundamental Research Fund of Science and Technology Department of Yunnan Province (202201AU070172).
The authors declare that there are no conflicts of interest.
[1] |
Baron JC, Yamauchi H, Fujioka M, et al. (2014) Selective neuronal loss in ischemic stroke and cerebrovascular disease. J Cereb Blood Flow Metab 34: 2-18. doi: 10.1038/jcbfm.2013.188
![]() |
[2] |
Zhang J, Zhang Y, Xing S, et al. (2012) Secondary neurodegeneration in remote regions after focal cerebral infarction: a new target for stroke management? Stroke 43: 1700-1705. doi: 10.1161/STROKEAHA.111.632448
![]() |
[3] |
Levine DA, Galecki AT, Langa KM, et al. (2015) Trajectory of Cognitive Decline After Incident Stroke. JAMA 314: 41-51. doi: 10.1001/jama.2015.6968
![]() |
[4] |
Yang J, Wong A, Wang Z, et al. (2015) Risk factors for incident dementia after stroke and transient ischemic attack. Alzheimers Dement 11: 16-23. doi: 10.1016/j.jalz.2014.01.003
![]() |
[5] |
Querfurth HW, LaFerla FM (2010) Alzheimer's disease. N Engl J Med 362: 329-344. doi: 10.1056/NEJMra0909142
![]() |
[6] |
Caughey B, Lansbury PT (2003) Protofibrils, pores, fibrils, and neurodegeneration: separating the responsible protein aggregates from the innocent bystanders. Annu Rev Neurosci 26: 267-298. doi: 10.1146/annurev.neuro.26.010302.081142
![]() |
[7] | Haass C, Selkoe DJ (2007) Soluble protein oligomers in neurodegeneration: lessons from the Alzheimer's amyloid beta-peptide. Nat Rev Mol Cell Biol 8: 101-112. |
[8] |
Roychaudhuri R, Yang M, Hoshi MM, et al. (2009) Amyloid beta-protein assembly and Alzheimer disease. J Biol Chem 284: 4749-4753. doi: 10.1074/jbc.R800036200
![]() |
[9] | Giuffrida ML, Caraci F, De Bona P, et al. (2010) The monomer state of beta-amyloid: where the Alzheimer's disease protein meets physiology. Rev Neurosci 21: 83-93. |
[10] |
Gilbert BJ (2013) The role of amyloid beta in the pathogenesis of Alzheimer's disease. J Clin Pathol 66: 362-366. doi: 10.1136/jclinpath-2013-201515
![]() |
[11] | Carrillo-Mora P, Luna R, Colin-Barenque L (2014) Amyloid beta: multiple mechanisms of toxicity and only some protective effects? Oxid Med Cell Longev 2014: 795375. |
[12] |
Viola KL, Klein WL (2015) Amyloid beta oligomers in Alzheimer's disease pathogenesis, treatment, and diagnosis. Acta Neuropathol 129: 183-206. doi: 10.1007/s00401-015-1386-3
![]() |
[13] | Bruggink KA, Muller M, Kuiperij HB, et al. (2012) Methods for analysis of amyloid-beta aggregates. J Alzheimers Dis 28: 735-758. |
[14] |
Pryor NE, Moss MA, Hestekin CN (2012) Unraveling the early events of amyloid-beta protein (Abeta) aggregation: techniques for the determination of Abeta aggregate size. Int J Mol Sci 13: 3038-3072. doi: 10.3390/ijms13033038
![]() |
[15] |
van Groen T, Puurunen K, Maki HM, et al. (2005) Transformation of diffuse beta-amyloid precursor protein and beta-amyloid deposits to plaques in the thalamus after transient occlusion of the middle cerebral artery in rats. Stroke 36: 1551-1556. doi: 10.1161/01.STR.0000169933.88903.cf
![]() |
[16] |
Aho L, Jolkkonen J, Alafuzoff I (2006) Beta-amyloid aggregation in human brains with cerebrovascular lesions. Stroke 37: 2940-2945. doi: 10.1161/01.STR.0000248777.44128.93
![]() |
[17] |
Makinen S, van Groen T, Clarke J, et al. (2008) Coaccumulation of calcium and beta-amyloid in the thalamus after transient middle cerebral artery occlusion in rats. J Cereb Blood Flow Metab 28: 263-268. doi: 10.1038/sj.jcbfm.9600529
![]() |
[18] |
Ly JV, Rowe CC, Villemagne VL, et al. (2012) Subacute ischemic stroke is associated with focal 11C PiB positron emission tomography retention but not with global neocortical Abeta deposition. Stroke 43: 1341-1346. doi: 10.1161/STROKEAHA.111.636266
![]() |
[19] |
Lipsanen A, Kalesnykas G, Pro-Sistiaga P, et al. (2013) Lack of secondary pathology in the thalamus after focal cerebral ischemia in nonhuman primates. Exp Neurol 248: 224-227. doi: 10.1016/j.expneurol.2013.06.016
![]() |
[20] |
Sahathevan R, Linden T, Villemagne VL, et al. (2016) Positron Emission Tomographic Imaging in Stroke: Cross-Sectional and Follow-Up Assessment of Amyloid in Ischemic Stroke. Stroke 47: 113-119. doi: 10.1161/STROKEAHA.115.010528
![]() |
[21] | Ong LK, Zhao Z, Kluge M, et al. (2016) Chronic stress exposure following photothrombotic stroke is associated with increased levels of Amyloid beta accumulation and altered oligomerisation at sites of thalamic secondary neurodegeneration in mice. J Cereb Blood Flow Metab. |
[22] |
Lesne S, Koh MT, Kotilinek L, et al. (2006) A specific amyloid-beta protein assembly in the brain impairs memory. Nature 440: 352-357. doi: 10.1038/nature04533
![]() |
[23] |
Selkoe DJ, Hardy J (2016) The amyloid hypothesis of Alzheimer's disease at 25 years. EMBO Mol Med 8: 595-608. doi: 10.15252/emmm.201606210
![]() |
[24] | Sheikh S, Safia, Haque E, et al. (2013) Neurodegenerative Diseases: Multifactorial Conformational Diseases and Their Therapeutic Interventions. J Neurodegener Dis 2013: 563481. |
[25] |
Walsh DM, Klyubin I, Fadeeva JV, et al. (2002) Naturally secreted oligomers of amyloid beta protein potently inhibit hippocampal long-term potentiation in vivo. Nature 416: 535-539. doi: 10.1038/416535a
![]() |
[26] |
Buell AK, Dobson CM, Knowles TP, et al. (2010) Interactions between amyloidophilic dyes and their relevance to studies of amyloid inhibitors. Biophys J 99: 3492-3497. doi: 10.1016/j.bpj.2010.08.074
![]() |
[27] |
Nesterov EE, Skoch J, Hyman BT, et al. (2005) In vivo optical imaging of amyloid aggregates in brain: design of fluorescent markers. Angew Chem Int Ed Engl 44: 5452-5456. doi: 10.1002/anie.200500845
![]() |
[28] |
Klunk WE, Engler H, Nordberg A, et al. (2004) Imaging brain amyloid in Alzheimer's disease with Pittsburgh Compound-B. Ann Neurol 55: 306-319. doi: 10.1002/ana.20009
![]() |
[29] |
Cohen AD, Rabinovici GD, Mathis CA, et al. (2012) Using Pittsburgh Compound B for in vivo PET imaging of fibrillar amyloid-beta. Adv Pharmacol 64: 27-81. doi: 10.1016/B978-0-12-394816-8.00002-7
![]() |
[30] |
Liu W, Wong A, Au L, et al. (2015) Influence of Amyloid-beta on Cognitive Decline After Stroke/Transient Ischemic Attack: Three-Year Longitudinal Study. Stroke 46: 3074-3080. doi: 10.1161/STROKEAHA.115.010449
![]() |
[31] | Rowe CC, Villemagne VL (2011) Brain amyloid imaging. J Nucl Med 52: 1733-1740. |
[32] |
Glass CK, Saijo K, Winner B, et al. (2010) Mechanisms underlying inflammation in neurodegeneration. Cell 140: 918-934. doi: 10.1016/j.cell.2010.02.016
![]() |
[33] |
Martin JB (1999) Molecular basis of the neurodegenerative disorders. N Engl J Med 340: 1970-1980. doi: 10.1056/NEJM199906243402507
![]() |
[34] | Kovacs GG (2016) Molecular Pathological Classification of Neurodegenerative Diseases: Turning towards Precision Medicine. Int J Mol Sci 17. |
[35] |
Tamura A, Tahira Y, Nagashima H, et al. (1991) Thalamic atrophy following cerebral infarction in the territory of the middle cerebral artery. Stroke 22: 615-618. doi: 10.1161/01.STR.22.5.615
![]() |
[36] |
Ogawa T, Yoshida Y, Okudera T, et al. (1997) Secondary thalamic degeneration after cerebral infarction in the middle cerebral artery distribution: evaluation with MR imaging. Radiology 204: 255-262. doi: 10.1148/radiology.204.1.9205256
![]() |
[37] |
Nakane M, Tamura A, Sasaki Y, et al. (2002) MRI of secondary changes in the thalamus following a cerebral infarct. Neuroradiology 44: 915-920. doi: 10.1007/s00234-002-0846-3
![]() |
[38] |
Li C, Ling X, Liu S, et al. (2011) Early detection of secondary damage in ipsilateral thalamus after acute infarction at unilateral corona radiata by diffusion tensor imaging and magnetic resonance spectroscopy. BMC Neurol 11: 49. doi: 10.1186/1471-2377-11-49
![]() |
[39] |
Buffon F, Molko N, Herve D, et al. (2005) Longitudinal diffusion changes in cerebral hemispheres after MCA infarcts. J Cereb Blood Flow Metab 25: 641-650. doi: 10.1038/sj.jcbfm.9600054
![]() |
[40] |
Herve D, Molko N, Pappata S, et al. (2005) Longitudinal thalamic diffusion changes after middle cerebral artery infarcts. J Neurol Neurosurg Psychiatry 76: 200-205. doi: 10.1136/jnnp.2004.041012
![]() |
[41] |
Pappata S, Levasseur M, Gunn RN, et al. (2000) Thalamic microglial activation in ischemic stroke detected in vivo by PET and [11C]PK1195. Neurology 55: 1052-1054. doi: 10.1212/WNL.55.7.1052
![]() |
[42] |
Gerhard A, Schwarz J, Myers R, et al. (2005) Evolution of microglial activation in patients after ischemic stroke: a [11C](R)-PK11195 PET study. Neuroimage 24: 591-595. doi: 10.1016/j.neuroimage.2004.09.034
![]() |
[43] |
Thiel A, Heiss WD (2011) Imaging of microglia activation in stroke. Stroke 42: 507-512. doi: 10.1161/STROKEAHA.110.598821
![]() |
[44] |
Fujie W, Kirino T, Tomukai N, et al. (1990) Progressive shrinkage of the thalamus following middle cerebral artery occlusion in rats. Stroke 21: 1485-1488. doi: 10.1161/01.STR.21.10.1485
![]() |
[45] |
Iizuka H, Sakatani K, Young W (1990) Neural damage in the rat thalamus after cortical infarcts. Stroke 21: 790-794. doi: 10.1161/01.STR.21.5.790
![]() |
[46] |
Dihne M, Grommes C, Lutzenburg M, et al. (2002) Different mechanisms of secondary neuronal damage in thalamic nuclei after focal cerebral ischemia in rats. Stroke 33: 3006-3011. doi: 10.1161/01.STR.0000039406.64644.CB
![]() |
[47] |
Justicia C, Ramos-Cabrer P, Hoehn M (2008) MRI detection of secondary damage after stroke: chronic iron accumulation in the thalamus of the rat brain. Stroke 39: 1541-1547. doi: 10.1161/STROKEAHA.107.503565
![]() |
[48] |
Bihel E, Pro-Sistiaga P, Letourneur A, et al. (2010) Permanent or transient chronic ischemic stroke in the non-human primate: behavioral, neuroimaging, histological, and immunohistochemical investigations. J Cereb Blood Flow Metab 30: 273-285. doi: 10.1038/jcbfm.2009.209
![]() |
[49] |
Ross DT, Ebner FF (1990) Thalamic retrograde degeneration following cortical injury: an excitotoxic process? Neuroscience 35: 525-550. doi: 10.1016/0306-4522(90)90327-Z
![]() |
[50] |
Jones KA, Zouikr I, Patience M, et al. (2015) Chronic stress exacerbates neuronal loss associated with secondary neurodegeneration and suppresses microglial-like cells following focal motor cortex ischemia in the mouse. Brain Behav Immun 48: 57-67. doi: 10.1016/j.bbi.2015.02.014
![]() |
[51] |
Patience MJ, Zouikr I, Jones K, et al. (2015) Photothrombotic stroke induces persistent ipsilateral and contralateral astrogliosis in key cognitive control nuclei. Neurochem Res 40: 362-371. doi: 10.1007/s11064-014-1487-8
![]() |
[52] |
Hiltunen M, Makinen P, Peraniemi S, et al. (2009) Focal cerebral ischemia in rats alters APP processing and expression of Abeta peptide degrading enzymes in the thalamus. Neurobiol Dis 35: 103-113. doi: 10.1016/j.nbd.2009.04.009
![]() |
[53] |
Zhang Y, Xing S, Zhang J, et al. (2011) Reduction of beta-amyloid deposits by gamma-secretase inhibitor is associated with the attenuation of secondary damage in the ipsilateral thalamus and sensory functional improvement after focal cortical infarction in hypertensive rats. J Cereb Blood Flow Metab 31: 572-579. doi: 10.1038/jcbfm.2010.127
![]() |
[54] |
Sarajarvi T, Lipsanen A, Makinen P, et al. (2012) Bepridil decreases Abeta and calcium levels in the thalamus after middle cerebral artery occlusion in rats. J Cell Mol Med 16: 2754-2767. doi: 10.1111/j.1582-4934.2012.01599.x
![]() |
[55] |
Zhang J, Zhang Y, Li J, et al. (2012) Autophagosomes accumulation is associated with beta-amyloid deposits and secondary damage in the thalamus after focal cortical infarction in hypertensive rats. J Neurochem 120: 564-573. doi: 10.1111/j.1471-4159.2011.07496.x
![]() |
[56] |
Mitkari B, Kerkela E, Nystedt J, et al. (2015) Unexpected complication in a rat stroke model: exacerbation of secondary pathology in the thalamus by subacute intraarterial administration of human bone marrow-derived mesenchymal stem cells. J Cereb Blood Flow Metab 35: 363-366. doi: 10.1038/jcbfm.2014.235
![]() |
Channel | Central wavelength/μm | Detection targets |
1 | 0.46 | aerosol, coastline |
2 | 0.51 | phytoplankton |
3 | 0.64 | vegetation, aerosol |
4 | 0.86 | stratus cloud |
5 | 1.6 | cloud top phase, snow |
6 | 2.3 | surface, cloud, snow |
7 | 3.9 | surface, wildfire |
8 | 6.2 | upper atmospheric water vapor, rainfall |
9 | 7.0 | mid-level atmospheric water vapor, rainfall |
10 | 7.3 | sulfur dioxide |
11 | 8.6 | water, sulfur dioxide, rainfall |
12 | 9.6 | ozone, airflow, wind |
13 | 10.4 | surface, cloud |
14 | 11.2 | surface, cloud |
15 | 12.4 | water, sea surface temperature |
16 | 13.3 | temperature, cloud altitude, cloudiness |
Algorithm | HR | KSS |
Multi-spectral threshold method | 0.763 | 0.572 |
RF | 0.811 | 0.65 |
KNN | 0.753 | 0.606 |
GMM | 0.973 | 0.958 |
Proposed algorithm | 0.975 | 0.973 |
Channel | Central wavelength/μm | Detection targets |
1 | 0.46 | aerosol, coastline |
2 | 0.51 | phytoplankton |
3 | 0.64 | vegetation, aerosol |
4 | 0.86 | stratus cloud |
5 | 1.6 | cloud top phase, snow |
6 | 2.3 | surface, cloud, snow |
7 | 3.9 | surface, wildfire |
8 | 6.2 | upper atmospheric water vapor, rainfall |
9 | 7.0 | mid-level atmospheric water vapor, rainfall |
10 | 7.3 | sulfur dioxide |
11 | 8.6 | water, sulfur dioxide, rainfall |
12 | 9.6 | ozone, airflow, wind |
13 | 10.4 | surface, cloud |
14 | 11.2 | surface, cloud |
15 | 12.4 | water, sea surface temperature |
16 | 13.3 | temperature, cloud altitude, cloudiness |
Algorithm | HR | KSS |
Multi-spectral threshold method | 0.763 | 0.572 |
RF | 0.811 | 0.65 |
KNN | 0.753 | 0.606 |
GMM | 0.973 | 0.958 |
Proposed algorithm | 0.975 | 0.973 |