
Citation: Fernando Arias, Maytee Zambrano, Kathia Broce, Carlos Medina, Hazel Pacheco, Yerenis Nunez. Hyperspectral imaging for rice cultivation: Applications, methods and challenges[J]. AIMS Agriculture and Food, 2021, 6(1): 273-307. doi: 10.3934/agrfood.2021018
[1] | Zhongnan Ran, Mingfeng Jiang, Yang Li, Zhefeng Wang, Yongquan Wu, Wei Ke, Ling Xia . Arrhythmia classification based on multi-feature multi-path parallel deep convolutional neural networks and improved focal loss. Mathematical Biosciences and Engineering, 2024, 21(4): 5521-5535. doi: 10.3934/mbe.2024243 |
[2] | Eleonora Sulas, Monica Urru, Roberto Tumbarello, Luigi Raffo, Danilo Pani . Systematic analysis of single- and multi-reference adaptive filters for non-invasive fetal electrocardiography. Mathematical Biosciences and Engineering, 2020, 17(1): 286-308. doi: 10.3934/mbe.2020016 |
[3] | Peng Zhang, Mingfeng Jiang, Yang Li, Ling Xia, Zhefeng Wang, Yongquan Wu, Yaming Wang, Huaxiong Zhang . An efficient ECG denoising method by fusing ECA-Net and CycleGAN. Mathematical Biosciences and Engineering, 2023, 20(7): 13415-13433. doi: 10.3934/mbe.2023598 |
[4] | Enes Efe, Emrehan Yavsan . AttBiLFNet: A novel hybrid network for accurate and efficient arrhythmia detection in imbalanced ECG signals. Mathematical Biosciences and Engineering, 2024, 21(4): 5863-5880. doi: 10.3934/mbe.2024259 |
[5] | Chunkai Zhang, Yingyang Chen, Ao Yin, Xuan Wang . Anomaly detection in ECG based on trend symbolic aggregate approximation. Mathematical Biosciences and Engineering, 2019, 16(4): 2154-2167. doi: 10.3934/mbe.2019105 |
[6] | Yuni Zeng, Hang Lv, Mingfeng Jiang, Jucheng Zhang, Ling Xia, Yaming Wang, Zhikang Wang . Deep arrhythmia classification based on SENet and lightweight context transform. Mathematical Biosciences and Engineering, 2023, 20(1): 1-17. doi: 10.3934/mbe.2023001 |
[7] | Diguo Zhai, Xinqi Bao, Xi Long, Taotao Ru, Guofu Zhou . Precise detection and localization of R-peaks from ECG signals. Mathematical Biosciences and Engineering, 2023, 20(11): 19191-19208. doi: 10.3934/mbe.2023848 |
[8] | Xiaowen Jia, Jingxia Chen, Kexin Liu, Qian Wang, Jialing He . Multimodal depression detection based on an attention graph convolution and transformer. Mathematical Biosciences and Engineering, 2025, 22(3): 652-676. doi: 10.3934/mbe.2025024 |
[9] | Mingming Zhang, Huiyuan Jin, Ying Yang . ECG classification efficient modeling with artificial bee colony optimization data augmentation and attention mechanism. Mathematical Biosciences and Engineering, 2024, 21(3): 4626-4647. doi: 10.3934/mbe.2024203 |
[10] | Wei Zhong, Li Mao, Wei Du . A signal quality assessment method for fetal QRS complexes detection. Mathematical Biosciences and Engineering, 2023, 20(5): 7943-7956. doi: 10.3934/mbe.2023344 |
Data from the World Heart Federation shows that the number of people with cardiovascular disease worldwide has exceeded 500 million. Known as the number one killer that threatens human health, cardiovascular disease is increasingly occurring in young people, including cardiac arrhythmia. An effective way to diagnose arrhythmia is to use electrocardiogram (ECG) signals, which contain important information for diagnosing the type of abnormal heart rhythm. But some information is less differentiated and an experienced physician is required. Therefore, automatic arrhythmia detection is particularly important.
To date, there have been numerous studies on arrhythmia detection, including ECG signal preprocessing, feature extraction and classification models. In ECG signal preprocessing, noise reduction techniques are widely used to remove artifacts from baseline wander, powerline interference, electrode motion, muscle artifacts, etc. Feature extraction is crucial in ECG signal classification, but the classification results are heavily dependent on the complex feature extraction process, which cannot adapt to different application environments and may result in some important information loss. As deep learning techniques evolve, features will become automatically selectable and classification tasks will become processable end-to-end. Such techniques are also widely used in arrhythmia detection. However, plenty of studies focus on single lead information, and the interpretability is not strong. At the same time, multi-lead data requires analysis of the relationship between leads and the heartbeats within the lead as shown in Figure 1. In this figure, the ECG signals include 12 leads, i.e., Ⅰ, Ⅱ, Ⅲ, AVR, AVL, AVF, V1, V2, V3, V4, V5 and V6. Each lead (marked in green) contains information on multiple heartbeats, which are called single-lead beats (marked in red). Single-lead beats in the same time interval in multiple leads constitute one multi-lead beat (marked in blue). However, many studies have focused on the diagnosis of one-dimensional lead information or simple multi-lead information fusion, which cannot make good use of the comprehensive information in multi-lead signals, such as the relationship between the leads and heartbeats, or the relationship between multi-lead beats and single-lead beats.
In summary, it is critical to simultaneously construct intra-lead and inter-lead correlations in the model for multi-lead arrhythmia detection. The instances (nodes) in the graph are associated with other nodes through complex correlations that are available to capture their interrelationships. Heterogeneous graphs take into account the rich interactions between multiple types of nodes and edges. Multi-lead ECG signals include different node types such as multi-lead beats, single-lead beats, and individual leads, meanwhile, they have complicated relationships among them. Therefore, in a study on arrhythmia detection using multi-lead ECG signals, heterogeneous graphs are suitable for modeling heartbeat data. The embedding of each node in the graph is jointly represented by the characteristics of the node and related nodes. In addition, interaction-related information is derived by means of information propagation between neighboring nodes in the graph.
To capture the rich interactions in multi-lead ECG signal data, an arrhythmia detection framework based on a heterogeneous graph attention network (MADHGAT) is developed in this work. A heterogeneous graph has been modeled for multi-lead ECG signals to integrate diverse data, which contains single-lead beats, multi-lead beats and multiple leads as nodes, additionally, intra-lead and inter-lead correlations are constructed. Heterogeneous graph attention networks (HGATs) are utilized for heterogeneous graph embedding to perform multi-lead beat classification as arrhythmia detection. Implementing a dual-level attention strategy in the HGAT allows for the learning of node-level and type-level contributions of adjacent nodes. The proposed framework does not need require any feature extraction for the ECG signals. We describe extensive experiments and show the effectiveness of the proposed model. The following key contributions of this work are summarized.
1) A heterogeneous graph has been modeled for multi-lead ECG signals to capture rich information, which can integrate diverse signal information and consider the inter-lead and intra-lead correlations at the same time.
2) An arrhythmia detection framework with HGATs is proposed for heterogeneous graph embedding and automatic cardiac arrhythmia classification. The implementation of a dual-level attention strategy in the HGAT permits the capture of the complex interactions between different heartbeat nodes and different node types.
3) To the best of our knowledge, this is the first time that multi-lead ECG signals have been modeled as a heterogeneous graph and HGATs have been employed for cardiac arrhythmia classification.
4) The results of extensive experiments indicate that the proposed model is able to well integrate the information of multiple leads, and results in significant improvements when applied to the INCART dataset without any feature extraction.
Automatic arrhythmia detection algorithms for ECG signal data usually have three main tasks, which include data preprocessing, feature extraction and classification [1]. The preprocessing operations include data denoising [2] and heartbeat segmentation [3]. Feature extraction is crucial in ECG signal classification. Features can be extracted from the ECG signal's shape in the spatiotemporal domain, such as the QRS complex [4], R-peak [5], ST segment [6], Shannon entropy [7] and T duration [8]. Finally, signals are categorized into different types by machine learning methods, such as multilayer perceptrons (MLPs) [9], k-nearest neighbors [10], deep learning methods [11,12] and support vector machines [13]. However, the classification results heavily rely on a complex feature extraction process, which cannot be adapted to various environments. Additionally, missing important information may lead to misdiagnosis. Deep learning methods play an important role in automatic feature extraction, which are widely utilized for feature extraction [14,15,16] or arrhythmia detection [17,18,19] in an end-to-end mode without any handcrafted features. Deep learning methods reduce the amount of manual work and improve the detection performance. As shown in Figure 1, correlations among multi-lead ECG signals constitute important information for arrhythmia detection. However, most existing works regard beats as independent information, ignoring the intra-lead correlations among beats and the inter-lead correlations among leads.
ECG signals are shown as time-series data, so the intra-lead correlations among beats are mainly reflected in the temporal properties, which include statistical time features [20,21,22] and temporal dependency [23,24,25]. The temporal properties have been considered for both single-lead signals and multi-lead signals. Mahajan et al. [22] collected time-frequency data and linear and nonlinear features for single-lead ECG signal classification. Yao et al. [26] developed a time-incremental network based on an attention mechanism. It considered the spatial and temporal features of multi-lead ECG signals. Che et al. [27] developed a deep learning framework that integrated a transformer network into a convolutional neural network to capture the temporal features of multi-lead ECG signals. Compared to single-lead ECG signals, multi-lead signals contain more complex and abundant information on heartbeats, which is useful for achieving a more accurate cardiac disease diagnosis. However, some works only consider the temporal properties in ECG signals, ignoring the inter-lead correlations in multi-lead data as well as the correlations between multi lead beats and single-lead beats.
Various works have been proposed for automatic arrhythmia diagnosis based on multi-lead ECG signal data, such as studies focused on feature fusion [28,29,30] and inter-lead correlations [31,32,33]. Han and Shi [34] detected and located myocardial infarction through the use of a multi-lead residual neural network and feature fusion for 12-lead ECG signal data. Sepahvand and Abdali-Mohammadi [35] developed a multi-lead ECG personal recognition system, which calculated intra-correlations and cross-correlations among multi-lead signal data in the time-frequency domain to estimate functional dependencies. Different leads correspond to different aspects of heart activity. These works take multi-lead signals into consideration, thus improving the accuracy of arrhythmia diagnoses. However, feature fusion of multiple single-lead signals cannot systematically exploit the inter-lead signal correlations. Additionally, a majority of the aforementioned studies did not consider the intra-lead correlations between beats.
Graph neural networks include graph convolutional networks [36], graph attention networks [37], graph spatiotemporal networks [38], etc. They are applied successfully for many tasks, such as computer vision [39], traffic forecasting [38], protein interface prediction [40] and disease prediction [41]. However, there is no related work on automatic arrhythmia diagnosis with multi-lead ECG signals. Each element in a graph neural network is taken as a node while the relationships between elements are denoted by edges. The structure of homogeneous graphs [42,43], which only contain a single type of nodes and edges, is relatively simple. Therefore, graph neural networks for homogeneous graphs just need to aggregate a single type of neighbors to update the embedding of nodes. For example, a graph convolutional neural network was proposed to diagnose pneumonia caused by COVID-19 [44]. However, a majority of graphs in the real world should be constructed into heterogeneous graphs in consideration of the complex interactions among diverse nodes and edges, thus avoiding information loss. Compared to homogeneous graphs, the difference between neighbors in different relationships should be taken into consideration in heterogeneous graphs [45,46]. For example, a heterogeneous network with a variational graph auto-encoder is presented for miRNA-disease association prediction [47]. Therefore, in consideration of the various information types and relationships associated with multi-lead ECG signals, a heterogeneous graph can be constructed to capture the complex correlations within multi-lead ECG signal data for the task of automatic arrhythmia diagnosis, thereby improving diagnosis performance.
We present an automatic arrhythmia detection framework based on an HGAT, which fully exploits both intra-lead correlations between beats and inter-lead correlations between leads based on the information spread along the graph. A flexible heterogeneous graph is presented first to model multi-lead ECG signal data and integrate abundant information as well as to capture the dependencies among multi-lead beats, single-lead beats and multiple leads, as shown in Figure 2. Then, we present a heterogeneous graph network with a dual-level attention strategy to embed the heterogeneous graph for arrhythmia detection. HGATs take into account information heterogeneity. Moreover, the attention strategy is able to learn the weights of different heartbeat nodes and node types.
Multi-lead beats were merely regarded as data objects for arrhythmia detection via multi-lead ECG signal data analysis in previous studies. Excluding for multi-lead beats, two types of additional information are considered, i.e., multiple leads and single-lead beats. As illustrated in Figure 2, the heterogeneous graph G=(V,E) is constructed to model multiple leads ML={ml1,…,mlD}, multi-lead beats MB={mb1,…,mbN} and single-lead beats SB={sb1,…,sbK} as nodes, where V=ML⋃MB⋃SB. The collection of edges E indicates their correlations and K=D×N.
In detail, for the nodes of multiple leads, mean pooling is performed for the corresponding samples of all single-lead beats from the same lead, which realizes the embedding of every individual lead. For each multi-lead beat node, its representation is obtained by pooling the corresponding samples of all single-lead beats which constitute this multi-lead beat. For a better understanding of the constructed graph, an example is shown in Figure 3. Let us suppose that there are three multi-lead beats in the dataset, i.e., A1, A2 and A3. The whole graph contains 12 leads and each red box marks a single-lead beat. For each single-lead beat node, its feature is represented by its samples. Regarding each multi-lead heartbeat node–with A3 taken as an example, it is represented by pooling all single-lead beats comprising this multi-lead heartbeat. For the nodes of multiple leads, individual leads are represented by pooling the single-lead beats of A1, A2 and A3 in this lead.
So far, the maximum number of multiple leads for ECG signals is typically 12, so the maximum value of D is 12. We assign each multi-lead beat to these D leads. Therefore, the edge between one multi-lead beat and every individual lead has been built to enable analysis of the global significance of individual leads for arrhythmia detection and inter-lead correlations through the use of information propagation along the graph.
The beat segments in every lead are regarded as the nodes of single-lead beats. The embedding of one single-lead beat is the time series of samples and the embedding of one multi-lead beat is the mean pooling of multiple single-lead beats in this multi-lead beat. The edge between one multi-lead beat and one single-lead beat is built if one multi-lead beat contains this single-lead beat, providing information on the correlations between multi-lead beats and their corresponding single-lead beats, as well as the local significance of every single-lead beat for arrhythmia detection.
To further enhance multi-lead information and information propagation, the relationships between single-lead beats are considered. The edge between two single-lead beats is built if they belong to the same lead, which provides information on the inter-lead correlations. By incorporating multiple leads, single-lead beats and the correlations, the information of multi-lead beats is enriched, which greatly facilitates arrhythmia detection.
The HGAT [48] is utilized for heterogeneous graph embedding and then arrhythmia detection, as shown in Figure 4. The HGAT considers the heterogeneity in the information on leads and beats via heterogeneous graph convolution. Furthermore, a dual-layer attention mechanism in the HGAT can capture and learn node-level and type-level importance between adjacent nodes in heterogeneous graphs. Finally, the labels of the multi-lead beats are predicted by a softmax layer.
For heterogeneous graph construction, three types of nodes have been modeled: multiple leads, multi-lead beats and single-lead beats. Though their feature space can be the same and a graph convolutional network can be adopted for heterogeneous graph embedding via the feature space concatenation of these three types of nodes P={p1,p2,p3}. Nevertheless, the performance will be degraded because the heterogeneity of distinct node types is not considered. To solve this problem, we adopt a heterogeneous graph convolutional network, which takes into account the distinction between different types of nodes and projects each of them with respective adjacency and transformation matrices into hidden space, as described by Eq (4.1).
H(l+1)=σ(∑pi∈P˜Api⋅H(l)pi⋅W(l)pi) | (4.1) |
where σ(·) is the activation function. ˜Api∈R|V|×|Vpi| is the submatrix of the normalized adjacency matrix ˜A and is the adjacency matrix of each type pi, whose rows and columns respectively denote all nodes and their adjacent nodes of the type pi. The hidden embeddings of the nodes represented by H(l+1)∈R|V|×q(l+1) are learned by assembling information from their adjacent nodes of H(l)pi∈R|Vpi|×q(l) with different types pi by using the corresponding transformation matrix W(l)pi∈Rq(l)×q(l+1). q(l) is the dimension of hidden embeddings in the lth layer. The transformation matrix W(l)pi considers the distinctiveness of different feature spaces that are projected into the hidden space Rq(l+1).
For a node in our graph, its adjacent nodes may include different types of nodes. In general, the information contained in different kinds of nodes has different contribution degrees. As we know, each lead in the ECG contributes differently to arrhythmia detection. Different weights are beneficial to better highlight important features in the ECG signal data. To capture more effective information about neighboring nodes, a dual-level attention strategy is applied to determine the different contributions at the node level and type level.
Type-level attention is used to capture the importance of different node types, such as multi-lead beats or single-lead beats. Specifically, given a node v, the sum of the adjacent node embeddings hv′ with the type pi is used to express the embedding of type pi as hpi=∑v′∈Nv˜Avv′hv′. ˜Avv′ is one element of the normalized adjacency matrix ˜A, where the row is v and the column is v′. Then, the type-level attention score is computed by using the current node embedding hv and type embedding hpi according to Eq (4.2):
api=σ(μTpi⋅[hv‖hpi]) | (4.2) |
where σ(⋅) is the activation function and ‖ represents "concatenate". μpi is the attention vector for type pi, which is learned in the process of model training.
Finally, the attention score for each type is normalized by Eq (4.3) to obtain the attention weights at the type level.
αpi=exp(api)∑pj∈Pexp(apj) | (4.3) |
Different neighbors also have different contributions. The node-level attention is developed to capture the weights of different neighbour nodes while reducing the importance of noisy nodes. This is a good way to minimize useless information in the ECG signal data. In detail, for a certain node v of the type pi and its neighbor node v′∈Nv of the type pj, the node-level attention score for node v′ is calculated by using hv, hv′ and the type-level attention weight αpj according to Eq (4.4).
bvv′=σ(νTpj⋅αpj[hv‖hv′]) | (4.4) |
where νpj is the node-level attention vector for the type pj, and it is learned in the process of model training. Then, node-level attention score for each neighbour node is normalized by Eq (4.5).
βvv′=exp(bvv′)∑k∈Nvexp(bvk) | (4.5) |
In the end, the dual-level attention strategy with node-level and type-level attentions is integrated into the heterogeneous graph convolutional network by substituting the propagation rule in Eq (4.6) for that in Eq (4.1).
H(l+1)=σ(∑pi∈PBpi⋅H(l)pi⋅W(l)pi) | (4.6) |
where Bpi denotes the attention matrix wherein every element is βvv′ in Eq (4.5).
The embeddings of nodes in the constructed graph are represented by the L-layer HGAT. So, the multi-lead beat nodes are supplemented by neighbor node information. Finally, multi-lead beat embeddings H(L) are input to the softmax layer for arrhythmia classification, as described by Eq (4.7). The cross-entropy loss with the L2-norm is adopted for model training, as described by Eq (4.8).
Z=softmax(H(L)) | (4.7) |
L=−∑i∈DtrainC∑j=1Yij⋅logZij+η‖Θ‖2 | (4.8) |
where C represents the number of categories, Dtrain is the training set of multi-lead beat data, Y is the category matrix, Θ indicates the parameters of this model and η is the regularization coefficient. At last, the gradient descent algorithm is applied for model parameter optimization.
We focus on multi-lead heartbeat classification. The ECG arrhythmia records used in this study are obtained from the St Petersburg INCART 12-lead arrhythmia database (INCART database) [49], which includes 75 annotated recordings collected from 32 patients under investigation for coronary artery disease. Every recording is 30 minutes long sampled at 257 Hz and consists of 12 standard leads, additionally each beat in the recording contains a beat annotation in the middle of QRS complex. Most features of the ECG signals are centered around the QRS complex. If the heartbeat is sampled over a long time interval, more noise information will be included, alternatively, a relatively short interval will lead to insufficient features. Therefore, 150 samples centered about the middle of the QRS complex are taken as the representation of each single-lead beat, which can cover the features of a QRS complex. To verify the proposed model, 1720 heartbeats from the INCART database with seven detailed categories are selected randomly. The statistics of the dataset are shown in Table 1.
Type | Description | Quantity |
N | Normal beat | 500 |
V | Premature ventricular contraction | 500 |
A | Atrial premature beat | 200 |
F | Fusion of ventricular and normal beat | 200 |
n | Supraventricular escape beat | 30 |
R | Right bundle branch block beat | 200 |
j | Nodal (junctional) escape beat | 90 |
Total | 1720 |
Five-fold cross-validation is used for the experimental analysis. The heartbeats are sampled randomly in the dataset described above. In each experiment, one part is used as test data and the other four parts as training data. After five experiments, five confusion matrices are added together to obtain an overall confusion matrix, which is then used to compute the evaluation metrics in accordance with the corresponding formulas, including the overall accuracy (Ov.Acc), accuracy (Acc), sensitivity (Sen), precision (Ppv), specificity (Spe) and F1-score.
Ov.Acc=n∑i=1(TPr)iM | (5.1) |
Acc=TP+TNTP+FP+TN+FN | (5.2) |
Sen=TPTP+FN | (5.3) |
Ppv=TPTP+FP | (5.4) |
Spe=TNFP+TN | (5.5) |
F1−score=2×Ppv×SenPpv+Sen | (5.6) |
where TP, TN, FP and FN denote the true positive, true negative, false positive, and false negative rates, respectively. And, in Eq (5.1), n means the number of arrhythmia categories, (TPr)i means the number of correct predictions in each category, and M means the number of all samples.
Following previous studies, we choose 1, 2, 3 and 12 as respective values of D, which is the number of leads for experimental analysis. The ECG signals from Leads Ⅱ and V1 were found to have clear QRS and T waves, as well as large P waves. P waves can indicate potential changes in the atria and reflect atrial excitation. T waves reflect the depolarization process after ventricular excitation, and they have diagnostic significance. The ECG signals from Lead V5 are distinct with clear QRS waves. The R waves observed for Lead Ⅱ are always larger than the S waves. Lead V1 is dominated by negative waves, and Lead V5 is dominated by the R wave. Information fusion for these three leads can be applied to better detect different contents of QRS wave groups. Thus, the single leads, Lead Ⅱ, V1 and V5, are selected for experimental analysis, and their combination is used for multi-lead ECG classification.
For model training, the hidden dimension of the MADHGAT is set to be 512 and the number of layers is set as 2. The learning rate is set to be 0.003, while the dropout rate is 0.8 and the regularization coefficient η is set to be 5 ×10−8. To avoid overfitting, the L2 regularization method is adopted.
Some baseline methods are described as follows, including the DL-CCANet [50], TL-CCANet [50], RandNet [50], PCANet [51], 1-D ResNet [52] and 1-D CCANet-SVD [52].
1) DL-CCANet is a model for two-lead arrhythmia detection, and TL-CCANet is used for three leads. Their main ideas are the same, and the core lies in the CCANet model, which includes two cascaded convolutional layers for the feature extraction of different leads, as well as an output layer. The difference between TL-CCANet and DL-CCANet is that TL-CCANet has an additional input channel.
2) RandNet is widely used in image learning. It is an unsupervised neural network with compressed inputs for deriving learning representations from compressed data to enable the accommodation of large amounts of data. RandNet was employed to process one-lead ECG signals for heartbeat classification.
3) PCANet uses principal component analysis (PCA), which is a linear subspace projection technique for downsampling high-dimensional datasets and minimizing reprojection errors. By applying PCA filtering for continuous convolution and iteration, the vector representations of the ECG signals were obtained and finally used for heartbeat classification.
4) 1-D ResNet is also widely used in image learning. The model has a deep residual learning framework, which includes residual representations and shortcut connections to solve the degradation problem in a deep learning environment and determine the accuracy gain of arrhythmia detection. Through the use of convolutional layers and residual blocks, two input channels can fuse the information of two leads.
5) 1-D CCANet-SVD is an improved variant of CCANet. For each set of two leads, three types of artificial features including the one-dimensional spectrogram, autoregressive model and time-domain features are applied as the input of CCANet-SVD for neural feature extraction. Then, neural features are fed to an SVM for heartbeat classification.
Single-lead signals are a special case of multi-lead signal. We perform arrhythmia detection for three single-lead signals, i.e., Lead Ⅱ, V1 and V5. Tables 2–4 show their confusion matrices and performance results, which are obtained by applying our model for single-lead experiments. The overall accuracy is determined to be 0.9413, 0.9378 and 0.9378, respectively. Compared with PCANet and RandNet (Table 5), which do not incorporate any handcrafted features, our method improves the overall accuracy for all single leads by about 2% on average. And in the case of Lead V1, we improve the performance by 3% relative to RandNet, which shows that our method takes advantage of single lead data and the heterogeneous graph we constructed works. Since multi-lead beats are composed of their single-lead beats, the relationships between multi-lead beats and their corresponding single-lead beats are useless in the single-lead experiment. However, we still consider the relationship between heartbeats of the same lead as well as the relationship between the lead and the heartbeat, so that the model can still consider the global and local information. In the baseline model, PCANet and RandNet can only consider their own situation in the case of one lead, which reflects the advantages of our model.
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 487 | 4 | 5 | 3 | 1 | 0 | 0 | 0.9645 | 0.9740 | 0.9103 | 0.9607 | 0.9411 |
V | 13 | 477 | 0 | 10 | 0 | 0 | 0 | 0.9715 | 0.9540 | 0.9483 | 0.9787 | 0.9511 |
A | 20 | 1 | 176 | 1 | 0 | 1 | 1 | 0.9814 | 0.8800 | 0.9565 | 0.9947 | 0.9167 |
F | 10 | 17 | 2 | 170 | 0 | 0 | 1 | 0.9738 | 0.8500 | 0.9189 | 0.9901 | 0.8831 |
n | 3 | 0 | 1 | 0 | 26 | 0 | 0 | 0.9971 | 0.8667 | 0.9630 | 0.9994 | 0.9123 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 2 | 0 | 0 | 1 | 0 | 0 | 87 | 0.9971 | 0.9667 | 0.9775 | 0.9988 | 0.9721 |
Average | 0.9832 | 0.9245 | 0.9528 | 0.9888 | 0.9377 | |||||||
Overall accuracy | 0.9413 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 477 | 7 | 4 | 9 | 2 | 0 | 1 | 0.9657 | 0.9540 | 0.9298 | 0.9705 | 0.9418 |
V | 4 | 483 | 2 | 11 | 0 | 0 | 0 | 0.9715 | 0.9660 | 0.9379 | 0.9738 | 0.9517 |
A | 12 | 2 | 183 | 3 | 0 | 0 | 0 | 0.9860 | 0.9150 | 0.9632 | 0.9954 | 0.9385 |
F | 17 | 21 | 1 | 157 | 0 | 2 | 2 | 0.9610 | 0.7850 | 0.8674 | 0.9842 | 0.8241 |
n | 2 | 0 | 0 | 1 | 27 | 0 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 1 | 0 | 0 | 0 | 199 | 0 | 0.9977 | 0.9950 | 0.9851 | 0.9980 | 0.9901 |
j | 1 | 1 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9822 | 0.9260 | 0.9402 | 0.9884 | 0.9326 | |||||||
Overall accuracy | 0.9378 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 488 | 6 | 3 | 1 | 0 | 0 | 2 | 0.9581 | 0.9760 | 0.8905 | 0.9508 | 0.9313 |
V | 11 | 482 | 0 | 6 | 0 | 0 | 1 | 0.9733 | 0.9640 | 0.9451 | 0.9770 | 0.9545 |
A | 25 | 5 | 169 | 1 | 0 | 0 | 0 | 0.9779 | 0.8450 | 0.9602 | 0.9954 | 0.8989 |
F | 15 | 14 | 3 | 168 | 0 | 0 | 0 | 0.9767 | 0.8400 | 0.9545 | 0.9947 | 0.8936 |
n | 2 | 0 | 0 | 0 | 28 | 0 | 0 | 0.9988 | 0.9333 | 1.0000 | 1.0000 | 0.9655 |
R | 0 | 3 | 0 | 0 | 0 | 196 | 1 | 0.9977 | 0.9800 | 1.0000 | 1.0000 | 0.9899 |
j | 7 | 0 | 1 | 0 | 0 | 0 | 82 | 0.9930 | 0.9111 | 0.9535 | 0.9975 | 0.9318 |
Average | 0.9822 | 0.9213 | 0.9577 | 0.9879 | 0.9379 | |||||||
Overall accuracy | 0.9378 |
Method | Year | #Types | #Samples | Ov.ACC |
PCANet (Ⅱ) | 2018 | 7 | 1720 | 93.72% |
PCANet (V1) | 2018 | 7 | 1720 | 92.10% |
PCANet (V5) | 2018 | 7 | 1720 | 93.37% |
RandNet (Ⅱ) | 2019 | 7 | 1720 | 92.67% |
RandNet (V1) | 2019 | 7 | 1720 | 90.76% |
RandNet (V5) | 2019 | 7 | 1720 | 92.91% |
MADHGAT (Ⅱ) | 2022 | 7 | 1720 | 94.13% |
MADHGAT (V1) | 2022 | 7 | 1720 | 93.78% |
MADHGAT (V5) | 2022 | 7 | 1720 | 93.78% |
The experimental results of multi-lead classification will be presented for two-lead, three-lead and 12-lead ECG signal data. Tables 6–8 show the confusion matrices and classification results for the MADHGAT model with two-lead signals. The overall accuracy is found to be 0.9494, 0.9523 and 0.9547, respectively. It is clear that the performance with two-lead signals is better than that with single-lead signals.
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 7 | 2 | 5 | 2 | 0 | 0 | 0.9727 | 0.9680 | 0.9398 | 0.9746 | 0.9537 |
V | 5 | 488 | 0 | 6 | 0 | 0 | 1 | 0.9767 | 0.9760 | 0.9457 | 0.9770 | 0.9606 |
A | 13 | 1 | 181 | 4 | 0 | 1 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 5 | 11 | 1 | 183 | 0 | 0 | 0 | 0.9797 | 0.9150 | 0.9104 | 0.9882 | 0.9127 |
n | 3 | 0 | 0 | 1 | 25 | 1 | 0 | 0.9959 | 0.8333 | 0.9259 | 0.9988 | 0.8772 |
R | 0 | 8 | 0 | 0 | 0 | 192 | 0 | 0.9936 | 0.9600 | 0.9846 | 0.9980 | 0.9722 |
j | 5 | 1 | 1 | 2 | 0 | 1 | 80 | 0.9936 | 0.8889 | 0.9877 | 0.9994 | 0.9357 |
Average | 0.9855 | 0.9209 | 0.9532 | 0.9905 | 0.9360 | |||||||
Overall accuracy | 0.9494 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 8 | 1 | 6 | 1 | 0 | 0 | 0.9698 | 0.9680 | 0.9308 | 0.9705 | 0.9490 |
V | 5 | 486 | 1 | 7 | 0 | 1 | 0 | 0.9785 | 0.9720 | 0.9548 | 0.9811 | 0.9633 |
A | 14 | 2 | 181 | 3 | 0 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 11 | 9 | 2 | 177 | 0 | 0 | 1 | 0.9773 | 0.8850 | 0.9171 | 0.9895 | 0.9008 |
n | 3 | 0 | 0 | 0 | 27 | 0 | 0 | 0.9977 | 0.9000 | 0.9643 | 0.9994 | 0.9310 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 3 | 0 | 0 | 0 | 0 | 0 | 87 | 0.9977 | 0.9667 | 0.9886 | 0.9994 | 0.9775 |
Average | 0.9864 | 0.9395 | 0.9613 | 0.9909 | 0.9499 | |||||||
Overall accuracy | 0.9523 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9750 | 0.9780 | 0.9386 | 0.9738 | 0.9579 |
V | 4 | 489 | 0 | 7 | 0 | 0 | 0 | 0.9785 | 0.9780 | 0.9495 | 0.9787 | 0.9635 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
R | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
j | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
n | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9870 | 0.9381 | 0.9595 | 0.9914 | 0.9483 | |||||||
Overall accuracy | 0.9547 |
Table 9 shows the confusion matrix and classification results for the MADHGAT model with three-lead signals. The overall accuracy is 0.9552. It can be noticed that the three-lead results are better than the two-lead results. However, we can find that the three-lead experiment yields similar results as the experiment with Leads V1 and V5. In this regard, we think that the key reason is limited sampled heartbeats, which has led slightly different experimental results. Table 10 shows the classification results for the MADHGAT model with 12-lead signals. The overall accuracy is 0.9721. It outperforms the three-lead experiment by almost 2%.
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9756 | 0.9780 | 0.9404 | 0.9746 | 0.9588 |
V | 3 | 490 | 0 | 7 | 0 | 0 | 0 | 0.9791 | 0.9800 | 0.9496 | 0.9787 | 0.9646 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
n | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
j | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 | |||||||
Overall accuracy | 0.9552 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 2 | 9 | 4 | 0 | 0 | 1 | 0.9820 | 0.9680 | 0.9699 | 0.9877 | 0.9690 |
V | 1 | 496 | 0 | 3 | 0 | 0 | 0 | 0.9907 | 0.9920 | 0.9764 | 0.9902 | 0.9841 |
A | 8 | 0 | 191 | 0 | 0 | 0 | 1 | 0.9890 | 0.9550 | 0.9502 | 0.9934 | 0.9526 |
F | 3 | 6 | 1 | 190 | 0 | 0 | 0 | 0.9901 | 0.9500 | 0.9645 | 0.9954 | 0.9572 |
n | 1 | 0 | 0 | 0 | 29 | 0 | 0 | 0.9994 | 0.9667 | 1.0000 | 1.0000 | 0.9831 |
R | 0 | 4 | 0 | 0 | 0 | 195 | 1 | 0.9965 | 0.9750 | 0.9949 | 0.9993 | 0.9848 |
j | 2 | 0 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9920 | 0.9676 | 0.9747 | 0.9949 | 0.9711 | |||||||
Overall accuracy | 0.9721 |
Figure 5 integrates our results above, and it can be clearly seen that as we apply more leads, the classification results gradually improved. This is because in the case of multiple leads, the heterogeneous graph we constructed can utilize the associations of leads so that the MADHGAT model can obtain more information on the multi-lead ECG signals to strengthen its learning ability and more accurately assign a multi-lead heartbeat to a certain category. It is proved that our model can consider the correlations between different leads and improve the performance of arrhythmia diagnosis.
Table 11 shows the results of comparing several models with multiple leads. DL-CCANet, TL-CCANet and MADHGAT are end-to-end models without any handcrafted features. The 1-D CCANet-SVD and 1-D ResNet models need handcrafted feature extraction. Our model with 12 leads achieves better performance than the others. Compared with DL-CCANet on the same leads, all three of our two-lead combinations outperform its results by approximately 1.3% on average. In the two-lead Ⅱ and V1, V1 and V5 combinations, the improvement is approximately 1.2% and 1.6%, respectively. Although DL-CCANet considers information fusion for two leads, it joins and reconstructs the signals of two leads separately to obtain features, which are then fed into CCANet to extract the fused features. However, it not only ignores the correlations between single-lead beats for the same lead, but also does not consider the associations between the overall leads and two-lead beats. It shows that our method can better utilize intra-lead and inter-lead correlations.
Method | Year | #Types | #Samples | Ov.ACC |
DL-CCANet (Ⅱ and V1) | 2019 | 7 | 1720 | 94.01% |
DL-CCANet (Ⅱ and V5) | 2019 | 7 | 1720 | 94.07% |
DL-CCANet (V1 and V5) | 2019 | 7 | 1720 | 93.90% |
TL-CCANet (Ⅱ, V1 and V5) | 2019 | 7 | 1720 | 95.52% |
1-D ResNet (Ⅱ and V1) | 2021 | 7 | 1720 | 86.25% |
1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 2021 | 7 | 1720 | 93.60% |
1-D CCANet-SVD (w/o ar) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.12% |
1-D CCANet-SVD (w/o 1D-spec) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.77% |
1-D CCANet-SVD (w/o time-domain feature) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.35% |
1-D CCANet-SVD (w/o stack) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.83% |
1-D CCANet-SVD (Ⅱ and V1) | 2021 | 7 | 1720 | 95.70% |
MADHGAT (Ⅱ and V1) | 2022 | 7 | 1720 | 94.94% |
MADHGAT (Ⅱ and V5) | 2022 | 7 | 1720 | 95.23% |
MADHGAT (V1 and V5) | 2022 | 7 | 1720 | 95.47% |
MADHGAT (Ⅱ, V1 and V5) | 2022 | 7 | 1720 | 95.52% |
MADHGAT (12 leads) | 2022 | 7 | 1720 | 97.21% |
At the same time, the results for Leads Ⅱ and V1 indicate that our model still outperforms several variants of 1-D CCANet-SVD which need some artificial features. For example, 1-D CCANet-SVD without singular value decomposition (SVD) adds time-domain features, autoregressive modeling and a one-dimensional spectrogram, but lacks SVD assistance. Nonetheless, we can still achieve better arrhythmia diagnosis by obtaining information from the leads. In the case of 1-D CCANet-SVD, one-dimensional spectrogram features capture the frequency representation of the signal over time. Autoregressive modeling is applied to capture features that can fit their own time series. Time-domain features take into account the information from the previous heartbeat signal from the same lead, as well as include higher-order centroids and kurtosis. These manual features enable the capture of information about the previous heartbeat, the global frequency domain, etc. All of these features only consider each lead itself. Finally, the CCANet model was utilized to fuse the information of two leads for classification. However, this method aims to mine the content of each lead and heartbeat, missing the associations between different heartbeats of the same lead, as well as the associations between multi-lead heartbeats and single-lead heartbeats. Although 1-D CCANet-SVD achieved 95.70% with all of the handcrafted features, our overall accuracy of 97.21% with 12 leads, which is achieved without any feature extraction, is still higher by 1.51%. We have greatly improved the performance by taking into account inter-lead and intra-lead correlations. The performance of arrhythmia detection can still be excellent without cumbersome handcrafted feature extraction.
Overall, the MADHGAT model achieves the best results among all of the models for all different numbers of leads. Meanwhile, it is obvious that multiple leads are important for arrhythmia detection, as overall detection performance is improved significantly with more leads. This indicates the effectiveness of the proposed MADHGAT model to capture inter-lead and intra-lead information.
For the constructed graph, we have defined three kinds of nodes. In this experiment, we decide to consider the importance of two kinds of nodes: multiple leads and single-lead beats. Three leads, i.e., Leads Ⅱ, V1 and V5 are applied in this experiment with seven categories. We eliminate the nodes of multiple leads and single-lead beats individually and in combination. The results in Table 12 indicate that the full graph with three types of nodes performs better than the graphs with missing nodes, proving the effectiveness of our graph structure.
Graph | Ov.ACC | Avg-Acc | Avg-Sen | Avg-Ppv | Avg-Spe | Avg-F1 |
without multiple leads | 0.9384 | 0.9824 | 0.9152 | 0.9345 | 0.9884 | 0.9243 |
without single-lead beats | 0.8273 | 0.9507 | 0.7660 | 0.8507 | 0.9670 | 0.7975 |
only multi-lead beats | 0.8070 | 0.9449 | 0.7070 | 0.8542 | 0.9623 | 0.7568 |
full graph | 0.9552 | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 |
We can find that the result of removing single-lead beat nodes is a deterioration for every metric relative to the full graph. The most obvious difference is in overall accuracy, which differs by about 13%. This reflects the importance of the relationship between single-lead beats and multi-lead beats, as well as the intra-lead correlations between single-lead beats, which are used to allow the MADHGAT model to capture the contribution of each single-lead beat.
When removing the nodes of multiple leads, relative to the results for the full graph, the overall accuracy is reduced by about 2%. At the same time, it is found that other indicators on average also decrease accordingly. The nodes of multiple leads consider the global information for all heartbeats from a certain lead and transmit the information to the multi-lead heartbeat nodes. Meanwhile, the inter-lead correlations between multiple leads are constructed, improving the performance of arrhythmia detection.
Finally, we remove the nodes of multiple leads and single-lead beats. Only multi-lead beat nodes remain in the heartbeat graph. The final result is the worst because it does not take into account any additional information. Compared to the other cases, it also proves the contributions of the inter-lead and intra-lead correlations.
The proposed MADHGAT model is based on an HGAT, which is transductive and therefore observes all of the data beforehand for classification. To evaluate the performance of MADHGAT with different training set proportions, experiments are conducted using Leads Ⅱ and V1 with different ratios of training data to test data, i.e., 1:2, 1:1, 2:1 and 3:1. The results have been compared with the five-fold results, as shown in Table 13. The experimental results of some baseline methods are also listed.
Ratio of training set to test set | Method | Ov.ACC |
4:1 | DL-CCANet (Ⅱ and V1) | 94.01% |
4:1 | 1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 93.60% |
4:1 | 1-D ResNet (Ⅱ and V1) | 86.25% |
4:1 | MADHGAT | 94.94% |
3:1 | MADHGAT | 94.36% |
2:1 | MADHGAT | 94.48% |
1:1 | MADHGAT | 92.38% |
1:2 | MADHGAT | 90.20% |
When the ratio of training set to test set is 1:2, our model is still able to achieve an overall accuracy of 90.20%, which is still better than the ResNet method. Meanwhile, when the ratio is greater than or equal to 2:1, the overall accuracy rate is above 94%. We find that after reducing the number of training sets, our results are still better than some methods. For example, in the case of a 2:1 ratio, our model achieves 94.48% which is higher than that for several methods listed in Table 13. Moreover, 1-D CCANet-SVD needs handcrafted features. Therefore, it is not difficult to conclude that MADHGAT can still achieve better results when the ratio of training data to test data is decreased. And it shows that our proposed MADHGAT framework can capture effective information, making the model more robust and generalizable.
A novel multi-lead arrhythmia detection model based on an HGAT has been proposed in this paper. Multi-lead ECG signals are modeled as a heterogeneous graph to integrate rich information. Additionally, intra-lead and inter-lead correlations are constructed to capture intricate interactions among multi-lead signals. This work provides a reasonable and effective data model that utilizes a heterogeneous graph neural network with dual-level attention strategy to quantify the contributions of different types of information, thereby improving detection performance. At the same time, our model does not require any feature extraction process for ECG signal analysis, which avoids complex feature engineering.
Though the proposed model performs excellently for automatic arrhythmia detection, it is transductive. An extremely massive amount of memory is needed for the training process. In light of this problem, in the future, we will focus on developing an inductive learning model based on a graph neural network that can handle more data.
This work is supported by the Natural Science Foundation of Guangdong Province (No. 2021A1515012290), Guangdong Provincial Key Laboratory of Cyber-Physical Systems (No. 2020B1212060069) and National & Local Joint Engineering Research Center of Intelligent Manufacturing Cyber-Physical Systems.
The authors declare that there is no conflict of interest.
[1] | Chauhan BS, Jabran K, Mahajan G (2017) Rice production worldwide, Springer. |
[2] |
Hu EA, Pan A, Malik V, et al. (2012) White rice consumption and risk of type 2 diabetes: Meta-analysis and systematic review. Bmj 344: e1454. doi: 10.1136/bmj.e1454
![]() |
[3] |
Seck PA, Diagne A, Mohanty S, et al. (2012) Crops that feed the world 7: Rice. Food Secur 4: 7-24. doi: 10.1007/s12571-012-0168-1
![]() |
[4] |
Fuller DQ, Sato YI, Castillo C, et al. (2010) Consilience of genetics and archaeobotany in the entangled history of rice. Archaeol Anthropol Sci 2: 115-131. doi: 10.1007/s12520-010-0035-y
![]() |
[5] | Food and Agriculture Organization of the United Nations (2018) Rice market monitor 2018, Rice Market Monitor XXI. |
[6] | Tan BL, Norhaizan ME (2020) Rice demands: A brief description, Rice by-products: Phytochemicals and food products application, Springer, 7-11. |
[7] |
Peng S, Tang Q, Zou Y (2009) Current status and challenges of rice production in China. Plant Prod Sci 12: 3-8. doi: 10.1626/pps.12.3
![]() |
[8] |
Van Nguyen N, Ferrero A (2006) Meeting the challenges of global rice production. Paddy Water Environ 4: 1-9. doi: 10.1007/s10333-005-0031-5
![]() |
[9] | Barker R, Dawe D, Tuong T, et al. (2020) The outlook for water resources in the year 2020: Challenges for research on water management in rice production. Southeast Asia 1: 1-5. |
[10] | Mo C, Lim J, Kwon SW, et al. (2017) Hyperspectral imaging and partial least square discriminant analysis for geographical origin discrimination of white rice. J Biosyst Eng 42: 293-300. |
[11] |
Tong Q, Swallow B, Zhang L, et al. (2019) The roles of risk aversion and climate-smart agriculture in climate risk management: Evidence from rice production in the Jianghan Plain, China. Clim Risk Manage 26: 100199. doi: 10.1016/j.crm.2019.100199
![]() |
[12] |
Balasubramanian V, Sie M, Hijmans R, et al. (2007) Increasing rice production in sub-Saharan Africa: Challenges and opportunities. Adv Agron 94: 55-133. doi: 10.1016/S0065-2113(06)94002-4
![]() |
[13] |
Chen J, Zou W, Meng L, et al. (2019) Advances in the uptake and transport mechanisms and QTLs mapping of cadmium in rice. Int J Mol Sci 20: 3417. doi: 10.3390/ijms20143417
![]() |
[14] |
Jin XB, Yu XH, Wang XY, et al. (2020) Deep learning predictor for sustainable precision agriculture based on internet of things system. Sustainability 12: 1433. doi: 10.3390/su12041433
![]() |
[15] | Salam A, Shah S (2019) Internet of things in smart agriculture: Enabling technologies, 2019 IEEE 5th world forum on internet of things (WF-IoT), 692-695. |
[16] |
Kundu M, Krishnan P, Kotnala R, et al. (2019) Recent developments in biosensors to combat agricultural challenges and their future prospects. Trends Food Sci Technol 88: 157-178. doi: 10.1016/j.tifs.2019.03.024
![]() |
[17] |
Thakur D, Kumar Y, Kumar A, et al. (2019) Applicability of wireless sensor networks in precision agriculture: A review. Wireless Pers Commun 107: 471-512. doi: 10.1007/s11277-019-06285-2
![]() |
[18] | International Telecommunications Union and XPRIZE Foundation (2017) AI for good global summit report, Geneva, Switzerland, United Nations. |
[19] | Bejaoui C, Guitton A, Kachouri A (2017) Efficient data monitoring of rice-fields based on WMSN, 2017 IEEE/ACS 14th international conference on computer systems and applications (AICCSA), 1090-1094. |
[20] |
Chen WL, Lin YB, Ng FL, et al. (2019) RiceTalk: Rice blast detection using internet of things and artificial intelligence technologies. IEEE Internet Things J 7: 1001-1010. doi: 10.1109/JIOT.2019.2947624
![]() |
[21] |
Gao Z, Li W, Zhu Y, et al. (2018) Wireless channel propagation characteristics and modeling research in rice field sensor networks. Sensors 18: 3116. doi: 10.3390/s18093116
![]() |
[22] |
Sabet FS, Hosseini M, Khabbaz H, et al. (2017) FRET-based aptamer biosensor for selective and sensitive detection of aflatoxin B1 in peanut and rice. Food Chem 220: 527-532. doi: 10.1016/j.foodchem.2016.10.004
![]() |
[23] |
Hung T, Byun W (2006) Assessment of rice leaf growth and nitrogen status by hyperspectral canopy reflectance and partial least square regress. Eur J Agron 24: 349-356. doi: 10.1016/j.eja.2006.01.001
![]() |
[24] |
Intaravanne Y, Sumriddetchkajorn S (2015) Android-based rice leaf color analyzer for estimating the needed amount of nitrogen fertilizer. Comput Electron Agric 116: 228-233. doi: 10.1016/j.compag.2015.07.005
![]() |
[25] |
Wang L, Liu D, Pu H, et al. (2015) Use of hyperspectral imaging to discriminate the variety and quality of rice. Food Anal Methods 8: 515-523. doi: 10.1007/s12161-014-9916-5
![]() |
[26] | Caballero D, Calvini R, Amigo JM (2020) Hyperspectral imaging in crop fields: Precision agriculture, Data handling in science and technology, Elsevier, 453-473. |
[27] |
Lu B, Dao PD, Liu J, et al. (2020) Recent advances of hyperspectral imaging technology and applications in agriculture. Remote Sens 12: 2659. doi: 10.3390/rs12162659
![]() |
[28] |
Foster DH, Amano K (2019) Hyperspectral imaging in color vision research: tutorial. JOSA A 36: 606-627. doi: 10.1364/JOSAA.36.000606
![]() |
[29] |
Khan MJ, Khan HS, Yousaf A, et al. (2018) Modern trends in hyperspectral image analysis: A review. IEEE Access 6: 14118-14129. doi: 10.1109/ACCESS.2018.2812999
![]() |
[30] |
Paoletti M, Haut J, Plaza J, et al. (2019) Deep learning classifiers for hyperspectral imaging: A review. ISPRS J Photogramm Remote Sens 158: 279-317. doi: 10.1016/j.isprsjprs.2019.09.006
![]() |
[31] |
Zhang K, Liu X, Ma Y, et al. (2020) A comparative assessment of measures of leaf nitrogen in rice using two leaf-clip meters. Sensors 20: 175. doi: 10.3390/s20010175
![]() |
[32] |
Zhang M, Lin H, Wang G, et al. (2018) Mapping paddy rice using a convolutional neural network (CNN) with Landsat 8 datasets in the Dongting Lake Area, China. Remote Sens 10: 1840. doi: 10.3390/rs10111840
![]() |
[33] |
Zheng H, Zhou X, He J, et al. (2020) Early season detection of rice plants using RGB, NIR-GB and multispectral images from unmanned aerial vehicle (UAV). Comput Electron Agric 169: 105223. doi: 10.1016/j.compag.2020.105223
![]() |
[34] |
Baek I, Kim MS, Cho BK, et al. (2019) Selection of optimal hyperspectral wavebands for detection of discolored, diseased rice seeds. Appl Sci 9: 1027. doi: 10.3390/app9051027
![]() |
[35] |
Fan Y, Wang T, Qiu Z, et al. (2017) Fast detection of striped stem-borer (Chilo suppressalis Walker) infested rice seedling based on visible/near-infrared hyperspectral imaging system. Sensors 17: 2470. doi: 10.3390/s17112470
![]() |
[36] |
Huang J, Liao H, Zhu Y, et al. (2012) Hyperspectral detection of rice damaged by rice leaf folder (Cnaphalocrocis medinalis). Comput Electron Agric 82: 100-107. doi: 10.1016/j.compag.2012.01.002
![]() |
[37] |
Schneider P, Asch F (2020) Rice production and food security in Asian Mega deltas-A review on characteristics, vulnerabilities and agricultural adaptation options to cope with climate change. J Agron Crop Sci 206: 491-503. doi: 10.1111/jac.12415
![]() |
[38] |
Gebbers R, Adamchuk VI (2010) Precision agriculture and food security. Science 327: 828-831. doi: 10.1126/science.1183899
![]() |
[39] |
Ehleringer JR, Cerling TE, Helliker BR (1997) C 4 photosynthesis, atmospheric CO2, and climate. Oecologia 112: 285-299. doi: 10.1007/s004420050311
![]() |
[40] |
Sonnino A (1994) Agricultural biomass production is an energy option for the future. Renewable Energy 5: 857-865. doi: 10.1016/0960-1481(94)90105-8
![]() |
[41] |
Lichtenthaler HK, Rinderle U (1988) The role of chlorophyll fluorescence in the detection of stress conditions in plants. CRC Crit Rev Anal Chem 19: S29-S85. doi: 10.1080/15476510.1988.10401466
![]() |
[42] | Mafakheri A, Siosemardeh A, Bahramnejad B, et al. (2010) Effect of drought stress on yield, proline and chlorophyll contents in three chickpea cultivars. Aust J Crop Sci 4: 580. |
[43] |
Bannister T (1974) Production equations in terms of chlorophyll concentration, quantum yield, and upper limit to production. Limnol Oceanogr 19: 1-12. doi: 10.4319/lo.1974.19.1.0001
![]() |
[44] | Wood C, Reeves D, Himelrick D (1993) Relationships between chlorophyll meter readings and leaf chlorophyll concentration, N status, and crop yield: A review, Proceedings of the agronomy society of New Zealand, 1-9. |
[45] |
Makino A (2011) Photosynthesis, grain yield, and nitrogen utilization in rice and wheat. Plant Physiol 155: 125-129. doi: 10.1104/pp.110.165076
![]() |
[46] |
Zhou K, Cheng T, Zhu Y, et al. (2018) Assessing the impact of spatial resolution on the estimation of leaf nitrogen concentration over the full season of paddy rice using near-surface imaging spectroscopy data. Frontiers in plant science 9: 964. doi: 10.3389/fpls.2018.00964
![]() |
[47] | Matsushima S (1993) Researches on the requirements for achieving high yields in rice. Sci Rice Plant 2: 737-747. |
[48] |
Inoue Y, Sakaiya E, Zhu Y, et al. (2012) Diagnostic mapping of canopy nitrogen content in rice based on hyperspectral measurements. Remote Sens Environ 126: 210-221. doi: 10.1016/j.rse.2012.08.026
![]() |
[49] |
Lee YJ, Yang CM, Chang KW, et al. (2008) A simple spectral index using reflectance of 735 nm to assess nitrogen status of rice canopy. Agron J 100: 205-212. doi: 10.2134/agronj2007.0018
![]() |
[50] |
Sun J, Lu X, Mao H, et al. (2017) A method for rapid identification of rice origin by hyperspectral imaging technology. J Food Process Eng 40: e12297. doi: 10.1111/jfpe.12297
![]() |
[51] |
Raines CA (2011) Increasing photosynthetic carbon assimilation in C3 plants to improve crop yield: Current and future strategies. Plant Physiol 155: 36-42. doi: 10.1104/pp.110.168559
![]() |
[52] |
Nowak B, Nesme T, David C, et al. (2015) Nutrient recycling in organic farming is related to diversity in farm types at the local level. Agric, Ecosyst Environ 204: 17-26. doi: 10.1016/j.agee.2015.02.010
![]() |
[53] |
Song YQ, Zhao X, Su HY, et al. (2018) Predicting spatial variations in soil nutrients with hyperspectral remote sensing at regional scale. Sensors 18: 3086. doi: 10.3390/s18093086
![]() |
[54] |
Luo JS, Huang J, Zeng DL, et al. (2018) A defensin-like protein drives cadmium efflux and allocation in rice. Nat Commun 9: 1-9. doi: 10.1038/s41467-017-02088-w
![]() |
[55] |
Muehe EM, Wang T, Kerl CF, et al. (2019) Rice production threatened by coupled stresses of climate and soil arsenic. Nat Commun 10: 1-10. doi: 10.1038/s41467-019-12946-4
![]() |
[56] | Sharma RK, Agrawal M (2005) Biological effects of heavy metals: An overview. J Environ Biol 26: 301-313. |
[57] |
Xie P, Deng J, Zhang H, et al. (2015) Effects of cadmium on bioaccumulation and biochemical stress response in rice (Oryza sativa L.). Ecotoxicol Environ Saf 122: 392-398. doi: 10.1016/j.ecoenv.2015.09.007
![]() |
[58] |
Xue D, Jiang H, Deng X, et al. (2014) Comparative proteomic analysis provides new insights into cadmium accumulation in rice grain under cadmium stress. J Hazard Mater 280: 269-278. doi: 10.1016/j.jhazmat.2014.08.010
![]() |
[59] |
Hanak E, Lund JR (2012) Adapting California's water management to climate change. Clim Change 111: 17-44. doi: 10.1007/s10584-011-0241-3
![]() |
[60] |
Iglesias A, Garrote L (2015) Adaptation strategies for agricultural water management under climate change in Europe. Agric Water Manage 155: 113-124. doi: 10.1016/j.agwat.2015.03.014
![]() |
[61] | Ahangarha M, Seydi ST, Shahhoseini R (2019) Hyperspectral change detection in wetland and water-body areas based on machine learning. Int Arch Photogramm, Remote Sens Spat Inf Sci XLⅡ-4/W18: 1-19. |
[62] | León CD, Kosow H, Zahumensky Y, et al. (2019) Solutions and planning tools for water supply and wastewater management in prosperous regions tackling water scarcity, Mid-term Conference-Frankfurt am main, Germany 20-21 February 2019, 28. |
[63] |
Liew S, Choo C, Lau J, et al. (2019) Monitoring water quality in Singapore reservoirs with hyperspectral remote sensing technology. Water Pract Technol 14: 118-125. doi: 10.2166/wpt.2018.119
![]() |
[64] |
Prošek J, Gdulová K, Barták V, et al. (2020) Integration of hyperspectral and LiDAR data for mapping small water bodies. Int J Appl Earth Obs Geoinformation 92: 102181. doi: 10.1016/j.jag.2020.102181
![]() |
[65] |
Ge X, Wang J, Ding J, et al. (2019) Combining UAV-based hyperspectral imagery and machine learning algorithms for soil moisture content monitoring. Peer J 7: e6926. doi: 10.7717/peerj.6926
![]() |
[66] |
Huo H, Li Z-L, Xing Z (2019) Temperature/emissivity separation using hyperspectral thermal infrared imagery and its potential for detecting the water content of plants. Int J Remote Sens 40: 1672-1692. doi: 10.1080/01431161.2018.1513668
![]() |
[67] | Hurley SP, Horney M, Drake A (2019) Using hyperspectral imagery to detect water stress in vineyards, Autonomous air and ground sensing systems for agricultural optimization and phenotyping IV, 1100807. |
[68] |
Krishna G, Sahoo RN, Singh P, et al. (2019) Comparison of various modelling approaches for water deficit stress monitoring in rice crop through hyperspectral remote sensing. Agric Water Manage 213: 231-244. doi: 10.1016/j.agwat.2018.08.029
![]() |
[69] |
Zhang Z, Lou Y, Li R, et al. (2019) Estimation of relative water content in rice panicle based on hyperspectral vegetation indexes under water saving irrigation. Spectrosc Lett 52: 150-158. doi: 10.1080/00387010.2019.1594309
![]() |
[70] | Manolakis DG, Lockwood RB, Cooley TW (2016) Hyperspectral imaging remote sensing: Physics, sensors, and algorithms, Cambridge University Press. |
[71] |
Ahmed MR, Yasmin J, Mo C, et al. (2016) Outdoor applications of hyperspectral imaging technology for monitoring agricultural crops: A review. J Biosyst Eng 41: 396-407. doi: 10.5307/JBE.2016.41.4.396
![]() |
[72] | Fei B (2020) Hyperspectral imaging in medical applications, Data handling in science and technology, Elsevier, 523-565. |
[73] |
Ma J, Sun DW, Pu H, et al. (2019) Advanced techniques for hyperspectral imaging in the food industry: principles and recent applications. Annu Rev Food Sci Technol 10: 197-220. doi: 10.1146/annurev-food-032818-121155
![]() |
[74] |
Zheng H, Cheng T, Li D, et al. (2018) Evaluation of RGB, color-infrared and multispectral images acquired from unmanned aerial systems for the estimation of nitrogen accumulation in rice. Remote Sens 10: 824. doi: 10.3390/rs10060824
![]() |
[75] |
Mahajan G, Pandey R, Sahoo R, et al. (2017) Monitoring nitrogen, phosphorus and sulphur in hybrid rice (Oryza sativa L.) using hyperspectral remote sensing. Precis Agric 18: 736-761. doi: 10.1007/s11119-016-9485-2
![]() |
[76] |
Zhu Y, Tian Y, Yao X, et al. (2007) Analysis of common canopy reflectance spectra for indicating leaf nitrogen concentrations in wheat and rice. Plant Prod Sci 10: 400-411. doi: 10.1626/pps.10.400
![]() |
[77] |
Behmann J, Steinrücken J, Plümer L (2014) Detection of early plant stress responses in hyperspectral images. ISPRS J Photogramm Remote Sens 93: 98-111. doi: 10.1016/j.isprsjprs.2014.03.016
![]() |
[78] |
Römer C, Wahabzada M, Ballvora A, et al. (2012) Early drought stress detection in cereals: Simplex volume maximisation for hyperspectral image analysis. Funct Plant Biol 39: 878-890. doi: 10.1071/FP12060
![]() |
[79] |
Pearlman JS, Barry PS, Segal CC, et al. (2003) Hyperion, a space-based imaging spectrometer. IEEE Trans Geosci Remote Sens 41: 1160-1173. doi: 10.1109/TGRS.2003.815018
![]() |
[80] |
Green RO, Eastwood ML, Sarture CM, et al. (1998) Imaging spectroscopy and the airborne visible/infrared imaging spectrometer (AVIRIS). Remote Sens Environ 65: 227-248. doi: 10.1016/S0034-4257(98)00064-9
![]() |
[81] |
Banerjee BP, Raval S, Cullen P (2020) UAV-hyperspectral imaging of spectrally complex environments. Int J Remote Sens 41: 4136-4159. doi: 10.1080/01431161.2020.1714771
![]() |
[82] |
Wiegmann M, Backhaus A, Seiffert U, et al. (2019) Optimizing the procedure of grain nutrient predictions in barley via hyperspectral imaging. PloS One 14: e0224491. doi: 10.1371/journal.pone.0224491
![]() |
[83] |
Smith ML, Ollinger SV, Martin ME, et al. (2002) Direct estimation of aboveground forest productivity through hyperspectral remote sensing of canopy nitrogen. Ecol Appl 12: 1286-1302. doi: 10.1890/1051-0761(2002)012[1286:DEOAFP]2.0.CO;2
![]() |
[84] |
Yao L, Wang Q, Yang J, et al. (2019) UAV-Borne dual-band sensor method for monitoring physiological crop status. Sensors 19: 816. doi: 10.3390/s19040816
![]() |
[85] |
Fan L, Zhao J, Xu X, et al. (2019) Hyperspectral-based estimation of leaf nitrogen content in corn using optimal selection of multiple spectral variables. Sensors 19: 2898. doi: 10.3390/s19132898
![]() |
[86] |
Hruska Z, Yao H, Kincaid R, et al. (2020) Spectral-based screening approach evaluating two specific maize lines with divergent resistance to invasion by aflatoxigenic fungi. Front Microbiol 10: 3152. doi: 10.3389/fmicb.2019.03152
![]() |
[87] |
Biswas DK, Coulman B, Biligetu B, et al. (2019) Advancing bromegrass breeding through imaging phenotyping and genomic selection: A review. Front Plant Sci 10: 1673. doi: 10.3389/fpls.2019.01673
![]() |
[88] |
Feng X, Zhan Y, Wang Q, et al. (2020) Hyperspectral imaging combined with machine learning as a tool to obtain high-throughput plant salt-stress phenotyping. Plant J 101: 1448-1461. doi: 10.1111/tpj.14597
![]() |
[89] |
Liu B, Li R, Li H, et al. (2019) Crop/Weed discrimination using a field imaging spectrometer system. Sensors 19: 5154. doi: 10.3390/s19235154
![]() |
[90] |
Wang D, Vinson R, Holmes M, et al. (2019) Early Detection of tomato spotted wilt virus by hyperspectral imaging and outlier removal auxiliary classifier generative adversarial nets (OR-AC-GAN). Sci Rep 9: 1-14. doi: 10.1038/s41598-018-37186-2
![]() |
[91] |
Nie P, Zhang J, Feng X, et al. (2019) Classification of hybrid seeds using near-infrared hyperspectral imaging technology combined with deep learning. Sens Actuators B: Chem 296: 126630. doi: 10.1016/j.snb.2019.126630
![]() |
[92] |
Bangelesa F, Adam E, Knight J, et al. (2020) Predicting soil organic carbon content using hyperspectral remote sensing in a degraded mountain landscape in lesotho. Appl Environ Soil Sci 2020: 2158573. doi: 10.1155/2020/2158573
![]() |
[93] |
Chen Y, Wang J, Liu G, et al. (2019) Hyperspectral estimation model of forest soil organic matter in northwest Yunnan Province, China. Forests 10: 217. doi: 10.3390/f10030217
![]() |
[94] |
Guo L, Zhang H, Shi T, et al. (2019) Prediction of soil organic carbon stock by laboratory spectral data and airborne hyperspectral images. Geoderma 337: 32-41. doi: 10.1016/j.geoderma.2018.09.003
![]() |
[95] |
Shen Q, Xia K, Zhang S, et al. (2019) Hyperspectral indirect inversion of heavy-metal copper in reclaimed soil of iron ore area. Spectrochim Acta Part A 222: 117191. doi: 10.1016/j.saa.2019.117191
![]() |
[96] |
Zhang S, Shen Q, Nie C, et al. (2019) Hyperspectral inversion of heavy metal content in reclaimed soil from a mining wasteland based on different spectral transformation and modeling methods. Spectrochim Acta Part A 211: 393-400. doi: 10.1016/j.saa.2018.12.032
![]() |
[97] |
Hu J, Peng J, Zhou Y, et al. (2019) Quantitative estimation of soil salinity using UAV-borne hyperspectral and satellite multispectral images. Remote Sens 11: 736. doi: 10.3390/rs11070736
![]() |
[98] |
Grafton M, Kaul T, Palmer A, et al. (2019) Regression analysis of proximal hyperspectral data to predict soil pH and Olsen P. Agriculture 9: 55. doi: 10.3390/agriculture9030055
![]() |
[99] | Patel AK, Ghosh JK (2019) Soil fertility status assessment using hyperspectral remote sensing, Remote sensing for agriculture, Ecosystems, and Hydrology XXI, 111490E. |
[100] |
Gold KM, Townsend PA, Chlus A, et al. (2020) Hyperspectral measurements enable pre-symptomatic detection and differentiation of contrasting physiological effects of late blight and early blight in potato. Remote Sens 12: 286. doi: 10.3390/rs12020286
![]() |
[101] | Gorretta N, Nouri M, Herrero A, et al. (2019) Early detection of the fungal disease "apple scab" using SWIR hyperspectral imaging, 2019 10th workshop on hyperspectral imaging and signal processing: Evolution in remote sensing (WHISPERS), 1-4. |
[102] |
Mahlein AK, Alisaac E, Al Masri A, et al. (2019) Comparison and combination of thermal, fluorescence, and hyperspectral imaging for monitoring Fusarium head blight of wheat on spikelet scale. Sensors 19: 2281. doi: 10.3390/s19102281
![]() |
[103] | Yan Y, Yu W (2019) Early detection of rice blast (pyricularia) at seedling stage based on near-infrared hyper-spectral image, Proceedings of the 2019 8th international conference on bioinformatics and biomedical science, 64-68. |
[104] | Gustavsson J, Cederberg C, Sonesson U, et al. (2011) Global food losses and food waste. |
[105] |
Parfitt J, Barthel M, Macnaughton S (2010) Food waste within food supply chains: Quantification and potential for change to 2050. Philos Trans R Soc, B 365: 3065-3081. doi: 10.1098/rstb.2010.0126
![]() |
[106] |
Liu S, Liu X, Liu M, et al. (2017) Extraction of rice phenological differences under heavy metal stress using EVI time-series from HJ-1A/B Data. Sensors 17: 1243. doi: 10.3390/s17061243
![]() |
[107] |
Xing F, Yao H, Liu Y, et al. (2019) Recent developments and applications of hyperspectral imaging for rapid detection of mycotoxins and mycotoxigenic fungi in food products. Crit Rev Food Sci Nutr 59: 173-180. doi: 10.1080/10408398.2017.1363709
![]() |
[108] |
Liu T, Liu X, Liu M, et al. (2018) Classification of rice heavy metal stress levels based on phenological characteristics using remote sensing time-series images and data mining algorithms. Sensors 18: 4425. doi: 10.3390/s18124425
![]() |
[109] |
Jeong S, Ko J, Yeom J-M (2018) Nationwide projection of rice yield using a crop model integrated with geostationary satellite imagery: a case study in South Korea. Remote Sens 10: 1665. doi: 10.3390/rs10101665
![]() |
[110] |
Lu J, Miao Y, Shi W, et al. (2020) Developing a proximal active canopy sensor-based precision nitrogen management strategy for high-yielding rice. Remote Sens 12: 1440. doi: 10.3390/rs12091440
![]() |
[111] |
Paleari L, Movedi E, Vesely FM, et al. (2019) Estimating crop nutritional status using smart apps to support nitrogen fertilization. A case study on paddy rice. Sensors 19: 981. doi: 10.3390/s19040981
![]() |
[112] |
Park S, Im J, Park S, et al. (2018) Classification and mapping of paddy rice by combining Landsat and SAR time series data. Remote Sens 10: 447. doi: 10.3390/rs10030447
![]() |
[113] |
Ranghetti L, Cardarelli E, Boschetti M, et al. (2018) Assessment of water management changes in the Italian rice paddies from 2000 to 2016 using satellite data: A contribution to agro-ecological studies. Remote Sens 10: 416. doi: 10.3390/rs10030416
![]() |
[114] |
Sianturi R, Jetten VG, Ettema J, et al. (2018) Distinguishing between hazardous flooding and non-hazardous agronomic inundation in irrigated rice fields: A case study from west java. Remote Sens 10: 1003. doi: 10.3390/rs10071003
![]() |
[115] |
Zhang S, Li J, Wang S, et al. (2020) Repaid identification and prediction of Cadmium-Lead cross-stress of different stress levels in rice canopy based on visible and near-infrared spectroscopy. Remote Sens 12: 469. doi: 10.3390/rs12030469
![]() |
[116] |
Barnes PW, Searles PS, Ballaré CL, et al. (2000) Non-invasive measurements of leaf epidermal transmittance of UV radiation using chlorophyll fluorescence: field and laboratory studies. Physiol Plant 109: 274-283. doi: 10.1034/j.1399-3054.2000.100308.x
![]() |
[117] |
Du L, Shi S, Yang J, et al. (2016) Using different regression methods to estimate leaf nitrogen content in rice by fusing hyperspectral LiDAR data and laser-induced chlorophyll fluorescence data. Remote Sens 8: 526. doi: 10.3390/rs8060526
![]() |
[118] |
Du L, Yang J, Chen B, et al. (2020) Novel combined spectral indices derived from hyperspectral and laser-induced fluorescence LiDAR spectra for leaf nitrogen contents estimation of rice. Remote Sens 12: 185. doi: 10.3390/rs12010185
![]() |
[119] | Yang J, Du L, Sun J, et al. (2016) Estimating the leaf nitrogen content of paddy rice by using the combined reflectance and laser-induced fluorescence spectra. Opt Express 24: 19354-19365. |
[120] |
Yang J, Song S, Du L, et al. (2018) Analyzing the effect of fluorescence characteristics on leaf nitrogen concentration estimation. Remote Sens 10: 1402. doi: 10.3390/rs10091402
![]() |
[121] | Fowler JE (2014) Compressive pushbroom and whiskbroom sensing for hyperspectral remote-sensing imaging, 2014 IEEE international conference on image processing (ICIP), 684-688. |
[122] | Kerekes JP, Schott JR (2007) Hyperspectral imaging systems. Hyperspectral data exploitation: Theory and applications, 19-45. |
[123] | Iturrino C, Arias FX, Sierra H, et al. (2019) Single-shot multispectral image acquisition for low-altitude remote sensing using light diffraction techniques, 2019 10th workshop on hyperspectral imaging and signal processing: Evolution in remote sensing (WHISPERS), 1-5. |
[124] |
Blasinski H, Farrell J, Wandell B (2020) Simultaneous surface reflectance and fluorescence spectra estimation. IEEE Trans Image Process 29: 8791-8804. doi: 10.1109/TIP.2020.2973810
![]() |
[125] |
He X, Feng X, Sun D, et al. (2019) Rapid and nondestructive measurement of rice seed vitality of different years using near-infrared hyperspectral imaging. Molecules 24: 2227. doi: 10.3390/molecules24122227
![]() |
[126] | Yuan J, Su Z, Jia Y, et al. (2016) Identification of rice leaf blast and nitrogen deficiency in cold region using hyperspectral imaging. Trans Chin Soc Agric Eng 32: 155-160. |
[127] | Gewali UB, Monteiro ST (2016) A novel covariance function for predicting vegetation biochemistry from hyperspectral imagery with Gaussian processes, 2016 IEEE international conference on image processing (ICIP), 2216-2220. |
[128] | Kampe T, Leisso N, Musinsky J, et al. (2013) The NEON 2013 airborne campaign at domain 17 terrestrial and aquatic sites in California. NEON Technical Memorandum Series, TM-005. |
[129] |
Ding M, Guan Q, Li L, et al. (2020) Phenology-based rice paddy mapping using multi-source satellite imagery and a fusion algorithm applied to the poyang lake plain, Southern China. Remote Sens 12: 1022. doi: 10.3390/rs12061022
![]() |
[130] |
Du L, Gong W, Shi S, et al. (2016) Estimation of rice leaf nitrogen contents based on hyperspectral LIDAR. Int J Appl Earth Obs Geoinformation 44: 136-143. doi: 10.1016/j.jag.2015.08.008
![]() |
[131] |
Gnyp ML, Miao Y, Yuan F, et al. (2014) Hyperspectral canopy sensing of paddy rice aboveground biomass at different growth stages. Field Crops Res 155: 42-55. doi: 10.1016/j.fcr.2013.09.023
![]() |
[132] |
Cheng T, Song R, Li D, et al. (2017) Spectroscopic estimation of biomass in canopy components of paddy rice using dry matter and chlorophyll indices. Remote Sens 9: 319. doi: 10.3390/rs9040319
![]() |
[133] |
Li D, Wang X, Zheng H, et al. (2018) Estimation of area-and mass-based leaf nitrogen contents of wheat and rice crops from water-removed spectra using continuous wavelet analysis. Plant Methods 14: 76. doi: 10.1186/s13007-018-0344-1
![]() |
[134] |
Liu Z, Shi J, Zhang L, et al. (2010) Discrimination of rice panicles by hyperspectral reflectance data based on principal component analysis and support vector classification. J Zhejiang Univ Sci B 11: 71-78. doi: 10.1631/jzus.B0900193
![]() |
[135] |
Zheng H, Ma J, Zhou M, et al. (2020) Enhancing the nitrogen signals of rice canopies across critical growth stages through the integration of textural and spectral information from unmanned aerial vehicle (UAV) multispectral imagery. Remote Sens 12: 957. doi: 10.3390/rs12060957
![]() |
[136] | Starr C, Evers C, Starr L (2010) Biology: Concepts and applications, Cengage Learning. |
[137] |
Das B, Sahoo RN, Biswas A, et al. (2020) Discrimination of rice genotypes using field spectroradiometry. Geocarto Int 35: 64-77. doi: 10.1080/10106049.2018.1506507
![]() |
[138] |
Yang M-D, Huang K-S, Kuo Y-H, et al. (2017) Spatial and spectral hybrid image classification for rice lodging assessment through UAV imagery. Remote Sens 9: 583. doi: 10.3390/rs9060583
![]() |
[139] |
Yang C-M (2010) Assessment of the severity of bacterial leaf blight in rice using canopy hyperspectral reflectance. Precis Agric 11: 61-81. doi: 10.1007/s11119-009-9122-4
![]() |
[140] |
Dunn BW, Dehaan R, Schmidtke L, et al. (2016) Using field-derived hyperspectral reflectance measurement to identify the essential wavelengths for predicting nitrogen uptake of rice at panicle initiation. J Near Infrared Spectrosc 24: 473-483. doi: 10.1255/jnirs.1246
![]() |
[141] |
Feng H, Jiang N, Huang C, et al. (2013) A hyperspectral imaging system for an accurate prediction of the above-ground biomass of individual rice plants. Rev Sci Instrum 84: 095107. doi: 10.1063/1.4818918
![]() |
[142] |
Kawamura K, Ikeura H, Phongchanmaixay S, et al. (2018) Canopy hyperspectral sensing of paddy fields at the booting stage and PLS regression can assess grain yield. Remote Sens 10: 1249. doi: 10.3390/rs10081249
![]() |
[143] |
Fabiyi SD, Vu H, Tachtatzis C, et al. (2020) Varietal classification of rice seeds using RGB and hyperspectral images. IEEE Access 8: 22493-22505. doi: 10.1109/ACCESS.2020.2969847
![]() |
[144] |
Sun J, Shi S, Yang J, et al. (2018) Estimating leaf chlorophyll status using hyperspectral lidar measurements by PROSPECT model inversion. Remote Sens Environ 212: 1-7. doi: 10.1016/j.rse.2018.04.024
![]() |
[145] |
Kong W, Zhang C, Liu F, et al. (2013) Rice seed cultivar identification using near-infrared hyperspectral imaging and multivariate data analysis. Sensors 13: 8916-8927. doi: 10.3390/s130708916
![]() |
[146] |
Devia CA, Rojas JP, Petro E, et al. (2019) High-throughput biomass estimation in rice crops using UAV multispectral imagery. J Intell Rob Syst 96: 573-589. doi: 10.1007/s10846-019-01001-5
![]() |
[147] |
Zhao X, Yuan Y, Song M, et al. (2019) Use of unmanned aerial vehicle imagery and deep learning unet to extract rice lodging. Sensors 19: 3859. doi: 10.3390/s19183859
![]() |
[148] | Zheng H, Zhou X, Cheng T, et al. (2016) Evaluation of a UAV-based hyperspectral frame camera for monitoring the leaf nitrogen concentration in rice, 2016 IEEE international geoscience and remote sensing symposium (IGARSS), 7350-7353. |
[149] |
Zhang Y, Gao J, Cen H, et al. (2019) Automated spectral feature extraction from hyperspectral images to differentiate weedy rice and barnyard grass from a rice crop. Comput Electron Agric 159: 42-49. doi: 10.1016/j.compag.2019.02.018
![]() |
[150] | Witt C, Pasuquin J, Mutters R, et al. (2005) New leaf color chart for effective nitrogen management in rice. Better Crops 89: 36-39. |
[151] |
Sun J, Shi S, Gong W, et al. (2017) Evaluation of hyperspectral LiDAR for monitoring rice leaf nitrogen by comparison with multispectral LiDAR and passive spectrometer. Sci Rep 7: 40362. doi: 10.1038/srep40362
![]() |
[152] |
Zhou W, Zhang J, Zou M, et al. (2019) Feasibility of using rice leaves hyperspectral data to estimate CaCl 2-extractable concentrations of heavy metals in agricultural soil. Sci Rep 9: 1-9. doi: 10.1038/s41598-018-37186-2
![]() |
[153] |
Uto K, Seki H, Saito G, et al. (2013) Characterization of rice paddies by a UAV-mounted miniature hyperspectral sensor system. IEEE J Sel Top Appl Earth Obs Remote Sens 6: 851-860. doi: 10.1109/JSTARS.2013.2250921
![]() |
[154] |
Inoue Y, Penuelas J (2001) An AOTF-based hyperspectral imaging system for field use in ecophysiological and agricultural applications. Int J Remote Sens 22: 3883-3888. doi: 10.1080/01431160110069863
![]() |
[155] |
Li D, Wang R, Xie C, et al. (2020) A recognition method for rice plant diseases and pests video detection based on deep convolutional neural network. Sensors 20: 578. doi: 10.3390/s20030578
![]() |
[156] |
Liu Z, Zhang Q, Han T, et al. (2016) Heavy metal pollution in a soil-rice system in the Yangtze river region of China. Int J Environ Res Pub Health 13: 63. doi: 10.3390/ijerph13010063
![]() |
[157] | Shwetank JK, Bhatia K (2010) Review of rice crop identification and classification using hyper-spectral image processing system. Int J Comput Sci Commun 1: 253-258. |
[158] |
Onoyama H, Ryu C, Suguri M, et al. (2018) Estimation of rice protein content before harvest using ground-based hyperspectral imaging and region of interest analysis. Precis Agric 19: 721-734. doi: 10.1007/s11119-017-9552-3
![]() |
[159] |
Nagelkerke NJ, others (1991) A note on a general definition of the coefficient of determination. Biometrika 78: 691-692. doi: 10.1093/biomet/78.3.691
![]() |
[160] |
Samat A, Gamba P, Abuduwaili J, et al. (2016) Geodesic flow kernel support vector machine for hyperspectral image classification by unsupervised subspace feature transfer. Remote Sens 8: 234. doi: 10.3390/rs8030234
![]() |
[161] |
Tehrany MS, Pradhan B, Mansor S, et al. (2015) Flood susceptibility assessment using GIS-based support vector machine model with different kernel types. Catena 125: 91-101. doi: 10.1016/j.catena.2014.10.017
![]() |
[162] |
Liu P, Choo K-KR, Wang L, et al. (2017) SVM or deep learning? A comparative study on remote sensing image classification. Soft Comput 21: 7053-7065. doi: 10.1007/s00500-016-2247-2
![]() |
[163] | Hecht-Nielsen R (1992) Theory of the backpropagation neural network, Neural networks for perception, Elsevier, 65-93. |
[164] |
Chen Y, Jiang H, Li C, et al. (2016) Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans Geosci Remote Sens 54: 6232-6251. doi: 10.1109/TGRS.2016.2584107
![]() |
[165] |
Wang J, Zhang C, Shi Y, et al. (2020) Evaluation of quinclorac toxicity and alleviation by salicylic acid in rice seedlings using ground-based visible/near-infrared hyperspectral imaging. Plant Methods 16: 1-16. doi: 10.1186/s13007-019-0534-5
![]() |
[166] |
Thenkabail PS, Enclona EA, Ashton MS, et al. (2004) Accuracy assessments of hyperspectral waveband performance for vegetation analysis applications. Remote Sens Environ 91: 354-376. doi: 10.1016/j.rse.2004.03.013
![]() |
[167] | Chang YL, Chang L, Xu M-X, et al. (2017) Impurity function band prioritization based on particle swarm optimization and gravitational search algorithm for hyperspectral images, 2017 IEEE international geoscience and remote sensing symposium (IGARSS), 1788-1791. |
[168] |
Knyazikhin Y, Schull MA, Stenberg P, et al. (2013) Hyperspectral remote sensing of foliar nitrogen content. Proc the Natl Acad Sci 110: E185-E192. doi: 10.1073/pnas.1210196109
![]() |
1. | Jiayu Fu, Haiyan Wang, Risu Na, A JISAIHAN, Zhixiong Wang, Yuko OHNO, Recent advancements in digital health management using multi-modal signal monitoring, 2023, 20, 1551-0018, 5194, 10.3934/mbe.2023241 | |
2. | Qing Lin, Dino Oglić, Hak-Keung Lam, Michael J. Curtis, Zoran Cvetkovic, 2024, A Hybrid GCN-LSTM Model for Ventricular Arrhythmia Classification Based on ECG Pattern Similarity, 979-8-3503-7149-9, 1, 10.1109/EMBC53108.2024.10781976 | |
3. | Tuan Hoang Vu, Anh Nguyen, Minh Tuan Nguyen, 2024, Graph Attention Networks Based Cardiac Arrest Diagnosis, 979-8-3503-5398-3, 762, 10.1109/ATC63255.2024.10908172 |
Type | Description | Quantity |
N | Normal beat | 500 |
V | Premature ventricular contraction | 500 |
A | Atrial premature beat | 200 |
F | Fusion of ventricular and normal beat | 200 |
n | Supraventricular escape beat | 30 |
R | Right bundle branch block beat | 200 |
j | Nodal (junctional) escape beat | 90 |
Total | 1720 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 487 | 4 | 5 | 3 | 1 | 0 | 0 | 0.9645 | 0.9740 | 0.9103 | 0.9607 | 0.9411 |
V | 13 | 477 | 0 | 10 | 0 | 0 | 0 | 0.9715 | 0.9540 | 0.9483 | 0.9787 | 0.9511 |
A | 20 | 1 | 176 | 1 | 0 | 1 | 1 | 0.9814 | 0.8800 | 0.9565 | 0.9947 | 0.9167 |
F | 10 | 17 | 2 | 170 | 0 | 0 | 1 | 0.9738 | 0.8500 | 0.9189 | 0.9901 | 0.8831 |
n | 3 | 0 | 1 | 0 | 26 | 0 | 0 | 0.9971 | 0.8667 | 0.9630 | 0.9994 | 0.9123 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 2 | 0 | 0 | 1 | 0 | 0 | 87 | 0.9971 | 0.9667 | 0.9775 | 0.9988 | 0.9721 |
Average | 0.9832 | 0.9245 | 0.9528 | 0.9888 | 0.9377 | |||||||
Overall accuracy | 0.9413 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 477 | 7 | 4 | 9 | 2 | 0 | 1 | 0.9657 | 0.9540 | 0.9298 | 0.9705 | 0.9418 |
V | 4 | 483 | 2 | 11 | 0 | 0 | 0 | 0.9715 | 0.9660 | 0.9379 | 0.9738 | 0.9517 |
A | 12 | 2 | 183 | 3 | 0 | 0 | 0 | 0.9860 | 0.9150 | 0.9632 | 0.9954 | 0.9385 |
F | 17 | 21 | 1 | 157 | 0 | 2 | 2 | 0.9610 | 0.7850 | 0.8674 | 0.9842 | 0.8241 |
n | 2 | 0 | 0 | 1 | 27 | 0 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 1 | 0 | 0 | 0 | 199 | 0 | 0.9977 | 0.9950 | 0.9851 | 0.9980 | 0.9901 |
j | 1 | 1 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9822 | 0.9260 | 0.9402 | 0.9884 | 0.9326 | |||||||
Overall accuracy | 0.9378 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 488 | 6 | 3 | 1 | 0 | 0 | 2 | 0.9581 | 0.9760 | 0.8905 | 0.9508 | 0.9313 |
V | 11 | 482 | 0 | 6 | 0 | 0 | 1 | 0.9733 | 0.9640 | 0.9451 | 0.9770 | 0.9545 |
A | 25 | 5 | 169 | 1 | 0 | 0 | 0 | 0.9779 | 0.8450 | 0.9602 | 0.9954 | 0.8989 |
F | 15 | 14 | 3 | 168 | 0 | 0 | 0 | 0.9767 | 0.8400 | 0.9545 | 0.9947 | 0.8936 |
n | 2 | 0 | 0 | 0 | 28 | 0 | 0 | 0.9988 | 0.9333 | 1.0000 | 1.0000 | 0.9655 |
R | 0 | 3 | 0 | 0 | 0 | 196 | 1 | 0.9977 | 0.9800 | 1.0000 | 1.0000 | 0.9899 |
j | 7 | 0 | 1 | 0 | 0 | 0 | 82 | 0.9930 | 0.9111 | 0.9535 | 0.9975 | 0.9318 |
Average | 0.9822 | 0.9213 | 0.9577 | 0.9879 | 0.9379 | |||||||
Overall accuracy | 0.9378 |
Method | Year | #Types | #Samples | Ov.ACC |
PCANet (Ⅱ) | 2018 | 7 | 1720 | 93.72% |
PCANet (V1) | 2018 | 7 | 1720 | 92.10% |
PCANet (V5) | 2018 | 7 | 1720 | 93.37% |
RandNet (Ⅱ) | 2019 | 7 | 1720 | 92.67% |
RandNet (V1) | 2019 | 7 | 1720 | 90.76% |
RandNet (V5) | 2019 | 7 | 1720 | 92.91% |
MADHGAT (Ⅱ) | 2022 | 7 | 1720 | 94.13% |
MADHGAT (V1) | 2022 | 7 | 1720 | 93.78% |
MADHGAT (V5) | 2022 | 7 | 1720 | 93.78% |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 7 | 2 | 5 | 2 | 0 | 0 | 0.9727 | 0.9680 | 0.9398 | 0.9746 | 0.9537 |
V | 5 | 488 | 0 | 6 | 0 | 0 | 1 | 0.9767 | 0.9760 | 0.9457 | 0.9770 | 0.9606 |
A | 13 | 1 | 181 | 4 | 0 | 1 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 5 | 11 | 1 | 183 | 0 | 0 | 0 | 0.9797 | 0.9150 | 0.9104 | 0.9882 | 0.9127 |
n | 3 | 0 | 0 | 1 | 25 | 1 | 0 | 0.9959 | 0.8333 | 0.9259 | 0.9988 | 0.8772 |
R | 0 | 8 | 0 | 0 | 0 | 192 | 0 | 0.9936 | 0.9600 | 0.9846 | 0.9980 | 0.9722 |
j | 5 | 1 | 1 | 2 | 0 | 1 | 80 | 0.9936 | 0.8889 | 0.9877 | 0.9994 | 0.9357 |
Average | 0.9855 | 0.9209 | 0.9532 | 0.9905 | 0.9360 | |||||||
Overall accuracy | 0.9494 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 8 | 1 | 6 | 1 | 0 | 0 | 0.9698 | 0.9680 | 0.9308 | 0.9705 | 0.9490 |
V | 5 | 486 | 1 | 7 | 0 | 1 | 0 | 0.9785 | 0.9720 | 0.9548 | 0.9811 | 0.9633 |
A | 14 | 2 | 181 | 3 | 0 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 11 | 9 | 2 | 177 | 0 | 0 | 1 | 0.9773 | 0.8850 | 0.9171 | 0.9895 | 0.9008 |
n | 3 | 0 | 0 | 0 | 27 | 0 | 0 | 0.9977 | 0.9000 | 0.9643 | 0.9994 | 0.9310 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 3 | 0 | 0 | 0 | 0 | 0 | 87 | 0.9977 | 0.9667 | 0.9886 | 0.9994 | 0.9775 |
Average | 0.9864 | 0.9395 | 0.9613 | 0.9909 | 0.9499 | |||||||
Overall accuracy | 0.9523 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9750 | 0.9780 | 0.9386 | 0.9738 | 0.9579 |
V | 4 | 489 | 0 | 7 | 0 | 0 | 0 | 0.9785 | 0.9780 | 0.9495 | 0.9787 | 0.9635 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
R | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
j | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
n | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9870 | 0.9381 | 0.9595 | 0.9914 | 0.9483 | |||||||
Overall accuracy | 0.9547 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9756 | 0.9780 | 0.9404 | 0.9746 | 0.9588 |
V | 3 | 490 | 0 | 7 | 0 | 0 | 0 | 0.9791 | 0.9800 | 0.9496 | 0.9787 | 0.9646 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
n | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
j | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 | |||||||
Overall accuracy | 0.9552 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 2 | 9 | 4 | 0 | 0 | 1 | 0.9820 | 0.9680 | 0.9699 | 0.9877 | 0.9690 |
V | 1 | 496 | 0 | 3 | 0 | 0 | 0 | 0.9907 | 0.9920 | 0.9764 | 0.9902 | 0.9841 |
A | 8 | 0 | 191 | 0 | 0 | 0 | 1 | 0.9890 | 0.9550 | 0.9502 | 0.9934 | 0.9526 |
F | 3 | 6 | 1 | 190 | 0 | 0 | 0 | 0.9901 | 0.9500 | 0.9645 | 0.9954 | 0.9572 |
n | 1 | 0 | 0 | 0 | 29 | 0 | 0 | 0.9994 | 0.9667 | 1.0000 | 1.0000 | 0.9831 |
R | 0 | 4 | 0 | 0 | 0 | 195 | 1 | 0.9965 | 0.9750 | 0.9949 | 0.9993 | 0.9848 |
j | 2 | 0 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9920 | 0.9676 | 0.9747 | 0.9949 | 0.9711 | |||||||
Overall accuracy | 0.9721 |
Method | Year | #Types | #Samples | Ov.ACC |
DL-CCANet (Ⅱ and V1) | 2019 | 7 | 1720 | 94.01% |
DL-CCANet (Ⅱ and V5) | 2019 | 7 | 1720 | 94.07% |
DL-CCANet (V1 and V5) | 2019 | 7 | 1720 | 93.90% |
TL-CCANet (Ⅱ, V1 and V5) | 2019 | 7 | 1720 | 95.52% |
1-D ResNet (Ⅱ and V1) | 2021 | 7 | 1720 | 86.25% |
1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 2021 | 7 | 1720 | 93.60% |
1-D CCANet-SVD (w/o ar) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.12% |
1-D CCANet-SVD (w/o 1D-spec) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.77% |
1-D CCANet-SVD (w/o time-domain feature) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.35% |
1-D CCANet-SVD (w/o stack) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.83% |
1-D CCANet-SVD (Ⅱ and V1) | 2021 | 7 | 1720 | 95.70% |
MADHGAT (Ⅱ and V1) | 2022 | 7 | 1720 | 94.94% |
MADHGAT (Ⅱ and V5) | 2022 | 7 | 1720 | 95.23% |
MADHGAT (V1 and V5) | 2022 | 7 | 1720 | 95.47% |
MADHGAT (Ⅱ, V1 and V5) | 2022 | 7 | 1720 | 95.52% |
MADHGAT (12 leads) | 2022 | 7 | 1720 | 97.21% |
Graph | Ov.ACC | Avg-Acc | Avg-Sen | Avg-Ppv | Avg-Spe | Avg-F1 |
without multiple leads | 0.9384 | 0.9824 | 0.9152 | 0.9345 | 0.9884 | 0.9243 |
without single-lead beats | 0.8273 | 0.9507 | 0.7660 | 0.8507 | 0.9670 | 0.7975 |
only multi-lead beats | 0.8070 | 0.9449 | 0.7070 | 0.8542 | 0.9623 | 0.7568 |
full graph | 0.9552 | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 |
Ratio of training set to test set | Method | Ov.ACC |
4:1 | DL-CCANet (Ⅱ and V1) | 94.01% |
4:1 | 1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 93.60% |
4:1 | 1-D ResNet (Ⅱ and V1) | 86.25% |
4:1 | MADHGAT | 94.94% |
3:1 | MADHGAT | 94.36% |
2:1 | MADHGAT | 94.48% |
1:1 | MADHGAT | 92.38% |
1:2 | MADHGAT | 90.20% |
Type | Description | Quantity |
N | Normal beat | 500 |
V | Premature ventricular contraction | 500 |
A | Atrial premature beat | 200 |
F | Fusion of ventricular and normal beat | 200 |
n | Supraventricular escape beat | 30 |
R | Right bundle branch block beat | 200 |
j | Nodal (junctional) escape beat | 90 |
Total | 1720 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 487 | 4 | 5 | 3 | 1 | 0 | 0 | 0.9645 | 0.9740 | 0.9103 | 0.9607 | 0.9411 |
V | 13 | 477 | 0 | 10 | 0 | 0 | 0 | 0.9715 | 0.9540 | 0.9483 | 0.9787 | 0.9511 |
A | 20 | 1 | 176 | 1 | 0 | 1 | 1 | 0.9814 | 0.8800 | 0.9565 | 0.9947 | 0.9167 |
F | 10 | 17 | 2 | 170 | 0 | 0 | 1 | 0.9738 | 0.8500 | 0.9189 | 0.9901 | 0.8831 |
n | 3 | 0 | 1 | 0 | 26 | 0 | 0 | 0.9971 | 0.8667 | 0.9630 | 0.9994 | 0.9123 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 2 | 0 | 0 | 1 | 0 | 0 | 87 | 0.9971 | 0.9667 | 0.9775 | 0.9988 | 0.9721 |
Average | 0.9832 | 0.9245 | 0.9528 | 0.9888 | 0.9377 | |||||||
Overall accuracy | 0.9413 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 477 | 7 | 4 | 9 | 2 | 0 | 1 | 0.9657 | 0.9540 | 0.9298 | 0.9705 | 0.9418 |
V | 4 | 483 | 2 | 11 | 0 | 0 | 0 | 0.9715 | 0.9660 | 0.9379 | 0.9738 | 0.9517 |
A | 12 | 2 | 183 | 3 | 0 | 0 | 0 | 0.9860 | 0.9150 | 0.9632 | 0.9954 | 0.9385 |
F | 17 | 21 | 1 | 157 | 0 | 2 | 2 | 0.9610 | 0.7850 | 0.8674 | 0.9842 | 0.8241 |
n | 2 | 0 | 0 | 1 | 27 | 0 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 1 | 0 | 0 | 0 | 199 | 0 | 0.9977 | 0.9950 | 0.9851 | 0.9980 | 0.9901 |
j | 1 | 1 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9822 | 0.9260 | 0.9402 | 0.9884 | 0.9326 | |||||||
Overall accuracy | 0.9378 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 488 | 6 | 3 | 1 | 0 | 0 | 2 | 0.9581 | 0.9760 | 0.8905 | 0.9508 | 0.9313 |
V | 11 | 482 | 0 | 6 | 0 | 0 | 1 | 0.9733 | 0.9640 | 0.9451 | 0.9770 | 0.9545 |
A | 25 | 5 | 169 | 1 | 0 | 0 | 0 | 0.9779 | 0.8450 | 0.9602 | 0.9954 | 0.8989 |
F | 15 | 14 | 3 | 168 | 0 | 0 | 0 | 0.9767 | 0.8400 | 0.9545 | 0.9947 | 0.8936 |
n | 2 | 0 | 0 | 0 | 28 | 0 | 0 | 0.9988 | 0.9333 | 1.0000 | 1.0000 | 0.9655 |
R | 0 | 3 | 0 | 0 | 0 | 196 | 1 | 0.9977 | 0.9800 | 1.0000 | 1.0000 | 0.9899 |
j | 7 | 0 | 1 | 0 | 0 | 0 | 82 | 0.9930 | 0.9111 | 0.9535 | 0.9975 | 0.9318 |
Average | 0.9822 | 0.9213 | 0.9577 | 0.9879 | 0.9379 | |||||||
Overall accuracy | 0.9378 |
Method | Year | #Types | #Samples | Ov.ACC |
PCANet (Ⅱ) | 2018 | 7 | 1720 | 93.72% |
PCANet (V1) | 2018 | 7 | 1720 | 92.10% |
PCANet (V5) | 2018 | 7 | 1720 | 93.37% |
RandNet (Ⅱ) | 2019 | 7 | 1720 | 92.67% |
RandNet (V1) | 2019 | 7 | 1720 | 90.76% |
RandNet (V5) | 2019 | 7 | 1720 | 92.91% |
MADHGAT (Ⅱ) | 2022 | 7 | 1720 | 94.13% |
MADHGAT (V1) | 2022 | 7 | 1720 | 93.78% |
MADHGAT (V5) | 2022 | 7 | 1720 | 93.78% |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 7 | 2 | 5 | 2 | 0 | 0 | 0.9727 | 0.9680 | 0.9398 | 0.9746 | 0.9537 |
V | 5 | 488 | 0 | 6 | 0 | 0 | 1 | 0.9767 | 0.9760 | 0.9457 | 0.9770 | 0.9606 |
A | 13 | 1 | 181 | 4 | 0 | 1 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 5 | 11 | 1 | 183 | 0 | 0 | 0 | 0.9797 | 0.9150 | 0.9104 | 0.9882 | 0.9127 |
n | 3 | 0 | 0 | 1 | 25 | 1 | 0 | 0.9959 | 0.8333 | 0.9259 | 0.9988 | 0.8772 |
R | 0 | 8 | 0 | 0 | 0 | 192 | 0 | 0.9936 | 0.9600 | 0.9846 | 0.9980 | 0.9722 |
j | 5 | 1 | 1 | 2 | 0 | 1 | 80 | 0.9936 | 0.8889 | 0.9877 | 0.9994 | 0.9357 |
Average | 0.9855 | 0.9209 | 0.9532 | 0.9905 | 0.9360 | |||||||
Overall accuracy | 0.9494 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 8 | 1 | 6 | 1 | 0 | 0 | 0.9698 | 0.9680 | 0.9308 | 0.9705 | 0.9490 |
V | 5 | 486 | 1 | 7 | 0 | 1 | 0 | 0.9785 | 0.9720 | 0.9548 | 0.9811 | 0.9633 |
A | 14 | 2 | 181 | 3 | 0 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 11 | 9 | 2 | 177 | 0 | 0 | 1 | 0.9773 | 0.8850 | 0.9171 | 0.9895 | 0.9008 |
n | 3 | 0 | 0 | 0 | 27 | 0 | 0 | 0.9977 | 0.9000 | 0.9643 | 0.9994 | 0.9310 |
R | 0 | 4 | 0 | 0 | 0 | 196 | 0 | 0.9971 | 0.9800 | 0.9949 | 0.9993 | 0.9874 |
j | 3 | 0 | 0 | 0 | 0 | 0 | 87 | 0.9977 | 0.9667 | 0.9886 | 0.9994 | 0.9775 |
Average | 0.9864 | 0.9395 | 0.9613 | 0.9909 | 0.9499 | |||||||
Overall accuracy | 0.9523 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9750 | 0.9780 | 0.9386 | 0.9738 | 0.9579 |
V | 4 | 489 | 0 | 7 | 0 | 0 | 0 | 0.9785 | 0.9780 | 0.9495 | 0.9787 | 0.9635 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
R | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
j | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
n | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9870 | 0.9381 | 0.9595 | 0.9914 | 0.9483 | |||||||
Overall accuracy | 0.9547 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 489 | 6 | 2 | 3 | 0 | 0 | 0 | 0.9756 | 0.9780 | 0.9404 | 0.9746 | 0.9588 |
V | 3 | 490 | 0 | 7 | 0 | 0 | 0 | 0.9791 | 0.9800 | 0.9496 | 0.9787 | 0.9646 |
A | 16 | 1 | 181 | 1 | 1 | 0 | 0 | 0.9866 | 0.9050 | 0.9784 | 0.9974 | 0.9403 |
F | 7 | 16 | 1 | 175 | 0 | 1 | 0 | 0.9791 | 0.8750 | 0.9409 | 0.9928 | 0.9067 |
n | 2 | 0 | 0 | 0 | 27 | 1 | 0 | 0.9971 | 0.9000 | 0.9310 | 0.9988 | 0.9153 |
R | 0 | 3 | 0 | 0 | 1 | 195 | 1 | 0.9959 | 0.9750 | 0.9898 | 0.9987 | 0.9824 |
j | 3 | 0 | 1 | 0 | 0 | 0 | 86 | 0.9971 | 0.9556 | 0.9885 | 0.9994 | 0.9718 |
Average | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 | |||||||
Overall accuracy | 0.9552 |
Predicted | Acc | Sen | Ppv | Spe | F1-score | |||||||
N | V | A | F | n | R | j | ||||||
N | 484 | 2 | 9 | 4 | 0 | 0 | 1 | 0.9820 | 0.9680 | 0.9699 | 0.9877 | 0.9690 |
V | 1 | 496 | 0 | 3 | 0 | 0 | 0 | 0.9907 | 0.9920 | 0.9764 | 0.9902 | 0.9841 |
A | 8 | 0 | 191 | 0 | 0 | 0 | 1 | 0.9890 | 0.9550 | 0.9502 | 0.9934 | 0.9526 |
F | 3 | 6 | 1 | 190 | 0 | 0 | 0 | 0.9901 | 0.9500 | 0.9645 | 0.9954 | 0.9572 |
n | 1 | 0 | 0 | 0 | 29 | 0 | 0 | 0.9994 | 0.9667 | 1.0000 | 1.0000 | 0.9831 |
R | 0 | 4 | 0 | 0 | 0 | 195 | 1 | 0.9965 | 0.9750 | 0.9949 | 0.9993 | 0.9848 |
j | 2 | 0 | 0 | 0 | 0 | 1 | 87 | 0.9965 | 0.9667 | 0.9667 | 0.9982 | 0.9667 |
Average | 0.9920 | 0.9676 | 0.9747 | 0.9949 | 0.9711 | |||||||
Overall accuracy | 0.9721 |
Method | Year | #Types | #Samples | Ov.ACC |
DL-CCANet (Ⅱ and V1) | 2019 | 7 | 1720 | 94.01% |
DL-CCANet (Ⅱ and V5) | 2019 | 7 | 1720 | 94.07% |
DL-CCANet (V1 and V5) | 2019 | 7 | 1720 | 93.90% |
TL-CCANet (Ⅱ, V1 and V5) | 2019 | 7 | 1720 | 95.52% |
1-D ResNet (Ⅱ and V1) | 2021 | 7 | 1720 | 86.25% |
1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 2021 | 7 | 1720 | 93.60% |
1-D CCANet-SVD (w/o ar) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.12% |
1-D CCANet-SVD (w/o 1D-spec) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.77% |
1-D CCANet-SVD (w/o time-domain feature) (Ⅱ and V1) | 2021 | 7 | 1720 | 95.35% |
1-D CCANet-SVD (w/o stack) (Ⅱ and V1) | 2021 | 7 | 1720 | 94.83% |
1-D CCANet-SVD (Ⅱ and V1) | 2021 | 7 | 1720 | 95.70% |
MADHGAT (Ⅱ and V1) | 2022 | 7 | 1720 | 94.94% |
MADHGAT (Ⅱ and V5) | 2022 | 7 | 1720 | 95.23% |
MADHGAT (V1 and V5) | 2022 | 7 | 1720 | 95.47% |
MADHGAT (Ⅱ, V1 and V5) | 2022 | 7 | 1720 | 95.52% |
MADHGAT (12 leads) | 2022 | 7 | 1720 | 97.21% |
Graph | Ov.ACC | Avg-Acc | Avg-Sen | Avg-Ppv | Avg-Spe | Avg-F1 |
without multiple leads | 0.9384 | 0.9824 | 0.9152 | 0.9345 | 0.9884 | 0.9243 |
without single-lead beats | 0.8273 | 0.9507 | 0.7660 | 0.8507 | 0.9670 | 0.7975 |
only multi-lead beats | 0.8070 | 0.9449 | 0.7070 | 0.8542 | 0.9623 | 0.7568 |
full graph | 0.9552 | 0.9872 | 0.9384 | 0.9598 | 0.9915 | 0.9485 |
Ratio of training set to test set | Method | Ov.ACC |
4:1 | DL-CCANet (Ⅱ and V1) | 94.01% |
4:1 | 1-D CCANet-SVD (w/o SVD) (Ⅱ and V1) | 93.60% |
4:1 | 1-D ResNet (Ⅱ and V1) | 86.25% |
4:1 | MADHGAT | 94.94% |
3:1 | MADHGAT | 94.36% |
2:1 | MADHGAT | 94.48% |
1:1 | MADHGAT | 92.38% |
1:2 | MADHGAT | 90.20% |