
Citation: Anna Kostiukow, Wojciech Strzelecki, Piotr Poniewierski, Włodzimierz Samborski. The estimation of the functioning of families with ASD children[J]. AIMS Public Health, 2019, 6(4): 587-599. doi: 10.3934/publichealth.2019.4.587
[1] | Maria Pia Riccio, Gennaro Catone, Rosamaria Siracusano, Luisa Occhiati, Pia Bernardo, Emilia Sarnataro, Giuseppina Corrado, Carmela Bravaccio . Vitamin D deficiency is not related to eating habits in children with Autistic Spectrum Disorder. AIMS Public Health, 2020, 7(4): 792-803. doi: 10.3934/publichealth.2020061 |
[2] | Anastasia Stathopoulou, Georgios F. Fragkiadakis . Assessment of psychological distress and quality of life of family caregivers caring for patients with chronic diseases at home. AIMS Public Health, 2023, 10(2): 456-468. doi: 10.3934/publichealth.2023032 |
[3] |
Dorota Zarnowiecki, Meaghan S Christian, James Dollman, Natalie Parletta, Charlotte E.L Evans, Janet E Cade .
Comparison of school day eating behaviours of 8–11 year old children from Adelaide, South Australia, and London, England . AIMS Public Health, 2018, 5(4): 394-410. doi: 10.3934/publichealth.2018.4.394 |
[4] | Casey T. Harris, Kevin Fitzpatrick, Michael Niño, Priya Thelapurath, Grant Drawve . Examining disparities in the early adoption of Covid-19 personal mitigation across family structures. AIMS Public Health, 2022, 9(3): 589-605. doi: 10.3934/publichealth.2022041 |
[5] | Christos Sikaras, Maria Tsironi, Sofia Zyga, Aspasia Panagiotou . Anxiety, insomnia and family support in nurses, two years after the onset of the pandemic crisis. AIMS Public Health, 2023, 10(2): 252-267. doi: 10.3934/publichealth.2023019 |
[6] | Waled Amen Mohammed Ahmed, Sara Boutros Shokai, Insaf Hassan Abduelkhair, Amira Yahia Boshra . Factors Affecting Utilization of Family Planning Services in a Post-Conflict Setting, South Sudan: A Qualitative Study. AIMS Public Health, 2015, 2(4): 655-666. doi: 10.3934/publichealth.2015.4.655 |
[7] | Erin Nolen, Catherine Cubbin, Mackenzie Brewer . The effect of maternal food insecurity transitions on housing insecurity in a population-based sample of mothers of young children. AIMS Public Health, 2022, 9(1): 1-16. doi: 10.3934/publichealth.2022001 |
[8] | J. Nwando Olayiwola, Melanie Raffoul . Saving Women, Saving Families: An Ecological Approach to Optimizing the Health of Women Refugees with S.M.A.R.T Primary Care. AIMS Public Health, 2016, 3(2): 357-374. doi: 10.3934/publichealth.2016.2.357 |
[9] | Dominique Meekers , Raseliarison Ratovonanahary , Tokinirina Andrianantoandro , Hiangotiana Randrianarisoa . Using Survey Data to Identify Opportunities to Reach Women with An Unmet Need for Family Planning: The Example of Madagascar. AIMS Public Health, 2016, 3(3): 629-643. doi: 10.3934/publichealth.2016.3.629 |
[10] | Argyro Pachi, Maria Anagnostopoulou, Athanasios Antoniou, Styliani Maria Papageorgiou, Effrosyni Tsomaka, Christos Sikaras, Ioannis Ilias, Athanasios Tselebis . Family support, anger and aggression in health workers during the first wave of the pandemic. AIMS Public Health, 2023, 10(3): 524-537. doi: 10.3934/publichealth.2023037 |
According to the World Health Organization (WHO), an increasing number of glaucoma patients have been reported worldwide in recent years [1]. Glaucoma prevention and treatment has been a major focus of international directives including the WHO's Vision 2020 campaign. It is estimated that the number of people with glaucoma is expected to rise from 64 million to 76 million in 2020 and 111.8 million in 2040, with Africa and Asia being affected more heavily than the rest of the world as there is a shortage of trained ophthalmologist for its diagnosis [2,3]. Hence, this disease is considered as a major public health concern, and its early diagnosis is important for preventing blindness.
Glaucoma is the second most common cause of irreversible vision loss and its diagnosis is a challenging research field because real world glaucoma images are acquired in the presence of several factors such as, illumination changes, background interference, light variation, etc [4]. With the rapid advancement in imaging technologies, several retinal imaging modalities like Heidelberg Retina Tomography (HRT) and Optical Coherence Tomography (OCT) have been developed. Although, these are used in developed and under-developed countries for image based diagnosis of various ocular diseases, i.e., Macular degeneration, Diabetic retinopathy and Glaucoma. However, these are costlier technologies and are not affordable for smaller public health units. Therefore, the most popular and widely used imaging device is the Fundoscopy [5].
Fundoscopy enables ophthalmologist to examine the Optic Nerve Head (ONH) for the glaucoma diagnosis and a typical image of ONH is shown in Figure 1. There are various parts of ONH which can be considered to classify between normal and glaucoma eyes. Four main changes can be observed in ONH associated with glaucoma, including ONH cupping, Neuroretinal rim thinning, Nerve Fibre Layer (NFL) thickness and Parapapillary Atrophy (PPA). These changes are detected manually through analysis of ONH images to diagnose glaucoma [6]. However, identification of glaucomatous signs in ONH require specialist ophthalmologists with years of experience and practice. Therefore, the development of automatic glaucoma assessment algorithms based on fundus image analysis, will be very helpful in reducing overall workload of ophthalmologists and also make the diagnosis more feasible and efficient even in smaller health units.
Recently, Deep CNNs have attracted a lot of interest and are seen to have great potential for the solution of computer vision tasks like image classification and semantic segmentation [7]. These networks are more robust to discover meaningful features in the images that are usually ignored in conventional image processing techniques. Moreover, different intermediate steps such as feature extraction and selection are embedded within the networks and these networks can perform feature learning and classification, simultaneously. Due to their popularity, numerous algorithms have been developed during the past few years on the detection and classification for glaucomatous fundus images. However, to train deep CNNs a large amount of annotated data is required [8,9]. Therefore, a lot of research effort is being put on the training methodology of the networks. We can divide these methodologies into two categories, i.e., training a deep network from scratch (full training) and transfer learning with fine-tuning.
Training from scratch requires large amount of labeled data which is extremely difficult to find in medical imaging. It is expensive to collect images both in terms of time and budget for disease identification. Besides, the training of Deep CNNs is time consuming that usually requires extensive memory and computational resources. Furthermore, the designing and adjustment of the hyper-parameters are the challenging tasks with reference to over-fitting and other issues. On the contrary, transfer learning with fine-tuning [10,11,12] is the easiest way to overcome such problems. The transfer learning is commonly defined as the ability of the network to exploit the knowledge learned in one domain, to another that share some common characteristics. Therefore, it is more popular method in machine learning and data mining for regression, clustering, and classification problems.
In an ensemble framework, various classifiers are trained to solve the same problem. The classification capability of an individual optimized Deep CNN is not generalized for diverse databases. Generally, the classification performance of grouping of various Deep CNNs is better than single architecture[13]. To make the most of the single classification capability, a promising solution would be to create an ensemble of several Deep CNN models.
During the past few years, extensive research has been carried out for the development of automatic glaucoma assessment systems based on fundus image analysis through transfer learning [14,15,16]. In the ImageNet Large Scale Visual Recognition Challenge (ILSVRC), millions of labeled images selected from 1000 different classes have been successfully tested on different medical image analysis [17]. It is tested that CNNs outperform the previous implementations with computer aided screening systems for medical images. Two individual CNN architectures are used in [18] to segment the optic disc and optic cup to find the cup-to-disc ratio (CDR). In [19], the authors have developed a CNN architecture to automate the detection process of glaucoma. Another Deep CNN algorithm has been developed in [20] to detect glaucoma through extraction of different features with the combination of different classifiers, i.e., Random Forest, Support Vector Machine and Neural Network. Feature learning through deep learning algorithm from retinal fundus images has also proposed in [21] to detect glaucoma. The authors have considered CNN model to learn features with linear and nonlinear activation functions. In the study [22], authors have implemented a Glaucoma-Deep system. The deep-belief network has been used to select the most discriminative features. In the work [23], two different types of CNNs, i.e., Overfeat and VGG-S, have been used as feature extractors. Contrast-Limited Adaptive Histogram Equalization (CLAHE) and vessels deletion have considered to investigate the performance of these networks. The authors have proposed a joint segmentation of optic disc, optic cup and glaucoma prediction in [24]. CNN feature sharing for different tasks ensured better learning and over-fitting prevention. Inception-v3 architecture has presented in [25] to detect glaucomatous optic neuropathy. Local space average color subtraction has been applied in pre-processing to accommodate for varying illumination. A framework on a dataset of fundus images collecting from different hospitals has presented in [26] by incorporating both domain knowledge and feature learned from a deep learning model. The assessment of deep learning algorithms with transfer learning have also been addressed in [27,28], with greater number of images than previous methods with high accuracy, sensitivity and specificity. Recently, in [29], authors have implemented deep learning based segmentation to identify glaucomatous optic disc. In [30], the authors have combined CNN and Recurrent Neural Network (RNN) to extract the spatial features in a fundus image and also the temporal features embedded in a fundus video, i.e., sequential images. ResNet50 deep CNN model has been explored with 48 full convolutional neural network layers. A deep learning approach based on deep residual neural network (ResNet101) for automated glaucoma detection using fundus images is proposed in [31]. Chronic eye disease diagnosis using ensemble-based classifier has presented in [32]. This study tends to achieve an early and accurate diagnosis of glaucoma based on an ensemble classifier by integrating the principal component analysis with rotation forest tree. Ensemble learning based CNN has been proposed in [33] to segment retinal images. The output of the classifier is subject to an unsupervised graph cut algorithm followed by a convex hull transformation to obtain the final optic cup and disc segmentation. A novel disc-aware ensemble network based on the application of different CNNs is presented in [34] for automatic glaucoma screening. The authors have introduced a deep learning technique to gain additional image-relevant information and screen glaucoma from the fundus image directly. More recently, a multi-class multi-label ophthalmological disease detection using transfer learning has been investigated in [35] with four pre-trained Deep CNN architectures. ResNet, InceptionV3, MobileNet, and VGG16 are implemented to detect eight types of ocular diseases. Besides, in [36], an artificial intelligence and telemedicine based screening tool has been developed to identify glaucoma suspects from color fundus Images. An ensemble of five pre-trained Deep CNN architectures is presented to detect glaucoma on the basis of cup to disc ratio (CDR). Two Xception architectures, InceptionResNetV2, NasNet, and InceptionV3 are used in the proposed ensemble to calculate the final CDR value for glaucoma detection.
It is evident from the above literature that almost all previously proposed methods employ CNN for the detection of glaucoma. However, limited work has been carried out to ensemble Deep CNNs for glaucoma diagnosis.
The research contributions of presented work are summarized as follows:
$ 1) $ We propose a new two-staged scheme based on Deep CNNs architectures for glaucoma classification using fundus images. This new scheme comprises four Deep architectures, i.e., AlexNet, InceptionV3, InceptionResNetV2 and NasNet-Large obtained through extensive experimental results search. We also assess their individual performance on both publicly and local hospital datasets. Our results clearly demonstrate the effectiveness of the newly proposed scheme based on Deep CNN architecture.
$ 2) $ We propose a new ensemble framework for better glaucoma classification. Four different pre-trained Deep CNN models are fused together in parallel and the output scores / probabilities are optimized using five different voting techniques to find the best Deep CNNs combination. We also propose two novel voting techniques to achieve much better results as compared to existing state-of-the-art for glaucoma classification.
The rest of the paper is organized as follows: the proposed methodology is presented in Section 2. Experiments and results are given in Section 3. Results related discussion is presented in Section 4, while conclusion and future work are summarized in Section 5.
The proposed automatic glaucoma diagnosis using fundus images is shown in Figure 2. There are five steps: 1) data collection; 2) pre-processing; 3) Deep CNN feature learning; 4) Ensemble ($ 4x $ Deep CNNs); 5) classification/diagnosis. These steps are explained in the following subsections:
Both public and private datasets are used to train, validate and test the different Deep CNNs. In this way, we can create the diversity in the images and the results will be generalized. The description about each dataset is given as follows:
ACRIMA is the most newly available public dataset for the classification of glaucoma through Deep learning. It consists of 705 fundus images (309 normal and 396 glaucomatous). All images are annotated by two glaucoma experts with eight years of experience. It is only be used for classification tasks because optic disc and optic cup annotations are not provided [37].
The ORIGA-light is another public dataset which is annotated by experienced professionals from Singapore Eye Research Institute. There are 650 fundus images (482 normal and 168 glaucomatous) with a resolution of $ 3072 \times 2048 $ pixels. This dataset is widely used as a benchmark for the diagnosis of glaucoma [38].
The RIM-ONE is a very popular publicly available dataset for ONH segmentation and glaucoma detection. The database was created by collaboration of three Spanish hospitals, i.e., Hospital Universitario de Canarias, Hospital Clínico San Carlos and Hospital Universitario Miguel Servet. There are 455 fundus images, 261 normal and 194 glaucoma [39].
Another 124 fundus images, are collected from local private hospital, i.e., Armed Forces Institute of Ophthalmology (AFIO), Military Hospital, Rawalpindi, Pakistan. Similarly, 55 images are also acquired from local private hospital, i.e., Hayatabad Medical Complex (HMC), Peshaware, Pakistan. These images are annotated by two glaucoma experts with ten years of experience in glaucoma department. The number of normal and glaucoma images in each database are listed in Table 1.
Dataset | Normal | Glaucoma | Total |
ACRIMA | 309 | 396 | 705 |
ORIGA-light | 482 | 168 | 650 |
RIM-ONE | 261 | 194 | 455 |
AFIO | 85 | 39 | 124 |
HMC | 40 | 15 | 55 |
1177 | 812 | 1989 |
The images of different datasets are shown in Figure 3. The difference between normal and glaucoma images are also presented in Figure 4. The first row shows the normal while the second row is the glaucoma images.
The data pre-processing steps include the image patch extraction centered at ONH and data augmentation for the training of Deep CNNs.
First of all, we process all the images into a standard format that are used to train Deep CNNs. Image patches, centered at ONH, are extracted at same size according to the requirement of different Deep CNNs. Bicubic interpolation [40] is considered for resizing. It is examined that output pixel values are weighted average of pixels in 4-by-4 neighborhood pixels. The fundus images are cropped by evaluating the bounding box of 1.5 times the optic disc radius. Meanwhile, the illumination and contrast enhancement procedures are avoided to make the Deep CNNs learning more dynamic. In case of images from the local hospitals, i.e., AFIO and HMC, we have to localize ONH at the center of fundus images. It is to note that glaucoma disease mainly affects the ONH and its surrounding area. The cropping of images around ONH turned out to be more effective as compared to whole image, used for the training of Deep CNNs [41]. Moreover, the computational cost is also reduced during network learning.
The data augmentation technique is also explored during training of Deep CNNs to increase the training images and minimize over-fitting. The fundus images are invariant for flipping, rotation and translation. Hence, these three steps have been considered to increase the data for training of Deep CNNs.
Currently, Deep CNNs are applied to a wide variety of applications in image segmentation and classification tasks. In this work, we also implement ImageNet trained Deep CNNs for the diagnosis of glaucoma through retinal fundus images. We have explored four types of Deep CNNs, i.e., AlexNet, InceptionV3, InceptionResNetV2 and NasNet-Large for the classification of normal and glaucomatous images. The basic architecture as well as feature extraction scheme of each network are illustrated in the forthcoming paragraphs.
AlexNet, first proposed in [42], is an extremely powerful model in achieving high accuracies on the challenging databases. It is very similar architecture as LeNet [43] but much deeper with more filters per layer with stacked CNNs layers. In this model, there are 5 convolutional and 3 Fully Connected (FC) layers with max pooling, Rectified Linear Unit (ReLu), and dropout layers. The input to this model is an RGB image of size $ 256 \times 256 $. The basic structure of this network is shown in Figure 5.
InceptionV3 [44] is an extended network of the GoogLeNet [45] with good classification performance in several biomedical applications with transfer learning [46,47]. InceptionV3 originally contains 11 Inception modules and one FC layer. Each Inception module also has 4 to 10 multiscale convolutional operations conducting with $ 1\times1, 3\times3 $, or $ 5\times5 $ filters. Max-pooling is used as the spatial pooling operation in the inception modules and FC is for final classification. This design reduces the computational complexity as well as number of parameters to be trained. It is trained on more than a million images from the ImageNet database. The network has an image input size of $ 299\times299 $. The fundamental architecture of InceptionV3 is illustrated in Figure 6.
We also consider the state-of-the-art InceptionResnetV2 network [48] to extract deep spatial features. It is a combination of two recent networks, one is Residual Connections [49] and another is Inception structure [44]. The engaged InceptionResNetV2 network includes Stem, InceptionResNet, Reduction layers, average pooling layer and a FC layer. The Stem includes preliminary convolution operations executed before entering the Inception blocks. The InceptionResNet layers have residual connections with convolution operations while the Reduction layers are responsible for changing the width and height of the image. The spatial features are extracted from the convolutional layers and the pooling layers decrease the dimensionality of individual feature map, but hold the most significant features. Furthermore, the convolutional layers are followed by the batch normalization layer and ReLU, which is a nonlinearity function and helped to decrease the training time. The network has an image input size of $ 299\times299 $. Figure 7 illustrates the architecture of deep spatial feature extraction using InceptionResNetV2 network.
Neural Architecture Search Network Large (NASNet-Large) is basically composed of two kinds of layers or cells, i.e., Normal and Reduction Cells. During the forward path, the width and height of feature map is reduced half by Reduction Cell while the Normal Cell retain these two dimensions same as the input feature map. The Normal Cells are stacked between Reduction Cells as shown in Figure 8. Each cell in Normal / Reduction Cell is composed of a number of blocks. Each block is built from the set of popular operations in Deep CNNs with various kernel size e.g.: convolutions, max pooling, average pooling, dilated convolution, depth-wise separable convolutions. Finding best architecture for Normal Cell and Reduction Cell with 5 blocks is described in [50]. NASNet-Large with N equals to 6 aims to get maximum possible accuracy. With the help of this arrangement, the network is able to learn rich feature representations for a wide range of images [51]. This network has an image input size of $ 331 \times 331 $.
It is to note that the inputs in the above mentioned Deep CNN architectures are pre-processed color fundus images with centered at ONH. The region of interest is also kept same for all the models while the last FC layer is modified with softmax classifier to study only two classes, i.e., normal and glaucoma.
In case of an ensemble framework, a set of diverse classifiers are aggregated and are trained to solve the same problem. All the above mentioned architectures are combined in one classifier as presented in Figure 9. The final decision is given by any of the five voting techniques, i.e., Majority Voting (MV), Proportional Voting (PV), Averaging (AV), Accuracy based Weighted Voting (AWV), and Accuracy/Score based Weighted Averaging (ASWA).
In MV [52], each model makes a prediction (vote) for each test instance and the final output prediction is the one that receives more than half of the votes. Here $ V_{ij} $ is the vote for $ jth $ class with reference to $ ith $ classifier and $ N(V_{j}) $ is the total number of votes for $ jth $ class. We predict the class label, $ O $, via highest number of votes. Mathematically, it is written as
$ N(Vj)=N∑i=1Vij $
|
(2.1) |
$ O=Max{N(V1),N(V2),.............,N(Vj)} $
|
(2.2) |
where $ N(V_{1}) $, $ N(V_{2}) $, and $ N(V_{3}) $ are the total number of votes for class 1, class 2, and class 3, respectively.
In PV, the training accuracy of each classifier are summed-up with respect to their prediction and the maximum result will be the final outcome. $ A_{ij} $ is the accuracy for $ jth $ class with respect to $ ith $ classifier and $ T(A_{j}) $ is the sum of training accuracies for $ jth $ class. We can write as
$ T(Aj)=N∑i=1Aij $
|
(2.3) |
$ O=Max{T(A1),T(A2),.............,T(Aj)} $
|
(2.4) |
where $ T(A_{1}), T(A_{2}) $, and $ T(A_{3}) $ are the sum of training accuracies for class 1, class 2, and class 3, respectively.
However, in case of equal votes, the AV is considered. The score vector for each predicted class are summed-up and averaged. $ S_{ij} $ is the score for $ jth $ class according to $ ith $ classifier and $ S_{j} $ is the average score for $ jth $ class. The output class, $ O $, is the one corresponding to the highest value, such as
$ Sj=1NN∑i=1Sij $
|
(2.5) |
$ O=Max{S1,S2,.............,Sj} $
|
(2.6) |
where $ S_{1}, S_{2} $, and $ S_{3} $ are the averaged score for class 1, class 2, and class 3, respectively.
In AWV, accuracy of each classifier is modified according to the following relation,
$ αi=e−10(1−ACCi)∑ki=1e−10(1−ACCi) $
|
(2.7) |
where $ \alpha_{i} $ is the updated accuracy of $ i $th classifier, i.e., $ ACC{i} $. The final decision, $ O $, is assigned to the maximum value, given as
$ (AWV)j=N∑i=1αi×Wij $
|
(2.8) |
$ O=Max{(AWV)1,(AWV)2,.............,(AWV)j} $
|
(2.9) |
where $ W_{ij} $ is the weight of $ jth $ class with reference to $ ith $ classifier and $ (AWV)_{1}, (AWV)_{2} $, and $ (AWV)_{3} $ are the accuracy based weighted votes for class 1, class 2, and class 3, respectively.
In ASWA, the probabilities / Scores are used to calculate the weighted average based on accuracy. Mathematically, it is written as
$ (ASWA)j=N∑i=1αi×Sij $
|
(2.10) |
$ O=Max{(ASWA)1,.............,(ASWA)j} $
|
(2.11) |
where $ \alpha_{i} $ is already calculated from Eq (2.7), and $ S_{ij} $ is the probability / score of $ jth $ class with respect to $ ith $ classifier. The final output decision is evaluated according to the Eq (2.11). $ (ASWA)_{1}, (ASWA)_{2} $, and $ (ASWA)_{3} $ are the accuracy/score based weighted averaging of class 1, class 2, and class 3, respectively.
The classification task is performed by the Softmax layer. During training a network, this layer updates the weights through back propagation. This process is based on the loss function used in the training stage. Keeping in view the underlying binary classification problem, the Cross Entropy (C.E) has been used as a loss function, as shown in Eq (2.12).
$ C.E=−C=2∑t(tilog(Si))=−t1log(S1)−(1−t1)log(1−S1) $
|
(2.12) |
where $ t_{i} $ and $ S_{i} $ are the ground truth and the Deep CNNs score for each class $ i $ in $ C $.
We have analyzed the fundus images in two stages. In stage one, the performance of each Deep CNN has been evaluated for glaucoma diagnosis. In stage two, all four Deep CNNs are grouped together as a single classifier to further improve the accuracy of the system. Five voting schemes with two newly proposed techniques are considered to calculate the final decision.
First, all fundus images are divided into three groups, training, validation and testing images. The sixty percent of all is for training while thirty percent of remaining for validation and rest for testing purpose. Table 2 shows the distribution of images in each group.
Class | Training | Validation | Testing |
Normal | 706 | 141 | 330 |
Glaucoma | 487 | 98 | 227 |
The validation group is selected to monitor the number of epochs in the training process of different deep architectures. All the images used in validation, shared with the training set after selecting hyper-parameters and cross validation experiments. Thus, the training set has 847 normal and 585 glaucoma images. But, the testing sets are kept same during all experiments. The distribution of test images in different datasets is presented in Table 3.
Dataset | Normal | Glaucoma |
ACRIMA | 87 | 110 |
ORIGA-Light | 135 | 47 |
RIM-ONE | 73 | 55 |
AFIO | 24 | 11 |
HMC | 11 | 4 |
Commonly, the transfer learning strategy is applied in Deep CNNs to use the knowledge learned while classifying natural images to classify retinal images with glaucoma. Hence, the transfer learning has been employed in this work that involves, replacing and retraining the softmax layer and also fine-tuning the weights of the pre-trained network. We have carried out several experiments to achieve the optimal performance of each Deep CNNs with different number of fine-tuned layers and number of epochs. Initially, we have considered the last weighted layers of Deep CNNs for the number of fine-tuned layers, while keeping the initial layers in a freezing state. After that, the number of fine-tuned layers is increased until updating all the remaining layers in the Deep CNNs. Secondly, the effect of number of epochs have been evaluated for the best performance of each model. It is to note that we get maximum performance for each Deep CNNs for $ 30 $ epochs. The Stochastic Gradient Descent (SGD) is used as the optimizer. The other hyper-parameters, the learning rate, the batch size, and the momentum are optimally selected to get the best results in different sets of experiments. The objective of optimal setting is to adjust the weights of each Deep CNN such that the training loss is minimum. By minimizing the loss, we can achieve the optimal parameters resulting in the best model performance. Batch size is the number of training examples used in the estimation of error gradient for the learning algorithm. We take its value smaller because smaller batch size make it easier to fit one batch worth of training data in memory, offering a regularizing effect and lower the generalization error. The learning rate decides how far to move the weights in the direction of gradient to get the minimum loss. However, an optimal value is required to reach a minimum loss quickly. This is because smaller learning rate will take tiny steps to and hence a large time is required to reach the minimum loss function. While, the higher learning rate will result in overshooting the minimum loss and the Deep CNN may not converge. Besides, momentum is another parameter used to optimize the learning rate. It aims to calculate the weighted average of weights between the average of previous values and the current value. Thus, we have set the batch size to $ 6 $, the learning rate to $ 1\times e^{-4} $ and the momentum to $ 0.9 $ for all the networks.
We have also assessed the performances of Deep CNNs using 10 cross validation technique. Due to limited training data, over-fitting is a well-known problem, occurred in Deep CNNs. To avoid over-fitting and increase the robustness of the architectures, we have pondered the dropout technique as proposed in [53] that temporally remove units along with all its incoming and outgoing connections in deep neural networks. Similarly, we have also employed data augmentation technique during training of all the networks. The fundus images are augmented by using random rotations between $ 0 $ to $ 360 $ degrees, and random translation of maximum $ 10 $ pixels in both $ x $ and $ y $ direction of the image. The input images are also resized according to the default input size of each Deep CNNs. Hence, we have evaluated the performance of each Deep CNNs using all datasets and experiments have been carried out to compare the best of the above mentioned four Deep CNNs. The training and testing procedures are displayed in Figures 10 and 11, respectively.
The number of epochs has been evaluated for each of the selected Deep CNN during the training process. The validation accuracy/loss during fine-tuning process of all four Deep CNNs for ACRIMA dataset are illustrated in Figure 12. It is observed that after 30 epochs, the validation accuracy reaches its maximum values, i.e., 99.10, 99.28, 94.72 and 100% for AlexNet, InceptionV3, InceptionResNetV2 and NasNet-Large, respectively. Similarly, the validation accuracy/loss results for other Deep CNNs are also evaluated.
It is observed that the NasNet-Large has maximum accuracy, i.e., 100% for ACRIMA dataset while for ORIGA-Light it has minimum value, i.e., 87.10% as compared with other networks. On average, all Deep CNNs perform well on ACRIMA dataset and the lowest results have been considered on ORIGA-Light dataset. This is because ACRIMA is a newly developed dataset for classification of glaucoma images while ORIGA-Light is designed for the segmentation of optic cup and optic disc. The superiority of the NasNet-Large model has also been observed during training for all other datasets. The training accuracy based comparison among $ 4x $ Deep CNNs for ACRIMA dataset is illustrated in Figure 13. It is investigated that the NasNet-Large outperforms all other networks during training. It is also applicable for other datasets.
Furthermore, Deep CNNs training time has been reduced through transfer learning. The weights in pre-trained networks are used as the starting point for the training process. Hence, we have minimized the training epochs, i.e., 30 for all Deep CNNs and achieved best validation results, as displayed in Table 4.
Deep CNNs | ACRIMA | ORIGA-Light | RIM-ONE | AFIO | HMC |
AlexNet | 99.10 | 75.50 | 86.70 | 90.90 | 98.50 |
InceptionV3 | 99.28 | 78.50 | 86.40 | 84.50 | 90.50 |
InceptionResNetV2 | 94.72 | 76.50 | 90.60 | 85.50 | 95.50 |
NasNet-Large | 100.00 | 87.10 | 94.70 | 97.70 | 98.60 |
Commonly, a single evaluation metric is not appropriate to evaluate the performance of a given algorithm due to the presence of some imbalanced classes in the dataset or a large number of training labels [54]. Therefore, the performance of Deep CNNs are reported in terms of five distinct metrics including Accuracy (ACC), Sensitivity (SEN), Specificity (SP), F1 score and Area Under the Curve (AUC) as proposed in the previous studies [55]. These performance parameters are calculated using the following equations:
$ ACC=TP+TNTP+FP+TN+FN $
|
(3.1) |
$ SEN=TPTP+FN $
|
(3.2) |
$ SP=TNTN+FP $
|
(3.3) |
$ F1=2precision.recallprecision+recall $
|
(3.4) |
where the precision and recall are expressed as
$ Precision=TPTP+FP $
|
(3.5) |
$ Recall=TPTP+FN $
|
(3.6) |
In the above equations, the True Positive (TP) is defined as the number of glaucoma images classified as glaucoma and True Negative (TN) is the number of normal images classified as normal. False Positive (FP) is the number of normal images identified as glaucoma images and False Negative (FN) is the number of glaucoma images classified as normal.
AUC is the area under the Receiver Operating Curve (ROC) and it provides the probability that the model ranks a positive example more highly than a negative example. ROC is a plot between two parameters, i.e., True Positive Rate (TPR) and False Positive Rate (FPR). TPR is synonym for recall while FPR can be calculated as
$ FPR=FPFP+TN $
|
(3.7) |
The confusion matrices of each Deep CNNs have been evaluated for the test images in each dataset. Figures 14 and 15 show the confusion matrices for ACRIMA test images and total test images, respectively. Similarly, for other test images, ACC, SEN and SP results are also calculated according to Eqs (3.1), (3.2) and (3.3).
It is observed that NasNet-Large achieves best results over ACRIMA dataset, i.e., ACC (99.5%), SP (100%) and SEN (99%). The InceptionResNetV2, gives ACC (99%), SP (97.7%) and SEN (100%). Now for ORIGA-Light test images, it can be seen that all the Deep CNNs show poor performance except NasNet-Large with ACC (88%), SP (91%) and SEN (79%). The AlexNet, InceptionV3, and InceptionResNetV2 achieve worst results in the range of 60–66% SEN.
All the Deep CNNs perform well on RIM-ONE test images. Like, NasNet-Large again gives better results in terms of ACC (94.5%), SP (96%) and SEN (92.7%). While, the AlexNet shows lowest results with ACC (87.5%), SP (90.4%) and SEN (83.6%) as compared with other networks.
The performance metrics are also evaluated for both the local datsets, i.e., AFIO and HMC. It is noticed that NasNet-Large provides maximum results in terms of ACC, SP, and SEN for AFIO test images while InceptionResNetV2 has 88.6% ACC, 83.3% SP, and 100% SEN. Similarly, NasNet-Large gives maximum results, i.e., 100% in all performance metrics while AlexNet has 93.3% ACC, 100% SP, and 75.0% SEN with HMC test images.
In case of total test set of images, the NasNet-Large again performs well as compared to other networks. It provides 99.3% ACC, 99.4% SP, and 99.1% SEN, while InceptionV3 shows lowest results in terms of ACC, SP, and SEN. These results are displayed in Table 5.
Deep CNNs | Datasets | ||||||||
ACRIMA | ORIGA-Light | RIM-ONE | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 99.5 | 98.9 | 100.0 | 75.8 | 81.5 | 60.0 | 87.5 | 90.4 | 83.6 |
InceptionV3 | 98.5 | 97.7 | 99.1 | 78.6 | 83.0 | 66.0 | 92.2 | 94.5 | 89.1 |
InceptionResNetV2 | 99.0 | 97.7 | 100.0 | 76.9 | 82.2 | 61.7 | 90.6 | 94.5 | 85.5 |
NasNet-Large | 99.5 | 100.0 | 99.1 | 87.9 | 91.1 | 78.7 | 94.5 | 95.9 | 92.7 |
AFIO | HMC | TOTAL | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 91.4 | 95.8 | 81.8 | 93.3 | 100.0 | 75.0 | 97.3 | 97.3 | 97.4 |
InceptionV3 | 91.4 | 91.7 | 91.0 | 86.7 | 100.0 | 50.0 | 96.6 | 97.0 | 96.0 |
InceptionResNetV2 | 88.6 | 83.3 | 100.0 | 93.3 | 91.0 | 100.0 | 97.1 | 97.3 | 96.9 |
NasNet-Large | 94.3 | 91.7 | 100.0 | 99.9 | 100.0 | 100.0 | 99.3 | 99.4 | 99.1 |
It is noted that the classification accuracies have been improved using an ensemble of all four Deep CNNs. For ACRIMA dataset, NasNet-Large and AlexNet give 99.5% accuracy while ensemble with AWV provides 99.6% accuracy. In case of ORIGA-Light, the accuracy has been increased from 87.9 to 88.3% with ASWA. Similarly, for RIM-ONE, AFIO, and HMC datasets, the results have been improved. Now, if we consider total test set of images, the result is increased from 99.3 to 99.5% with ASWA. The AV provides lowest results. This is because it is simple averaging. However, the AWV and ASWA provide better results. This is because the weights are updated according to the accuracy and score of each Deep CNNs. The results of five voting schemes are displayed in Table 6. Figure 16 also provides an overview of newly developed AWV and ASWA schemes that makes it different from MV, PV, and AV.
Dataset | MV | PV | AV | AWV | ASWA |
ACRIMA | 99.5 | 99.4 | 99.1 | 99.6 | 99.5 |
ORIGA-Light | 87.9 | 88.0 | 79.8 | 88.2 | 88.3 |
RIM-ONE | 94.5 | 94.5 | 91.2 | 95.1 | 95.2 |
AFIO | 94.3 | 94.3 | 91.4 | 96.1 | 96.2 |
HMC | 99.9 | 99.9 | 96.3 | 99.8 | 99.9 |
Total | 99.3 | 99.3 | 98.5 | 99.4 | 99.5 |
Besides, we have also considered the combinations of two and three Deep CNNs to increase the persuasiveness of the proposed ensemble framework. The results of different voting schemes with ensemble of $ 2x $, $ 3x $, and the proposed $ 4x $ Deep CNNs for total test set of images are presented in Table 7. It is observed that the AlexNet and NasNet-Large provide better results as comapred to other networks in $ 2x $ Ensemble framework. While, in the case of $ 3x $ Ensemble network, AlexNet, InceptionResNetV2 and NasNet-Large show superior results with other combinations. However, our proposed $ 4x $ Ensemble framework outperforms in all five voting schemes as compared to $ 2x $, and $ 3x $ Deep CNNs ensemble.
Deep CNNs | MV | PV | AV | AWV | ASWA | |
$ 2x $ Ensemble | A+B | 98.70 | 98.60 | 97.80 | 99.00 | 98.80 |
A+C | 98.80 | 98.50 | 97.60 | 99.00 | 98.90 | |
A+D | 99.00 | 98.90 | 97.80 | 99.20 | 99.00 | |
B+C | 98.80 | 99.00 | 96.90 | 98.90 | 98.80 | |
B+D | 98.80 | 98.70 | 97.50 | 98.90 | 98.90 | |
$ 3x $ Ensemble | A+B+C | 98.90 | 98.80 | 98.10 | 98.80 | 98.90 |
A+B+D | 98.80 | 99.00 | 98.10 | 98.80 | 99.00 | |
A+C+D | 99.00 | 99.10 | 98.20 | 99.00 | 99.00 | |
B+C+D | 98.90 | 99.10 | 98.10 | 98.80 | 98.90 | |
$ 4x $ Ensemble | A+B+C+D | 99.30 | 99.30 | 98.50 | 99.40 | 99.50 |
A = AlexNet, B = InceptionV3, C = InceptionResNetV2, & D = NasNet-Large |
The evaluation parameters of classification performance, i.e., sensitivity, specificity, accuracy, and AUC of the ImageNet trained Deep CNNs have been displayed in Table 8, where the performance comparison of proposed work with [22,26,37], and [56,57,58,59,60,61] is presented. In [57,58,59,60], the authors have used CNN based architectures for the classification of glaucoma using RIM-ONE, and ORIGA-Light images. Similarly, in another study proposed by [61], the authors have considered newly developed ACRIMA dataset for the glaucoma classification and achieved highest 96.0% AUC. While, in our study, the NasNet-Large performed well and gives 99.1% sensitivity, 99.4% specificity, 99.3% accuracy and 97.8% AUC for the total test set of images. This is because the cropped images centered at ONH are to be more effective as compared to whole image as well as computational cost is reduced during network learning. It is also helpful in improving the identification of glaucomatous damages in early stages. Furthermore, a data augmentation technique is also considered during training of Deep CNNs to increase the training images and minimize over-fitting problem. The classification accuracy has been further increase from 99.3 to 99.5% with ensemble framework. The results show that AWV and ASWA provide better results as compared with other voting techniques. This is because weights are updated according to accuracy and scores of each Deep CNNs. These results indicate that the proposed study provides better performance than its previous state-of-the-art. To the best of our knowledge, there is no existing related Deep CNNs based ensemble framework for the diagnosis of glaucoma using fundus images.
Systems | Methods | Dataset | Sensitivity (%) | Specificity (%) | Accuracy (%) |
[22] | CNN | Public & Private | 84.5 | 98.0 | 99.0 |
[26] | Multi-branch NN | Private | 92.0 | 90.0 | 91.0 |
[37] | Deep CNNs | Public | 93.4 | 85.8 | 96.0 (AUC) |
[56] | CNN | Private | 98.0 | 98.0 | 98.0 |
[57] | CNN | RIM-ONE | 80.0 | 88.0 | 85.0 |
[58] | RCNN | ORIGA-Light | NR | NR | 87.4 (AUC) |
[59] | GoogleNet | HRF & RIM-ONE | NR | NR | 87.6 |
[60] | AlexNet | RIM-ONE | 87.0 | 85.0 | 88.0 |
[61] | Deep Models | Public & Private | NR | NR | 85.0 (AUC)ACRIMA |
[62] | InceptionV3 | Private | NR | NR | 84.5 & 93.0 (AUC) |
[63] | Ensemble Classifier | Private | 86.0 | 90.0 | 88.0 & 94.0 (AUC) |
Our | NasNet-Large$ \ast $ | Public & Private | 99.1 | 99.4 | 99.3 & 97.8% (AUC) |
Our | Ensemble Classifier$ \ast $ | Public & Private | 99.1 | 99.7 | 99.5 |
The experimental results presented in this study suggest the following key observations:
● From the above mentioned results, the proposed automatic glaucoma diagnosis system is more useful and effective. This is because Deep CNNs have the ability to learn glaucoma specific features automatically. While in traditional methods, the feature extraction strategies are manual that limit the success of overall system. Moreover, the illumination and textural variations in retinal fundus images are also the problems in classification of glaucoma. Conversely, the presented work uses automatic glaucoma features extraction and thus achieve superior classification accuracy across a wide range of publicly and locally available datasets. Hence, the results are generalized for diverse set of images.
● In pre-processing step, the region ONH has been extracted and used as the input image to Deep CNNs models. This is due to the fact that most of the initial changes have been occurred in ONH during early stage of glaucoma. Hence, the cropped images centered at ONH are to be more effective as compared to whole image as well as computational cost is reduced during network learning. It is also helpful in improving identification of glaucomatous damages in early stages. Furthermore, a data augmentation technique is also considered during training of Deep CNNs to increase the training images and minimize over-fitting problem.
● Transfer learning generally refers to a process where a model trained on one problem is used in some way on a second related problem. In our study, Deep CNNs training time has been reduced through transfer learning. The weights in pre-trained networks are used as the starting point for the training process. Hence, we have minimized the training epochs, i.e., 30 for all Deep CNNs and achieved best validation results, as displayed in Table 4. Additionally, a graphical comparison is also being presented among these models during the training process. It is examined that our proposed model, i.e., NasNet-Large achieves best results as compared to other Deep CNNs. These results are displayed in Figure 13.
● Generally, a single performance metrics can lead to inappropriate classification results due to some imbalance classes in the dataset or too small or large number of training subjects. From the literature survey of the existing methods on fundus images such as [37,58,61] show classification performance in terms of AUC only. In contrast, we have evaluated four distinct metrics including SEN, SP, ACC, and AUC. The results show the steady performance in glaucomatous classification across different metrics.
● We have also performed extensive experiments to evaluate the performance of four Deep CNNs (AlexNet, InceptionV3, InceptionResNetV2 and NasNet-Large). Comparing the results of these architectures, it is noted that NasNet-Large is significantly better than that of others networks. The results are consistent with the relative performance of these architectures on wide range of public and private fundus images. The AlexNet also provides better results as compared with other networks. However, it shows lower performance on ORIGA-Light and RIM-ONE datasets as compared with others. From the results, it is observed that ORIGA-Light is more challenging dataset for each type of network. On the contrary, all the networks provide good results for ACRIMA dataset because it is a newly developed dataset for Deep learning classification tasks.
● We have also presented an ensemble classifier to further improve the classification accuracy. The final results have been evaluated with five voting techniques. Two newly developed voting schemes, i.e., AWV and ASWA provide the better results as compared with others as presented in Table 6. Moreover, we have also carried out ablation experiments to increase the persuasiveness of the experimental results as well as the validation of four Deep CNNs combination. The experimental results are displayed in Table 7. It is clearly observed that ensemble of four networks show better results as compared to ensemble of $ 2x $ and $ 3x $ networks. These results clearly demonstrate the effectiveness of the proposed ensemble framework to diagnose the glaucoma.
● The experimental results are also compared with other deep learning based diagnostic systems developed by other researchers. In [57,58,59,60], the authors have used CNN based architectures for the classification of glaucoma using RIM-ONE, and ORIGA-Light images. Similarly, in another study proposed by [61], the authors have considered newly developed ACRIMA dataset for the glaucoma classification and achieved highest 96.0% AUC. In [62], the authors have implemented InceptionV3 architecture for glaucoma detection and achieved maximum 84.5% accuracy with 93% AUC. More recently, an ensemble of AlexNet, ResNet-50, and ResNet-152 have investigated in [63] and achieved the highest accuracy of 88% with 0.94 AUC. However, our proposed ensemble framework have achieved superior results as compared with previously proposed methods. The results of SEN, SP, and ACC proposed in this study are displayed in Table 5. A detailed comparisons with existing state-of-the-art has been presented in Table 8.
In this work, we have proposed an ensemble framework based on pre-trained Deep CNNs for glaucoma classification using fundus images. At first, four Deep CNNs, i.e., AlexNet, InceptionV3, InceptionResNetV2, and NasNet-Large are tested on five different datasets, three publicly available, i.e., ACRIMA, ORIGA-Light, and RIM-ONE, and others two collected from the local hospitals. Dropout and data augmentation techniques are also considered to improve the performance of Deep CNNs models. NasNet-Large is the best option with transfer learning and fine-tuning, with AUC (97.8%), SEN (99.1%), SP (99.4%) and ACC (99.3%). Secondly, for even better results, we also proposed an ensemble framework for automatic glaucoma classification. The AWV and ASWA based ensembling methods improve the accuracy with all datasets and total test images to $ 0.3\% $. Moreover, the proposed ensemble classifier has considerably better accuracy and robustness than the individual optimized Deep CNN models for automatic glaucoma diagnosis.
As a future work, the new architectures with more data can be explored and assessed to confirm the presented line of work. The performance of Deep CNNs can also be enhanced to extract deep features for the classification tasks. In this way, we can train even more robust glaucoma classifiers.
We would like to mention here the great struggle of Dr. Yousaf Jamal Mahsood, Associate Professor, Glaucoma Khyber Girls Medical College, HMC Peshawar, in collecting the fundus images and AFIO department, Military Hospital, Rawalpindi, Pakistan.
The authors declare no conflict of interest.
[1] |
Ingersoll B, Schreibman L, Stahmer A (2001) Brief Report: Differential Treatment Outcomes for Children with Autistic Spectrum Disorder Based on Level of Peer Social Avoidance. J Autism Dev Disord 31: 343–349. doi: 10.1023/A:1010703521704
![]() |
[2] |
Farley MA, McMahon WM, Fombonne E, et al. (2009) Twenty-year outcome for individuals with autism and average or near-average cognitive abilities. Autism Res Off J Int Soc Autism Res 2: 109–118. doi: 10.1002/aur.69
![]() |
[3] | Association AP (2013) Diagnostic and Statistical Manual of Mental Disorders (DSM-5®). American Psychiatric Pub. |
[4] | Li Y, Yu D (2018) Variations of the Functional Brain Network Efficiency in a Young Clinical Sample within the Autism Spectrum: A fNIRS Investigation. Front Physiol 9. |
[5] | Levels of Autism: Symptoms and Outlook of Severity Levels 1, 2, and 3 Healthline. Available from: https://www.healthline.com/health/levels-of-autism. |
[6] |
Sherer MR, Schreibman L (2005) Individual behavioral profiles and predictors of treatment effectiveness for children with autism. J Consult Clin Psychol 73: 525–538. doi: 10.1037/0022-006X.73.3.525
![]() |
[7] | Cicchetti D, Curtis WJ (2015) The Developing Brain and Neural Plasticity: Implications for Normality, Psychopathology, and Resilience. Developmental Psychopathology. Wiley-Blackwell, 1–64. |
[8] |
Dawson G (2008) Early behavioral intervention, brain plasticity, and the prevention of autism spectrum disorder. Dev Psychopathol 20: 775–803. doi: 10.1017/S0954579408000370
![]() |
[9] | Baio J, Wiggins L, Christensen DL, et al. (2018) Prevalence of Autism Spectrum Disorder Among Children Aged 8 Years-Autism and Developmental Disabilities Monitoring Network, 11 Sites, United States, 2014. MMWR Surveill Summ 67: 1–23. |
[10] |
Medeiros K, Winsler A (2014) Parent-child gesture use during problem solving in autistic spectrum disorder. J Autism Dev Disord 44: 1946–1958. doi: 10.1007/s10803-014-2069-y
![]() |
[11] |
Elsabbagh M, Divan G, Koh YJ, et al. (2012) Global prevalence of autism and other pervasive developmental disorders. Autism Res Off J Int Soc Autism Res 5: 160–179. doi: 10.1002/aur.239
![]() |
[12] | Rice CE, Rosanoff M, Dawson G, et al. (2012) Evaluating Changes in the Prevalence of the Autism Spectrum Disorders (ASDs). Public Health Rev 34: 1–22. |
[13] |
Hansen SN, Schendel DE, Parner ET (2015) Explaining the increase in the prevalence of autism spectrum disorders: the proportion attributable to changes in reporting practices. JAMA Pediatr 169: 56–62. doi: 10.1001/jamapediatrics.2014.1893
![]() |
[14] | Blumberg SJ, Bramlett MD, Kogan MD, et al. (2013) Changes in prevalence of parent-reported autism spectrum disorder in school-aged U.S. children: 2007 to 2011–2012. National Center for Health Statistics Reports. Number 65. NCHS. |
[15] |
McConkey R, Truesdale-Kennedy M, Cassidy A (2009) Mothers' Recollections of Early Features of Autism Spectrum Disorders. Child Adolesc Ment Health 14: 31–36. doi: 10.1111/j.1475-3588.2008.00495.x
![]() |
[16] |
Zeedyk SM, Rodriguez G, Tipton LA, et al. (2014) Bullying of youth with autism spectrum disorder, intellectual disability, or typical development: Victim and parent perspectives. Res Autism Spectr Disord 8: 1173–1183. doi: 10.1016/j.rasd.2014.06.001
![]() |
[17] |
Chamak B, Bonniau B, Oudaya L, et al. (2011) The autism diagnostic experiences of French parents. Autism Int J Res Pract 15: 83–97. doi: 10.1177/1362361309354756
![]() |
[18] |
Desai MU, Divan G, Wertz FJ, et al. (2012) The discovery of autism: Indian parents' experiences of caring for their child with an autism spectrum disorder. Transcult Psychiatry 49: 613–637. doi: 10.1177/1363461512447139
![]() |
[19] | Gray DE (2003) Gender and coping: the parents of children with high functioning autism. Soc Sci Med 1982 56: 631–642. |
[20] |
Ludlow A, Skelly C, Rohleder P (2012) Challenges faced by parents of children diagnosed with autism spectrum disorder. J Health Psychol 17: 702–711. doi: 10.1177/1359105311422955
![]() |
[21] |
Davis NO, Carter AS (2008) Parenting stress in mothers and fathers of toddlers with autism spectrum disorders: associations with child characteristics. J Autism Dev Disord 38: 1278–1291. doi: 10.1007/s10803-007-0512-z
![]() |
[22] |
Zhou T, Yi C (2014) Parenting styles and parents' perspectives on how their own emotions affect the functioning of children with autism spectrum disorders. Fam Process 53: 67–79. doi: 10.1111/famp.12058
![]() |
[23] |
Dale E, Jahoda A, Knott F (2006) Mothers' attributions following their child's diagnosis of autistic spectrum disorder: exploring links with maternal levels of stress,depression and expectations about their child's future. Autism Int J Res Pract 10: 463–479. doi: 10.1177/1362361306066600
![]() |
[24] |
Meirsschaut M, Roeyers H, Warreyn P (2010) Parenting in families with a child with autism spectrum disorder and a typically developing child: Mothers' experiences and cognitions. Res Autism Spectr Disord 4: 661–669. doi: 10.1016/j.rasd.2010.01.002
![]() |
[25] | Khan TM, Ooi K, Ong YS, et al. (2016) A meta-synthesis on parenting a child with autism. Neuropsychiatr Dis Treat 745. |
[26] |
Poslawsky IE, Naber FBA, Van Daalen E, et al. (2014) Parental reaction to early diagnosis of their children's autism spectrum disorder: an exploratory study. Child Psychiatry Hum Dev 45: 294–305. doi: 10.1007/s10578-013-0400-z
![]() |
[27] | Bitsika V, Sharpley C (2004) Stress, Anxiety and Depression Among Parents of Children With Autism Spectrum Disorder. Humanit Soc Sci Pap 14: 151–161. |
[28] |
Dabrowska A, Pisula E (2010) Parenting stress and coping styles in mothers and fathers of pre-school children with autism and Down syndrome. J Intellect Disabil Res JIDR 54: 266–280. doi: 10.1111/j.1365-2788.2010.01258.x
![]() |
[29] |
Pisula E, Porębowicz-Dörsmann A (2017) Family functioning, parenting stress and quality of life in mothers and fathers of Polish children with high functioning autism or Asperger syndrome. PloS One 12: e0186536. doi: 10.1371/journal.pone.0186536
![]() |
[30] |
Benson P, Karlof KL, Siperstein GN (2008) Maternal involvement in the education of young children with autism spectrum disorders. Autism Int J Res Pract 12: 47–63. doi: 10.1177/1362361307085269
![]() |
[31] |
Bitsika V, Sharpley CF, Bell R (2013) The Buffering Effect of Resilience upon Stress, Anxiety and Depression in Parents of a Child with an Autism Spectrum Disorder. J Dev Phys Disabil 25: 533–543. doi: 10.1007/s10882-013-9333-5
![]() |
[32] |
Zung WWK (1965) A Self-Rating Depression Scale. Arch Gen Psychiatry 12: 63–70. doi: 10.1001/archpsyc.1965.01720310065008
![]() |
[33] |
Zung WWK (1971) A Rating Instrument For Anxiety Disorders. Psychosomatics 12: 371–379. doi: 10.1016/S0033-3182(71)71479-0
![]() |
[34] |
Connor KM, Davidson JRT (2003) Development of a new resilience scale: The Connor-Davidson Resilience Scale (CD-RISC). Depress Anxiety 18: 76–82. doi: 10.1002/da.10113
![]() |
[35] | Margasiński A (2013) Skala Oceny Rodziny. Polska adaptacja FACES IV- Flexibility and Cohesion Evaluation Scales Davida H. Olsona Family Assessment Scales, Polish adaptation FACES IV-Flexibility and Cohesion Evaluation Scaleń by David H. Olson. Warszawa. |
[36] |
Olson D (2011) FACES IV and the Circumplex Model: validation study. J Marital Fam Ther 37: 64–80. doi: 10.1111/j.1752-0606.2009.00175.x
![]() |
[37] |
Olson DH (2000) Circumplex Model of Marital and Family Sytems. J Fam Ther 22: 144–167. doi: 10.1111/1467-6427.00144
![]() |
[38] | Olson DH, Gorall DM, Tiesel J (2007) FACES IV & the Circumplex Model: Validation Study: Recuperado el 01/04/09 en www. facesiv. com Life Innovations. |
[39] |
Brobst JB, Clopton JR, Hendrick SS (2009) Parenting Children With Autism Spectrum Disorders: The Couple's Relationship. Focus Autism Dev Disabil 24: 38–49. doi: 10.1177/1088357608323699
![]() |
[40] |
Hoffman CD, Sweeney DP, Hodge D, et al. (2009) Parenting stress and closeness: Mothers of typically developing children and mothers of children with autism. Focus Autism Other Dev Disabil 24: 178–187. doi: 10.1177/1088357609338715
![]() |
[41] |
Lee GK, Lopata C, Volker MA, et al. (2009) Health-related quality of life of parents of children with high-functioning autism spectrum disorders. Focus Autism Other Dev Disabil 24: 227–239. doi: 10.1177/1088357609347371
![]() |
[42] |
Benson PR, Karlof KL (2009) Anger, Stress Proliferation, and Depressed Mood Among Parents of Children with ASD: A Longitudinal Replication. J Autism Dev Disord 39: 350–362. doi: 10.1007/s10803-008-0632-0
![]() |
[43] |
Hayes SA, Watson SL (2013) The Impact of Parenting Stress: A Meta-analysis of Studies Comparing the Experience of Parenting Stress in Parents of Children With and Without Autism Spectrum Disorder. J Autism Dev Disord 43: 629–642. doi: 10.1007/s10803-012-1604-y
![]() |
[44] |
Abbeduto L, Seltzer MM, Shattuck P, et al. (2004) Psychological well-being and coping in mothers of youths with autism, Down syndrome, or fragile X syndrome. Am J Ment Retard AJMR 109: 237–254. doi: 10.1352/0895-8017(2004)109<237:PWACIM>2.0.CO;2
![]() |
[45] |
Pisula E (2007) A comparative study of stress profiles in mothers of children with autism and those of children with Down's syndrome. J Appl Res Intellect Disabil 20: 274–278. doi: 10.1111/j.1468-3148.2006.00342.x
![]() |
[46] |
Cochrane JJ, Goering PN, Rogers JM (1997) The mental health of informal caregivers in Ontario: an epidemiological survey. Am J Public Health 87: 2002–2007. doi: 10.2105/AJPH.87.12.2002
![]() |
[47] |
Pinquart M, Sörensen S (2003) Differences between caregivers and noncaregivers in psychological health and physical health: a meta-analysis. Psychol Aging 18: 250–267. doi: 10.1037/0882-7974.18.2.250
![]() |
[48] |
Lovell B, Moss M, Wetherell M (2012) The psychosocial, endocrine and immune consequences of caring for a child with autism or ADHD. Psychoneuroendocrinology 37: 534–542. doi: 10.1016/j.psyneuen.2011.08.003
![]() |
[49] |
Ruiz-Robledillo N, Moya-Albiol L (2013) Self-reported health and cortisol awakening response in parents of people with asperger syndrome: the role of trait anger and anxiety, coping and burden. Psychol Health 28: 1246–1264. doi: 10.1080/08870446.2013.800517
![]() |
[50] |
Bebko JM, Konstantareas MM, Springer J (1987) Parent and professional evaluations of family stress associated with characteristics of autism. J Autism Dev Disord 17: 565–576. doi: 10.1007/BF01486971
![]() |
[51] | Liwag MECD (2002) Mothers and Fathers of Autistic Children: An Exploratory Study of Family Stress and Coping. Philipp J Psychol 35. |
[52] |
Bishop SL, Richler J, Cain AC, et al. (2007) Predictors of Perceived Negative Impact in Mothers of Children With Autism Spectrum Disorder. Am J Ment Retard 112: 450–461. doi: 10.1352/0895-8017(2007)112[450:POPNII]2.0.CO;2
![]() |
[53] |
Konstantareas MM, Papageorgiou V (2006) Effects of temperament, symptom severity and level of functioning on maternal stress in Greek children and youth with ASD. Autism 10: 593–607. doi: 10.1177/1362361306068511
![]() |
[54] |
Lecavalier L, Leone S, Wiltz J (2006) The impact of behaviour problems on caregiver stress in young people with autism spectrum disorders. J Intellect Disabil Res JIDR 50: 172–183. doi: 10.1111/j.1365-2788.2005.00732.x
![]() |
[55] |
Rivard M, Terroux A, Parent-Boursier C, et al. (2014) Determinants of stress in parents of children with autism spectrum disorders. J Autism Dev Disord 44: 1609–1620. doi: 10.1007/s10803-013-2028-z
![]() |
[56] |
Zaidman-Zait A, Mirenda P, Duku E, et al. (2014) Examination of bidirectional relationships between parent stress and two types of problem behavior in children with autism spectrum disorder. J Autism Dev Disord 44: 1908–1917. doi: 10.1007/s10803-014-2064-3
![]() |
[57] | Bitsika V, Sharpley CF, Andronicos NM, et al. (2017) What worries parents of a child with Autism? Evidence from a biomarker for chronic stress. Res Dev Disabil 62: 209–217. |
[58] | Fredrickson BL, Tugade MM, Waugh CE, et al. (2003) What good are positive emotions in crises? A prospective study of resilience and emotions following the terrorist attacks on the United States on September 11th, 2001. J Pers Soc Psychol 84: 365–376. |
[59] | Bristol MM, Schopler E (1983) Stress and Coping in Families of Autistic Adolescents. Autism in Adolescents and Adults, Springer, Boston, MA, 251–278. |
[60] |
Kuhn JC, Carter AS (2006) Maternal self-efficacy and associated parenting cognitions among mothers of children with autism. Am J Orthopsychiatry 76: 564–575. doi: 10.1037/0002-9432.76.4.564
![]() |
[61] |
Pottie CG, Ingram KM (2008) Daily stress, coping, and well-being in parents of children with autism: a multilevel modeling approach. J Fam Psychol 22: 855–864. doi: 10.1037/a0013604
![]() |
[62] | Vidyasagar N, Koshy S (2010) Stress and coping in mothers of autistic children. J Indian Acad Appl Psychol 36: 245–248. |
[63] |
Sharpley CF, Bitsika V, Efremidis B (1997) Influence of gender, parental health, and perceived expertise of assistance upon stress, anxiety, and depression among parents of children with autism. J Intellect Dev Disabil 22: 19–28. doi: 10.1080/13668259700033261
![]() |
[64] |
Das S, Das B, Nath K, et al. (2017) Impact of stress, coping, social support, and resilience of families having children with autism: A North East India-based study. Asian J Psychiatry 28: 133–139. doi: 10.1016/j.ajp.2017.03.040
![]() |
[65] |
Diener ED, Emmons RA, Larsen RJ, et al. (1985) The satisfaction with life scale. J Pers Assess 49: 71–75. doi: 10.1207/s15327752jpa4901_13
![]() |
[66] |
Tobin DL, Holroyd KA, Reynolds RV, et al. (1989) The hierarchical factor structure of the Coping Strategies Inventory. Cognit Ther Res 13: 343–361. doi: 10.1007/BF01173478
![]() |
[67] |
Godoy-Izquierdo D, Sola HC, García JFG (2011) Assessing coping with stress self-efficacy: English validation of the CSSES. Health Outcomes Res Med 2: e105–e118. doi: 10.1016/j.ehrm.2011.06.001
![]() |
[68] |
Folkman S, Lazarus RS (1980) An Analysis of Coping in a Middle-Aged Community Sample. J Health Soc Behav 21: 219–239. doi: 10.2307/2136617
![]() |
[69] | Luque Salas B, Yáñez Rodríguez V, Tabernero Urbieta C, et al. (2017) The role of coping strategies and self-efficacy as predictors of life satisfaction in a sample of parents of children with autism spectrum disorder. Psicothema 29: 55–60. |
[70] |
Matud M (2004) Gender differences in stress and coping styles. Pers Individ Differ 37: 1401–1415. doi: 10.1016/j.paid.2004.01.010
![]() |
[71] |
Smith LE, Hong J, Seltzer MM, et al. (2010) Daily Experiences Among Mothers of Adolescents and Adults with Autism Spectrum Disorder. J Autism Dev Disord 40: 167–178. doi: 10.1007/s10803-009-0844-y
![]() |
[72] |
Tehee E, Honan R, Hevey D (2009) Factors Contributing to Stress in Parents of Individuals with Autistic Spectrum Disorders. J Appl Res Intellect Disabil 22: 34–42. doi: 10.1111/j.1468-3148.2008.00437.x
![]() |
[73] |
Beck A, Daley D, Hastings RP, et al. (2004) Mothers' expressed emotion towards children with and without intellectual disabilities. J Intellect Disabil Res 48: 628–638. doi: 10.1111/j.1365-2788.2003.00564.x
![]() |
[74] |
Gray DE (2006) Coping over time: the parents of children with autism. J Intellect Disabil Res 50: 970–976. doi: 10.1111/j.1365-2788.2006.00933.x
![]() |
[75] |
Hastings RP, Brown T (2002) Behavior problems of children with autism, parental self-efficacy, and mental health. Am J Ment Retard 107: 222–232. doi: 10.1352/0895-8017(2002)107<0222:BPOCWA>2.0.CO;2
![]() |
[76] |
Jones TL, Prinz RJ (2005) Potential roles of parental self-efficacy in parent and child adjustment: a review. Clin Psychol Rev 25: 341–363. doi: 10.1016/j.cpr.2004.12.004
![]() |
[77] |
King GA, Zwaigenbaum L, King S, et al. (2006) A qualitative investigation of changes in the belief systems of families of children with autism or Down syndrome. Child Care Health Dev 32: 353–369. doi: 10.1111/j.1365-2214.2006.00571.x
![]() |
[78] |
Lightsey Jr OR, Sweeney J (2008) Meaning in life, emotion-oriented coping, generalized self-efficacy, and family cohesion as predictors of family satisfaction among mothers of children with disabilities. Fam J 16: 212–221. doi: 10.1177/1066480708317503
![]() |
[79] |
Woodman AC, Hauser-Cram P (2013) The role of coping strategies in predicting change in parenting efficacy and depressive symptoms among mothers of adolescents with developmental disabilities: Role of coping strategies. J Intellect Disabil Res 57: 513–530. doi: 10.1111/j.1365-2788.2012.01555.x
![]() |
[80] | Gardner David G, Shoback Dolores (2007) Glucocorticoids and Adrenal Androgens. In: Gardner DG, Shoback D, Greenspan FS, Greenspan's Basic & Clinical Endocrinology, 9Eds., McGraw-Hill Medical. |
[81] |
Buono MJ, Yeager JE, Hodgdon JA (1986) Plasma adrenocorticotropin and cortisol responses to brief high-intensity exercise in humans. J Appl Physiol 61: 1337–1339. doi: 10.1152/jappl.1986.61.4.1337
![]() |
[82] |
Chrousos GP (2009) Stress and disorders of the stress system. Nat Rev Endocrinol 5: 374–381. doi: 10.1038/nrendo.2009.106
![]() |
[83] |
Whittingham K, Sofronoff K, Sheffield J, et al. (2009) Behavioural Family Intervention with parents of children with ASD: What do they find useful in the parenting program Stepping Stones Triple P? Res Autism Spectr Disord 3: 702–713. doi: 10.1016/j.rasd.2009.01.009
![]() |
[84] |
Kaminski JW, Valle LA, Filene JH, et al. (2008) A meta-analytic review of components associated with parent training program effectiveness. J Abnorm Child Psychol 36: 567–589. doi: 10.1007/s10802-007-9201-9
![]() |
[85] | Bitsika V, Sharpley C (1999) An exploratory examination of the effects of support groups on the well-being of parents of children with autism-I: General counselling. J Appl Health Behav 1: 16–22. |
1. | Gwendoline DESQUENNE GODFREY, Naomi DOWNES, Emilie CAPPE, A Systematic Review of Family Functioning in Families of Children on the Autism Spectrum, 2023, 0162-3257, 10.1007/s10803-022-05830-6 | |
2. | Emma Chad-Friedman, Karen A. Kuhlthau, Rachel A. Millstein, Giselle K. Perez, Christina M. Luberto, Lara Traeger, Jacqueline Proszynski, Elyse Park, Characteristics and Experiences of Parents of Children with Learning and Attention Disabilities and Autism Spectrum Disorder: A Mixed Methods Study, 2022, 30, 1066-4807, 427, 10.1177/10664807211052304 | |
3. | Anna Kostiukow, Piotr Poniewierski, Dominika Janowska, Włodzimierz Samborski, Levels of happiness and depression in parents of children with autism spectrum disorder in Poland, 2021, 81, 0065-1400, 279, 10.21307/ane-2021-026 | |
4. | Talal E. Alhuzimi, Family Functioning and Strengths in Families of Children With Autism Spectrum Disorder in Saudi Arabia, 2024, 32, 1066-4807, 230, 10.1177/10664807231217061 | |
5. | Fátima El‐Bouhali‐Abdellaoui, Núria Voltas, Paula Morales‐Hidalgo, Josefa Canals, Examining the Relationship Between Parental Broader Autism Phenotype Traits, Offspring Autism, and Parental Mental Health, 2024, 1939-3792, 10.1002/aur.3295 |
Dataset | Normal | Glaucoma | Total |
ACRIMA | 309 | 396 | 705 |
ORIGA-light | 482 | 168 | 650 |
RIM-ONE | 261 | 194 | 455 |
AFIO | 85 | 39 | 124 |
HMC | 40 | 15 | 55 |
1177 | 812 | 1989 |
Class | Training | Validation | Testing |
Normal | 706 | 141 | 330 |
Glaucoma | 487 | 98 | 227 |
Dataset | Normal | Glaucoma |
ACRIMA | 87 | 110 |
ORIGA-Light | 135 | 47 |
RIM-ONE | 73 | 55 |
AFIO | 24 | 11 |
HMC | 11 | 4 |
Deep CNNs | ACRIMA | ORIGA-Light | RIM-ONE | AFIO | HMC |
AlexNet | 99.10 | 75.50 | 86.70 | 90.90 | 98.50 |
InceptionV3 | 99.28 | 78.50 | 86.40 | 84.50 | 90.50 |
InceptionResNetV2 | 94.72 | 76.50 | 90.60 | 85.50 | 95.50 |
NasNet-Large | 100.00 | 87.10 | 94.70 | 97.70 | 98.60 |
Deep CNNs | Datasets | ||||||||
ACRIMA | ORIGA-Light | RIM-ONE | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 99.5 | 98.9 | 100.0 | 75.8 | 81.5 | 60.0 | 87.5 | 90.4 | 83.6 |
InceptionV3 | 98.5 | 97.7 | 99.1 | 78.6 | 83.0 | 66.0 | 92.2 | 94.5 | 89.1 |
InceptionResNetV2 | 99.0 | 97.7 | 100.0 | 76.9 | 82.2 | 61.7 | 90.6 | 94.5 | 85.5 |
NasNet-Large | 99.5 | 100.0 | 99.1 | 87.9 | 91.1 | 78.7 | 94.5 | 95.9 | 92.7 |
AFIO | HMC | TOTAL | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 91.4 | 95.8 | 81.8 | 93.3 | 100.0 | 75.0 | 97.3 | 97.3 | 97.4 |
InceptionV3 | 91.4 | 91.7 | 91.0 | 86.7 | 100.0 | 50.0 | 96.6 | 97.0 | 96.0 |
InceptionResNetV2 | 88.6 | 83.3 | 100.0 | 93.3 | 91.0 | 100.0 | 97.1 | 97.3 | 96.9 |
NasNet-Large | 94.3 | 91.7 | 100.0 | 99.9 | 100.0 | 100.0 | 99.3 | 99.4 | 99.1 |
Dataset | MV | PV | AV | AWV | ASWA |
ACRIMA | 99.5 | 99.4 | 99.1 | 99.6 | 99.5 |
ORIGA-Light | 87.9 | 88.0 | 79.8 | 88.2 | 88.3 |
RIM-ONE | 94.5 | 94.5 | 91.2 | 95.1 | 95.2 |
AFIO | 94.3 | 94.3 | 91.4 | 96.1 | 96.2 |
HMC | 99.9 | 99.9 | 96.3 | 99.8 | 99.9 |
Total | 99.3 | 99.3 | 98.5 | 99.4 | 99.5 |
Deep CNNs | MV | PV | AV | AWV | ASWA | |
$ 2x $ Ensemble | A+B | 98.70 | 98.60 | 97.80 | 99.00 | 98.80 |
A+C | 98.80 | 98.50 | 97.60 | 99.00 | 98.90 | |
A+D | 99.00 | 98.90 | 97.80 | 99.20 | 99.00 | |
B+C | 98.80 | 99.00 | 96.90 | 98.90 | 98.80 | |
B+D | 98.80 | 98.70 | 97.50 | 98.90 | 98.90 | |
$ 3x $ Ensemble | A+B+C | 98.90 | 98.80 | 98.10 | 98.80 | 98.90 |
A+B+D | 98.80 | 99.00 | 98.10 | 98.80 | 99.00 | |
A+C+D | 99.00 | 99.10 | 98.20 | 99.00 | 99.00 | |
B+C+D | 98.90 | 99.10 | 98.10 | 98.80 | 98.90 | |
$ 4x $ Ensemble | A+B+C+D | 99.30 | 99.30 | 98.50 | 99.40 | 99.50 |
A = AlexNet, B = InceptionV3, C = InceptionResNetV2, & D = NasNet-Large |
Systems | Methods | Dataset | Sensitivity (%) | Specificity (%) | Accuracy (%) |
[22] | CNN | Public & Private | 84.5 | 98.0 | 99.0 |
[26] | Multi-branch NN | Private | 92.0 | 90.0 | 91.0 |
[37] | Deep CNNs | Public | 93.4 | 85.8 | 96.0 (AUC) |
[56] | CNN | Private | 98.0 | 98.0 | 98.0 |
[57] | CNN | RIM-ONE | 80.0 | 88.0 | 85.0 |
[58] | RCNN | ORIGA-Light | NR | NR | 87.4 (AUC) |
[59] | GoogleNet | HRF & RIM-ONE | NR | NR | 87.6 |
[60] | AlexNet | RIM-ONE | 87.0 | 85.0 | 88.0 |
[61] | Deep Models | Public & Private | NR | NR | 85.0 (AUC)ACRIMA |
[62] | InceptionV3 | Private | NR | NR | 84.5 & 93.0 (AUC) |
[63] | Ensemble Classifier | Private | 86.0 | 90.0 | 88.0 & 94.0 (AUC) |
Our | NasNet-Large$ \ast $ | Public & Private | 99.1 | 99.4 | 99.3 & 97.8% (AUC) |
Our | Ensemble Classifier$ \ast $ | Public & Private | 99.1 | 99.7 | 99.5 |
Dataset | Normal | Glaucoma | Total |
ACRIMA | 309 | 396 | 705 |
ORIGA-light | 482 | 168 | 650 |
RIM-ONE | 261 | 194 | 455 |
AFIO | 85 | 39 | 124 |
HMC | 40 | 15 | 55 |
1177 | 812 | 1989 |
Class | Training | Validation | Testing |
Normal | 706 | 141 | 330 |
Glaucoma | 487 | 98 | 227 |
Dataset | Normal | Glaucoma |
ACRIMA | 87 | 110 |
ORIGA-Light | 135 | 47 |
RIM-ONE | 73 | 55 |
AFIO | 24 | 11 |
HMC | 11 | 4 |
Deep CNNs | ACRIMA | ORIGA-Light | RIM-ONE | AFIO | HMC |
AlexNet | 99.10 | 75.50 | 86.70 | 90.90 | 98.50 |
InceptionV3 | 99.28 | 78.50 | 86.40 | 84.50 | 90.50 |
InceptionResNetV2 | 94.72 | 76.50 | 90.60 | 85.50 | 95.50 |
NasNet-Large | 100.00 | 87.10 | 94.70 | 97.70 | 98.60 |
Deep CNNs | Datasets | ||||||||
ACRIMA | ORIGA-Light | RIM-ONE | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 99.5 | 98.9 | 100.0 | 75.8 | 81.5 | 60.0 | 87.5 | 90.4 | 83.6 |
InceptionV3 | 98.5 | 97.7 | 99.1 | 78.6 | 83.0 | 66.0 | 92.2 | 94.5 | 89.1 |
InceptionResNetV2 | 99.0 | 97.7 | 100.0 | 76.9 | 82.2 | 61.7 | 90.6 | 94.5 | 85.5 |
NasNet-Large | 99.5 | 100.0 | 99.1 | 87.9 | 91.1 | 78.7 | 94.5 | 95.9 | 92.7 |
AFIO | HMC | TOTAL | |||||||
ACC | SP | SEN | ACC | SP | SEN | ACC | SP | SEN | |
AlexNet | 91.4 | 95.8 | 81.8 | 93.3 | 100.0 | 75.0 | 97.3 | 97.3 | 97.4 |
InceptionV3 | 91.4 | 91.7 | 91.0 | 86.7 | 100.0 | 50.0 | 96.6 | 97.0 | 96.0 |
InceptionResNetV2 | 88.6 | 83.3 | 100.0 | 93.3 | 91.0 | 100.0 | 97.1 | 97.3 | 96.9 |
NasNet-Large | 94.3 | 91.7 | 100.0 | 99.9 | 100.0 | 100.0 | 99.3 | 99.4 | 99.1 |
Dataset | MV | PV | AV | AWV | ASWA |
ACRIMA | 99.5 | 99.4 | 99.1 | 99.6 | 99.5 |
ORIGA-Light | 87.9 | 88.0 | 79.8 | 88.2 | 88.3 |
RIM-ONE | 94.5 | 94.5 | 91.2 | 95.1 | 95.2 |
AFIO | 94.3 | 94.3 | 91.4 | 96.1 | 96.2 |
HMC | 99.9 | 99.9 | 96.3 | 99.8 | 99.9 |
Total | 99.3 | 99.3 | 98.5 | 99.4 | 99.5 |
Deep CNNs | MV | PV | AV | AWV | ASWA | |
$ 2x $ Ensemble | A+B | 98.70 | 98.60 | 97.80 | 99.00 | 98.80 |
A+C | 98.80 | 98.50 | 97.60 | 99.00 | 98.90 | |
A+D | 99.00 | 98.90 | 97.80 | 99.20 | 99.00 | |
B+C | 98.80 | 99.00 | 96.90 | 98.90 | 98.80 | |
B+D | 98.80 | 98.70 | 97.50 | 98.90 | 98.90 | |
$ 3x $ Ensemble | A+B+C | 98.90 | 98.80 | 98.10 | 98.80 | 98.90 |
A+B+D | 98.80 | 99.00 | 98.10 | 98.80 | 99.00 | |
A+C+D | 99.00 | 99.10 | 98.20 | 99.00 | 99.00 | |
B+C+D | 98.90 | 99.10 | 98.10 | 98.80 | 98.90 | |
$ 4x $ Ensemble | A+B+C+D | 99.30 | 99.30 | 98.50 | 99.40 | 99.50 |
A = AlexNet, B = InceptionV3, C = InceptionResNetV2, & D = NasNet-Large |
Systems | Methods | Dataset | Sensitivity (%) | Specificity (%) | Accuracy (%) |
[22] | CNN | Public & Private | 84.5 | 98.0 | 99.0 |
[26] | Multi-branch NN | Private | 92.0 | 90.0 | 91.0 |
[37] | Deep CNNs | Public | 93.4 | 85.8 | 96.0 (AUC) |
[56] | CNN | Private | 98.0 | 98.0 | 98.0 |
[57] | CNN | RIM-ONE | 80.0 | 88.0 | 85.0 |
[58] | RCNN | ORIGA-Light | NR | NR | 87.4 (AUC) |
[59] | GoogleNet | HRF & RIM-ONE | NR | NR | 87.6 |
[60] | AlexNet | RIM-ONE | 87.0 | 85.0 | 88.0 |
[61] | Deep Models | Public & Private | NR | NR | 85.0 (AUC)ACRIMA |
[62] | InceptionV3 | Private | NR | NR | 84.5 & 93.0 (AUC) |
[63] | Ensemble Classifier | Private | 86.0 | 90.0 | 88.0 & 94.0 (AUC) |
Our | NasNet-Large$ \ast $ | Public & Private | 99.1 | 99.4 | 99.3 & 97.8% (AUC) |
Our | Ensemble Classifier$ \ast $ | Public & Private | 99.1 | 99.7 | 99.5 |