Classification using supervised learning requires annotating a large amount of classes-balanced data for model training and testing. This has practically limited the scope of applications with supervised learning, in particular deep learning. To address the issues associated with limited and imbalanced data, this paper introduces a sample-efficient co-supervised learning paradigm (SEC-CGAN), in which a conditional generative adversarial network (CGAN) is trained alongside the classifier and supplements semantics-conditioned, confidence-aware synthesized examples to the annotated data during the training process. In this setting, the CGAN not only serves as a co-supervisor but also provides complementary quality examples to aid the classifier training in an end-to-end fashion. Experiments demonstrate that the proposed SEC-CGAN outperforms the external classifier GAN (EC-GAN) and a baseline ResNet-18 classifier. For the comparison, all classifiers in above methods adopt the ResNet-18 architecture as the backbone. Particularly, for the Street View House Numbers dataset, using the 5% of training data, a test accuracy of 90.26% is achieved by SEC-CGAN as opposed to 88.59% by EC-GAN and 87.17% by the baseline classifier; for the highway image dataset, using the 10% of training data, a test accuracy of 98.27% is achieved by SEC-CGAN, compared to 97.84% by EC-GAN and 95.52% by the baseline classifier.
Citation: Hao Zhen, Yucheng Shi, Jidong J. Yang, Javad Mohammadpour Vehni. Co-supervised learning paradigm with conditional generative adversarial networks for sample-efficient classification[J]. Applied Computing and Intelligence, 2023, 3(1): 13-26. doi: 10.3934/aci.2023002
Related Papers:
[1]
Yantao Song, Wenjie Zhang, Yue Zhang .
A novel lightweight deep learning approach for simultaneous optic cup and optic disc segmentation in glaucoma detection. Mathematical Biosciences and Engineering, 2024, 21(4): 5092-5117.
doi: 10.3934/mbe.2024225
[2]
Jun Gao, Qian Jiang, Bo Zhou, Daozheng Chen .
Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Mathematical Biosciences and Engineering, 2019, 16(6): 6536-6561.
doi: 10.3934/mbe.2019326
[3]
Meenu Garg, Sheifali Gupta, Soumya Ranjan Nayak, Janmenjoy Nayak, Danilo Pelusi .
Modified pixel level snake using bottom hat transformation for evolution of retinal vasculature map. Mathematical Biosciences and Engineering, 2021, 18(5): 5737-5757.
doi: 10.3934/mbe.2021290
[4]
Aziz-ur-Rehman, Imtiaz A. Taj, Muhammad Sajid, Khasan S. Karimov .
An ensemble framework based on Deep CNNs architecture for glaucoma classification using fundus photography. Mathematical Biosciences and Engineering, 2021, 18(5): 5321-5346.
doi: 10.3934/mbe.2021270
[5]
Yali Ouyang, Zhuhuang Zhou, Weiwei Wu, Jin Tian, Feng Xu, Shuicai Wu, Po-Hsiang Tsui .
A review of ultrasound detection methods for breast microcalcification. Mathematical Biosciences and Engineering, 2019, 16(4): 1761-1785.
doi: 10.3934/mbe.2019085
[6]
Yu Li, Meilong Zhu, Guangmin Sun, Jiayang Chen, Xiaorong Zhu, Jinkui Yang .
Weakly supervised training for eye fundus lesion segmentation in patients with diabetic retinopathy. Mathematical Biosciences and Engineering, 2022, 19(5): 5293-5311.
doi: 10.3934/mbe.2022248
[7]
Auwalu Saleh Mubarak, Zubaida Said Ameen, Fadi Al-Turjman .
Effect of Gaussian filtered images on Mask RCNN in detection and segmentation of potholes in smart cities. Mathematical Biosciences and Engineering, 2023, 20(1): 283-295.
doi: 10.3934/mbe.2023013
[8]
Jianzhong Peng, Wei Zhu, Qiaokang Liang, Zhengwei Li, Maoying Lu, Wei Sun, Yaonan Wang .
Defect detection in code characters with complex backgrounds based on BBE. Mathematical Biosciences and Engineering, 2021, 18(4): 3755-3780.
doi: 10.3934/mbe.2021189
[9]
Naigong Yu, Hongzheng Li, Qiao Xu .
A full-flow inspection method based on machine vision to detect wafer surface defects. Mathematical Biosciences and Engineering, 2023, 20(7): 11821-11846.
doi: 10.3934/mbe.2023526
[10]
Qiaokang Liang, Jianyong Long, Yang Nan, Gianmarc Coppola, Kunlin Zou, Dan Zhang, Wei Sun .
Angle aided circle detection based on randomized Hough transform and its application in welding spots detection. Mathematical Biosciences and Engineering, 2019, 16(3): 1244-1257.
doi: 10.3934/mbe.2019060
Abstract
Classification using supervised learning requires annotating a large amount of classes-balanced data for model training and testing. This has practically limited the scope of applications with supervised learning, in particular deep learning. To address the issues associated with limited and imbalanced data, this paper introduces a sample-efficient co-supervised learning paradigm (SEC-CGAN), in which a conditional generative adversarial network (CGAN) is trained alongside the classifier and supplements semantics-conditioned, confidence-aware synthesized examples to the annotated data during the training process. In this setting, the CGAN not only serves as a co-supervisor but also provides complementary quality examples to aid the classifier training in an end-to-end fashion. Experiments demonstrate that the proposed SEC-CGAN outperforms the external classifier GAN (EC-GAN) and a baseline ResNet-18 classifier. For the comparison, all classifiers in above methods adopt the ResNet-18 architecture as the backbone. Particularly, for the Street View House Numbers dataset, using the 5% of training data, a test accuracy of 90.26% is achieved by SEC-CGAN as opposed to 88.59% by EC-GAN and 87.17% by the baseline classifier; for the highway image dataset, using the 10% of training data, a test accuracy of 98.27% is achieved by SEC-CGAN, compared to 97.84% by EC-GAN and 95.52% by the baseline classifier.
1.
Introduction
Content-based image analysis is applied in different applications of computer vision and digital images are a major part of multimedia data [1]. In last few decades, significant research is performed on different applications of medical image analysis. The main aim of research in medical image analysis is to assist the doctors to detect diseases on the basis of image contents [2,3]. Glaucoma is a persistent eye disorder that gradually damages the optic nerve. It is a neurodegeneration illness and contemplates as a reason for blindness. As debasement of nerves is irrevocable action hence it results in permanent loss of vision [3,4]. Numerous research studies have been performed to estimate the number of affected people with this chronic disease [5,6,7]. Glaucoma is declared as the second leading cause of the loss of sight [8,9]. Results indicate an estimate of 60 million people suffered from glaucoma in 2010, while the number will be increased to approximately 80 million by 2020 [9]. Glaucoma causes irrevocable vision damage due to elevated Intraocular Pressure (IOP) and optic nerve damage[10]. Glaucoma is commonly referred to as "hushed burglar of vision" since the symptoms at an early stage of glaucoma are not explicitly defined and are hard to quantify. If the succession of glaucoma is not stopped at the initial stages it results in the severe demolition of the optic nerve and as a consequence, it will lead to incurable blindness [11]. It is necessary to detect glaucoma at an early stage due to following facts:
1. There are no perceptible indications in its preliminary stages.
2. It is a severe disorder as the damage it causes is irremediable.
3. It leads to perpetual loss of sight if not cured promptly.
4. There is no prophylactic treatment for glaucoma, but it is possible to avoid blindness by detecting, treating and managing glaucoma at an initial stage [12]. The list of abbreviations used in this paper are shown in the table 1.
Computer-Aided Diagnosis(CAD) helps in the rehabilitation of the affected eyes of the disease. Numerous research studies have been performed on fundus images to detect glaucoma at an early phase and to avoid any severe damages. Glaucoma alludes to a group of conditions and states characterized by the variations in the retinal nerve fiber cushion and nervous optics head leading to reduced vision and imparity [13]. Researchers are attempting to decrease the damage through ameliorate disease detection and more effective treatment like instant Trabecular Micro-Bypass (TMB). Figure 1 represents an overview of the anatomy of human eye [14]. The authors[15] gave a brief survey on state of art techniques of automatic detection of glaucoma in early stages. Critical evaluation of the methods have been conducted which include Optic cup disc ratio, Retinal Nerve Fiber Layer (RNFL) etc. These researches add value to the efficient diagnosis of glaucoma. They further summarizes the survey into 2 categories of segmentation based approaches and non-segmentation based approaches. Sarhan et al.[16] conducted the survey of current approaches to detect glaucoma that are valuable for the researchers to predict the proficiency and precision. This is an effective survey that lays the guiding details for the researcher in terms of approach of glaucoma and dataset selection for further research [16].
Glaucoma is a multiplex ailment and a single test cannot give much information about the detection of glaucoma. A constant eye assessment habitually necessitates concealing for glaucoma and indicates that additional inspection is needed. Glaucoma's sightlessness is inevitable, that is why initial detection is pivotal as glaucoma cure rescue endure vision but not refine eyesight [17]. Glaucoma can occur due to the enlarging of IOP, the changes in the inter-papillary, and the para-papillary area of the optic nerve head [10]. Glaucoma succession directs to few structural alterations in the region of retinal layers and optic nerve. Cup to Disc Ratio (CDR) and Rim to Disc Ratio (RDR) are numerous structural alterations that are used for glaucoma detection. Observing the values of these parameters may help in detection of glaucoma [11]. Figure 2 illustrates the fundus image of normal human eye [18]. The increase in IOP in the eye is considered as the origin of nerve optics injure. This pertains to mechanical squeezing or reduces blood passage of the optic nerve and occasionally eye pressure results in glaucoma [19]. Increased pressure changes the CDR and RDR that are considered to be the most important parameter for glaucoma diagnosis [20]. The circular zone is optic nerve concentred by the optic cup and the area in middle is called Neuro-Retinal Rim (NRR). CDR and RDR are constructional variations consume in the detection of glaucoma [21]. In eyes with glaucoma, the growth of IOP results in the enlargement of an optic cup that causes a reduction in NRR, leading to a large CDR and a small RDR. Figure 3 illustrates a labelled fundus image [22].
Patients suffering from glaucoma do not have any sort of symptoms at early stages and have no signal of illness but at the advanced stages people having glaucoma experience vision loss. Optic nerve is a blemish to work due to high pressure in eye which is affected by the high production of aqueous humor or degassing system of the liquid gets blocked. Somehow this disease is also occurred because of genetics having this disease prior in the family. Several tests are performed with patients having suspected glaucoma like tonometry i.e. used for measuring the pressure of eyes, gonioscopy is used to check out either the angle is open or closed. Moreover, fundus imaging as well as optical coherence tomography is used for the optic nerve and the location of the retina. The thickness of retinal nerve fiber layer is diagnosed by the retinal fundus imaging, it helps in the demonstration of glaucoma. This technique is very common and used by most of the ophthalmologists. The major advantage of using this technique is that the size of retina can be measured very easily with little effort and it is portable to use.
1.2. Risks cognate with glaucoma
Since last few years, a considerable amount of studies have been conducted and results have been reported about the risks associated with glaucoma [23]. These studies stated many risk factors and their adverse effects. Figure 4 represents the risks affiliated with glaucoma. The occurrence of any of these symptoms in an individual does not indicate that the one is infected with glaucoma, only an ophthalmologist can determine the risk of disease onset and progression. Some of the high risk factors associated with glaucoma are (ⅰ) increased eye pressure (ⅰ) age (ⅲ) thin cornea (ⅳ) hypertensive (ⅴ) diabetic patients (ⅵ) history of injuries to eyes (ⅶ) a family history of glaucoma (ⅷ) nearsightedness (ⅸ) cardiovascular diseases (ⅹ) steroid consumption (ⅹⅰ) rate of blood pressure (ⅹⅱ) severe anemia [25]. Research studies reported that 17% hypertensive, 35% diabetic and 48% individuals suffering from both diseases are at a high risk of developing Open-Angle Glaucoma (OAG). Commonly associated conditions associated to the metabolic syndrome's components comprise of hyperlipidemia (high levels of triglyceride and cholesterol), hypertension, diabetes and obesity. Research studies indicated that diabetes and hypertension increase the risk of OAG in individuals. More research efforts are being made for glaucoma treatment and to assess the impact of hyperlipidemia as glaucoma risk factor. Risk factors are important to consider since they specify a potential target for a specific treatment.
Glaucoma is categorized into various types depending on the causes and symptoms as revealed by the researcher during the last decade. The two pre-eminent types of glaucoma are OAG further branched as Normal-Tension Glaucoma (NTG) [26] and Angle-Colure Glaucoma (ACG). Figure 5 represents the types of glaucoma. These fundamental types encompasses of 90% among all glaucoma causes [27]. The types and the causes of glaucoma are described below:
1. Primary Open Angle Glaucoma (POAG): Freakishness in trabecular reticulum and slow clogging of the drainage canals causes IOP to increase which in turn causes open angle glaucoma. It develops over the time and is referred to as slow poison that is lifelong lasting. Optic nerve damage and optic defect are the measuring metrics for OAG [28].
2. Primary Angle Closure Glaucoma (PACG): This is less common type of glaucoma as compared to open angle glaucoma. The causes include the damage of optic nerve and high IOP. Unlike POAG, angle closure glaucoma grows fast and needs immediate medical attention. The metrics of detection are almost similar as that of POAG i.e. optic nerve head damage, glaucomatous visual filed defect and narrow camber angle. One important measure is occludable angle used to estimate glaucoma. An occludable angle correspond to PACG if the rare trabecular meshwork is seen in less then 90∘ of the angle boundary [29].
3. Low Tension Glaucoma (LTG): It is also caused by the optic nerve damage but the elevation of IOP is not obligatory. Even though the IOPs of some patients are normal but still they have normal tension glaucoma often called low tension glaucoma. People with normal IOP can suffer LTG since they may be sensitive to normal eye pressure. The main cause can be the low blood supply to optic nerve head. The metrics of detection are same as the both POAG and PACG [28]. Table 2 provides an overview of the other types of glaucoma.
Ocular inflammatory disorders (uveitis) can occur for various reasons and unfortunately, glaucoma can soften up to 20% of patients. Some of these disorders tend to affect only one eye. Other forms can affect both eyes
The backside of cornea contains cells that scatter over the eye drainage tissue and across the surface of the iris. This leads to an increased IOP in an eye that can damage the optic nerve. These cells form accession that glue the cornea and iris and block the drainage channel system. And iris and block the drainage channel system.
Normally babies suffer from this type of glaucoma. When a woman is pregnant than if her baby's eye drainage canal system development is not correctly building this leads to Congenital Glaucoma.
Pseudoexfoliative glaucoma occurs when ash-like material peels off from the external side of the lens in the eye. This ash-like material collects within the angle between the iris and cornea and damage the whole system of the eye.
This type of glaucoma always related to other abnormalities, most often diabetes. Blood vessels are blocked and the eye's fluid can leave the eye's drainage canals and this will cause an increase in IOP and this kind of glaucoma is very hard to treat.
Treatment of glaucoma involves certain after-effects and financial expenses. The rudimentary treatment which is also called first-line treatment [41] is the use of eye-drops [42]. Second-line treatments encompass trabeculectomy or microsurgery [43]. Third-line treatment is given to the patients not responding to medications and are at a severe stage of glaucoma. These include laser surgery, drainage implants and other methods that help lower intraocular pressure [44]. Figure 6 demonstrates the different types of treatments associated with glaucoma.
Glaucoma is considered to be an irremediable blindness in people with an age of more than 40 years. Carrillo et al. [45] proposed a computational tool to assist automatic detection of glaucoma. The most widely applied technique for glaucoma detection is fundus imaging since it provides a trade-off between cost, size and portability. The authors reported improved results for disc segmentation as compared to the related research. Disc segmentation was done by thresholding, vessel segmentation was done by edge detection and for cup segmentation a strategy is followed that make use of vessels and cup intensities. Results obtained from a set of fundus images having percentage of glaucoma detection success was 88.5%. According to authors [45], the future work perturbs to use a larger dataset of fundus images to ensure a pro-found evaluation. Sengar et al. [46] suggested a method for the detection of suspected glaucoma by practicing existence or deficiency of hemorrhages in specific area near optic disc in image having fundus features. This method is considered to be computationally efficient and provides effective and accurate results on the diagnosis of suspected glaucoma. Salam et al. [11] proposed procedure that uses an algorithm for the detection of glaucoma having a combination of CDR and hybrid textual and different intensities features. Image organization is processed using the result that comes from both CDR as well as the classifier. According to the results [11], the proposed fusion improved sensitivity, specificity and accuracy. Poshtyar et al. [47] suggested the procedure of automatic detection of glaucoma i.e. without the aid of consultant and advance machinery. Quantitative determination is used for glaucoma detection. The method used is efficient. Khan et al. [48] proposed the techniques for detecting glaucoma in the early stage. In the proposed study [48], glaucoma is classified by CDR and NRR i.e. ISNT quadrants for the purpose of checking out that it follow or unfollow ISNT rule. Table 3 provides a description of the datasets and the reported results in term of accuracy for the approaches that are discussed above.
2.1. Glaucoma detection through peripapillary atrophy
Peripapillary atrophy is said to be divided into 2 categories [49]. These categories are a-zone and b-zone. The authors states that b-zone category is associated with glaucoma progression. They further elaborate that glaucomatous eyes with PPA show faster thinning of retinal nerve then the regular eyes. In this study OCT B scan images were utilized to gain optic disc sizes. Total of 115 subjects were examined in the study. The presence of glaucoma was classified in the basis of Peripapillary atrophy. Authors argue that peripapillary atrophy is considered to be very progressive morphological diagnosis for open angular glaucoma [50]. Peripapillary atrophy is a variable that can be helpful for the detection of glaucoma. Eyes with glaucoma show large peripapillary atrophy. The technique used is confocal scanning laser tomography and automated static threshold perimetry for measurement of peripapillary atrophy [51]. The test was carried out on 102 eyes samples. Here the entropy was divided into 2 zones. First zone was central zone with visible and large vessel and second zone called peripheral with hyper pigmentation. The authors concluded that the peripapillary atrophy is associated with optic nerve damage which is the major cause of glaucoma [51].
2.2. Glaucoma detection through Retinal Nerve fiber
Here researchers used retinal nerve fiber layer RNFL, optic nerve head and muscular thickness parameters to classify healthy and effected eyes [52]. The total eyes under examination were 80. Stratus OCT was used for the scanning of the patients. Authors state that RFNL parameter has higher AUC then other parameters under consideration [52]. Authors tested 64 patients either with healthy eyes or glaucomatous eyes [53]. Each patient went through five oct scans. Linear regression slopes of the patients were measured with glaucoma and non-glaucomatous eyes. Kaplan-Meier survival curve analysis was done for the patients and a significant progression rate was recorded [53].
2.3. Glaucoma detection through Vessel displacement
The authors describe the relationship between cup width, cup depth and vessel displacement. The sample on which the study was conducted consisted of 826 eyes [54]. Fixed image separation method was used to obtain the colored graphs. The vessel displacement was estimated on a scale from 0 to 4 positive numbers. Finally they conclude that the depth of the cup and the vessel displacement are very much correlated [54]. The authors studied 16 eyes of minipig. using fundoscopic photography and fluorescein angiography [55]. The change in glaucomatous eyes was studied. These changes were compared with the previous results. The results found in human and minipig were different and interesting. Angiography changes in minipig were negligible as compared with human eyes diagnosed glaucoma. The eyes with glaucoma of minipig showed no changes. However in the eyes of human with open angle glaucoma both the arterioles and the venues are displaced [55]. The automation of glaucoma has been an important research area for the researcher for the last two decades. Segmentation approaches discussed above are being widely used for the detection of glaucoma due to their improved results. The segmentation approaches have the advantage of being cost effective and providing accurate result hence the trade-off between size, cost and portability are achieved. Image segmentation is done by selecting appropriate feature that needs to be examined however, how to select the suitable image feature and classifier is the challenging issue that are faced by the researchers using segmentation approaches [56]. Pattern based approaches are very helpful for the detection of glaucoma. Since ONH have the structural characteristics which are well understood and verified in several studies and medical institutions [57][58]. Since the parameters are derived from the structural features and characteristics the results are very meaningful and efficient. These approaches are low cost, efficient and reliable [59]. In last few decades great effort has been done on automation of detection of glaucoma using different Machine Learning approaches and techniques. Machine learning deals with the large amount of data and provides the analysis of data [60]. It is a faster process in acquiring information about the data. They have a feature of learning from their mistakes and experiences. Detection of glaucoma is been achieved by using neural networks [61], SVM [62], Naive Bayes classifier, k- nearest neighbor [63], Canny edge detector [64] Fuzzy min-max neural network [65]. The main advantages of the machine learning approaches in detection of glaucoma is that they can handle large amount of datasets and can provide beneficial results. Since the machine learning approaches do not need much human interference hence the human resource is saved. The data is handled and is trained and made error free using AI. The results obtained using machine learning classifiers show significant improvement compared to the traditional segmentation approaches.
The diagram 7 illustrates the architecture called AGLAIA-MII to detect the features for glaucoma detection and pass the features to multiple kernel learning (MKL) framework which will finally make the diagnosis of Glaucoma. This is a very effective framework used by researchers for glaucoma detection. The only disadvantage associated with ML techniques is the data acquisition, which should be good quality and unbiased. Some of the glaucoma ML detection techniques do not provide the excellent results only due to low data quality and less data sometimes. Cost and resources are some factors which are considered sometimes before using ML classifiers as they require large amount of resources and cost. This section covers the technical information about the image processing techniques, methodologies and algorithms proposed for the automatic detection of glaucoma. The researchers have done a lot of work for the diagnostics of glaucoma which resulted in number of types of approaches and algorithms. Figure 8 precisely demonstrates the commonly used computer vision techniques for glaucoma detection. In this article, we aim to present a detailed review about recent trends to detect glaucoma by highlighting the main contributions, available benchmarks for research evaluation, upcoming challenges. The limitations and drawbacks associated with the published research to detect glaucoma are discussed in detail and possible future directions are also mentioned. We aim to discuss the glaucoma detection based on machine learning, simple segmentation approaches, fusion based approaches, multi-scale-based approaches, texture feature-based approaches, OCT-based approaches, deep-learning-based approaches, a summary about datasets/standard benchmark to evaluate glaucoma-based research is also presented, and a comparison between deep-learning and traditional approaches is also presented in this review article. The performance evaluation metrics are also presented and discussed in detail. At the end of each section, a detailed comparison is presented in the form of tables that conclude about the research methodology, datasets and performance percentage for each of the selected research model.
Figure 8.
Classification of computer vision techniques for glaucoma detection [42].
3.
Machine Learning approaches for glaucoma detection
Machine learning is an entreaty of artificial intelligence that has been given great attention. This provides the system with the ability to learn and improve the by the experience and in turn enhance the efficiency of the output without telling it what to do by programming. Since last two decades, ML approaches are used for the detection of glaucoma [67]. Dey at el. [68] in the year 2016 proposed ML approach for automatic glaucoma detection using SVM as a classifier. In this approach [68], the authors applied to pre-processing on the image which improved the contrast of the images and removed the noise. The authors observed that the pre-processing has a major impact on the overall efficiency of the method. They took 100 images for tanning while using SVM as classifier. After pre-processing, image normalization and color conversion were performed. Afterward, feature extraction was done using the principal component analysis method. Finally, SVM was applied which had the training and testing data to classify the images as glaucomatous or non-glaucomatous. In this approach, the final results were stated as follows SVM classifier has accuracy rate 96%, sensitivity 100%, specificity 92%, positive predictive accuracy 92.59% and negative predictive accuracy 100%. Silve at el. [69] proposed the use of a hybrid feature set made up of texture and structural features of retinal image for accurate glaucoma detection. Optical Coherence Tomography (OCT) was used for early-stage detection of glaucoma and ML approaches were used for automated glaucoma detection whereas KNN shows its 90% accuracy in results.
Akram et al. [70] presented an approach that relies on the extraction of features from the OD region. Both spatial and spectral features were extracted to construct the multivariate model and LFDA is used to conduct a supervised intensification of features. The recognition of m-medoids is a cluster of data, another component that was added to analysis was ROI. The authors erect multifaceted with MMM to codify glaucoma. The proposed approach was compared to other classifiers such as SVM, GMM, KNN, and Multilayered Perceptron's. On average, their approach attained 91.7% accuracy, a positive predictive value of 78%, 92.7% specificity and a sensitivity of 84.6. Moreover, GMM is another classifier that achieved almost similar results as LFDA with a Gaussian distribution of 620. The main focus of the study is to enhance the performance of classification by integrating feature enhancement through supervised learning using LFDA.
Acharya et al. [5] presented an approach to extract features of the whole image for glaucoma classification. In the first step, images were converted to gray-scale via adaptive histogram equalization. Thereafter, to obtain textons, a total of four filter banks were utilized respectively. Then, features were extracted via local configuration pattern and this was to be done by using previously extracted textons. Authors used four filters and after extraction of results from all of them authors decided one between them. The authors obtained a total of 17808 features. Therefore, the authors wants to reduce this number so they decided to apply a new method. By applying a sequential floating forward search they decided to select the set of features. Afterwards authors applied student t-test to rank the selected features and choose those features that have a p-value of < 0.05. They used total of Five classifiers, namely discriminate classifier, KNN, decision tree, probabilistic neural network and SVM with 10-fold cross-validation applied. The proposed approach resulted in 95.7% accuracy, a sensitivity of 96.2% and a specificity of 93.7%.
In this paper [71], choose mass screening because it would diagnosis glaucoma in early stages in the vast population. An automated diagnosis technique is required for mass screening. Authors proposed an automation that is consists of optimal wavelet transformation, preprocessing, ,classification modules and feature extraction. From fundus images statistical features are extracted via hyper analytic wavelet transformation (HWT). As phase information is preserves by the HWT and is appropriate for feature extraction. Features are classified via SVM with the help of a Radial Basis Function (RBF) kernel. To avoid premature convergence area screening and Group Search Optimizer (GSO) are embedded in Particle Swarm Optimization (PSO) framework. hybrid algorithm increased the fitness by 9.4% and 6.1%, in the test runs as Compare to conventional PSO. This success rate is achieved with the help of search ability of PSO. Furthermore, algorithm also maintains PSO population's identity.
In this paper [72], the authors discussed the automatic identification of glaucoma and normal classes with the help of Discrete Wavelet Transform (DWT) and Higher Order Spectra (HOS) features. Then features are given to the SVM classifier either in linear or polynomial order that is 1, 2, 3 and Radial Basis Function (RBF) to select the best kernel function for the selection of automated decision making. to identify the normal and glaucoma images automatically. Authors use Support Vector Machine classifier with the kernel having polynomial order 2 and achieve an accuracy of 95%, sensitivity and specificity of 93.33% and 96.67% respectively. In this paper [73], the authors used automatic screening of Glaucoma from fundal retinal images in a person via complex wavelet features and trispectrum. At first, gray-scale conversion is done and then Random transform and histogram equalization is done to enhance the input retinal image. optic disk segmentation is carried out using region separation and binary conversion. After that features are extracted with the compensation of trispectrum, complex wavelet and segmented area features. For the classification of the images to check whether it's normal or glaucoma training of neural network will be processed via conjugate gradient descent algorithm. Rim-One database is used in this experiment. comparison was also made in this paper with the exiting methods and Proposed technique has outperformed existing method by having better accuracy value of 81%.
3.1. Fundus image applications of ML
Lim et al. [74] implemented the deep learning approach on OD segmentation. This technique allows to obtain correctness of OD segmentation and that is used to govern the most biding situations for manual examination. Maninis et al. [75] inspected fundus images to segment OD and blood vessels respectively. Al-Bander et al. [76] segmented both fovea and OD and used the technique based on deep learning approach with an accuracy of 97.5%. Mitra et al. [77] recently reported some drawbacks of Al-Bander et al. [76] work while arguing that he uses gray-scale images data loss occurs. To avoid these drawbacks like in [75][76] of OD segmentation, the authors used batch-normalization in CNN. Mitra at el. [77] took fundus images to detect glaucomatous OD. This technique used a trained model with dataset i-e ImageNet and its output is replaced by the new layer for output 2 nodes for glaucoma and normal classes respectively. When this work was compared with the pervious work we found out that this work gathers a large amount of data and is considered the most robust one compared to the rest of the works.
FCNN [78] is a very famous and re-known technique for classification-based problems associated with biomedical images [79]. X Sun et al. [80] make use of much faster RCNN architecture for OD segmentation from fundus images that are based on deep object detection architecture respectively. Figure 9 below shows the flow of basic neural network and how it works.
Ghassabi et al. [81] integrated approach of cup segmentation and ONH using SOM i-e self-organizing maps for the assessment of glaucoma.it was an effective approach when there were unclear NRR, low intensities of OC, and non-obvious para-papillary shrivel. Table 4 represents a summary about various techniques that are discussed above.
Table 4.
Summary of glaucoma detection approaches based on ML techniques.
In this section different segmentation approaches will be discussed i.e. optic disc segmentation approaches, optic cup segmentation approaches, and a combination of optic disc and optic cup segmentation. While in most of the published research, optic disc segmentation approaches are described and in others, only optic cup segmentation approaches are used while some other published research models have only concerned with a mixture of the optic cup and optic disc segmentation approaches. But here we have focused on the optic disc, optic cup and their mixture of approaches.
4.1. Optic disc segmentation approaches
According to Cheng et al. [84], segmentation of disc is very crucial in the detection of different applications concerned with computer-aided systems, including detection of glaucoma using fundus images. Ahmad et al.[85] stated that the segmentation mainly depends on searching disc pixels that are mostly centered. It has been specially studied for the It has been extensively studied for solicitation in the diabetic shielding procedure in Kavitha et al.[86]. Zhang et al.[87] stated the problems related to segmentation and the disc is track down. According to experimental results, the proposed research outperformed the state-of-the-art to detection of glaucoma because the segmentation techniques measures the disc boundary which is a critical task because there are several clotting between blood vessels.
Welfare et al. [88] proposed an adaptive method for the segmentation of fundus images. In this research the experiments are performed on two different sets of databases i.e. DRIVE and DIARETDB1 with a dataset of 89 images and an accuracy of 100% is achieved. Tjandrasa et al. [89] applied Hough transform at the initial level for the segmentation of optic disk. According to Tavakoli et al. [90], in retinal analysis the identification of OD or ONH are important for vessel tracking-ness and for localization of anatomical objects. In the fundus images ONH are brightest in appearance therefore localization is done with high intensity pixels with high gray scale values. Tavakoli et al. [90] presented three retinal vessel segmentation techniques for automatic ONH detection. The three databases named Mashhad University Medical Science Database (MUMS-DB), DRIVE and CHASE DB1. Although the fundus images are colored images but green channel has been used as it has clear contrast. Morphological operations have been used to increase the contrast between vessel and background as a pre-processing [90]. Image is inverted to deal with the background brightness variation in the whole fundus image so that it doesn't cause false or missed vessel detection. Multi-overlapping window has been used, its size and window overlapping have great impact on method's accuracy. For retinal vessel segmentation the edge detector used are Laplacian of Gaussian, Canny and Matched filter. Sensitivity and specificity are used as an evaluation criteria, higher values of both parameter indicates the best performance.
In an automatic diagnosis, retinal vessel detection and segmentation is a basic step and is very important subject in the field of medical image processing [91]. Although numerous methods and techniques have been proposed but none of them is fast enough to be used in real automatic detection as manual detection is very tedious process. So Bibiloni et al.[91] developed an algorithm for real-time retinal segmentation based on fuzzy morphological operations by taking into consideration the usage of cost and operations. Retinal images are being used in the detection of many medical problems such that glaucoma or diabetic retinopathy. Retinal images are acquired from fundus camera. These images are very noisy, blurred and low contrast so there is uncertainty which can cause lack of information and indefinite. To enhance contrast authors used top-hat transformation and by taking the advantage from fuzzy logic and set theory in fuzzy mathematical morphology has been used to handle it the uncertainty present in the images. The DRIVE and STARE databases has been used and precision, accuracy, specificity and sensitivity are used as a performance evaluation criteria. The results have been evaluated and compared with other methods of state-of-the-art. According to Bibiloni et al. [91] proposed method improves segmentation and has better performance when comparing with real-time systems. Agarwal et al. [92] presented an automatic techniques for strategic OD Segmentation. They used morphology-based and edge detection techniques for correctly segmentation of OD and snake-based active contour fitting has been used for exact boundary detection. For experiments 60 fundus images were used. Computational time and overlapping score have been calculated between segmented optic disc and ground truth [92]. Results shows that proposed method has more than 90% overlapping score which shows the accuracy of the method and very helpful for automatic screening of patients.
According to Pal et al. [93], the segmentation of OD in retinal fundus images is important in diagnosis process for the detection of many diseases. Pal et al. [93] presented a technique for OD segmentation based on morphological operations that has been applied on different channels of the fundus images. Contrast Limited Adaptive Histogram Equalization (CLAHE) has been performed and discussed for the choice of clip limit to enhance the contrast. RIM-ONE dataset of 290 images of glaucoma has been used. To make segmentation process effective by using simple canny edge detection approach various morphological techniques have been combined. The performance has been evaluated on ground truth and the parameters such as sensitivity, specificity, accuracy and overlapping ratio are used [93]. The comparison has been performed with state-of-art methods. Table 5 below summarizes all the discussed techniques for OD segmentation.
Table 5.
Summary of segmentation-based approaches.
Wang et al.[94] proposed model based segmentation for early detection of glaucoma. In this method the images are taken locally not globally as it works efficiently by taking a small region of the defective area. Fourier transform is used to detect the center of glaucomatous images by taking the linear operations as dimensional function of gaussian filter while the multiple features are taken as superposition of the gaussian filter Deng et al. [95]. Then features are separated by K-Means clustering technique. Different models are selected based upon the size of the block which varies and it depends on the researcher criteria. The akakie information criteria is used to select the best model over the rest. The dataset used here is 20 and it is obtained by STARE dataset. The image is first converted in to gray scale. The model used is cheap and easy to use but it has some difficulty as it is quite robust in its functioning and probably uncertainty occur due to low scale ratio. The algorithm works best with the specificity of 93% and having a sensitivity factor of 82%.
According to Vermeer et al. [96] there are so many methods that are used for the early detection of glaucoma in fundus images but the result are not always accurate and satisfactory but he present a unique model for this purpose. The approach used here is the Laplace. The image used in this model are selected by Gx, a laser based device and it selected two kinds of images. The blood vessels obtained by this method are of both types i.e. light as well as dark. A gaussian filter is applied here to smoothen the corresponding image. This model first find out the dark portion and then comes to the inner portion. Laplacian filter is applied and objects are divided by thresholding one to one, two or more objects. Noise is removed by separating small objects but this process will remove fine details of important features, so to sort out this problem fragment vessels and then perform separation. Orlando et al. [97] proposed a conditional random field model for the diagnosis of glaucoma. The different parameters are used here and are learned automatically by the help of vector machine. Different datasets are used for this purpose. The summary of the model-based approaches is shown in Table 6.
Segmentation in the optic cup is more difficult than the segmentation of optic disk because the density of blood vessels is high in the optic cup compared to the optic disk. Moreover, the gradual change of intensity in the optic cup and NRR make the segmentation more difficult in optic cup. When a person has glaucoma then due to its presence cup size automatically changes. According to Ingle et al. [98], the detection of glaucoma CDR is a crucial method having a region of the cup to disc. In this proposed methodology region of the cup is detected from the dataset. ROI varies differently in different images. An automatic cup region procedure has been used here in this research[98].
Joshi et al. [99] proposed an automatic technique for early detection of glaucoma. In this research, an automatic OD parametrization technique is applied to segment monocular retinal images. Cup segmentation is done by vessel bends which are quite close to the detection of glaucoma. There exist a variation between the mathematical and photometric changes. The OC segmentation obtained in this way is quite efficient and effective. Damon et al. [100] proposed a method for automatic detection of the optic cup fused with kinks Finkelstein et al.[101] approach. Here blotches are extracted from the OD by the help of kinks and for vessel detection, the color and edges are determined with the help of patches. Xu et al.[102] proposed super-pixel-based framework by using segmentation of optic cup in glaucoma detection. Xu et al. [103] applied super pixel and different tasks are formulated with their solution using LRR approach. Table 7 concludes the above studies.
Table 7.
Summary of optic-cup segmentation approaches.
In computer vision and image analysis, fusion-based approaches are used to enhance the performance of image classification model [104,105]. According to Ho et al. [106], many features are extracted for the identification and detection of glaucoma. To differentiate between optic disc and optic cup histogram, dissemination is calculated from the previous studies of optic images proposed by Chang et al. [107]. The active contour model is used for early detection using different vessels. Here for the identification of major peaks erosion and dilation techniques are used. Moreover eroded and dilated image differentiation was used for the extraction of the main peak using the reverse histogram. Wong et al. [108] proposed an automatic technique for the detection of glaucoma i.e. ARGALI and it is used for the segmentation of optic disc and optic cup. This paper makes use of SVM and NN fusion instead of ARGALI. Here the results show that SVM has greater consistency compared to NN which proves that SVM could be the best option in the replacement of the ARGALI approach. Figure 10 shows the architecture of SVM based technique.
Figure 10.
Simple SVM based technique architecture.
Yin et al. [109] proposed a statistical model-based methodology for performing the segmentation of OD and OC. This method makes a fusion of Hough transform and optimal channels for OD segmentation. It is then enhanced to OC segmentation which is assumed to be a more critical and time taking task. Here 325 images are used for testing. The results obtained by using these approaches are promising efficient because the methods used are quite intelligent. Chandrika et al. [110] proposed K-means pixel clustering technique and gabor wavelet transform to segment images and CDR is calculated using different parameters of quantitative measures. In this research, the normal CDR ratio is 0.3 and if the CDR ratio enhances from 0.3 then it is considered to be glaucomatous otherwise it is considered to be non-glaucomatous. Annu et al. [111] proposed a Probabilistic Neural Network(PNN) for the segmentation of optic cup and optic disc. PNN classification is further divided into two different phases i.e. training and testing. The accuracy encounter here is 95% which shows the effectiveness of this approach. Fu et al. [112] proposed a deep learning approach that is M-Net which solves the problem of OD and OC segmentation together on a single system. It has several layers and these layers are responsible for the formation of the pyramid of image. Polar transformation Dhar et al. [113] also used it. The dataset used here was ORIGA.
Wong et al. [114] formulated a morphological attribute for identification of glaucoma which is termed as kinks. The different features are extracted from the images with help of edge detection approaches and the different edges of the vessels provide kinks and consequently with the help of these kinks the cup and disc surroundings are measured. The results show that these kinks provide very little error in measurement which proves that these methods are quite beneficial to use. Figure 12 illustrates the basic deep learning architecture that is being used for the detection. Murthi et al. [115] proposed an automatic approach for the extraction of features from colored images. For the detection of glaucoma, CDR is a core feature that should be calculated very carefully. An algorithm is proposed here which is used to enhance the accuracy of the boundary and is known as least square fitting. ARGALI method is used for the segmentation of optic disc and optic cup together. This algorithm improves the clinical elucidation of glaucoma and all implementation is done in MATLAB. According to Khalid et al. [116], CDR or measurement is important for the detection of glaucoma in fundus images and due to the ambiguous color pattern between the optical cup and the disk and its calculation is difficult. Khalid et al. [116] proposed Fuzzy c-Mean(FCM) techniques for the segmentation of optic cup and disc with aim to measure CDR. From the RGB fundus images the green channel has been used as green channel has better contrast than others. Basic morphological operations such as erosion and dilation are used to reduce vernacular. The elimination of vernacular part plays vital role to get effective segmented results and the values of accuracy, sensitivity and specificity are improved. Ground truth and ROC analysis have been used as an evaluation of proposed method for segmentation [116]. The result shows the effectiveness of FCM segmentation algorithm with morphology.
According to Nugroho et al. [117], the retinal abnormality such as glaucoma is diagnosed by measuring the CDR (cup-to-disc ratio) so segmentation of both optic cup and disc is important. Two approaches for segmentation of fundus images are proposed by Nugroho et al.[117]. For optic disc segmentation active contour with integration of reconstruction of morphology has been used and for optic cup segmentation reconstruction of morphology with convex hull has been utilized [117]. In segmentation process retinal vessel removal for cup segmentation and channel extraction with some contrast improvement to reduce noise for optic disc segmentation has been performed. The experiments have been performed on 50 colored images of DRISHTI-GS dataset. The parameters such as accuracy, sensitivity and specificity are used to evaluate the performance of the proposed approach [117]. Table 8 summarizes the methodology, datasets and the metrics that are used for the evaluation of the research models that are discussed here.
Zhang et al. [118], presented a segmentation algorithm for recognition of vessel or non-vessel pixels based on a texton dictionary. In contrast to the previous researches that learned filter parameters from image pixels that were labeled manually, the proposed approach derived filter parameters from a small collection of image features termed as keypoints. Gabor Filter bank was used for the extrication of keypoints that represented features based on a SIFT inspired approach. The keypoints were determined from the validation set, the seeds derived from the validation set are used to initialize k-means clustering technique which is ultimately used to create a texton dictionary for some other training set. Testing is performed using NN classifier and DRIVE database is used for evaluation. The experiments demonstrated that the robustness of extracted clusters based on kerpoints as compared to the manually labelled pixels. The textons mitigate the problems associated with inter and intra-observer variability and appeared to be a better representative of vessel/non-vessel classes. According to Li et al. [119], ground truth is critical for evaluation of segmentation algorithms, however, the ground truth estimation from manual segmentation collections remains a challenging task. To deal with inter-rater variation there is a need to create a proper estimation approach. The authors performed an analysis in order to develop a better understanding and investigate the significance of patterns collected from manual segmentation on ground truth. The authors proposed an approach based on level-set to estimate ground truth using a probabilistic formulation. Segmentation is considered to be the first step according to the authors for computer-based applications and its accuracy is also very important. Computer-based skin lesion procedures that are used for the segmentation also become crucial. Preceding design information is used in the model. Experiments were done on both artificial as well as natural data and it is proved that preceding information gives more close, original and accurate results toward GT.
Ricci et al. [119] introduced vessel segmentation for the diagnosis of the retinal diseases that are computer-aided i.e. glaucoma. Here a line detector is applied on the green channel from RGP of the retinal image. The purpose of using the green channel is it provides the best result towards vessel background. The vessels have a shiny texture compared to backcloth. Moreover, the vessels seem blazing over the green channel so it is preferred here to use a green channel. Two different segmentation procedures are also used here. Both the segmentation procedures give efficient results. Fraz et al.[120] proposed a supervised algorithm for segmentation of retinal images. A supervised algorithm is used for the instinctive process of segmentation. Two databases that are DRIVE Niemeijer et al. [121] and STRIVE Choi et al. [122] are used for image manipulation as well as the segmentation of retinal fundus images. The algorithm has a training image at the start, then vessel features are extracted from the training image and samples are labeled. The labeled samples have manual segmentation initially then these labeled samples are trained with the help of the classifier. Then test images are taken and again vessels are extracted the same as extracted for the training images. Then vessel features and classifiers together perform vessel classification and finally the image is segmented. The average accuracy achieved is 95.79%.
According to Zhang et al.[123], vessel segmentation problems in the retinal fundus images are considered to be a very challenging and critical scenario. Here machine learning paradigm is proposed for the identification of the texture features of the retinal images that are termed as texons. This paper also proposed a filter bank that is used for the extraction of features, edges and different changes that occur in an image. DRIVE and STARE databases are used and give efficient results when 20 images are encounter here. Table 9 summarizes the methodology, datasets and the metrics used to evaluate the research discussed.
According to Septiarini et al.[124], the area of RNFL has textured surface with a number of streaks or ridges of white color and their texture is uniformly disturbed and clear in normal human eye but in the eyes of person having glaucoma the layer of nerve fiber disoriented. Septiarini et al.[124] proposed a technique based on texture attributes for automatic detection of RNFL. Co-occurrence matrix has been acquired from the outer area of optical nerve head and whether to remove blood vessels they are considered as the detection of RNFL is easy in this part. For the detection of RNFL the different classifiers are applied but Back Propagation Neural Network (BPNN) classification approach get better accuracy [124]. 40 retinal fundus images have been used in experiment and for the evaluation purpose sensitivity, specificity, and accuracy are calculated.
Kirar et al. [125] proposed a computer aided diagnosis (CAD) method for glaucoma from fundus images based on Empirical Wavelet Transform (EWT) and Discrete Wavelet Transform (DWT). After pre-processing, the input images are decomposed by EWT, DWT, EWTDWT and DWTEWT operations and feature layers has been formed which is concatenated in next step and then normalization and selection of features has been performed. To find out best features these selected features are given in to SVD(Singular Value Decomposition) and then classification with SVM has been performed [125]. 505 images of MIAG image database has been used and many performance evaluation parameters like accuracy, sensitivity, specificity, Negative predictive rate (NPR), Rate of Negative Predictions (RNP), Matthews Correlation Coefficient (MCC) etc. are used [125]. The performance comparison of the proposed method and classification comparison with other methods has been reported. According to Nirmala et al.[126], the presence of glaucoma cause vision loss so Nirmala et al. presented wavelet-based contourlet transformation (WBCT) method for detection of glaucoma. To enhance the image contrast adaptive gamma correction technique with weighted distribution has been used. OD is segmented and blood vessels are removed by using Gabor filter. Coefficients are extracted by applying wavelet-based contourlet transformation on segmented OD region and then features are extracted from these computed coefficients. For detection of glaucoma SVM classifier is used [126]. The experiments have been performed on FAU dataset and for evaluation purpose accuracy, specificity, sensitivity and positive predictive ratio are evaluated but only accuracy has been compared with some other methods.
According to Elseid et al. [127], low contrast of RNFL in fundus images cause defects in early detection of glaucoma. Examiner may be miss the accurate detection as it relies on the experience of the inspector who is examining. So Elseid et al. [127] proposed computer aided algorithm based on RNFL texture features that accurately detects the defect in very early stage of disease. OD segmentation has been performed and then from this segmented region texture features are extracted and selected for classification purpose. For validation of the proposed algorithm RIM-ONE dataset has been used. The ROC (Receiver Operating Characteristic), accuracy, specificity and sensitivity are used as performance evaluation [127]. The reported result shows the effectiveness of the proposed algorithm that it has high accuracy. According to Claro et al. [128], the examination of glaucoma can be done in either two ways as periodic examination in which specialized instruments are necessary or it can be done via examining the fundus image through the application of digital image analysis. Claro et al. [128] presented an automatic detection approach for glaucoma. The segmentation process for optic disc has been performed fist then texture features are extracted in different color modes of image based on gray-level co-occurrence matrix and entropy [128]. Lastly, the classification by using different classifiers has been done. The DRISTHI-GS, DRIONS-DB and RIM-ONE datasets are used comparison has been performed for evaluation of the proposed method.
By taking advantage of multi-resolution wavelet transformation of localization in frequency as well as time Abdel-Hamid et al. [129] presented an approach for real time screening of glaucoma patients. The proposed method acquire textural wavelet and statistical features from the optic disc. Then selection of these feature has been performed and kNN classifier is used for classification of healthy or non-healthy fundus images [129]. GlaucomaDB and HRF datasets has been used to validate the proposed method. Specificity, sensitivity and accuracy have been computed and used as performance evaluation criteria. The reported result shows that the methods in computationally less expensive for processing of high resolution images and suitable for real time system [129]. Table 10 summarizes the methodology, datasets and the metrics used to evaluate the research discussed.
Table 10.
Summary of texture feature-based approaches.
7.
Optical coherence tomography(OCT) based approaches
Approaches based on OCT are commonly employed for the diagnosis and monitoring of glaucoma. OCT based techniques are applied by many researcher using different classifiers and great results and high accuracy is obtained [130]. Further more the incorporation of IOP measurements and visual field tests, performed along with OCT are highly related with glaucoma detection, these factors likely contributed to their higher prediction accuracy. Despite of the development and the advantage of OCT most of the clinics and physicians are not taking proper benefit of this powerful and efficient tool. The OCT scans of optic discs are widely taken and the scan of the parts that contain macula are usually missed due to which vital parts to be examined are missed [131]. Using OCT not only the existence of RNFL damage is uncovered but the pattern of RFNL damage is also analyzed. OCT scans if done properly can be used for screening [132]. The following table 11 summary for OCT-based approaches.
8.
Glaucoma classification via deep learning approaches
In last few years, DL based models are used in various applications relevant to medical image analysis [133,134,135]. Chen at el. [134] stated that CNN are associated with DL systems that are used by the authors to discriminate between the glaucomatous and non-glaucomatous images in the proposed study. CNN net contains six layers out of which the first four are convolutional and the two are fully connected. The output of the last fully connected layer for glaucoma prediction. The authors applied the pre-processing on the image to remove the extra illuminations and get a better image. Later on, ROI is specified for the image and further computation is done on ROI. The proposed architecture of CNN consists of 6 layers through which the image is passed and final evaluation is performed. Finally, dropout and data augmentation techniques are applied to the ROI to enhance glaucoma detection accuracy. The ORIGA dataset was used which comprised of 168 glaucoma and 482 normal fundus images. The SCES dataset contains 1676 fundus images, and 46 images are glaucoma cases. The results showed that the technique provided greater results than other states of art methods applied to detect glaucoma. ORIGA dataset images were used for training and testing: 99 images were used for training and the remaining 551 images were tested.
Nguyen et al. [135] presented an unsupervised technique that is a line detector for the extraction of blood vessels. The authors focused on each pixel's position and calculate it's the corresponding gray-scale and also assigned a window over there. In the region, twelve lines were drawn down and it is noticed that all of them passing through its center. Thereafter, across each line it's average gray levels were computed. The winning line is the one with more votes and had the highest gray value amongst all and was utilized in the segmentation of blood vessels. The line detector had some draw backs i.e. it gravitate to merge close vessels. It produces an appendix at crossover points. It produces false vessel responses at background pixels near strong vessels. These drawbacks are addressed by varying the length of basic line detector. Three publicly available datasets i.e. DRIVE, STARE, and REVIEW were used for the evaluation of the proposed method. The proposed method had false positive detection i.e. 515 around the OD region. To address this problem a post-processing step can be applied to reduce the false positive and improve the performance of the method even more.
Feedforward neural network (FNN): JH Tan et al. [136] Drive dataset is used with 40 glaucoma images. First of all image normalization is done before any classification that is a pre-process applied by the authors. The CNN proposed consist of 7 layers. Standard back propagation technique is used to train the data. The learning rate is set accordingly. The algorithm performs testing at the end of each training on the DRIVE dataset which achieves the accuracy of 92.68%. CNN based methods: Chen et al. [134] were the first who use a deep learning approach for glaucoma detection. He implements the CNN and CNN with dropout and CNN with 4 convolutional layers that are six-layer deep with filter size (11, 5, 3, 3) that is followed by the 2 dense layers. Chai et al. [137] dispense a structure on the dataset of fundus images that are acquired from different hospitals by integrating both feature learned and domain knowledge from the deep learning model. Authors implied 2554 retinal fundus images from 1542 patients, out of which 1023 images are from glaucoma patients and 1531 are from people non-glaucoma patients. The authors proposed a method in which the advantage of both deep learning and domain knowledge is taken. A multi-branched neural network model is proposed to achieve the glaucoma detection. A combination of Faster RCNN, CNN and Fully convolutional network is used.
Auto encoder: Pal et al. [138] prove his architecture to be more robust and given results even in low-quality images i.e. G-Eye Net architecture. Pal et al. proposed a multi- task learning procedure to detect glaucoma. Authors obtained the ROI from a colored fundus image and applies a preprocessing step using contrast limited adaptive histogram equalization. An encoder De-coder framework is applied to the data to get automated glaucoma detection i.e. G-EyeNet. This framework consists of three main parts which are 1. Encoder 2. Decoder 3. Classifier. Authors used HRF, RIM ONE v.3., DRISHTI-GS and DRIONS-DB datasets. 400 distinct images were obtained using all these datasets. The G-EyeNet showed great and significant improvement as compared to the state-of-art deep learning approaches Different glaucoma detection technique's results summary is given in Table 12.
Table 12.
Summary of deep learning-based approaches.
The authors used a deep learning algorithm to detect glaucoma [139]. The algorithm named Residual Network was used and a training dataset was used. The size of dataset used was 1364 color fundus images. Image augmentation was done and a prominent increase in the accuracy was noted. The accuracy of 94.8% was calculated using the first dataset. When the same technique was applied to the second dataset the accuracy was calculated to be 99%. Hence the authors concluded that previously used DL methods can achieve higher accuracy using image augmentation. Authors introduced an automatic feature leaning based on deep learning with the usage of convolutional neural network for feature learning [140]. As traditional convolution layers use linear filter followed by non-linear filters, this convolution neural network uses more complex structures to identify the features and introduced a deep learning structures to discriminates between glaucomatous and non-glaucomatous. The datasets used by the authors are SCES and ORIGA. The authors concluded that the accuracy is much better than the state-of-art algorithms. The proposed method gives the accuracy of 89%. This method is much more efficient then the traditional methods as authors argued that the deep learning approaches gave higher results then the state-of-art approaches. Authors argue that it is difficult of compare deep learning algorithms and techniques used for glaucoma detection as many of the deep learning techniques used by different researchers use private datasets and the ones who use public datasets face the issue of overfitting as the size of datasets in many cases is small and the deep learning algorithm cannot work efficiently on small size datasets [141]. An existing model called inception-v3 was finetuned by using different methods and approaches. Data augmentation along with Contrast Limited Adaptive Histogram Equalization and normalization were applied on the dataset called RIM-ONE.Hence the results showed that problem of overfitting was addressed and now the method was more effective.
Authors proposed a deep learning approach based on convolution neural network [142]. This technique involves transfer learning and local binary pattern-based data augmentation. The CNN model used by the authors is Alex net which is responsible for transfer learning. The proposed CNN model is trained via transfer learning and the images are classified as glaucomatous and non-glaucomatous with the help of level fusion technique. They uses RIM-ONE dataset. Authors argues that the performance of the proposed method shows prominent improvement by applying LBP-based augmentation technique. Authors also adds that the approach gives promising results and these deep learning methods helps the experts during mass screening making it automatic and very accurate. Authors give a brief overview of deep learning methods and traditional methods that are used in glaucoma detection [143]. Authors states that deep learning approaches outperformed traditional methodologies in terms of accuracy and reliability. These automatic detection techniques are also considered as support tools for the clinicians. Table 12 below summarizes the approaches.
9.
Evaluation criteria for the glaucoma detection techniques
Automatic detection results are evaluated by using large variations of evaluation metrics. The metrics used for the evaluation must be properly defined and should co-related to the domain of the system which is under consideration. A confusion matrix as shown in Figure 12, commonly referred to as the contingency matrix or error matrix [147] is a table that gives the visual representation of the performance of an algorithm. This matrix is commonly used in the field of machine learning and classification algorithms. Figure 12 shows the confusion matrix.
9.1. Performance metrics from the confusion matrix
Some of the terms used in the confusion matrix need a little explanation for the better understanding of the evaluation [67]. Some important terms are described below and their formulas are given in Table 13.
1. Accuracy: is termed as the percentage of the correct predictions made.
2. Precision: also termed as Positive Predictive Value and determines the percentage of correctly classified positive instances.
3. Negative Predictive Value: gives the percentage of correctly classified negative instances.
4. Recall: also referred as sensitivity, gives the percentage of actual positive instances identified correctly.
5. Specificity: the percentage of actual negative instances identified correctly.
6. True positive TrP: refers to correct prediction of event values.
7. False-positive FlP: determines incorrect prediction of event values.
8. True negative TrN: is used for correct prediction of no-event values.
9. False-negative FlN: gives incorrect prediction of no-event values.
9.2. Datasets available for retinal images
Most of the techniques that are discussed in this survey use the publicly available datasets. These datasets are mentioned in the table below which gives a brief description and understanding of the datasets. These are the datasets which are commonly used by the researches to carry out the proposed approaches and conclude the findings. Table 14 summarizes the datasets that are available and can be used for glaucoma detection research.
Table 14.
Datasets available for automatic glaucoma detection.
This dataset is partitioned into training and test image sets respectively. The training set comprise of 50 images, with ground truth providing information regarding region boundary, segmentation soft map and Cup-to-Disk ratio for each image. Whereas the testing set contains 51 images and the ground truth for test set is available only upon registration.Image size and resolution is 2896×1944 pixels and PNG uncompressed image format is used. The retinal image type of the dataset is Fundus Images .
HRF is publically available database consisting of high-resolution fundus images. The dataset is composed of three sets of fundus images: comprising of the healthy retinas, glaucomatous and DR retinas. In the first set, 15 images of healthy patients are included which do not contain any instance of retinal pathology. The next set is comprised of 15 retinal images of patients with DR exhibiting pathological changes, such as bright lesions, neovascular nets, spots after laser treatment and hemorrhages. The last set includes 15 images of patients with advanced stage glaucoma and showing symptoms of focal and diffuse nerve fiber layer loss. The retail type of the dataset is Fundus images. The image quality of the database is 3304 x 2336 pixels.
There are 99 fundus images taken from 50 patients.The dataset type is Fundus images. Images were captured using Kowa VX-10 alpha digital fundus camera with 50 degree field of view (FOV). The images have resolution of 4288x2848 pixels and are stored in jpg file format. The size of each image is about 800 KB.
IDRiD is a recently introduced retinal fundus image database publicly available and consists of 516 images organized in two categoeies: 1.Retinal images that exhibit symptoms of DME and/or DR 2. Normal retinal images without any indications of DME and/or DR.
The RIGA dataset comprises of 750 images collected from three sources and 4500 manual marked images. The images are saved in JPG and TIFF format and contain both glaucomatous and normal images.The resolution of 3 sources are a. MESSIDOR 2240x1488 PIXELS (460 Images) B. Magrabi eye center Riyadh 2743x1936 (95 Images) c. Bin Rushed Ophthalmic center in Riyhad 2376x1584 (195 images) The retinal image type of the dataset is Fundus images
It comprises of 110 retinal images with image resolution of 600×400 pixels and the optic Disc annotated by two experts with 36 landmarks.The resolution of the dataset is 600x400. The retinal image type is Fundus images .
It consists of 258 images annotated manually alonwith their clinical diagnosis details. The images in ACHIKO-K are collected from patients suffering from glaucoma and contain rich information related to glaucoma pathological symptoms , e.g. optic nerve drusen, hemorrhage and optic cup notching, etc. Multiple images are taken from patients suffering from glaucoma having age between 27 to 85 years during 2008 to 2012 time span. A total of 258 fundus images are available in database, where 144 are from glaucoma tic eyes and 114 are captured from normal eyes.
The dataset consists of 2258 individuals with genome data, image data and clean screening data. Among the 2258 subjects, 100 were diagnosed to have glaucoma and the rest were normal.
10.
Deep learning approaches vs traditional approaches
A brief comparison has been conducted among the approaches with deep learning methodologies and traditional methodologies. Mostly the results of DL approaches seem to have high accuracy and performance. In the domain of image processing and method automation DL has pushed the limits and have given extra ordinary results. ML exhibited high sensitivity and specificity for the detection of glaucoma and hence traditional approaches are considered less effective than the former one. DL is largely based on Artificial Neural networks which is a brain like structure and works according to how brain does. Since the trained neural networks are used and are not programmed hence they give greater accuracy than the traditional approaches. Table 15 summarizes the study and shows the quantitative results.
Table 15.
Summary of deep learning approaches vs traditional detection approaches.
Medical imaging systems are used to create the pictorial representation of human body to monitor various types of diseases and public health-care systems reply on these technologies. Digital image processing and computer vision techniques are used in various health-care systems to detect the diseases. Glaucoma is a type of persistent eye disorder that gradually damages the optic nerve and can lead to permanent blindness. The inappropriate intraocular pressure within the human eye is reported as the main cause of this disease and glaucoma is declared as the second leading cause to the loss of sight. The use of image analysis with CAD tools can assist the clinicians to execute the work more effectively. The main aim of this review is to present and discuss different types of research models that can perform automated glaucoma detection. We have discussed optic disc segmentation approaches, optic cup segmentation approaches, machine learning-based approaches, fusion-based approaches, multi-scale-based approaches, texture features-based approaches and current trends about deep learning models. A comprehensive detail about the types of standard benchmarks, number of samples used in the referred research and performance evaluation criteria is discussed in detail. 1) Feature fusion based approaches have shown an effective performance than the use of one feature. 2) The hand-crafted features have shown an effective performance while using small-scale image benchmarks. 3) The experimental results show the superior performance of deep learning algorithms as compared to the hand-crafted features. 4) The computational cost and training time of deep learning-based approaches is more as compared to traditional machine learning approaches with hand-crafted features. 5) It is also observed that performance of some algorithms suffers in case when the patient is affected with multiple diseases including glaucoma. Therefore, the use of deep learning-based approaches can enhance the effectiveness of professionals when they are dealing with patients suffering from multiple eye diseases. 6) The use of deep learning algorithms with large-scale fundus image benchmarks seems to be a possible research direction in this field. 7) It is also observed that most of the researchers have used their own fundus image benchmarks for the evaluation of their research and there are few publicly available images benchmarks. Due to this reason, it is required to create a large-scale publicly available image benchmark that can be used for the evaluation of future research. This will be beneficial to identify the best performance of upcoming research evaluated on the same data-set and it can lead to development of an efficient CAD for glaucoma detection.
Conflict of interest
The authors have no conflict of interest.
References
[1]
A. Krizhevsky, I. Sutskever, G. E. Hinton, Imagenet classification with deep convolutional neural networks, Commun. ACM, 60 (2017), 84-90. https://doi.org/10.1145/3065386 doi: 10.1145/3065386
[2]
M. Frid-Adar, I. Diamant, E. Klang, M. Amitai, J. Goldberger, H. Greenspan, GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification, Neurocomputing, 321 (2018), 321-331. https://doi.org/10.1016/j.neucom.2018.09.013 doi: 10.1016/j.neucom.2018.09.013
[3]
B. Bhattarai, S. Baek, R. Bodur, T. K. Kim, Sampling strategies for GAN synthetic data, ICASSP 2020-2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), (2020), 2303-2307. IEEE. https://doi.org/10.1109/ICASSP40776.2020.9054677
[4]
M. Mirza, S. Osindero, Conditional generative adversarial nets, arXiv: 1411.1784, 2014.
[5]
J. J. Bird, C. M. Barnes, L. J. Manso, A. Ekárt, D. R. Faria, Fruit quality and defect image classification with conditional GAN data augmentation, Scientia Horticulturae, 293 (2022), 110684. https://doi.org/10.1016/j.scienta.2021.110684 doi: 10.1016/j.scienta.2021.110684
[6]
V. A. Fajardo, D. Findlay, C. Jaiswal, X. Yin, R. Houmanfar, H. Xie, H., et al., On oversampling imbalanced data with deep conditional generative models, Expert Syst. Appl., 169 (2021), 114463. https://doi.org/10.1016/j.eswa.2020.114463 doi: 10.1016/j.eswa.2020.114463
[7]
T. Salimans, I. Goodfellow, W. Zaremba, V. Cheung, A. Radford, X. Chen, Improved techniques for training GANs, Advances in neural information processing systems, 29 (2016).
[8]
V. Dumoulin, I. Belghazi, B. Poole, O. Mastropietro, A. Lamb, M. Arjovsky, et al., Adversarially learned inference, arXiv: 1606.00704, 2016.
[9]
A. Odena, Semi-supervised learning with generative adversarial networks, arXiv: 1606.01583, 2016.
[10]
C. Li, T. Xu, J. Zhu, B. Zhang, Triple generative adversarial nets, Advances in neural information processing systems, 30 (2017). https://doi.org/10.1007/978-3-319-70139-4 doi: 10.1007/978-3-319-70139-4
[11]
A. Haque, Ec-gan: Low-sample classification using semi-supervised algorithms and GANs, Proceedings of the AAAI Conference on Artificial Intelligence,35 (2021), 15797-15798. https://doi.org/10.1609/aaai.v35i18.17895 doi: 10.1609/aaai.v35i18.17895
[12]
T. Chen, S. Kornblith, M. Norouzi, G. Hinton, A simple framework for contrastive learning of visual representations, International conference on machine learning, (2020), 1597-1607. PMLR.
[13]
X. Chen, K. He, Exploring simple siamese representation learning, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2021), 15750-15758. https://doi.org/10.1109/CVPR46437.2021.01549
[14]
K. He, X. Chen, S. Xie, Y. Li, P. Dollár, R. Girshick, Masked autoencoders are scalable vision learners, Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, (2022), 16000-16009. https://doi.org/10.1109/CVPR52688.2022.01553 doi: 10.1109/CVPR52688.2022.01553
[15]
A. Odena, C. Olah, J. Shlens, Conditional image synthesis with auxiliary classifier GANs, International conference on machine learning, 70 (2017), 2642-2651.
[16]
I. J. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, et al., Generative adversarial networks, Commun. ACM, 63 (2020), 139-144. https://doi.org/10.1145/3422622 doi: 10.1145/3422622
[17]
A. Radford, L. Metz, S. Chintala, Unsupervised representation learning with deep convolutional generative adversarial networks, arXiv: 1511.06434, 2015.
[18]
A. Brock, J. Donahue, K. Simonyan, Large scale GAN training for high fidelity natural image synthesis, arXiv preprint arXiv: 1809.11096, 2018.
[19]
X. Chen, Y. Duan, R. Houthooft, J. Schulman, I. Sutskever, P. Abbeel, Infogan: Interpretable representation learning by information maximizing generative adversarial nets, Advances in neural information processing systems, 29 (2016).
[20]
J. Donahue, P. Krä henbühl, T. Darrell, Adversarial feature learning, arXiv: 1605.09782, 2016.
[21]
T. Karras, S. Laine, T. Aila, A style-based generator architecture for generative adversarial networks, Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, (2019), 4401-4410. https://doi.org/10.1109/CVPR.2019.00453 doi: 10.1109/CVPR.2019.00453
[22]
J. Y. Zhu, T. Park, P. Isola, A. A. Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks, Proceedings of the IEEE international conference on computer vision, (2017), 2223-2232. https://doi.org/10.1109/ICCV.2017.244 doi: 10.1109/ICCV.2017.244
[23]
G. Perarnau, J. Van De Weijer, B. Raducanu, J. M. Á lvarez, Invertible conditional GANs for image editing, arXiv preprint arXiv: 1611.06355, 2016.
[24]
G. L. Grinblat, L. C. Uzal, P. M. Granitto, Class-splitting generative adversarial networks, arXiv preprint arXiv: 1709.07359, 2017.
[25]
S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele, H. Lee, Generative adversarial text to image synthesis, International conference on machine learning, (2016), 1060-1069.
[26]
H. Zhang, T. Xu, H. Li, X. Zhang, X. Wang, X. Huang, et al., Stackgan: Text to photo-realistic image synthesis with stacked generative adversarial networks, Proceedings of the IEEE international conference on computer vision, (2017), 5907-5915. https://doi.org/10.1109/ICCV.2017.629 doi: 10.1109/ICCV.2017.629
[27]
T. Kim, M. Cha, H. Kim, J. K. Lee, J. Kim, Learning to discover cross-domain relations with generative adversarial networks, International conference on machine learning, (2017), 1857-1865.
[28]
V. Dumoulin, J. Shlens, M. Kudlur, A learned representation for artistic style, arXiv: 1610.07629, 2016.
[29]
M. Patel, X. Wang, S. Mao, Data augmentation with conditional GAN for automatic modulation classification, Proceedings of the 2nd ACM Workshop on Wireless Security and Machine Learning, (2020), 31-36. https://doi.org/10.1145/3395352.3402622 doi: 10.1145/3395352.3402622
[30]
D. H. Lee, Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks, Workshop on challenges in representation learning, ICML, 3 (2013), 896.
[31]
Y. Grandvalet, Y. Bengio, Semi-supervised learning by entropy minimization, Advances in neural information processing systems, 17 (2004).
[32]
K. He, X. Zhang, S. Ren, J. Sun, Deep residual learning for image recognition, Proceedings of the IEEE conference on computer vision and pattern recognition, (2016), 770-778. https://doi.org/10.1109/CVPR.2016.90 doi: 10.1109/CVPR.2016.90
[33]
N. V. Chawla, K. W. Bowyer, L. O. Hall, W. P. Kegelmeyer, SMOTE: synthetic minority over-sampling technique, J.Artif. Intell. Res., 16 (2002), 321-357. https://doi.org/10.1613/jair.953 doi: 10.1613/jair.953
[34]
Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, A. Y. Ng, Reading digits in natural images with unsupervised feature learning, NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011.
[35]
D. P. Kingma, J. Ba, Adam: A method for stochastic optimization, arXiv: 14126980, 2014.
This article has been cited by:
1.
Giovana A. Benvenuto, Marilaine Colnago, Maurício A. Dias, Rogério G. Negri, Erivaldo A. Silva, Wallace Casaca,
A Fully Unsupervised Deep Learning Framework for Non-Rigid Fundus Image Registration,
2022,
9,
2306-5354,
369,
10.3390/bioengineering9080369
2.
Hansi Gunasinghe, James McKelvie, Abigail Koay, Michael Mayo,
2022,
Chapter 34,
978-3-031-21966-5,
421,
10.1007/978-3-031-21967-2_34
3.
Aidan Pucchio, Saffire Krance, Daiana R Pur, Arshpreet Bassi, Rafael Miranda, Tina Felfeli,
The role of artificial intelligence in analysis of biofluid markers for diagnosis and management of glaucoma: A systematic review,
2022,
1120-6721,
112067212211409,
10.1177/11206721221140948
Mehwish Saqlain, Saddaf Rubab, Malik M. Khan, Nouman Ali, Shahzeb Ali, Abdul Qadeer Khan,
Hybrid Approach for Shelf Monitoring and Planogram Compliance (Hyb-SMPC) in Retails Using Deep Learning and Computer Vision,
2022,
2022,
1563-5147,
1,
10.1155/2022/4916818
Aqsa Rasheed, Nouman Ali, Bushra Zafar, Amsa Shabbir, Muhammad Sajid, Muhammad Tariq Mahmood,
Handwritten Urdu Characters and Digits Recognition Using Transfer Learning and Augmentation With AlexNet,
2022,
10,
2169-3536,
102629,
10.1109/ACCESS.2022.3208959
8.
Iqra Ashraf Kiyani, Adeela Razaq,
A Comprehensive Review on Sentiment Perception Using Electroencephalography (EEG),
2022,
3,
2662-995X,
10.1007/s42979-022-01155-4
9.
Shervan Fekri-Ershad, Mustafa Jawad Al-Imari, Mohammed Hayder Hamad, Marwa Fadhil Alsaffar, Fuad Ghazi Hassan, Mazin Eidan Hadi, Karrar Salih Mahdi, Nouman Ali,
Cell Phenotype Classification Based on Joint of Texture Information and Multilayer Feature Extraction in DenseNet,
2022,
2022,
1687-5273,
1,
10.1155/2022/6895833
10.
Muhammad Zamir, Nouman Ali, Amad Naseem, Areeb Ahmed Frasteen, Bushra Zafar, Muhammad Assam, Mahmoud Othman, El-Awady Attia,
Face Detection & Recognition from Images & Videos Based on CNN & Raspberry Pi,
2022,
10,
2079-3197,
148,
10.3390/computation10090148
11.
Amsa Shabbir, Nouman Ali, Jameel Ahmed, Bushra Zafar, Aqsa Rasheed, Muhammad Sajid, Afzal Ahmed, Saadat Hanif Dar, Muazzam Maqsood,
Satellite and Scene Image Classification Based on Transfer Learning and Fine Tuning of ResNet50,
2021,
2021,
1563-5147,
1,
10.1155/2021/5843816
12.
Yudong Zhang, Juan Manuel Gorriz, Deepak Ranjan Nayak,
Optimization Algorithms and Machine Learning Techniques in Medical Image Analysis,
2023,
20,
1551-0018,
5917,
10.3934/mbe.2023255
13.
Ch. Madhu Babu, G. Prabaharan, R. Pitchai,
Efficient detection of glaucoma using double tier deep convolutional neural network,
2022,
1617-4909,
10.1007/s00779-022-01673-1
14.
Surbhi Bhatia, Shadab Alam, Mohammed Shuaib, Mohammed Hameed Alhameed, Fathe Jeribi, Razan Ibrahim Alsuwailem,
Retinal Vessel Extraction via Assisted Multi-Channel Feature Map and U-Net,
2022,
10,
2296-2565,
10.3389/fpubh.2022.858327
15.
Aqsa Tahir, Muhammad Asif, Maaz Bin Ahmad, Toqeer Mahmood, Muhammad Adnan Khan, Mushtaq Ali, Muhammad Sajid,
Brain Tumor Detection using Decision-Based Fusion Empowered with Fuzzy Logic,
2022,
2022,
1563-5147,
1,
10.1155/2022/2710285
16.
Jia Gu, Fangzheng Tian, Il-Seok Oh,
Retinal vessel segmentation based on self-distillation and implicit neural representation,
2022,
0924-669X,
10.1007/s10489-022-04252-2
Lina Elwiya, Alaa Hamid Mohammed, Zinah Khalid Jasim Jasim,
2021,
Retinal Fundus Images Of Optical Disk Detection,
978-1-6654-4930-4,
664,
10.1109/ISMSIT52890.2021.9604599
22.
Maryam Bukhari, Sadaf Yasmin, Saima Sammad, Ahmed A. Abd El-Latif, Nouman Ali,
A Deep Learning Framework for Leukemia Cancer Detection in Microscopic Blood Samples Using Squeeze and Excitation Learning,
2022,
2022,
1563-5147,
1,
10.1155/2022/2801227
23.
Najdavan A. Kako, Adnan M. Abdulazeez,
Peripapillary Atrophy Segmentation and Classification Methodologies for
Glaucoma Image Detection: A Review,
2022,
18,
15734056,
1140,
10.2174/1573405618666220308112732
24.
Divya Parthasarathy Rao, Sujani Shroff, Florian M. Savoy, Shruthi S, Chao-Kai Hsu, Kalpa Negiloni, Zia Sultan Pradhan, Jayasree P V, Anand Sivaraman, Harsha L. Rao,
Evaluation of an offline, artificial intelligence system for referable glaucoma screening using a smartphone-based fundus camera: a prospective study,
2024,
38,
0950-222X,
1104,
10.1038/s41433-023-02826-z
25.
R. Thamilselvan, P. Natesan, R.R. Rajalaxmi, S. Mohan, S. Muruganantham, R. Vanitha,
2023,
An Elaborate Review of Diagnosing Glaucoma Disease Using Machine Learning and Deep Learning Techniques,
979-8-3503-4821-7,
1,
10.1109/ICCCI56745.2023.10128598
26.
Vijaya Kumar Velpula, Lakhan Dev Sharma,
Multi-stage glaucoma classification using pre-trained convolutional neural networks and voting-based classifier fusion,
2023,
14,
1664-042X,
10.3389/fphys.2023.1175881
27.
Sandeep Angara, Jongwoo Kim,
2024,
Deep Ensemble Learning for Classification of Glaucoma from Smartphone Fundus Images,
979-8-3503-8472-7,
412,
10.1109/CBMS61543.2024.00074
28.
S. Nandhini, J. Jagadeesan, Anitha E,
2023,
U-Net Architecture for Detecting Glaucoma with Retinal Fundus Images,
979-8-3503-4279-6,
1,
10.1109/ICAEECI58247.2023.10370979
29.
Mohammad J. M. Zedan, Mohd Asyraf Zulkifley, Bashar Ayad Saeed, Siti Raihanah Abdani,
2024,
Deep Learning-based Model Benchmarking of Glaucoma Segmentation Using a Novel Ibn Al-Haitham Fundus Image Dataset,
979-8-3503-6410-1,
244,
10.1109/SIML61815.2024.10578202
Sivarathri Susrutha, Robin Prakash Mathur,
2023,
Review On Ocular Disease Recognition Using Deep Learning,
979-8-3503-9648-5,
316,
10.1109/InCACCT57535.2023.10141747
32.
Mohammad Zedan, Mohd Zulkifley, Ahmad Ibrahim, Asraf Moubark, Nor Kamari, Siti Abdani,
Automated Glaucoma Screening and Diagnosis Based on Retinal Fundus Images Using Deep Learning Approaches: A Comprehensive Review,
2023,
13,
2075-4418,
2180,
10.3390/diagnostics13132180
33.
Shiwen Zeng, Xianwei Zhang, Hongxing Zhang, Fan Meng, Kun Zhao,
2023,
Label Update Based Partial Label Learning for Remote Sensing Scene Classification,
979-8-3503-9905-9,
872,
10.1109/ICEICT57916.2023.10245602
34.
Shalli Rani,
2023,
Analysis of VGG16 based Support Vector Classifier for Automated Glaucoma Detection and Classification*,
978-1-6654-9099-3,
767,
10.1109/SPIN57001.2023.10116670
35.
Iqra Ashraf Kiyani, Tehmina Shehryar, Samina Khalid, Uzma Jamil, Adeel Muzaffar Syed,
Deep learning based Glaucoma Network Classification (GNC) using retinal images,
2024,
34,
0899-9457,
10.1002/ima.23003
36.
Tautvydas Kvietkauskas, Ernest Pavlov, Pavel Stefanovič, Birutė Pliuskuvienė,
The Efficiency of YOLOv5 Models in the Detection of Similar Construction Details,
2024,
14,
2076-3417,
3946,
10.3390/app14093946
37.
Laureen Wegert, Stefan Schramm, Alexander Dietzel, Dietmar Link, Sascha Klee,
Three-Dimensional Light Field Fundus Imaging: Automatic Determination of Diagnostically Relevant Optic Nerve Head Parameters,
2023,
12,
2164-2591,
21,
10.1167/tvst.12.7.21
38.
S Ananya, G Apoorva, T N Mohan Babu, K P Bharath,
2024,
Glaucoma Detection using HOG based features with Feed-forward Neural Networks and Extreme Learning Machine,
979-8-3503-7289-2,
1,
10.1109/NMITCON62075.2024.10698815
39.
Tasleem Kausar, Yun Lu, Adeeba Kausar,
Breast Cancer Diagnosis Using Lightweight Deep Convolution Neural Network Model,
2023,
11,
2169-3536,
124869,
10.1109/ACCESS.2023.3326478
40.
Siddhi Chourasia, Rhugved Bhojane, Rushikesh Patil, Deepali M. Kotambkar,
2023,
Domain Adaptation Using DCGAN For Glaucoma Diagnosis,
979-8-3503-3401-2,
1,
10.1109/I2CT57861.2023.10126413
41.
B. S. Sujithra, S. Albert Jerome,
Identification of glaucoma in fundus images utilizing gray wolf optimization with deep convolutional neural network-based resnet50 model,
2023,
83,
1573-7721,
49301,
10.1007/s11042-023-17506-z
42.
B. S. Sujithra, S. Albert Jerome,
Glaucoma detectıon and classıfıcatıon by usıng hysteresıs thresholdıng based IAOAWTO and ICNNBTL classıfıer,
2023,
83,
1573-7721,
42519,
10.1007/s11042-023-17148-1
43.
Sai Sudha Gadde, K. V. D. Kiran,
Entropy-Based Feature Extraction Model for Fundus Images with Deep Learning Model,
2024,
24,
0219-4678,
10.1142/S0219467823400065
Manpinder Singh, Saiba Dalmia, Ranjeet Kumar Ranjan,
Detection of diabetic retinopathy and age-related macular degeneration using DenseNet based neural networks,
2024,
1573-7721,
10.1007/s11042-024-18701-2
46.
Diya Wang, Qiao Wang, Qiang Su, Shun Wang, Zhijie Jian, Jianpu Li, Fang Ye, Yanli Hou, Mingxi Wan,
Multi-Parametric Retinal Microvascular Functional Perfusion Imaging Based on Dynamic Fundus Fluorescence Angiography,
2024,
71,
0018-9294,
3123,
10.1109/TBME.2024.3408636
47.
S. Ramkumar, Mohanthi K, Deepika P, Martin Joel Rathnam,
2024,
Automation of Cataloging an Abnormal Condition of Glaucoma Using Fundas Image of the Retina,
979-8-3503-6908-3,
1,
10.1109/ICEEICT61591.2024.10718487
48.
Arindam Chowdhury, Ankit Lodh, Rohit Agarwal, Rahul Garai, Ahana Nandi, Narayan Murmu, Sumit Banerjee, Debashis Nandi,
Rim learning framework based on TS-GAN: A new paradigm of automated glaucoma screening from fundus images,
2025,
187,
00104825,
109752,
10.1016/j.compbiomed.2025.109752
49.
Vincenza Ferrara, Anders Wästfelt,
An ancient olive tree in the garden. Mapping the deep history of land use from a single image,
2025,
40,
1010-6049,
10.1080/10106049.2025.2471090
Ocular inflammatory disorders (uveitis) can occur for various reasons and unfortunately, glaucoma can soften up to 20% of patients. Some of these disorders tend to affect only one eye. Other forms can affect both eyes
The backside of cornea contains cells that scatter over the eye drainage tissue and across the surface of the iris. This leads to an increased IOP in an eye that can damage the optic nerve. These cells form accession that glue the cornea and iris and block the drainage channel system. And iris and block the drainage channel system.
Normally babies suffer from this type of glaucoma. When a woman is pregnant than if her baby's eye drainage canal system development is not correctly building this leads to Congenital Glaucoma.
Pseudoexfoliative glaucoma occurs when ash-like material peels off from the external side of the lens in the eye. This ash-like material collects within the angle between the iris and cornea and damage the whole system of the eye.
This type of glaucoma always related to other abnormalities, most often diabetes. Blood vessels are blocked and the eye's fluid can leave the eye's drainage canals and this will cause an increase in IOP and this kind of glaucoma is very hard to treat.
This dataset is partitioned into training and test image sets respectively. The training set comprise of 50 images, with ground truth providing information regarding region boundary, segmentation soft map and Cup-to-Disk ratio for each image. Whereas the testing set contains 51 images and the ground truth for test set is available only upon registration.Image size and resolution is 2896×1944 pixels and PNG uncompressed image format is used. The retinal image type of the dataset is Fundus Images .
HRF is publically available database consisting of high-resolution fundus images. The dataset is composed of three sets of fundus images: comprising of the healthy retinas, glaucomatous and DR retinas. In the first set, 15 images of healthy patients are included which do not contain any instance of retinal pathology. The next set is comprised of 15 retinal images of patients with DR exhibiting pathological changes, such as bright lesions, neovascular nets, spots after laser treatment and hemorrhages. The last set includes 15 images of patients with advanced stage glaucoma and showing symptoms of focal and diffuse nerve fiber layer loss. The retail type of the dataset is Fundus images. The image quality of the database is 3304 x 2336 pixels.
There are 99 fundus images taken from 50 patients.The dataset type is Fundus images. Images were captured using Kowa VX-10 alpha digital fundus camera with 50 degree field of view (FOV). The images have resolution of 4288x2848 pixels and are stored in jpg file format. The size of each image is about 800 KB.
IDRiD is a recently introduced retinal fundus image database publicly available and consists of 516 images organized in two categoeies: 1.Retinal images that exhibit symptoms of DME and/or DR 2. Normal retinal images without any indications of DME and/or DR.
The RIGA dataset comprises of 750 images collected from three sources and 4500 manual marked images. The images are saved in JPG and TIFF format and contain both glaucomatous and normal images.The resolution of 3 sources are a. MESSIDOR 2240x1488 PIXELS (460 Images) B. Magrabi eye center Riyadh 2743x1936 (95 Images) c. Bin Rushed Ophthalmic center in Riyhad 2376x1584 (195 images) The retinal image type of the dataset is Fundus images
It comprises of 110 retinal images with image resolution of 600×400 pixels and the optic Disc annotated by two experts with 36 landmarks.The resolution of the dataset is 600x400. The retinal image type is Fundus images .
It consists of 258 images annotated manually alonwith their clinical diagnosis details. The images in ACHIKO-K are collected from patients suffering from glaucoma and contain rich information related to glaucoma pathological symptoms , e.g. optic nerve drusen, hemorrhage and optic cup notching, etc. Multiple images are taken from patients suffering from glaucoma having age between 27 to 85 years during 2008 to 2012 time span. A total of 258 fundus images are available in database, where 144 are from glaucoma tic eyes and 114 are captured from normal eyes.
The dataset consists of 2258 individuals with genome data, image data and clean screening data. Among the 2258 subjects, 100 were diagnosed to have glaucoma and the rest were normal.
Ocular inflammatory disorders (uveitis) can occur for various reasons and unfortunately, glaucoma can soften up to 20% of patients. Some of these disorders tend to affect only one eye. Other forms can affect both eyes
The backside of cornea contains cells that scatter over the eye drainage tissue and across the surface of the iris. This leads to an increased IOP in an eye that can damage the optic nerve. These cells form accession that glue the cornea and iris and block the drainage channel system. And iris and block the drainage channel system.
Normally babies suffer from this type of glaucoma. When a woman is pregnant than if her baby's eye drainage canal system development is not correctly building this leads to Congenital Glaucoma.
Pseudoexfoliative glaucoma occurs when ash-like material peels off from the external side of the lens in the eye. This ash-like material collects within the angle between the iris and cornea and damage the whole system of the eye.
This type of glaucoma always related to other abnormalities, most often diabetes. Blood vessels are blocked and the eye's fluid can leave the eye's drainage canals and this will cause an increase in IOP and this kind of glaucoma is very hard to treat.
This dataset is partitioned into training and test image sets respectively. The training set comprise of 50 images, with ground truth providing information regarding region boundary, segmentation soft map and Cup-to-Disk ratio for each image. Whereas the testing set contains 51 images and the ground truth for test set is available only upon registration.Image size and resolution is 2896×1944 pixels and PNG uncompressed image format is used. The retinal image type of the dataset is Fundus Images .
HRF is publically available database consisting of high-resolution fundus images. The dataset is composed of three sets of fundus images: comprising of the healthy retinas, glaucomatous and DR retinas. In the first set, 15 images of healthy patients are included which do not contain any instance of retinal pathology. The next set is comprised of 15 retinal images of patients with DR exhibiting pathological changes, such as bright lesions, neovascular nets, spots after laser treatment and hemorrhages. The last set includes 15 images of patients with advanced stage glaucoma and showing symptoms of focal and diffuse nerve fiber layer loss. The retail type of the dataset is Fundus images. The image quality of the database is 3304 x 2336 pixels.
There are 99 fundus images taken from 50 patients.The dataset type is Fundus images. Images were captured using Kowa VX-10 alpha digital fundus camera with 50 degree field of view (FOV). The images have resolution of 4288x2848 pixels and are stored in jpg file format. The size of each image is about 800 KB.
IDRiD is a recently introduced retinal fundus image database publicly available and consists of 516 images organized in two categoeies: 1.Retinal images that exhibit symptoms of DME and/or DR 2. Normal retinal images without any indications of DME and/or DR.
The RIGA dataset comprises of 750 images collected from three sources and 4500 manual marked images. The images are saved in JPG and TIFF format and contain both glaucomatous and normal images.The resolution of 3 sources are a. MESSIDOR 2240x1488 PIXELS (460 Images) B. Magrabi eye center Riyadh 2743x1936 (95 Images) c. Bin Rushed Ophthalmic center in Riyhad 2376x1584 (195 images) The retinal image type of the dataset is Fundus images
It comprises of 110 retinal images with image resolution of 600×400 pixels and the optic Disc annotated by two experts with 36 landmarks.The resolution of the dataset is 600x400. The retinal image type is Fundus images .
It consists of 258 images annotated manually alonwith their clinical diagnosis details. The images in ACHIKO-K are collected from patients suffering from glaucoma and contain rich information related to glaucoma pathological symptoms , e.g. optic nerve drusen, hemorrhage and optic cup notching, etc. Multiple images are taken from patients suffering from glaucoma having age between 27 to 85 years during 2008 to 2012 time span. A total of 258 fundus images are available in database, where 144 are from glaucoma tic eyes and 114 are captured from normal eyes.
The dataset consists of 2258 individuals with genome data, image data and clean screening data. Among the 2258 subjects, 100 were diagnosed to have glaucoma and the rest were normal.
Figure 3. Left: images in the SVHN dataset; Right: synthesized images generated by SEC-CGAN; From left to right, images on each column denote class 0 to 9
Figure 4. Sample images of road weather conditions
Figure 5. Three examples of synthesized images from the proposed SEC-CGAN. For each grid, the columns from left-to-right are: foggy, rainy, snowy, sunny