Case report Topical Sections

Changes in FGFR2 amino-acid residue Asn549 lead to Crouzon and Pfeiffer syndrome with hydrocephalus

  • Received: 01 August 2016 Accepted: 27 September 2016 Published: 09 October 2016
  • Mutations in Fibroblast Growth Factor Receptor II (FGFR2) have been identified in patients with Crouzon and Pfeiffer syndrome, among which rare mutations of the intracellular tyrosine kinase domain. Correlating subtle phenotypes with each rare mutation is still in progress. In Necker-Enfants Malades Hospital, we identified three patients harboring three different pathogenic variants of the same amino acid residue Asn-549 located in this domain: in addition to a very typical crouzonoid appearance, they all developed clinically relevant hydrocephalus, which is an inconstant feature of Crouzon and Pfeiffer syndrome. Overall, FGFR2 tyrosine kinase domain mutations account for 5/67 (7.4%) cases in our hospital. We describe a novel mutation, p.Asn549Ser, and new cases of p.Asn549His and p.Asn549Thr mutations, each reported once before. Our three cases of Asn-549 mutations, alongside with rare previously reported cases, show that these patients are at higher risk of hydrocephalus. Clinical and imaging follow-up, with possible early surgery, may help prevent secondary intellectual disability.

    Citation: Caroline Apra, Corinne Collet, Eric Arnaud, Federico Di Rocco. Changes in FGFR2 amino-acid residue Asn549 lead to Crouzon and Pfeiffer syndrome with hydrocephalus[J]. AIMS Genetics, 2016, 3(4): 205-211. doi: 10.3934/genet.2016.4.205

    Related Papers:

    [1] Mingzhou Xu, Xuhang Kong . Note on complete convergence and complete moment convergence for negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(4): 8504-8521. doi: 10.3934/math.2023428
    [2] He Dong, Xili Tan, Yong Zhang . Complete convergence and complete integration convergence for weighted sums of arrays of rowwise m-END under sub-linear expectations space. AIMS Mathematics, 2023, 8(3): 6705-6724. doi: 10.3934/math.2023340
    [3] Mingzhou Xu, Kun Cheng, Wangke Yu . Complete convergence for weighted sums of negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2022, 7(11): 19998-20019. doi: 10.3934/math.20221094
    [4] Lunyi Liu, Qunying Wu . Complete integral convergence for weighted sums of negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(9): 22319-22337. doi: 10.3934/math.20231138
    [5] Mingzhou Xu . On the complete moment convergence of moving average processes generated by negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2024, 9(2): 3369-3385. doi: 10.3934/math.2024165
    [6] Shuyan Li, Qunying Wu . Complete integration convergence for arrays of rowwise extended negatively dependent random variables under the sub-linear expectations. AIMS Mathematics, 2021, 6(11): 12166-12181. doi: 10.3934/math.2021706
    [7] Mingzhou Xu . Complete convergence of moving average processes produced by negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(7): 17067-17080. doi: 10.3934/math.2023871
    [8] Mingzhou Xu . Complete convergence and complete moment convergence for maximal weighted sums of extended negatively dependent random variables under sub-linear expectations. AIMS Mathematics, 2023, 8(8): 19442-19460. doi: 10.3934/math.2023992
    [9] Qingfeng Wu, Xili Tan, Shuang Guo, Peiyu Sun . Strong law of large numbers for weighted sums of m-widely acceptable random variables under sub-linear expectation space. AIMS Mathematics, 2024, 9(11): 29773-29805. doi: 10.3934/math.20241442
    [10] Baozhen Wang, Qunying Wu . Almost sure convergence for a class of dependent random variables under sub-linear expectations. AIMS Mathematics, 2024, 9(7): 17259-17275. doi: 10.3934/math.2024838
  • Mutations in Fibroblast Growth Factor Receptor II (FGFR2) have been identified in patients with Crouzon and Pfeiffer syndrome, among which rare mutations of the intracellular tyrosine kinase domain. Correlating subtle phenotypes with each rare mutation is still in progress. In Necker-Enfants Malades Hospital, we identified three patients harboring three different pathogenic variants of the same amino acid residue Asn-549 located in this domain: in addition to a very typical crouzonoid appearance, they all developed clinically relevant hydrocephalus, which is an inconstant feature of Crouzon and Pfeiffer syndrome. Overall, FGFR2 tyrosine kinase domain mutations account for 5/67 (7.4%) cases in our hospital. We describe a novel mutation, p.Asn549Ser, and new cases of p.Asn549His and p.Asn549Thr mutations, each reported once before. Our three cases of Asn-549 mutations, alongside with rare previously reported cases, show that these patients are at higher risk of hydrocephalus. Clinical and imaging follow-up, with possible early surgery, may help prevent secondary intellectual disability.


    Common macular and vascular diseases include age-related macular degeneration (ARMD), diabetic macular edema (DME), branch retinal vein occlusion (BRVO), central retinal vein occlusion (CRVO), and central serous chorioretinopathy (CSCR), which are the leading causes of visual impairment and blindness worldwide [1,2,3]. According to the World Health Organization (WHO), DME, which primarily affects working-age adults, affected 425 million people worldwide in 2017 and is expected to affect 629 million people by 2045 [4]. The WHO also estimates that 196 million people had ARMD in 2020; this number is expected to rise to 288 million by 2040 [5]. The prevalence of ARMD in elderly people is 40% at the age of 70 years, rising to 70% at the age of 80 years. Rogers et al. [6] discovered that BRVO and CRVO affected 13.9 million and 2.5 million of the world's population aged 30 years and older, respectively, in 2008. Men have a higher prevalence of CSCR compared to women [7]. A large population is afflicted by these diseases, and projections suggest that this number will escalate in the future. However, the first stage of these diseases can be treated, and patients can recover their vision loss through early detection and treatment [8,9,10].

    Optical coherence tomography (OCT) is a noninvasive imaging modality that provides high-resolution information within a cross sectional area. OCT retinal imaging enables visualization of the thickness, structure, and detail of various layers of the retina. In addition, when the retina develops a disease, OCT enables the visualization of abnormal features and damaged retinal structures [11]. Therefore, retinal OCT images are widely used in the medical field to monitor information in medical images prior to treatment or for the diagnosis of various diseases.

    For several years, ophthalmologists have analyzed the comprehensive information inside the retina for retinal care services, treatment, and diagnosis using retinal OCT images in clinical settings. The clinician performs these tasks manually and wait for each process. As a result, manual analysis is time consuming when there are numerous OCT images. Even if the clinician has great expertise, this analysis may not be accurate [12]. An automated technique based on deep learning (DL) or machine learning using artificial intelligence has been proposed as a solution to overcome this limitation.

    Recently, computer algorithms based on artificial intelligence, DL, and machine learning have been proposed for the automatic diagnosis of various retinal diseases and have been applied in clinical health care. Han et al. [13] modified three well known convolutional neural network (CNN) models to gain access to normal and three subtypes of neovascular age-related macular degeneration (nAMD). The classification layers of the original CNN models were replaced by new layers: four fully connected layers and three dropout layers, along with a Leaky rectified linear activation unit (ReLU) as an activation function. The modified models were trained using the transfer learning technique and tested on 920 OCT images; the VGG-16 model achieved an accuracy of 87.4%. Sotoudeh-Paima et al. [14] classified OCT images to identify normal, AMD, and choroidal neovascularization (CNV) using a multiscale CNN. This CNN was evaluated and achieved a classification accuracy of 93.40% on the public dataset. Elaziz et al. [15] developed a four-class classification method for accessing retinal diseases from OCT images based on an ensemble DL model and machine learning. First, the features are extracted from the two models, MobileNet and DenseNet, and were concatenated as full features of the input images. Then, feature selection was performed to remove irrelevant features and to input the useful features into machine learning to classify the input data. A total of 968 OCT images were used to evaluate classification performance, and an accuracy of 94.31% was achieved. Another study by Liu et al. [16] used a DL model to extract attention features from OCT images. It used the extracted features as guiding features for CNV, DME, drusen, and normal. The classification performance was assessed using public datasets, and an average accuracy of 95.10% was achieved. Minagi et al. [17] used transfer learning with universal adversarial perturbations (UAPs) for classification with a limited dataset. Three types of medical images, including OCT images, were used to assess diseases, and DL models were trained using the ImageNet dataset. The UAPs algorithm was used to generate a training set based on the data provided to train the DL model. There were 11,200 OCT images utilized in training and assessing the model's performance, and a classification accuracy of 95.3% was achieved for the four classes: CNV, DME, drusen, and normal. Tayal et al. [18] presented four ocular disease classifications based on three CNN models using OCT images. Images were enhanced before being fed to CNN models. To assess the performance of the presented method, 6,678 publicly available OCT images were evaluated. The method achieved an accuracy of 96.50% with a CNN model which compressed nine layers. The performance of the CNN models with nine layers outperformed the experimented CNN models with five and seven layers. Adversarial retraining is an algorithm used to improve the performance of DL models based on classification.

    According to the literature, retinal OCT classification was developed using DL and DL based methods such as transfer learning, smoothing generative adversarial networks, adversarial retraining, and multi-scale CNN. This method is used to improve the model's performance by fine-tuning previous task knowledge using the OCT image problem, increasing the dataset size for training, applying the technique of inputting data for the training model, and changing the training input image sizes. However, the classification methods can achieve an accuracy of less than 97.00%, indicating their potential for further improvement. Moreover, these studies classify retinal diseases into fewer than five classes. This study aims to improve the classification accuracy and detect five classes of retinal diseases, which are more than the previous studies highlighted in the literature.

    In this study, we propose an automatic method based on a hybrid of deep learning and ensemble machine learning for screening five different retinal diseases from OCT images to improve the performance of OCT image classification. The proposed method improves classification accuracy, outperforming standalone classifiers without a hybrid. In addition, it can be trained using a smaller dataset from our hospital, which has been strictly labelled by experts. Moreover, the proposed method enables deployment with a web server for open access to test the evaluation performance within seconds.

    All OCT images were collected from Soonchunhyang University's Bucheon Hospital. The OCT images were collected and normalized after approval by the Bucheon Hospital's Institutional Review Board (IRB). OCT images were captured using DRI-OCT (Topcon Medical System, Inc., Oakland, NJ, USA). The scan range was 3–12 mm in the horizontal and vertical directions, with a lateral resolution of 20 μm and an in-depth resolution of 8 μm. The shooting speed was 100,000 A-scans per second. The OCT images utilized were collected twice; the first comprised 2,000 images that were captured between April and September 2021, while the second consisted of 998 images, and took place over a period of approximately five months from September 2021 to January 2022. Therefore, the total number of OCT images collected twice was 2,998; these were labeled by ophthalmologists for five retinal diseases (ARMD:740, BRVO:450, CRVO:299, CSCR:749, and DME:760) as the ground truth.

    This study was approved by the Institutional Review Board (IRB) from Soonchunhyang University Bucheon Hospital, Bucheon, Republic of Korea (IRB approval number: 2021-05-001). All methods were performed in accordance with relevant guidelines and regulations. Informed consent was obtained from all subjects.

    Image processing is a technique for performing various operations on the original images to convert it into a format suitable for DL models or to extract useful features. In image classification based on deep learning, image processing is an essential initial process to change an image before feeding it to the CNN model. The CNN model requires a unique size for the image input, and higher-resolution images demand longer computing times. To shorten the operating time and the suitable size required by the CNN models, all OCT images were downsized to 300 pixels in height and 500 pixels in width. The OCT image dataset was split into an 80% training set and 20% testing set. The training set was used to train the deep learning model and the testing set was used to assess performance.

    The size of the dataset has a significant impact on the DL performance. Therefore, a larger dataset may enable a better performance. However, in the medical field, most medical dataset has size limits. Data augmentation is a technique developed to overcome the limitations of a dataset by performing different operations on the data provided and creating new data, thereby enhancing the dataset size. Additionally, data augmentation is used to enhance performance [19], generalize the model [20], and avoid overfitting [21]. We utilize data augmentation techniques from the Python library imgaug including like vertical flip, rotation, scale, brightness, saturation, contrast, enhance and contrast, and equalization. The OCT images were augmented at angles of 170,175,185, and 190. The selected angle is suitable for rectangle shape representation without loss of information from the original OCT images; scale image with a random range between 0.01 to 0.12; the level of brightness from 1 to 3; the saturate operation, which ranges from 1 to 5, increases by one with each level; random contrast with contrast values ranging from 0.2 to 3; enhance and contrast with levels ranging from 1 to 1.5; and image equalization with levels ranging from 0.9 to 1.4. At the end of the data augmentation process, one OCT image can serve as the basic for generating 29 augmented images. Therefore, the training set comprised a total of 69,455 OCT images, including samples. The acquired OCT and augmented images are shown in Figure 1. Applying data augmentation, only the training set is used for training the proposed method. After finishing the augmentation operation, the OCT images are passed through the 10-fold cross-validation technique to partition the data into folds for the training model (training data) and to test the model after finishing every epoch (validation data).

    Figure 1.  OCT images before and after performing data augmentation. (a) represents the original OCT image. (b), (c), and (d) illustrate brightness adjustments. (e) and (f) demonstrate contrast modifications. (g), (h), and (i) display contrast enhancement. (j) and (k) depict equalization. (l) represents a vertical flip. (m), (n), (o), (p), (q), and (r) indicate angle rotations. (s), (t), (u), (v), and (w) illustrate saturation changes. (x), (y), (z), (A), (B), and (C) represent scaling variations.

    Figure 2 shows the architecture of the proposed method that comprises three significant blocks: feature extraction, classification, and boosting performance. First, transfer learning based on CNN models extracts one thousand features from the OCT images. Second, various machine learning algorithms are used to classify the OCT images based on the features extracted by the CNN model. Finally, the ensemble algorithm fuses the distribution probabilities of the same class and predicts the retinal disease class based on probability. Each block of the proposed architecture is described in detail in the following subsections.

    Figure 2.  System architecture overview of the proposed method. The proposed method accepts images with resolution of 500 pixels in width and 300 pixels in height. CNN models extract features from OCT images and classify them using machine learning algorithms. Voting classifier ensemble output probabilities for predicting retinal disease.

    Transfer learning is a technique used to transform the knowledge of a related task that has already been studied to improve the learning of a new task. Training a CNN model from scratch is computationally expensive and time consuming; moreover, an extensive dataset is required to achieve a better performance. Therefore, transfer learning has been developed to overcome DL's drawbacks [22]. To retrain the model with new tasks based on prior knowledge, pretrain was refined, small top layers were trained, and the final layers were frozen. In this study, the transfer learning CNN (TL-CNN) models EfficientNetB0 [23], InceptionResNetV2 [24], InceptionV3 [25], ResNet50 [26], VGG16 [27], and VGG19 [28] are selected and updated. The modification names of the CNN models start with TL, indicating transfer learning, and ends with the original names of the CNN models, including TL-EfficientNetB0, TL -InceptionResNetV2, TL-InceptionV3, TL-ResNet50, TL-VGG16, and TL-VGG19. The original CNN models were created for generic image classification tasks. They were trained and tested on a large dataset (ImageNet) to categorize 1000 different types of images. To use a CNN model with the transfer learning technique and classify retinal OCT images, each CNN model must modify its classification layers to adapt to the target classes. One specific problem is the categorization of OCT images. The new classification layer is modified with continued stacking of GlobalAveragePooling2D, one Normalization layer, and two Dense layers. The first Dense layer consists of 1,024 with the ReLU activation function and the final dense layer with a five output- vectors. Finally, the updated model is pretrained, and the pretrain model is retrained to fine-tune the previous feature representation in the base model to make it more relevant for OCT image classification. The output consists of five vectors representing the distribution class probabilities using the Softmax activation function. As mentioned previously, a CNN model based on transfer learning is used to extract convolutional features from the OCT images. Therefore, the convolutional features were extracted from the TL-CNNs models where the GlobalAveragePooling2D layers of the classification layer. These features are one-dimensional. Different models provide various features and numbers based on the structure and convolution filters of the model.

    Six TL-CNN models independently extracted the features. At the GlobalAverage-Pooling2D layers, the features were extracted (TL-EfficientNetB0: 1,280 features, TL-InceptionResnetV2: 1,536 features, TL-InceptionV3: 2,048 features, TL-ResNet50: 2,048 features, TL-VGG16: 512 features, TL-VGG19: 512 features). Then, the extracted features of each TL-CNN model were used as the input to five popular machine learning classifiers: support vector machine (SVM) [29], k-nearest neighbors (k-NN) [30], decision tree (DT) [31], Random Forests (RF) [32], Naïve Bayes [33], and XGBoost [34]. Various machine learning classifiers use different techniques for learning and distinguishing the different classes of data.

    Individual machine learning classifiers provide different identification accuracies. This is because each classifier has its own learning ability to identify classes based on the given features. Therefore, an ensemble method is used to aggregate the distribution probabilities of the two classifiers. The proposed method selects two higher prediction classifiers (k-NN and XGBoost) based on an experiment to perform aggregation. An ensemble is a type of soft voting that performs better than other models [35]. Soft voting predicts the final class label as the class label most frequently predicted by classifiers. In soft voting, class labels are predicted by averaging the probability p of the class. Table 1 presents the proposed algorithm, which includes image processing, splitting data, data augmentation, feature extraction, classification, and an ensemble of classifiers:

    yFC=argmaximk=1wkpik (1)
    Table 1.  Algorithm of the proposed method.
    Algorithm 1: Proposed OCT images Classification
    1: procedure OCT IMAGES PROCESSING
    2:       return preprocessed-images

    3: procedure SPLIT-DATA (OCT-data)
    4: train-data, test-data, train-labels, test-labels = split (OCT-images, labels)

    5: procedure DATA AUGMENTATION (train-data)
    6: augmented images = augmentation (vertical flip, rotation, scale, brightness, saturation, contrast, enhance and
    7: contrast, and equalization)
    8:       return augmented-images

    9: procedure 10-FOLD_CROSS_VALIDATION (augmented images, labels)
    10: Fold1, Fold2, ……Fold10 = train_test_split(augmented images, labels)
    11:       return Fold1-10

    12: procedure FEATURE_EXTRACTION (Fold1-10, test-data, test-labels)
    13: TL-CNN models = modify the convolutional neural network (CNN) models
    14: pre-train the TL-CNN models, small top layers are trained, and the final layers are frozen.
    15: extracted features = TL-CNN model at GlobalAveragePooling2D layers
    16: return extracted features saved in csv format
    17: procedure CLASSIFICATION (extracted features, labels)
    18: classifiers = [svm, k-NN, DT, RF, Naïve-Bayes, and XGBoost]
    19:   for clsf in range (0, 6):
    20:       predicted-labels = classifiers[clsf]. fit (extracted-features)
    21:       training-accuracy = accuracy (predicted-labels, labels)
    22:       save_train_weight
    23: voting = "soft"
    24: ML1 = k-NN (train-data, train-labels, test-data)
    25: ML2 = XGBoost (train-data, train-labels, test-data)
    26: procedure ENSMEBLE_CLASSIFIERS (train-data, train-labels, test-data)
    27: ensemble-classifiers = concadenate(ML1, ML2)
    28: ensemble-classifers.fit (train-data, train-labels)
    29: predictions = ensemble-classifers.predict(test-data)
    30: save_training_weights, results_visualization

     | Show Table
    DownLoad: CSV

    where w_k is the weight of the machine learning classifiers, which can be either k-NN or XGBoost; it automatically learns from disease features in OCT images and then identifies the type of disease based on the input data; i represents the class label of the retinal diseases, where i ∈{0: ARMD, 1: BRVO, 2: CRVO, 3: CSCR, 4: DME}; and p_ik represents the probability of machine-learning weight k for class i.

    The proposed OCT image classification method was developed using Python 3.7 and TensorFlow 2.6.0. In addition, Scikit Learn was operated on a personal computer running the Windows 10 operating system powered by an Intel(R) Xeon (R) Silver 4114 @ 2.20GHz CPU, 192GB RAM, and an NVIDIA TITAN RTX 119GB GPU.

    The proposed OCT image classification method was trained using augmented OCT images and evaluated using a test set. There were two types of training. First, six TL-CNN models were trained to perform feature extraction from OCT images.

    Six TL-CNN models were separately trained using a combination of the training set and the augmented images of the training set. The combination data were split using a 10-fold cross-validation algorithm to separate the images for training, validate the model during training, and prevent overfitting. Furthermore, the TL-CNN models were individually trained with a fixed batch size of 64, epochs of size 100, and an Adam optimizer with a learning rate of 0.0001. The learning rate was selected based on the standard learning rate provided by the TensorFlow library. For example, with a setting of 100 epochs, each model must be trained 100 times on the same data. Therefore, the performance is improved by updating the weight based on the information lost through repetitions of a training session. The weights of each TL-CNN model were recorded in a separate file after training and were utilized to extract features from the training and testing sets. Then, the machine learning models were trained with the convolution features extracted by the TL-CNN models to access the class probabilities. Six machine learning models were separately trained, and the weights were recorded after the training completed. Finally, an ensemble method based on soft voting was applied to the average class probabilities of the two classifiers to obtain an effective final class prediction.

    The results of the proposed OCT image classification are divided into three parts: classification results, deployment of the classification results to web services, and a comparison of the results with similar studies in terms of classification accuracy.

    A test set was used to evaluate the performance of the proposed method after training the model. The same preprocessing was performed on both the test dataset and the training dataset without data augmentation. The test set contained 601 OCT images, which were used to assess the classification performance. Six TL-CNN models were individually trained to extract features from the OCT images and store the extracted features in pickle format. Six machine learning classifiers were utilized to discriminate the classes of OCT images based on the features extracted by the TL-CNN. Statistical theories were analyzed to measure the classification ability among the classes, sensitivity, specificity, precision, and accuracy. The relationship between the sensitivity and specificity of various categories was shown through a receiver operating characteristic (ROC) curve. Moreover, the confusion matrix was analyzed, which indicated the correct and incorrect class predictions. Table 2 lists the test results of using TL-EfficientNetB0 as an extractor and seven types of classifiers, including an ensemble classifier, the classification result outperformed the ensemble classifier with a sensitivity, specificity, precision, and accuracy of 96.17, 98.92, 95.89 and 95.85%, respectively. The second highest performance was achieved with the k-NN classifier, which achieved a sensitivity, specificity, precision, and accuracy of 87.37, 96.95, 88.82 and 88.89%, respectively. The classification results for the other machine learning classifiers are unstable, both increasing and decreasing randomly.

    Table 2.  Shown are OCT images before and after performing data augmentation. (a) represents the original OCT image. (b), (c), and (d) illustrate brightness adjustments. (e) and (f) demonstrate contrast modifications. (g), (h), and (i) display contrast enhancement. (j) and (k) depict equalization. (l) represents a vertical flip. (m), (n), (o), (p), (q), and (r) indicate angle rotations. (s), (t), (u), (v), and (w) illustrate saturation changes. (x), (y), (z), (A), (B), and (C) represent scaling variations.
    TL-CNN model Machine Learning Sensitivity Specificity Precision Accuracy
    TL-EfficientNetB0 SVM 86.79% 96.78% 88.64% 88.39%
    k-NN 87.37% 96.95% 88.82% 88.89%
    DT 85.80% 96.41% 84.95% 86.90%
    RF 66.20% 92.60% 81.08% 75.95%
    Naive Bayes 86.11% 96.40% 86.58% 86.90%
    XGBoost 85.86% 96.45% 86.38% 87.23%
    Ensemble 96.17% 98.92% 95.89% 95.85%

     | Show Table
    DownLoad: CSV

    Table 3 shows the classification results when using TL-InceptionResnetV2 as an extractor and seven classifiers, showing that the result outperforms the ensemble classifier with a sensitivity, specificity, precision, and accuracy of 97.42, 99.40, 97.49 and 97.68%, respectively. The second highest performance was achieved with the k-NN classifier, with a sensitivity, specificity, precision, and accuracy of 87.37, 96.48, 88.19 and 87.56%, respectively. In addition, with the same extractor, the classification performance of XGBoost was similar to that of the k-NN classifier. Table 4 lists the evaluation results when using the TL-InceptionV3 extractor and seven machine learning classifiers, including the ensemble classifier, which outperformed other methods with a sensitivity, specificity, precision, and accuracy of 91.34, 97.59, 91.03 and 91.04%, respectively. The second highest performance was achieved by XGBoost, with a sensitivity, specificity, precision, and accuracy of 84.42, 95.10, 82.88, and 82.91%, respectively. Table 5 lists the classification results when using the TL-ResNet50 model as a feature extractor and classifying those features by seven different classifiers, which indicates that using ensemble classifiers outperforms the obtained a sensitivity, specificity, precision, and accuracy of 96.46, 99.14, 96.76 and 96.68%, respectively. The second highest performance was achieved by XGBoost, with a sensitivity, specificity, precision, and accuracy of 87.63, 96.59%, 88.27 and 87.73%, respectively. The performances of the other two classifiers, SVM and k-NN, were comparable and better than those of the three classifiers in the experiments. Table 6 lists the test results of the proposed classification with VGG-16 as a feature extractor and seven machine learning classifiers, the ensemble classifier exhibited the best performance, with a sensitivity, specificity, precision, and classification accuracy of 92.07, 98.00, 92.60 and 92.54%, respectively. The XGBoost classifier had the second highest performance for TL-VGG16 as a feature extractor; it obtained a sensitivity, specificity, precision, and accuracy of 80.48, 94.91, 81.44 and 82.26%, respectively. A similar performance was observed for SVM and k-NN. Table 7 lists the classification test results of the TL-VGG19 model for feature extraction and classification using these features by various classifiers. Ensemble classifiers algorithm outperformed the five other classifiers; its sensitivity, specificity, precision, and accuracy are 93.86, 93.40, 93.44 and 93.86%, respectively. The second- and third-highest performances were achieved by XGBoost and SVM, respectively.

    Table 3.  Performance summary of proposed classification through feature extraction using TL-InceptionResnetV2, six classifiers, and ensemble voting classifiers. Various sensitivities, specificities, precisions, and accuracies are obtained using different classifiers. The proposed classification method with ensemble classifiers outperforms all statistic measurements.
    TL-CNN model Machine Learning Sensitivity Specificity Precision Accuracy
    TL-InceptionResnetV2 SVM 86.27% 96.13% 86.86% 86.40%
    k-NN 87.37% 96.48% 88.19% 87.56%
    DT 83.77% 95.54% 83.37% 84.41%
    RF 72.93% 93.79% 80.67% 79.27%
    Naive Bayes 79.66% 93.41% 78.25% 77.78%
    XGBoost 87.29% 96.47% 88.05% 87.56%
    Ensemble 97.42% 99.40% 97.49% 97.68%

     | Show Table
    DownLoad: CSV
    Table 4.  Performance summary of proposed classification through feature extraction using TL-InceptionV3, six classifiers, and ensemble voting classifiers. Various sensitivities, specificities, precisions, and accuracies are obtained when using different classifiers. The proposed classification method with ensemble classifiers outperforms all statistic measurements.
    TL-CNN models Machine Learning Sensitivity Specificity Precision Accuracy
    TL-InceptionV3 SVM 82.42% 94.50% 80.85% 81.09%
    k-NN 83.05% 94.61% 80.94% 81.43%
    DT 81.54% 94.18% 79.65% 80.09%
    RF 79.50% 94.01% 80.52% 79.77%
    Naive Bayes 65.72% 86.52% 2.58% 61.33%
    XGBoost 84.42% 95.10% 82.88% 82.91%
    Ensemble 91.34% 97.59% 91.03% 91.04%

     | Show Table
    DownLoad: CSV
    Table 5.  Performance summary of proposed classification through features extraction using TL-ResNet50, six classifiers, and ensemble voting classifiers. Various sensitivities, specificities, precisions, and accuracies are obtained using different classifiers. The proposed classification method with ensemble classifiers outperforms all statistic measurements.
    TL-CNN model Machine Learning Sensitivity Specificity Precision Accuracy
    TL-ResNet50 SVM 86.25% 96.02% 85.12% 85.95%
    k-NN 85.75% 96.00% 86.26% 85.74%
    DT 82.04% 94.94% 82.34% 82.59%
    RF 52.90% 88063% 71.77% 65.67%
    Naive Bayes 67.71% 89.82% 72.49% 64.68%
    XGBoost 87.63% 96.59% 88.27% 87.73%
    Ensemble 96.46% 99.14% 96.76% 96.68%

     | Show Table
    DownLoad: CSV
    Table 6.  Performance summary of proposed classification through features extraction using TL-VGG16, six classifiers, and ensemble voting classifiers. Various sensitivities, specificities, precisions, and accuracies are obtained using different classifiers. The proposed classification method with ensemble classifiers outperforms all statistic measurements.
    TL-CNN model Machine Learning Sensitivity Specificity Precision Accuracy
    TL-VGG16 SVM 76.39% 93.49% 76.96% 78.28%
    k-NN 74.55% 92.33% 75.70% 74.96%
    DT 57.42% 84.86% 55.72% 58.37%
    RF 50.55% 86.58% 38.39% 63.18%
    Naive Bayes 59.53% 85.08% 58.84% 59.20%
    XGBoost 80.48% 94.91% 81.44% 82.26%
    Ensemble 92.07% 98.00% 92.60% 92.54%

     | Show Table
    DownLoad: CSV
    Table 7.  Performance summary of proposed classification through features extraction using TL-VGG19, six classifiers, and ensemble voting classifiers. Various sensitivities, specificities, precisions, and accuracies are obtained using different classifiers. The proposed classification method with ensemble classifiers outperforms all statistic measurements.
    TL-CNN model Machine Learning Sensitivity Specificity Precision Accuracy
    TL-VGG19 SVM 79.90% 93.74% 78.77% 78.82%
    k-NN 69.16% 90.99% 70.56% 71.64%
    DT 53.23% 82.94% 53.41% 54.73%
    RF 48.29% 85.62% 37.71% 60.36%
    Naive Bayes 56.41% 82.96% 54.70% 54.89%
    XGBoost 82.44% 95.30% 81.90% 83.58%
    Ensemble 93.86% 93.40% 93.44% 93.86%

     | Show Table
    DownLoad: CSV

    Six TL-CNN models were compared, and TL-InceptionResNetV2 achieved a better performance than the other five models used in this study, with a sensitivity, specificity, precision, and accuracy of 97.42, 99.40, 97.49 and 97.68 %, respectively. The ensemble algorithm always outperformed all the TL-CNN models. The individual k-NN and XGBoost classifiers performed better than the three individual classifiers. Thus, ensembled k-NN and XGBoost also achieved better performance than k-NN and XGBoost.

    Figure 3 shows the ROC result of the proposed classification method, which outperforms TL- InceptionResnetV2 with ensemble classifiers (k-NN and XGBoost). The ROC among each class ARMD, BRVO, CRVO, CSCR, and DME is 0.99, 0.96, 0.99, 0.99, and 0.98, respectively. The relationship between sensitivity and specificity of the five classes is most important. The confusion matrix is implemented by using the Sklearn library in Python. The size of test data is essential to present the robustness of classification. The confusion matrix shows the number of correct and incorrect predictions among all classes. Figure 4 shows the confusion matrix of the proposed method which exhibited best performance; 148 of 149 OCT images of ARMD class are correctly predicted, 85 of 91 images of BRVO class are correctly predicted (ARMD:3 and DME:3 are incorrect predictions), 59 of 60 images are correctly predicted as CRVO and one image that is incorrectly predicted as BRVO, 148 of 150 images are correctly predicted and two images are incorrectly predicted as ARMD, and 149 of 153 are correctly predicted, and four are incorrectly predicted (ARMD: 1, BRVO:1, and CRVO:2 are incorrect prediction).

    Figure 3.  ROC curve of the proposed classification method, which exhibits best accuracy on TL-InceptionResnetV2 model and ensemble classifiers.
    Figure 4.  Confusion matrix of the proposed method when it exhibits best performance on TL-InceptionResNetV2 and ensemble classifiers.

    To render the proposed method applicable and accessible from outside through an Internet connection, we deployed the proposed OCT image classification to a web server using the Flask framework. The web server receives one image input at a time and inputs it into the proposed classification method to predict retinal diseases. The input image is an OCT image consisting of three channels with a resolution of 300 pixels in height and 500 pixels in width. When inputting an OCT image through a web service user interface (UI), the image is transferred to a computer server that runs a DL classification model. First, the computer server performs image processing which is the same to the processes used in both the train and test sets. Second, the preprocessed image is inputted into the proposed classification weights for prediction. Finally, the predicted results are forwarded to the web service using the Flask framework. The prediction results consist of the image input, distribution probabilities among the five classes, the retinal disease diagnosis class, and prediction times of an image. The prediction time is the time taken to input an image to a web service to predict and return the prediction result. Figure 5 shows the initial UI of the web server. The prediction results obtained after inputting the OCT images are shown in Figure 6.

    Figure 5.  Initial user interface of the developed web service for OCT image classification. The "Select an Image" button allows the user to browse to the location of a stored image and upload it to the webservice, and the "Predict" button sends the image to a deep learning server and receives the diagnosis class.
    Figure 6.  Prediction results from the development web service for OCT image classification. The predicted OCT image, distribution probabilities among five classes of retinal diseases in percent, a final predicted class based on higher probability, and time prediction are represented.

    The higher accuracy of the proposed OCT image classification method is compared with that of the recent studies reviewed in the literature review section, as listed in Table 8. These studies focused on transfer learning, developing new models, and combining well known CNN models with machine learning. All the listed studies used either different OCT databases or a combination of these datasets. Moreover, the number and type of classification classes were different, with at most four classes. We classify retinal diseases into five classes using a dataset obtained from a hospital. An additional number of classes can affect the performance of the classification methods. Table 8 lists the methods and algorithms that have been presented, including the suggested model with transfer learning, the multiscale DL model, and transfer learning using existing CNN models. However, the results as listed in the literature review have shown an accuracy of < 97%. Instead of focusing on a single classifier, this study combines two machine-learning classifiers and the DL as a feature extractor. Our study exhibits an accuracy of 97.68%, which is greater than the accuracy of the aforementioned studies. In addition, the number of classification classes is higher than that of the studies reviewed.

    Table 8.  Results comparison.
    Author Year Method Disease type Dataset size Accuracy
    Han et al. [13] 2022 Transfer learning with a modification of the well-known CNN models 4-class: PCV, RAP, nAMD, and NORMAL 4749 87.4%
    Sotoudeh-Paima et al. [14] 2022 Deep learning: multi-scale convolutional neural network 3-class: AMD, CNV, NORMAL 120,961 93.4%
    Elaziz et al. [15] 2022 Ensemble deep learning model for feature extraction, features selection, machine learning as classifier. 4-class: DME, CNV, DRUSEN, and NORMAL 84,484 94.32%
    Liu et al. [16] 2022 Deep learning based on method and lesions segmentation model. 4-class: CNV, DME, DRUSEN, and NORMAL 86,134 95.10%
    Minagi et al. [17] 2022 Transfer learning with DNN models 4-class: CNV, DME, DRUSEN, and NORMAL 11,200 95.3%
    Tayal et al. [18] 2022 Deep learning-based method 4-class: DME, CNV, DRUSEN, NORMAL 84,484 96.5%
    Proposed method Hybrid of deep learning and machine learning + ensemble machine learning classifiers. 5-class: ARMD, BRVO, CRVO, CSCR, DME 2,998 97.68%

     | Show Table
    DownLoad: CSV

    Our study classifies retinal OCT images with disease classes that differ from the reviewed studies and are not available in the public dataset. We hope that these retinal diseases will become available in the future, and we will evaluate the proposed OCT image classification system using a public dataset.

    This study presents a hybrid ensemble OCT image classification method for the diagnosis of five classes of retinal diseases. The proposed method employs an ensemble machine learning classifier as the classifier and a hybrid deep learning model as the feature extractor. We identified the deep learning model and ensemble classifiers that were most suitable for OCT image classification. The proposed model outperformed an individual classifier. With an accuracy of 97.68%, the best deep learning model and ensemble machine learning classifiers of the proposed method were TL- InceptionResnetV2 and the aggregation of KNN and XGBoost. This classification can be deployed to web services for convenient access to diagnose retinal disease from outside the Internet. Moreover, the prediction time in seconds was short, reducing the time required for prediction. This study contributes to the development of accurate multiclass OCT image classification. In the future, we aim to improve the classification performance. If datasets with the same class as in our study are made public, we will assess the proposed method on these datasets to broaden their applicability. In the medical field, improved performance can be used to automatically classify OCT images and eliminate time-consuming tasks, and this classification can also aid in the prevention of vision loss.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    The data used to support this study have not been made available access because they are real clinical data from Soonchunhyang Bucheon Hospital, and patient's privacy should be protected, it enables to detect people through this data, but they are available from the corresponding author on reasonable request.

    This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (No. NRF-2021R1A2C1010362) and the Soonchunhyang University Research Fund.

    The authors declare no competing interests.

    [1] Agochukwu NB, Solomon BD, Muenke M (2012) Impact of genetics on the diagnosis and clinical management of syndromic craniosynostoses. Childs Nerv Syst 28: 1447-1463. doi: 10.1007/s00381-012-1756-2
    [2] Rocco FD, Arnaud E, Renier D (2009) Evolution in the frequency of nonsyndromic craniosynostosis. J Neurosurg Pediat 4: 21-25. doi: 10.3171/2009.3.PEDS08355
    [3] Crouzon O (1912) Dysostose cranio-faciale hereditaire. Bull Mem Soc Med Hop Paris 33: 545-555.
    [4] Reardon W, Winter RM, Rutland P, et al. (1994) Mutations in the fibroblast growth factor receptor 2 gene cause Crouzon syndrome. Nature Genetics 8: 98-103. doi: 10.1038/ng0994-98
    [5] Collet C, Alessandri JL, Arnaud E, et al. (2014) Crouzon syndrome and Bent bone dysplasia associated with mutations at the same Tyr-381 residue in FGFR2 gene. Clin Genet 85: 598-599.
    [6] Lajeunie E, Heuertz S, Ghouzzi VE, et al. (2006) Mutation screening in patients with syndromic craniosynostoses indicates that a limited number of recurrent FGFR2 mutations accounts for severe forms of Pfeiffer syndrome. Eur J Hum Genet 14: 289-298. doi: 10.1038/sj.ejhg.5201558
    [7] Lajeunie E, Ma HW, Bonaventure J, et al. (1995) FGFR2 mutations in Pfeiffer syndrome. Nat Genet 9: 108.
    [8] Tartaglia M, Di Rocco C, Lajeunie E, et al. (1997) Jackson-Weiss syndrome: identification of two novel FGFR2 missense mutations shared with Crouzon and Pfeiffer craniosynostotic disorders. Hum Genet 101: 47-50. doi: 10.1007/s004390050584
    [9] Kan S, Elanko N, Johnson D, et al. (2002) Genomic screening of fibroblast growth-factor receptor 2 reveals a wide spectrum of mutations in patients with syndromic craniosynostosis. Am J Hum Genet 70: 472-486.
    [10] Meyers GA, Day D, Goldberg R, et al. (1996) FGFR2 exon IIIa and IIIc mutations in Crouzon, Jackson-Weiss, and Pfeiffer syndromes: evidence for missense changes, insertions, and a deletion due to alternative RNA splicing. Am J Hum Genet 58: 491-498.
    [11] Chen H, Ma J, Li W, et al. (2007) A molecular brake in the kinase hinge region regulates the activity of receptor tyrosine kinases. Mol Cell 27: 717-730. doi: 10.1016/j.molcel.2007.06.028
    [12] Wilkie AOM, Bochukova EG, Hansen RMS, et al. (2007) Clinical dividends from the molecular genetic diagnosis of craniosynostosis. Am J Med Genet 143: 1941-1949.
    [13] de Ravel TJL, Taylor IB, Van Oostveldt AJT, et al. (2005) A further mutation of the FGFR2 tyrosine kinase domain in mild Crouzon syndrome. Eur J Hum Genet 13: 503-505. doi: 10.1038/sj.ejhg.5201325
    [14] Zankl A, Jaeger G, Bonafé L, et al. (2004) Novel mutation in the tyrosine kinase domain of FGFR2 in a patient with Pfeiffer syndrome. Am J Med Genet A 131: 299-300.
    [15] Roscioli T, Elakis G, Cox TC, et al. (2013) Genotype and clinical care correlations in craniosynostosis: findings from a cohort of 630 Australian and New Zealand patients. Am J Med Genet C Semin Med Genet 163C: 259-270.
    [16] Collmann H, Sörensen N, Krauss J (2005) Hydrocephalus in craniosynostosis: a review. Childs Nerv Syst 21: 902-912. doi: 10.1007/s00381-004-1116-y
    [17] Raybaud C, Di Rocco C (2007) Brain malformation in syndromic craniosynostoses, a primary disorder of white matter: a review. Childs Nerv Syst 23: 1379-1388. doi: 10.1007/s00381-007-0474-7
    [18] McGillivray G, Savarirayan R, Cox T, et al. (2005) Familial scaphocephaly syndrome caused by a novel mutation in the FGFR2 tyrosine kinase domain. J Med Genet 42: 656-662. doi: 10.1136/jmg.2004.027888
  • This article has been cited by:

    1. Li Wang, Qunying Wu, Almost sure convergence theorems for arrays under sub-linear expectations, 2022, 7, 2473-6988, 17767, 10.3934/math.2022978
    2. He Dong, Xili Tan, Yong Zhang, Complete convergence and complete integration convergence for weighted sums of arrays of rowwise m-END under sub-linear expectations space, 2023, 8, 2473-6988, 6705, 10.3934/math.2023340
    3. Yansong Bai, Yong Zhang, Congmin Liu, Zhiming Wang, Moderate deviation principle for different types of classical likelihood ratio tests, 2022, 0361-0926, 1, 10.1080/03610926.2022.2144376
    4. Peiyu Sun, Dehui Wang, Xue Ding, Xili Tan, Yong Zhang, Moderate Deviation Principle for Linear Processes Generated by Dependent Sequences under Sub-Linear Expectation, 2023, 12, 2075-1680, 781, 10.3390/axioms12080781
    5. TAN Xili, DONG He, SUN Peiyu, ZHANG Yong, Almost Sure Convergence of Weighted Sums for m-END Sequences under Sub-linear Expectations, 2024, 1, 3006-0656, 10.59782/sidr.v1i1.26
    6. Xue Ding, Yong Zhang, Chover’s law of the iterated logarithm for weighted sums under sub-linear expectations, 2024, 53, 0361-0926, 6055, 10.1080/03610926.2023.2239399
    7. Peiyu Sun, Dehui Wang, Xili Tan, Equivalent Conditions of Complete p-th Moment Convergence for Weighted Sum of ND Random Variables under Sublinear Expectation Space, 2023, 11, 2227-7390, 3494, 10.3390/math11163494
    8. Xue Ding, Complete convergence for moving average process generated by extended negatively dependent random variables under sub-linear expectations, 2024, 53, 0361-0926, 8166, 10.1080/03610926.2023.2279924
    9. Xue Ding, Yong Zhang, Precise asymptotics for maxima of partial sums under sub-linear expectation, 2024, 0361-0926, 1, 10.1080/03610926.2024.2435585
  • Reader Comments
  • © 2016 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(5466) PDF downloads(1064) Cited by(0)

Figures and Tables

Figures(2)

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog