
Radiology experts often face difficulties in mammography mass lesion labeling, which may lead to conclusive yet unnecessary and expensive breast biopsies. This paper focuses on building an automated diagnosis tool that supports radiologists in identifying and classifying mammography mass lesions. The paper's main contribution is to design a hybrid model based on Pulse-Coupled Neural Networks (PCNN) and Deep Convolutional Neural Networks (CNN). Due to the need for large datasets to train and tune CNNs, which are not available for medical images, Transfer Learning (TL) was exploited in this research. TL can be an effective approach when working with small-sized datasets. The paper's implementation was tested on three public benchmark datasets: DDMS, INbreast, and BCDR datasets for training and testing and MIAS for testing only. The results indicated the enhancement that PCNN provides when combined with CNN compared to other methods for the same public datasets. The hybrid model achieved 98.72% accuracy for DDMS, 97.5% for INbreast, and 96.94% for BCDR. To avoid overfitting, the proposed hybrid model was tested on an unseen MIAS dataset, achieving 98.77% accuracy. Other evaluation metrics are reported in the results section.
Citation: Meteb M. Altaf. A hybrid deep learning model for breast cancer diagnosis based on transfer learning and pulse-coupled neural networks[J]. Mathematical Biosciences and Engineering, 2021, 18(5): 5029-5046. doi: 10.3934/mbe.2021256
[1] | Kun Lan, Jianzhen Cheng, Jinyun Jiang, Xiaoliang Jiang, Qile Zhang . Modified UNet++ with atrous spatial pyramid pooling for blood cell image segmentation. Mathematical Biosciences and Engineering, 2023, 20(1): 1420-1433. doi: 10.3934/mbe.2023064 |
[2] | Jinzhu Yang, Meihan Fu, Ying Hu . Liver vessel segmentation based on inter-scale V-Net. Mathematical Biosciences and Engineering, 2021, 18(4): 4327-4340. doi: 10.3934/mbe.2021217 |
[3] | Yu Li, Meilong Zhu, Guangmin Sun, Jiayang Chen, Xiaorong Zhu, Jinkui Yang . Weakly supervised training for eye fundus lesion segmentation in patients with diabetic retinopathy. Mathematical Biosciences and Engineering, 2022, 19(5): 5293-5311. doi: 10.3934/mbe.2022248 |
[4] | Yue Li, Hongmei Jin, Zhanli Li . A weakly supervised learning-based segmentation network for dental diseases. Mathematical Biosciences and Engineering, 2023, 20(2): 2039-2060. doi: 10.3934/mbe.2023094 |
[5] | Yantao Song, Wenjie Zhang, Yue Zhang . A novel lightweight deep learning approach for simultaneous optic cup and optic disc segmentation in glaucoma detection. Mathematical Biosciences and Engineering, 2024, 21(4): 5092-5117. doi: 10.3934/mbe.2024225 |
[6] | Danial Sharifrazi, Roohallah Alizadehsani, Javad Hassannataj Joloudari, Shahab S. Band, Sadiq Hussain, Zahra Alizadeh Sani, Fereshteh Hasanzadeh, Afshin Shoeibi, Abdollah Dehzangi, Mehdi Sookhak, Hamid Alinejad-Rokny . CNN-KCL: Automatic myocarditis diagnosis using convolutional neural network combined with k-means clustering. Mathematical Biosciences and Engineering, 2022, 19(3): 2381-2402. doi: 10.3934/mbe.2022110 |
[7] | Zhaoxuan Gong, Jing Song, Wei Guo, Ronghui Ju, Dazhe Zhao, Wenjun Tan, Wei Zhou, Guodong Zhang . Abdomen tissues segmentation from computed tomography images using deep learning and level set methods. Mathematical Biosciences and Engineering, 2022, 19(12): 14074-14085. doi: 10.3934/mbe.2022655 |
[8] | Duolin Sun, Jianqing Wang, Zhaoyu Zuo, Yixiong Jia, Yimou Wang . STS-TransUNet: Semi-supervised Tooth Segmentation Transformer U-Net for dental panoramic image. Mathematical Biosciences and Engineering, 2024, 21(2): 2366-2384. doi: 10.3934/mbe.2024104 |
[9] | Xiangfen Song, Yinong Wang, Qianjin Feng, Qing Wang . Improved graph cut model with features of superpixels and neighborhood patches for myocardium segmentation from ultrasound image. Mathematical Biosciences and Engineering, 2019, 16(3): 1115-1137. doi: 10.3934/mbe.2019053 |
[10] | Dongwei Liu, Ning Sheng, Tao He, Wei Wang, Jianxia Zhang, Jianxin Zhang . SGEResU-Net for brain tumor segmentation. Mathematical Biosciences and Engineering, 2022, 19(6): 5576-5590. doi: 10.3934/mbe.2022261 |
Radiology experts often face difficulties in mammography mass lesion labeling, which may lead to conclusive yet unnecessary and expensive breast biopsies. This paper focuses on building an automated diagnosis tool that supports radiologists in identifying and classifying mammography mass lesions. The paper's main contribution is to design a hybrid model based on Pulse-Coupled Neural Networks (PCNN) and Deep Convolutional Neural Networks (CNN). Due to the need for large datasets to train and tune CNNs, which are not available for medical images, Transfer Learning (TL) was exploited in this research. TL can be an effective approach when working with small-sized datasets. The paper's implementation was tested on three public benchmark datasets: DDMS, INbreast, and BCDR datasets for training and testing and MIAS for testing only. The results indicated the enhancement that PCNN provides when combined with CNN compared to other methods for the same public datasets. The hybrid model achieved 98.72% accuracy for DDMS, 97.5% for INbreast, and 96.94% for BCDR. To avoid overfitting, the proposed hybrid model was tested on an unseen MIAS dataset, achieving 98.77% accuracy. Other evaluation metrics are reported in the results section.
Coronary artery disease is the leading cause of cardiovascular mortality. Cardiac imaging has a pivotal role in preventing, diagnosing and treating ischemic heart disease. In recent years, non-invasive clinical cardiac imaging techniques have been rapidly developed; they are commonly used to assess myocardial ischemia and quantitative perfusion parameters, including single-photon emission computed tomography (SPECT), magnetic resonance imaging, computed tomography, myocardial contrast echocardiography (MCE), etc. Compared with the other cardiac imaging techniques, MCE has the advantages of being radiation-free, convenient and inexpensive (about 3–4 times less expensive than SPECT) [1]. MCE has been validated as an effective myocardial perfusion imaging method- to evaluate myocardial perfusion and infarction size [2,3], the microvascular changes after coronary revascularization [4] and the outcome in those undergoing a heart transplant [5].
In the clinical application of MCE, the ultrasound contrast agents (UEAs) containing the gas cores and lipid shells are injected intravenously into the myocardium. Contrast imaging can be generated from the signals produced by the resonance of microbubbles; blood containing microbubbles appear to be a bright white region [6]. Shown in Figure 1, after the distribution of the UEA reaches a stable state, high mechanical index impulses will be applied to clear all of the microbubbles, and then several end-systolic frames during the destruction-replenishment of UEA are selected to fit the time-intensity curve (shown in Figure 2), from which myocardial perfusion parameters such as blood volume, and flux rate can be obtained.
However, the analysis of MCE is very time-consuming, as it includes two steps that need to be done manually by experienced echocardiographers. Several end-systolic frames need to be extracted, and then the myocardium, i.e., the region of interest, needs to be segmented in the cardiac frames too. Therefore, automatic myocardial segmentation methods are desired for efficiency and operator-independence of the MCE perfusion analysis. Nevertheless, myocardial segmentation faces the following challenges. First, the concentration of UEA changes over time during destruction and replenishment, causing great intensity variations in images [7]. Second, the variations in shape and position of the myocardium according to different chambers, heart motions, patient individual differences, etc. Moreover, unclear myocardial borders and misleading structures such as papillary muscle have similar appearances to the myocardium.
Existing myocardial segmentation methods could be broadly classified into traditional image segmentation algorithms and machine learning algorithms. Traditional image segmentation algorithms define the segmentation task as a contour finding problem by using optimization methods based on image information, such as active contour [8] and an active shape model [9]. Malpica et al. [10] proposed a coupled active contour model guided by optical flow estimates to track the myocardium in MCE. Pickard et al. [11] applied principal component analysis with an active shape model algorithm to model the shape variability; they proposed a specialized gradient vector flow field to guide the contours to the myocardial borders, Guo et al. [12] proposed an automatic myocardial segmentation method based on an active contours model and neutrosophic similarity score; they applied a clustering algorithm to detect the initial ventricle region to speed up the evolution procedure and increase accuracy. However, due to the low complexity of the traditional image segmentation algorithm, it does not perform well on the MCE myocardial segmentation task with a large intensity variation [13] and it still needs manual tracing of myocardial contours; in addition, the optimization algorithm would be easily stuck in the local optimal solution without a good manual initial contour. Machine learning algorithms for myocardial segmentation tasks are often defined as pixel-level classification tasks, known as semantic segmentation. Li et al. [13] combined a random forest with a shape model, achieving notable improvement in segmentation accuracy compared with the classic random forest and active shape model. In recent years, deep learning has shown superior performance and great potential in medical image analysis; the majority of these deep learning approaches in cardiac ultrasound focus on left ventricle segmentation. Azarmehr et al. [14] experimented with three deep learning left ventricle (LV) segmentation models (U-Net, SegNet and fully connected DenseNets) on 992 echocardiograms, the U-net model outperformed the other models and achieved an average dice coefficient of 0.93. Veni et al. [15] proposed a U-net combined with a shape-driven deformable model in the form of a level set. The U-net model is used to produce the segmentation of LV, which is considered as a prior shape; then, the prior shape drives the level set to converge the final shape; the model produced a 0.86 dice coefficient on a private 2D echocardiographic dataset. Hu et al. [16] proposed a segmentation model based on a bilateral segmentation network (BiSeNet); it consists of two paths, a spatial path for capturing low-level spatial features and a context path for exploiting high-level context semantic features; they also used a fusion module to fuse the features of those two paths, achieving a dice coefficient of 0.932 and 0.908 in the left ventricle and left atrium, respectively. To the best of our knowledge, Li et al. [17] was the first to apply deep learning methods to MCE segmentation; they proposed an encoder-decoder architecture based on a U-net, introduced a bi-directional training schema incorporating temporal information in MCE sequences and achieved the highest segmentation precision compared to the traditional U-net model.
However, we believe that the MCE segmentation accuracy still has great improvement space due to the rapid development of new deep learning algorithms. Among all of the deep learning algorithms, DeepLabv3+ [18] has become an excellent algorithm in the field of medical segmentation by virtue of its ability to extract multi-scale information and its encoder-decoder structure. Thus, in this paper, we propose a semantic segmentation method based on DeepLabV3+ to solve the segmentation problem in the MCE automatic perfusion quantification.
Li et al. [17] have made the MCE dataset publicly available; it consists of MCE data from 100 patients from Guangdong Provincial People's Hospital. Apical two-chamber view (A2C), apical three-chamber view (A3C) and apical four-chamber view (A4C) MCE data were collected from each patient. Every MCE sequence has 30 end-systolic frames. In summary, there are 100 (patients) × 3 (chamber views) × 30 (end-systolic frames) = 9000 frames. The manual annotations of the myocardium were performed by an experienced echocardiographer. We split the dataset into the training dataset and test dataset at a proportion of 7:3. The segmentation models were trained for each chamber view separately; the detailed data information is illustrated in Table 1.
Patient number | MCE sequence number | Frame number | |
training data | 70 | 210 | 6300 |
testing data | 30 | 90 | 2700 |
As shown in Figure 3, the segmentation model was modified based on Deeplabv3+; it consists of a dilated ResNet backbone to extract feature maps, an atrous spatial pyramid pooling (ASPP) module to convert feature maps into multi-scale information and a decode module to generate the final predictions.
The backbone network was based on a modified 101 depth ResNet [19]. First, we replaced the 7 × 7 convolution in the input stem with a 3 × 3 convolution to improve the performance and accelerate the training process [20]. the standard ResNet uses downsampling operations such as a convolutional layer with a stride greater than 1 to increase feature maps. However, it would cause receptive field reduction; thus, DeepLabV3+ utilizes dialated convolutions [21] to alleviate spatial information losses from downsampling operations, also known as atrous convolution. The implementation of atrous convolution involves adding zeros between weights in the convolutional kernel with a stride of 1, as shown in Figure 4. In this way, features can be extracted across pixels, increasing the receptive field without introducing redundant parameters that need to be learned. Figure 5 shows a comparison of the original ResNet and dilated ResNet in the final two groups of the ResNet.
The ASPP module applies atrous convolution to extract multi-scale information by using atrous convolution with different dilation factors. The ASPP module consists of one 1 × 1 standard convolution and 3 × 3 atrous convolutions in parallel. The original DeepLabV3+ model proposed dilation factors of 6, 12 and 18 for atrous convolutions; however, the original structure may not be suitable because the myocardial border is not very clear due to the huge intensity variation in MCE; so, we added convolution with dilation factor 4 to the ASPP module to obtain more detailed spatial information. In conclusion, the ASPP module gets five feature maps from five parallel atrous convolutions and concatenates them together; it then sends them to the decoder. The detailed architecture of the ASPP module is shown in Figure 3.
The decoder decodes features aggregated by the encoder at multiple levels and generates a semantic segmentation mask from high dimensional feature vectors. The decoder is simple but effective, and it is the most most significant improvement of the DeepLabV3+ compared to the predecessor DeepLabV3 [22]; in this way, the detailed boundaries of the myocardium can be recovered faster and stronger [18]. In the decoder module, the feature map from the encoder is first upsampled by 4 bilinearly and then concatenated with the lower-level feature map from the backbone after channel reduction from 1 × 1 convolution; since the lower-level feature map contains more spatial information, the fusion of the lower-level feature map and high-level feature map improves the segmentation accuracy. After a 3 × 3 convolution, the segmentation prediction is obtained by upsampling by 4 bilinearly.
The model was implemented by using a Pytorch 1.11.0 deep learning framework and trained using NVIDIA RTX 2060 SUPER with 8 GB of memory. All images were center cropped to 256 × 256 and RGB pixels were normalized using the following: mean = [0.485, 0.456, 0.406], standard deviation = [0.229, 0.224, 0.225].
For data augmentation, during the training phase, all images were randomly scaled by [0.8, 1.2], rotated by [-5◦, 5◦] and randomly flipped by a probability of 0.5. During the testing phase, we did not apply any augmentations.
Every model was trained for 80 epochs that contained 84000 iterations. Stochastic gradient descent was used as the optimizer, where the momentum was set to 0.9 and the weight decay was set to 3e-5. The initial learning rate was set to 0.01, and the minimum learning rate was set to 0.001 and followed the polynomial decay policy, which is defined as
lr=initial_lr×(1−iterationnum_iteration)power | (1) |
where iteration represents the current iteration, num_iteration represents the total iteration, initial_lr = 0.01, power = 0.9.
Moreover, we used dice loss [23] as our loss function due to the imbalance problem, because the myocardium is small compared to the large heart chamber; it is defined as
lossdice(P,T)=1−2|P∩T||P|+|T| | (2) |
where P is the predicted myocardium area and T is the ground truth of the myocardium area. In addition, we added an auxiliary loss [24] only in the training phase to help optimize the learning process.
The Dice coefficient and intersection over union (IoU) were used as evaluation criteria to evaluate the performance of the model; they are defined as
dice(P,T)=2|P∩T||P|+|T| | (3) |
IOU(P,T)=|P∩T|P∪T | (4) |
where P is the predicted myocardium area and T is the ground truth.
The visualization of the segmentation results are illustrated in Figure 6; six apical four-chamber view MCE images were randomly selected from a subject in the test dataset and input into our trained model. It can be seen from the figure that the proposed model gets the correct prediction of myocardium in the presence of the misleading structure, papillary muscle. The boundaries of the predicted segmentation area have a great match with the ground truth.
Moreover, we compared the modified DeepLabV3+ to the original DeepLabV3+, the results of Li et al. [17] and other state-of-the-art models, e.g., a U-net [25] with a Deeplabv3 backbone and PSPnet [24] with a ResNet-101 backbone; the results are shown in Table 2.
Modifed | Original |
||||
DeepLabV3+ | DeepLabV3+ | ||||
Dice | |||||
A2C | 0.84 | 0.84 | 0.81 | 0.83 | 0.82 |
A3C | 0.84 | 0.83 | 0.81 | 0.84 | 0.82 |
A4C | 0.86 | 0.84 | 0.82 | 0.83 | 0.80 |
IoU | |||||
A2C | 0.74 | 0.72 | 0.69 | 0.73 | 0.69 |
A3C | 0.72 | 0.71 | 0.65 | 0.72 | 0.70 |
A4C | 0.75 | 0.75 | 0.71 | 0.72 | 0.72 |
The modified DeepLabV3+ improved the segmentation results for the dice coefficient in A3C by 0.01 and in A4C by 0.02, and for the IoU in A2C by 0.02 and A3C by 0.01. Moreover, the modified DeepLabV3+ also outperformed other state-of-the-art models in both metrics.
To see the trade-off between model performance and complexity, we also tested the model with different depths of ResNet, i.e., 18 and 50, and compared it with the PSPnet and U-net. The number of parameters and GFlops (giga floating-point operations per second) were selected to indicate the model complexity, and the average IoU of all chamber views was selected to evaluate the model performance. In addition, the MCE frame per second processed (FPS) was evaluated on a personal computer with the RTX 2060 super and an Intel® Core™ i5-9600 processor; the results are shown in Table 3 and the comparisons of the number of parameters and GFlops to the average IoU are illustrated in Figure 7 and Figure 8, respectively.
No. of parameters | GFlops | Average IoU (%) | FPS | |
modified DeepLabV3 + (ResNet18) | 12.47 | 54.21 | 70.94 | 39.6 |
modified DeepLabV3 + (ResNet50) | 43.58 | 176.25 | 72.81 | 21.2 |
modified DeepLabV3 + (ResNet101) | 62.68 | 255.14 | 74.23 | 15.2 |
PSPnet (ResNet101) | 68.07 | 256.44 | 72.68 | 15.7 |
U-net | 29.06 | 203.43 | 71.02 | 20.5 |
Although the parameter count of PSPnet (ResNet101) increased 36.87% and GFlops increased 31.27% relative to the modified DeepLabV3+ (ResNet50), DeepLabV3+ (ResNet50) still outperformed PSPnet, which proves the efficiency of the proposed model. Comparing different ResNet backbone depths of the modified DeepLabV3+ and 18 depth only had 28.61% and 19.89% of the parameter count of the 50 depth and 101 depth, respectively, and 30.76% and 21.25% of GFlops of the 50 depth and 101 depth, respectively; it still had a 97.43% IoU for the 50 depth and 95.56% IoU for the 101 depth.
The feasibility of model application takes both performance and computational complexity into consideration; we believe balance can be made based on the depth of the backbone ResNet.
This paper proposed a modified architecture DeepLabV3+ model for MCE segmentation. The model consists of three main modules: the backbone, ASPP module and decoder. The backbone utilizes ResNet with atrous convolution, which allows the algorithm to find the best balance between the receptive field from a large field of view and the resolution of the feature map from a small field of view. The modified ASPP module also applies atrous convolution with different dilation factors to resample multi-scale patterns from the feature maps extracted from the backbone. The decoder combines the lower-level feature map from the encoder and high-level feature map from the decoder to generate the final prediction. A comparison between the proposed model and other state-of-the-art models, i.e., the PSPnet and U-net was conducted; the proposed model has achieved the best scores for both the dice and IoU. Moreover, we also did a performance and complexity analysis for all of the models, including the proposed model with different backbone depths, PSPnet and U-net; the results show the efficiency of the proposed architecture, and a comparison of the different depths of ResNet backbone illustrated the application feasibility of the proposed model. In the future study, we will focus on the balance of performance and complexity of the model to seek opportunities for the application in clinical analysis. Moreover, the Li et al. data only provides MCE frames rendered by a coloring mapping procedure; however, the rendering schema depends on settings from different companies and radiologists, so it will reduce the objectives and make the segmentation lose detail, which might bring variation to the algorithm's performance. Thus, we hope to experiment with our model on original MCE frames to improve the robustness and generalization ability in the future.
This work was supported by the Natural Science Foundation of China (NSFC) under grant number 62171408, and the Key Research and Development Program of Zhejiang Province (2020C03060, 2020C03016, 2022C03111).
The authors declare that there is no conflict of interest.
[1] |
S. Negoita, E. Feuer, A. Mariotto, K. A Cronin, V. I Petkov, S. K Hussey, et al., Annual report to the nation on the status of cancer, part Ⅱ: Recent changes in prostate cancer trends and disease characteristics, Cancer, 124 (2018), 2801-2814. doi: 10.1002/cncr.31549
![]() |
[2] |
R. L. Siegel, K. D. Miller, A. Jemal, Cancer statistics, 2016, CA. Cancer J. Clin., 66 (2016), 7-30. doi: 10.3322/caac.21332
![]() |
[3] |
R. L. Siegel, K. D. Miller, A. Jemal, Cancer statistics, 2017, CA. Cancer J. Clin., 67 (2017), 7-30. doi: 10.3322/caac.21387
![]() |
[4] | L. Berlin, Radiologic errors, past, present and future, Diagnosis, 1 (2014), 79-84. |
[5] |
Y. Guo, X. Shang, Z. Li, Identification of cancer subtypes by integrating multiple types of transcriptomics data with deep learning in breast cancer, Neurocomputing, 324, (2019), 20-30. doi: 10.1016/j.neucom.2018.03.072
![]() |
[6] |
B. Sahiner, A. Pezeshk, L. M. Hadjiiski, X. Wang, K. Drukker, K. H. Cha, et al., Deep learning in medical imaging and radiation therapy, Med. Phys., 46 (2019), e1-e36. doi: 10.1002/mp.13264
![]() |
[7] |
H. M. Ahmad, M. J. Khan, A. Yousaf, S. Ghuffar, K. Khurshid, Deep learning: a breakthrough in medical imaging, Curr. Med. Imaging, 16, (2020), 946-956. doi: 10.2174/1573405615666191219100824
![]() |
[8] |
G. Litjens, T. Kooi, B. Bejnordi, A. Setio, F. Ciompi, M. Ghafoorian, et al., A survey on deep learning in medical image analysis, Med. Image Anal., 42 (2017), 60-88. doi: 10.1016/j.media.2017.07.005
![]() |
[9] |
A. A. Mohamed, W. A. Berg, H. Peng, Y. Luo, R. C. Jankowitz, S. Wu, A deep learning method for classifying mammographic breast density categories, Med. Phys., 45 (2018), 314-321. doi: 10.1002/mp.12683
![]() |
[10] |
F. F. Ting, Y. J. Tan, K. S. Sim, Convolutional neural network improvement for breast cancer classification, Expert Syst. Appl., 120 (2019), 103-115. doi: 10.1016/j.eswa.2018.11.008
![]() |
[11] |
D. M. Vo, N. -Q. Nguyen, S.-W. Lee, Classification of breast cancer histology images using incremental boosting convolution networks, Inf. Sci. (Ny)., 482 (2019), 123-138. doi: 10.1016/j.ins.2018.12.089
![]() |
[12] |
F. Ciompi, B. de Hoop, V. R Sarah, K. Chung, E. Scholten, M. Oudkerk, et al., Automatic classification of pulmonary peri-fissural nodules in computed tomography using an ensemble of 2D views and a convolutional neural network out-of-the-box, Med. Image Anal., 26 (2015), 195-202. doi: 10.1016/j.media.2015.08.001
![]() |
[13] |
H. Shin, H. Roth, M. Gao, L. Lu, Z. Xu, I. Nogues, et al., Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning, IEEE Trans. Med. Imaging, 35 (2016), 1285-1298. doi: 10.1109/TMI.2016.2528162
![]() |
[14] | A. Krizhevsky, I. Sutskever, G. E. Hinton, Machine learning and computer vision group deep learning with tensorflow, 2012. |
[15] | K. Simonyan, A. Zisserman, Very deep convolutional networks for large-scale image recognition, in Computer Vision and Pattern Recognition, 2014. |
[16] | Z. Cömert, A. F. Kocamaz, Fetal hypoxia detection based on deep convolutional neural network with transfer learning approach, Comput. Sci. On-line Conf., (2018), 239-248. |
[17] | M. Oquab, L. Bottou, I. Laptev, J. Sivic, Learning and transferring mid-level image representations using convolutional neural networks, in Conference on Computer Vision and Pattern Recognition, (2014), 1717-1724. |
[18] | M. S. Abdel-Wahab, M. Aboul-Ela, A. Samir, Arabic sign language recognition using neural network and graph matching techniques, Proceed. Int. Conf. App. Inf. Commun., (2006), 163-168. |
[19] |
A. SamirElons, M. Abull-ela, M. F. Tolba, Pulse-coupled neural network feature generation model for Arabic sign language recognition, IET Image Process., 7 (2013), 829-836. doi: 10.1049/iet-ipr.2012.0222
![]() |
[20] |
M. F. Akay, Support vector machines combined with feature selection for breast cancer diagnosis, Expert Syst. Appl., 36 (2009), 3240-3247. doi: 10.1016/j.eswa.2008.01.009
![]() |
[21] |
M. Karabatakm, M. C. Ince, An expert system for detection of breast cancer based on association rules and neural network, Expert Syst. Appl., 36 (2009), 3465-3469. doi: 10.1016/j.eswa.2008.02.064
![]() |
[22] | M. B. Rodrigues, R. D. NóBrega, S. Alves, P. Fillho, J. Durate, A. Sangaiah, et al., Health of things algorithms for malignancy level classification of lung nodules, IEEE Access, 6 (2012), 18592-18601. |
[23] | N. Arunkumar, M. A. Mohammed, S. A. Mostafa, D. A. Ibrahim, J. Rodrigues, V. H. C. de Albuquerque, Fully automatic model‐based segmentation and classification approach for MRI brain tumor using artificial neural networks, Concurr. Comput. Pract. Expert, 32 (2020), e4962. |
[24] | C. M. J. M. Dourado, S. P. P. D. Silva, R. V. M. D. Nóbrega, P. P. R. Filho, K. Muhammad, V. H. C. D. Albuquerque, An open IoHT-based deep learning framework for online medical image recognition, IEEE J. Sel. Areas Commun., 39 (2020), 541-548. |
[25] | K. Muhammad, S. Khan, J. D. Ser, V. H. C. de Albuquerque, Deep learning for multigrade brain tumor classification in smart healthcare systems: A prospective survey, IEEE Trans. Neural Networks Learn. Syst., (2020), 1-16. |
[26] |
C. A. Pena-Reyes, M. Sipper, A fuzzy-genetic approach to breast cancer diagnosis, Artif. Intell. Med., 17 (1999), 131-155. doi: 10.1016/S0933-3657(99)00019-6
![]() |
[27] | B. N. Dontchos, A. Yala, R. Barzilay, J. Xiang, C. D. Lehman, External validation of a deep learning model for predicting mammographic breast density in routine clinical practice, Acad. Radiol., 28 (2020), 475-480. |
[28] | T. D. Truong, H. T. -T. Pham, Breast cancer histopathological image classification utilizing convolutional neural network, " in International Joint Conference on Neural Networks (IJCNN), 69 (2019), 531. |
[29] |
M. Abdar, M. Zomorodi-Moghadam, X. Zhou, R. Gururajan, X. Tao, P. D. Barua, et al., A new nested ensemble technique for automated diagnosis of breast cancer, Pattern Recognit. Lett., 132 (2020), 123-131. doi: 10.1016/j.patrec.2018.11.004
![]() |
[30] |
P. Ferroni, F. M. Zanzotto, S. Riondino, N. Scarpato, F. Guadagni, M. Roselli, Breast cancer prognosis using a machine learning approach, Cancers (Basel), 11 (2019), 328. doi: 10.3390/cancers11030328
![]() |
[31] | D. A. Omondiagbe, S. Veeramani, A. S. Sidhu, Machine learning classification techniques for breast cancer diagnosis, in IOP Conference Series: Materials Science and Engineering., 495 (2019), 12033. |
[32] |
Y Tseng, C. Huang, C. Wen, P. Lai, M. Wu, Y. Sun, et al., Predicting breast cancer metastasis by using serum biomarkers and clinicopathological data with machine learning technologies, Int. J. Med. Inform., 128 (2019), 79-86. doi: 10.1016/j.ijmedinf.2019.05.003
![]() |
[33] |
M. Abdar, V. Makarenkov, CWV-BANN-SVM ensemble learning classifier for an accurate diagnosis of breast cancer, Measurement, 146 (2019), 557-570. doi: 10.1016/j.measurement.2019.05.022
![]() |
[34] | E. H. Houssein, M. M. Emam, A. A. Ali, P. N. Suganthan, Deep and machine learning techniques for medical imaging-based breast cancer: A comprehensive review, Expert Syst. Appl., (2020), 114161. |
[35] | J. Yosinski, J. Clune, Y. Bengio, H. Lipson, How transferable are features in deep neural networks?, Adv. Neural Inf. Process. Syst., 4 (2014), 3320-3328. |
[36] | K. Bowyer, D. Kopans, W. Kegelmeyer, R. Moore, M. Sallam, K. Chang, et al., The digital database for screening mammography, in Proceedings of the Fifth International Workshop on Digital Mammography, 58 (1996), 27. |
[37] |
I. C. Moreira, I. Amaral, I. Domingues, A. Cardoso, M. J. Cardoso, J. S. Cardoso, INbreast: toward a full-field digital mammographic database, Acad. Radiol., 19 (2012), 236-248. doi: 10.1016/j.acra.2011.09.014
![]() |
[38] | J. Suckling, J. Parker, D. Dance, S. Astley, I. Hutt, C. Boggis, et al., Mammographic Image Analysis Society (MIAS) Database v1. 21[Dataset], 2015. |
[39] | Y. LeCun, The MNIST Database of Handwritten Digits, 1998. |
[40] | Y. Jia, E. Shelhamer, J. Donahue, S. Karayev, J. Long, R. Girshick, et al., Caffe: Convolutional architecture for fast feature embedding, in Proceedings of the 22nd ACM international conference on Multimedia, (2014), 675-678. |
[41] |
O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, et al., ImageNet large scale visual recognition challenge, Int. J. Comput. Vis., 115 (2015), 211-252. doi: 10.1007/s11263-015-0816-y
![]() |
[42] |
B. Q. Huynh, H. Li, M. L. Giger, Digital mammographic tumor classification using transfer learning from deep convolutional neural networks, J. Med. Imaging, 3 (2016), 034501. doi: 10.1117/1.JMI.3.3.034501
![]() |
[43] | J. D. Gallego-Posada, D. A. Montoya-Zapata, O. L. Quintero-Montoya, D. A. Montoya-Zapa, Detection and diagnosis of breast tumors using deep convolutional neural networks, in Conference Proceedings of XVⅡ Latin American Conference in Automatic Control, (2016), 17. |
[44] | P. Xi, C. Shu, R. Goubran, Abnormality detection in mammography using deep convolutional neural networks, Comput. Sci., 2018. |
[45] | D. Lévy, A. Jain, Breast mass classification from mammograms using deep convolutional neural networks, Comput. Sci., 2016. |
[46] | H. S. Ranganath, G. Kuntimad, J. L. Johnson, Pulse coupled neural networks for image processing, in Proceedings IEEE Southeastcon 95 Visualize the Future, (1995), 37-43. |
[47] |
H. Jia, X. Peng, L. Kang, Y. Li, Z. Jiang, K. Sun, Pulse coupled neural network based on Harris hawks optimization algorithm for image segmentation, Multimed. Tools Appl., 79 (2020), 28369-28392. doi: 10.1007/s11042-020-09228-3
![]() |
[48] | T. Kalaiselvi, K. Rahimunnisa, A. S. Begum, Department of electronics and instrumentation engineering, easwari, Pulse, 29 (2020), 4411-4415. |
[49] |
A. Binder, M. Bockmayr, M. Hägele, S. Wienert, D. Heim, K. Hellweg, et al., Morphological and molecular breast cancer profiling through explainable machine learning, Nat. Mach. Intell., 3 (2021), 355-366. doi: 10.1038/s42256-021-00303-4
![]() |
[50] |
Y. -D. Zhang, S. C. Satapathy, D. S. Guttery, J. M. Górriz, S. -H. Wang, Improved breast cancer classification through combining graph convolutional network and convolutional neural network, Inf. Process. Manag., 58 (2021), 102439. doi: 10.1016/j.ipm.2020.102439
![]() |
[51] | M. Abdar, F. Pourpanah, S. Hussain, D. Rezazadegan, L. Liu, M. Ghavamzadeh, et al., A review of uncertainty quantification in deep learning: techniques, applications and challenges, Inf. Fus., 2021. |
[52] | R. Alizadehsani, M. Roshanzamir, S. Hussain, A. Khosravi, A. Koohestani, M. H Zangooei, et al., Handling of uncertainty in medical data using machine learning and probability theory techniques: a review of 30 years (1991-2020), Ann. Oper. Res., (2021), 1-42. |
1. | Yuchen Wu, Jin Li, Junkai Yang, 2023, Using Improved DeepLabV3+ for Complex Scene Segmentation, 979-8-3503-0562-3, 855, 10.1109/AUTEEE60196.2023.10408693 | |
2. | Rongpu Cui, Shichu Liang, Weixin Zhao, Zhiyue Liu, Zhicheng Lin, Wenfeng He, Yujun He, Chaohui Du, Jian Peng, He Huang, A Shape-Consistent Deep-Learning Segmentation Architecture for Low-Quality and High-Interference Myocardial Contrast Echocardiography, 2024, 50, 03015629, 1602, 10.1016/j.ultrasmedbio.2024.06.001 | |
3. | Tomonari Yamada, Takaaki Yoshimura, Shota Ichikawa, Hiroyuki Sugimori, Improving Cerebrovascular Imaging with Deep Learning: Semantic Segmentation for Time-of-Flight Magnetic Resonance Angiography Maximum Intensity Projection Image Enhancement, 2025, 15, 2076-3417, 3034, 10.3390/app15063034 | |
4. | Saurabhi Samant, Anastasios Nikolaos Panagopoulos, Wei Wu, Shijia Zhao, Yiannis S. Chatzizisis, Artificial Intelligence in Coronary Artery Interventions: Preprocedural Planning and Procedural Assistance, 2025, 4, 27729303, 102519, 10.1016/j.jscai.2024.102519 |
Patient number | MCE sequence number | Frame number | |
training data | 70 | 210 | 6300 |
testing data | 30 | 90 | 2700 |
Modifed | Original |
||||
DeepLabV3+ | DeepLabV3+ | ||||
Dice | |||||
A2C | 0.84 | 0.84 | 0.81 | 0.83 | 0.82 |
A3C | 0.84 | 0.83 | 0.81 | 0.84 | 0.82 |
A4C | 0.86 | 0.84 | 0.82 | 0.83 | 0.80 |
IoU | |||||
A2C | 0.74 | 0.72 | 0.69 | 0.73 | 0.69 |
A3C | 0.72 | 0.71 | 0.65 | 0.72 | 0.70 |
A4C | 0.75 | 0.75 | 0.71 | 0.72 | 0.72 |
No. of parameters | GFlops | Average IoU (%) | FPS | |
modified DeepLabV3 + (ResNet18) | 12.47 | 54.21 | 70.94 | 39.6 |
modified DeepLabV3 + (ResNet50) | 43.58 | 176.25 | 72.81 | 21.2 |
modified DeepLabV3 + (ResNet101) | 62.68 | 255.14 | 74.23 | 15.2 |
PSPnet (ResNet101) | 68.07 | 256.44 | 72.68 | 15.7 |
U-net | 29.06 | 203.43 | 71.02 | 20.5 |
Patient number | MCE sequence number | Frame number | |
training data | 70 | 210 | 6300 |
testing data | 30 | 90 | 2700 |
Modifed | Original |
||||
DeepLabV3+ | DeepLabV3+ | ||||
Dice | |||||
A2C | 0.84 | 0.84 | 0.81 | 0.83 | 0.82 |
A3C | 0.84 | 0.83 | 0.81 | 0.84 | 0.82 |
A4C | 0.86 | 0.84 | 0.82 | 0.83 | 0.80 |
IoU | |||||
A2C | 0.74 | 0.72 | 0.69 | 0.73 | 0.69 |
A3C | 0.72 | 0.71 | 0.65 | 0.72 | 0.70 |
A4C | 0.75 | 0.75 | 0.71 | 0.72 | 0.72 |
No. of parameters | GFlops | Average IoU (%) | FPS | |
modified DeepLabV3 + (ResNet18) | 12.47 | 54.21 | 70.94 | 39.6 |
modified DeepLabV3 + (ResNet50) | 43.58 | 176.25 | 72.81 | 21.2 |
modified DeepLabV3 + (ResNet101) | 62.68 | 255.14 | 74.23 | 15.2 |
PSPnet (ResNet101) | 68.07 | 256.44 | 72.68 | 15.7 |
U-net | 29.06 | 203.43 | 71.02 | 20.5 |