
In order to improve the accuracy of liquid level detection in intravenous left auxiliary vein infusion and reduce the pain of patients with blood returning from intravenous infusion, we propose a deep learning based liquid level detection model of infusion levels to facilitate this operation.
We implemented a Yolo v3-based detection model of infusion level images in intravenous infusion, and at the same time, compare it with SURF image processing technique, RCNN, and Fast-RCNN methods.
The model in this paper is better than the comparison algorithm in Intersection over Union (IoU), precision, recall and test time. The liquid level detection model based on Yolo v3 has a precision of 0.9768, a recall rate of 0.9688, an IoU of 0.8943, and a test time of 2.9 s.
The experimental results prove that the liquid level detection method based on deep learning has the characteristics of high accuracy and good real-time performance. This method can play a certain auxiliary role in the hospital environment and improve work efficiency of medical workers.
Citation: Zeyong Huang, Yuhong Li, Tingting Zhao, Peng Ying, Ying Fan, Jun Li. Infusion port level detection for intravenous infusion based on Yolo v3 neural network[J]. Mathematical Biosciences and Engineering, 2021, 18(4): 3491-3501. doi: 10.3934/mbe.2021175
[1] | Venkat Anil Adibhatla, Huan-Chuang Chih, Chi-Chang Hsu, Joseph Cheng, Maysam F. Abbod, Jiann-Shing Shieh . Applying deep learning to defect detection in printed circuit boards via a newest model of you-only-look-once. Mathematical Biosciences and Engineering, 2021, 18(4): 4411-4428. doi: 10.3934/mbe.2021223 |
[2] | Hassan Ali Khan, Wu Jue, Muhammad Mushtaq, Muhammad Umer Mushtaq . Brain tumor classification in MRI image using convolutional neural network. Mathematical Biosciences and Engineering, 2020, 17(5): 6203-6216. doi: 10.3934/mbe.2020328 |
[3] | Lorand Gabriel Parajdi, Radu Precup, Marcel-Adrian Şerban, Ioan Ştefan Haplea . Analysis of the effectiveness of the treatment of solid tumors in two cases of drug administration. Mathematical Biosciences and Engineering, 2021, 18(2): 1845-1863. doi: 10.3934/mbe.2021096 |
[4] | Junjing Cai, Qiwei Wang, Ce Wang, Yu Deng . Research on cell detection method for microfluidic single cell dispensing. Mathematical Biosciences and Engineering, 2023, 20(2): 3970-3982. doi: 10.3934/mbe.2023185 |
[5] | Rebeccah E. Marsh, Jack A. Tuszyński, Michael Sawyer, Kenneth J. E. Vos . A model of competing saturable kinetic processes with application to the pharmacokinetics of the anticancer drug paclitaxel. Mathematical Biosciences and Engineering, 2011, 8(2): 325-354. doi: 10.3934/mbe.2011.8.325 |
[6] | Tingxi Wen, Hanxiao Wu, Yu Du, Chuanbo Huang . Faster R-CNN with improved anchor box for cell recognition. Mathematical Biosciences and Engineering, 2020, 17(6): 7772-7786. doi: 10.3934/mbe.2020395 |
[7] | Jun Gao, Qian Jiang, Bo Zhou, Daozheng Chen . Convolutional neural networks for computer-aided detection or diagnosis in medical image analysis: An overview. Mathematical Biosciences and Engineering, 2019, 16(6): 6536-6561. doi: 10.3934/mbe.2019326 |
[8] | Siyu Chen, Yin Zhang, Yuhang Zhang, Jiajia Yu, Yanxiang Zhu . Embedded system for road damage detection by deep convolutional neural network. Mathematical Biosciences and Engineering, 2019, 16(6): 7982-7994. doi: 10.3934/mbe.2019402 |
[9] | Kangjian Sun, Ju Huo, Qi Liu, Shunyuan Yang . An infrared small target detection model via Gather-Excite attention and normalized Wasserstein distance. Mathematical Biosciences and Engineering, 2023, 20(11): 19040-19064. doi: 10.3934/mbe.2023842 |
[10] | Lei Yang, Guowu Yuan, Hao Wu, Wenhua Qian . An ultra-lightweight detector with high accuracy and speed for aerial images. Mathematical Biosciences and Engineering, 2023, 20(8): 13947-13973. doi: 10.3934/mbe.2023621 |
In order to improve the accuracy of liquid level detection in intravenous left auxiliary vein infusion and reduce the pain of patients with blood returning from intravenous infusion, we propose a deep learning based liquid level detection model of infusion levels to facilitate this operation.
We implemented a Yolo v3-based detection model of infusion level images in intravenous infusion, and at the same time, compare it with SURF image processing technique, RCNN, and Fast-RCNN methods.
The model in this paper is better than the comparison algorithm in Intersection over Union (IoU), precision, recall and test time. The liquid level detection model based on Yolo v3 has a precision of 0.9768, a recall rate of 0.9688, an IoU of 0.8943, and a test time of 2.9 s.
The experimental results prove that the liquid level detection method based on deep learning has the characteristics of high accuracy and good real-time performance. This method can play a certain auxiliary role in the hospital environment and improve work efficiency of medical workers.
Intravenous infusion is a common treatment method in the medical process [1,2,3]. It is widely used because of the fast infusion speed and good treatment effect. At present, medical technology continues to improve, and intravenous infusion is used to deliver nutrient solutions to the human body to improve resistance. In medical infusion, the problem of not being able to check the liquid level of the infusion bottle in time has always been a headache for patients and nursing staff [4]. If it is not handled in time, the infusion bottle will be empty, and the blood may return. If the situation is light, the patient will have swelling and pain at the injection site. Worse, it may also cause the patient to shock [5].
The application of intravenous infusion in the medical field has a long history, and unintelligent ways cause frequent problems. Therefore, the problem of infusion monitoring has caused many scholars to study. Mainly divided into the following categories: capacitive infusion monitoring [6,7], photoelectric infusion monitoring [8,9], laser sensor monitoring [10,11,12] and gravity infusion monitoring [13,14,15].
This paper uses image processing and deep learning methods to detect the liquid level of the infusion bottle. Our facility installs a high-definition camera that can perform image processing in the ward to take images of the infusion in real time. First, the target detection is performed on the position of the infusion bottle, then the level line of the infusion bottle is detected, and finally the position of the bit line is judged. When the liquid level line in the infusion bottle is lower than the pre-set warning line threshold, it indicates that the infusion is about to end, and an alarm is issued so that the nurse can go to the ward to change the liquid or draw the needle to the patient. This can improve work efficiency and allocate time reasonably.
Figure 1 shows a block diagram of the liquid level detection process of the entire infusion bottle. We put the image to be detected into the trained Yolo v3 target detection model [16,17] for target detection, then save the result image of the detection and perform area cropping. The crop size is the position size of the target frame, because the target frame is cropped, it is equivalent to the extraction of the infusion bottle.
Next, we send the cropped image into the trained model, perform classification judgment, and judge whether there is a liquid level. If the liquid level is over, a warning will be issued, otherwise, it will not be used.
As shown in Figure 2, our technique is depicted by a block diagram of the liquid level detection flow based on image processing. First, we preprocess the image, including grayscale processing, image filtering and pooling operations; then we use Canny edge [18,19,20] to detect the overall contour; and finally we use projection method to coarsely locate the liquid level. When the liquid level is below the set threshold, an alarm will be sent out, and if the liquid level is greater than the threshold, it means that the infusion has not ended.
The Canny algorithm calculates the gradient by calculating the first derivative of the Gaussian function to find the local maximum. For weak edge and strong edge detection, thresholds will be set separately for output. Therefore, the Canny algorithm has good anti-noise ability and can distinguish noise from edge detection and detect weak edges.
Integral projection method [21,22] is a mathematical statistical method that needs to accumulate the pixel values of each row or column and then present it. The method is simple to operate, fast in detection speed, and high in accuracy.
The Yolo series of target detection algorithms improve the accuracy of target detection. This paper uses the Yolo v3 algorithm [23]. The Yolo v3 algorithm is formed on the basis of v1 [24] and v2 [25,26]. The structure of Yolo v3 is shown in Figure 3.
Darknetconv2d-BN-Leaky (DBL) is the basic component of Yolo v3 [27]. Concat is to splice the upsampling of the darknet middle layer and a later layer.
Regions with CNN features (RCNN) is a milestone in the application of the CNN method to the target detection problem. It was proposed by Girshick [28]. With the help of CNN's good feature extraction and classification performance, the Region Estimation method is used to achieve the transformation of target detection problems. Region Estimation is a type of traditional region extraction method, which can be regarded as sliding windows of different widths and heights, and potential target images are obtained through window sliding. The network structure of RCNN is shown in Figure 4.
The main network of Fast-RCNN algorithm [29,30] is still VGG16 [31,32]. After inputting the image, after 5 convolutional layers and 2 downsampling layers, it enters the ROI Pooling layer. The input is the output of the conv5 layer and the region Estimation. Then go through two fully connected layers whose output is 4096. Finally, it passes through two fully connected layers with 21 and 84 outputs respectively. The network structure of Fast-RCNN is shown in Figure 5.
Speeded Up Robust Features (SURF) is a method based on traditional image processing [33,34]. In the field of computer vision, the scale space is the result of repeated convolution and sub-sampling. The more vivid image is the image pyramid. The SURF algorithm uses the method of changing the size of the filter to process multi-level spatial images, so the performance of the algorithm is improved.
As shown in Figure 6, the traditional pyramid and the SURF pyramid are compared. Image (A) is a Gaussian pyramid, changing the image size without changing the size of the filter; Image (B) is a pyramid constructed by the SURF algorithm, changing the size of the filter without changing the size of the original image.
Since there is no relevant standard data set available on the Internet, all the data in this paper are obtained by hand-photographing on-site using Apple mobile phones in our hospital. We collected photos of patients who came to our hospital for intravenous infusion treatment from September 2019 to September 2020. We obtained 280 original images.
In order to solve the problem of insufficient sample set, we carried out data enhancement processing on each image, as much as possible to simulate the noise generated by the sensor material property or circuit structure or image decoding when the image is collected in the real use scene, and the exposure caused by the photo. Perform the following 6 operations on each original image: adding salt and pepper noise, Gaussian noise, blurring, perspective transformation, contrast enhancement, and contrast reduction in sequence.
Finally, a total of 1680 data sets were obtained, and we randomly divided them into 80% training set and 20% test set. The specific numbers of training set and test set are shown in Table 1.
Classification | Quantity |
Training set | 1344 |
Test set | 336 |
We use the Bard three valve type intravenous infusion port produced by Bard Access Systems, Inc. of the United States, product number: 0603654, placed under the skin as a catheter and an injection seat. The catheter is made of radiographic silicone, with an inner diameter of 1.3 mm and a flow rate greater than 500 mL/hr. Other auxiliary equipment mainly include guide wire, puncture needle, curved non-damaged steel needle, vascular dilator and intubation sheath.
Precision refers to the proportion of true positive classes among all the positive classes judged to be positive. The computation is as shown in Eq (1).
(1) |
Recall refers to the proportion of all real positive classes that are judged to be positive. As shown in Eq (2).
(2) |
The IoU indicator refers to the intersection and ratio, and has been used as a standard metric in semantic segmentation. Cross-combination is not only used in semantic segmentation, but is also one of the commonly used indicators in target detection and other directions. As shown in Eq (3).
(3) |
We count and compare the test time based on traditional image processing methods and deep learning-based detection methods, and then determine which method is more advantageous.
Figure 7 shows the detection images of Yolo v3, RCNN, Fast-RCNN and SURF. It can be seen from the results that our proposed method has the highest accuracy.
We note that the liquid level detection based on Yolo v3 is the clearest, while the liquid level lines of the other three methods are fuzzy and intermittent. Yolo v3 is the closest to the liquid level in the infusion bottle, while the liquid level of RCNN, Fast-RCNN and SURF fluctuates up and down.
Table 2 shows the performance comparison of Yolo v3, RCNN, Fast-RCNN and SURF. We use Precision, Recall, IOU and test time to evaluate the four methods.
Method | Performance index | |||
Precision | Recall | IOU | Test time (s) | |
Yolo v3 | 0.9768 | 0.9688 | 0.8943 | 2.9 |
RCNN | 0.8427 | 0.8133 | 0.7986 | 9.4 |
Fast-RCNN | 0.8233 | 0.8004 | 0.7899 | 9.3 |
SURF | 0.7302 | 0.7211 | 0.7112 | 16.7 |
Intravenous infusion is the most common treatment in hospitals. However, because of the speed of infusion, the leakage of needles, and the untimely removal of needles, they may have unpredictable consequences for patients. Therefore, doctors and nurses need to pay close attention to the infusion process. This will undoubtedly increase the burden on medical staff, so it is very necessary to introduce computer technology into the process of intravenous infusion and realize the intelligentization of infusion.
Yolo regards target detection as a regression problem. Its basic idea is to first extract the input features through a feature extraction network to obtain a certain size of output, and then divide the image into several grids. There will be a fixed number of bounding boxes for the target, and the coordinates of the bounding box, the confidence of the object contained in the bounding box, and the probability of classification can be directly obtained by the target through the Yolo network. The most notable feature of Yolo is its relatively fast speed. Secondly, when Yolo processes the input, it is aimed at the entire image rather than a local area. In this case, the context information of the input image can be used well, so that the interference of the background is significantly reduced. As a target detection network, Yolo has a relatively strong generalization ability, and it can often achieve better results in target detection in natural environments.
In the process of liquid level detection of the infusion bottle, the solution proposed in this paper has achieved good detection results. It can be used for reference in the real hospital environment, which is helpful for the development of intelligent medical management and has broad application prospects. However, there are still some shortcomings in this paper that need further improvement.
First, we need to collect a large number of infusion bottle samples, and it takes more time to manually create a data set. The number of sample sets collected in this paper is limited. In order to achieve better detection results, the data set should be diversified; the object of this paper is to detect static images, and the detection of real-time video can be improved later.
Second, in order to improve the accuracy and real-time performance of target detection, one can try to introduce the latest target detection network framework proposed by authoritative scholars in the field of deep learning.
Third, when the liquid level line of the infusion bottle is severely covered and obscured by the medical trademark, it will affect the accurate detection of the liquid level line. The integral projection method needs to be improved, and the projection characteristics of the vertical edge and the horizontal edge are comprehensively used to set the line segment length reasonably. The threshold value can be used for further study.
Based on the problem of detecting the liquid level of the infusion bottle, we adopted four methods to compare the detection results. The Yolo v3 liguid level detection that is proposed in this paper has the advantages of simple operation and fast calculation speed. It can project the characteristics of the pixels in the image in the horizontal and vertical directions, so this method can accurately and quickly detect the liquid in the infusion bottle. Next, this paper divides the liquid level into two categories: anhydrous and with liquid, as a dual-classification problem, and then conducts training and learning, and finally can more accurately classify whether the liquid level category is liquid or anhydrous. After detecting the liquid level state, it can be judged whether the liquid level of the infusion bottle is in the no-liquid alarm state, so as to realize the function of intravenous infusion no-liquid detection.
The liquid level detection method based on Yolo v3 has certain practical value. This method can play a certain auxiliary role in the hospital environment and improve work efficiency. In addition, this method can also be applied to other liquid level detection environments, such as lake level detection, milk bottle level detection, etc.
There is no ethics approval required for this paper.
The authors declare no conflicts of interest.
[1] | M. Jordinson, A. J. Fitzgerald, R. A. Goodlad, A. Brynes, G. Grant, M. Pignatelli, et al., Systemic effect of peanut agglutinin following intravenous infusion into rats, Aliment. Pharm. Ther., 14 (2000), 835. |
[2] | J. Nemunaitis, C. Cunningham, A. Buchanan, A. Blackburn, G. Edelman, P. Maples, et al., Intravenous infusion of a replication-selective adenovirus (ONYX-015) in cancer patients: safety, feasibility and biological activity, Gene Ther., 8 (2001), 746-759. |
[3] | D. Pallares, M. R. Testelli, B. L. Fishleder, A. Bisteni, G. A. Medrano, C. Friedland, et al., Effects of an intravenous infusion of a potassium-glucose-insulin solution on the electrocardiographic signs of myocardial infarction, J. Am. Coll. Cardiol., 5 (1965), 166-181. |
[4] | K. B. Degn, B. Brock, C. B. Juhl, C. B. Djurhuus, J. Grubert, D. Kim, et al., Effect of intravenous infusion of exenatide (synthetic exendin-4) on glucose-dependent insulin secretion and counterregulation during hypoglycemia, Diabetes, 53 (2004), 2397-2403. |
[5] | E. Raymond, J. Alexandre, S. Faivre, K. Vera, E. Materman, J. Boni, et al., Safety and pharmacokinetics of escalated doses of weekly intravenous infusion of CCI-779, a novel mTOR inhibitor, in patients with cancer, J. Clin. Oncol., 22 (2004), 2336-2347. |
[6] | B. T. Tedner, H. S. Jacobson, D. Linnarsson, L. E. Lins, Impedance fluid volume monitoring during intravenous infusion in healthy subjects, Acute Care, 10 (1984), 200-206. |
[7] | D. B. Goodie, J. H. Philip, An analysis of the effect of venous resistance on the performance of gravity-fed intravenous infusion systems, J. Clin. Comput., 75 (1994), 127-132. |
[8] | J. W. Cui, H. Zhang, Design of intravenous infusion monitoring and alarm system based on wireless communication technology, IEEE Int. Conf. Mechatron. Automat., 2011. |
[9] | Y. Zhang, S. F. Zhang, Y. Ji, G. X. Wu, Wireless sensor network-enabled intravenous infusion monitoring, IET Wirel. Sens. Syst., 124 (2011), 107-112. |
[10] |
Y. C. Kim, S. F. Ahmad, H. D. Kim, Intravenous infusion monitoring sensor based on longitudinal electric field proximity sensing technique, J. Sens. Sci. Technol., 26 (2017), 101-106. doi: 10.5369/JSST.2017.26.2.101
![]() |
[11] | S. M. Bhavnani, S. L. Preston, Monitoring of intravenous quinidine infusion in the treatment of Plasmodium falciparum malaria, Ann. Pharmacotherl. 29 (1995), 33-35. |
[12] |
G. Rudnitsky, K. D. Miller, T. Padua, T. L. Stull, Continuous-infusion quinidine gluconate for treating children with severe plasmodium falciparum malaria, J. Infect. Dis., 155 (1987), 1040-1043. doi: 10.1093/infdis/155.5.1040
![]() |
[13] |
D. B. Goodie, J. H. Philip, An analysis of the effect of venous resistance on the performance of gravity-fed intravenous infusion systems, J. Clin. Monit. Comput., 10 (1994), 222-228. doi: 10.1007/BF02899506
![]() |
[14] | A. Meess, Platelet transfusion in chemotherapy patients: comparison of the effect of intravenous infusion pumps versus gravity transfusion, Brit. J. Biomed. Sci., 157 (2015), 213-220. |
[15] | N. Simon, B. Décaudin, D. Lannoy, M. F. Odou, M. D. Broucker, C. Barthélémy, et al., Impact of infusion method on amikacin serum levels in humans, Pulm. Pharmacol. Ther., 23 (2010), 324-326. |
[16] | X. Gao, D. Ge, Z. Chen, The research on autopilot system based on lightweight Yolo-V3 target detection algorithm, J. Phys. Conf. Ser., 1486 (2020), 032028. |
[17] | T. Q. Vinh, N. T. N. Anh, Real-time face mask detector using Yolo-V3 algorithm and haar cascade classifier, Advanced Comput. Appl. (ACOMP) Int. Conf., (2020), 146-149. |
[18] | P. Bao, L. Zhang, X. Wu, Canny edge detection enhancement by scale multiplication, IEEE T. Pattern Anal., 45 (2005), 78-82. |
[19] |
Q. Xu, S. Varadarajan, C. Chakrabarti, L. J. Karam, A distributed canny edge detector: algorithm and FPGA implementation, IEEE T. Image Process., 23 (2014), 2944-2960. doi: 10.1109/TIP.2014.2311656
![]() |
[20] | M. Amiri, F. M. Siddiqui, C. Kelly, R. Woods, K. Rafferty, B. Bardak., FPGA-based soft-core processors for image processing applications, J. Signal Process. Syst., 87 (2017), 1-18. |
[21] | P. A. Zuidema, E. Jongejans, P. D. Chien, H. J. During, F. Schieving, Integral projection models for trees: a new parameterization method and a validation of model output, J. Ecol., 55 (2010), 66-72. |
[22] |
C. Sert, A. Beskok, Spectral element formulations on non-conforming grids: a comparative study of pointwise matching and integral projection methods, J. Comput. Phys., 211 (2006), 300-325. doi: 10.1016/j.jcp.2005.05.019
![]() |
[23] | M. Ju, J. Luo, Z. Wang, H. Luo. Adaptive feature fusion with attention mechanism for multi-scale target detection, Neural. Comput. Appl., 7 (2020), 333-338. |
[24] | X. Mou, X. Chen, J. Guan, W. Zhou, N. Liu, Y. Dong, Clutter suppression and marine target detection for radar images based on INet, J. Radars, 9 (2020), 640-653. |
[25] | X. Wu, B. Xu, X. Chen, C. Jin, Visual target detection based on Yolo network structure, Boletín Técnico, 55 (2017), 6-13. |
[26] | H. Chen, L. Zhang, J. Ma, J. Zhang, Target heat-map network: an end-to-end deep network for target detection in remote sensing images, Neurocomputing, 331 (2018), 375-387. |
[27] | P. Adarsh, P. Rathi, M. Kumar, Yolo v3-tiny: object detection and recognition using one stage improved model, in 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS), (2020), 68-694. |
[28] | R. Girshick, J. Donahue, T. Darrell, J. Malik, Region-based convolutional networks for accurate object detection and segmentation, IEEE T. Pattern Anal., 38 (2015), 142-158. |
[29] | L. Quan, D. Pei, B. Wang, W. Ruan, Research on human target recognition algorithm of home service robot based on fast-RCNN, in 2017 10th International Conference on Intelligent Computation Technology and Automation (ICICTA), (2017), 369-373. |
[30] | R. Pan, W. Sun, Deep learning target detection based on pre-segmentation and regression, Prec. Eng., (2005), 221-227. |
[31] | A. K. Rangarajan, R. Purushothaman, Disease classification in eggplant using pre-trained VGG16 and MSVM, Rep-UK, 10 (2020), 2322. |
[32] |
C. D. Cunha, M. do Rosário, A. S. Rosado, S. G. F. Leite, Serratia SVGG16: a promising biosurfactant producer isolated from tropical soil during growth with ethanol-blended gasoline, Process. Biochem., 39 (2004), 2277-2282. doi: 10.1016/j.procbio.2003.11.027
![]() |
[33] | H. Bay, A. Ess, T. Tuytelaars, L. V. Gool, Speeded-up robust features (SURF), Comput. Vis Image Understanding, 110 (2008), 346-359. |
[34] | H. Bay, T. Tuytelaars, L. V. Gool, SURF: Speeded Up Robust Features, in European Conference on Computer Vision, Springer, Berlin, Heidelberg, 2006. |
1. | Yaqin Peng, Xin Zhang, Dandan Li, Yifei Chen, Yi Shen, 2022, An Infusion Liquid Level Detection Method Based on Improved ROI Segmentation and Horizontal Lines Modification, 978-988-75815-3-6, 6208, 10.23919/CCC55666.2022.9902495 | |
2. | Bowen Sun, Xiaofeng Wang, Ammar Oad, Amjad Pervez, Feng Dong, Automatic Ship Object Detection Model Based on YOLOv4 with Transformer Mechanism in Remote Sensing Images, 2023, 13, 2076-3417, 2488, 10.3390/app13042488 | |
3. | Yuan Gong, Li Yu, Fang Wang, Lixia Wang, Mengru Sun, Huiyu Li, Sorayouth Chumnanvej, The Radiographic Changes of Catheter Tip in Different Positions of Arm Infusion Port, 2022, 2022, 1555-4317, 1, 10.1155/2022/9374774 | |
4. | Miaolong Cao, Hao Fu, Jiayi Zhu, Chenggang Cai, Lightweight tea bud recognition network integrating GhostNet and YOLOv5, 2022, 19, 1551-0018, 12897, 10.3934/mbe.2022602 | |
5. | Zhengzhi Sun, Mayire Ibrayim, Askar Hamdulla, Detection of Pine Wilt Nematode from Drone Images Using UAV, 2022, 22, 1424-8220, 4704, 10.3390/s22134704 | |
6. | Isabel Pinheiro, André Aguiar, André Figueiredo, Tatiana Pinho, António Valente, Filipe Santos, Nano Aerial Vehicles for Tree Pollination, 2023, 13, 2076-3417, 4265, 10.3390/app13074265 | |
7. | Jaehwang Seol, Sangyun Lee, Jeongyun Park, Kwang Gi Kim, Automatic Control System for Precise Intravenous Therapy Using Computer Vision Based on Deep Learning, 2023, 11, 2169-3536, 121870, 10.1109/ACCESS.2023.3328568 | |
8. | Zhe Chen, Shihao Huang, Hui Lv, Zhixue Luo, Jinhao Liu, Defect detection in automotive glass based on modified YOLOv5 with multi-scale feature fusion and dual lightweight strategy, 2024, 40, 0178-2789, 8099, 10.1007/s00371-023-03225-x | |
9. | Akanksha Soni, Avinash Rai, 2024, YOLO for Medical Object Detection (2018–2024), 979-8-3503-9072-8, 1, 10.1109/ICEPES60647.2024.10653506 | |
10. | Harold Lucero, Viviana Moya, Andrea Pilco, Alexander Tirira, Qin Lei, 2023, Restricted Area Sign Detector Using YOLO v5, 979-8-3503-2331-3, 268, 10.1109/ICAICA58456.2023.10405395 | |
11. | Mohammed Gamal Ragab, Said Jadid Abdulkadir, Amgad Muneer, Alawi Alqushaibi, Ebrahim Hamid Sumiea, Rizwan Qureshi, Safwan Mahmood Al-Selwi, Hitham Alhussian, A Comprehensive Systematic Review of YOLO for Medical Object Detection (2018 to 2023), 2024, 12, 2169-3536, 57815, 10.1109/ACCESS.2024.3386826 | |
12. | Chutamas Deepho, Viphava Khlaisuwan, Chidsanuphong Pengchai, Warunporn Intarachana, Phatchara Rakchuai, Kullaphat Kajhan, Akkharawoot Takhom, 2024, Toward the Development of an Oral-diagnosis Framework: A Case Study of Teeth Segmentation and Numbering in Bitewing Radiographs via YOLO Models, 979-8-3503-7009-6, 1, 10.1109/ICCI60780.2024.10532585 |
Classification | Quantity |
Training set | 1344 |
Test set | 336 |
Method | Performance index | |||
Precision | Recall | IOU | Test time (s) | |
Yolo v3 | 0.9768 | 0.9688 | 0.8943 | 2.9 |
RCNN | 0.8427 | 0.8133 | 0.7986 | 9.4 |
Fast-RCNN | 0.8233 | 0.8004 | 0.7899 | 9.3 |
SURF | 0.7302 | 0.7211 | 0.7112 | 16.7 |