Research article

Real-Time personnel safety detection in chemical parks using YOLOv8-ARR

  • Published: 16 May 2025
  • With the rapid development of the chemical industry, personnel safety has become a critical concern. Traditional monitoring systems, which rely heavily on human surveillance, are inefficient and often inaccurate. To address this issue, we proposed a real-time detection and identification system for personnel safety in chemical parks based on an advanced artificial intelligence algorithm named YOLOv8-ARR. The key contributions of this system include: (1) The introduction of Adaptive Powerful-IoU(APIoU) as a network optimization for bounding box regression loss, which effectively balances gradient gains between high-quality and low-quality samples, enhancing model localization; (2) a novel attention mechanism, Reinforced Channel Prioritized Contextual Attention(RCPCA), to improve background information extraction; (3) replacing traditional convolution with RFAConv to assign different weights to each receptive field position and feature channel, highlighting crucial details; (4) the use of a bidirectional feature pyramid network (BiFPN) for the weighted fusion of multi-scale feature maps; and (5) the addition of a small object detection layer in the YOLOv8 network to enhance the detection of small targets. Experimental results on a custom dataset of chemical park workers showed that our model improves the mean Average Precision (mAP@0.5) by 5.475% compared to the standard model. This system provides a more accurate solution for identifying abnormal behaviors and potential risks in chemical parks compared to traditional methods. Additionally, it significantly reduces dependency on human resources, minimizes false positives and negatives, and enhances monitoring efficiency and safety.

    Citation: Zhong Wang, Lanfang Lei, Tong Li, Peibei Shi. Real-Time personnel safety detection in chemical parks using YOLOv8-ARR[J]. AIMS Electronics and Electrical Engineering, 2025, 9(3): 260-287. doi: 10.3934/electreng.2025013

    Related Papers:

  • With the rapid development of the chemical industry, personnel safety has become a critical concern. Traditional monitoring systems, which rely heavily on human surveillance, are inefficient and often inaccurate. To address this issue, we proposed a real-time detection and identification system for personnel safety in chemical parks based on an advanced artificial intelligence algorithm named YOLOv8-ARR. The key contributions of this system include: (1) The introduction of Adaptive Powerful-IoU(APIoU) as a network optimization for bounding box regression loss, which effectively balances gradient gains between high-quality and low-quality samples, enhancing model localization; (2) a novel attention mechanism, Reinforced Channel Prioritized Contextual Attention(RCPCA), to improve background information extraction; (3) replacing traditional convolution with RFAConv to assign different weights to each receptive field position and feature channel, highlighting crucial details; (4) the use of a bidirectional feature pyramid network (BiFPN) for the weighted fusion of multi-scale feature maps; and (5) the addition of a small object detection layer in the YOLOv8 network to enhance the detection of small targets. Experimental results on a custom dataset of chemical park workers showed that our model improves the mean Average Precision (mAP@0.5) by 5.475% compared to the standard model. This system provides a more accurate solution for identifying abnormal behaviors and potential risks in chemical parks compared to traditional methods. Additionally, it significantly reduces dependency on human resources, minimizes false positives and negatives, and enhances monitoring efficiency and safety.



    加载中


    [1] Girshick R (2015) Fast r-cnn. Proceedings of the IEEE international conference on computer vision.
    [2] Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 28.
    [3] Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu CY, et al. (2016) Ssd: Single shot multibox detector. In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I 14, 21‒37. Springer. https://doi.org/10.1007/978-3-319-46448-0_2
    [4] Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. Proceedings of the IEEE conference on computer vision and pattern recognition, 779‒788. https://doi.org/10.1109/CVPR.2016.91
    [5] Rubaiyat AH, Toma TT, Kalantari-Khandani M, Rahman SA, Chen L, Ye Y, et al. (2016) Automatic detection of helmet uses for construction safety. 2016 IEEE/WIC/ACM International Conference on Web Intelligence Workshops (WIW), 135‒142. IEEE. https://doi.org/10.1109/WIW.2016.045
    [6] Sun X, Xu K, Wang S, Wu C, Zhang W, Wu H (2021) Detection and tracking of safety helmet in factory environment. Meas Sci Technol 32: 105406. https://doi.org/10.1088/1361-6501/ac06ff doi: 10.1088/1361-6501/ac06ff
    [7] Huang X, Jia K, Liu P (2020) Automatic Detection of Taxi Driver Smoking Behavior Based on Traffic Monitoring. Computer Simulation 37: 337‒344.
    [8] Park S, Yoon S, Heo J (2019) Image-based automatic detection of construction helmets using R-FCN and transfer learning. KSCE Journal of Civil and Environmental Engineering Research 39: 399‒407.
    [9] Widiarsini KU, Khrisne DC, Suyadnya IMA (2021) Automatic Cigarette Object Concealment in Video using R-CNN.
    [10] Son H, Choi H, Seong H, Kim C (2019) Detection of construction workers under varying poses and changing background in image sequences via very deep residual networks. Automat Constr 99: 27‒38. https://doi.org/10.1016/j.autcon.2018.11.033 doi: 10.1016/j.autcon.2018.11.033
    [11] Wang X, Niu D, Luo P, Zhu C, Ding L, Huang K (2020) A safety helmet and protective clothing detection method based on improved-yolo v 3. 2020 Chinese automation congress (CAC), 5437‒5441. IEEE. https://doi.org/10.1109/CAC51589.2020.9327187
    [12] Benyang D, Xiaochun L, Miao Y (2020) Safety helmet detection method based on YOLO v4. 2020 16th International conference on computational intelligence and security (CIS), 155‒158. IEEE. https://doi.org/10.1109/CIS52066.2020.00041
    [13] Tan S, Lu G, Jiang Z, Huang L (2021) Improved YOLOv5 network model and application in safety helmet detection. 2021 IEEE International Conference on Intelligence and Safety for Robotics (ISR), 330‒333. IEEE. https://doi.org/10.1109/ISR50024.2021.9419561
    [14] Fu Y, Ran T, Xiao W, Yuan L, Zhao J, He L, et al. (2024) GD-YOLO: An improved convolutional neural network architecture for real-time detection of smoking and phone use behaviors. Digital Signal Processing 151: 104554. https://doi.org/10.1016/j.dsp.2024.104554 doi: 10.1016/j.dsp.2024.104554
    [15] Li H, Wu D, Zhang W, Xiao C (2024) YOLO-PL: Helmet wearing detection algorithm based on improved YOLOv4. Digital Signal Processing 144: 104283. https://doi.org/10.1016/j.dsp.2023.104283 doi: 10.1016/j.dsp.2023.104283
    [16] Nath ND, Behzadan AH, Paal SG (2020) Deep learning for site safety: Real-time detection of personal protective equipment. Automat Constr 112: 103085. https://doi.org/10.1016/j.autcon.2020.103085 doi: 10.1016/j.autcon.2020.103085
    [17] Li L, Zhang P, Yang S, Jiao W (2023) YOLOv5-SFE: An algorithm fusing spatio-temporal features for detecting and recognizing workers' operating behaviors. Adv Eng Inform 56: 101988. https://doi.org/10.1016/j.aei.2023.101988 doi: 10.1016/j.aei.2023.101988
    [18] Redmon J, Farhadi A (2017) YOLO9000: better, faster, stronger. Proceedings of the IEEE conference on computer vision and pattern recognition, 7263‒7271. https://doi.org/10.1109/CVPR.2017.690
    [19] Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv: 1804.02767.
    [20] Bochkovskiy A, Wang CY, Liao HYM (2020) YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv preprint arXiv: 2004.10934.
    [21] Ge Z, Liu S, Wang F, Li Z, Sun J (2021) YOLOX: Exceeding YOLO Series in 2021. arXiv preprint arXiv: 2107.08430.
    [22] Wang CY, Bochkovskiy A, Liao HYM (2022) YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 7464‒7475. https://doi.org/10.1109/CVPR52729.2023.00721
    [23] Huang H, Chen Z, Zou Y, Lu M, Chen C, Song Y, et al. (2024) Channel prior convolutional attention for medical image segmentation. Comput Biol Med 178: 108784.
    [24] Zhang X, Liu C, Yang D, Song T, Ye Y, Li K, et al. (2023) Rfaconv: Innovating spatital attention and standard convolutional operation. arXiv preprint arXiv: 2304.03198.
    [25] He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE T Pattern Anal 37: 1904‒1916. https://doi.org/10.1109/TPAMI.2015.2389824 doi: 10.1109/TPAMI.2015.2389824
    [26] Zheng Z, Wang P, Liu W, Li J, Ye R, Ren D (2020) Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. Proceedings of the AAAI conference on artificial intelligence, 34: 12993‒13000. https://doi.org/10.1609/aaai.v34i07.6999 doi: 10.1609/aaai.v34i07.6999
    [27] Gevorgyan Z (2022) SIoU loss: More powerful learning for bounding box regression. arXiv preprint arXiv: 2205.12740.
    [28] Rezatofighi H, Tsoi N, Gwak J, Sadeghian A, Reid I, Savarese S (2019) Generalized intersection over union: A metric and a loss for bounding box regression. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 658‒666. https://doi.org/10.1109/CVPR.2019.00075
    [29] Zhang YF, Ren W, Zhang Z, Jia Z, Wang L, Tan T (2022) Focal and Efficient IOU Loss for Accurate Bounding Box Regression. Neurocomputing 506: 146‒157. https://doi.org/10.1016/j.neucom.2022.07.042 doi: 10.1016/j.neucom.2022.07.042
    [30] Zhang H, Xu C, Zhang S (2023) Inner-IoU: more effective intersection over union loss with auxiliary bounding box. arXiv preprint arXiv: 2311.02877.
    [31] Zhang H, Zhang S (2023) Shape-IoU: More Accurate Metric considering Bounding Box Shape and Scale. arXiv preprint arXiv: 2312.17663.
    [32] Siliang M, Yong X (2023) Mpdiou: a loss for efficient and accurate bounding box regression. arXiv preprint arXiv: 2307.07662.
    [33] Wang J, Xu C, Yang W, Yu L (2021) A normalized Gaussian Wasserstein distance for tiny object detection. arXiv preprint arXiv: 2110.13389.
    [34] Tong Z, Chen Y, Xu Z, Yu R (2023) Wise-IoU: bounding box regression loss with dynamic focusing mechanism. arXiv preprint arXiv: 2301.10051.
    [35] Liu C, Wang K, Li Q, Zhao F, Zhao K, Ma H (2024) Powerful-IoU: More straightforward and faster bounding box regression loss with a nonmonotonic focusing mechanism. Neural Networks 170: 276‒284. https://doi.org/10.1016/j.neunet.2023.11.041 doi: 10.1016/j.neunet.2023.11.041
    [36] Misra D, Nalamada T, Arasanipalai AU, Hou Q (2021) Rotate to Attend: Convolutional Triplet Attention Module. Proceedings of the IEEE/CVF winter conference on applications of computer vision, 3139‒3148. https://doi.org/10.1109/WACV48630.2021.00318
    [37] Woo S, Park J, Lee JY, Kweon IS (2018) Cbam: Convolutional block attention module. Proceedings of the European conference on computer vision (ECCV), 3‒19. https://doi.org/10.1007/978-3-030-01234-2_1
    [38] Yang L, Zhang RY, Li L, Xie X (2021) SimAM: A Simple, Parameter-Free Attention Module for Convolutional Neural Networks. International Conference on Machine Learning, 11863‒11874. PMLR.
    [39] Liu H, Liu F, Fan X, Huang D (2022) Polarized Self-Attention: Towards High-quality Pixel-wise Regression. Neurocomputing 506: 158‒167. https://doi.org/10.1016/j.neucom.2022.07.054 doi: 10.1016/j.neucom.2022.07.054
    [40] Zhu L, Wang X, Ke Z, Zhang W, Lau RW (2023) Biformer: Vision transformer with bi-level routing attention. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 10323‒10333. https://doi.org/10.1109/CVPR52729.2023.00995
    [41] Li X, Hu X, Yang J (2019) Spatial group-wise enhance: Improving semantic feature learning in convolutional networks. arXiv preprint arXiv: 1905.09646.
    [42] Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. (2017) Attention Is All You Need. Advances in neural information processing systems 30.
    [43] Yang J, Li C, Dai X, Gao J (2022) Focal modulation networks. Advances in Neural Information Processing Systems 35: 4203‒4217.
    [44] Wang H, Guo P, Zhou P, Xie L (2024) MLCA-AVSR: Multi-Layer Cross Attention Fusion based Audio-Visual Speech Recognition. ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 8150‒8154. https://doi.org/10.1109/ICASSP48485.2024.10446769
    [45] Li Y, Hou Q, Zheng Z, Cheng MM, Yang J, Li X (2023) Large selective kernel network for remote sensing object detection. Proceedings of the IEEE/CVF International Conference on Computer Vision, 16794‒16805. https://doi.org/10.1109/ICCV51070.2023.01540
    [46] Azad R, Niggemeier L, Hüttemann M, Kazerouni A, Aghdam EK, Velichko Y, et al. (2024) Beyond self-attention: Deformable large kernel attention for medical image segmentation. Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 1287‒1297. https://doi.org/10.1109/WACV57701.2024.00132
    [47] Li X, Wang W, Hu X, Yang J (2019) Selective kernel networks. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 510‒519. https://doi.org/10.1109/CVPR.2019.00060
    [48] Hu J, Shen L, Sun G (2018) Squeeze-and-Excitation Networks. 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 7132‒7141. https://doi.org/10.1109/CVPR.2018.00745
    [49] Goyal A, Bochkovskiy A, Deng J, Koltun V (2022) Non-deep Networks. Advances in neural information processing systems 35: 6789‒6801.
    [50] Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) ECA-Net: Efficient channel attention for deep convolutional neural networks. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 11534‒11542. https://doi.org/10.1109/CVPR42600.2020.01155
    [51] Chen Y, Kalantidis Y, Li J, Yan S, Feng J (2018) A.2-nets: Double attention networks. Advances in neural information processing systems 31.
    [52] Li Y, Yao T, Pan Y, Mei T (2022) Contextual Transformer Networks for Visual Recognition. IEEE T Pattern Anal 45: 1489‒1500.
    [53] Lee Y, Park J (2020) Centermask: Real-time anchor-free instance segmentation. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 13906‒13915. https://doi.org/10.1109/CVPR42600.2020.01392
    [54] Xia Z, Pan X, Song S, Li LE, Huang G (2022) Vision transformer with deformable attention. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 4794‒4803. https://doi.org/10.1109/CVPR52688.2022.00475
    [55] Ouyang D, He S, Zhang G, Luo M, Guo H, Zhan J, et al. (2023) Efficient multi-scale attention module with cross-spatial learning. ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1‒5. IEEE. https://doi.org/10.1109/ICASSP49357.2023.10096516
    [56] Hu L, Li Y (2021) Micro-YOLO: Exploring Efficient Methods to Compress CNN based Object Detection Model. In ICAART (2), 151‒158. https://doi.org/10.5220/0010234401510158
    [57] Li C, Li L, Jiang H, Weng K, Geng Y, Li L, et al. (2022) YOLOv6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv: 2209.02976.
    [58] Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-End Object Detection with Transformers. European conference on computer vision, 213‒229. Cham: Springer International Publishing. https://doi.org/10.1007/978-3-030-58452-8_13
    [59] Duan K, Bai S, Xie L, Qi H, Huang Q, Tian Q (2019) Centernet: Keypoint triplets for object detection. Proceedings of the IEEE/CVF international conference on computer vision, 6569‒6578. https://doi.org/10.1109/ICCV.2019.00667
  • Reader Comments
  • © 2025 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1418) PDF downloads(174) Cited by(0)

Article outline

Figures and Tables

Figures(9)  /  Tables(7)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog