Research article Special Issues

Pseudospectral method for fourth-order fractional Sturm-Liouville problems

  • Fourth-order fractional Sturm-Liouville problems are studied in this work. The numerical simulation uses the pseudospectral method, utilizing Chebyshev cardinal polynomials. The presented algorithm is implemented after converting the desired equation into an associated integral equation and gives us a linear system of algebraic equations. Then, we can find the eigenvalues by calculating the roots of the corresponding characteristic polynomial. What is most striking is that the proposed scheme accurately solves this type of equation. Numerical experiments confirm this claim.

    Citation: Haifa Bin Jebreen, Beatriz Hernández-Jiménez. Pseudospectral method for fourth-order fractional Sturm-Liouville problems[J]. AIMS Mathematics, 2024, 9(9): 26077-26091. doi: 10.3934/math.20241274

    Related Papers:

    [1] Eminugroho Ratna Sari, Fajar Adi-Kusumo, Lina Aryati . Mathematical analysis of a SIPC age-structured model of cervical cancer. Mathematical Biosciences and Engineering, 2022, 19(6): 6013-6039. doi: 10.3934/mbe.2022281
    [2] Tufail Malik, Jody Reimer, Abba Gumel, Elamin H. Elbasha, Salaheddin Mahmud . The impact of an imperfect vaccine and pap cytologyscreening on the transmission of human papillomavirus and occurrenceof associated cervical dysplasia and cancer. Mathematical Biosciences and Engineering, 2013, 10(4): 1173-1205. doi: 10.3934/mbe.2013.10.1173
    [3] Ilse Domínguez-Alemán, Itzel Domínguez-Alemán, Juan Carlos Hernández-Gómez, Francisco J. Ariza-Hernández . A predator-prey fractional model with disease in the prey species. Mathematical Biosciences and Engineering, 2024, 21(3): 3713-3741. doi: 10.3934/mbe.2024164
    [4] Adnan Sami, Amir Ali, Ramsha Shafqat, Nuttapol Pakkaranang, Mati ur Rahmamn . Analysis of food chain mathematical model under fractal fractional Caputo derivative. Mathematical Biosciences and Engineering, 2023, 20(2): 2094-2109. doi: 10.3934/mbe.2023097
    [5] Simphiwe M. Simelane, Phumlani G. Dlamini, Fadekemi J. Osaye, George Obaido, Blessing Ogbukiri, Kehinde Aruleba, Cadavious M. Jones, Chidozie W. Chukwu, Oluwaseun F. Egbelowo . Modeling the impact of public health education on tungiasis dynamics with saturated treatment: Insight through the Caputo fractional derivative. Mathematical Biosciences and Engineering, 2023, 20(5): 7696-7720. doi: 10.3934/mbe.2023332
    [6] Jutarat Kongson, Chatthai Thaiprayoon, Apichat Neamvonk, Jehad Alzabut, Weerawat Sudsutad . Investigation of fractal-fractional HIV infection by evaluating the drug therapy effect in the Atangana-Baleanu sense. Mathematical Biosciences and Engineering, 2022, 19(11): 10762-10808. doi: 10.3934/mbe.2022504
    [7] Noura Laksaci, Ahmed Boudaoui, Seham Mahyoub Al-Mekhlafi, Abdon Atangana . Mathematical analysis and numerical simulation for fractal-fractional cancer model. Mathematical Biosciences and Engineering, 2023, 20(10): 18083-18103. doi: 10.3934/mbe.2023803
    [8] Ritu Agarwal, Pooja Airan, Mohammad Sajid . Numerical and graphical simulation of the non-linear fractional dynamical system of bone mineralization. Mathematical Biosciences and Engineering, 2024, 21(4): 5138-5163. doi: 10.3934/mbe.2024227
    [9] Najat Ziyadi . A male-female mathematical model of human papillomavirus (HPV) in African American population. Mathematical Biosciences and Engineering, 2017, 14(1): 339-358. doi: 10.3934/mbe.2017022
    [10] Hardik Joshi, Brajesh Kumar Jha, Mehmet Yavuz . Modelling and analysis of fractional-order vaccination model for control of COVID-19 outbreak using real data. Mathematical Biosciences and Engineering, 2023, 20(1): 213-240. doi: 10.3934/mbe.2023010
  • Fourth-order fractional Sturm-Liouville problems are studied in this work. The numerical simulation uses the pseudospectral method, utilizing Chebyshev cardinal polynomials. The presented algorithm is implemented after converting the desired equation into an associated integral equation and gives us a linear system of algebraic equations. Then, we can find the eigenvalues by calculating the roots of the corresponding characteristic polynomial. What is most striking is that the proposed scheme accurately solves this type of equation. Numerical experiments confirm this claim.



    The quality of tobacco products is an essential indicator of tobacco enterprises' production level and brand establishment. Cigarettes inevitably have appearance defects on the production lines, such as stains, scratches and wrinkles. There is an urgent need for tobacco companies to adopt strict quality control measures to prevent cosmetically defective products from entering the sales market. In the past, manual inspection was the main method for cigarette factories to detect appearance defects. However, this method has problems such as strong subjectivity, slow detection speed, low efficiency and high missed detection rate, and it cannot adapt to the high-speed and high-precision detection requirements of modern industry. Therefore, there is an urgent need to develop an efficient, accurate and cost-effective automated detection method to replace manual detection.

    Product appearance defect detection based on conventional computer vision algorithms often relies on template matching to identify defect locations, entailing complications in template design and yielding suboptimal accuracy. With the advancements in deep learning technology, deep learning methods have also been harnessed for product appearance defect detection, resulting in enhanced detection accuracy.

    Product appearance defect detection represents a specific application of object detection models. Based on deep learning, object detection models can be broadly categorized into two types: two-stage detectors and one-stage detectors. Two-stage detectors, exemplified by region-based convolutional neural networks (RCNN) [1] and Faster-RCNN [2,3], and one-stage detectors, represented by You Only Look Once (YOLO) [4,5,6,7] and Single Shot MultiBox Detector (SSD) [8], have been employed in product appearance defect detection. For instance, Liu et al. [9] utilized an improved YOLOv5 model to detect surface defects on diode metal bases, incorporating attention mechanisms and the K-means++ clustering algorithm, achieving mAP@0.5 84.0%. Chen et al. [10] applied an improved YOLOv3 model to detect surface defects on surface-mounted device (SMD) LED chips, replacing the backbone network with DenseNet and achieving mAP@0.5 95.28%. Hu and Wang [11] employed an improved Faster R-CNN model to detect surface defects on printed circuit boards (PCB), replacing the model's backbone network and feature fusion pyramid, achieving mAP@0.5 94.2%. Li and Xu [12] used an improved SSD model for appearance defect detection on electronic products, replacing the backbone network with MobileNet and optimizing parameters to achieve mAP@0.5 88.6%. Duan et al. [13] utilized an improved YOLOv3 model for surface defect detection in casting, introducing dual-density convolutions and low-resolution detection layers to achieve mAP@0.5 88.02%. Kou et al. [14] employed an improved YOLOv3 model for surface defect detection on steel strips, modifying the detection head to Anchor-Free and achieving mAP@0.5 71.3% on the GC10-DET dataset.

    In defect detection research for cigarette appearance, Yuan et al. [15] employed an improved ResNeSt model, achieving a classification accuracy of 94.01% through transfer learning, multi-scale learning methods and modified activation functions. Qu et al. [16] utilized an improved SSD model for cigarette appearance defect detection, achieving mAP@0.5 90.26% by improving the pyramid convolution. However, the detection speed still fell short of 100 frames per second (FPS). Liu and Yuan [17] employed an improved YOLOv5s model for cigarette appearance defect detection, achieving mAP@0.5 94.0% by introducing attention mechanisms and modifications to the loss function. Nonetheless, the recall rate reached only 86.8%. Yuan et al. [18] employed an improved YOLOv4 model for cigarette appearance defect detection, achieving mAP@0.5 91.46% through introducing attention mechanisms, dilated pyramid spatial pooling and modifications to anchor clustering algorithms and loss functions. However, the recall rate remained at 88.81%. Liu et al. [19] utilized an improved CenterNet model for cigarette appearance defect detection, achieving mAP@0.5 95.01% by introducing attention mechanisms, deformable convolutions, feature fusion pyramids and modifications to activation functions and data augmentation methods. Nonetheless, the recall rate reached only 85.96%.

    Although there has been considerable research on industrial appearance defect detection using deep learning, studies specifically addressing cigarette appearance defect detection remain limited. Existing research on cigarette appearance defect detection still suffers from low detection recall rates and slow detection speeds. With the advancement of automation in tobacco production, cigarette production lines can now achieve speeds of up to 200 cigarettes per second, posing challenges for real-time detection using existing methods. To address this practical demand, we propose a real-time defect detection method for cigarette appearance based on YOLOv5n. Building upon the YOLOv5 model, our approach incorporates the C2F convolution module, Jump Concat fusion pyramid and SIoU loss function. Experimental results demonstrate that our improved model enhances the precision and recall rate of cigarette appearance defect detection, effectively reducing the presence of defective cigarettes in the consumer market. This model provides robust support for tobacco companies in improving the quality of their cigarette products.

    The cigarette images used in our experiments are from the Yunnan Branch of China Tobacco Industry Company Limited. The high-speed industrial cameras capture the images on the automated production line. The front and back cigarettes can be captured at different production line positions. A standard cigarette is 84 mm in length and 7.8 mm in diameter.

    Cigarettes are primarily composed of two parts: the cigarette rod, depicted as the white portion on the left side in Figure 1, and the filter, represented by the dark portion on the right side. During cigarette manufacturing, excessive tobacco stems can puncture cigarette rod packaging during rolling [20]. During the twisting process, a certain amount of pressure is applied to ensure a secure bond between the tipping paper and the cigarette rod. However, variations in filter elasticity, roundness of non-filter cigarettes, adhesive properties of latex and the absorptive characteristics of the tipping paper can result in insufficient bonding, misalignment, filter detachment, as well as folding, creasing and misalignment of the tipping paper [21]. Furthermore, improper printing on the cigarette packaging paper or staining during production and transportation can result in visual contamination of the cigarette's appearance.

    Figure 1.  Normal cigarette examples.

    To improve the detection of cigarette appearance defects, we categorized defective cigarettes into seven types: misplacement, stain, scratched, wrinkle, unpacked, no filter and bend, as shown in Figure 2.

    Figure 2.  Cigarette examples with appearance defects.

    YOLOv5 comprises five network variants: YOLOv5n, YOLOv5s, YOLOv5m, YOLOv5l and YOLOv5x. These models share the same network architecture but differ in module depth and width. YOLOv5n and YOLOv5s have the same depth, with YOLOv5n being the shallowest among the YOLOv5 series. To improve the network's runtime on small-scale devices, YOLOv5n reduces the network's width by half compared to YOLOv5s, significantly reducing the scale of parallel computations and enhancing runtime performance on low-capacity devices [22]. To meet the real-time detection requirements and improve recognition speed, we select YOLOv5n as the baseline model.

    YOLOv5 comprises four components: input, backbone network, neck network and detection head. The input component incorporates mosaic data augmentation, adaptive anchor calculation and adaptive image scaling. The backbone network performs feature extraction. The neck network employs a top-down and bottom-up feature fusion pathway to merge image features. The detection head includes three layers corresponding to feature maps of sizes 80 × 80, 40 × 40 and 20 × 20 for detecting objects of different scales. Finally, the Complete IoU (CIoU [23]) loss function is utilized to calculate the distance between predicted and ground truth bounding boxes and non-maximum suppression (NMS) is applied to remove redundant boxes while retaining those with the highest confidence.

    Due to the smaller depth and width of YOLOv5n, its ability to fit complex features is relatively limited. Therefore, this research introduces the C2F module proposed in the state-of-the-art object detection model YOLOv8 [24]. Compared to the original C3 (CSP Bottleneck with 3 convolutions) module [25], the C2F module incorporates bottleneck skip connections, allowing for better feature extraction by increasing the gradient flow and channel capacity. Additionally, only half of the feature matrix participates in subsequent multiple bottleneck operations by splitting the channel number equally after one convolution. This design enhances feature extraction capabilities without compromising GPU inference speed, thus avoiding false detections and omissions.

    The original YOLOv5's feature fusion pyramid involves multiple fusion operations on the P4 feature layer, which can lead to occlusion and coverage of subtle texture folds in cigarettes, increasing the difficulty of detection. Therefore, this research uses Jump Concat to the P4 feature layer to mitigate the occlusion of fine-textured crease defects resulting from feature map fusion.

    The original YOLOv5 utilizes the CIoU localization loss function, which calculates the localization loss based on IoU (Intersection over Union), center point distance and aspect ratio between predicted and ground truth bounding boxes. However, it does not account for mismatched orientations between the predicted and ground truth boxes. This limitation slows down convergence and efficiency, as predicted boxes may "hover" during training, ultimately leading to suboptimal model performance. Therefore, SIoU [26] is introduced, which considers the angle between the center point vectors and includes angle penalty metrics. This loss function enables predicted boxes to converge rapidly towards the nearest axis and converge only on a single axis. This loss function effectively accelerates the convergence of predicted boxes, improving the model's localization accuracy and confidence in object detection.

    Our improved model structure is shown in Figure 3.

    Figure 3.  CJS-YOLOv5n network structure.

    The C2F module, proposed in the latest object detection model YOLOv8, enhances feature extraction capabilities compared to the original C3 module by introducing bottleneck skip connections and chunk operations, resulting in increased parameter quantity, gradient flow and channel capacity. In Figure 4(a), the C2F module is depicted, while Figure 4(b) represents the original C3 module. The C2F module possesses more channels and gradient flow than the C3 module. The chunk operation performs slicing on the input feature map, dividing it equally into two parts along the channel dimension. Only half of the feature map participates in the bottleneck convolution operation, effectively reducing computational load. Additionally, to maintain a lightweight structure, the C2F module reduces the computation of the right branch CBS (Convolution, Batch normalization and SiLU activation function) module.

    Figure 4.  Comparison of convolutional module structures.

    The YOLOv5n model incorporates both FPN (Feature Pyramid Network) [27] and PANet (Path Aggregation Network) [28] for multi-scale feature fusion. FPN enhances semantic information in a top-down manner, while PANet reinforces positional information in a bottom-up way. This combination enhances the feature fusion capability of the neck layer. However, in the P4 feature layer, multiple feature fusion operations may cause certain feature information to be overshadowed. To address this issue, this study introduces an additional Jump Concat at the end of the P4 feature layer to prevent the coverage of subtle textures with minimal pixel variations, such as small wrinkles in cigarette appearances.

    Figure 5(a) illustrates the original bi-directional feature fusion pyramid of YOLOv5, while Figure 5(b) depicts the enhanced bi-directional feature fusion pyramid with the additional skip connection.

    Figure 5.  Comparison of feature pyramid.

    For P4, the two fusion feature processes are formed as follows:

    P4out=Conv(Resize(P3out)+P4in+Conv(P4in+Resize(Conv(P5in)))) (1)

    where, Resize(·) is usually an upsampling or downsampling operation for resolution matching, and Conv(·) is usually a convolutional operation for feature processing. Piout is the output of the i-th feature map and Piin is the input of the i-th feature map.

    The SIoU loss function is employed for bounding box regression. Compared to CIoU, DIoU (Distance IoU) [29] and GIoU (General IoU) [30], SIoU further considers the vector angle between the ground truth and predicted anchor boxes, leading to a redefinition of the loss function. It facilitates faster regression of anchor boxes towards the nearest axis (x or y), thereby enhancing convergence speed and localization accuracy.

    In Figure 6, BB represents the predicted anchor box, GT denotes the ground truth box, σ represents the Euclidean distance between the centers of the predicted and ground truth boxes, Ch represents the projection distance of the center point on the y-axis, Cw represents the projection distance of the center point on the x-axis, α represents the angle between the center line and the x-axis and β represents the angle between the center line and the y-axis.

    Figure 6.  Principle of SIoU loss function.

    The SIoU loss function involves three components in its computation: angle loss, distance loss and intersection over union loss:

    IoU=|BBGT||BBGT| (2)

    The above equation represents the calculation of the intersection over union, where BB denotes the area of the predicted bounding box and GT represents the area of the ground truth box. The intersection ratio to the predicted and ground truth box union yields the IoU value.

    Δ=t=x,y1eγρt (3)

    The formula above calculates the angle, Δ, which comprises the angle and distance calculations involved in SIoU. It represents the final result of the angle calculation. x and y represent the sine values of the angles between the center points of the ground truth and predicted boxes. At the same time, ρ is the ratio of the distance between the center points of the ground truth and predicted boxes relative to the width and height of their minimum enclosing rectangle squared. Here, e denotes the Euler's number.

    Ω=t=w,h(1eωt)θ (4)

    The above equation computes the Ω shape loss, where w and h represent the width and height of the predicted box, respectively, and θ denotes the attention coefficient in the shape calculation formula. Finally, the SIoU loss consists of the three components above, as depicted by the following formula:

    Lbox=1IoU+Ω+Δ2 (5)

    On high-speed cigarette production lines, the probability of encountering cigarettes with visual defects is approximately 1%. The manual screening process to obtain a substantial amount of defective cigarette data is labor-intensive. We gathered 900 valid images of appearance defect cigarettes through meticulous manual selection. Due to the limited availability of the appearance defect dataset, we employed data augmentation techniques, including flipping, brightness adjustment and Gaussian noise addition, to expand the dataset. Consequently, the augmented dataset now comprises 6200 images. It is important to note that since the images were captured in pairs, the actual research dataset consists of 12,400 cigarette images. To ensure the model's effectiveness, we divide the dataset into training, validation and testing sets in a 6:2:2 ratio, as detailed in Table 1.

    Table 1.  Dataset partition.
    Defect type Training set Validation set Testing set
    Misplacement 480 160 160
    Stain 840 280 280
    Scratched 1680 560 560
    Wrinkle 1800 600 600
    Unpacked 1476 492 492
    No filter 876 292 292
    Bend 420 140 140
    Total 7572 2524 2524

     | Show Table
    DownLoad: CSV

    The model was trained and tested on a Windows 10 system running PyTorch 1.12.1, using the following hardware specifications: an AMD R5600 processor @ 3.50 GHz, 32 GB of memory and an NVIDIA GeForce RTX3060 graphics card with 12 GB of VRAM. The software was CUDA 11.6, Torch vision 0.13.1 and Python 3.7. The integrated development environments are PyCharm and Anaconda.

    The initial learning rate was 0.01 during training, and a cosine annealing strategy was employed to reduce the learning rate. Additionally, the neural network parameters were optimized using the stochastic gradient descent (SGD) method with a momentum value of 0.937 and a weight decay score of 0.0005. The training process consisted of 300 epochs, with a batch size of 64 images. The input image resolution was uniformly adjusted to 640 × 640. The adjusted training parameters are summarized in Table 2.

    Table 2.  Training parameter.
    Parameter Value
    Image size 640 × 640
    Batch size 64
    Epoch 300
    Learning rate 0.01
    Optimizer SGD
    Weight decay 0.0005

     | Show Table
    DownLoad: CSV

    The experimental evaluation encompasses two aspects: performance evaluation and complexity evaluation. Performance evaluation metrics for the model include accuracy, recall, mAP@0.5 and mAP@0.5-0.95. Complexity evaluation metrics for the model consist of the size of model parameters, floating point operations (FLOPs) and frames per second (FPS), which assess the computational efficiency and image processing speed of the model.

    Precision measures the proportion of correctly predicted positive samples out of the total positive samples, assessing the model's classification ability. Conversely, recall measures the proportion of correctly predicted positive samples out of the whole of positive samples. AP is the integral of precision and recall, and mAP represents the average AP, reflecting the model's overall performance for object detection and classification. The calculation formulas for these metrics are shown in Eqs (6)–(9).

    Precision=TPTP+FP (6)
    Recall=TPTP+FN (7)

    where, TP represents the number of true positive samples correctly detected, FP represents the number of false positive samples incorrectly detected, and FN represents the number of false negative samples incorrectly not detected.

    AP=ò10P(R)dR (8)
    mAP=åni=1APin (9)

    where, n is the number of categories.

    FLOPs(Conv)=(2×Cin×K21)×Wout×Hout×Cout (10)
    FLOPs(Liner)=(2?Cin1)?Cout (11)

    Model size refers to the amount of memory required to store the model. FLOPs, on the other hand, measure the complexity of the model by quantifying the total number of multiplication and addition operations performed during model execution. A lower FLOPs value indicates lower computational requirements for model inference, resulting in faster model computation speed. Here, Cin represents the number of input channels, Cout represents the number of output channels, K represents the convolutional kernel size and Wout and Hout represent the width and height of the output feature map, respectively.

    Figure 8 presents the training loss curves for the CJS-YOLOv5n method and YOLOv5n on the cigarette appearance defect dataset. The graph illustrates the overall loss values during the training process. In the initial 30 epochs, the model experiences a rapid decline in loss. Subsequently, from epoch 50 to 200, a gradual decrease in loss is observed. Between epochs 200 and 300, the loss values stabilize and approach convergence. Thus, 300 epochs are determined as the appropriate training iteration count for the model.

    Figure 7.  Training loss curve.
    Figure 8.  YOLOv5n detection result.

    Furthermore, the dashed line represents the loss curve of the CJS-YOLOv5n, while the solid line represents the loss curve of YOLOv5n. It can be observed from the graph that, under the same conditions of rapid loss reduction and convergence, the improved model exhibits a lower final convergence value compared to the original model. Additionally, our improved model shows more minor fluctuations in the loss curve during training. Consequently, our improved model demonstrates superior performance during the training process.

    This section verifies the effectiveness of different optimization modules through ablation experiments. Several improved models are constructed sequentially, adding the C2F module, Jump Concat and SIoU localization loss function to the baseline model YOLOv5n. The results are compared using the same test data, and the gains in model performance due to the added modules are presented in Table 3.

    Table 3.  Ablation experiment result.
    C2F Jump Concat SIoU Precision
    (%)
    Recall
    (%)
    mAP@0.5 (%) mAP@0.5:0.95 (%) FLOPs
    (G)
    FPS
    (GPU)
    FPS
    (CPU)
    94.8 93 94.2 56 4.2 556 43
    95.1 93.6 94.8 56.4 5.6 556 40
    94.9 94 95.2 56.5 4.2 556 43
    94.2 94.1 95.2 56.9 4.2 556 43
    95.5 93.9 95.7 57.7 5.6 556 40
    95.5 94.2 95.1 57.2 5.6 556 40
    94.7 95 95.5 56.8 4.2 556 43
    95.3 95.3 95.9 57.9 5.6 556 40

     | Show Table
    DownLoad: CSV

    Table 3 shows the gain of model performance after adding each module. Due to the small depth and width of the original YOLOv5n network, the ability to fit complex features is poor. In order to enhance the feature extraction ability of the network, this paper replaces the C3 convolution module with the C2F convolution module, which increases mAP@0.5 by 0.6%. The rate increased by 0.6%. Later, in order to prevent the feature fusion from covering up the fine texture features, this paper adds a Jump Concat to the P4 feature layer, which increases mAP@0.5 by 1% and the recall rate by 1%. Finally, in order to enhance the positioning accuracy and avoid false detection, this paper replaces the CIoU loss function with the SIoU loss function, which increases mAP@0.5 by 1% and the recall rate by 1.1%. After combining the three modules, the improved model achieves the best performance. The increased computational load of the improved module in this paper does not affect the performance of the model running on the GPU, but the increased parallel computing of the C2F convolution module makes the running speed of the model on the CPU drop by nearly 10%.

    The proposed method in this study was trained using the same parameters as other advanced lightweight methods. The experimental results are compared in Table 4.

    Table 4.  Comparison of different detection models.
    Model Precision
    (%)
    Recall
    (%)
    mAP@0.5
    (%)
    mAP@0.5:0.95
    (%)
    Parameters
    (M)
    FPS
    YOLOv3tiny 82.3 93.7 92.9 54.3 8.6 500
    YOLOv4tiny 88.4 93.2 92.8 54.0 5.9 500
    YOLOv5n 94.8 93.0 94.2 56.0 1.7 556
    YOLOv7tiny 91.5 87.0 91.6 51.3 6.2 278
    YOLOv8n 95.7 93.3 95.4 57.2 3.0 286
    SSD-improved [16] 87.4 88.7 90.3 50.1 26 84
    YOLOv4-improved [18] 95.2 92.8 94.8 55.7 66 161
    CJS-YOLOv5n 95.3 95.3 95.9 57.9 2.3 556

     | Show Table
    DownLoad: CSV

    Table 4 shows the performance comparison with other target detection models. The model selected in this paper is not optimal in every performance indicator, but it is faster than the better performance models like YOLOv8n, and detection accuracy is higher than models like YOLOv3tiny that are close in speed. Since the number of model parameters selected in this paper is the least among all models, it is beneficial to reduce the deployment cost of practical applications and achieve a balance between detection speed and accuracy. While maintaining the detection speed, the improved CJS-YOLOv5n in this paper has improved mAP@0.5 by 1.7% and the recall rate by 2.3% compared with the original model. YOLOv8 is a state-of-the-art object detection model with high accuracy and speed in many applications. Compared with the most advanced YOLOv8n in average detection accuracy, recall rate and detection speed achieved an all-around lead, which can meet most of the current detection needs. The experimental results show that the model proposed in this paper is a good detection method for cigarette appearance defects, which can meet the needs of cigarette appearance defect detection on the production line.

    Our ablation experiments and comparative experiments demonstrate the effectiveness of our improvements. Among them, the C2F module achieves better feature extraction at the cost of a small amount of calculation, and various performance indicators have different improvements. Jump Concat reduces the degree of occlusion of subtle features in feature fusion, effectively improving the recall rate. The SIoU loss function improves the localization accuracy of anchor boxes by optimizing the anchor box loss calculation, thus increasing the confidence and recall of predicted anchor boxes. Comparing the latest research results of YOLOv4-improved in detecting cigarette appearance defects and the current state-of-the-art object detection model YOLOv8, the improved model CJS-YOLOv5n has different degrees of increase in detection performance and detection speed.

    Figure 9(a)(h) shows the detection results of YOLO-v5n for seven kinds of defects, and Figure 10(a)–(h) shows the detection results of CJS-YOLOv5n for seven kinds of defects. The detection results prove the improvement of this paper. Compared with the original model, the latter model has better performance in seven kinds of defect detection. Comparing Figure 9 and (b), (d) and (h) in Figure 10, it can be found that the improved model can better identify the subtle wrinkle and stain defects that were not detected by the original network, and the positioning anchor box of the fold defect is more complete. In complex scenarios with multiple defects, the improved model also avoids the false detection of the original model. Comparing the detection results of other defects, it can be seen that the improved network has different confidence improvements compared with the original network in the detection of different defects.

    Figure 9.  CJS-YOLOv5n detection result.

    According to the fact that the actual production speed of cigarette appearance defects is fast, the improved YOLOv5n network is used to detect cigarette appearance defects, and good detection results and detection speed can be achieved in the case of insufficient data sets. In this paper, the C2F module with stronger feature extraction ability than the C3 convolution module is selected as the network basic convolution module, and the P4 feature layer Jump Concat is added to further strengthen the feature fusion, prevent information coverage and use the SIoU positioning loss function to help the model improve. The positioning accuracy and convergence speed in the experimental results prove the effectiveness of the algorithm in the task of cigarette appearance defect detection. This method can achieve better detection speed and recall rate, and has certain robustness. The model can effectively improve the recall rate and detection accuracy without affecting the detection speed.

    Our model in this paper still has shortcomings. For example, in terms of detection effect, it is still not as good as YOLOv8n with a larger network. Although the improvement in this paper can effectively increase the recall rate and help control the quality of cigarette sales, it is limited by the depth and width of the network. The ability to fit complex defects still needs to be improved. In the future, we will focus on how to improve the accuracy of network detection while maintaining the detection speed and reducing the deployment cost, such as adding a lightweight attention mechanism CBAM (Convolutional block attention module) [31], a detection head based on Anchor-Free [32], further expand the dataset, and use lightweight convolution. In addition, we will use higher-order moment time series analysis in the Caputo sense [33] and fractional differential equations [34] to improve post-processing. We hope we can achieve better performance in cigarette appearance defect detection.

    The authors declare they have not used Artificial Intelligence (AI) tools in the creation of this article.

    Thank you to Yunnan Branch of China Tobacco Industry Company Limited for providing the cigarette defect dataset.

    This research was funded by the Yunnan Provincial Department of Science and Technology-Yunnan University Joint Special Project for Double-Class Construction (Grant No. 202201BF070001-005), the Natural Science Foundation of China (Grant No. 62061049, 12263008), the Key R & D Projects of Yunnan Province (Grant No. 202202AD080004), the Application and Foundation Project of the Yunnan Province (Grant No. 202001BB050032) and the Postgraduate Practice and Innovation Project of Yunnan University (Grant no. 22221382).

    The authors declare there is no conflict of interest.



    [1] A. Afarideh, F. D. Saei, M. Lakestani, B. N. Saray, Pseudospectral method for solving fractional Sturm-Liouville problem using Chebyshev cardinal functions, Phys. Scr., 96 (2021), 125267. https://doi.org/10.1088/1402-4896/ac3c59 doi: 10.1088/1402-4896/ac3c59
    [2] A. Afarideh, F. D. Saei, B. N. Saray, Eigenvalue problem with fractional differential operator: Chebyshev cardinal spectral method, J. Math. Model., 11 (2023), 343–355. https://doi.org/10.22124/JMM.2023.24239.2169 doi: 10.22124/JMM.2023.24239.2169
    [3] M. A. Al-Gwaiz, Sturm-Liouville theory and its applications, London: Springer, 2008. https://doi.org/10.1007/978-1-84628-972-9
    [4] Q. M. Al-Mdallal, An efficient method for solving fractional Sturm-Liouville problems, Chaos Solitons Fract., 40 (2009), 183–189. https://doi.org/10.1016/j.chaos.2007.07.041 doi: 10.1016/j.chaos.2007.07.041
    [5] A. L. Andrew, J. W. Paine, Correcttion of Numerov's eigenvalue estimates, Numer. Math., 47 (1985), 289–300. https://doi.org/10.1007/BF01389712 doi: 10.1007/BF01389712
    [6] M. Arif, F. Ali, I. Khan, K. S. Nisar, A time fractional model with non-singular kernel the generalized couette flow of couple stress nanofluid, IEEE Access, 8 (2020), 77378–77395. https://doi.org/10.1109/ACCESS.2020.2982028 doi: 10.1109/ACCESS.2020.2982028
    [7] M. Asadzadeh, B. N. Saray, On a multiwavelet spectral element method for integral equation of a generalized Cauchy problem, BIT Numer. Math., 62 (2022), 1383–1416. https://doi.org/10.1007/s10543-022-00915-1 doi: 10.1007/s10543-022-00915-1
    [8] B. S. Attili, D. Lesnic, An efficient method for computing eigenelements of Sturm-Liouville fourth-order boundary value problems, Appl. Math. Comput., 182 (2006), 1247–1254. https://doi.org/10.1016/j.amc.2006.05.011 doi: 10.1016/j.amc.2006.05.011
    [9] A. Benkerrouche, D. Baleanu, M. S. Souid, A. Hakem, M. Inc, Boundary value problem for nonlinear fractional differential equations of variable order via Kuratowski MNC technique, Adv. Differ. Equ., 2021 (2021), 1–19. https://doi.org/10.1186/s13662-021-03520-8 doi: 10.1186/s13662-021-03520-8
    [10] E. S. Baranovskii, Analytical solutions to the unsteady Poiseuille flow of a second grade fluid with slip boundary conditions, Polymers, 16 (2024), 1–16. https://doi.org/10.3390/polym16020179 doi: 10.3390/polym16020179
    [11] J. P. Boyd, Chebyshev and Fourier spectral methods, 2 Eds., Mineola: Dover Publications, 2001.
    [12] B. Chanane, Accurate solutions of fourth order Sturm-Liouville problems, J. Comput. Appl. Math., 234 (2010), 3064–3071. https://doi.org/10.1016/j.cam.2010.04.023 doi: 10.1016/j.cam.2010.04.023
    [13] A. L. Chang, H. G. Sun, C. M. Zheng, B. Q. Lu, C. P. Lu, R. Ma, et al., A time fractional convection-diffusion equation to model gas transport through heterogeneous soil and gas reservoirs, Phys. A, 502 (2018), 356–369. https://doi.org/10.1016/j.physa.2018.02.080 doi: 10.1016/j.physa.2018.02.080
    [14] C. Canuto, M. Y. Hussaini, A. Quarteroni, T. A. Zang, Spectral methods: fundamentals in single domains, Berlin, Heidelberg: Springer, 2006. https://doi.org/10.1007/978-3-540-30726-6
    [15] L. Chen, H. P. Ma, Approximate solution of the Sturm-Liouville problems with Legendre-Galerkin-Chebyshev collocation method, Appl. Math. Comput., 206 (2008), 748–754. https://doi.org/10.1016/j.amc.2008.09.038 doi: 10.1016/j.amc.2008.09.038
    [16] V. Daftardar-Gejji, H. Jafari, Adomian decomposition: a tool for solving a system of fractional differential equations, J. Math. Anal. Appl., 301 (2005), 508–518. https://doi.org/10.1016/j.jmaa.2004.07.039 doi: 10.1016/j.jmaa.2004.07.039
    [17] G. J. Fix, J. P. Roof, Least squares finite-element solution of a fractional order two-point boundary value problem, Comput. Math. Appl., 48 (2004), 1017–1033. https://doi.org/10.1016/j.camwa.2004.10.003 doi: 10.1016/j.camwa.2004.10.003
    [18] P. Ghelardoni, Approximations of Sturm-Liouville eigenvalues using boundary value methods, Appl. Numer. Math., 23 (1997), 311–325. https://doi.org/10.1016/S0168-9274(96)00073-6 doi: 10.1016/S0168-9274(96)00073-6
    [19] M. A. Hajji, Q. M. Al-Mdallal, F. M. Allan, An efficient algorithm for solving higher-order fractional Sturm-Liouville eigenvalue problems, J. Comput. Phys., 272 (2014), 550–558. https://doi.org/10.1016/j.jcp.2014.04.048 doi: 10.1016/j.jcp.2014.04.048
    [20] Y. Huang, J. Chen, Q. Z. Luo, A simple approach for determining the eigenvalues of the fourth-order Sturm-Liouville problem with variable coefficients, Appl. Math. Lett., 26 (2013), 729–734. https://doi.org/10.1016/j.aml.2013.02.004 doi: 10.1016/j.aml.2013.02.004
    [21] A. A. Kilbas, H. M. Srivastava, J. J. Trujillo, Theory and applications of fractional differential equations, Elsevier, 2006.
    [22] M. Lakestani, M. Dehghan, The use of Chebyshev cardinal functions for the solution of a partial differential equation with an unknown time-dependent coefficient subject to an extra measurement, J. Comput. Appl. Math., 235 (2010), 669–678. https://doi.org/10.1016/j.cam.2010.06.020 doi: 10.1016/j.cam.2010.06.020
    [23] K. Marynets, Analysis of a Sturm-Liouville problem arising in atmosphere, J. Math. Fluid Mech., 26 (2024), 38. https://doi.org/10.1007/s00021-024-00873-4 doi: 10.1007/s00021-024-00873-4
    [24] K. Marynets, A Weighted Sturm-Liouville problem related to ocean flows, J. Math. Fluid Mech., 20 (2018), 929–935. https://doi.org/10.1007/s00021-017-0347-0 doi: 10.1007/s00021-017-0347-0
    [25] J. A. T. Machado, M. F. Silva, R. S. Barbosa, I. S. Jesus, C. M. Reis, M. G. Marcos, et al., Some applications of fractional calculus in engineering, Math. Probl. Eng., 2010 (2010), 639801. https://doi.org/10.1155/2010/639801 doi: 10.1155/2010/639801
    [26] F. Mainardi, Fractional calculus and waves in linear viscoelasticity, Imperial College Press, 2010. https://doi.org/10.1142/p614
    [27] K. S. Miller, B. Ross, An introduction to the fractional calculus and fractional differential equations, New York: Wiley, 1993.
    [28] K. B. Oldham, J. Spanier, The fractional calculus, New York: Academic Press, 1974.
    [29] I. Podlubny, Fractional differential equations, Academic Press, 1999.
    [30] K. Sayevand, H. Arab, An efficient extension of the Chebyshev cardinal functions for differential equations with coordinate derivatives of non-integer order, Comput. Methods Differ. Equ., 6 (2018), 339–352.
    [31] M. I. Syam, H. I. Siyyam, An efficient technique for finding the eigenvalues of fourth-order Sturm-Liouville problems, Chaos Solitons Fract., 39 (2009), 659–665. https://doi.org/10.1016/j.chaos.2007.01.105 doi: 10.1016/j.chaos.2007.01.105
    [32] M. Shahriari, B. N. Saray, B. Mohammadalipour, S. Saeidian, Pseudospectral method for solving the fractional one-dimensional Dirac operator using Chebyshev cardinal functions, Phys. Scr., 98 (2023), 055205. https://doi.org/10.1088/1402-4896/acc7d3 doi: 10.1088/1402-4896/acc7d3
    [33] L. Shi, B. N. Saray, F. Soleymani, Sparse wavelet Galerkin method: application for fractional Pantograph problem, J. Comput. Appl. Math., 451 (2024), 116081. https://doi.org/10.1016/j.cam.2024.116081 doi: 10.1016/j.cam.2024.116081
    [34] Z. Shi, Y. Y. Cao, Application of Haar wavelet method to eigenvalue problems of high order differential equations, Appl. Math. Model., 36 (2012), 4020–4026. https://doi.org/10.1016/j.apm.2011.11.024 doi: 10.1016/j.apm.2011.11.024
    [35] W. Weaver Jr., S. P. Timoshenko, D. H. Young, Vibration problems in engineering, John Wiley & Sons, 1991.
    [36] Q. Yuan, Z. Q. He, H. N. Leng, An improvement for Chebyshev collocation method in solving certain Sturm-Liouville problems, Appl. Math. Comput., 195 (2008), 440–447. https://doi.org/10.1016/j.amc.2007.04.113 doi: 10.1016/j.amc.2007.04.113
    [37] U. Yücel, B. Boubaker, Differential quadrature method (DQM) and Boubaker polynomials expansion scheme (BPES) for efficient computation of the eigenvalues of fourth-order Sturm-Liouville problems, Appl. Math. Model., 36 (2012), 158–167. https://doi.org/10.1016/j.apm.2011.05.030 doi: 10.1016/j.apm.2011.05.030
  • This article has been cited by:

    1. Shichao Wu, Xianzhou Lv, Yingbo Liu, Ming Jiang, Xingxu Li, Dan Jiang, Jing Yu, Yunyu Gong, Rong Jiang, Enhanced SSD framework for detecting defects in cigarette appearance using variational Bayesian inference under limited sample conditions, 2024, 21, 1551-0018, 3281, 10.3934/mbe.2024145
    2. Zheng Zhang, Xiang Lu, Shouqi Cao, An efficient detection model based on improved YOLOv5s for abnormal surface features of fish, 2024, 21, 1551-0018, 1765, 10.3934/mbe.2024076
    3. Haixia Xu, Fanxun Ding, Wei Zhou, Feng Han, Yanbang Liu, Jiang Zhu, CFF-YOLO: cross-space feature fusion based YOLO model for screw detection in vehicle chassis, 2024, 18, 1863-1703, 8537, 10.1007/s11760-024-03474-w
    4. Youliang Zhang, Guowu Yuan, Hao Wu, Hao Zhou, MAE-GAN: a self-supervised learning-based classification model for cigarette appearance defects, 2024, 4, 2771-392X, 253, 10.3934/aci.2024015
    5. Houde Wu, Ting Chen, Longshuang Wang, Li Guo, Speed and accuracy in Tandem: Deep Learning-Powered Millisecond-Level pulmonary embolism detection in CTA, 2025, 106, 17468094, 107792, 10.1016/j.bspc.2025.107792
    6. Jianhua Liao, Jun Cao, Wenjie Long, Guozhong Wu, Yang Li, Shihao Tang, Yang Cao, Jing Yang, 2024, A Filter Capsule Detection and Correction Method Based on Visual Detection and PID Control Linkage, 979-8-3315-0707-7, 851, 10.1109/ICCC62609.2024.10942241
    7. Xi Hu, Xianghua Zeng, Jinshan Lei, Xinan Yang, Chunguang Li, Rui Chen, Wei Zhang, Qiuling Wang, Wenkui Zhu, 2024, YOLOWEN:An Efficient Multi-object Detection Model for Cigar Appearance Defects, 979-8-3503-5541-3, 1303, 10.1109/ICICML63543.2024.10957923
    8. Dejin Zhao, Rui Sun, Wei Li, Yunjie Ma, Tong Tong, Xiaolong Yuan, Dechao Wang, Peng Liu, Bo Li, Dexin Kong, Jianhai Zhang, LAM-YOLOv10: lightweight and multiscale feature enhancement for paint surface defect detection model, 2025, 34, 1017-9909, 10.1117/1.JEI.34.3.033022
  • Reader Comments
  • © 2024 the Author(s), licensee AIMS Press. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0)
通讯作者: 陈斌, bchen63@163.com
  • 1. 

    沈阳化工大学材料科学与工程学院 沈阳 110142

  1. 本站搜索
  2. 百度学术搜索
  3. 万方数据库搜索
  4. CNKI搜索

Metrics

Article views(1559) PDF downloads(34) Cited by(0)

Figures and Tables

Figures(2)  /  Tables(4)

Other Articles By Authors

/

DownLoad:  Full-Size Img  PowerPoint
Return
Return

Catalog