Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Insect Sci., 26 September 2025

Sec. Pest Management

Volume 5 - 2025 | https://doi.org/10.3389/finsc.2025.1635439

This article is part of the Research TopicModeling, Remote Sensing, and Machine Learning in Pest ManagementView all 3 articles

GIWT-YOLO: an efficient multi-scale framework for real-time Scolytinae pests detection

Jingwei Liu,,Jingwei Liu1,2,3Yongke Li,,*Yongke Li1,2,3*Lei Wang,,Lei Wang1,2,3Yunjie Zhao,,*Yunjie Zhao1,2,3*Bowen Mao,,Bowen Mao1,2,3Pengying Wang,,Pengying Wang1,2,3
  • 1College of Computer and Information Engineering, Xinjiang Agricultural University, Urumqi, China
  • 2Xinjiang Agricultural Informatization Engineering Technology Research Center, Xinjiang Agricultural University, Urumqi, China
  • 3Research Center for Intelligent Agriculture, Ministry of Education Engineering, Urumqi, China

The broad range of Scolytinae pests sizes and their subtle visual similarities, especially in smaller species, continue to challenge the accuracy of mainstream object detection models. To address these challenges, this paper proposes GIWT-YOLO, a lightweight detection model based on YOLOv11s, specifically tailored for Scolytinae pests detection. (1) We designed a lightweight multi-scale convolution module, GIConv, to improve the model’s ability to extract features at different pest scales. This module enhances the accuracy of small-object detection while reducing the computational cost and parameter complexity of the backbone. (2) The WTConv module inspired by wavelet transform is introduced into the backbone. This enlarges the effective receptive field and improves the model’s ability to distinguish pests with similar textures. (3) An SE attention mechanism is incorporated between the Neck and Head to enhance the model’s focus on key feature regions. Experimental results show that GIWT-YOLO achieves 84.7% in Precision, 88.7% in mAP@50, and 63.4% in mAP@50~95, which are improvements of 2.2%, 4.0%, and 3.1%, respectively, compared to the baseline YOLOv11s. Additionally, the model’s parameters and GFLOPs are reduced by 11.3% and 13.4%, respectively. Our proposed model surpasses the state-of-the-art (SOTA) performance in small-sized pest detection while maintaining a lightweight architecture, and its generalization ability has been validated on other public datasets. Our model provides an efficient solution for detecting Scolytinae pests. In future work, we plan to collect additional images of various pest species to expand the dataset, further enhancing the model’s applicability to a wider range of pest detection scenarios.

1 Introduction

As one of the largest forestry countries in the world, China has experienced a sharp increase in harmful organisms in recent years due to the impact of pests and diseases, resulting in significant economic losses and posing severe challenges to forestry development (1). However, due to the wide variety of pests, the low accuracy of pest identification has been plaguing personnel (2). The traditional method is mainly identified by professionals, which has a huge workload, low efficiency. Additionally, it lacks real-time monitoring of pest outbreaks, making it difficult to implement preventive strategies in a timely manner. Therefore, timely and accurate detection and identification of forest pests are crucial for effective forest management (3).

The subfamily Scolytinae is among the most destructive forest pests, widely distributed across forest ecosystems worldwide, and poses serious threats particularly to coniferous species such as pine and spruce (4). With the development of deep learning technology, it has become a more effective scheme to use convolutional neural network technology to realize the automatic detection of forestry pests. Applying it to pest identification can not only significantly improve recognition accuracy and processing efficiency, but also enable the automated analysis of large volumes of trap images, thereby overcoming the limitations of manual identification in terms of time, labor, and accuracy. However, due to the wide variety of pests and large differences in morphology and size, there are problems such as false detection, and overlapping detection.

Traditional pest identification methods primarily rely on image processing and feature engineering, combined with machine learning algorithms for pest classification. Wang et al. (5) proposed a whitefly counting algorithm based on K-means clustering and ellipse fitting. Deng et al. (6) extended the HMAX model, utilizing Scale-Invariant Feature Transform (SIFT) and Non-negative Sparse Coding (NNSC) to extract features, followed by the use of SVM for pest identification. Ma et al. (7) introduced an SVM classifier based on a combination of global features and HOG features for pest classification. Yang et al. (8) proposed an image recognition algorithm for pests on greenhouse whitefly and thrips trap boards, using Prewitt and Canny edge detection operators for segmentation, followed by SVM. However, this method is for a single pest species and does not consider pest detection in complex environments. Peng and Jinlan (9) proposed a small green leafhopper recognition method based on PCA-LDA-SVM. Despite progress in feature extraction, these methods still struggle to meet practical requirements. Machine learning-based vision algorithms face issues in complex environments, such as low recognition accuracy, slow inference speed, and poor robustness. Furthermore, they tend to rely heavily on the color of static pests for identification, over-depend on sample-specific features, and exhibit poor generalization capability.

In recent years, deep convolutional neural networks have gained widespread attention in the field of forestry pest detection. YOLO (You Only Look Once) (1016) series algorithms have achieved an effective balance between detection accuracy and computational complexity, garnering significant interest and widespread application among researchers. Zhong et al. (17) proposed a method to count the number of flying insects in images using the YOLO model for object detection, followed by further classification through SVM. However, the study primarily focused on simple backgrounds and lacked research on insect recognition in more complex environments. Bai et al. (18) proposed a MOG2-YOLOv4 detection model for East Asian migratory locusts, which effectively solved the problem of low recognition accuracy caused by high speed movement and occlusion of East Asian migratory locusts. Wang (19) used an improved ALexNet model to identify agricultural pests, which effectively accelerated the inference speed of the model. Zhang et al. (20) improved the YOLOv3 model and combined Spatial Pyramid Pooling (SPP) with it to improve the accuracy of small-size pests. However, the model’s complexity was not considered, which limited its applicability for deployment on edge devices. Bhatt et al. (21) used YOLOv3 model for object detection in tea garden pest images, and the mAP reached 86% while maintaining the same inference performance. Liu and Wang (22) improved the YOLOv3 model by adding Spatial Pyramid Pooling (SPP) and utilizing multi-scale techniques to better capture tomato pest features, achieving a detection accuracy of 92.39%. Wen et al. (23) enhanced the YOLOv4 model on the Pest24 dataset and proposed the PEST-YOLO detection method for detecting multiple types of dense, tiny pests. However, the focus was primarily on improving missed detection issues, neglecting the balance between detection accuracy and mAP. Zhongzhu et al. (24) insert a Global Attention Upsampling (GAU) module into the output layer of the backbone network. This module uses global information from high-level features to help the model to extract features from complex backgrounds, improving recognition accuracy. However, it still suffers from high false detection rates. Yuan et al. (25) added a convolutional attention module to improve the feature expression of tea garden pests. Jiang and Yang (26) proposed a lightweight pest detection algorithm based on YOLOv8n (Bm-YOLO), designed the MCCA attention mechanism module to integrate with C2f, and performed well on the IP102 dataset, but the recognition accuracy is still insufficient in complex environments. There are some problems such as false detection, missed detection, and overlapping detection. Although the above studies have achieved certain improvements in detection accuracy, the adopted models are computationally complex and lack of lightweight optimization, which limits their application on edge devices.

Due to the limited lightweight capability of existing models in pest detection tasks, deploying them on edge devices is challenging. In addition, the similarity of features among Scolytinae pests makes feature extraction difficult. This is especially true for small pests, whose tiny size and fuzzy texture make detection harder. To address the above issues, we propose GIWT-YOLO, an improved object detection model specifically designed for tiny pests. Its architecture is based on YOLOv11s (13), a recent framework introduced by Khanam and Hussain in 2024, which serves as a strong and modern baseline. The following improvements are made in this paper:

1. We proposed a lightweight model, GIWT-YOLO, for detecting small and visually similar pest targets. The model improves detection accuracy for small pests and those with similar appearances.

2. We proposed GIConv, a convolution module suitable for multi-scale pest detection, to replace the standard convolution in the YOLOv11s backbone. This module improves the model’s ability to detect pests at different scales while making the backbone network more lightweight.

3. In the C3K2 structure, the WTConv module, inspired by wavelet transform, is introduced to construct the C3K2_WT structure. This modification enlarges the effective receptive field, improving the model’s ability to distinguish pests with similar textures. At the same time, it reduces model complexity and maintains inference efficiency.

4. Considering the high morphological similarity among different pest species, this paper introduces the SE attention mechanism module between the Neck and Head networks. This module enhances the model’s focus on key feature regions, enabling it to more accurately capture differences between pest categories and improve its ability to differentiate similar pests.

The structure of this paper is as follows: Section 2 introduces the dataset acquisition and the GIWT-YOLO model architecture, with detailed explanations of the improvements made to each module. Section 3 covers the experimental design and result analysis. Finally, Sections 4 and 5 provide the discussion and conclusion.

2 Materials and methods

2.1 Dataset introduction

The dataset utilized in this research was sourced from the Baidu AI Studio platform, an open online environment powered by PaddlePaddle, which facilitates data sharing, collaborative development, and participation in machine learning challenges. The dataset used in this study is the Scolytinae pests dataset provided by Beijing Forestry University (4) (https://aistudio.baidu.com/datasetdetail/51399, accessed on 15 January 2025). The dataset contains 2,183 image samples. Some sample images from the dataset are shown in Figure 1. These images include six insect categories: Boerner, Leconte, Linnaeus, Acuminatus, Armandi, and Coleoptera. In this study, the pest images are manually annotated using the LabelImg tool, with each image’s annotations stored in an Extensible Markup Language (XML) (27) file following the VOC format.

Figure 1
Six images of beetles are shown. The top row has three beetles labeled Leconte, Boerner, and Linnaeus. The second row includes beetles labeled acuminatus, armandi, and coleoptera. Below are two petri dishes containing multiple beetles, highlighted with colored rectangles.

Figure 1. Data sample image.

To ensure the effectiveness of model training, the original dataset of 2,183 image samples was divided into a training set (1,693 images), a validation set (245 images), and a test set (245 images). The pest samples in this dataset exhibit significant differences in body size, and the images were captured under varying lighting conditions. As illustrated in Figures 1A–C, these images show different levels of illumination. Considering that data diversity can enhance training stability, this study adopts the Mixup (28) and Mosaic (10) data augmentation strategies, without employing additional offline augmentation techniques. This approach enhances data diversity and improves the generalization ability of the model, while avoiding the potential noise introduced by excessive augmentation.

2.2 GIWT-YOLO pest detection model

The YOLOv11s (13) model has achieved significant improvements in both inference performance and accuracy. However, it still involves high computational complexity and performs poorly in detecting small objects. To solve this problem, this study proposes an improved lightweight GIWT-YOLO model based on YOLOv11s, and its model structure is shown in Figure 2. In order to achieve a lightweight network and high pest detection accuracy, GIWT-YOLO mainly makes the following improvements. First, the GIConv module is introduced to replace the ordinary convolution in the YOLOv11s backbone. This enhances the model’s ability to detect pests at different scales, especially improving recognition accuracy for small-object pests. It reduces the amount of computation and the number of parameters at the same time. Secondly, in the C3K2 structure, the WTConv module is introduced to construct the C3K2_WT structure. This modification increases the effective receptive field, thereby enhancing the model’s ability to detect pests with similar feature textures. Finally, Considering the high morphological similarity among different pest species, this paper introduces an SE attention mechanism module between the Neck and Head networks. This module enhances the model’s focus on key feature regions, enabling it to more accurately capture differences between pest categories and improve its ability to differentiate similar pests.

Figure 2
Diagram illustrating a neural network architecture with two main components. In part A, an Inception Block receives input of dimensions H x W x c, processes it through multiple convolution layers and pooling operations, and outputs the same dimensions. In part B, a GIConv block takes input, applies a 3x3 convolution, and processes it through an Inception Block and an Identity function before concatenating the results for the output.

Figure 2. The proposed GIWT-YOLO algorithm model. The red dashed line represents the added improvement module.

2.3 GIConv module design

As the pests in the Scolytinae pests dataset show more obvious body size differences, it is difficult for a single-scale convolution to take into account the feature extraction of pests of different sizes. Inspired by the Inception (29) module and GhostConv (30), we propose Ghost Inception Convolution(GIConv), which eliminates the use of simple linear operation. Firstly, a part of the core feature map is obtained using the 3x3 convolution module. Then, the Inception module (shown in Figure 3A) is utilized to obtain a multi-scale feature map. Finally, the core feature map and multi-scale feature map are spliced by Concat to finally obtain a multi-scale feature map. The convolution makes full use of the advantages of multi-scale depth convolution. It effectively solves the problem of missing feature information caused by the difference in pest body size by paralleling convolution layers of different scales. As a result, the model can extract global information from large object pests. At the same time, it pays more attention to the detailed features of tiny pests, thus improving detection accuracy. In addition, compared with ordinary convolution, the GIConv module significantly reduces the computational volume, and improves the computational efficiency and overall performance of the model. Its structure is shown in Figure 3B.

Figure 3
Diagram illustrating a neural network architecture with two main components. In part A, an Inception Block receives input of dimensions H x W x c, processes it through multiple convolution layers and pooling operations, and outputs the same dimensions. In part B, a GICOnv block takes input, applies a 3x3 convolution, and processes it through an Inception Block and an Identity function before concatenating the results for the output.

Figure 3. The structure of the GIConv module. (A) shows a parallel multi-branch structure with multi-scale convolutions; (B) shows the structure of GIConv.

The amount of computation for each parallel path of the Inception module and the total computation cost, which is calculated as follows:

cost1=14×H×w×c2(1)
cost2=H×w×c×c8+3×3×c8×H×w×c4=1332×H×w×c2(2)
cost3=H×w×c×c8+3×3×c8×H×w×c4+3×3×c4×H×w×c4=3132×H×w×c2(3)
cost4=14×H×w×c2(4)
cost=cost1+cost2+cost3+cost4=158×H×w×c2(5)

In Equations 1-5, H, w, c are the height, width and number of channels of the input and output feature maps; The computation amount of the four parallel paths is: cost1,cost2,cost3,cost4, and the total computation amount of the final Inception module is:cost.

The amount of computation required for ordinary convolution in YOLO11s is:

costa=H×W×n×k×k×c(6)

If the ordinary convolution in YOLO11s is replaced by the GIConv module, the computational cost of the module is as follows:

costb=H×W×n2×k×k×c+cost=H×W×n2×k×k×c+158×H×W×(n2)2(7)
S=costbcosta=12+15n32×k2×c(8)

In Equations 6, 7, and 8, h, w, c are the height, width, and number of channels of the input feature maps; H, W, n are the height, width, and number of channels of the output feature maps; and S is the ratio of the computation amount of the GIConv module to the computation amount of the ordinary convolution. In this study, k=3 and n=2c, S is approximately equal to 0.6, so the computation of GIConv module is reduced to 60% of the ordinary convolution module.

2.4 C3K2_WT module

Due to the similar texture and color of some pests in the data samples. It is difficult for the C3K2 module to extract enough key features, which affects the classification and detection accuracy of the pests. Therefore, a large size convolutional kernel is used to expand the receptive field to capture more detailed features. However, the use of larger convolutional kernels results in a quadratic growth in computational complexity, which presents a significant challenge for achieving model lightweighting. To address this issue, WTConv (31) introduces the Wavelet Transform (WT). It uses a set of small-sized convolutional modules, with each convolution focusing on a different frequency bands of the input, resulting in an increasingly larger receptive field. The structural diagram of WTConv is shown in Figure 4. Notably, the number of parameters in WTConv increases logarithmically with the expansion of the receptive field. Compared with standard convolutional methods, WTConv achieves a larger receptive field while effectively avoiding the problem of excessive parameter growth.

Figure 4
Diagram illustrating a wavelet transform (WT) process on image data. It shows an input image being processed through two iterations of WT, labelled as \(i=1\) and \(i=2\), producing sub-components \(X_{LL}\), \(X_{LH}\), \(X_{HL}\), \(X_{HH}\) at each stage. The components are passed through a convolution (Conv) and are eventually recombined using an inverse wavelet transform (IWT) to produce the output image. Each step is connected by arrows.

Figure 4. WTConv module structure. Wavelet Transform (WT) Principle, X is the input feature map, XLL is the low-frequency component, and XLH,XHL,XHHis the horizontal, vertical and diagonal high-frequency components. Among them, when i = 0, XLL(0)=X, i represents the current level. "+" represents the Concat operation. IWT represents the Inverse Wavelet Transform operation.

In this study, the WTConv module is integrated into the C3K2 structure, resulting in a novel structure named C3K2_WT, as illustrated in Figure 5. Firstly, inspired by the principles of the Wavelet Transform (WT), the standard convolutional layers within the Bottleneck block are replaced with WTConv modules. This leads to the design of a new Bottleneck_WT module (Figure 5E). This module effectively enlarges the receptive field and enables the extraction of more discriminative features, particularly for pests with highly similar textures and colors, thereby enhancing the model’s feature extraction capability. Secondly, leveraging the branching characteristics of the C3K2 structure, we further embed the Bottleneck_WT module specifically into the C3K block. This replaces the original Bottleneck components and forms a new C3K_WT structure (Figure 5C). Meanwhile, to maintain model flexibility and adaptability, the original Bottleneck block is also retained (Figure 5D). Ultimately, the new C3K2_WT structure is constructed(Figures 5A, B). This design enhances the model’s capacity for fine-grained feature extraction and detection accuracy for Scolytinae pests detection, while also effectively reducing the number of parameters.

Figure 5
Five diagrams labeled A to E illustrate different neural network architectures with variations in convolutional (Conv) and bottleneck layers. Diagram A shows a split unit with bottleneck layers and contact. Diagram B includes C3K_WT units. Diagram C uses Bottleneck_WT layers. Diagram D highlights a shortcut with bottleneck architecture. Diagram E features a Bottleneck_WT with WTConv and shortcut integration.

Figure 5. C3K2_WT module structure. (A) C3K2_WT structure when C3K_WT is True; (B) C3K2_WT structure when C3K_WT is False; (C) C3K_WT structure; (D) Bottleneck structure; (E) Bottleneck_WT structure.

2.5 SE attention mechanism module

The traditional YOLO network architecture directly transmits the output pest feature maps to the detection head for pest detection after feature fusion. However, in this process, the importance of channels is not effectively distinguished. So the key feature information of pests is not fully utilized, which directly affects the accuracy of the model pest detection.

Considering the high morphological similarity between Scolytinae pests species, which causes issues such as false positives and false negatives during the detection process. We introduce targeted optimizations to the YOLO network architecture to enhance the model’s ability to detect Scolytinae pests. Specifically, the SE(Squeeze and Excitation) (32) attention mechanism is integrated between the Neck and Head networks. This module utilizes the Squeeze operation to extract global contextual information, and then the Excitation operation adaptively learns the importance weights of each channel. This significantly improves the model’s responsiveness to discriminative features in pest images, thereby enhancing its ability to distinguish between categories. As a result, the detection head can focus more effectively on the critical regions of the pests, improving the model’s sensitivity to fine-grained differences between Scolytinae pests species. This enhancement significantly boosts the model’s capability to differentiate between pests in complex backgrounds and highly similar species, thereby improving overall detection accuracy and robustness. The structure is illustrated in Figure 6.

Figure 6
Diagram illustrating a squeeze-and-excitation network structure. It shows the process starting from “Neck” with a convolutional block, leading to a squeeze step using Global Average Pooling (GAP), followed by two fully connected layers with ReLU and Sigmoid functions. The excitation step produces a weighted channel output, leading to the “Head.

Figure 6. Schematic of the SE attention mechanism.

3 Results

3.1 Experimental design

3.1.1 Experimental setup and hyperparameter configuration

This experiment uses PyTorch as the deep learning framework, with the detailed experimental environment shown in Table 1. To optimize model training, we combine cosine annealing and early stopping strategies. Cosine annealing dynamically adjusts the learning rate, gradually decreasing it during training to help the model fine-tune parameters in the later stages and avoid the issue of local minima. At the same time, the early stopping strategy monitors the validation loss to prevent overfitting. Training will be stopped early if the validation loss does not significantly converge within 100 consecutive epochs, thus reducing training time and avoiding overfitting. The combination of these two strategies optimizes the model training process and ensures the best performance on the validation set. During training, the batch size is set to 32, the number of epochs is set to 300, the initial learning rate is set to 0.01, weight decay is set to 0.0005, and the momentum factor is set to 0.937. To prevent getting stuck in local optima, the SGD optimizer is used in this experiment.

Table 1
www.frontiersin.org

Table 1. Hardware and software environment.

3.1.2 Evaluation metrics

To achieve a balance between detection accuracy and model complexity in the Scolytinae pests detection task. This study employs Precision(P), Recall (R), F1 score and mean Average Precision (mAP) as the primary evaluation metrics for assessing the detection performance of the proposed model. Meanwhile, the model complexity is measured in terms of parameters, and GFLOPs. The corresponding calculation formulas are shown in Equations 9-14:

Precision=TPTP+FP×100%(9)
Recall=TPTP+FN×100%(10)
F1=2×Precision×RecallPrecision+Recall×100%(11)

In Equations 9 and 10: TP (True Positive) refers to the number of correctly detected pests; FN(False Negative) denotes the number of pests that were not correctly detected; TN (True Negative) indicates the number of non-pest instances that were correctly classified as non-pests; FP (False Positive) refers to the number of non-pest instances that were incorrectly detected as pests. In Equation 11, the F1 score is defined as the harmonic mean of Precision and Recall, which comprehensively evaluates the trade-off between accuracy and completeness of the model.

mAP=1n1nPi=1n(P1+P2++Pn)(12)
mAP@50~95=1ck=1cmAP@50k(13)
mAP@50~95=110(mAP@50+mAP@55+mAP@95)(14)

In Equations 12-14: mAP@0.5 represents the mean Average Precision when the Intersection over Union(IoU) threshold is set to 0.5. mAP@0.5~0.95 refers to the average of mAP values calculated at multiple IoU thresholds ranging from 0.5 to 0.95 with a step size of 0.05. This metric adopts a stricter and more comprehensive evaluation standard, offering a more accurate reflection of the model’s overall detection performance.

3.1.3 Model architecture and parameter details

GIWT-YOLO is primarily used for the detection of Scolytinae pests. First, the input image is resized to 640×640, and Mosaic data augmentation is applied to improve the model’s generalization ability. Next, the preprocessed image is input into the backbone network for feature extraction, capturing multi-scale pest features. The extracted features then enter the neck network, where features from different scales are further fused to enhance the detection capability for small pests. Finally, the fused feature maps are refined using the SE attention mechanism for channel weight adjustment. The detection head then performs object classification and bounding box regression, producing the final detection results. The detailed network architecture of GIWT-YOLO is shown in Table 2.

Table 2
www.frontiersin.org

Table 2. Detailed Parameters of GIWT-YOLO Network.

3.2 Result and analysis

3.2.1 Performance comparison and analysis of various object detection models

Since GIWT-YOLO is developed upon the YOLO11 architecture, comparing it with earlier versions of YOLO provides a direct and meaningful reference to assess the effectiveness of the proposed improvements. To comprehensively evaluate the detection performance of the GIWT-YOLO model, we conducted comparative experiments using the same experimental setting and training hyperparameters with YOLOv5s (11), YOLOv8s (14), YOLOv9s (16), YOLOv10s (15), RT-DETR-l (33), RT-DETR-resnet50 (33), and the baseline model YOLOv11s (13). The experimental results are shown in Table 3.

Table 3
www.frontiersin.org

Table 3. The performance comparison of different object detection models.

The main performance of GIWT-YOLO surpasses that of YOLOv5s, YOLOv8s, YOLOv9s, YOLOv10s, RT-DETR-l, RT-DETR-ResNet50, and YOLOv11s models. Specifically, it achieves a Precision of 84.7%, Recall of 82.2%, F1-Score of 83.4%, mAP@50 of 88.7%, and mAP@50~95 of 63.4%. The model’s GFLOPs and Parameters are only 18.7G and 8.4M, respectively, achieving a balance between detection accuracy and model lightweighting.

RT-DETR-l and RT-DETR-resnet50 are Transformer-based (34) object detection models, but their model sizes are relatively large, with model complexity and parameter counts significantly exceeding those of other detection algorithms. RT-DETR-l and RT-DETR-resnet50 have similar complexity and detection accuracy. The GFLOPs and Parameters of RT-DETR-Resnet50 are as high as 130.5G and 42.8M, approximately 7 times and 5 times greater than those of GIWT-YOLO, respectively. However, GIWT-YOLO outperforms RT-DETR-resnet50 by 2% in accuracy, 13% in mAP@50, and 9.6% in mAP@50~95. Overall, the performance of the GIWT-YOLO model far exceeds that of RT-DETR-l and RT-DETR-resnet50, achieving a balance between detection accuracy and model lightweighting.

Compared to the baseline model YOLOv11s, the GIWT-YOLO model improved by 2.2%, 2.4%, 2.3%, 4%, and 3.1% in Precision, Recall, F1-Score, mAP@50, and mAP@50~95, respectively, while reducing model GFLOPs and Parameters by 13.4% and 11.3%, respectively. Additionally, although YOLOv8s, YOLOv10s, and GIWT-YOLO have similar parameters, GIWT-YOLO significantly outperforms the other models in Precision, Recall, F1-Score, and mAP, fully demonstrating the effectiveness and superiority of the proposed improvements. These experimental results show that GIWT-YOLO achieves higher detection accuracy and stronger generalization ability while maintaining low GFLOPs. The model, designed to address the significant size differences among Scolytinae pests, adopts a multi-scale feature extraction strategy. This enables it to effectively capture features of pests of various sizes. To further comprehensively showcase the superiority of GIWT-YOLO’s performance, we present a comparison of Precision with mAP@50, mAP@50~95, GFLOPs, and F1-Score, as shown in Figure 7.

Figure 7
Scatter plots comparing precision against various metrics for different detection models. Top-left: Precision vs. mAP@50. Top-right: Precision vs. GFLOPs. Bottom-left: Precision vs. mAP@50~95. Bottom-right: Precision vs. F1-Score. Models include YOLOv5s, YOLOv8s, YOLOv9s, YOLOv10s, YOLOv11s, RT-DETR-I, RT-DETR-resnet50, and GIWT-YOLO, each identified by distinct shapes and colors.

Figure 7. Scatter plot of the performance of different models.

3.2.2 Ablation experiment

To evaluate the performance of the GIConv, C3K2_WT, and SEAttention modules integrated into the model, we conducted ablation experiments on the Scolytinae pests dataset from Beijing Forestry University. Starting with the YOLOv11s model, we progressively added each improvement module. Model1 replaced the standard convolution modules in the backbone network of YOLOv11s with the GIConv module. Model2 integrated the C3K2_WT module into the original YOLOv11s model. Model3 combined the GIConv module and the C3K2_WT module in the original model. Finally, based on Model3, we integrated the SEAttention module was integrated into both the Neck and Head networks, forming the final GIWT-YOLO model.

As shown in Table 4, based on the YOLOv11s model, Model1 introduced the innovative GIConv module. The training results demonstrate that the model achieved improvements of 1.8%, 2.6%, and 2.0% in Recall, mAP@50, and mAP@50~95, respectively, while the parameter count and GFOLPs decreased by 8.5% and 11.6%. This indicates that by addressing the size differences of Scolytinae pests in the images, the model effectively captures more relevant pest features through multi-scale convolution, resulting in more accurate detection and a reduced false negative rate. Model2, built upon the baseline YOLOv11s model, integrated the C3K2_WT module. By effectively expanding the receptive field, it enhanced the detection accuracy of pests with similar textures and colors. This improvement led to mAP@50 and mAP@50~95 values of 85.9% and 63.4%, respectively, which are 1.2% and 3.1% higher than those of the baseline model. Model3, based on Model1, integrated the C3K2_WT module. Through multi-scale convolution and expanded receptive fields, this model improved Scolytinae pests detection accuracy, with increases of 4.2%, 0.4%, and 0.9% in Precision, mAP@50, and mAP@50~95, respectively. Finally, compared to Model3, GIWT-YOLO enhanced the feature extraction capability by introducing the SE module. This strengthened the model’s focus on key feature regions, allowing it to more accurately capture the differences between different Scolytinae pests. As a result, Recall increased by 1.2%, mAP@50 increased by 1.0%, and mAP@50~95 increased by 0.2%. However, due to the impact of channel weighting, the model placed greater emphasis on key feature regions while neglecting some feature information, leading to a slight decrease of 1.6% in Precision. Nevertheless, the improvement in mAP@50~95 indicates enhanced object bounding box regression performance at different IoU thresholds, leading to an overall improvement in detection performance.

Table 4
www.frontiersin.org

Table 4. Ablation Experiment of GIWT-YOLO Model.

In summary, compared to the baseline model YOLOv11s, GIWT-YOLO achieves improvements of 2.2%, 2.4%, 4%, and 3.1% in Precision, Recall, mAP@50, and mAP@50~95, respectively. At the same time, the model’s parameters and GFLOPs are reduced by 11.3% and 13.4%, enhancing pest detection accuracy while achieving a lightweight design, enabling efficient detection of differently sized Scolytinae pests under limited computational resources. As shown in Figure 8, a comprehensive comparison of different model performances is presented.

Figure 8
Three-dimensional scatter plot depicting models represented as colored spheres at different points. Axes are Precision, mAP at 50 to 95, and mIOU at 50. Models listed in the legend include GIWT_YOLO, Mode13, Mode12, Mode11, and YOLOv11s, each color-coded.

Figure 8. Performance Comparison Among Different Models. Precision as the X-axis, mAP@50 as the Y-axis, and mAP@5095 as the Z-axis. The color in the legend represents different models, and the size of the balls corresponds to GFLOPs.

3.2.3 Comparison of different lightweight convolutions

When designing GIConv module, we verified its effectiveness through comparative analysis with the standard Convolution and GhostConv (30) modules. As shown in Table 5, compared with the GhostConv module, although the amount of GFLOPs is slightly increased by 0.3G and the number of parameters is 0.15M, the improvement in Precision, Recall, F1-Score, mAP@50 and mAP@50~95 indicators is more significant. Compared with standard Convolution, GIConv has significant performance improvement except for a slight decrease of 0.4% in Precision. Overall, the proposed GIConv module demonstrates superior performance. Figure 9 illustrates the performance curves using different convolution modules.

Table 5
www.frontiersin.org

Table 5. Experimental results for the lightweight convolutions.

Figure 9
Bar and line chart comparing the performance and parameters of different models. Three models are shown: YOLOv11s (84.7% mAP@50, 9.43M parameters), YOLOv11s+GhostConv (85.8% mAP@50, 8.47M parameters), YOLOv11s+GIConv (87.3% mAP@50, 8.62M parameters). Performance increases with each model while parameters vary.

Figure 9. Comparison of different convolutional performance.

3.2.4 Comparison of different attention modules

To investigate the impact of attention mechanisms on model performance and determine their optimal placement. This study inserts attention modules between the neck network and head network of the model, including SE (32), CBAM (35), CA (36), ECA (37), GAM (38), and CCA (39). A comparative analysis of their performance is conducted. As shown in Table 6, among the six attention mechanisms, the SE module achieves the best performance in terms of Precision, F1-Score, mAP@50, and mAP@50–95, while also maintaining the lowest computational complexity. Although the ECA module achieves the highest Recall, it performs the worst in terms of Precision and mAP@50. Overall, the SE module effectively enhances feature representation, improves detection accuracy and stability. It is maintains a low computational cost, making it more advantageous for practical applications. Figure 10 shows the performance curves of different attention mechanisms.

Table 6
www.frontiersin.org

Table 6. Comparison of different attention models’ performance.

Figure 10
Line chart comparing the performance of different models in terms of mAP@50, mAP@50-95, and GFLOPs. Model3+GAM shows the highest mAP@50 of 88.5% and mAP@50-95 of 63.1% with 34.4 GFLOPs. Other models show varying performances with GFLOPs ranging from 18.7 to 20.0.

Figure 10. Comparison chart of various attention mechanisms. Model3 integrates GIConv module and C3K2_WT module into YOLOv11s.

We investigated different placement strategies for attention modules within the network to evaluate their effects on detection performance, as shown in Table 7. The SE module inserted between neck network and head network compared with the backbone network, it improves the Precision, Recall, F1-Score, mAP@50, and mAP@50~95 by 2.3%, 2.5%, 2.4%, 1.7%, and 1.2%, respectively. This phenomenon is related to the feature requirements at different stages of the network. Inserting the SE module at the backbone network will lead to the squeeze and excitation of low-level features, prematurely affecting the weight of feature channels. This results in the loss of part of feature information and the decline of detection ability.

Table 7
www.frontiersin.org

Table 7. Experimental results on the effects of inserting attention modules at different positions.

3.2.5 Model visualization results

To provide a more comprehensive demonstration of the model’s detection performance, this study compares the annotated image, YOLOv8s, YOLOv10s, RT-DETR-l, RT-DETR-resnet50, YOLOv11s, and GIWT-YOLO models. The visualization results are shown in Figure 11, and the PR performance curves of each model are presented in Figure 12.

Figure 11
Annotated images showing a comparison of different object detection models: YOLOv8s, YOLOv10s, RT-DETR-I, RT-DETR-resnet50, YOLOv11s, and GIWT-YOLO. Each model's detection performance is displayed in a grid format with labels and confidence scores overlaid on images of beetles in circular containers.

Figure 11. Visualization of different models.

Figure 12
Six panels of precision-recall curves for different models: A) YOLOv8s, B) YOLOv10s, C) RT-DETR-l, D) RT-DETR-resnet50, E) YOLOv11s, F) GIWT-YOLO. Each curve displays performance for classes: Leconte, Boerner, coleoptera, armandi, Linnaeus, acuminatus, and all classes with mAP at an Intersection over Union threshold of 0.5.

Figure 12. Training PR curves for different models.

Combined with Figures 10 and 11, it can be clearly seen that the YOLOv8s (Figure 11B, Figure 12A) and YOLOv10s (Figure 11C, Figure 12B) models have weak feature extraction capabilities for small object pests. These models exhibit many false positives and low confidence in bounding boxes, which often result in misclassifying small object pests. For example, incorrectly detecting Linnaeus as belonging to the armandi category. This leads to missed detections for the correct Linnaeus class, causing an increase in false negatives (FN) and a decrease in true positives (TP), thus lowering Recall and the AP for the Linnaeus category. Additionally, misclassifying Linnaeus as armandi leads to an increase in false positives (FP), causing a decrease in Precision for the armandi category and a reduction in AP. On the other hand, RT-DETR-l (Figure 11D, Figure 12C) and RT-DETR-resnet50 (Figure 11E, Figure 12D) avoid the NMS (Non-Maximum Suppression) (40) operation. They achieve this by introducing region-aware and global optimization strategies. This effectively prevents redundant detections and the generation of duplicate bounding boxes. However, these models still experience category misclassification, which results in a low mAP.

The YOLOv11s (Figure 11F, Figure 12E) model exhibits category misclassification along with multiple overlapping bounding boxes. The main reason for this issue is the insufficient detection capability of the model, which lacks focus on small and similarly Scolytinae pests. As a result, some coleoptera are incorrectly classified as acuminatus, leading to AP of only 83.2% and 79.3% for coleoptera and acuminatus, respectively. In contrast, the improved GIWT-YOLO model (Figure 11G, Figure 12F) utilizes the innovative GIConv module and introduces wavelet transform concepts. It enhances the feature extraction capability of the backbone network through multi-scale convolutions and increased receptive field size. Additionally, an SE attention mechanism is incorporated between the neck network and head network to boost the model’s focus on pest objects, significantly reducing misclassifications and overlapping bounding box issues. As shown in Table 8, our method outperforms the state-of-the-art (SOTA) models in terms of AP across multiple pest categories. When compared to the baseline YOLOv11s, the APs for small Scolytinae pests like coleoptera, Linnaeus, armandi, and acuminatus improve by 1.8%, 2.4%, 7.7%, and 8.4%, respectively. The AP for larger Scolytinae pests, such as leconte and Boerner, also show improvements of 1.1% and 2.7%, reaching excellent performances of 97.6% and 95.2%, respectively. In conclusion, the GIWT-YOLO model not only achieves a lightweight design but also significantly outperforms other models in detecting pests of various sizes, particularly small Scolytines pests. This results in a notable enhancement in the accuracy and reliability of forestry pest detection.

Table 8
www.frontiersin.org

Table 8. Comparison of average precision (AP) for different pest categories across various models.

3.2.6 Validating the generalization of the GIWT-YOLO model

To further evaluate the generalization capability of the proposed model, this study employs the publicly available rice pest dataset RP11. Which refines the rice pest subset of IP102 and supplements it with additional images collected via web crawling (41)(https://www.kaggle.com/datasets/dingbiao11/rp11-a-dataset-focus-on-adult-rice-pest, accessed on 22 March 2025).

The dataset comprises 11 categories and a total of 4,559 images. It is randomly divided into training, validation, and test sets in a ratio of 8:1:1. All experimental environments and hyperparameter settings were kept consistent with Section 3.1.1 to ensure fair comparison. The experimental results are shown in Table 9. Compared to the YOLOv11s model, GIWT-YOLO demonstrated improvements across all performance metrics. Precision increased by 1.3%, significantly enhancing the detection accuracy for pests of different sizes. Additionally, mAP@50 and mAP@50~95 improved by 0.6% and 0.9%, respectively, indicating more stable detection capability across different IoU thresholds.

Table 9
www.frontiersin.org

Table 9. Comparison of generalization verification experimental results.

In real-world field conditions, we visualized the detection results on the RP11 pest dataset. As shown in Figure 13, despite challenges such as natural backgrounds and illumination changes, the GIWT-YOLO model was still able to detect most pests with good accuracy, demonstrating strong robustness. In contrast, the baseline YOLOv11s misclassified Delphacidae as Cicadellidae, and its confidence scores were overall lower than those of GIWT-YOLO. These findings highlight the practical applicability of GIWT-YOLO in complex agricultural environments, but it is also evident that complex backgrounds can lead to a decrease in detection performance. In future work, we will place greater emphasis on supplementing datasets collected under real field conditions to further enhance the generalization ability of GIWT-YOLO.

Figure 13
Top row displays images of two insects on a leaf, processed through different AI models. The original image is on the left, YOLOv11s in the center identifies them as “Cicadellidae” and “Delphacidae,” and GIWT_YOLO on the right identifies both as “Delphacidae” with higher confidence. Bottom row shows insects on leaves with boxes. The left image has unlabelled boxes, while the right image uses GIWT_YOLO to label them as “Curculionidae” with confidence scores ranging from 0.83 to 0.92.

Figure 13. Model Visualization Results.

4 Discussion

This study proposes a lightweight GIWT-YOLO model based on the improved YOLOv11s, specifically designed for the efficient detection of Scolytinae pests. It provides a valuable technical reference for automated pest monitoring and effectively addresses the detection challenges caused by variations in pest sizes, demonstrating significant application value in forestry pest surveillance and prevention.

Although this study has made progress in the detection of small pests, several limitations remain. Our improved model demonstrates a more balanced performance across different pest sizes. As shown in the PR curves in Figure 12, for larger pests, the model maintains a high precision even at around 90% recall. However, for smaller pests, the recall drops to about 85%, accompanied by a faster decline in precision, resulting in more False Negatives and False Positives.

Another limitation lies in the discrepancy between our experimental dataset and real-world field conditions. The Scolytinae pest dataset used in this study was primarily collected in controlled environments with clean backgrounds and relatively simple visual features. However, in actual field conditions, images often contain cluttered and diverse backgrounds such as soil, plant debris, and shadows. Additionally, pests frequently appear in high-density clusters with overlap. As a result, When applied to more complex background environments, our model may experience a decrease in detection accuracy, along with an increase in false positives and false negatives, thereby limiting its effectiveness in real-world applications.

In future work, we plan to enhance the model’s adaptability to complex environmental conditions by supplementing the dataset with more field-acquired imagery. In addition, we aim to further improve the detection accuracy of small pests and reduce False Negatives. Beyond methodological improvements, we also envision deploying the proposed GIWT-YOLO model onto edge devices such as the NVIDIA Jetson Nano. To further reduce model complexity, we plan to explore channel pruning and knowledge distillation techniques. Integrating lightweight detection models with embedded platforms holds great potential for the future development of scalable pest monitoring systems.

5 Conclusions

Achieving a balance between accuracy and lightweight design has been challenging for previous studies. In this work, we constructed a GIConv module to enhance multi-scale feature extraction capabilities, particularly for small object pests. This module significantly improves detection performance while reducing the computational cost to only 60% of that of standard convolutional modules, thus achieving a better balance between accuracy and efficiency. Additionally, to address the challenges of detecting pests with similar colors and textures, a WTConv module inspired by wavelet transform was introduced into the C3K2 module. So as to expand the effective receptive field with only a small amount of parameters, improve the detection accuracy of pests with similar color and texture, and further reduce the amount of parameters and calculation. Furthermore, considering the morphological similarities among Scolytinae pests species, an SE attention mechanism was incorporated between the neck and head components of the network. This enhancement increases the model’s focus on key feature regions, enabling more accurate discrimination between similar pest classes. As a result, the proposed GIWT-YOLO model achieved a Precision of 84.7%, Recall of 82.2%, mAP@50 of 88.7%, and mAP@50–95 of 63.4%, with only 18.7 GFLOPs and a final model size of 17.1 MB. Compared to the baseline YOLOv11s model, GIWT-YOLO improves Precision and mAP@50 by 2.2% and 4%, respectively, while reducing the parameter count, computation, and model size by 11.3%, 13.4%, and 10.9%. These results demonstrate that GIWT-YOLO effectively balances detection accuracy and model complexity, offering a more accurate and efficient solution for forestry pest monitoring applications.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found in the article/Supplementary Material.

Author contributions

JL: Funding acquisition, Validation, Formal Analysis, Writing – original draft, Conceptualization, Methodology, Investigation, Data curation. YL: Resources, Supervision, Data curation, Methodology, Writing – review & editing, Conceptualization. LW: Formal Analysis, Writing – review & editing, Investigation. YZ: Writing – review & editing, Supervision. BM: Resources, Writing – review & editing. PW: Writing – review & editing, Validation.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This work is supported by the National Science and Technology Major Special Project of New Generation Artificial Intelligence “Demonstration of Integrated Application of Key Technologies in Smart Farm(No. 2022ZD0115805).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/finsc.2025.1635439/full#supplementary-material

References

1. Zhao Z, Yang M, Yang L, Yuan Q, Chi X, and Liu W. Predicting the spread of forest diseases and pests. IEEE Access. (2020) 8:199803–12. doi: 10.1109/Access.6287639

Crossref Full Text | Google Scholar

2. Zheng J, Xu Y, Zhang H, Zhou H, and Li Q. Advances and prospects of target recognition techniques for forest pest control at home and abroad. Scientia Silvae Sinicae. (2023) 59:152–66. doi: 10.11707/j.1001-7488.LYKX20210996

Crossref Full Text | Google Scholar

3. Zhao Y, Liu Y, Ye Q, and Zhou X. Forestry pest detection based on deep learning. Chin J Liquid Crystals Displays. (2022) 37:1216–27. doi: 10.37188/CJLCD.2022-0077

Crossref Full Text | Google Scholar

4. Zuo Y. Pest recognition system based on deep learning. Beijing, China: Beijing Forestry University (2018). doi: 10.26949/d.cnki.gblyu.2018.000148

Crossref Full Text | Google Scholar

5. Wang Z, Wang K, Zhang S, Liu Z, and Mu C. Whiteflies counting with K-means clustering and ellipse fitting. Trans Chin Soc Agric Eng. (2014) 30:105–12. doi: 10.3969/j.issn.1002-6819.2014.01.014

Crossref Full Text | Google Scholar

6. Deng L, Wang Y, Han Z, and Yu R. Research on insect pest image detection and recognition based on bio-inspired methods. Biosyst Eng. (2018) 169:139–48. doi: 10.1016/j.biosystemseng.2018.02.008

Crossref Full Text | Google Scholar

7. Ma P, Zhou A, Yao Q, Yang B, Tang J, and Pan X. Influence of image features and sample sizes on rice pest identification. Chin J Rice Sci. (2018) 32:405. doi: 10.16819/j.1001-7216.2018.7116

Crossref Full Text | Google Scholar

8. Yang X, Liu M, Xu J, Zhao L, Wei S, Li W, et al. Image segmentation and recognition algorithm of greenhouse whitefly and thrip adults for automatic monitoring device. Trans Chin Soc Agric Eng. (2018) 34:164–70. doi: 10.11975/j.issn.1002-6819.2018.01.022

Crossref Full Text | Google Scholar

9. Peng W and Jinlan L. Recognition of Empoasca Flavescens based on PCA-LDA-SVM algorithm. J Chin Agric Mechanization. (2024) 45:295. doi: 10.13733/j.jcam.issn.2095-5553.2024.01.040

Crossref Full Text | Google Scholar

10. Bochkovskiy A, Wang C-Y, and Liao H-YM. Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934. (2020). doi: 10.48550/arXiv.2004.10934

Crossref Full Text | Google Scholar

11. Jocher G, Stoken A, Borovec J, Changyu L, Hogan A, Diaconu L, et al. ultralytics/yolov5: v3. 0. Geneva, Switzerland: Zenodo (2020). doi: 10.5281/zenodo.3983579

Crossref Full Text | Google Scholar

12. Li C, Li L, Jiang H, Weng K, Geng Y, Li L, et al. YOLOv6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976. (2022). doi: 10.48550/arXiv.2209.02976

Crossref Full Text | Google Scholar

13. Khanam R and Hussain M. Yolov11: An overview of the key architectural enhancements. arXiv 2024 arXiv preprint arXiv:2410.17725. (2024). doi: 10.48550/arXiv.2410.17725

Crossref Full Text | Google Scholar

14. Varghese R and Sambath M. (2024). Yolov8: A novel object detection algorithm with enhanced performance and robustness, in: 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS). Chennai, India: IEEE. pp. 1–6. doi: 10.1109/ADICS58448.2024.10533619

Crossref Full Text | Google Scholar

15. Wang A, Chen H, Liu L, Chen K, Lin Z, and Han J. Yolov10: Real-time end-to-end object detection. Adv Neural Inf Process Syst. (2024) 37:107984–8011. doi: 10.48550/arXiv.2405.14458

Crossref Full Text | Google Scholar

16. Wang C-Y, Yeh I-H, and Mark Liao H-Y. (2024). Yolov9: Learning what you want to learn using programmable gradient information, in: European conference on computer vision. Cham: Springer Nature Switzerland. pp. 1–21. doi: 10.1007/978-3-031-72751-1_1

Crossref Full Text | Google Scholar

17. Zhong Y, Gao J, Lei Q, and Zhou Y. A vision-based counting and recognition system for flying insects in intelligent agriculture. Sensors. (2018) 18:1489. doi: 10.3390/s18051489

PubMed Abstract | Crossref Full Text | Google Scholar

18. Bai Z, Tang Z, Diao L, Lu S, Guo X, Zhou H, et al. Video target detection of East Asian migratory locust based on the MOG2-YOLOv4 network. Int J Trop Insect Sci. (2022) 42(1):793–806. doi: 10.1007/s42690-021-00602-8

Crossref Full Text | Google Scholar

19. Wang B. Identification of crop diseases and insect pests based on deep learning. Sci Programming. (2022) 2022:9179998. doi: 10.1155/2022/9179998

Crossref Full Text | Google Scholar

20. Zhang B, Zhang M, and Chen Y. Crop pest identification based on spatial pyramid pooling and deep convolution neural network. Trans Chin Soc Agric Eng. (2019) 35:209–15. doi: 10.11975/j.issn.1002-6819.2019.19.025

Crossref Full Text | Google Scholar

21. Bhatt PV, Sarangi S, and Pappula S. Detection of diseases and pests on images captured in uncontrolled conditions from tea plantations. In: Autonomous air and ground sensing systems for agricultural optimization and phenotyping IV. Baltimore, MD, United States: SPIE (2019). p. 73–82. doi: 10.1117/12.2518868

Crossref Full Text | Google Scholar

22. Liu J and Wang X. Tomato diseases and pests detection based on improved Yolo V3 convolutional neural network. Front Plant Sci. (2020) 11:898. doi: 10.3389/fpls.2020.00898

PubMed Abstract | Crossref Full Text | Google Scholar

23. Wen C, Chen H, Ma Z, Zhang T, Yang C, Su H, et al. Pest-YOLO: A model for large-scale multi-class dense and tiny pest detection and counting. Front Plant Sci. (2022) 13:973985. doi: 10.3389/fpls.2022.973985

PubMed Abstract | Crossref Full Text | Google Scholar

24. Zhongzhu L, Shaojian S, and Xiuhua L. An agricultural pest detection method based on YOLOv5. Plant Prot. (2025) 51:111–22. doi: 10.16688/j.zwbh.2024089

Crossref Full Text | Google Scholar

25. Yuan W, Lan L, Xu J, Sun T, Wang X, Wang Q, et al. Smart agricultural pest detection using I-YOLOv10-SC: an improved object detection framework. Agronomy. (2025) 15:221. doi: 10.3390/agronomy15010221

Crossref Full Text | Google Scholar

26. Jiang Y and Yang X. Lightweight detection algorithm of agricultural pests based on improved YOLOv8. Comput Appl software. (2024), 1–10.

Google Scholar

27. Everingham M, Van Gool L, Williams CK, Winn J, and Zisserman A. The pascal visual object classes (voc) challenge. Int J Comput Vision. (2010) 88:303–38. doi: 10.1007/s11263-009-0275-4

Crossref Full Text | Google Scholar

28. Zhang H, Cisse M, Dauphin YN, and Lopez-Paz D. mixup: Beyond empirical risk minimization. arXiv preprint arXiv:1710.09412. (2017). doi: 10.48550/arXiv.1710.09412

Crossref Full Text | Google Scholar

29. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. (2024). Going deeper with convolutions, in: Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 1–9. doi: 10.1109/CVPR.2015.7298594

Crossref Full Text | Google Scholar

30. Han K, Wang Y, Tian Q, Guo J, Xu C, and Xu C. (2020). Ghostnet: More features from cheap operations, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 1580–9. doi: 10.1109/CVPR42600.2020.00165

Crossref Full Text | Google Scholar

31. Finder SE, Amoyal R, Treister E, and Freifeld O. (2024). Wavelet convolutions for large receptive fields, in: European Conference on Computer Vision. Cham: Springer Nature Switzerland. pp. 363–80. doi: 10.1007/978-3-031-72949-2_21

Crossref Full Text | Google Scholar

32. Hu J, Shen L, and Sun G. (2019). Squeeze-and-excitation networks, in: Proceedings of the IEEE conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 7132–41. doi: 10.1109/CVPR.2018.00745

Crossref Full Text | Google Scholar

33. Zhao Y, Lv W, Xu S, Wei J, Wang G, Dang Q, et al. (2024). Detrs beat yolos on real-time object detection, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 16965–74. doi: 10.1109/CVPR52733.2024.01605

Crossref Full Text | Google Scholar

34. Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, et al. Attention is all you need. Adv Neural Inf Process Syst. (2017) 30:5998–6008. doi: 10.48550/arXiv.1706.03762

Crossref Full Text | Google Scholar

35. Woo S, Park J, Lee J-Y, and Kweon IS. (2018). Cbam: Convolutional block attention module, in: Proceedings of the European conference on computer vision (ECCV). Cham: Springer Nature Switzerland. pp. 3–19. doi: 10.1007/978-3-030-01234-2_1

Crossref Full Text | Google Scholar

36. Hou Q, Zhou D, and Feng J. (2021). Coordinate attention for efficient mobile network design, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 13713–22. doi: 10.1109/CVPR46437.2021.01350

Crossref Full Text | Google Scholar

37. Wang Q, Wu B, Zhu P, Li P, Zuo W, and Hu Q. (2020). ECA-Net: Efficient channel attention for deep convolutional neural networks, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 11534–42. doi: 10.1109/CVPR42600.2020.01155

Crossref Full Text | Google Scholar

38. Liu Y, Shao Z, and Hoffmann N. Global attention mechanism: Retain information to enhance channel-spatial interactions. arXiv preprint arXiv:2112.05561. (2021). doi: 10.48550/arXiv.2112.05561

Crossref Full Text | Google Scholar

39. Zhou J, Roy SK, Fang P, Harandi M, and Petersson L. Cross-correlated attention networks for person re-identification. Image Vision Computing. (2020) 100:103931. doi: 10.1016/j.imavis.2020.103931

Crossref Full Text | Google Scholar

40. Neubeck A and Van Gool L. (2006). Efficient non-maximum suppression, in: 18th international conference on pattern recognition (ICPR'06). Piscataway, NJ: IEEE. pp. 850–5. doi: 10.1109/ICPR.2006.479

Crossref Full Text | Google Scholar

41. Wu X, Zhan C, Lai Y-K, Cheng M-M, and Yang J. (2019). Ip102: A large-scale benchmark dataset for insect pest recognition, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Piscataway, NJ: IEEE. pp. 8787–96. doi: 10.1109/CVPR.2019.00899

Crossref Full Text | Google Scholar

Keywords: pest detection, Scolytinae pests, lightweight model, YOLOv11s, multi-scale convolutional, effective receptive field, SE attention mechanism

Citation: Liu J, Li Y, Wang L, Zhao Y, Mao B and Wang P (2025) GIWT-YOLO: an efficient multi-scale framework for real-time Scolytinae pests detection. Front. Insect Sci. 5:1635439. doi: 10.3389/finsc.2025.1635439

Received: 26 May 2025; Accepted: 12 September 2025;
Published: 26 September 2025.

Edited by:

José Bruno Malaquias, Federal University of Paraíba, Brazil

Reviewed by:

Angelica Salustino, Federal University of Paraíba, Brazil
Magali Haidée Pereira Martínez, Federal University of Paraíba, Brazil
Adrian Fuxman, Biobest Belgium NV, Belgium

Copyright © 2025 Liu, Li, Wang, Zhao, Mao and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yongke Li, bHlrQHhqYXUuZWR1LmNu; Yunjie Zhao, enlqQHhqYXUuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.