Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 09 October 2025

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1687282

This article is part of the Research TopicSmart Plant Pest and Disease Detection Machinery and Technology: Innovations for Sustainable AgricultureView all 6 articles

A counting method of whiteflies on crop leave images captured by AR glasses based on segmentation and improved YOLOv11 models

  • 1School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Sci-Tech University, Hangzhou, China
  • 2School of Information Science and Engineering, Zhejiang Sci-Tech University, Hangzhou, China
  • 3State Key Laboratory of Rice Biology, China National Rice Research Institute, Hangzhou, China
  • 4School of Information and Control, Keyi College of Zhejiang Sci-Tech University, Shaoxing, China
  • 5Jinhe Nengyuan Technology Co., Ltd., Changzhou, China

The whitefly (Bemisia tabaci) is a globally distributed agricultural pest. While accurate monitoring of this species is crucial for early warning systems and efficient pest control, traditional manual monitoring methods suffer from subjectivity, low accuracy with large populations, and arduous data traceability. To surmount these challenges, this paper proposes an automatic counting method for whitefly adults and late-instar nymphs, based on whitefly images acquired using augmented reality (AR) glasses and a segmentation-then-detection approach. Acquired by the surveyors wearing AR glasses, the images of whiteflies on the undersides of crop leaves are transmitted to a server via Wi-Fi/5G. The system enables the automatic whitefly counting model to enumerate the adult and late-instar nymph populations, and the results can be viewed on both the AR glasses and mobile devices. The study utilizes Mask2Former-Leaf to segment the foreground primary leaf in pursuit of the minimal influence of non-primary leaf areas and background noise in the images, and detects tiny whitefly adults and late-instar nymphs in high-resolution images by involving the YOLOv11-Whitefly detection model. This model integrates Slicing Aided Hyper Inference (SAHI) capability, and can enormously amplify the feature representation of tiny objects by slicing large images through overlapping windows for both training and inference. Furthermore, DyCM-C3K2 is introduced into the YOLOv11 backbone network, which enhances the detection capability for small whitefly targets by dynamically generating input-dependent convolutional kernels, and injecting global contextual information into local convolution operations. Also, a Multi-Branch Channel Re-Weighted Feature Pyramid Network (MCRFPN) is designed to replace the original neck network, optimizing the fusion between superficial and deep features. In contrast to mainstream detection models such as YOLO, RTMDet, Cascade-CNN, DETR, and DINO, YOLOv11-Whitefly demonstrates superior performance—attaining an average recall rate of 86.20%, an average precision of 84.25%, and an mAP50 of 91.60% for whitefly adults and late-instar nymphs. With the purpose of visualizing the whitefly infestation data, this paper developed an intelligent whitefly survey system that provides on-site visualization of whitefly images integrated with their adult and late-instar nymph counting results. This facilitates surveyors in understanding pest populations and formulating scientific control decisions.

1 Introduction

The whitefly (Bemisia tabaci) refers to a globally distributed agricultural pest incurring substantial economic losses in agriculture. It poses formidable challenges to integrated pest management based on the following factors: a broad host range, the habit of damaging the undersides of crop leaves, diverse transmission pathways, rapid generation turnover, and strong pesticide resistance (Chu and Zhang, 2018). Timely and accurate assessment of whitefly population density in the field is a prerequisite for efficient pest control. According to the national technical standard NY/T 2950-2016 (2016) in China, whitefly nymphs are classified into two groups: low-instar nymphs (1st and 2nd instars) and late-instar nymphs (3rd, 4th, and pseudo-pupae). These standards dictate that only adults and late-instar nymphs are to be used as survey targets and statistical metrics. Currently, two primary methods are employed for whitefly surveys. The first is manual field surveying, where investigators physically enter fields, quickly turn over leaves, and visually estimate and manually record the number of adult and late-instar nymph whiteflies on the leaf’s underside. The second is the yellow sticky trap method, which involves using traps to attract and manually count flying adults. However, due to the tiny size of the insect, both manual identification and counting methods endure inefficient survey, myriad estimation errors (especially during peak periods), and impedance in data traceability.

At the current stage, automatic counting of tiny insect pests using yellow sticky traps is emerging as a principal research direction for bolstering the intelligence and accuracy of pest monitoring. Traditional machine vision methods for pest counting primarily lean on handcrafted features combined with machine learning classifiers for identification. For instance, (Deqin et al., 2018) developed a whitefly counting method for vegetable pests by engaging a structured random forest-based image segmentation algorithm, an irregular structural feature extraction algorithm, and sub-algorithms for interference target removal and detection. Similarly, (Zhong et al., 2018) leveraged manually extracted global features such as shape, texture, and color, in conjunction with HOG local features and Support Vector Machines. The target was to identify and classify images of yellow sticky traps containing six types of pests, including whiteflies and thrips. Despite their abilities to guarantee high recognition accuracy under specific conditions, the reliance of these methods on manual feature engineering suppresses their adaptability to intricate and variable field circumstances, resulting in poor capabilities of generalization. Thanks to the successful application of deep learning in image recognition and object detection, researchers commence with employing improved object detection models for pest identification and counting on yellow sticky traps. For example, to detect whiteflies and thrips on yellow sticky traps, (Li et al., 2021) proposed the TPest-RCNN model, marking an improvement based on Faster R-CNN. (Bai et al., 2024) put forth a multi-insect recognition framework termed MS-P2P in light of point regression by integrating YOLOv7-tiny, LAHead, and the Hungarian matching algorithm. The average F1 score for whitefly detection in their self-built cotton field yellow sticky board dataset was 80.9%. Nevertheless, the yellow sticky trap method attracts merely flying pests and is non-applicable to surveying the number of nymphs. Besides, during periods of high pest populations, the saturated sticky surface exacerbates counting accuracy, and the sticky traps compel regular replacement.

Although methods based on yellow sticky traps and computer vision have accomplished progress in the automatic counting of specific flying pests, an immense portion of pest surveys still entail manual field surveys. (de Castro Pereira et al., 2022) presented an improved YOLOv4 deep learning strategy for images of soybean leaf whiteflies captured in a laboratory setting. The strategy starts with training the model weights using cropped images, followed by image stitching to merge detection results. The model harvested an average F1 score of 0.87 for detecting whiteflies of different instars. Unfortunately, the dataset in the paper was collected under controlled laboratory conditions using single leaf images, and the model’s performance in natural field conditions still needs validation. (Bereciartua-Pérez et al., 2022) described a density map estimation method for counting whiteflies on eggplant leaves collected from Spanish greenhouses. They formulated a fully FCRN and a tailored GSP strategy to plot Gaussian density maps for precise insect localization. According to experiments, this method achieved a coefficient of determination R² of 0.97 for whitefly counting, and the model was deployed to a mobile application, providing an efficient tool for precise pest management. (Feng et al., 2024) collected whitefly images on cotton leaves. To modify a YOLOv8s-based model, they replaced its C2F module with a Swin-Transformer, then introduced a P2 branch structure in the detection head, and carried out Slicing Aided Hyper Inference (SAHI) for image preprocessing. This augmented model derived a mAP50 of 92.00% for whitefly identification, and was ultimately integrated into a Raspberry Pi edge computing terminal, offering a feasible solution for real-time field pest monitoring. Most of the aforementioned methods construct counting models based on whitefly adults on single crop leaves photographed in controlled or relatively simple scenarios. These approaches confront a confined image background that hinders the models from performing whitefly detection tasks on various crop leaves in convoluted field environments—not to mention they are normally unable to detect smaller whitefly late-instar nymphs.

To address the low efficiency and high error rates of manual field surveys, as well as the inability of yellow sticky traps to monitor nymphs, this paper proposes an automatic whitefly counting method based on augmented reality (AR) glasses and a segmentation-then-detection approach. The system leverages the advantages of AR glasses, including hands-free operation, real-time visual display, and voice control functionalities, which enables surveyors to efficiently and conveniently acquire high-resolution whitefly images in complex field environments. To prevent false positives from non-primary leaves and the influence of complex backgrounds, we utilize Mask2Former-Leaf to achieve the precise extraction of foreground primary leaves. Moreover, to achieve high-precision detection of tiny whitefly adults and late-instar nymphs in images, we designed the YOLOv11-Whitefly detection model. This model integrates SAHI capability to amplify the feature representation of tiny whiteflies by slicing large images through overlapping windows. Furthermore, the DyCM-C3K2 module and MCRFPN are introduced to enhance the model’s detection capabilities for small whitefly targets. Finally, we developed an intelligent whitefly survey system based on these core technologies. This system realizes efficient image acquisition, high-precision whitefly detection and counting, on-site result display, and data traceability. It ultimately provides surveyors with valuable insights into pest populations, thereby facilitating the formulation of scientific control decisions.

2 Materials and methods

2.1 Intelligent monitoring and survey method for whitefly

With the intent of enhancing the intelligence and precision of whitefly monitoring and surveying, surveyors wear AR glasses to quickly capture high-definition whitefly images via voice control. These images are then transmitted to a server via Wi-Fi/5G, which activates a segmentation-then-detection model to detect the whitefly adults and late-instar nymphs in the images. Finally, the detection and counting results are transmitted back and displayed on the AR glasses as well as the Web and APP terminals of the intelligent survey system for visualization and subsequent data management. The architecture of the intelligent survey method for whitefly monitoring and surveying is illustrated in Figure 1.

Figure 1
Diagram illustrating an intelligent investigation system for whiteflies. On the left, an image acquisition section shows AR glasses capturing leaf images, which are processed by application and algorithm servers. In the center, a segmentation and detection model is depicted with processes like Mask2Former and YOLOv11. The right section displays input and output images of leaves, processed by SAHI pre and post-processors. Web and app interfaces with whitefly data visualization are shown at the bottom left.

Figure 1. Intelligent monitoring and survey method for whitefly.

2.2 Image acquisition and dataset construction

This study selected the SUPERHEXA AR glasses from Superhexa Century Technology as the image acquisition device, intending to precisely capture high-definition images of tiny whiteflies on the undersides of crop leaves in the field and adapt to varying outdoor lighting conditions. The device features a 50-megapixel high-definition camera, a 3000-nits peak brightness optical display, and fast autofocus capabilities, ensuring image quality in intricate environments. Furthermore, its voice control function notably underpins the collection efficiency during field operations.

During the high-incidence seasons for whiteflies from May to October 2023 and May to October 2024, image collection was conducted in strict accordance with whitefly survey standards at the Jiangsu Academy of Agricultural Sciences and the Jinan Vegetable Research Science and Technology Park in Shandong Province. As depicted in Figure 2, with AR glasses, surveyors captured images using voice commands, while gently flipping and flattening the leaves with both hands. Across various times of day and weather conditions, a total of 5124 images of whiteflies were finally garnered on the undersides of five different crops (i.e. pepper, cotton, cucumber, eggplant and tomato), all of which were 4080×3072 pixels and included the whiteflies in both adult and late-instar nymph stages. The detailed data information is listed in Table 1. The per-image counts for these pest stages were characterized by a mean of 15.50 for adults and 5.46 for late-instar nymphs, with corresponding standard deviations of 29.15 and 24.31. The number of adults ranged from 0 to 380 per image, while the number of late-instar nymphs varied from 0 to 514. This wide range in pest density, including instances of extremely low and high populations, reflects the diverse and complex nature of whitefly distribution in field environments.

Figure 2
A person inspecting plants uses smart glasses with a camera and screen. The glasses identify whitefly adults and nymphs on a leaf. Close-ups show whitefly adults and late-instar nymphs. A device display shows detection results: fifty whitefly adults and four nymphs.

Figure 2. An example of whitefly image acquisition. (A) Surveyors wearing AR glasses and capturing whitefly images; (B) Image of adult and late-instar nymph whiteflies on leaf underside.

Table 1
www.frontiersin.org

Table 1. Whitefly pest information on the undersides of five crop leaves.

To reduce interference from complex field backgrounds and non-primary leaf whiteflies, accurate extraction is essential for the foreground primary leaf. First, we randomly sampled a portion of images from each of the five collected crops. We then used the SAM (Kirillov et al., 2023) for auxiliary annotation and performed manual correction with the Labelme tool. The annotated data was saved in JSON format, and subsequently used to construct a primary leaf semantic segmentation dataset. Subsequently, for the whitefly object detection task, we used the LabelImg tool to annotate the categories and location information of whitefly adults and late-instar nymphs on all image data from the five crops, saving the data as XML files. Both the segmentation and detection datasets were then split into training, validation, and test sets with an 8:1:1 ratio, respectively.

2.3 Whitefly counting model

In natural field environments, images of whiteflies captured using AR glasses often contain more than a single leaf, so directly applying a detection model to the entire image would detect and count the whiteflies on other non-primary leaves. This contradicts the standard on agricultural whitefly pest forecasting, which specifies that the pests should be quantified on a single leaf. To address the flaw, this paper proposes a segmentation-then-detection whitefly automatic counting method. More precisely, the Mask2Former-Leaf foreground primary leaf segmentation model was established in the first stage, which extracts the foreground primary leaf from the image, and discards other non-primary leaves and irrelevant backgrounds to prevent false positive whitefly detection. The second stage relies on an object detection method to create the YOLOv11-Whitefly detection model that detects and counts the adult and late-instar nymph whiteflies specifically on the segmented foreground primary leaf.

2.3.1 Mask2Former-leaf model

In this study, semantic segmentation is employed as a prerequisite step for our core task of whitefly detection, specifically for precisely extracting the foreground primary leaf region in convoluted field environments. The model classifies image pixels as primary leaf or background, filtering out non-primary leaves and irrelevant background interference, thus providing premium target input images for subsequent detection. Mainstream semantic segmentation models were involved for experimentation in conjunction with post-processing optimization to achieve foreground primary leaf segmentation at minimal cost.

To build our foreground primary leaf segmentation model, the study ultimately chose Mask2Former (Cheng et al., 2022). Its highly versatile architecture, which leverages a mask attention mechanism and an efficient multi-scale feature fusion strategy, is particularly well-suited for our segmentation task. To further reinforce the foreground primary leaf segmentation masks, a mask post-processing optimizer was developed that encompasses connected component analysis, morphological operations, and edge enhancement techniques to alleviate noise and rough boundaries. This three-stage process commences by performing connected component analysis to filter out small noisy regions based on a predefined area threshold, so that the main leaf body can retained; afterwards, it applies a morphological opening operation with a fixed-size structural element to smooth mask edges, aimed at removing irregularities and filling small holes for a more regular contour; finally, bilateral filtering is used to internally smooth non-edge regions and enormously raise edge clarity between the primary leaf and background. This yields a highly refined segmentation with accurate boundaries, as depicted in Figure 3.

Figure 3
Diagram showing a leaf image processing workflow. The input image of a leaf transitions to an original mask, then passes through a Mask Optimizer with three steps: small noise removal, boundary smoothing, and boundary enhancement. These steps transform the original mask into an optimized mask, leading to the final output image of the leaf.

Figure 3. Structure of the foreground primary leaf mask post-processing optimizer.

2.3.2 YOLOv11-whitefly model

YOLOv11 (Khanam and Hussain, 2024), the latest iteration in the YOLO series, marks a number of innovations across its model architecture, feature fusion, and computational efficiency. It incorporates the C3K2 module that combines depth-wise separable convolutions with residual connections, relieving computational load, and optimizing small object detection; the SPFF module dynamically weights and fuses multi-scale features, dramatically strengthening feature extraction capabilities in complicated scenarios; and the dynamic inference strategy adaptively chooses pathways based on image complexity, fostering a balance between inference speed and accuracy.

To better suit the whitefly detection task, this study tuned the YOLOv11s object detection model to launch the YOLOv11-Whitefly detection model. First, whiteflies are less than 1 mm in length, and the images in the dataset are all 4080×3072 pixels. This implies an extremely small proportion of pixels occupied by the whitefly adults and late-instar nymphs, incurring insufficiency in feature information. To conquer this problem, SAHI was integrated into both the training and inference stages of YOLOv11. The method divides the large images into overlapping sub-images using a sliding window, performs inference on each sub-image, and combines the results to obtain the final detection for the original high-resolution image. Second, given the dense and often overlapping distribution of whiteflies on leaves, coupled with interference from non-target pests such as dead whiteflies, DyCM-Conv was proposed to intensify YOLOv11’s backbone network by modifying its C3K2 module. Such a module dynamically generates convolution kernels adapting to input features, significantly underpinning the model’s ability to extract traits from tiny whiteflies. Its context-aware mixing mechanism captures long-range dependencies within images, boosting detection robustness for dense and overlapping targets, while also suppressing interference from elements like dead insects. Finally, because whitefly sizes vary drastically across images of different crop leaves, this study introduced the MCRFPN to replace the original neck network, which tailors the fusion of different scale features through its SCRF and ACRF modules, mitigating the issue of insufficient receptive fields for small objects and reinforcing the model’s multi-scale feature extraction and fusion capabilities. The structure of the YOLOv11-Whitefly model is visualized in Figure 4.

Figure 4
Diagram illustrating a model architecture for leaf disease detection. The process begins with an image input of a diseased leaf, passes through a backbone network featuring CBS, DyCM-C3k2, and SPPF modules, and progresses through a neck section called MCRFPN with C3k2 and ACRF components. The head section performs detection, leading to the output post-processed by NMS, resulting in an image of the leaf. Definitions for components like CBS, SPPF, and detection modules are included in the diagram.

Figure 4. The structure of the YOLOv11-Whitefly model.

2.3.2.1 SAHI

As per the standard on MS COCO classification, objects are categorized by absolute pixel area into the small objects less than 32×32 pixels, the medium objects between 32×32 and 96x96 pixels, and the large objects greater than 96×96 pixels. A statistical analysis of the whitefly dataset, based on these criteria, reveals that the median width and height are 24×25 pixels for whitefly adults, and 19×19 pixels for late-instar nymphs. Critically, small objects account for 72.01% and 74.21% of adult and late-instar nymphs, respectively. Given a resolution of 4080×3072 pixels for all the images, this dataset is characteristic of a high-resolution, small object detection task. The distribution of object scales and width-height distributions for each category is shown in Figure 5.

Figure 5
Two plots depicting whitefly data. The left bar chart shows the number of small, medium, and large objects for whitefly adults and nymphs. Small adults: 64,985; medium: 24,688; large: 560. Small nymphs: 22,748; medium: 7,857; large: 45. The right violin plot displays the width and height distribution of bounding box instances for both classes. Whitefly adults width: 4 to 193; height: 5 to 203. Nymphs width and height: 4 to 139 and 7 to 139, respectively.

Figure 5. Distribution of object scale counts and aspect ratios for each class.

SAHI (Akyon et al., 2022) is an enhancement strategy specifically designed for high-resolution small object detection, whose core philosophy is to divide large images into smaller sub-images via overlapping sliding windows. Each sub-image is then independently analyzed, with the results merged. This process intends to upgrade the model’s sensitivity to small targets by expanding the relative size of objects within each sub-image. The execution flow is as follows:

1. Image slicing: The input image is decomposed into multiple sub-images of fixed size S×S. To preclude edge targets from being cropped, adjacent sub-images retain a certain overlap area, with the overlap ratio controlled by an overlap ratio α.

2. Sub-image detection: Each sub-image undergoes independent inference through the detection model. For every sub-image, a set of local object detection bounding boxes, class labels, and confidence scores are generated.

3. Result merging: After mapping the sub-image detection boxes from their local coordinates to the global coordinates of the original image, the application of NMS ensues to eliminate redundant detection boxes originated from the same target in overlapping areas, while preserving the prediction results with the highest confidence.

SAHI employs a divide and conquer strategy to transform tiny objects in high-resolution images into prominent targets in sub-images, which evidently mitigates the issue of feature loss provoked by extremely small object sizes. In this study, the sub-image size is 640×640 pixels, with an overlap rate α of 0.2. During the phase of model training, images were sliced according to these parameters, and any sub-images without targets were discarded as invalid. For the inference phase, the same parameters were adopted to slice the sub-images—each detected individually, and the eventual detection outcomes were obtained by merging with NMS. The workflow is shown in Figure 6.

Figure 6
Diagram showing the process from an input image of a plant leaf to an output image. The input image is divided into grid sections, highlighting an overlap area. The post-processed output image shows the leaf with enhanced clarity, focusing on features or defects. Arrows and labels guide through the transition steps.

Figure 6. SAHI workflow in the whitefly dataset.

2.3.2.2 DyCM-C3K2

Confronting the drawbacks of the C3K2 module in conventional YOLOv11, such as poor adaptability of static convolution kernels, confined receptive fields, and paucity of feature selectivity, this paper enhances the backbone network by replacing the standard convolutional layer in the bottleneck of the C3K2 module with Dynamic Context-Mixed Convolution (DyCM-Conv). While retaining the cross-stage residual connections, DyCM-Conv dynamically gives rise to convolution kernels, allowing the model to fuse multi-scale contextual information and model long-range dependencies while preserving the inductive bias of traditional convolutions. Figure 7 unveils its structure and how it captures long-range dependencies during the convolution process. Its essential concept lies in computing the correlation between each spatial position of a feature map and multiple global region centers, which engenders a relevance matrix that designates the strength of inter-position correlations. The matrix is then learnably transformed into a dynamic convolution kernel, injecting global contextual information directly into the convolution kernel. The key implementation steps are elucidated as follows:

Figure 7
Diagram showing a complex process with two parts labeled A and B. Part A illustrates a flow of data involving reshaping, transposing, and matrix operations, with components like adaptive pooling, convolutions, and projections. Part B depicts a visual with a region center and global-aware weights interconnected by dashed lines. Annotations explain operations and symbols used.

Figure 7. (A) A schematic diagram of DyCM-Conv; (B) DyCM-Conv’s ability to capture long-range dependencies and maintain inductive bias.

1. Formation of the Relevance Matrix: In computer vision tasks, tokens are often generalized as feature units with independent semantic or contextual associations. Given an input feature map X=RC×H×W, each spatial point is treated as a token. A 1x1 convolution spawns a query (Q) that preserves spatial information. Concurrently, S×S region center adaptive average pooling and another 1x1 convolution create a key (K) that aggregates global information, followed by the channel-wise division of both Q and K into G groups. Within each group, a matrix multiplication is performed to compute a relevance matrix Rg=QgTKgRHW×SS, where the i-th row indicates the correlation between the i-th pixel and all region centers.

2. Generation of the Dynamic Convolution Kernels: Each token’s relevance matrix row is processed by a linear layer to integrate weights from an array of region centers. This output is then split to represent the properties of predefined Klarge and Ksmall convolution kernels. Thereafter, relevance values are aggregated and normalized by a direct projection which includes a learnable linear layer and a softmax function. In the end, these weighted rows are reshaped into dynamic convolution kernels of sizes Klarge×Klarge and Ksmall×Ksmall.

3. Context-Mixed Dynamic Convolution Workflow: This module operates by virtue of a grouped parallel processing approach. Each group arouses one large and one small dynamic convolution kernels, which are both shared across the channels within that group. Afterwards, the generated large and small kernels perform convolution operations on the input feature map independently. As a result, another branch processes the same original input feature map using a DWConv. Finally, the module’s final output is formed by merging the output feature maps from the large dynamic convolution kernel branch, the small dynamic convolution kernel branch, and the DWConv branch.

2.3.2.3 MCRFPN

YOLOv11 enhances multi-scale feature fusion via top-down and bottom-up pathways. However, its original neck network’s focus on high-level semantic features and simple, fixed-weight fusion mechanism often leads to insufficient integration of high-resolution features from superficial layers, causing a significant loss of crucial details for small objects. To address this, this paper proposes Multi-Branch Channel Re-Weighted Feature Pyramid Network (MCRFPN) as an alternative to the original neck network. The MCRFPN architecture can be seen in the neck portion of the YOLOv11-Whitefly model in Figure 4. The MCRFPN merges the superficial re-weighted fusion (SCRF) and the advanced re-weighted fusion (ACRF), and embeds the channel re-weighted (CRW) strategy to optimize feature fusion. Figure 8 shows the structures of SCRF, ACRF, and CRW.

Figure 8
Flowchart depicting two neural network modules: SCRF and ACRF, each with downsampling and upsampling paths. The SCRF combines feature maps and applies a CRW operation. The ACRF similarly processes inputs, with additional parallel downsampling streams. A detailed CRW section shows adaptive pooling followed by an MLP for feature refinement.

Figure 8. SCRF, ACRF, and CRW structures.

The SCRF module is embedded at the superficial connection point between the backbone and neck of the network. It enhances feature representation by fusing outputs from the same level of the backbone Pn with superficial feature outputs Pn1 and deeper outputs from the neck Pn+1'. Embedded deeper within the Neck, the ACRF module improves multi-scale gradient information interaction by fusing superficial Pn1', mid-level Pn', deep Pn+1' from the earlier part of the neck and superficial feature output Pn1'' from the end of the neck. These operations of SCRF and ACRF are expressed in (Equations 1 and 2).

Pn=CRW(Concat(Down(Pn1), Pn, Up(Pn+1')))(1)
Pn''=CRW(Concat(Down(Pn1'), Pn', Up(Pn+1'), Down(Pn1'')))(2)

Here, Pn, Pn' and Pn'' represent the n level feature outputs from the backbone network, the first half of the neck, and the second half of the neck, respectively; Up(·) signifies an upsampling operation; Down(·) denotes a downsampling operation; Concat(·) refers to concatenation along the channel dimension; and CRW(·) serves as a channel re-weighting strategy.

In a feature map with a few parameters, Channel Re-Weighting (CRW) is a feature fusion optimization method to adaptively adjust the importance of different channels, with the purpose of boosting the semantic representation capabilities of multi-scale feature fusion. To commence its implementation, GMP and GAP are used on the input feature map to capture channel-level global contextual information. The pooled results are then concatenated and fed into a lightweight fully connected network. Including two fully connected layers and an activation function layer, the network outputs a normalized weight vector that quantifies each channel’s semantic importance. Finally, this weight vector is multiplied element-wise with the original feature map to enhance the contribution of crucial channels. These operations of CRW are expressed in (Equation 3).

Xout=Xin(σ(W2·δ(W1·Concat(GMP(Xin), GAP(Xin)))))(3)

Here, Xin means the input feature map; Xout denotes the output feature map; GMP(·) and GAP(·) represent global max pooling and global average pooling respectively; W1Rc4×2c and W2Rc×c4 are learnable parameters; δ signifies an ReLU activation function; and σ refers to a Sigmoid function.

2.4 Experimental configuration

The computational framework for all experiments in this study was built on an Intel Core i9–13900 processor with a base frequency of 3.0 GHz, 128 GB of RAM, and two NVIDIA GeForce RTX 4090 GPUs. The software environment comprised the Ubuntu 24.04.2 operating system, PyTorch 2.2.0 as the deep learning framework, CUDA 11.8, and Python 3.8.17.

The semantic segmentation models were trained for 50,000 iterations with a batch size of 4, and an input image size of 2048 x 1536 pixels. The target detection models were trained for 200 epochs with momentum set to 0.937, weight decay set to 0.0005, and a batch size of 16. The initial learning rate was 0.01, using the SGD algorithm. All models were trained from scratch on our custom dataset without the use of pretrained weights from external sources.

2.5 Performance evaluation

To provide a robust and objective evaluation of our foreground primary leaf segmentation model’s performance, we utilize a comprehensive set of widely-adopted metrics in semantic segmentation. These metrics include Pixel Accuracy (PA), Intersection over Union (IoU), and the Dice coefficient (F1-score), with their formulas defined as in (Equations 4-6), respectively.

PA=TP+TNTP+TN+FP+FN(4)
IoU=|AB||AB|(5)
F1=2|AB||A|+|B|(6)

Pixel Accuracy (PA) provides a fundamental measure of overall classification performance by calculating the ratio of correctly classified pixels to the total number of pixels. The IoU measures the accuracy of the segmentation algorithm by calculating the ratio of the intersection area between the ground truth and the segmentation result to their union area. When the F1-score assesses the segmentation precision of the algorithm, the area of overlap between the segmentation result and the ground truth are computed twice, then divided by the sum of their individual areas.

With the target of objectively evaluating the detection performance of the whitefly detection model, Precision (P), Recall (R), Average Precision (AP), and mean Average Precision (mAP) were engaged as evaluation indicators, and the formulas are expressed as (Equations 7-10), respectively.

Pk=TPkTPk+FPk(7)
Rk=TPkTPk+FNk(8)
APk=01Pk(Rk)dRk(9)
mAP=k=1nAPkn(10)

Precision (P) indicates the proportion of correctly identified targets among all samples identified as that target. Recall (R) represents the proportion of correctly identified targets among all actually existing targets of that class. Average Precision (AP) evaluates model performance by integrating both precision and recall, specifically by computing the area under the precision-recall curve at various thresholds. Mean Average Precision (mAP) means the average of the AP values across all classes, serving as a crucial metric for comprehensively assessing model performance in multi-class tasks. A higher mAP suggests better detection performance across multiple categories. In this study, n signifies the total quantity of whitefly adult and late-instar nymph classes, where n = 2. Additionally, to provide a more direct assessment of the model’s counting performance, the Root Mean Square Error (RMSE) was adopted as a supplementary metric. RMSE quantifies the average magnitude of the discrepancy between the predicted counts and the true counts, offering a clear measure of the model’s counting accuracy. The formula for RMSE is given by (Equation 11), where yipred and yigt represent the predicted and ground truth counts for the i-th image, respectively, and N is the total number of images in the dataset.

RMSE=i=1N(yipredyigt)2N(11)

2.6 Intelligent monitoring and survey system for whitefly

In pursuit of automation, precision, and digital management of whitefly surveys, an intelligent survey system was developed for whitefly monitoring and surveying. This system uses a Browser/Server architecture that supports mobile applications—all established on a front-end/back-end separation design. The front-end interaction layer is developed via Vue for the web interface and Android for the mobile app, enabling functions like batch image uploads, visualization of detection results, and statistical analysis of detection data. The back-end service layer, built with the SpringBoot framework, is responsible for handling requests from both the web and mobile fronts, calling model inference interfaces, and performing data analysis. For data storage, MySQL is applied for model detection results, whereas Object Storage Service (OSS) saves original pest images and result images. Finally, the algorithm model service layer is developed based on the Flask framework to provide an automated whitefly counting interface, and is deployed to a dedicated algorithm server by virtue of Docker containerization technology.

3 Results

3.1 Comparison of foreground leaf segmentation models

To assess the performance of the Mask2Former-Leaf for foreground primary leaf segmentation, comparative experiments were conducted against mainstream semantic segmentation models, such as Fast-SCNN (Poudel et al., 2019), PSPNet (Zhao et al., 2017), DPT (Ranftl et al., 2021), Mask2Former (Cheng et al., 2022) and DeepLabv3+ (Chen et al., 2018). All models were trained and tested under identical experimental conditions. Specifically, Mask2Former was implemented with a ResNet-18 backbone, while DeepLabv3+ and PSPNet utilized a ResNetV1c-18 backbone. Fast-SCNN was implemented with its native customized backbone, and DPT employed a Vision Transformer backbone. As summarized by Table 2, the Mask2Former-Leaf model attained the best performance, with mIoU of 97.34%, mFscores of 98.65% and PA of 98.69%, respectively. This remarkably outperforms the other models, affirming its effectiveness in foreground primary leaf segmentation.

Table 2
www.frontiersin.org

Table 2. Model comparison experiment for foreground leaf segmentation task.

3.2 Ablation experiment of the detection model

In order to ascertain the effectiveness of the three improved strategies in the YOLOv11-Whitefly model, this paper tested each of them using the same test set and compared the Precision (P), Recall (R), Average Precision (AP), and the mAP50. Table 3 summarizes the results of this ablation study, where “√” implies the applied improvement strategy; “Adult” refers to whitefly adults; and “Nymph” means whitefly late-instar nymphs.

Table 3
www.frontiersin.org

Table 3. Performance of three improvement strategies in the YOLOv11-Whitefly model for whitefly detection.

As listed in Table 3, the original YOLOv11s model performed poorly in detecting whiteflies in high-resolution images, with recall rates of just 55.6% for adults and 51.3% for late-instar nymphs, and an mAP50 of 56.8%. While SAHI is a general-purpose, architecture-agnostic strategy, it serves as an indispensable baseline for our high-resolution imagery. Integrating SAHI into the YOLOv11s model prominently upgraded performance, with recall rates for adults and late-instar nymphs jumped to 91.6% and 71.3%, respectively, while precision ascended to 92.7% and 75.1%. The AP for adults and late-instar nymphs rose to 97.2% and 80.2%, arousing an overall mAP50 of 88.7%. This demonstrates that, by slicing large images into sub-images, SAHI augmented the pixel proportion and morphological features of whiteflies, allowing the model to grasp richer feature representations during both training and inference. To further demonstrate the independent contribution and synergistic effect of our internal modules, we conducted additional experiments. Introducing DyCM-Conv into the backbone network to underpin the feature extraction module further boosted performance, particularly for the smaller late-instar nymphs. The late-instar nymph recall improved to 75.6%; precision grew by 1.5%; and AP rose to 84.0%—these results confirm that DyCM-Conv dynamically generates convolution kernels that adapt to input context, and this enhances feature extraction and captures long-range dependencies, so that small object detection can be fostered. Concurrently, replacing YOLOv11’s original neck network with the MCRFPN also demonstrated a significant performance boost, intensifying the average precision for whitefly adults and late-instar nymphs to 97.1% and 83.5%, respectively, spawning an mAP of 90.3%. This experiment corroborates that MCRFPN’s multi-scale feature fusion of superficial and deep features strengthens feature information interaction, and enhances the model’s ability to detect whiteflies of varying scales on different crop leaves. The full integration of all three strategies yielded the highest performance, achieving an mAP50 of 91.6%.

3.3 Comparison of different detection models

In this paper, to evaluate the detection performance of the YOLOv11-Whitefly, comparative experiments were carried out against several prominent detection models, which involved the one-stage models YOLOv10 (Wang et al., 2024) and RTMDet (Lyu et al., 2022), the two-stage model Cascade R-CNN (Cai and Vasconcelos, 2018), and the Transformer-based end-to-end object detection models DETR (Carion et al., 2020) and DINO (Zhang et al., 2022). Specifically, we chose ResNet-50 as the backbone for Cascade R-CNN, DETR, and DINO, while YOLOv10 and RTMDet utilized their native backbone architectures. For a fair comparison, all models were trained and tested on identical datasets using the same parameter settings. Given the difficulty of detecting tiny targets in our dataset, all models were also implemented with SAHI. The results of this evaluation are exhibited in Table 4.

Table 4
www.frontiersin.org

Table 4. Performance comparison between YOLOv11-Whitefly and other detection models.

It can be seen from Table 4 that the YOLOv11-Whitefly model attained the highest mAP50, surpassing RTMDet, YOLOv10s, Cascade R-CNN, DETR, and DINO by 5.45%, 4.68%, 10.35%, 8.56%, and 8.92%, respectively. For a more intuitive comparison, Figure 9 displays localized magnified views of the whitefly detection results on various crop leaves generated by the aforementioned models. As shown in Figure 9B-D), traditional single-stage and two-stage detection models exhibit subpar performance in detecting whitefly late-instar nymphs, instigating noticeable missed detection. Concurrently, Figure 9E, F reflect that Transformer-based end-to-end object detection models generate numerous redundant prediction boxes, provoking a pronounced elevation in false positive rates. In contrast, the YOLOv11-Whitefly model proposed in this paper demonstrates superior robustness and stronger detection capabilities in intricate backgrounds, and lessens both missed detection and false positives.

Figure 9
Seven rows labeled A to G display close-up images of leaf surfaces with varying degrees of markings. Each row is segmented into multiple columns, showing detailed views of spots and rectangular outlines in colors like orange, blue, and purple, indicating areas of interest or analysis on the leaves.

Figure 9. Examples of whitefly detection results on crop leaves using different models. (A) Manual annotation; (B) RTMDet; (C) YOLOv10s; (D) Cascade-RCNN; (E) DETR; (F) DINO; (G) YOLOv11-Whitefly (the proposed); Red and blue boxes indicate whitefly adults and late-instar nymphs, respectively; Purple and orange dashed boxes designate model misses and misdetects, respectively.

To provide a more practical evaluation of the models’ final counting performance in real-world applications, we used the aforementioned detection models to count whiteflies on the original large images from the test set. For this experiment, all models were inferred with a confidence threshold of 0.3 and an IoU threshold of 0.5, and the SAHI method was applied to all predictions. The RMSE was calculated by comparing the results against the ground truth annotations. The RMSE for RTMDet, YOLOv10s, Cascade R-CNN, DETR, and DINO were 7.32, 7.07, 12.58, 9.74, and 10.51, respectively. Our model demonstrated superior counting accuracy, achieving the lowest overall RMSE of 4.49. A more granular analysis reveals that this high precision holds for both adult and late-instar nymph stages, achieving an RMSE of 3.99 for adult whiteflies and 1.39 for late-instar nymphs. This substantiates that our method’s high detection performance directly translates to a minimal counting error, which is a critical advantage for practical pest population surveys.

3.4 YOLOv11-whitefly model generalization experiment

To assess the cross-crop generalization capability of the YOLOv11-Whitefly model, a leave-out crop strategy was employed. Specifically, for this experiment, the model was trained exclusively on whitefly images from pepper, eggplant, and cucumber, ensuring that all images of cotton and tomato were strictly excluded from the training set. Subsequently, the model was evaluated on a complete test set comprising all five crops, with the test results summarized in Table 5. The model performed exceptionally on the crops it was trained on, and the mAP50 scores were 92.4% for pepper, 90.5% for eggplant, and 93.2% for cucumber. Crucially, the model also demonstrated a commendable detection ability on cotton and tomato, which was completely unexposed during the training stage. The model achieved 86.1% and 88.7% mAP50 on cotton and tomato crops, respectively. These results substantiate that the YOLOv11-Whitefly detection model is able to transfer its learning, and reliably detect whiteflies on unknown crop leaves, showcasing advantageous performance for cross-crop generalization.

Table 5
www.frontiersin.org

Table 5. Performance of the YOLOv11-Whitefly model for detecting leaf whitefly in five different crops.

3.5 Visualization of the intelligent survey system for whitefly

The functionalities of the intelligent survey system for whitefly pests designed in this study are illustrated in Figure 10, which encompass batch image upload for detection, image list display, and data analysis. Users may upload whitefly images in batches via the Web or mobile APP client interface. A whitefly counting interface call is sent by the backend server to the algorithm model, which sequentially processes the images through the Mask2Former-Leaf and YOLOv11-Whitefly models. The results are then returned to the back-end to undergo data persistence and encapsulation before being sent back to the front-end for final display.

Figure 10
Four screenshots from a Chinese agricultural app. A: Interface for entering plant survey details, showing data fields and image upload options. B: Image list with details about plant inspections, including dates and types of pests. C: Gallery view displaying multiple photos of plant leaves. D: Data visualization page with graphs and a table summarizing agricultural inspections and data points over time.

Figure 10. Interfaces of intelligent survey system for whitefly. (A) Intelligent survey interface in APP; (B) Results display interface in APP; (C) Image query interface in Web; (D) Whitefly data analysis interface in Web.

4 Discussion

Benefited from the broadly permeated adoption of high-resolution image acquisition devices, the surging advancements in deep learning algorithms, and the bolstered accessibility and cost-effectiveness of computational resources, this paper focuses on the task of whitefly survey on crop leaves in natural field environments. An automatic whitefly counting method is put forth based on a segmentation-then-detection approach, utilizing AR glasses to acquire images from five types of crop leaves. Hence, an appropriate intelligent survey system has been developed, delivering a smart, efficient, and precise solution for monitoring and surveying the agricultural pest.

High-quality image data underlies the deep learning algorithms. From fixed cameras (Gao et al., 2025), smartphones (Xu et al., 2024), to professional cameras (Li et al., 2022; Ciampi et al., 2023), and custom-built pole-mounted photographic devices (Du et al., 2022; Li et al., 2023), a variety of portable, high-resolution image acquisition devices have been prevailing across numerous fields such as agricultural pest and disease detection. However, these machines confront some drawbacks. While fixed cameras save labor, their monitoring capabilities are confined to specific areas. Handheld devices (e.g. smartphones and cameras) offer solutions for manual field surveys, but are often perplexed by factors like strong illumination, fluctuating resolutions, and the necessity for manual operation. In contrast, AR glasses have been successfully applied to agricultural pest and disease image acquisition in studies by (Sheng et al., 2024; Chen et al., 2025; Ye et al., 2025), thanks to their hands-free operation, real-time visual display, and voice control functionalities. Therefore, this paper opted for the SUPERHEXA monocular AR glasses from Superhexa Century Technology for image acquisition, which is designed with a 50-megapixel main camera and a 3000-nits peak brightness optical display, and supports rapid autofocus and voice-controlled photography. By wearing these AR glasses, surveyors can utilize both hands to gently turn over and flatten leaves, which ensures a clear and stable view of the whiteflies for subsequent image capture. They also enable surveyors to view and adjust the content in real-time on the optical display, and directly trigger image capture via voice commands, so that the operational efficiency can be considerably enhanced. The AR glasses selected in this study are distinguished by powerful image acquisition capabilities and high imaging quality, capable of fulfilling the demands for automated counting of tiny whiteflies. Nevertheless, their detection accuracy for late-instar nymphs still compels improvement, which necessitates future work involving the AR glasses equipped with higher-resolution cameras. Another flaw lies in the AR glasses’ relatively small optical screen which could be replaced by a larger, higher-resolution display to ensure easier capturing of high-quality images.

Given that the whitefly agricultural pest monitoring standard NY/T 2950-2016 (2016) requires the counting of pests on single leaves, yet the images garnered by AR glasses in natural field environments often contain more than one leaf, thus precise extraction of the foreground primary leaf from the image implies immense importance for accurate whitefly enumeration. In recent years, semantic segmentation methods have demonstrated prominent merits in the field of crop disease detection. For instance, (Wang et al., 2021) utilized DeepLabV3+ and U-Net models for segmenting diseased areas on cucumber leaves in complex backgrounds; (Zhang et al., 2023) proposed the Locally Reversible Transformer model for grape leaf disease segmentation in natural scenes; and (Ye et al., 2025) introduced the DeepLab-Leafminer model for segmenting leafminer damage regions. Inspired by these accomplishments, this paper harnesses the Mask2Former model combined with a mask post-processing optimizer to precisely segment the foreground primary leaf from whitefly images across various crop types in natural environments. This initial segmentation stage effectively attenuates interference from intricate backgrounds, and minimizes false positives arising from whiteflies not located on the primary leaf, so that a cleaner and more focused region for subsequent detection is provided.

In the domain of tiny agricultural pest detection, several advanced object detection models marking notable progress have been presented in existing research, as exemplified by YOLO-Pest (Xiang et al., 2023), DRB (Jiao et al., 2022), MDM (Wang et al., 2020), and Cascade-RCNN-PH (Sheng et al., 2024). To enable precise detection of tiny whitefly adults and late-instar nymphs in the foreground primary leaf, this paper introduces the DyCM-C3K2 module based on the YOLOv11 model, so as to construct the YOLOv11-Whitefly model by reinforcing the backbone network, substituting the original neck network with MCRFPN, and merging SAHI capabilities. Notwithstanding that the segmentation-then-detection method for whitefly counting realizes high-accuracy detection on various crop leaves, the sequential execution of two independent deep learning models inevitably deteriorates the model’s computational load and inference time. To address this, future efforts could be invested in a more comprehensive exploration of model lightweighting strategies. In addition to model pruning (Sun et al., 2023) and knowledge distillation (Hsieh et al., 2023), techniques such as model quantization could be employed to reduce the computational complexity and accelerate inference speed without compromising the high level of feature extraction and detection performance. Furthermore, a crucial step for improving throughput would be to deploy the model directly onto edge devices, such as the AR glasses themselves, thereby eliminating network latency and ensuring a more responsive system for on-site applications.

Despite the fruits harvested in whitefly detection in this study, a number of defects perpetuate. Our dataset reflects a class imbalance with adults being more prevalent than late-instar nymphs, which is a direct consequence of the inherent challenges of professional field data acquisition where pest populations naturally exhibit a non-uniform and imbalanced distribution. This also resulted in substantial variability in pest density, as reflected by high standard deviations and a wide range of counts, from low-density scenarios to high-density clusters. To mitigate these issues and further enhance model performance, a key focus for future endeavors will be on data expansion. This involves collecting a more balanced dataset with additional nymph-rich images and employing more systematic sampling protocols to ensure a more uniform representation of pest populations across various densities. Furthermore, we will incorporate images from a wider variety of host crops to improve the model’s generalization capabilities. While the present study focuses exclusively on whitefly (Bemisia tabaci), a dominant species in our data collection areas, agricultural practices indicate that greenhouse whitefly (Trialeurodes vaporariorum) is also an important pest that frequently co-occurs in other regions. The morphological similarity presents a significant challenge for machine vision-based discrimination. Accordingly, a crucial direction for future work is to develop a robust model capable of differentiating between these two species, which will substantially enhance the model’s practical utility in regions where they co-exist.

5 Conclusions

Aimed at efficiently surveying whitefly populations on crop leaves in natural field environments, this study proposes an automatic whitefly counting method based on AR glasses and a segmentation-then-detection model. Acquired by the voice-control of the surveyors wearing AR glasses, the images of whiteflies on the undersides of crop leaves are transmitted to a server that activates an automatic whitefly detection model to count the insect population. In pursuit of the minimal interference from non-primary leaf areas and intricate field backgrounds, Mask2Former-Leaf is engaged to segment the foreground primary leaf in the images. For the sake of accurately detecting tiny whitefly adults and late-instar nymphs in the images, SAHI was integrated into the YOLOv11 model, and enhanced by incorporating DyCM-C3K2 and MCRFPN. The final model attained an average precision of 84.25% and an average recall rate of 86.20% for whitefly detection, apart from an mAP50 of 91.60%. Furthermore, the Web/APP client of the intelligent survey system for whitefly monitoring and surveying can synchronously display whitefly images and their counts. Its data analysis features help surveyors understand whitefly occurrences and elaborate feasible prevention and control strategies. The proposed method presented can also be extended to forecast and survey cotton aphid, thrips and other agricultural pests in natural fields, thereby advancing the intelligent development of agricultural pest surveys.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

SZ: Validation, Data curation, Visualization, Methodology, Investigation, Conceptualization, Writing – review & editing, Writing – original draft. JLy: Visualization, Investigation, Conceptualization, Writing – review & editing. SL: Investigation, Funding acquisition, Writing – review & editing, Data curation. ZF: Formal Analysis, Methodology, Writing – review & editing. HL: Resources, Investigation, Writing – review & editing. JJ: Software, Data curation, Writing – review & editing. ZN: Investigation, Software, Writing – review & editing. BY: Resources, Investigation, Writing – review & editing. QY: Writing – review & editing, Methodology, Funding acquisition, Validation, Conceptualization. JLu: Investigation, Writing – review & editing, Supervision.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. The work was supported by the Zhejiang Provincial Natural Science Foundation of China (Grant No. LTGN24C140007) and the “San Nong Jiu Fang” Sciences and Technologies Cooperation Project of Zhejiang Province, China (Grant No. 2024SNJF010).

Conflict of interest

Authors HL was employed by Jinhe Nengyuan Technology Co., Ltd.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Akyon, F. C., Altinuc, S. O., and Temizel, A. (2022). “Slicing aided hyper inference and fine-tuning for small object detection,” in 2022 IEEE international conference on image processing (ICIP). 966–970 (Bordeaux, France: IEEE). doi: 10.1109/ICIP46576.2022.9897990

Crossref Full Text | Google Scholar

Bai, M., Chen, T., Yuan, J., Zhou, G., Wang, J., and Jia, Z. (2024). A point-based method for identification and counting of tiny object insects in cotton fields. Comput. Electron. Agric. 227, 109648. doi: 10.1016/j.compag.2024.109648

Crossref Full Text | Google Scholar

Bereciartua-Pérez, A., Gómez, L., Picón, A., Navarra-Mestre, R., Klukas, C., and Eggers, T. (2022). Insect counting through deep learning-based density maps estimation. Comput. Electron. Agric. 197, 106933. doi: 10.1016/j.compag.2022.106933

Crossref Full Text | Google Scholar

Cai, Z. and Vasconcelos, N. (2018). “Cascade r-cnn: Delving into high quality object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. 6154–6162. doi: 10.48550/arXiv.1712.00726

Crossref Full Text | Google Scholar

Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., and Zagoruyko, S. (2020). “End-to-end object detection with transformers,” in European conference on computer vision. 213–229 (Glasgow, UK: Springer). doi: 10.1007/978-3-030-58452-8_13

Crossref Full Text | Google Scholar

Chen, X., Yang, B., Liu, Y., Feng, Z., Lyu, J., Luo, J., et al. (2025). Intelligent survey method of rice diseases and pests using AR glasses and image-text multimodal fusion model. Comput. Electron. Agric. 237, 110574. doi: 10.1016/j.compag.2025.110574

Crossref Full Text | Google Scholar

Chen, L.-C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018). “Encoder-decoder with atrous separable convolution for semantic image segmentation,” in Proceedings of the European conference on computer vision (ECCV). 801–818. doi: 10.48550/arXiv.1802.02611

Crossref Full Text | Google Scholar

Cheng, B., Misra, I., Schwing, A. G., Kirillov, A., and Girdhar, R. (2022). “Masked-attention mask transformer for universal image segmentation,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 1290–1299. doi: 10.1109/CVPR52688.2022.00135

Crossref Full Text | Google Scholar

Chu, D. and Zhang, Y. (2018). Research progress on the damages and management of Bemisia tabaci (Gennadius) in China over the past 10 years. Plant Prot. 44, 51–55. doi: 10.16688/j.zwbh.2018321

Crossref Full Text | Google Scholar

Ciampi, L., Zeni, V., Incrocci, L., Canale, A., Benelli, G., Falchi, F., et al. (2023). A deep learning-based pipeline for whitefly pest abundance estimation on chromotropic sticky traps. Ecol. Inf. 78, 102384. doi: 10.1016/j.ecoinf.2023.102384

Crossref Full Text | Google Scholar

de Castro Pereira, R., Hirose, E., de Carvalho, O. L. F., da Costa, R. M., and Borges, D. L. (2022). Detection and classification of whiteflies and development stages on soybean leaves images using an improved deep learning strategy. Comput. Electron. Agric. 199, 107132. doi: 10.1016/j.compag.2022.107132

Crossref Full Text | Google Scholar

Deqin, X., Yukang, Z., Meihong, F., Chunhua, P., Yaowen, Y., and Jiahao, C. (2018). Vegetable pest counting algorithm based on visual perception. Nongye Jixie Xuebao/Transact. Chin. Soc. Agric. Machinery 49, 51–58. doi: 10.6041/j.issn.1000-1298.2018.03.006

Crossref Full Text | Google Scholar

Du, J., Liu, L., Li, R., Jiao, L., Xie, C., and Wang, R. (2022). Towards densely clustered tiny pest detection in the wild environment. Neurocomputing 490, 400–412. doi: 10.1016/j.neucom.2021.12.012

Crossref Full Text | Google Scholar

Feng, Z., Wang, N., Jin, Y., Cao, H., Huang, X., Wen, S., et al. (2024). Enhancing cotton whitefly (Bemisia tabaci) detection and counting with a cost-effective deep learning approach on the Raspberry Pi. Plant Methods 20, 161. doi: 10.1186/s13007-024-01286-0

PubMed Abstract | Crossref Full Text | Google Scholar

Gao, Y., Yin, F., Hong, C., Chen, X., Deng, H., Liu, Y., et al. (2025). Intelligent field monitoring system for cruciferous vegetable pests using yellow sticky board images and an improved Cascade R-CNN. J. Integr. Agric. 24, 220–234. doi: 10.1016/j.jia.2024.06.017

Crossref Full Text | Google Scholar

Hsieh, C.-Y., Li, C.-L., Yeh, C.-K., Nakhost, H., Fujii, Y., Ratner, A., et al. (2023). Distilling step-by-step! outperforming larger language models with less training data and smaller model sizes 02301. doi: 10.48550/arXiv.2305.02301

Crossref Full Text | Google Scholar

Jiao, L., Li, G., Chen, P., Wang, R., Du, J., Liu, H., et al. (2022). Global context-aware-based deformable residual network module for precise pest recognition and detection. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.895944

PubMed Abstract | Crossref Full Text | Google Scholar

Khanam, R. and Hussain, M. (2024). Yolov11: An overview of the key architectural enhancements 17725. doi: 10.48550/arXiv.2410.17725

Crossref Full Text | Google Scholar

Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., et al. (2023). “Segment anything,” in Proceedings of the IEEE/CVF international conference on computer vision. 4015–4026. doi: 10.48550/arXiv.2304.02643

Crossref Full Text | Google Scholar

Li, H., Liang, Y., Liu, Y., Xian, X., Xue, Y., Huang, H., et al. (2023). Development of an intelligent field investigation system for Liriomyza using SeResNet-Liriomyza for accurate identification. Comput. Electron. Agric. 214, 108276. doi: 10.1016/j.compag.2023.108276

Crossref Full Text | Google Scholar

Li, W., Wang, D., Li, M., Gao, Y., Wu, J., and Yang, X. (2021). Field detection of tiny pests from sticky trap images using deep learning in agricultural greenhouse. Comput. Electron. Agric. 183, 106048. doi: 10.1016/j.compag.2021.106048

Crossref Full Text | Google Scholar

Li, R., Wang, R., Xie, C., Chen, H., Long, Q., Liu, L., et al. (2022). A multi-branch convolutional neural network with density map for aphid counting. Biosyst. Eng. 213, 148–161. doi: 10.1016/j.biosystemseng.2021.11.020

Crossref Full Text | Google Scholar

Lyu, C., Zhang, W., Huang, H., Zhou, Y., Wang, Y., Liu, Y., et al. (2022). Rtmdet: An empirical study of designing real-time object detectors 07784. doi: 10.48550/arXiv.2212.07784

Crossref Full Text | Google Scholar

NY/T 2950-2016 (2016). Technical Specification for Investigation and Forecast of Bemisia tabaci (Gennadius) (China: Ministry of Agriculture of the People’s Republic of China).

Google Scholar

Poudel, R. P., Liwicki, S., and Cipolla, R. (2019). Fast-scnn: Fast semantic segmentation network 04502. doi: 10.48550/arXiv.1902.04502

Crossref Full Text | Google Scholar

Ranftl, R., Bochkovskiy, A., and Koltun, V. (2021). “Vision transformers for dense prediction,” in Proceedings of the IEEE/CVF international conference on computer vision. 12179–12188. doi: 10.48550/arXiv.2103.13413

Crossref Full Text | Google Scholar

Sheng, H., Yao, Q., Luo, J., Liu, Y., Chen, X., Ye, Z., et al. (2024). Automatic detection and counting of planthoppers on white flat plate images captured by AR glasses for planthopper field survey. Comput. Electron. Agric. 218, 108639. doi: 10.1016/j.compag.2024.108639

Crossref Full Text | Google Scholar

Sun, M., Liu, Z., Bair, A., and Kolter, J. Z. (2023). A simple and effective pruning approach for large language models 11695. doi: 10.48550/arXiv.2306.11695

Crossref Full Text | Google Scholar

Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., and Han, J. (2024). Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 37, 107984–108011. doi: 10.48550/arXiv.2503.07465

Crossref Full Text | Google Scholar

Wang, C., Du, P., Wu, H., Li, J., Zhao, C., and Zhu, H. (2021). A cucumber leaf disease severity classification method based on the fusion of DeepLabV3+ and U-Net. Comput. Electron. Agric. 189, 106373. doi: 10.1016/j.compag.2021.106373

Crossref Full Text | Google Scholar

Wang, F., Wang, R., Xie, C., Yang, P., and Liu, L. (2020). Fusing multi-scale context-aware information representation for automatic in-field pest detection and recognition. Comput. Electron. Agric. 169, 105222. doi: 10.1016/j.compag.2020.105222

Crossref Full Text | Google Scholar

Xiang, Q., Huang, X., Huang, Z., Chen, X., Cheng, J., and Tang, X. (2023). YOLO-pest: An insect pest object detection algorithm via CAC3 module. Sensors 23, 3221. doi: 10.3390/s23063221

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, W., Yang, R., Karthikeyan, R., Shi, Y., and Su, Q. (2024). GBiDC-PEST: A novel lightweight model for real-time multiclass tiny pest detection and mobile platform deployment. J. Integr. Agricult. 24, 2749–2769 doi: 10.1016/j.jia.2024.12.017

Crossref Full Text | Google Scholar

Ye, Z., Liu, Y., Ye, F., Li, H., Luo, J., Guo, J., et al. (2025). Automatic diagnosis of agromyzid leafminer damage levels using leaf images captured by AR glasses. J. Integr. Agric. 24, 1–13. doi: 10.1016/j.jia.2025.02.008

Crossref Full Text | Google Scholar

Zhang, X., Li, F., Jin, H., and Mu, W. (2023). Local Reversible Transformer for semantic segmentation of grape leaf diseases. Appl. Soft Comput. 143, 110392. doi: 10.1016/j.asoc.2023.110392

Crossref Full Text | Google Scholar

Zhang, H., Li, F., Liu, S., Zhang, L., Su, H., Zhu, J., et al. (2022). Dino: Detr with improved denoising anchor boxes for end-to-end object detection 03605. doi: 10.48550/arXiv.2203.03605

Crossref Full Text | Google Scholar

Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017). “Pyramid scene parsing network,” in Proceedings of the IEEE conference on computer vision and pattern recognition. 2881–2890. doi: 10.48550/arXiv.1612.01105

Crossref Full Text | Google Scholar

Zhong, Y., Gao, J., Lei, Q., and Zhou, Y. (2018). A vision-based counting and recognition system for flying insects in intelligent agriculture. Sensors 18, 1489. doi: 10.3390/s18051489

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: whitefly, AR glasses, image segmentation, object detection, automatic counting

Citation: Zhao S, Lyu J, Liu S, Feng Z, Ling H, Jiao J, Ni Z, Yang B, Yao Q and Luo J (2025) A counting method of whiteflies on crop leave images captured by AR glasses based on segmentation and improved YOLOv11 models. Front. Plant Sci. 16:1687282. doi: 10.3389/fpls.2025.1687282

Received: 17 August 2025; Accepted: 24 September 2025;
Published: 09 October 2025.

Edited by:

Fan Yang, Jiangsu Normal University, China

Reviewed by:

Mikio Kamei, Hiroshima Prefectural Office, Japan
Kuang Yingchun, Hunan Agricultural University, China

Copyright © 2025 Zhao, Lyu, Liu, Feng, Ling, Jiao, Ni, Yang, Yao and Luo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Qing Yao, cS15YW9AenN0dS5lZHUuY24=; Ju Luo, bHVvanUwMkBjYWFzLmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.