Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 25 August 2025

Sec. Plant Bioinformatics

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1616165

YOLOv10-kiwi: a YOLOv10-based lightweight kiwifruit detection model in trellised orchards

  • College of Mathematics and Computer Science, Yan’an University, Yan’an, Shaanxi, China

To address the challenge of real-time kiwifruit detection in trellised orchards, this paper proposes YOLOv10-Kiwi, a lightweight detection model optimized for resource-constrained devices. First, a more compact network is developed by adjusting the scaling factors of the YOLOv10n architecture. Second, to further reduce model complexity, a novel C2fDualHet module is proposed by integrating two consecutive Heterogeneous Kernel Convolution (HetConv) layers as a replacement for the traditional Bottleneck structure. This replacement enables parallel processing and enhances feature extraction efficiency. By combining heterogeneous kernels in sequence, C2fDualHet captures both local and global features while significantly lowering parameter count and computational cost. To mitigate potential accuracy loss due to lightweighting, a Cross-Channel Fusion Module (CCFM) is introduced in the neck network. This module incorporates four additional convolutional layers to adjust channel dimensions and strengthen cross-channel information flow, thereby enhancing multi-scale feature integration. In addition, a MPDIoU loss function is introduced to overcome the limitations of the traditional CIoU in terms of aspect ratio mismatch and bounding box regression, accelerating convergence and improving detection accuracy. Experimental results demonstrate that YOLOv10-Kiwi achieves a model size of only 2.02 MB, with 0.51M parameters and 2.1 GFLOPs, representing reductions of 80.34%, 81.11%, and 68.18%, respectively, compared to the YOLOv10n baseline. On a self-constructed kiwifruit dataset, the model achieves 93.6% mAP@50 and an inference speed of 74 FPS. ​​YOLOv10-Kiwi offers an efficient solution for automated kiwifruit detection on low-power agricultural robots.

1 Introduction

Kiwifruit is widely welcomed for its abundance of vitamin C, dietary fiber, and a variety of trace elements (Richardson et al., 2018). Meixian County in Shaanxi Province, as the core kiwi-producing region in China, has seen its industry not only boost the local economy and rural revitalization but also gain a significant position in the global market. Modern T-shape cultivation enhances orchard management, but harvesting remains labor-intensive, facing labor shortages and high physical demands (Deng et al., 2023). Automation technologies can streamline harvesting processes, improving efficiency and reducing labor dependency (Fu et al., 2024).

Traditional kiwifruit detection methods mainly rely on color features, morphological features, and machine vision techniques, using handcrafted features for fruit segmentation and identification. For example (Scarfe, 2012), combined RGB color subtraction with Sobel filters to detect kiwis and their calyx edges, using template matching for identification, but did not consider fruit shape features (Zhan et al., 2013) used the Adaboost algorithm to optimize kiwifruit segmentation and localization in field conditions. By analyzing RGB, HSV, and Lab color spaces, they selected the most discriminative channels between fruit and background to improve detection accuracy. With the rising demand for nighttime automated harvesting (Longsheng et al., 2015), proposed a machine vision method based on artificial lighting to overcome the impact of natural light variability, thereby enhancing detection stability. Overall, these traditional methods are significantly affected by lighting conditions and fruit morphology variations, resulting in limited generalization ability and poor adaptability to complex environments. Consequently, researchers have begun exploring deep learning-based approaches to overcome these limitations.

Current deep learning-based kiwi fruit detection algorithms can be broadly categorized into two-stage and one-stage detection methods. Two-stage methods, such as Faster R-CNN (Faster r-cnn: Towards real-time object detection with region proposal networks, 2015), Cascade R-CNN (Cai and Vasconcelos, 2018), and Mask R-CNN (He et al., 2017), first generate region proposals followed by classification and bounding box refinement. These models achieve high detection accuracy and robustness but are computationally intensive and slower. In contrast, one-stage detectors—such as the YOLO series (Jiang et al., 2022), SSD (Liu et al., 2016), and EfficientDet (Tan et al., 2020)—offer faster inference and lower computational cost, gradually becoming the mainstream for real-time detection. For instance (Suo et al., 2021)used YOLOv4 for multi-class kiwifruit detection, achieving a maximum mAP of 91.9% with a per-image inference time of 25.5 ms, reducing robotic misoperations (Yao et al., 2021) improved YOLOv5 by adding a small object detection layer, introducing a SE attention mechanism, and optimizing the loss function to accurately detect kiwi defects (Xia et al., 2022) integrated an attention module into YOLOv7 and employed spectral techniques to improve detection precision.

With the increasing demand for lightweight models, researchers have proposed various optimizations to the YOLO series to reduce computational load while maintaining accuracy. In the fruit detection domain, numerous efficient improvements have been made. For example (Liu et al., 2024) proposed the Faster-YOLO-AP model, which used structural simplification and lightweight convolutions to compress the parameter count to 0.66M and FLOPs to 2.29G, while retaining high apple detection accuracy (Zhao et al., 2024). introduced YOLO-Granada, which maintained accuracy close to YOLOv5s while reducing parameters, computation, and model size to 54.7%, 51.3%, and 56.3% of the original, respectively, greatly enhancing pomegranate detection efficiency. In the context of kiwifruit detection, several lightweight models have also shown promising results. For example (Fu et al., 2021) optimized YOLOv3-tiny to develop the DY3TNet model, improving detection accuracy with 3×3 and 1×1 convolutions in layers 5 and 6, and compressing weights to 27 MB with an inference time of just 34 ms per image (Gao et al., 2022) used GhostNet to replace the original CSP-Darknet53 backbone and partially substituted standard convolutional layers, significantly reducing computational load (Zhou et al., 2022). proposed an enhanced YOLOX-S model, reducing parameters by 44.8% and increasing detection speed by 63.9% through feature map simplification, activation function optimization, and loss function improvements.

Despite these advancements in model lightweighting and detection performance, challenges remain in kiwifruit detection tasks. Existing models still face high parameter counts and computational burdens. While complex feature fusion and convolution operations improve accuracy, they exacerbate computational bottlenecks, limiting deployment and real-time application on edge devices. To address this, this paper proposes a lightweight kiwifruit detection model based on YOLOv10 (Wang et al., 2024), aiming to reduce computational complexity while maintaining high detection accuracy. The main contributions are as follows:

1. Construction of a diversified kiwifruit dataset. A total of 1,280 images were collected from real orchards under various angles and scenes, covering complex lighting conditions, occlusions, and diverse morphological appearances of the fruit.

2. Proposal of the lightweight YOLOv10-Kiwi model. Based on YOLOv10n, the network is compressed via scaling factor adjustments. The lightweight C2fDualHet module is designed for feature extraction, the CCFM structure is introduced to enhance feature fusion, and the MPDIoU loss function is adopted to improve bounding box regression accuracy.

3. Validation through comparative and ablation experiments. The proposed model’s effectiveness is demonstrated by extensive experiments against mainstream detection methods.

2 Materials and methods

2.1 Dataset construction

2.1.1 Data acquisition

The image dataset used in this study was collected from July to August 2024 in a commercial kiwifruit orchard in Meixian County, Shaanxi Province. Xuxiang kiwifruit images grown under trellis conditions were captured using iPhone 13 and 14 devices at various angles (top-down, oblique, upward) and lighting conditions (front lighting, backlighting, flash at night). Sampling was conducted on sunny and cloudy days, as well as at night (with flash), to simulate realistic harvesting scenarios. Rainy conditions were excluded since harvesting typically does not occur during such weather. This summer sampling phase coincided with the fruit maturation period, ensuring consistency in fruit size and color—crucial for effective model training in robotic applications. A total of 1,280 images were collected: 828 on sunny days (411 in the morning, 417 in the afternoon), 190 under cloudy skies, and 262 at night. All images were resized to 640 × 640 pixels and reflect challenging real-world conditions, including dense occlusion, uneven lighting, clustered fruit, and complex backgrounds. Representative samples are shown in Figure 1, which illustrates these visual challenges more intuitively. This figure highlights the diversity and complexity of orchard environments, underscoring the need for robust detection under variable field conditions.

Figure 1
Nine images of kiwi fruits in different conditions and angles. (a) Sunny day, bright lighting on fruits. (b) Cloudy day, dimmer lighting. (c) Night, highly obscured. (d) Leaf obstruction, covered by leaves. (e) Fruit overlap, two touching fruits. (f) Branch obstruction, partially obscured by branches. (g) Top-down shot, view from above. (h) Low-angle shot, viewed from below. (i) Eye-level shot, straight view.

Figure 1. Images of kiwifruit under different conditions: (a) Sunny day, (b) Cloudy day, (c) Night, (d) Leaf obstruction, (e) Fruit overlap, (f) Branch obstruction, (g) Top-down shot, (h) Low-angle shot, and (i) Eye-level shot.

2.1.2 Data annotation

In this experiment, we used LabelImg software to annotate kiwi fruit images with rectangular bounding boxes, assigning a single category label “kiwi.” After the annotation process was completed, the data was saved in YOLO format as.txt files, with each line containing the class label, the normalized center coordinates (x_center, y_center), width, and height—scaled to the [0, 1] range. The dataset was divided into training, validation, and test sets in a ratio of 8:1:1. Specifically, the 1,280 images were distributed as follows: 772 images in the training set (containing 3,318 instances), 253 images in the validation set (1,010 instances), and 255 images in the test set (1,022 instances). The statistical analysis results, as shown in Figure 2, indicate variations in both the scale and distribution of the annotated bounding boxes, which are consistent with the natural growth characteristics of kiwi fruits. Figure 2a illustrates the distribution of width-to-height ratios of the bounding boxes, reflecting significant differences in the sizes of the fruits. Figure 2b displays the distribution of object center points, with color intensity representing density, showing that most objects are concentrated near the center of the images. Figure 2c presents a scatter plot of width-to-height ratios, which tend to cluster around 1, indicating that the bounding boxes are regular in shape, with similar width and height. These characteristics provide valuable data support for subsequent algorithm optimization.

Figure 2
Three plots depict boundary box distributions: (a) a series of nested blue rectangles illustrating size distribution; (b) a scatter plot of blue squares showing target center point locations with x and y coordinates; (c) a density plot with blue gradient representing aspect ratio distribution, highlighting a trend along the diagonal between width and height.

Figure 2. Distribution of bounding box sizes and label scales in the kiwifruit dataset: (a) Boundary Box Size Distribution, (b) Distribution of object center points (color intensity represents density), and (c) Scatter plot of width-to-height ratios.

2.2 Novel network construction

YOLOv10 serves as the foundation for this study, offering a well-balanced trade-off between detection accuracy and inference speed through its lightweight classification head, spatial-channel decoupled downsampling, and rank-based architectural simplifications. These structural advantages make it suitable for real-time detection tasks. However, when applied directly to kiwifruit detection in trellised orchard environments, the default YOLOv10n model reveals several limitations. Its computational load and memory usage remain relatively high for resource-constrained edge devices, and the backbone structure lacks specialized design for lightweight feature extraction under conditions such as dense occlusion, variable lighting, and scale variation. In addition, the neck network provides limited cross-channel fusion, which weakens its ability to capture fine-grained multi-scale features, especially for small or overlapping fruits. Furthermore, the use of the CIoU loss function in bounding box regression often results in suboptimal localization performance when dealing with irregular fruit shapes and aspect ratio mismatches.

To address these challenges, we propose YOLOv10-Kiwi, a lightweight and efficient detection model specifically optimized for kiwifruit detection in complex orchard conditions. As illustrated in Figure 3, the model architecture incorporates three key structural refinements. The original C2f block in the YOLOv10n backbone is replaced with the C2fDualHet module, which uses two consecutive HetConv layers to improve feature representation while significantly reducing the number of parameters and computational cost. The neck is enhanced with a Cross-Channel Fusion Module (CCFM), which strengthens the integration of multi-scale and inter-channel information. Four additional convolutional layers are incorporated to refine channel dimensions and boost inter-channel interactions, thereby improving detection robustness in the presence of occlusion and size variability. Lastly, the head adopts a decoupled structure for classification and regression, with the conventional CIoU loss replaced by the Minimum Point Distance IoU (MPDIoU) loss. This loss function improves localization accuracy and convergence stability, particularly in scenes with densely packed or partially visible targets. With these targeted improvements, YOLOv10-Kiwi achieves a compact and efficient design that maintains real-time performance while enhancing detection accuracy, making it well-suited for deployment on low-power edge devices in agricultural applications.

Figure 3
Flowchart of a neural network architecture for object detection. It includes modules like CBS, C2fDualHet, SCDown, SPPF, and PSA. The process flows from input to backbone, then to neck (CCFM), and finally to the detection head (v10Detect), outlining convolution, upsampling, and regression/classification processes. An image of a jackfruit is shown as input.

Figure 3. Schematic of the proposed lightweight network structure.

2.2.1 C2fDualHet

To further reduce the number of model parameters, this study integrates HetConv (Heterogeneous Kernel Convolution) (Singh et al., 2019) into the YOLOv10 framework. HetConv combines two kernel sizes—3×3 and 1×1—and introduces a configurable parameter P to control the proportion of channels processed by each type, achieving an effective balance between spatial representation and computational efficiency.

As shown in Figure 4, the input channels M are divided into P parts: M/P channels are processed with 3×3 convolution kernels, while the remaining channels use 1×1 kernels. In a standard convolution operation, assume the input feature map has dimensions Di×Di×M, where Di is the width and height of the input feature map, and M is the number of input channels. The output feature map has dimensions Do×Do×N, where Do is the width and height of the output feature map, and N is the number of output channels. The output feature map is obtained using N convolution kernels, each of size K×K×M, where K is the kernel size. Based on this, the computational cost of this layer can be expressed as:

Figure 4
Diagram showing three vertical structures labeled P=1, P=2, and P=4, each composed of stacked blocks. P=1 features repeated 3Ă—3 blocks. P=2 intersperses 3Ă—3 and 1Ă—1 blocks. P=4 primarily uses 1Ă—1 blocks with some 3Ă—3 blocks. Arrows indicate flow direction, labeled with M.

Figure 4. HetConv convolutional structure.

FLs=D0×D0×M×N×K×K(1)

As shown in Equation 1, the computational cost of a convolutional layer is influenced by several key factors: the kernel size K, the image dimensions Di×Di, the number of input channels M, and the number of output channels N. In traditional standard convolution, this computational cost is often quite substantial, posing challenges for computational resources. In the pursuit of more efficient convolution methods, researchers have continuously explored and developed various convolutional models. Notable among these are depthwise convolution (DWC) (Howard, 2017), pointwise convolution (PWC) (Chollet, 2017), and group convolution (GWC) (Guo et al., 2019). While these models significantly reduce computational load, they typically introduce an additional unit of latency. In contrast, HetConv not only effectively reduces computational cost but also achieves zero latency, which is especially important in applications with high real-time requirements. Figure 5 illustrates a comparison between HetConv and other convolution filters in terms of computational efficiency and latency.

Figure 5
Illustration of four types of convolution methods for neural networks: (a) Depthwise Separable Convolution with blue and orange filters labeled with dimensions, (b) Groupwise Convolution using pink and orange filters, (c) Standard Convolution with orange filters, and (d) Heterogeneous Kernel Convolution featuring blue and orange filters with size indicators like “3x3”. Each method highlights different arrangements and connections between filters. Labels include “Depthwise Conv Filters”, “Groupwise Conv Filters”, “Standard Conv Filters”, and “HetConv Filter”.

Figure 5. Comparison between HetConv and other efficient convolutional filters: (a) Standard convolution, (b) Depthwise convolution, (c) Pointwise convolution, (d) Group convolution, and (e) HetConv.

In a specific HetConv layer L, the convolution operation uses K×K kernels over P channels, with its computational cost denoted as FLK.

FLK=(D0×D0×M×N×K×K)/P(2)

For the remaining (M - M/P) channels, 1×1 kernels are applied, with a corresponding computational cost denoted as FL1.

FL1=(D0×D0×N)×(MMP)(3)

Therefore, the total computational cost of layer L can be expressed as the sum of FLK and FL1.

FLHC= FLK+FL1(4)

Compared to standard convolution, the computational reduction ratio RHC of HetConv can be formulated accordingly.

RHC=FLK+FL1FLS=1P+11/PK2(5)

The computational formulations of Equations 25 provide a quantitative basis for understanding the efficiency advantages of HetConv over standard convolution, particularly in reducing FLOPs and parameter counts through mixed kernel usage. When P = 1, HetConv degenerates into standard convolution. HetConv retains K×K kernels (e.g., 3×3) only on a portion of the channels to capture critical spatial correlations, while applying 1×1 kernels on the remaining channels to reduce computational complexity. This design preserves essential spatial information while avoiding the high computational cost associated with applying large kernels across all channels. In terms of computational efficiency and parameter reduction, HetConv significantly outperforms standard convolution. The use of 1×1 kernels greatly reduces FLOPs and the number of model parameters, making the model more lightweight and vastly more efficient than standard convolution. Thus, we construct the DualHetBlock using two consecutive HetConv modules to replace the Bottleneck in the original C2f, naming it C2fDualHet.

2.2.2 Cross-scale convolutional fusion module

In the complex natural environment of kiwifruit orchards, fruits are often affected by factors such as occlusion, uneven lighting, and clustered distribution. These challenges lead to significant variations in the morphology and scale of the targets, thereby reducing the accuracy and robustness of detection models. To enhance the model’s capability for multi-scale feature representation and fusion, an improved Cross-scale Convolutional Fusion Module (CCFM) (Zhao et al., 2024) is introduced, along with a redesigned feature fusion pathway within the network.

As shown in Figure 6, this study adopts a processing strategy for the final layer of efficient hybrid encoder (F5) similar to that illustrated in the figure. Specifically, a 1×1 convolution is added after the PSA module in YOLOv10 to reduce the number of channels from 1024 to 256, thereby decreasing the parameter count while maintaining spatial dimensions and promoting inter-channel information interaction. In the diagram, shallow features S3 and S4 are compressed via 1×1 convolutions to ensure dimensional consistency before multi-scale feature fusion.

Figure 6
Diagram of an Efficient Hybrid Encoder consisting of AIFI and CCFM modules. The AIFI module processes layers S3, S4, and S5, feeding into the CCFM. The CCFM includes fusion operations and concatenations, leading to a sequence of convolutions and activations. The process ends with an element-wise addition and flattening.

Figure 6. Architecture of the original cross-channel fusion module (CCFM).

The improved CCFM structure is illustrated in Figure 7. The input_proj branch applies a 1×1 convolution to map the output features of the backbone network, achieving channel alignment and feature reorganization. The lateral_convs branch also uses 1×1 convolutions, but omits activation and normalization layers to simplify computation and adjust feature channels. This branch then upsamples the deep features to increase spatial resolution. Once the two feature streams are dimensionally aligned, they are fused through concatenation, effectively integrating high-resolution detail with low-resolution semantic information. In this improved architecture, the RepBlock in Figure 6 is replaced by an upsampling operation, which boosts both small-object detection capability and feature map resolution, facilitating efficient multi-scale feature fusion.

Figure 7
Flowchart of a neural network architecture with two main paths. The first path shows “Conv” followed by layers labeled as k=1, s=1, p=0, BN, and SiLU. The second path includes “input_proj” to “Conv”, “lateral_convs” to another “Conv”, an “Upsample” step, and a final “Concat” step merging both paths.

Figure 7. Architecture of the improved cross-channel fusion module (CCFM).

Compared to the traditional PAN + FPN structure (Guo et al., 2020), the improved neck architecture incorporates four convolution operations for lightweight feature fusion, significantly enhancing the model’s adaptability to multi-scale targets, particularly small objects.

2.2.3 Minimum point distance IoU

The CIoU (Complete Intersection over Union) loss function used in YOLOv10 takes into account the overlap area between the ground truth and predicted boxes, the Euclidean distance between their center points, and the aspect ratio difference. This improves the model’s ability to fit object boundaries. However, in practical applications, when the predicted box and the ground truth box share the same center and aspect ratio, the aspect ratio penalty term in CIoU may degrade to zero—even if there is a significant size mismatch between the two boxes. This reduces the loss function’s sensitivity to boundary dimensions, which can negatively affect model convergence speed and final accuracy.

To address this issue, the MPDIoU (Minimum Point Distance IoU) (Ma and Xu, 2023) loss function is introduced to further enhance the performance of bounding box regression. This method minimizes the Euclidean distance between the top-left and bottom-right corner points of the predicted and ground truth boxes, thereby directly constraining both the size and position of the bounding boxes. The corresponding loss function is expressed as follows:

LMPDIou=1IoU+d12h2+w2+d22h2+w2(6)
d12=(x1pred+x1gt)2+(y1pred+y1gt)2(7)
d22=(x2pred+x2gt)2+(y2pred+y2gt)2(8)

In Equations 68, (x1pred, y1pred)and( x2pred, y2pred) represent the top-left and bottom-right coordinates of the predicted bounding box, respectively. Similarly, (x1gt, y1gt) and (x2gt, y2gt) denote the top-left and bottom-right coordinates of the ground truth box. www and hhh represent the width and height of the image.

MPDIoU builds upon the optimization of IoU and center point distance by introducing constraints based on the distances between corresponding key points. This enables a more precise measurement of the geometric discrepancy between predicted and ground truth boxes. It effectively addresses the weakness of CIoU, which may fail when aspect ratios are identical but box sizes differ. As a result, MPDIoU enhances both the accuracy of bounding box regression and the model’s convergence speed. An example of MPDIoU metric parameters is illustrated in Figure 8.

Figure 8
Close-up of a fruit on a plant with annotated coordinates and bounding boxes indicating predicted and ground truth positions. Yellow lines show diagonal measurements labeled as \(d_1\) and \(d_2\). Blue arrows denote height (\(h\)) and width (\(w\)) dimensions.

Figure 8. Illustration of MPDIoU calculation process for bounding box regression.

3 Results and analysis

3.1 Experimental setup

To ensure the fairness and validity of the experimental results, all models were evaluated on the same standardized experimental platform. The system environment was Ubuntu 20.04, equipped with an Intel® Xeon® Gold 5418Y 12-core processor and an Nvidia GeForce RTX 4090 GPU with 24 GB of memory. The software environment included Python 3.8, PyTorch 1.11, and CUDA 11.3. The training parameters were configured as follows: the initial learning rate was set to 0.01, with a batch size of 16. The SGD optimizer was used, with a weight decay coefficient of 0.0005 and a total of 150 epochs. A warmup strategy with a momentum of 0.937 was adopted to gradually increase the learning rate, using one-dimensional linear interpolation. After the warmup phase, a cosine annealing algorithm was applied to adjust the learning rate throughout the training process. Additionally, the input images were resized to a resolution of 640×640 to ensure consistency in the training data.

3.2 Performance evaluation

In the task of kiwifruit detection, algorithm evaluation primarily focuses on two aspects: computational complexity and detection accuracy. Detection accuracy is commonly measured by Precision (P), Recall (R), and Mean Average Precision (mAP). Precision reflects the model’s accuracy in detecting targets, Recall assesses the detection coverage, while mAP, as a comprehensive metric for class-level detection, provides an overall evaluation of the model’s detection performance. Computational complexity is characterized by the number of parameters (Params) and floating point operations per second (FLOPs). A smaller number of parameters indicates a more lightweight model, and lower FLOPs imply reduced computational cost, making the model more suitable for deployment on mobile or edge devices. Meanwhile, the real-time performance of the algorithm is measured by Frames Per Second (FPS); the higher the FPS, the faster the model’s inference speed, and the stronger its real-time capability. These metrics comprehensively reflect the performance of a model from the perspectives of accuracy, computational speed, and complexity, thus providing important references for algorithm optimization and practical deployment. The following is a summary of the formulas for calculating P, R, AP, mAP, Params and GFLOPs.

P=TPTP+FP×100%(9)
R=TPTP+FN×100%(10)
AP=01P(R)dR(11)
mAP=1ni1nAPi×100%(12)
GFLOPS= O(i=1nKi2Ci2Ci+i=1nM2Ci)(13)
Params=O(i=1nM2K2Ci1Ci)(14)

In Equation 9, precision P represents the proportion of correctly predicted kiwi instances among all predicted kiwi instances. Here, TP denotes the number of correctly detected kiwis, FP refers to the number of incorrectly detected kiwis, and FN indicates the number of missed (undetected) kiwis.

In Equation 10, recall R represents the proportion of correctly detected kiwi instances out of the total number of actual kiwi instances. In Equation 11, AP denotes the area under the Precision-Recall (P-R) curve. Equation 12 defines mAP as the mean value of AP across all categories.

In this study, since there is only one kiwi category, n = 1. mAP@0.5 indicates the mean average precision when the IoU threshold is set to 0.5. Additionally, in Equations 13 and 14, K represents the kernel size, C is the number of channels, M is the size of the input image, and i denotes the number of iterations.

3.3 Comparative results

3.3.1 Comparison of different dataset split ratios

To evaluate the impact of different dataset split ratios on model training, three datasets were constructed with split ratios of 6:2:2, 7:2:1, and 8:1:1 after data annotation. All models were trained using the same parameters. As shown in Table 1, the 8:1:1 split achieved the best performance in terms of Recall (85.8%) and mAP@50 (93.6%). Although the 7:2:1 split achieved the highest Precision, its mAP@50 was 92.2%, slightly lower than that of 8:1:1. The 6:2:2 split showed weaker performance in both Recall and mAP@50. Overall, the 8:1:1 split demonstrated the best detection performance, and subsequent model training and validation will be based on this dataset configuration.

Table 1
www.frontiersin.org

Table 1. Comparison of model performance across different data split ratios.

3.3.2 Scaling factor experiment

To achieve the optimal balance between accuracy, inference speed, and resource consumption, we conducted experiments with different scaling factors on the YOLOv10n model. The experimental results are shown in Table 2. The original model configuration (depth 0.33, width 0.25) achieved an mAP@50 of 93.6% and the fastest inference speed (1.8 ms), but its parameter count (2.694M) and model size (10.28 MB) are relatively large, making it less suitable for deployment on edge devices.

Table 2
www.frontiersin.org

Table 2. Comparison of YOLOv10 model with different scaling factors.

Through comparison, the configuration with scaling factors (0.33, 0.125) demonstrated excellent performance in terms of mAP@50, parameter count, and model size. Although its inference speed slightly decreased to 2.9 ms, it remains within an acceptable range. This configuration strikes a favorable balance between accuracy and model lightweightness, making it more suitable for edge deployment. Other configurations, such as (0.33, 0.2) and (0.33, 0.15), achieved slightly higher mAP@50, but their overall optimization effect was inferior to that of (0.33, 0.125). The shallower (0.25, 0.2) configuration reached an mAP@50 of 93.3%, but did not significantly reduce the parameter count or model size. Therefore, the (0.33, 0.125) configuration was selected as the optimal choice, offering improved model compactness and inference efficiency while maintaining high accuracy. To aid comparison, the best-performing values in Tables 26 are highlighted in bold to indicate the optimal results under each metric.

Table 3
www.frontiersin.org

Table 3. Performance comparison of different neck networks.

Table 4
www.frontiersin.org

Table 4. Performance comparison of different loss functions.

Table 5
www.frontiersin.org

Table 5. Performance comparison of YOLOv10-kiwi with different detection models on kiwifruit dataset.

Table 6
www.frontiersin.org

Table 6. Ablation study results showing individual contributions of scaling factors, C2fDualHet, CCFM, and MPDIoU.

3.3.3 Performance comparison of neck network designs

This study compares the performance of lightweight neck modules—BiFPN (Chen et al., 2021), Slimneck (Li et al., 2022), and CCFM—in object detection tasks. As shown in Table 3, CCFM demonstrates a significant advantage in terms of lightweight design, with only 0.586 million parameters, a model size of 2.36 MB, and a computational complexity of 2.32 GFLOPs, outperforming both BiFPN and Slimneck. This makes CCFM particularly suitable for deployment on resource-constrained embedded or mobile devices. Meanwhile, CCFM also delivers excellent detection performance, achieving a recall rate of 87.0%, surpassing BiFPN (83.1%) and Slimneck (82.7%). In summary, CCFM strikes a favorable balance between lightweight design and detection performance, showcasing strong potential for application in resource-limited scenarios.

3.3.4 Comparison of bounding box loss functions

We compared six loss functions—DIoU (Zheng et al., 2020), EIoU (Zhang et al., 2022), GIoU (Rezatofighi et al., 2019), CIoU, MPDIoU, and SIoU (Gevorgyan, 2022)—to evaluate their impact on model performance. The experimental results are presented in Table 4, and the trends of the loss curves are illustrated in Figure 9. According to Table 4, MPDIoU demonstrates the most significant performance improvement, with an increase of 0.8% in mAP@50 and 2.7% in Recall, achieving the highest values among all loss functions. Additionally, SIoU yields a 0.8% gain in Precision, while EIoU improves Recall by 0.2%. The remaining loss functions show varying degrees of decline across these key metrics. Further analysis of Figure 9 reveals that MPDIoU achieves the lowest final values in both val/box_loss and val/cls_loss, with a stable convergence trend, indicating its outstanding robustness in bounding box regression tasks. In contrast, DIoU, EIoU, and GIoU exhibit fluctuations in the loss curves during the later stages of training, suggesting a risk of overfitting. Overall, MPDIoU not only achieves the lowest final loss and fastest convergence on the validation set, but also significantly enhances kiwifruit detection performance—especially in mitigating the impact of low-quality samples such as blurred or occluded instances—demonstrating a well-balanced and superior capability in object detection tasks.

Figure 9
Six line graphs comparing different loss functions during training and validation across three categories: box, classification, and DFL loss. Each graph plots loss against epochs from 0 to 150, showing trends for Diou, Eiou, Giou, Mpdiou, Shapeiou, and Siou. All loss functions generally decrease over time, with slight variations among methods.

Figure 9. Comparison of training and validation loss curves for various IoU-based loss functions.

3.3.5 Comparison with other models

To verify the effectiveness of the proposed model in kiwifruit detection, YOLOv10-Kiwi is compared with nine mainstream object detection algorithms. The results are presented in Table 5 and Figure 10, covering both accuracy and efficiency metrics. In terms of model complexity, YOLOv10-Kiwi contains only 0.51M parameters, which is 98.8%, 97.4%, 95.8%, and 91.5% fewer than those of Faster R-CNN, RT-DETR-R18, YOLOv3-tiny, and YOLOv7-tiny, respectively. Its model size is 2.02 MB, significantly smaller than that of Faster R-CNN and RT-DETR-R18, and 78.8% and 82.4% smaller than YOLOv5n and YOLOv8n. The GFLOPs is only 2.1, which is much lower than most compared models, including YOLOv8n (8.2) and YOLOv10n (6.6). Despite its compact structure, YOLOv10-Kiwi achieves competitive detection performance. It reaches an mAP@50 of 93.6%, comparable to RT-DETR-R18 and YOLOv10n, and 4.1% and 2.7% higher than Faster R-CNN and YOLOv3-tiny, respectively. The recall is 86.4%, second only to Faster R-CNN (92%). The inference speed reaches 74 FPS, close to YOLOv5n (76) and YOLOv7-tiny (71), demonstrating real-time capability.

Figure 10
Bar chart comparing nine models: Faster R-CNN, RT-DETR-R18, YOLOv3-tiny, YOLOv5n, YOLOv8n, YOLOv10n, YOLOv11n, and Ours in terms of precision, recall, mean Average Precision at 50, parameters, model size, and GFLOPs. Each model's performance is represented by colored bars, indicating normalized values for each category.

Figure 10. Normalized comparison of accuracy and efficiency among models.

Figure 10 provides a visual comparison of six normalized metrics, where detection-related indicators (precision, recall, mAP@50) are positively normalized and resource-related metrics (parameters, model size, GFLOPs) are reverse-normalized. YOLOv10-Kiwi exhibits a well-balanced profile, combining high detection accuracy with superior efficiency. These results highlight its strong potential for deployment on resource-constrained edge devices in agricultural scenarios.

3.4 Ablation experiments

To validate the effectiveness of the proposed lightweight algorithm in kiwifruit object detection tasks, we conducted a series of systematic ablation experiments. The results are shown in Table 6 and Figure 11.

Figure 11
Radar chart comparing different models based on various metrics: mAP at 50%, recall, precision, inverse parameters (M), inverse model size (MB), and inverse GFLOPs. The models are YOLOv10n, Model 1, Model 2, Model 3, Model 4, Model 5, and a model labeled “Ours” which generally outperforms others, especially in inverse parameters and precision.

Figure 11. Radar chart comparing accuracy and complexity of ablation models.

First, by adjusting the scaling factors of the YOLOv10n network, we obtained Model 1, which significantly reduced model complexity—parameters decreased from 2.7M to 0.998M (a reduction of 63.04%) and GFLOPs from 5.5 to 3.8 (a reduction of 30.91%)—demonstrating that scaling-based compression can effectively shrink the model footprint. However, detection performance declined, with Recall dropping by 2.9% and mAP@50 by 1.5%, indicating a degradation in representational capacity due to downscaled feature extraction. To address this, Model 2 introduced the C2fDualHet module as a lightweight replacement for the original C2f block in the backbone. The module was designed to further reduce computational overhead while enhancing the network’s ability to capture local and global features. Compared to Model 1, the parameter count and GFLOPs were further reduced by 8.12% and 13.16%, respectively—confirming that the compression objective was further achieved. However, detection accuracy remained relatively unchanged, and mAP@50 slightly decreased by 0.7%, suggesting that while C2fDualHet successfully reduces computation, its standalone impact on detection precision is limited without enhanced feature fusion. Model 3, which added the Cross-Channel Fusion Module (CCFM) to Model 1, significantly improved performance. Recall increased from 82.9% to 87.0%, and mAP@50 rose from 92.1% to 93.3%, with Precision reaching 88.2%. This confirms CCFM’s role in strengthening feature integration across scales and channels, compensating for information loss from earlier compression. Model 4 evaluated the impact of using MPDIoU loss in isolation (on top of Model 1). Compared with Model 1, Recall improved markedly from 82.9% to 85.2%, while mAP@50 increased to 92.7%, demonstrating that MPDIoU can significantly improve localization performance without structural changes. Model 5 combined both C2fDualHet and CCFM, resulting in the highest Precision (89.6%) among all sub-models and achieving strong efficiency with only 0.505M parameters and 2.1 GFLOPs, highlighting the synergistic benefits of both modules. Finally, our full model (Ours) integrated all proposed components, including MPDIoU loss. It achieved the best balance overall, with mAP@50 of 93.6%, Recall of 86.4%, and a compact model size of 2.02 MB. These results confirm that the full architecture of YOLOv10-Kiwi effectively balances detection accuracy and lightweight design.

To intuitively illustrate the performance trade-offs, Figure 11 presents a radar chart with normalized values (0–1), where computational metrics were inversely scaled. The progressive expansion of the radar area clearly visualizes how each component contributes to the overall model performance, with YOLOv10-Kiwi (Model Ours) demonstrating the best balance across all metrics.

3.5 Model detection effect analysis

To evaluate the performance of YOLOv10-Kiwi in the kiwifruit detection task, we present the detection results of RT-DETR-R18, YOLOv3-tiny, YOLOv8n, YOLOv10n, and YOLOv10-Kiwi in Figure 12. YOLOv10-Kiwi is capable of accurately identifying kiwifruit and maintains high detection precision even in complex backgrounds. Although occlusion from branches and leaves or lighting variations may cause occasional false detections or missed detections, the overall impact is minimal, indicating that YOLOv10-Kiwi can effectively meet the requirements of kiwifruit detection. In comparison, RT-DETR-R18 demonstrates limited detection accuracy and is prone to false detections under complex backgrounds; YOLOv3-tiny shows improvements in reducing false positives but still suffers from missed detections; YOLOv8n and YOLOv10n strike a balance between accuracy and real-time performance, yet exhibit deviations in detecting partially occluded targets. Notably, as a lightweight model, YOLOv10-Kiwi significantly reduces parameter count and computational complexity while maintaining detection accuracy comparable to YOLOv10n, making it more suitable for deployment on resource-constrained edge devices. YOLOv10-Kiwi effectively reduces the missed detection rate while ensuring high confidence, demonstrating outstanding detection performance.

Figure 12
Multiple panels show kiwi fruits on branches with overlapping red boxes and labels. Each row represents different images, and columns labeled (a) to (e) differentiate variations. The fruits and leaves appear in different lighting and shadow effects, indicating focus on fruit identification or distinction in various conditions.

Figure 12. Detection results of different models. (a) RT-DETR-R18; (b) YOLOv3-tiny; (c) YOLOv8n; (d) YOLOv10n; (e) YOLOv10-Kiwi.

3.6 Model visualization analysis

To further investigate the feature extraction capabilities of the model in kiwifruit detection, this study extracts features from multiple convolutional layers of different models and utilizes the Grad-CAM method (Selvaraju et al., 2017)to generate heatmaps for visual analysis, as shown in Figure 13. In these heatmaps, the regions of higher attention toward the kiwifruit targets are represented by deeper red areas, while lighter areas reflect lower attention. Experimental results indicate that YOLOv10-Kiwi demonstrates excellent performance in multi-layer feature extraction. Compared to YOLOv3-tiny and YOLOv8n, its heatmaps exhibit more focused and precise responses in the target regions, particularly in the identification of kiwifruit under complex backgrounds. This suggests that YOLOv10-Kiwi holds significant advantages in the hierarchical structure and accuracy of feature extraction, making it more suitable for the task of kiwifruit detection.

Figure 13
A series of images are arranged in a grid, showcasing kiwi fruit on trees. The first column (a) depicts clear images of kiwi clusters. Columns (b), (c), and (d) show the same images with increasing levels of highlighted areas in various colors, likely indicating heat map data or detection analysis to emphasize specific features of the kiwi and foliage.

Figure 13. Heat map for different models. (a) original images; (b) YOLOv3-tiny; (c) YOLOv8n; (d) YOLOv10-Kiwi.

4 Conclusion

In this study, we proposed YOLOv10-Kiwi, a lightweight and high-performance model specifically designed for kiwifruit detection in complex orchard environments. By integrating the C2fDualHet module, CCFM neck, and MPDIoU loss into the YOLOv10n framework, the model achieves an effective trade-off between accuracy and computational efficiency. Extensive experiments demonstrated that YOLOv10-Kiwi achieves an mAP@50 of 93.6%, recall of 86.4%, and precision of 88.3%, while reducing parameters to 0.51M, model size to 2.02 MB, and GFLOPs to 2.1. The real-time inference speed reaches 74 FPS, making it highly suitable for deployment on edge devices in agricultural scenarios. The proposed model exhibits robust detection under various conditions, such as occlusion, lighting variations, and complex backgrounds. It is also much more lightweight than comparable state-of-the-art detectors, enabling its practical use in resource-constrained environments.

In future work, we plan to expand the dataset across multiple seasons and kiwifruit varieties, and validate the model’s generalizability under varying orchard geometries. Moreover, we aim to integrate YOLOv10-Kiwi into robotic picking platforms to further explore its applicability in real-world agricultural automation tasks.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author.

Author contributions

WW: Writing – review & editing, Formal Analysis, Supervision, Conceptualization. JR: Writing – original draft, Methodology, Investigation, Conceptualization, Writing – review & editing. YT: Writing – review & editing, Investigation, Software. JH: Resources, Writing – review & editing, Validation.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This work is supported in part by National Natural Science Foundation of China under grant No. 62366053, and the “Unveiling and Commanding” Special Research Program of Yan ‘an university under grant No. 2023JBZR-021.

Acknowledgments

The authors would like to thank all the reviewers and editors for their valuable comments.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ren, S., He, K., Girshick, R., and Sun, J.(2015). Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 28, 91–9.

PubMed Abstract | Google Scholar

Cai, Z. and Vasconcelos, N. (2018). “Cascade r-cnn: Delving into high quality object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Los Angeles, CA, USA: IEEE. 6154–6162.

Google Scholar

Chen, J., Mai, H. S., Luo, L., et al. (2021). “Effective feature fusion network in BIFPN for small object detection,” in Computers and Electronics in Agriculture. Amsterdam, Netherlands: Elsevier, 699–703.

Google Scholar

Chollet, F. (2017). “Xception: Deep learning with depthwise separable convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Honolulu, HI, USA: IEEE. 1251–1258.

Google Scholar

Deng, S., He, R., Hu, Q., and Duan, Z. (2023). Significant increases in Donghong kiwifruit yield by a novel umbrella-shaped trellis system and identification of associated molecular mechanisms. Front. Plant Sci. 14, 1143525. doi: 10.3389/fpls.2023.1143525

PubMed Abstract | Crossref Full Text | Google Scholar

Fu, J., Wang, J., and Li, Y. (2021). Fast and accurate detection of kiwifruit in orchard using improved YOLOv3-tiny model. Precis. Agric. 22, 754–776. doi: 10.1007/s11119-020-09754-y

Crossref Full Text | Google Scholar

Fu, Y., Xu, J., Zhao, B., Xie, Y., Zeng, Y., and Peng, D. (2024). Design and experimentation of multi-fruit envelope-cutting kiwifruit picking robot. Front. Plant Sci. 15, 1338050. doi: 10.3389/fpls.2024.1338050

PubMed Abstract | Crossref Full Text | Google Scholar

Gao, X., Gao, Y., Li, Y., and Liu, Y. (2022). Kiwifruit detection method in orchard via an improved light-weight YOLOv4. Agronomy 12, 2081. doi: 10.3390/agronomy12092081

Crossref Full Text | Google Scholar

Gevorgyan, Z. (2022). SIoU loss: More powerful learning for bounding box regression. Mathematical Notes 112, 177–182. doi: 10.1134/S0001434622070215

Crossref Full Text | Google Scholar

Guo, Y., Zhang, S., and Wang, J. (2020). “Augfpn: Improving multi-scale feature learning for object detection,” in Proceedings of the 2020 International Conference on Artificial Intelligence and Computer Engineering (ICAICE). Beijing, China: IEEE. 12595–12604.

Google Scholar

Guo, Y., Zhang, S., and Wang, J. (2019). “Group-wise correlation stereo network,” in IEEE International Conference on Computer and Communications (ICCC). Chengdu, China: IEEE. 3273–3282.

Google Scholar

He, K., Gkioxari, G., Dollár, P., and Girshick, R. (2017). “Mask r-cnn,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV). Venice, Italy: IEEE, 2961–2969.

Google Scholar

Howard, A. G. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861. doi: 10.48550/arXiv.1704.04861

Crossref Full Text | Google Scholar

Jiang, Y., Chen, Y., and Li, Z. (2022). A Review of Yolo algorithm developments. Proc. Comput. Sci. 199, 1066–1073. doi: 10.1016/j.procs.2022.01.135

Crossref Full Text | Google Scholar

Li, X., Zhang, H., and Zhao, Q. (2022). Slim-neck by GSConv: A better design paradigm of detector architectures for autonomous vehicles. arXiv preprint arXiv:2206.02424 10.

Google Scholar

Liu, D., Zhang, H., Wang, B., Li, T., Wu, J., and Zhao, Y. (2024). Faster-YOLO-AP: A lightweight apple detection algorithm based on improved YOLOv8 with a new efficient PDWConv in orchard. Comput. Electron. Agric. 223, 109118. doi: 10.1016/j.compag.2024.109118

Crossref Full Text | Google Scholar

Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., and Fu, C-Ÿ (2016). “Ssd: Single shot multibox detector,” in European Conference on Computer Vision (ECCV). Amsterdam, Netherlands: Springer, 21–37.

Google Scholar

Longsheng, T., Jian, W., and Bing, L. (2015). Kiwifruit recognition at nighttime using artificial lighting based on machine vision. Int. J. Agric. Biol. Eng. 8, 52–59.

Google Scholar

Ma, S. and Xu, Y. (2023). Mpdiou: a loss for efficient and accurate bounding box regression. Comput. Electron. Agric. 210, 107973.

Google Scholar

Rezatofighi, H., Tsoi, N., Gwak, J-Ÿ, Sadeghian, A., Reid, I., and Savarese, S. (2019). “Generalized intersection over union: A metric and a loss for bounding box regression,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Long Beach, CA, USA: IEEE, 658–666.

Google Scholar

Richardson, D. P., Ansell, J., and Drummond, L. N. (2018). The nutritional and health attributes of kiwifruit: a review. Eur. J. Nutr. 57, 2659–2676. doi: 10.1007/s00394-018-1627-z

PubMed Abstract | Crossref Full Text | Google Scholar

Scarfe, A. J. (2012). Development of an autonomous kiwifruit harvester (Berlin, Germany: Massey University).

Google Scholar

Selvaraju, R R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV). Venice, Italy: IEEE, 618–626.

Google Scholar

Singh, P., Verma, V-K., Rai, P., Krishna, G., Singh, P., and Nishchal, N-K. (2019). “Hetconv: Heterogeneous kernel-based convolutions for deep cnns,” in IEEE 10th International Conference on Awareness Science and Technology (iCAST). Morioka, Japan: IEEE, 4835–4844.

Google Scholar

Suo, R., Gao, F., Zhou, Z., Fu, L., Song, Z., and Dhupia, J. (2021). Improved multi-classes kiwifruit detection in orchard to avoid collisions during robotic picking. Comput. Electron. Agric. 182, 106052. doi: 10.1016/j.compag.2021.106052

Crossref Full Text | Google Scholar

Tan, M., Pang, R., and Le, Q. V. (2020). “Efficientdet: Scalable and efficient object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA, USA: IEEE, 10781–10790.

Google Scholar

Wang, C-Y., Liao, H-Y-M., Wu, Y-H., Chen, P-Y., Hsieh, J-W., and Yeh, I-H. (2024). Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 37, 107984–108011. doi: 10.48550/arXiv.2405.14458

Crossref Full Text | Google Scholar

Xia, Y., Nguyen, M., and Yan, W. Q. (2022). “A real-time kiwifruit detection based on improved YOLOv7,” in Agriculture. Basel, Switzerland: MDPI, 48–61.

Google Scholar

Yao, J., Qi, J., Zhang, J., Zhang, H., Wang, T., and Liu, Y. (2021). A real-time detection algorithm for Kiwifruit defects based on YOLOv5. Electronics 10, 1711. doi: 10.3390/electronics10141711

Crossref Full Text | Google Scholar

Zhan, W., He, D., and Shi, S. (2013). Recognition of kiwifruit in field based on Adaboost algorithm. Trans. Chin. Soc. Agric. Eng. 29, 140–146. doi: 10.1523/JNEUROSCI.2742-12.2013

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, Y., Zhou, Q., Li, X., Chen, C., Wang, H., and Li, J. (2022). Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 506, 146–157. doi: 10.1016/j.neucom.2022.07.042

Crossref Full Text | Google Scholar

Zhao, J., Du, C., Li, Y., Mudhsh, M., Guo, D., Fan, Y., et al. (2024). YOLO-Granada: a lightweight attentioned Yolo for pomegranates fruit detection. Sci. Rep. 14, 16848. doi: 10.1038/s41598-024-67526-4

PubMed Abstract | Crossref Full Text | Google Scholar

Zhao, Y., Li, B., Zhang, R., Liu, C., Wang, J., and Chen, Z. (2024). “Detrs beat yolos on real-time object detection,” in Sensors. Basel, Switzerland: MDPI, 16965–16974.

Google Scholar

Zheng, Z., Wang, P., Liu, W., Li, J., Ye, R., and Ren, D. (2020). “Distance-IoU loss: Faster and better learning for bounding box regression,” in AAAI Conference on Artificial Intelligence (AAAI), New York, NY, USA: AAAI Press, Vol. 34. 12993–13000.

Google Scholar

Zhou, T., He, Y., and Lin, L. (2022). Lightweight detection algorithm of kiwifruit based on improved YOLOX-s. Agriculture 12, 993. doi: 10.3390/agriculture12070993

Crossref Full Text | Google Scholar

Keywords: kiwifruit detection, YOLOv10, lightweight network, HetConv, CCFM, MPDIou

Citation: Ren J, Wang W, Tian Y and He J (2025) YOLOv10-kiwi: a YOLOv10-based lightweight kiwifruit detection model in trellised orchards. Front. Plant Sci. 16:1616165. doi: 10.3389/fpls.2025.1616165

Received: 22 April 2025; Accepted: 23 July 2025;
Published: 25 August 2025.

Edited by:

Ye Liu, Nanjing Agricultural University, China

Reviewed by:

Zari Farhadi, University of Tabriz, Iran
Sutanu Nandi, University of Colorado Anschutz Medical Campus, United States

Copyright © 2025 Ren, Wang, Tian and He. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wendong Wang, eWFkeHd3ZDIwMTJAMTYzLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.