ORIGINAL RESEARCH article

Front. Plant Sci., 03 July 2025

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1588626

Identification of leaves of wild Ussurian Pear (Pyrus ussuriensis) based on YOLOv10n-MCS

Niman Li,Niman Li1,2Xingguang Dong*Xingguang Dong1*Yongqing Wu*Yongqing Wu2*Luming TianLuming Tian1Ying ZhangYing Zhang1Hongliang HuoHongliang Huo1Dan QiDan Qi1Jiayu XuJiayu Xu1Chao LiuChao Liu1Zhiyan ChenZhiyan Chen2Yulu MouYulu Mou1
  • 1Research Institute of Pomology, Chinese Academy of Agricultural Sciences (CAAS), Key Laboratory of Horticulture Crops Germplasm Resources Utilization, Ministry of Agriculture and Rural Affairs, Xingcheng, China
  • 2School of Software, Liaoning Technical University, Huludao, China

Introduction: Wild Ussurian Pear germplasm resource has rich genetic diversity, which is the basis for genetic improvement of pear varieties. Accurately and efficiently identifying wild Ussurian Pear accession is a prerequisite for germplasm conservation and utilization.

Methods: We proposed YOLOv10n-MCS, an improved model featuring: (1) Mixed Local Channel Attention (MLCA) module for enhanced feature extraction, (2) Simplified Spatial Pyramid Pooling-Fast (SimSPPF) for multi-scale feature capture, and (3) C2f_SCConv backbone to reduce computational redundancy. The model was trained on a self-made dataset of 16,079 wild Ussurian Pear leaves images.

Results: Experiment results demonstrate that the precision, recall, mAP50, parameters, FLOPs, and model size of YOLOv10n-MCS reached 97.7(95% CI: 97.18 to 98.16)%, 93.5(95% CI: 92.57 to 94.36)%, 98.8(95% CI: 98.57 to 99.03)%, 2.52M, 8.2G, and 5.4MB, respectively. The precision, recall, and mAP50 are significant improved of 2.9%, 2.3%, and 1.5% respectively over the YOLOv10n model (p<0.05). Comparative experiments confirmed its advantages in precision, model complexity, model size, and other aspects.

Discussion: This lightweight model enables real-time wild Ussurian Pear identification in natural environments, providing technical support for germplasm conservation and crop variety identification.

1 Introduction

The genus Pear is a worldwide fruit tree, which originated in the mountainous areas of southwest China in the Tertiary Period, and evolved into two groups of Occidental and Oriental pear, respectively and there are 22 primary Pyrus species recognized by the academic community (Bell, 1996). China is the center of origin of oriental pear, with 13 native species, including 5 basic species: P. pyrifolia, P. ussuriensis, P. pashia, P. betulaefolia, and P. calleryana (Pu and Wang, 1963). Among them, the Ussurian Pear (P. ussuriensis) is the most cold-tolerant species, with wild and cultivated types, which grow in northeast China, northern Hebei and northern Shaanxi, and also have wild distribution in the Russian Far East and North Korea. Northeast China is the characteristic production area of Ussurian Pear, and there are more than 150 cultivated varieties. The fruit of Ussurian Pear is generally small, round, rich in flavor, and eaten after ripening and softening, The representative varieties are Nanguo, Jingbai, Anli and so on, which are deeply loved by the local people. It is generally thought that the cultivars have been domesticated from wild species, and the diversity of wild species is more abundant, and our study of the genetic diversity of the Ussurian Pear has reached the same conclusion (Cao et al., 2012). The wild Ussurian Pear are mainly used as rootstocks in Northeast China, and their ecological and production value are far from being explored. Due to its long-term wild state, wild Ussurian Pear has strong adaptability to the environment, and the diversity of biological and botanical characteristics is also very rich, and these excellent traits are important gene sources for cultivated species improvement (Thakur et al., 2024).

Accurate and rapid identification of germplasm resources is the foundation for their preservation, research, and utilization. However, due to interspecific and intraspecific hybridization being the main mode of pear evolution, pear populations are large and with a high degree of heterozygosity. This leads to significant difficulties in identifying and classifying pear germplasm resources (Xue et al., 2017). Morphological identification is the most fundamental and important method for pear variety identification (Zhang et al., 2022a). The staff identifies pear varieties by analyzing their phenotypic characteristics, such as observing leaf morphology, branch color, and fruit characteristics. However, due to the susceptibility of this method to individual plant development, environmental conditions, cultivation measures, and human factors, the identification process is time-consuming and has a high error rate (Adão et al., 2025). With the rapid development of high-throughput sequencing and molecular biology techniques, pear varieties can be accurately identified using molecular markers. However, this method has limitations such as cumbersome operation, time-consuming, and high cost due to the high requirements for operating steps and equipment (Liu et al., 2024).

In the study of plant variety identification, leaves are easier to collect than other organs. The leaves of any plant have their unique characteristics, such as leaf shape, leaf bases, leaf apex, and leaf margin (Wang et al., 2023b). Pear leaves are mainly divided into circular, ovate, elliptical, lanceolate, and lobed shapes. Among them, ovate shaped resources are the most abundant. Leaf base refers to the proximal part of the leaf near the stem. The shapes of pear leaf bases are mainly divided into narrow cuneate, cuneate, broad cuneate, round, truncate, and cordate, with broad cuneate being the most prevalent. The shapes of pear leaf apices can be divided into acuminate, obtuse, acute, and caudate. And the leaf margins can be classified as entire, crenate, obtusely serrate, sharply serrate, and doubly serrate. In recent years, with the advancement of agricultural informatization, computer vision, machine learning and other technologies are widely used different areas in agriculture (Zhang et al., 2022b). More and more researchers are focusing on image classification and pattern recognition, and using them to quickly identify and classify plant leaves. Pan et al. (2024) used four deep learning networks, including GoogleNet, ResNet50, ResNet101, and VGG16, to identify and classify the leaf images of 23 wild grapes, and realized automatic real-time identification of wild grapes. Chen et al. (2022) used Convolutional Neural Network (CNN) to extract features and classify 30 types of apple leaves, significantly improving the classification accuracy of apple cultivars. Wei Tan et al. (2018) proposed a CNN model called the D-leaf model, which utilizes leaf vein morphometric to classify plant species. And three different CNN models (pre trained AlexNet, fine-tuned AlexNet, and D-Leaf) were used to preprocess the leaf images and extract feature information. To classify the extracted features, five machine learning approaches were utilized, including Artificial Neural Networks (ANN), Support Vector Machines (SVM), and CNN, etc. The results confirm that deep learning-based models are effective for tasks such as leaf recognition and variety categorization.

Although the use of deep learning for leaf classification is highly effective, many of the models used have limitations such as complex network structures and large parameter quantities, which consume a significant amount of computing resources and can only be deployed on the server side (Fang et al., 2019), resulting in certain limitations in leaf recognition tasks.”You Only Look Once” (YOLO) target detection algorithm is the leading single-stage algorithm in deep learning object detection methods (Adarsh et al., 2020). It can predict the category and location of objects in the image by recognizing image information. In recent years, with the continuous innovation of the YOLO series (Redmon, 2016; Redmon and Farhadi, 2017; Farhadi and Redmon, 2018; Bochkovskiy et al., 2020; Jocher, 2020; Li et al., 2022; Wang et al., 2023a; Jocher et al., 2023; Wang et al., 2025a, 2024), its performance has gradually improved, achieving a balance between recognition speed and accuracy. It has better performance and application prospects in quickly identifying plant leaves and deploying on mobile devices. The leaves of different pear varieties have differences in shape, leaf apex, and leaf margin. These differences can serve as the basis for YOLO classification. And the characteristics of pear leaves are highly matched with YOLO’s advantages such as real-time detection and multi-target processing. YOLO can extract these morphological features through CNN and learn the differentiation patterns of different varieties through training. In addition, by optimizing the model, YOLO’s potential for using leaves to classify pear varieties can be fully utilized. Many recent studies have demonstrated that YOLO series target detection algorithms can effectively achieve leaf species identification and classification tasks. Yang et al. (2024) used the YOLOX algorithm and combined with the self-made tea bud dataset to establish a tea bud classification model, which can recognize and classify four types of tea buds, in which the recognition accuracy for the yellow mountain species could reach 90.54%. Niu et al. (2024) improved the YOLOv8 algorithm and proposed a lightweight YOLOv8-EFS model based on their own soybean seedlings and weeds image datasets, which can quickly and accurately identify multiple weed species and provide support for intelligent weed control in farmland management and unmanned farms.

The rapid and accurate identification of wild pear accessions relies on high-quality datasets and high-performance detection models. Although significant achievements have been made in plant species recognition through existing research, there are still problems such as excessive consumption of computational resources and low recognition accuracy in the models used for recognition. In addition, there is currently limited research on the identification of wild pear germplasm resource, and there have been no reports on the use of target detection algorithms to identify wild pear leaves for variety classification.

In response to the above issues, this article takes wild Ussurian Pear leaves in natural environments as the research object, selects YOLOv10n as the baseline model, and proposes an improved leaves recognition and classification model YOLOv10n-MCS. Firstly, this study selected 30 wild Ussurian Pear accessions and constructed a dataset of wild Ussurian Pear leaves images. Collect 500–600 images for each accession to meet the requirements of multitasking classification. Secondly, based on YOLOv10n, Mixed Local Channel Attention (MLCA) module, C2f SCConv module, and Simplified Spatial Pyramid Pooling - Fast (SimSPPF) module were introduced to enhance the feature extraction ability of the model, reduced computational redundancy, and improve detection performance. Finally, the improved model was combined with the dataset of wild Ussurian Pear leaves images to establish a recognition and classification model covering 30 wild Ussurian Pear accessions, thereby achieving automatic recognition and classification of wild Ussurian Pear leaves. This study can accurately and quickly identify and classify wild Ussurian Pear leaves, reducing complicated labor costs. In addition, it also provides reference for the protection, utilization, classification research of wild pear germplasm resource, as well as the identification of other crop varieties.

2 Materials and methods

2.1 Image data acquisition

Training a wild pear leaves recognition and classification model requires inputting a large number of pear leaves images as sample images into the network model. Model accuracy and efficiency are significantly influenced by the quality of the dataset used for training. We selected 30 wild Ussurian Pear accessions as the research objects for wild pear leaves recognition and classification, and constructed a natural background wild Ussurian Pear leaves images dataset. The origin of these 30 wild Ussurian Pear accessions is shown in Figure 1. We went to the “National Pear and Apple Germplasm Resources Repository (Xingcheng)” to collect images of wild Ussurian Pear leaves. From October 5th to 12th, 2024, use a high-resolution camera on the same mobile phone to capture images of leaves from 30 wild pear resources under natural conditions. Place the camera at a distance of 20-60cm from the leaves, take frontal images of adult leaves, and try to ensure that the leaf shape, leaf bases, leaf apex, and leaf margin in the image are clear as much as possible. Collect 500–600 images for each variety. The collected images are in JPG format with a resolution of 3072 pixels × 4096 pixels. The weather conditions during the shooting process include sunny, cloudy, and rainy days. There are both shooting under soft lighting and strong afternoon lighting, covering various common weather conditions and natural lighting conditions, increasing the diversity and complexity of the natural conditions in which the leaves are located. Ultimately, a total of 16997 images of wild Ussurian Pear leaves were obtained through this process, and some of the leaves images dataset samples are shown in Figure 2.

Figure 1
Map of China highlighting provinces. An inset focuses on Liaoning, showing multiple red dots indicating specific locations. A small map of the South China Sea Islands is included.

Figure 1. Map of the origin regions of wild Ussurian Pear accession.

Figure 2
Twenty-seven images of Pyrus ussuriensis leaves arranged in a grid. Each image displays a different variant with accompanying labels, showcasing variations in leaf shape, size, and texture among the species.

Figure 2. Some representative samples of the leaves images dataset.

2.2 Images annotation and dataset construction

After careful manual screening, images with clear leaves contours and textures were selected, and a total of 16079 leaves images were used to form the wild Ussurian Pear leaves images dataset. Manually annotate the filtered leaves images using LabelImg software(can be in https://github.com/HumanSignal/labelImgget). Use the “Create RectBox” function in the software to annotate each leaf based on the minimum external rectangle in the image, and select a category for the annotated blade, such as’Sunwu-3’. In order to avoid inaccurate labeling and blurry targets affecting the quality and performance of the model, blades with an occlusion area exceeding 70% are not labeled. And ensure that the rectangular box contains as little background as possible to improve recognition precision. The annotation example is shown in Image 1 of the Supplementary File. Subsequently, after annotation, obtain the.txt format annotation file required by YOLO. This file contains the category of the blade, the coordinates x and y of the center point of the rectangle, the width w and height h relative to the image Fan et al. (2024). To facilitate model training and evaluation, the wild Ussurian Pear leaves images dataset was randomly split into training, validation, and test sets in a 7:2:1 ratio, corresponding to model training, validation, and prediction tasks, respectively. The training set contained 11243 samples, the validation set included 3371 samples, and the test set comprised 1465 samples. Detailed sample information is provided in Table 1.

Table 1
www.frontiersin.org

Table 1. Leaves samples information of 30 species of wild pear.

2.3 YOLOv10n model

YOLO is a single-stage target detection algorithm. From focusing on detection speed in YOLOv1 to optimizing the overall architecture, balancing speed, accuracy, and model size in YOLOv10, each version update is gradually addressing the shortcomings of the previous YOLO series. Among them, the YOLOv10 algorithm was proposed by researchers Wang et al. from Tsinghua University in 2024. It addresses the shortcomings in terms of post-processing and model architecture of previous versions of YOLO. Compared to previous versions, it improves both detection precision and speed.

The YOLOv10 network architecture consists of three parts: backbone, neck, and head. It selects YOLOv8 as the baseline model and proposes a new model design based on it. It introduces compact inverted block (CIB) and designs a C2fCIB structure to reduce computational costs and improve efficiency. YOLOv10 proposes an efficient partial self-attention (PSA) module design that improves detection performance and efficiency without increasing excessive computational costs. The head section employs a dual label assignment strategy, which combines a one-to-many approach with a one-to-one matching mechanism. This approach removes the need for non-maximum suppression (NMS), significantly reducing inference latency.

YOLOv10 performs well on COCO Lin et al. (2014), maintaining high accuracy in complex background detection while also being lightweight and easy to deploy on embedded devices. The YOLOv10 includes six variants, namely YOLOv10n, YOLOv10s, YOLOv10m, YOLOv10b, YOLOv10l, and YOLOv10x, to meet different application scenarios. Among them, YOLOv10n is the lightest in terms of parameters and floating-point operations (FLOPs), meeting the lightweight requirements.

2.4 Principle of YOLOv10n-MCS model

In target detection tasks based on deep learning, the size and complexity of the model directly affect the effectiveness of practical applications. Although YOLOv10 algorithm has high recognition accuracy and speed, for wild Ussurian Pear leaves targets in complex scenes, the model still has problems such as slow recognition speed and low precision due to the high similarity of leaf features and occlusion between leaves. Due to the fact that leaves recognition is an efficient and lightweight task, in order to ensure the feature extraction capability of the model for wild Ussurian Pear leaves and enhance the detection performance of the model, this paper selects YOLOv10n as the baseline model and proposes a new YOLOv10n-MCS model. Firstly, the MLCA module is introduced into the neck of the network to enhance the model’s feature extraction capability and improve recognition accuracy. Secondly, using SimSPPF module instead of the original network pyramid pooling layer can improve the detection efficiency of the model. Finally, design a C2f SCConv module to replace C2f in backbone, reducing computational redundancy and improving detection performance. The network structure of the improved YOLOv10n-MCS is shown in Figure 3.

Figure 3
Flowchart illustrating a neural network architecture. It has three main sections: Backbone, Neck, and Head, each outlined with dashed borders. The Backbone includes Conv, C2f_SCCConv, and SimSPPF modules. The Neck section features Concat, C2f, and MLCA components. The Head section displays MHSA, FFN, and PSA modules, with arrows indicating connections between components.

Figure 3. The architecture of the YOLOv10n-MCS model.

2.4.1 The MLCA module

In deep learning, attention mechanisms are designed to emulate the human visual and cognitive processes. By enabling models to concentrate on critical regions while disregarding irrelevant data, attention mechanisms enhance both efficiency and accuracy in information processing. Currently, most channel attention mechanisms, including Squeeze-and-Excitation (SE) Hu et al. (2018) and Efficient Channel Attention (ECA) Wang et al. (2020), focus solely on channel features and neglect spatial information within each channel. This limitation may result in the loss of crucial feature information, negatively impacting model performance in object detection and leading to inaccurate category predictions. Additionally, spatial attention modules, while effective, tend to be computationally intensive. To address this issue, Wan et al. Wan et al. (2023) developed the lightweight MLCA module, which balances performance and complexity. The MLCA module integrates channel and spatial information, along with local and global features, thereby preventing the loss of critical information and enhancing the expressive power and detection performance of object detection algorithms. A better balance has been achieved between detection accuracy, speed, and model parameters without significantly increasing computational costs. Therefore, we incorporated the MLCA module into the neck of our model to improve its feature extraction capabilities.

As illustrated in Figure 4, the feature vectors are fed into the MLCA module and converted into a 1 * C * ks * ks vector to capture local spatial details. The input is then processed through two parallel branches, each converting the data into one-dimensional vectors. The first branch focuses on local spatial features, while the second branch captures global information. One-dimensional convolution is applied independently to the vectors in both branches. The original resolution of the vectors is recovered through anti-pooling, followed by information fusion to achieve mixed attention. In the figure, k denotes the convolution kernel size, and C represents the channel dimension, with both being proportional. This indicates that local cross-channel interactions are captured by considering only the relationships between each channel and its k neighboring channels. The formula for calculating k is provided in Equation 1. Among them, γ and b are hyperparameters, odd means that k is only odd, and if k is even, add 1.

Figure 4
Diagram illustrating a deep learning process involving a series of transformations on a three-dimensional data cube. The process starts with an input cube labeled (C, W, H), where C, W, and H represent dimensions. The image shows steps including Local and Global Average Pooling (LAP, GAP), one-dimensional convolution (Conv1d), reshaping, and unpooling (UNAP). The cube transforms into smaller sections, undergoes calculations, and recombines to form an output cube identical in dimensions to the input. The text “Where ks = 5” indicates a kernel size parameter.

Figure 4. The principle of MLCA algorithm: LAP (Local Average Pooling) which divides the feature map into k * k patches and applies average pooling to each patch; GAP (Global Average Pooling), which uses adaptive pooling to reduce the feature map to a 1 * 1 output size; UNAP (Anti-average Pooling), which mainly focuses on the figure’s properties and scaling to the needed size.

k=Φ(C)=|log2C)γ+bγ|odd(1)

2.4.2 The C2f SCConv module

Building a deep network with few parameters and low computational complexity requires compact and highly efficient model design. The C2f module of YOLOv10 enhances its feature extraction capability through the bottleneck structure, but it introduces a large amount of irrelevant interference information for equivalent processing of all channel and position information. The C2f module is internally stacked with a large number of convolution operations, resulting in highly similar features between adjacent channels, leading to redundant features during iteration.

Spatial and Channel reconstruction Convolution (SCConv), introduced by Li et al. (2023), is an advanced convolutional module designed to minimize spatial and channel redundancy in feature maps, leading to more compact CNN models and enhanced performance. As illustrated in Figure 5, the SCConv module comprises two key components: the Spatial Reconstruction Unit (SRU) and the Channel Reconstruction Unit (CRU). The SRU employs a Separate-and-Reconstruct approach to effectively reduce spatial redundancy. For an input feature X, the SRU first performs a separate operation, dividing the feature maps into two categories based on their spatial information content: one category includes feature maps with rich spatial details, while the other contains maps with minimal or redundant information. These are then reconstructed to produce the spatially refined feature Xw. The CRU utilizes a Split-Transform-andFuse strategy to mitigate channel redundancy. It processes Xw by splitting its channels into two branches. The upper branch employs GWC and PWC convolutions to efficiently extract representative features, while the lower branch uses PWC convolutions to enhance hidden details. To adaptively fuse the features from both branches, the SKNet Li et al. (2019) technique is applied, producing a channel-refined output Y that significantly reduces redundancy, computational overhead, and memory usage. Together, the SRU and CRU synergistically reduce redundant information in convolutional layers, lower model complexity, and enhance feature extraction capabilities.

Figure 5
Diagram illustrating the SCConv block process in a neural network. It starts with a previous ConvBlock, followed by a one-by-one convolution to create Input Feature X. The Spatial-Refined Unit (SRU) refines it to Spatial-Refined Feature X^s. The Channel-Refined Unit (CRU) then refines it to Channel-Refined Feature Y. Another one-by-one convolution follows, leading to addition with the previous feature, completing the process before proceeding to the next ConvBlock.

Figure 5. The architecture of SCConv integrated with SRU and CRU.

Therefore, a new C2f SCConv module is designed by replacing the ordinary convolutions in C2f with SCConv. Replace C2f in YOLOv10n’s backbone with C2f SCConv module to improve the model’s learning ability and further reduce the number of model parameters. Reduce redundant features, decrease computational complexity, and improve detection performance. The C2f SCConv module is shown in Figure 6.

Figure 6
Flowchart of a convolutional neural network layer sequence. Steps include Conv (Convolution), Split, two SCCov (Sparsely Connected Convolution), Concat (Concatenation), followed by another Conv. Dotted arrow labeled 'N' indicates repetition of SCCov.

Figure 6. C2f-SCConv module structure diagram.

2.4.3 SimSPPF

Spatial Pyramid Pooling-Fast (SPPF) in YOLOv10n is a pooling operation that introduces three-layer max pooling and connects the outputs of each layer to extract multi-scale information from the input feature map. It extracts and aggregates input feature maps from different perspectives, achieving the fusion of local and global features. Improved computational efficiency while maintaining unchanged performance.

To further enhance the feature extraction capability and training speed of the model, this paper replaces SPPF in the backbone network with SimSPPF. The SimSPPF module is a simplified spatial pyramid pooling module proposed in YOLOv6, which can be used for feature extraction in computer vision tasks.

SimSPPF replaces the activation function Sigmoid Linear Unit (SiLU) of the CBS module in SPPF with a simpler Rectified Linear Unit (ReLU) function, eliminating complex operations and effectively avoiding unnecessary information interference, further improving the computational efficiency of the model. This module first performs convolution operation to compress the input feature map, and then sequentially uses three identical 5 × 5 max pooling layers to obtain feature maps with different receptive field information in a concatenated manner. Then, the feature maps from multiple dimensions are concatenated to finally output the fused feature map.

2.5 Evaluation of algorithm performance

In the task of identifying and classifying wild pear leaves, the precision, efficiency, and complexity of the model are the most important aspects. To assess the performance of different models in the task of wild pear leaf identification and classification, this study employed precision (P), recall (R), and mean Average Precision (mAP) as evaluation metrics. The definitions of these metrics are provided below.

P quantifies the accuracy of positive predictions by the model, defined as the ratio of correctly predicted positive samples to all samples predicted as positive. R measures the model’s ability to identify all actual positive samples, calculated as the ratio of correctly predicted positive samples to the total number of actual positive samples. The formulas for P and R are given in Equations 2, 3, Tang et al. (2023). Here, true positives (TP) are samples correctly predicted as positive, false positives (FP) are samples incorrectly predicted as positive, and false negatives (FN) are samples incorrectly predicted as negative when they are actually positive.

P=TPTP+FP(2)
R=TPTP+FN(3)

To evaluate model performance, a P-R curve is generated with R on the x-axis and P on the y-axis. The area under the P-R curve represents the Average Precision (AP), which measures the model’s detection accuracy for each category. A higher AP value indicates better detection performance. The mAP is the average of AP values across all categories and serves as a key metric for assessing overall model performance. mAP50 is the mAP value at an IoU threshold of 0.5 Yan et al. (2024), while mAP50–95 is the mAP value calculated across multiple IoU thresholds. Higher mAP values correspond to more accurate bounding box predictions. The formulas for AP and mAP are given in Equations 4, 5, where N represents the total number of categories.

AP=01 P(r)dr(4)
mAP=i=1NAP(i)N(5)

Additionally, we assess the lightweight characteristics of the model by considering its size, number of parameters, and FLOPs. Model size refers to the amount of memory the model occupies on the hardware. Parameters indicate the total count of trainable variables in the model, while FLOPs measure the computational workload required by the model. These metrics are crucial for evaluating model complexity and the computational resources needed Zhao et al. (2023).

2.6 Statistical analysis

Confidence intervals (CI) Shreffler and Huecker (2023) can be used to evaluate the reliability of the performance of the model and the significance of differences. The p-value is used to determine whether there are significant differences between models. P<0.05 indicates significant differences. In this study, IBM SPSS 20 Surendran et al. (2024) and a test set from the wild Ussurian Pear leaf dataset were used for the test. Compare the performance metrics of multiple models on the same test set to evaluate the performance of each model and verify whether the improved model in this paper has statistical significance.

3 Results and discussion

3.1 Experimental environment and parameter settings

This experiment is performed on a system running Ubuntu 20.04. Using Python 3.9.19 programming language. The development environment is CUDA 11.8. Use the PyTorch 2.0.0 deep learning framework. The GPU is NVIDIA GeForce RTX 3090. Equipped with 14 vCPU Intel (R) Xeon (R) Gold 6330 CPU @ 2.00GHz processor and 80GB memory.

The network training parameters are set as follows: the image input size is 640 × 640, the batch size is set to 16. The learning epoch is set to 200. Train the model using SGD as the optimizer and dynamically adjust the learning rate using the cosine annealing strategy. The initial learning rate is configured as 0.01, with a momentum factor of 0.937 and a weight decay coefficient of 0.0005. In this study, no pre-trained models are utilized; instead, all models are trained from scratch. To maintain experimental fairness, all models are trained under identical conditions. The ablation experiment uses the same hyperparameter settings, while the comparative experiments of different models use their own default hyperparameter settings.

3.2 Analysis of identification and classification results of wild Ussurian Pear leaves

The identification and classification of wild Ussurian Pear accessions in natural environments is crucial for the conservation of wild pear resources. To this end, the recognition and classification performance of the YOLOv10n-MCS model was tested using the self-made test set of wild Ussurian Pear leaves images dataset. Randomly select five images from the test set for heatmap visualization. Heatmap is a commonly used visual tool for displaying the importance of different regions in an image. The Gradient-weighted Class Activation Mapping (Grad-CAM) Selvaraju et al. (2017) method is employed to generate the model’s heatmap. Grad-CAM calculates the model classification weights and overlays them in the form of heatmap with the original image in equal proportions. To highlight important areas in the image. Figure 7 shows the visual detection results of YOLOv10n and YOLOv10n-MCS on wild Ussurian Pear leaves. Among them, the areas of contribution to model detection are indicated in red and yellow, and the areas with less contribution are blue Wang et al. (2025b). By comparing the heatmap performance of YOLOv10n and YOLOv10n-MCS models in pear leaf classification and detection, significant differences in feature extraction can be found. The attention of the YOLOv10n model is highly focused on the central and lateral vein structures of the leaves, and the heatmap shows a clear linear distribution pattern, indicating that the model mainly relies on leaf vein morphological features for variety classification. The YOLOv10nMCS model demonstrates a more comprehensive feature capture capability, with its heatmap not only focusing on leaf vein structure, but also extensively covering the leaf surface, forming a more uniform activation distribution. This multi-scale feature extraction method enables it to simultaneously capture leaf morphology, texture, and overall contour features. YOLOv10n-MCS has achieved richer feature representation through its multi-channel strategy, which may have stronger discriminative ability and robustness in pear leaf variety classification tasks. This indicates that the method proposed in this study can improve the feature extraction ability of the model and achieve accurate and efficient recognition of wild Ussurian Pear leaves.

Figure 7
Top row: original images of green leaves. Middle row: YOLOv10n processed images showing heatmaps and bounding boxes with labels. Bottom row: YOLOv10n-MCS processed images with more detailed heatmaps and bounding boxes. Images labeled (a) through (e).

Figure 7. Grad-CAM heatmaps of some test set images: (a) P. ussuriensis ‘kuandian-1’, (b) P. ussuriensis ‘dongning-3’, (c) P. ussuriensis ‘xilin-5’, (d) P. ussuriensis ‘Sunwu-7’, (e) P. ussuriensis ‘ningan-2’.

In order to further evaluate the recognition and classification ability of YOLOv10n-MCS on wild Ussurian Pear leaves, Table 2 shows the detection results of YOLOv10n and YOLOv10n-MCS on the recognition and classification of wild Ussurian Pear leaves. The results showed that YOLOv10n-MCS performed better in the task of identifying and classifying wild Ussurian Pear leaves. Compared with YOLOv10n, YOLOv10n-MCS has improved precision, recall, mAP50, and mAP50–95 by 2.9%, 2.3%, 1.5%, and 1.5%, respectively. In addition, the parameters, FLOPs, and model size of the model are reduced by 1.9M, 0.3G, and 0.4MB, respectively, and the model complexity is improved. Perform statistical analysis on precision, recall, and mAP separately. Among them, the precision of YOLOv10n is 94.8 (95% CI: 93.66 to 95.88)%, the recall is 91.2 (95% CI: 89.81 to 92.6)%, and the mAP is 97.3 (95% CI: 96.76 to 97.82)%. The precision of YOLOv10n-MCS is 97.7 (95% CI: 97.18 to 98.16)%, the recall is 93.5 (95% CI: 92.57 to 94.36)%, and the mAP is 98.8 (95% CI: 98.57 to 99.03)%. Compared with the original model, YOLOv10n-MCS showed better precision, recall, and mAP, and the difference is statistically significant (p<0.05). Figure 8 shows the detection results of YOLOv10n and YOLOv10n-MCS on 30 wild Ussurian Pear accessions leaves. From the figure, it can be seen that YOLOv10n has slightly higher recognition precision for the two varieties ‘Xilin-5’ and ‘Xinbin-1’ than YOLOv10n-MCS. In addition, the recognition precision of the other 28 accessions, YOLOv10n-MCS is higher than YOLOv10n. Overall, YOLOv10n-MCS performs better in the task of identifying and classifying wild pear leaves, with a precision of over 92% for all accessions, meeting the demand for high recognition accuracy.

Table 2
www.frontiersin.org

Table 2. The results of identifying and classifying wild Ussurian Pear leaves.

Figure 8
Radar chart comparing YOLOv10n-MCS and YOLOv10n across various categories like Wuchang-2, Sunwu-5, and others. Red squares represent YOLOv10n-MCS, and blue circles represent YOLOv10n. The chart shows different performance metrics ranging from 0.86 to 1.00, with distinct patterns for each model.

Figure 8. Precision results of 30 wild pear accessions.

3.3 Ablation experiments

To assess the performance and practicality of the proposed YOLOv10n-MCS model in wild Ussurian Pear leaves identification and classification, ablation experiments were performed using the YOLOv10n model as the baseline. To maintain experimental consistency, the same dataset, environment, and parameter configurations were utilized throughout the experiments. The YOLOv10n-MCS model consists of three improved methods: MLCA, C2f SCConv, and SimSPPF. The ablation study investigated the effects of three enhancement techniques on model performance. Figure 9 illustrates the loss curves for all models during training and validation. As shown in the figure, the loss function values exhibit a decreasing trend. With increasing epochs, both the training and validation loss curves gradually decline and stabilize, indicating that the models have effectively converged without underfitting or overfitting. The detailed results of the ablation experiments are presented in Table 3. Among them, “✓” represents the improvement methods that have been used.

Figure 9
Two line graphs showing the training loss and validation loss over 200 epochs for different YOLOv10n model variants. The left graph depicts training loss, and the right shows validation loss. Each line represents a different variant, including YOLOv10n, YOLOv10n+MLCA, and YOLOv10n-MCS, among others, with distinct colors. Both graphs show a steep decline initially, stabilizing as epochs increase.

Figure 9. Loss curves of different models.

Table 3
www.frontiersin.org

Table 3. Comparison of ablation experiment results.

According to the experimental results in Table 3, it is found that all combinations improved the performance of the model compared to YOLOv10n. After introducing the MLCA module into the baseline network YOLOv10n (model ①), the precision, recall, mAP50, and mAP50–95 of model ② are improved by 2.3%, 2.4%, 0.9%, and 0.9%, respectively. The FLOPs have increased by 0.2G, and there is no significant change in the parameters compared to the original. MLCA divides the input feature map into multiple local blocks and preserves important spatial information within each block separately. Then calculate the global channel attention of the entire feature map separately. Combine local and global attention weights to create a comprehensive attention map that includes both local details and global contextual information. Weighting the original feature map with the comprehensive attention map highlights important features and suppresses irrelevant information. Finally, the enhanced feature map is input into the network, which improves the accuracy of detection and classification. The experimental results demonstrate that the MLCA module can effectively enhance the feature extraction capability of the model and achieve a better balance between detection performance, speed, and model parameter quantity. Therefore, introducing MLCA module can improve model performance without excessively increasing the parameters and FLOPs. After introducing the SimSPPF module separately, compared with the baseline model, model ③ improved its precision, recall, mAP50, and mAP50–95 by 1.8%, 2.6%, 0.9%, and 1%, respectively, without significant changes in the parameters and FLOPs. SimSPPF uses SimConv convolutional layers to extract features from the input feature map, accelerating the training process and improving the stability of the model, enhancing its expressive power. And through multiple max pooling and concatenation operations, the fusion of features at different scales is achieved. The simplified design improves computational efficiency and enhances the training speed of the model without increasing complexity. Similarly, after introducing the C2f SCConv module separately, the precision, recall, mAP50, and mAP50–95 of model ④ are improved by 2.1%, 1.3%, 0.6%, and 0.5%, respectively. At the same time, the parameters in the model decrease by 0.19M, the FLOPs decrease by 0.6G, and the model size decreases by 0.4MB. C2f SCConv introduces spatial and channel reconstruction modules in convolution operations. Learn the spatial correlation of feature maps through the spatial reconstruction module, and learn the channel correlation of feature maps through the channel reconstruction module. The use of segmentation transformation fusion strategy reduces redundancy and computational costs. By separating and reconstructing redundant features, spatial redundancy is suppressed and feature representation is enhanced. The experimental results indicate that the C2f SCConv module can reduce redundant features, decrease parameters, and lower FLOPs. It not only improve the model’s detection capabilities but also lowers its computational demands.

In addition, SimSPPF and C2f SCConv modules are introduced on the basis of MLCA module. It is found that when the SimSPPF module is introduced, the precision, recall, and mAP of model ⑤ do not significantly improve compared to model ②. The parameters, FLOPs, and model size remain unchanged, and the complexity of the model is not improved. When the C2f SCConv module is introduced, compared with model ②, the precision, recall, mAP50, and mAP50–95 of model ⑥ are reduced by 0.6%, 0.6%, 0.2%, and 0.3%, respectively. However, the parameters, FLOPs, and model size are reduced by 0.19M, 0.5G, and 0.4M, respectively, and the complexity of the model is reduced. Experiment ⑦ introduced the C2f SCConv module on top of the SimSPPF module in the baseline model. Compared with Experiment ③, its precision, recall, and mAP have decreased, but the parameters, FLOPs, and model size have decreased by 0.19M, 0.6G, and 0.4M, respectively. These results indicate that introducing a single module alone in the task of identifying and classifying wild pear leaves doesn’t effectively improve model performance, but may instead lead to a decrease in performance. Therefore, a holistic consideration of multiple factors is essential to enhance model performance.

Finally, the model introduced MLCA module, SimSPPF module, and C2f SCConv module simultaneously. The precision, recall, mAP50, and mAP50–95 of model ⑧ reached 97.7%, 93.5%, 98.8%, and 94.7%, respectively, which are improved by 2.9%, 2.3%, 1.5%, and 1.5% compared to model ①. At the same time, the parameters, FLOPs, and model size of the model are reduced by 0.19M, 0.3G, and 0.4MB, respectively. While improving the accuracy of the model in identifying wild Ussurian Pear leaves, the complexity of the model is reduced, thereby improving the detection performance of the model.

3.4 Comparison and analysis of different network models

To further validate the effectiveness of the YOLOv10n-MCS model, we compared it with eight mainstream object detection models, including two common models YOLOv7n and YOLOv8n, two other models in the YOLOv10 series YOLOv10s and YOLOv10m, VGG16 Simonyan (2014), ResNet50 He et al. (2016), RT-DETR Zhao et al. (2024), and the baseline model YOLOv10n. And analyze the experimental results. Train these eight models using the wild pear leaves dataset in the same experimental environment, with each model undergoing 200 iterations. Subsequently, the model is evaluated using a validation set and the results are compared with YOLOv10n-MCS.

Figure 10 shows the mAP50 values for all models. It can be seen that the mAP50 value of YOLOv10nMCS is consistently better than other models. The precision and recall curves of all detection models are shown in Figure 11. Table 4 displays detailed information on the recognition and classification results of all models. From the table, it can be seen that YOLOv10n-MCS achieves higher detection accuracy compared to other models. Its precision reached 97.7% and mAP50 reached 98.8%, higher than the other eight models. Its mAP50 values are 11%, 3.5%, 1.5%, 2.3%, 2%, 0.4%, 9.6%, and 8.4% higher than YOLOv7n, YOLOv8n, YOLOv10n, YOLOv10s, YOLOv10m, RT-DETR, VGG16, and ResNet50, respectively. YOLOv7n and YOLOv8n use traditional C3 modules with limited receptive fields and insufficient feature extraction capabilities. YOLOv10n-MCS uses large kernel deep convolution to expand the receptive field and enhance the detection capability of targets. And by introducing MLCA and SimSPPF modules, the feature extraction capability of the model has been effectively enhanced. The architecture design concept of VGG16 and ResNet50 adopts a classification oriented static architecture, while YOLOv10n-MCS solves the inherent defects of traditional classification models in object detection tasks through detection specific architecture design and dynamic feature fusion. Therefore, it is possible to more accurately identify and classify wild pear leaves.

Figure 10
Line graph comparing performance of different models including YOLOv7n, YOLOv8n, YOLOv10n, YOLOv10s, YOLOv10m, RT-DETR, VGG16, ResNet50, and YOLOv10n-MCS. The x-axis represents epochs from 0 to 200, and the y-axis represents mAP50 from 0 to 1. Most models reach high performance after 20 epochs.

Figure 10. mAP50 curves of different detection models.

Figure 11
Two line graphs display precision and recall over training epochs for various models. The left graph shows precision, while the right graph shows recall. Each line represents a different model, including YOLO versions, RT-DETR, VGG16, and ResNet50. Precision and recall improve and stabilize as epochs increase. A legend indicates the color for each model line.

Figure 11. Precision and recall curves of different detection models.

Table 4
www.frontiersin.org

Table 4. Comparison of detection performance of different models.

Statistical analysis is conducted on the precision, recall, and mAP of YOLOv10n-MCS compared to 8 other models. Among them, its precision and mAP are significantly higher than YOLOv7n, YOLOv8n, YOLOv10n, YOLOv10s, YOLOv10m, VGG16, and ResNet50 (p<0.05). There is no significant difference comparad to RT-DETR (p=0.64 and 0.38, respectively). Its recall is significantly higher than YOLOv7n, YOLOv8n, YOLOv10n, VGG16, and ResNet50 (p<0.05). There is no significant difference between YOLOv10s and YOLOv10m (p=0.25 and 0.51, respectively). The recall of RT-DETR is significantly higher than that of YOLOv10n-MCS (p<0.05), but its parameters, FLOPs, and model size are 32.87M, 108.1G, and 66.3MB, respectively, which are too high compared to YOLOv10n-MCS. RT-DETR adopts an end-to-end detection architecture based on Transformer, which has computational redundancy and requires more computing resources and larger memory for training. High model complexity hinders its suitability for lightweight tasks. YOLOv10n-MCS uses depthwise separable convolution and channel compression techniques to reduce the model parameters and FLOPs. The FLOPs of YOLOv8n and YOLOv10n-MCS is the smallest, at 8.2G. However, YOLOv10n-MCS reduces redundancy and computational costs by introducing the C2f SCConv module. Compared to YOLOv8n, YOLOv10n-MCS has fewer parameters and lower model complexity. Its model size is 5.4MB, which is more lightweight compared to other models. Overall, the YOLOv10n-MCS model performs better. Therefore, it is more suitable for the task of identifying and classifying wild Ussurian Pear leaves.

4 Conclusion

Wild Ussurian Pear contains abundant genetic resources and is a good material for genetic improvement. Efficient and accurate identification and classification of wild Ussurian Pear accessions are the basis for resource collection, preservation, research and utilization. At present, there are no reports on the classification of wild pear germplasm resource by identifying leaves. Therefore, this article collected wild Ussurian Pear leaves images in the natural background and constructed a dataset of leaves images covering 30 accessions. And using YOLOv10n as the baseline model, a lightweight model called YOLOv10n-MCS was proposed for the recognition and classification of wild Ussurian Pear leaves in complex scenes. We have introduced the MLCA module based on YOLOv10n to enhance the feature extraction capability of the model. Use SimSPPF module instead of SPPF in the baseline model to improve the detection efficiency of the model. C2f SCConv module was designed to replace C2f in the original network backbone, reducing the computational redundancy of the model. We used a dataset of wild Ussurian Pear leaves images to validate the performance of the improved model. The experimental results showed that YOLOv10n-MCS achieved recognition precision, recall, mAP50, and mAP50–95 of 97.7%, 93.5%, 98.8%, and 94.7% for 30 accessions of wild Ussurian Pear leaves, respectively. The precision of 18 wild pear accessions can reach over 97%, while the precision of the other 12 accessions remains between 92% and 97%. Among them, the recognition precision of P. ussuriensis ‘Xinbin-1’ can reach 100%, meeting the accuracy requirements for wild Ussurian Pear leaves recognition and classification tasks. Statistical analysis was conducted on YOLOv10n and YOLOv10n-MCS, and the results showed that YOLOv10n-MCS exhibited statistically significant improvements in precision (2.9% improvement, p<0.05), recall (2.3% improvement, p<0.05), and mAP (1.5% improvement, p<0.05). Further validated the effectiveness of the improvement, particularly in terms of precision and recall.

Comparing the performance of the model with 8 mainstream models, the results show that YOLOv10nMCS has advantages in recognition precision, recall, and model size. It can improve detection performance while reducing parameter and FLOPs. This model can quickly and accurately identify wild Ussurian Pear leaves, outperforming baseline model and other mainstream models in automatic recognition and classification tasks of wild pear leaves.

This study demonstrates the feasibility of using object detection algorithms to identify wild Ussurian Pear leaves for accessions classification. This method can quickly identify wild Ussurian Pear germplasm resource while ensuring accuracy, reducing labor costs. The model proposed in this article effectively meets the requirements of accuracy and real-time performance, which helps to automate the identification of wild Ussurian Pear accessions and provides technical support and reference for the protection, utilization, classification research of wild pear germplasm resource. However, due to the limited sample size in this experiment and significant differences between different populations, its widespread use still requires continuous optimization of performance according to specific requirements and adaptation to new challenges. Nevertheless, the results of this study still provide a good template for achieving this goal.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

NL: Writing – original draft, Conceptualization. XD: Funding acquisition, Supervision, Writing – review & editing. YW: Conceptualization, Investigation, Supervision, Writing – review & editing. LT: Data curation, Formal Analysis, Writing – review & editing. YZ: Investigation, Resources, Writing – review & editing. HH: Visualization, Writing – review & editing. DQ: Project administration, Validation, Writing – review & editing. JX: Investigation, Software, Writing – review & editing. CL: Formal Analysis, Visualization, Writing – review & editing. ZC: Investigation, Software, Writing – review & editing. YM: Data curation, Resources, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by The Science and Technology Innovation Project of the Chinese Academy of Agricultural Sciences (CAAS-ASTIP-RIP) and the Earmarked Fund for the China Agriculture Research System (CARS-28-01).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2025.1588626/full#supplementary-material

References

Adão, T., Shahrabadi, S., Mendes, J., Bastardo, R., Magalhães, L., Morais, R., et al. (2025). Advancing digital ampelography: Automated classification of grapevine varieties. Comput. Electron. Agric. 229, 109675. doi: 10.1016/j.compag.2024.109675

Crossref Full Text | Google Scholar

Adarsh, P., Rathi, P., and Kumar, M. (2020). “Yolo v3-tiny: Object detection and recognition using one stage improved model,” in 2020 6th international conference on advanced computing and communication systems (ICACCS) (Piscataway, NJ: IEEE), 687–694. doi: 10.1109/ICACCS48705.2020.9074315

Crossref Full Text | Google Scholar

Bell, R. (1996). Fruit breeding, tree and tropical fruits[M]. New York, NY: John Wiley & Sons. doi: 10.3389/fnins.2013.12345

Crossref Full Text | Google Scholar

Bochkovskiy, A., Wang, C.-Y., and Liao, H.-Y. M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934. doi: 10.48550/arXiv.2004.10934

Crossref Full Text | Google Scholar

Cao, Y., Tian, L., Gao, Y., and Liu, F. (2012). Genetic diversity of cultivated and wild ussurian pear (pyrus ussuriensis maxim.) in China evaluated with m13-tailed ssr markers. Genet. Resour. Crop Evol. 59, 9–17. doi: 10.1007/s10722-011-9661-1

Crossref Full Text | Google Scholar

Chen, J., Han, J., Liu, C., Wang, Y., Shen, H., and Li, L. (2022). A deep-learning method for the classification of apple varieties via leaf images from different growth periods in natural environment. Symmetry 14, 1671. doi: 10.3390/sym14081671

Crossref Full Text | Google Scholar

Fan, X., Sun, T., Chai, X., and Zhou, J. (2024). Yolo-wdnet: A lightweight and accurate model for weeds detection in cotton field. Comput. Electron. Agric. 225, 109317. doi: 10.1016/j.compag.2024.109317

Crossref Full Text | Google Scholar

Fang, W., Wang, L., and Ren, P. (2019). Tinier-yolo: A real-time object detection method for constrained environments. IEEE Access 8, 1935–1944. doi: 10.1109/ACCESS.2019.2961959

Crossref Full Text | Google Scholar

Farhadi, A. and Redmon, J. (2018). “Yolov3: An incremental improvement,” in Computer vision and pattern recognition, vol. 1804. (Springer Berlin/Heidelberg, Germany), 1–6. doi: 10.48550/arXiv.1804.02767

Crossref Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 770–778. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

Hu, J., Shen, L., and Sun, G. (2018). “Squeeze-and-excitation networks,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 7132–7141. doi: 10.1109/CVPR.2018.00745

Crossref Full Text | Google Scholar

Jocher, G. (2020). Ultralytics YOLOv5 (v7.0). Available online at: https://github.com/ultralytics/yolov5.

Google Scholar

Jocher, G., Chaurasia, A., and Qiu, J. (2023). Ultralytics yolov8 (v8.0.0). Available online at: https://github.com/ultralytics/ultralytics.

Google Scholar

Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., et al. (2022). Yolov6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976. doi: 10.48550/arXiv.2209.02976

Crossref Full Text | Google Scholar

Li, X., Wang, W., Hu, X., and Yang, J. (2019). “Selective kernel networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 510–519. doi: 10.1109/CVPR.2019.00060

Crossref Full Text | Google Scholar

Li, J., Wen, Y., and He, L. (2023). “Scconv: Spatial and channel reconstruction convolution for feature redundancy,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Piscataway, NJ: IEEE), 6153–6162. doi: 10.1109/CVPR52729.2023.00596

Crossref Full Text | Google Scholar

Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., et al. (2014). “Microsoft coco: Common objects in context,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V, vol. 13. (Berlin-German: Springer), 740–755. doi: 10.1007/978-3-319-10602-1_48

Crossref Full Text | Google Scholar

Liu, J., Wang, H., Fan, X., Zhang, Y., Sun, L., Liu, C., et al. (2024). Establishment and application of a multiple nucleotide polymorphism molecular identification system for grape cultivars. Scientia Hortic. 325, 112642. doi: 10.1016/j.scienta.2023.112642

Crossref Full Text | Google Scholar

Niu, W., Lei, X., Li, H., Wu, H., Hu, F., Wen, X., et al. (2024). Yolov8-ecfs: A lightweight model for weed species detection in soybean fields. Crop Prot. 184, 106847. doi: 10.1016/j.cropro.2024.106847

Crossref Full Text | Google Scholar

Pan, B., Liu, C., Su, B., Ju, Y., Fan, X., Zhang, Y., et al. (2024). Research on species identification of wild grape leaves based on deep learning. Scientia Hortic. 327, 112821. doi: 10.1016/j.scienta.2023.112821

Crossref Full Text | Google Scholar

Pu, F. and Wang, Y. (1963). Pomology of China: Pears Vol. III (Shanghai, China: Shanghai Science & Technological Press).

Google Scholar

Redmon, J. (2016). “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (Piscataway, NJ: IEEE). doi: 10.1109/CVPR.2016.91

Crossref Full Text | Google Scholar

Redmon, J. and Farhadi, A. (2017). “Yolo9000: better, faster, stronger,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 7263–7271. doi: 10.1109/CVPR.2017.690

Crossref Full Text | Google Scholar

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). “Grad-cam: Visual explanations from deep networks via gradient-based localization,” in Proceedings of the IEEE international conference on computer vision (Piscataway, NJ: IEEE), 618–626. doi: 10.1109/ICCV.2017.74

Crossref Full Text | Google Scholar

Shreffler, J. and Huecker, M. R. (2023). “Hypothesis testing, p values, confidence intervals, and significance,” in StatPearls (Treasure Island, FL: StatPearls Publishing).

PubMed Abstract | Google Scholar

Simonyan, K. (2014). Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556. doi: 10.48550/arXiv.1409.1556

Crossref Full Text | Google Scholar

Surendran, R., Sree, C. P., and Sumathy, K. (2024). “Texture-and symptom-based coffee leaf disease classification using cnn and resnet algorithms,” in 2024 4th International Conference on Advancement in Electronics & Communication Engineering (AECE) (Piscataway, NJ: IEEE), 1287–1292. doi: 10.1109/AECE62803.2024.10911242

Crossref Full Text | Google Scholar

Tang, Z., He, X., Zhou, G., Chen, A., Wang, Y., Li, L., et al. (2023). A precise image-based tomato leaf disease detection approach using plpnet. Plant Phenomics 5, 42. doi: 10.34133/plantphenomics.0042

PubMed Abstract | Crossref Full Text | Google Scholar

Thakur, M., Ajender, Thakur, D., Thakur, K., Rimpika, and Thakur, K. (2024). “Pear genetic resources and utilization,” in Fruit and Nut Crops (Berlin, German: Springer), 119–140. doi: 10.1007/978-981-99-1586-64-1

Crossref Full Text | Google Scholar

Wan, D., Lu, R., Shen, S., Xu, T., Lang, X., and Ren, Z. (2023). Mixed local channel attention for object detection. Eng. Appl. Artif. Intell. 123, 106442. doi: 10.1016/j.engappai.2023.106442

Crossref Full Text | Google Scholar

Wang, C.-Y., Bochkovskiy, A., and Liao, H.-Y. M. (2023a). “Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 7464–7475. doi: 10.1109/CVPR52729.2023.00721

Crossref Full Text | Google Scholar

Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., et al. (2024). Yolov10: Real-time end-to-end object detection. arXiv preprint arXiv:2405.14458. doi: 10.48550/arXiv.2405.14458

Crossref Full Text | Google Scholar

Wang, Z., Cui, J., and Zhu, Y. (2023b). Review of plant leaf recognition. Artif. Intell. Rev. 56, 4217–4253. doi: 10.1007/s10462-022-10278-2

Crossref Full Text | Google Scholar

Wang, J., Qi, Z., Wang, Y., and Liu, Y. (2025b). A lightweight weed detection model for cotton fields based on an improved yolov8n. Sci. Rep. 15, 457. doi: 10.1038/s41598-024-84748-8

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., and Hu, Q. (2020). “Eca-net: Efficient channel attention for deep convolutional neural networks,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (Piscataway, NJ: IEEE), 11534–11542. doi: 10.1109/CVPR42600.2020.01155

Crossref Full Text | Google Scholar

Wang, C.-Y., Yeh, I.-H., and Mark Liao, H.-Y. (2025a). “Yolov9: Learning what you want to learn using programmable gradient information,” in European conference on computer vision (Berlin, German: Springer), 1–21. doi: 10.1007/978-3-031-72751-11

Crossref Full Text | Google Scholar

Wei Tan, J., Chang, S.-W., Abdul-Kareem, S., Yap, H. J., and Yong, K.-T. (2018). Deep learning for plant species classification using leaf vein morphometric. IEEE/ACM Trans. Comput. Biol. Bioinf. 17, 82–90. doi: 10.1109/TCBB.2018.2848653

PubMed Abstract | Crossref Full Text | Google Scholar

Xue, L., Liu, Q., Qin, M., Zhang, M., Wu, X., and Wu, J. (2017). Genetic variation and population structure of “zangli” pear landraces in tibet revealed by ssr markers. Tree Genet. Genomes 13, 1–11. doi: 10.1007/s11295-017-1110-7

Crossref Full Text | Google Scholar

Yan, C., Liang, Z., Yin, L., Wei, S., Tian, Q., Li, Y., et al. (2024). Afm-yolov8s: An accurate, fast, and highly robust model for detection of sporangia of plasmopara viticola with various morphological variants. Plant Phenomics 6, 246. doi: 10.34133/plantphenomics.0246

PubMed Abstract | Crossref Full Text | Google Scholar

Yang, M., Yuan, W., and Xu, G. (2024). Yolox target detection model can identify and classify several types of tea buds with similar characteristics. Sci. Rep. 14, 2855. doi: 10.1038/s41598-024-53498-y

PubMed Abstract | Crossref Full Text | Google Scholar

Zhang, Y., Cao, Y.-F., Huo, H.-L., Xu, J.-Y., Tian, L.-M., Dong, X.-G., et al. (2022a). An assessment of the genetic diversity of pear (pyrus l.) germplasm resources based on the fruit phenotypic traits. J. Integr. Agric. 21, 2275–2290. doi: 10.1016/S2095-3119(21)63885-6

Crossref Full Text | Google Scholar

Zhang, Y., Shi, N., Zhang, H., Zhang, J., Fan, X., and Suo, X. (2022b). Appearance quality classification method of huangguan pear under complex background based on instance segmentation and semantic segmentation. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.914829

PubMed Abstract | Crossref Full Text | Google Scholar

Zhao, R., Guan, Y., Lu, Y., Ji, Z., Yin, X., and Jia, W. (2023). Fcos-lsc: A novel model for green fruit detection in a complex orchard environment. Plant Phenomics 5, 69. doi: 10.34133/plantphenomics.0069

PubMed Abstract | Crossref Full Text | Google Scholar

Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., et al. (2024). “Detrs beat yolos on real-time object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (Piscataway, NJ: IEEE), 16965–16974. doi: 10.1109/CVPR52733.2024.01605

Crossref Full Text | Google Scholar

Keywords: identification, Ussurian Pear, leaves, YOLOv10n, target detection

Citation: Li N, Dong X, Wu Y, Tian L, Zhang Y, Huo H, Qi D, Xu J, Liu C, Chen Z and Mou Y (2025) Identification of leaves of wild Ussurian Pear (Pyrus ussuriensis) based on YOLOv10n-MCS. Front. Plant Sci. 16:1588626. doi: 10.3389/fpls.2025.1588626

Received: 14 March 2025; Accepted: 10 June 2025;
Published: 03 July 2025.

Edited by:

Kai Huang, Jiangsu Academy of Agricultural Sciences (JAAS), China

Reviewed by:

ChangQing Yu, Xijing University, China
Yulong Nan, Yancheng Institute of Technology, China

Copyright © 2025 Li, Dong, Wu, Tian, Zhang, Huo, Qi, Xu, Liu, Chen and Mou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xingguang Dong, ZG9uZ3hpbmdndWFuZ2RlQDEyNi5jb20=; Yongqing Wu, eXF3dXl5d3VAMTYzLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.