- 1School of Future Technology, Fujian Agriculture and Forestry University, Fuzhou, China
- 2Concore University College, Fujian Normal University, Fuzhou, China
- 3Minnan University of Science and Technology, Quanzhou, China
- 4Fujian Key Lab of Agricultural Internet of Things Applications, Sanming University, Sanming, China
With the development of smart agriculture, fruit picking robots have attracted widespread attention as one of the key technologies to improve agricultural productivity. Visual perception technology plays a crucial role in fruit picking robots, involving precise fruit identification, localization, and grasping operations. This paper reviews the research progress in the visual perception technology for fruit picking robots, focusing on key technologies such as camera types used in picking robots, object detection techniques, picking point recognition and localization, active vision, and visual servoing. First, the paper introduces the application characteristics and selection criteria of different camera types in the fruit picking process. Then, it analyzes how object detection techniques help robots accurately recognize fruits and achieve efficient fruit classification. Next, it discusses the picking point recognition and localization technologies, including vision-based 3D reconstruction and depth sensing methods. Subsequently, it elaborates on the adaptability of active vision technology in dynamic environments and how visual servoing technology achieves precise localization. Additionally, the review explores robot mobility perception technologies, focusing on V-SLAM, mobile path planning, and task scheduling. These technologies enhance harvesting efficiency across the entire orchard and facilitate better collaboration among multiple robots. Finally, the paper summarizes the challenges in current research and the future development trends, aiming to provide references for the optimization and promotion of fruit picking robot technology.
1 Introduction
With the continuous growth of the global population, agricultural production is facing increasingly severe challenges. Rising labor costs, increased labor intensity for farmers, and low agricultural productivity have become key factors limiting the development of modern agriculture (Zhang et al., 2024). To address these issues, agricultural automation technologies have emerged, with intelligent harvesting robots receiving widespread attention and research as an efficient and automated solution (Chunjiang et al., 2023). With the rapid advancements in artificial intelligence, robotics, and computer vision technologies, fruit harvesting robots have gradually become a focal point of research.
In order to provide a comprehensive understanding of the research trends in this field, we conducted a statistical analysis of related research articles from 2005 to 2024 based on the Web of Science database, as show in Figure 1. The results show a significant increase in the number of publications in the field of “Fruit Harvesting,” rising from 732 articles in 2005 to 2130 in 2024. This indicates that, with the rapid development of smart agriculture technologies, the research interest in this field has continued to grow, with visual perception and robotics technologies gradually becoming the core focus of research.
1.1 Development status of intelligent fruit harvesting robots
In 1968, the United States pioneered the study of fruit harvesting using mechanical or pneumatic vibration methods. Although these methods could perform basic harvesting tasks, vibration and pneumatics often caused significant damage to the fruit (Schertz and Brown, 1968). With the development of computer and control technologies, agricultural robots began to be applied in tasks such as harvesting, spraying, and weeding from the 1990s onward, assisted by computer vision. In particular, some robotic arms were able to simulate manual harvesting actions. However, due to the limitations of robot and sensor technologies at the time, automated harvesting robots still faced challenges such as high costs, low precision, and limited application scenarios. With the rapid development of Industry 4.0, advancements in artificial intelligence, the Internet of Things, and big data analysis have greatly propelled the progress of agricultural harvesting robots, especially in the precision of perception, autonomous decision-making, control, and execution (Oliveira et al., 2021). In particular, the continuous innovation of visual perception systems has provided harvesting robots with more powerful sensing capabilities. Modern intelligent fruit harvesting robots are now able to obtain real-time environmental information through devices such as cameras, LiDAR, and depth sensors, and identify the type, location, and status of fruits using image processing and pattern recognition technologies.
Harvesting robots can be divided into bulk harvesting robots and selective harvesting robots (Zhou et al., 2022). As shown in Figures 2a-c, bulk harvesting robots are typically large and perform one-time harvesting by applying vibration or forced separation to the fruit trees. Examples include apple harvesting by vibrating branches (De Kleine and Karkee, 2015), cherry harvesting by vibration (Zhou et al., 2016), and bulk grape harvesting for industrial use (Yan et al., 2023). Although bulk harvesting methods are efficient, they cause significant damage to the fruits and are difficult to distinguish based on ripeness, making them suitable only for industrial fruit, not for those intended for market sales.

Figure 2. Various types of harvesting robots. (a) Olive Shaking Bulk Harvesting Equipment (Sola-Guirado et al., 2023), (b) Apple Vibration Harvesting Robot (De Kleine and Karkee, 2015), (c) Apple Selective Harvesting Large-Scale Equipment (Jia et al., 2020), (d) Sweet Pepper Harvesting Robot (Arad et al., 2020), (e) Tomato Harvesting Robot (Rapado-Rincón et al., 2023), (f) Cucumber harvesting robot (Park et al., 2023).
Selective harvesting robots typically install the end effector on a robotic arm and use computer vision to identify ripe fruits, guiding the robotic arm and end effector to perform the harvesting task, as shown in Figures 2d-f. These devices are usually smaller in size and can move freely in agricultural environments. Since their harvesting method is the closest to human picking, they have already been applied in harvesting fruits such as apples (Jia et al., 2020), peppers (Arad et al., 2020), tomatoes (Rapado-Rincón et al., 2023), yellow peaches (Wang et al., 2023), and strawberries (Tafuro et al., 2022). Although selective harvesting robots have lower work efficiency, they support batch harvesting and effectively reduce fruit damage, thus preserving the commercial value of the fruits. This method is considered the most likely to fully replace human harvesters, which has led to widespread attention to selective harvesting robots in both academia and industry (Sanders, 2005).
1.2 The importance of visual perception technology in fruit harvesting
Visual perception technology plays a pivotal role in intelligent fruit harvesting robots, serving as one of the core technologies enabling automated picking. It facilitates the accurate identification and localization of target fruits through image processing and object detection, ensuring the efficient and precise execution of harvesting tasks. The visual system must adapt to varying lighting conditions, diverse fruit types, and complex background environments. The application of deep learning, 3D reconstruction, and image segmentation techniques enhances its robustness and accuracy. Furthermore, visual perception supports dynamic decision-making for the robot, allowing real-time adjustments to harvesting strategies in response to fruit displacement or occlusion, thereby ensuring operational continuity and stability. With technological advancements, the introduction of visual servo systems and closed-loop control has further improved manipulation precision and minimized fruit damage.
Scholars have developed models for detecting picking points using image analysis and deep learning techniques to guide robotic manipulators in the intelligent harvesting of fruits such as pepper (Arad et al., 2020; Babellahi et al., 2020), tomato (Jun et al., 2021; Wu et al., 2021), apple (Jia et al., 2020; Li C. et al., 2023), and grape (Yan et al., 2023; Wang J. et al., 2024). In intricate field settings, factors such as fluctuating illumination, fruit overlap, variations in fruit maturity, accurate peduncle/stem recognition, and precise localization of the picking point significantly impact the operational efficiency and harvesting accuracy of robots. Concurrently, when fruits are occluded, determining the optimal viewing angle for observation and planning effective manipulator trajectories become critical challenges for enhancing harvesting performance. Therefore, a thorough examination of the latest advancements, existing challenges, and future trends in visual perception technology for fruit harvesting robots holds substantial academic significance and practical value for advancing the field.
2 Common camera types for harvesting robots
Efficient visual perception systems are fundamental to intelligent fruit harvesting robots, with cameras serving as core components whose performance is determined by sensor type and design. Driven by advancements in computer vision, deep learning, and sensor technology, traditional RGB cameras are increasingly being supplemented or replaced by various advanced sensors. Combining different sensors proves particularly effective in complex agricultural environments, significantly enhancing system robustness and accuracy. Common vision sensors include monocular cameras, binocular (stereo) cameras, RGB-D cameras and event cameras, each possessing distinct advantages and suitable application scenarios. Table 1 presents a performance comparison of different types of cameras. The following will provide a detailed analysis of these camera types and explore their specific applications in fruit harvesting tasks.
2.1 Monocular camera
Monocular cameras, capturing color images through a single lens, are widely utilized for image acquisition in deep learning applications due to their simple structure and low cost, as shown in Figure 3a. However, they are incapable of directly capturing depth information, providing only two-dimensional scene data, and are primarily used for tasks like object detection and yield estimation. To address this limitation, researchers employ deep learning and other methods to process monocular images and estimate fruit positions (Khan et al., 2020; Cheng et al., 2021; Yin et al., 2023). For instance, Yang et al. proposed a self-supervised monocular depth estimation network (Yang et al., 2020), while Ban et al. tackled depth estimation in defocused images using Markov random fields and geometric constraints (Ban et al., 2022). Despite these efforts, the lack of inherent depth data means monocular depth estimation still relies on computationally intensive methods and achieves limited accuracy. This challenge is particularly pronounced in unstructured agricultural scenes, where environmental complexity and indistinct object features further complicate depth estimation.

Figure 3. Representatives of cameras from different technology types. (a) MOKOSE monocular camera, (b) ZED stereo camera, (c) Intel T265 stereo camera. (d) ORBBEC structured light camera, (e) SEERSENSE ToF (Time of Flight) camera. (f) iniVation event camera.
2.2 Stereo camera
Binocular cameras, as shown in Figures 3b, c, also referred to as stereo cameras, capture images of a scene using two lenses from different viewpoints. They compute object depth by leveraging the principle of parallax (Chao et al., 2023). By mimicking the human binocular vision system to acquire three-dimensional (3D) information, binocular cameras provide depth data more directly compared to monocular cameras. Consequently, they are widely adopted in agricultural robotics and automated harvesting scenarios due to their ability to deliver more accurate spatial localization in complex environments (Ling et al., 2019; Wu et al., 2021; Wen et al., 2022). However, binocular cameras also exhibit certain limitations. For instance, they exhibit a high dependency on texture features within the scene; depth estimation performance may degrade in texture-poor regions or under suboptimal lighting conditions. Furthermore, the hardware configuration of binocular cameras is inherently more complex than that of monocular cameras, demanding precise calibration and stringent synchronization between the two lenses.
2.3 RGB-D camera
To overcome the limitations of monocular and binocular cameras, RGB-D cameras have emerged as a solution. RGB-D cameras integrate an RGB color camera with a depth sensor, enabling simultaneous capture of color information and depth data from the scene, making them an increasingly popular choice for diverse applications. Beyond stereo vision, common methods for acquiring depth information with RGB-D cameras include structured light technology and Time of Flight (ToF) (Zhou et al., 2021) (as shown in Figures 3d, e). Structured light technology typically projects a known light pattern (e.g., stripes, dot arrays) onto object surfaces and captures the resulting deformation of this pattern using a camera to infer depth. Cameras employing this method offer high accuracy at close range and rapid depth acquisition, but depth measurement accuracy may decrease for objects with low reflectivity or lacking texture. Common structured light cameras include the Intel RealSense series, and the Intel RealSense D435 camera, valued for its compact size and high precision, is widely utilized in fruit harvesting tasks (Liu et al., 2024).
ToF calculates distance by emitting light pulses and measuring the time difference for the light to travel from the camera to the object and back. ToF can operate effectively under low-light conditions or significant illumination variations and provides rapid depth acquisition. However, its resolution is generally lower than that of structured light cameras, making it difficult to capture sufficiently detailed depth information in complex, close-range environments.
2.4 Event camera
In addition to conventional frame-based cameras, emerging vision sensors—such as event cameras—have demonstrated significant potential in agricultural applications, particularly in complex environments with high dynamic lighting conditions, as shown in Figure 3f. Unlike traditional cameras that capture entire images at fixed frame rates, event cameras operate using an asynchronous imaging mechanism that records data only when changes in pixel brightness occur (Gallego et al., 2021). Each event contains a timestamp, pixel location, and the polarity of brightness change, enabling ultra-high temporal resolution at the microsecond level, extremely low latency, and substantially reduced data redundancy. One of the most prominent advantages of event cameras is their exceptionally high dynamic range, often exceeding 120 dB, making them particularly effective in agricultural scenarios (Wan et al., 2024). For instance, event cameras can produce stable outputs under highly variable lighting conditions, such as when sunlight filters through foliage or when transitions occur rapidly between shaded and sunlit areas. Furthermore, their low power consumption and compact size make event cameras well-suited for integration into embedded systems and various field-deployed agricultural automation platforms.
In the context of precision agriculture, event cameras offer potential for a variety of tasks, including crop monitoring, where subtle structural changes in plants can be more effectively detected; real-time navigation of agricultural robots and UAVs in dynamically lit environments; and high-speed target detection (Gehrig and Scaramuzza, 2023), such as rapid identification of field animals, tracking of pest movements (Pohle-Fröhlich et al., 2024), or detection of fruit maturity status.
Compared to the aforementioned camera types, RGB-D cameras offer more stable depth perception in complex environments and exhibit reduced dependency on scene texture. They demonstrate superior performance in localization accuracy and computational efficiency (Zhou et al., 2022), making them well-suited for scenarios demanding high-precision depth information, such as agricultural robotics and autonomous driving. Given these advantages, RGB-D cameras have been successfully applied to the harvesting of various fruits (Yoshida et al., 2022).
2.5 Camera installation position
The installation position of the camera directly determines the perception ability of the picking robot toward the fruits. A reasonable installation position can maximize the coverage of the visual perception system, enhance the ability to capture image details, and reduce the interference of external factors on recognition accuracy. Generally, the camera installation positions on a picking robot can be divided into Eye-To-Hand and Eye-In-Hand. Eye-To-Hand means the camera is installed at a fixed position on the robotic arm, typically on the robot’s base, workbench, or another location that does not change with the movement of the robotic arm. For example, Birrell et al. (2020) fixed the camera on a bracket in their lettuce harvesting system, as shown in Figure 4. This method provides stable visual information, but the fixed camera may fail to detect all the fruits due to occlusion. Eye-In-Hand refers to the camera being directly installed at the end of the robotic arm, where each movement of the arm directly affects the camera’s view. For example, Junge et al. (2023) installed an RGB-D camera at the end of the robotic arm in their strawberry picking robot design, with the camera moving along with the arm, as shown in Figure 5. This method is better at handling target localization and manipulation tasks in complex or confined spaces. However, its drawbacks include a larger computational load and a higher risk of the camera being damaged due to accidental collisions.
3 Object detection technology in fruit picking
Objective detection technology is the core technology in intelligent fruit harvesting (Xiao et al., 2024). Due to the vast variety of fruits, which exhibit significant variations in morphology, size, and color, object detection enables the training and optimization of recognition capabilities for different fruit types. Within harvesting tasks, object detection must first precisely locate fruit positions, assess maturity levels, evaluate occlusion conditions, and identify pickable points. Furthermore, it determines the picking sequence by analyzing fruit clustering before robotic arm execution, thereby enhancing harvesting efficiency and accuracy. Object detection techniques are typically categorized into traditional feature-based machine learning methods and deep learning-based approaches.
3.1 Traditional object detection technology
Traditional object detection methods primarily rely on the sliding window strategy and manual feature extraction. These include color features (such as threshold segmentation in HSV, Lab, and other color spaces), texture features (e.g., Gray-Level Co-occurrence Matrix - GLCM and Local Binary Patterns - LBP), and shape features (e.g., edge detection and Hough transform). Due to their distinctiveness and stability, color features are widely employed in fruit recognition, particularly in scenarios with simple backgrounds and high contrast between the fruit and its surroundings. For instance, Arefi et al. achieved an accuracy of 96.36% by combining features extracted from the RGB, HIS, and YIQ color spaces for tomato recognition (Arefi et al., 2011). Tian et al. utilized components of the HIS and LAB color spaces for tomato leaf segmentation (Tian et al., 2019), while Yamamoto et al. implemented target identification for strawberry harvesting through color threshold analysis, achieving a harvest rate of 67% (Yamamoto et al., 2014). In complex agricultural environments, OTSU adaptive thresholding is extensively applied to extract target fruit locations based on color differences (Wei et al., 2014; Lv et al., 2016). While color models prove effective in distinguishing fruits from backgrounds, their performance deteriorates significantly in complex backgrounds or when encountering objects with similar colors.
Morphological characteristics also hold significant importance in traditional methods. Features such as shape can be extracted through edge detection (e.g., the Canny operator) and contour detection (e.g., Hough transform), proving particularly effective for regularly shaped fruits. For instance, Lv et al. achieved fruit recognition by combining RGB color features with the Canny operator and Hough transform (Lv et al., 2015), while Tan et al. utilized Canny edge detection to extract edge features from apples, lemons, and mangoes for subsequent classification using machine learning (Tan et al., 2021). However, the robustness of these traditional methods is often limited in complex scenarios or when detecting occluded fruits. To enhance accuracy, Rabby et al. successfully implemented fruit recognition and classification in controlled background settings by integrating color and morphological features (Rabby et al., 2018). Furthermore, texture features, including but not limited to those derived from the Gray-Level Co-occurrence Matrix (GLCM) and Local Binary Patterns (LBP), play a crucial role in fruit object detection (Aygün and Güneş, 2017; Gurubelli et al., 2020).
Furthermore, Haar-like features (Besnassi et al., 2020) and Histogram of Oriented Gradients (HOG) features (Zhou and Yu, 2021) are also widely employed for image description and fruit recognition. Haar-like features extract discriminative information by computing differences in pixel intensities within rectangular regions. While achieving notable success in facial recognition, this approach has also been effectively applied to fruit detection within the agricultural domain. Conversely, HOG features facilitate classifier recognition of fruits by quantifying the distribution of gradient orientations within localized image regions.
With the advancement of machine learning technologies, traditional methods have progressively been integrated with machine learning classifiers, forming feature-based + classifier frameworks for object detection. These classifiers encompass Support Vector Machines (SVM), Random Forests (RF), KNearest Neighbors (KNN), and Naïve Bayes, among others. For instance, Zhang et al. (Zhang and Wu, 2012) achieved the classification of multiple fruit types using an SVM, achieving an accuracy of 88.2%, while Lin et al. successfully identified six fruit types employing the Hough transform combined with an SVM (Lin et al., 2020). RF enhances classification stability by aggregating predictions from multiple decision trees (Ramisetty et al., 2022), whereas KNN classifies fruits such as apples and dragon fruit based on sample similarity (Aulia et al., 2023). Naïve Bayes performs well in relatively straightforward classification scenarios, demonstrating effectiveness in non-destructive testing applications for apples (Miriti, 2016; Yogesh et al., 2021).
Prior to the widespread adoption of deep learning, methods based on handcrafted features and machine learning classifiers constituted the mainstream approach in object detection. Although demonstrating satisfactory performance in simple scenarios, their heavy reliance on manually designed features resulted in suboptimal effectiveness when confronted with complex environments. However, the rise of deep learning has precipitated a paradigm shift, with automated feature learning progressively supplanting handcrafted feature engineering to become the dominant technology in object detection.
3.2 Object detection technology based on deep learning
Driven by the advancement of agricultural automation and intelligence, the application of deep learning technologies in fruit harvesting has emerged as a prominent research focus. Fruit harvesting confronts multiple challenges, including object recognition in complex environments, identification and localization of diverse fruit types, maturity assessment, and occlusion handling. Traditional manual or mechanical methods are often characterized by low efficiency, high costs, and significant environmental constraints. In contrast, deep learning techniques, particularly Convolutional Neural Networks (CNNs) and their extensions such as Faster R-CNN, DETR, and YOLO, have significantly propelled the intelligence and automation of fruit harvesting robots.
3.2.1 Two-stage object detection methods
Early object detection methods primarily relied on traditional CNN architectures like LeNet and AlexNet. While successful in image classification tasks, these networks inherently lacked the capability to directly output positional information. To address this limitation, the R-CNN approach proposed by Ross et al. pioneered the two-stage object detection paradigm by combining region proposal generation with deep feature extraction (Girshick, 2015). Subsequent advancements, namely Fast R-CNN and Faster R-CNN, substantially improved detection speed and accuracy through shared convolutional feature maps and the introduction of a Region Proposal Network (RPN) (Ren et al., 2016). The Feature Pyramid Network (FPN) further optimized Faster R-CNN by constructing a pyramid structure on feature maps of different scales, thereby enhancing multi-scale object detection capabilities (Lin et al., 2017). For example, Wan et al. achieved multiclass fruit detection using Faster R-CNN (Wan and Goudos, 2020), while Parvathi et al. applied Faster R-CNN for the detection of coconut maturity in complex backgrounds (Parvathi and Selvi, 2021).
Mask R-CNN is based on Faster R-CNN and achieves precise segmentation and localization of each instance object by adding pixel-level masks (He et al., 2017). This method has been applied to the identification of pick-up points, such as López-Barrios et al. (2023) who used Mask R-CNN to detect green bell peppers in greenhouses, successfully locating pick-up points. Despite the accuracy advantage of two-stage networks, they are computationally expensive and slow. Therefore, with the increasing demand for real-time performance, researchers have gradually shifted toward more efficient one-stage object detection methods.
3.2.2 One-stage object detection methods
The YOLO (You Only Look Once) family represents a milestone in one-stage object detection models by transforming object localization into a regression problem through a fully convolutional architecture, achieving high detection speed (Redmon, 2016). With successive iterations, YOLO models have steadily improved in both accuracy and efficiency. Among earlier versions, YOLOv5 gained widespread adoption in agricultural scenarios due to its streamlined architecture and training efficiency (Wang et al., 2022; Hou et al., 2022). For instance, Sozzi et al. (2022) validated YOLOv5’s reliability in grape cluster detection across YOLOv3, YOLOv4, and YOLOv5 models.
Recent versions have introduced more advanced designs tailored for real-time and complex environments. YOLOv6 incorporates cross-layer feature fusion strategies to enhance real-time performance in industrial contexts (Li C. et al., 2022), while YOLOv8 significantly improves multi-scale object detection and feature extraction (Hussain, 2024). In agricultural applications, Wang et al. (2025) proposed a customized YOLO-ALW model based on YOLOv8, achieving 99.1% mAP in pepper detection tasks.
Further developments from YOLOv9 to YOLOv12 introduced architectural innovations such as reversible branches, the GELAN backbone, and modules like C2f-faster and Area Attention, improving detection precision while reducing inference latency (Khanam and Hussain, 2024; Wang A. et al., 2024; Wang CY. et al., 2024; Tian et al., 2025). Figure 6 presents a comparison of latency (left) and computational complexity (FLOPs, right) against mAP on the MS COCO dataset. YOLOv12 achieves superior mAP while maintaining low latency and FLOPs, demonstrating outstanding overall efficiency. However, Sapkota et al. conducted a comprehensive evaluation of YOLOv8 through YOLOv12 in complex orchard environments and found that YOLOv9 delivered the best performance for green apple detection and counting (Sapkota and Karkee, 2025). Most recently, YOLOv13 introduced HyperACE (Hypergraph Adaptive Correlation Enhancement) and the FullPAD mechanism, further boosting detection performance (Lei et al., 2025). These advances suggest strong potential for future application in intelligent fruit harvesting.
In summary, while newer YOLO variants offer enhanced accuracy and speed, their effectiveness in agricultural environments depends on task-specific factors such as target size, occlusion level, and real-time requirements. Selecting the most suitable version requires careful consideration of these variables.
3.2.3 Transformer-based object detection methods
Originally achieving remarkable success in natural language processing, Transformer architectures have recently been introduced into the field of object detection due to their ability to model global dependencies via self-attention mechanisms. Representative models include DETR (Zhao et al., 2024), Deformable DETR (Zhu et al., 2020), Swin Transformer (Liu et al., 2022), and Vision Transformer (ViT) (Huang et al., 2022). Compared with convolutional neural networks (CNNs), Transformer-based models enable end-to-end training without relying on predefined anchor boxes and offer strong global modeling capabilities, making them particularly suitable for complex agricultural environments with background clutter or occlusion.
Despite these advantages, Transformers still face several challenges in practical applications, including high computational cost, slow convergence, and a strong dependence on large-scale labeled datasets. To address these limitations, Guo et al. proposed a Transformer-based fruit detection framework, which effectively captures long-range dependencies but still struggles with tasks such as small object detection and fruit localization at boundaries (Guo et al., 2024).
To provide a comparative view of detection performance across different fruit types and detection models, Tables 2, 3 summarize the results reported in recent studies. “Results” refers to the reported detection accuracy under specific datasets or field conditions, while “Cycle Time” indicates the average time to complete a full picking cycle for each fruit, including perception, motion planning and execution, and fruit placement. These comparisons help illustrate the trade-offs between detection performance and overall harvesting efficiency across various algorithms and application contexts. To balance real-time performance and accuracy, recent research has begun to explore hybrid models that integrate Transformer modules into YOLO frameworks. Additionally, fusing Transformer features with multi-modal sensor data—such as RGB-Depth or thermal imagery—has emerged as a promising direction for enhancing robustness and accuracy in agricultural detection tasks.

Table 2. Part 1 of the research progress on various fruit harvesting visual perception technologies.

Table 3. Part 2 of the research progress on various fruit harvesting visual perception technologies.
4 Data labeling methods and localization techniques for fruit picking
The localization of picking points determines whether the fruit can be successfully harvested, making it one of the core aspects of the fruit picking process. In recent years, many scholars have focused on the labeling and research of fruit picking points. The methods for data labeling of fruit picking points and their localization and recognition are crucial elements in the research of intelligent harvesting robots. The goal is to ensure the accurate identification and localization of picking points through efficient and precise labeling methods and localization technologies, thereby enhancing the automation and intelligence of the harvesting machinery.
Selective picking methods are classified into two categories based on the way the fruit is harvested: picking the fruit itself and picking the fruit stem. The terminal operation methods differ between these two categories, and there are also significant differences in data labeling approaches. In recent years, many researchers have noted variations in the labeling of data for picking the same type of fruit, and these differences affect the picking accuracy.
Wang et al. applied prior knowledge of apples and used the Hough transform method and contour curvature to propose a method for calculating the contours of occluded apples to enable picking localization (Wang et al., 2016). This method struggles to identify the fruit when they overlap. Yu et al. labeled the strawberry body with a bounding box and used R-YOLO to predict the rotational boundaries of the strawberry and the physical size estimation of the picking point based on the strawberry’s rotation angle to confirm the picking point (Yu et al., 2020), as shown in Figure 7a. Tafuro used instance segmentation to label the strawberry body and calculated the fruit stem position and picking point localization by recognizing the boundary of the strawberry (Figure 7b) (Tafuro et al., 2022). Zhong et al. in their lychee picking labeling, only labeled the main fruit branch and took the center point of the bounding box as the picking point (Zhong et al., 2021), as shown in Figure 7c. If the center point is not exactly on the branch or is blocked by leaves, it can cause significant errors. Figure 7d shows the sweet pepper picking labeling, where both the bounding box and the center point of the fruit are estimated to confirm the picking position (Ning et al., 2022).

Figure 7. Labeling method for picking points of different fruits. (a) Strawberry picking point calculation (Yu et al., 2020), (b) Strawberry picking point calculation (Tafuro et al., 2022), (c) Litchi picking point calculation (Zhong et al., 2021), (d) Sweet pepper picking point (Ning et al., 2022), (e) Viburnum picking point calculation (Lu et al., 2024), (f) Mango picking point calculation (Zheng et al., 2021), (g) Grape picking point calculation (Chen W. et al., 2024), (h) Pepper picking point calculation (Huang et al., 2024), (i) Mango picking point calculation (Li et al., 2024).
Lu et al. in their lotus pod picking used YOLOv5-based instance segmentation to label both the fruit region and the fruit stem region separately, and then calculated the key points from the segmented regions, inferring the picking position from those key points, as shown in Figure 7e (Lu et al., 2024). These various methods show the diversity in approaches to fruit picking point labeling and localization across different fruit types. The key challenge lies in ensuring high accuracy despite differences in fruit shapes, growth environments, and occlusions.
With the development of deep learning technologies, some researchers have shifted the fruit picking point localization from traditional geometric computations to regression-based calculations. Zheng et al. (2021) applied a combination of fruit instance segmentation and key point labeling for mango picking point localization, as shown in Figure 7f. They used the Mask RCNN model to simultaneously perform regression on the instance regions and multiple key points, with the picking point location ultimately determined by the key points. Chen et al (Chen W. et al., 2024), in their grape picking labeling work, used a fruit target bounding box and a fruit stem picking key point to label the data, and directly applied the YOLOv8pose model for regression calculations to achieve picking point localization, as shown in Figure 7g. To address the issue of chili picking points being occluded in complex scenarios, Huang et al. (2024) improved the YOLOv8-pose model by introducing a reversible network structure and a feature fusion module to achieve the recognition of multiple key points of the chili. The precise estimation of the picking points is realized through these key points, with the detection results shown in Figure 7h. Li et al (Li et al., 2024), in their mango picking work, combined object detection and instance segmentation. They first used two target bounding boxes to separately label the mango body and fruit stem, then applied instance segmentation to label the fruit stem region. After detecting the fruit stem using object detection, they performed instance segmentation on the stem region to obtain the skeleton line of the fruit stem, which was then used to calculate the picking point, as shown in Figure 7i.
In summary, we can observe that in recent years, there have been multiple labeling methods and picking point calculation approaches for the same fruit or different fruits with similar picking methods. The accuracy of the models trained or computed with different labeling methods also varies. In complex environments, how to develop a fruit labeling method that serves fruit picking tasks becomes particularly crucial. One of the key challenges in fruit picking work has always been how to minimize the position error of the fruit picking points.
5 Robot mobility and global environment perception technologies
5.1 Visual perception and navigation
Visual perception is one of the core technologies enabling fruit harvesting robots to achieve autonomous navigation and environmental understanding. By integrating Visual Simultaneous Localization and Mapping (V-SLAM) systems, robots can construct 3D maps and localize themselves in complex orchard environments, thereby enhancing their autonomous navigation capabilities. Chen et al. (2021) proposed a framework combining eye-in-hand stereo vision with SLAM, addressing the limitations of traditional SLAM methods in orchard environments and providing a solution for large-scale orchard harvesting that adapts to complex terrain and varying lighting conditions. Maud et al. (2023) utilized object detection and RTAB-Map algorithms to propose a real-time 3D mapping and localization system, optimizing the detection and management of palm oil trees and improving tree localization accuracy in large-scale plantations. Wang P. et al. (2025) based their approach on visual SLAM combined with semantic segmentation networks, improving the representation of point clouds and enhancing real-time processing speed, thus enabling more precise navigation and perception in greenhouse environments. These studies show that the combination of stereo vision with SLAM, particularly with the introduction of semantic SLAM, significantly enhances the robot’s perception and navigation accuracy in complex environments.
5.2 Path planning for mobile robots
Path planning is crucial for fruit harvesting robots to operate efficiently, particularly in complex orchard environments where optimizing paths to minimize time and energy consumption is essential. Urvina et al. (2024) proposed a combined global and local planning strategy, using the Traveling Salesman Problem (TSP) and the Informed Rapidly-exploring Random Tree (IRRT*) algorithm to optimize paths and avoid obstacles, improving navigation efficiency in complex terrain. Wang L. et al. (2022) introduced a full-coverage path planning method based on multi-objective constraints, which enhances the adaptability of path planning algorithms in irregular terrains, ensuring complete coverage. Wang et al. (2025a). developed a hybrid path planning approach, combining inner spiral and improved nested methods, significantly reducing non-work path length and improving operational coverage. These studies highlight the progression of path planning technologies toward combining global and local strategies, addressing path optimization challenges in complex agricultural environments.
5.3 Task scheduling
Task scheduling is vital for enhancing the efficiency of multi-tasking harvesting robots, especially when multiple tasks are performed simultaneously. Efficient task allocation and resource optimization are key to improving robot performance. Li et al. (2023a) proposed a Multi-Agent Reinforcement Learning (MARL)based scheduling method that dynamically adjusts task allocation based on real-time environment changes and task priorities, boosting operational efficiency. Wang et al. (2025b) addressed collaborative scheduling between harvesters and transport robots, introducing a task allocation and path planning method based on topological maps, significantly enhancing operational efficiency. Zhu et al. (2025) developed a task scheduling method for dual-arm robots using Mixed-Integer Linear Programming (MILP), optimizing task coordination and substantially improving strawberry harvesting throughput. These studies demonstrate that incorporating multi-agent systems and optimization algorithms into task scheduling can effectively enhance multi-task coordination and improve overall operational efficiency.
6 Optimal viewpoint planning for fruit picking
During the fruit picking process, environmental factors such as exposure, backlighting, shadows, occlusions, and vibrations may cause changes in the fruit’s position or lead to recognition failures. These factors not only result in the loss of visual information but may also prevent the accurate localization of picking points, ultimately reducing picking efficiency (Suresh Kumar and Mohan, 2023). For example, under strong sunlight or backlighting conditions, the camera may fail to clearly capture the fruit’s outline, while shadowed areas may obscure parts of the fruit, causing recognition errors. Vibration or mechanical movement can also shift the fruit’s position in the visual sensor, further affecting the accuracy and efficiency of the picking task. In addition, different viewpoints may produce varying picking outcomes. To address these issues, viewpoint planning, as an important technical measure, aims to maximize the fruit’s visibility and recognition rate by selecting the most appropriate angle, thereby minimizing the impact of external factors on recognition effectiveness (Yi et al., 2024). Viewpoint planning for fruit picking can be divided into four types based on the methods used: geometry-based viewpoint planning, information-based viewpoint planning, optimization-based viewpoint planning, and learning-based viewpoint planning.
6.1 Geometry-based viewpoint planning method
The geometric-based viewpoint planning method focuses on selecting the optimal viewpoint by calculating the spatial relationships between the environment and the target object. It typically involves using depth cameras or LiDAR to create an environmental model, which includes geometric shapes such as tree structures, fruit positions, and the locations of branches and leaves. The visual system then identifies the position of the target fruit and analyzes the feasibility of viewpoint selection based on the geometric relationship between the fruit and the environment. Once the best viewpoint is selected, it notifies the robotic arm to carry out the picking task. Menon et al. planned the optimal picking viewpoint based on the completeness of the fruit’s shape, as shown in Figure 8a (Menon et al., 2023). Hornung et al. proposed a 3D point cloud mapping based on octrees to simulate the robot’s 3D environment (Hornung et al., 2013). RVP constructed a voxel map of the fruit region and used a utility function based on expected information of the fruit region to evaluate candidate viewpoints (Zaenker et al., 2021). Burusa et al. drove next-best-view (NBV) planning through the tomato plant’s structural features and an attention mechanism (Burusa et al., 2024).

Figure 8. Different methods for calculating the optimal viewpoint. (a) Evaluate the picking point location through fruit shape completion, (b) Calculate the unobstructed areas of grape picking points using a scoring function, (c) Identify the optimal viewpoint through deep learning.
These methods have high computational complexity, are heavily dependent on equipment, and may become ineffective if the environment changes, such as when leaves or fruits sway, making pre-computed optimal viewpoints unsuitable.
6.2 Information-based and optimization-based viewpoint planning methods
Information-based and optimization-based viewpoint planning methods evaluate the characteristics of different viewpoints to select the ones that provide the maximum perceptual information or optimize task execution. These methods are widely applied in complex scenarios, such as fruit harvesting tasks. Yi et al. generated viewpoints randomly and guided the robotic arm to adjust its perspective by combining spatial coverage and motion cost to optimize the scoring function, as shown in Figure 8b (Yi et al., 2024); Menon et al. estimated missing information through shape completion and used an NBV-SC planner to find the best viewpoint (Menon et al., 2023); Akshay et al. made multi-viewpoint semantic perception decisions to determine the best viewpoint in tomato harvesting, achieving better results than active vision strategies (Burusa et al., 2024); Zaenker et al. designed a viewpoint motion planner to optimize the information gain for pepper detection (Zaenker et al., 2023). These methods require evaluating multiple viewpoints, resulting in a large computational load that affects real-time performance. Optimization-based viewpoint planning, on the other hand, uses optimization algorithms to select viewpoints, with objectives typically focused on minimizing occlusion, maximizing information gain, or improving task efficiency. These methods evaluate the quality of viewpoints by setting objective functions. For example, Li et al. improved YOLOv5 and combined it with the ant colony algorithm to optimize the harvesting sequence of citrus, addressing collision issues (Li C. et al., 2023); Li et al. used reinforcement learning to define a reward function for optimizing harvesting strategies in a multi-arm system (Li et al., 2023b); Yi et al. generated candidate viewpoints and scored them to select the best perspective (Yi et al., 2024). Optimization-based methods also require evaluating multiple viewpoints, which imposes a large computational burden, especially in large-scale and dynamic environments, affecting real-time performance.
6.3 Learning-based viewpoint planning methods
Learning-based planning methods utilize machine learning and deep learning techniques to train models that learn how to select the optimal viewpoint based on occlusion conditions. These methods offer high adaptability and flexibility, performing particularly well in complex and dynamic environments. Learning-based viewpoint planning works by automatically extracting features from a large amount of training data and making predictions using learned models. The models can include deep neural networks, reinforcement learning models, and others. The learning process typically involves using historical data to train the model, enabling it to generate reasonable viewpoint selection strategies based on input environmental information or task requirements. Zhang et al. applied deep learning techniques for multiview fruit detection in apple picking to determine the optimal picking location (Zhang et al., 2022). Wang et al. used a few-shot reinforcement learning approach to jointly train the Next Best View (NBV) and Next Best Point (NBP), with the model continuously optimizing viewpoint decisions through interaction with the environment (Wang G. et al., 2024). Chen et al. employed YOLOv8 for real-time object detection of longan fruits and guided a drone to perform fruit picking by establishing the relationship between the target points and the drone’s speed (Chen et al., 2025). Rehman et al. conducted viewpoint data collection by rotating 30 degrees from left to right around the target in a nighttime environment, using deep learning techniques to identify occluded areas and guide the harvesting robot in selecting the optimal viewpoint, as shown in Figure 8c (Rehman and Miura, 2021).
Overall, with the enhancement of perception and computational capabilities, significant progress has been made in fruit harvesting viewpoint planning technology. Geometric, information-based, optimization, and learning methods each have their advantages, adapting to different scenarios and requirements. Geometric methods are precise but complex and dependent on specific conditions; information-based methods optimize viewpoints but are computationally intensive; optimization methods are effective but burdensome in complex environments; and learning methods are highly adaptable but rely on training data and resources. Although existing research has improved recognition and harvesting efficiency, real-time performance, robustness, and accuracy in complex environments remain major challenges. Future research could explore the integration of multiple methods, such as combining optimization with deep learning, to enhance efficiency, reduce computational consumption, and improve the system’s adaptability and real-time adjustment capabilities.
7 Discussion
The fruit-picking robot has made significant advancements in visual perception technology, which is central to the automation of fruit harvesting. However, despite continuous technological progress, there are still many challenges when it comes to applying these systems in real agricultural environments.
7.1 Technical challenges and limitations
Various advanced cameras, such as monocular, binocular, and 3D depth sensors, have enhanced the precision of fruit recognition and localization for robots. Binocular cameras provide depth information through disparity, but they have limitations in calibration and adaptability. Complex depth sensors, such as Time-of-Flight (ToF) cameras and structured light cameras, offer excellent depth perception but are expensive and computationally intensive. Deep learning algorithms, such as YOLO, have improved fruit detection accuracy, but they require powerful computational resources, large training datasets, and depth data fusion. Striking a balance between computational efficiency and accuracy remains a key challenge for large-scale applications.
7.2 Impact of environmental variations
Intelligent fruit-picking robots face challenges such as lighting variations, plant positioning, and fruit occlusion in agricultural environments. These factors complicate the visual system’s ability to detect and localize fruits accurately. Even advanced sensors struggle when confronted with real-world agricultural settings. For instance, differences in the shape, color, and growth patterns of various fruits increase the difficulty of segmentation and classification. Ensuring high-precision recognition amidst these variations remains an unsolved problem.
7.3 Picking accuracy and efficiency
Picking accuracy is crucial, particularly in minimizing damage and improving fruit quality. Visual reconstruction and depth perception technologies assist in pinpointing the picking location, but the high computational cost remains a bottleneck in real-time data processing. Enhancing operational precision and preventing fruit damage are key considerations. Additionally, the introduction of active vision technology, which adjusts the visual angle based on real-time perception, can further improve picking accuracy.
7.4 Future development directions
Despite the challenges, the future of intelligent fruit-picking robots remains promising. Future research could explore sensor fusion, integrating visual, tactile, and force data to enhance the robot’s overall environmental perception. AI and machine learning, particularly unsupervised learning, hold the potential to reduce the reliance on large labeled datasets and improve the robot’s adaptability to new environments. By combining deep learning-based visual servoing techniques, path planning, and control strategies can be optimized. In the future, intelligent fruit-picking robots will achieve a better balance between real-time performance and accuracy.
8 Conclusion
In this paper, we reviewed the research progress of visual perception technology in intelligent fruit-picking robots. First, we introduced the advantages and disadvantages of different types of cameras: monocular cameras are suitable for simple scenarios, binocular cameras provide depth information for moderately complex environments, while structured light and ToF depth cameras perform excellently in high-precision depth perception and complex environments.
Next, we explored the application of object detection technology in fruit picking, comparing traditional image processing methods with modern deep learning methods such as YOLO and SSD. While deep learning methods offer higher accuracy and better adaptability, they require large amounts of training data and high-performance hardware. Traditional methods still have advantages when resources are limited.
Regarding the localization of picking points, we reviewed vision-based 3D reconstruction and depth perception methods, emphasizing the importance of accurate localization to improve the picking success rate and reduce fruit damage. Additionally, we explored technologies such as V-SLAM, mobile path planning, and task scheduling, which contribute to enhancing the robot’s operational efficiency throughout the entire orchard. We also discussed the combination of active vision and visual servoing techniques, showing that these two technologies can significantly enhance the robot’s adaptability and precision in dynamic environments. By adjusting the visual angle in real-time and optimizing control strategies, robots can more accurately locate and manipulate targets, especially when dealing with fruit occlusion and complex backgrounds.
Finally, we summarized the current status and future development directions of visual perception technology. Despite significant progress, challenges such as poor environmental adaptability, low system integration, and high costs still exist in real agricultural environments. With the continuous development of computer vision, deep learning, and sensor technologies, the future intelligent fruit-picking robots, combining active vision and visual servoing techniques, will make greater breakthroughs in efficiency and accuracy and will be capable of addressing more complex application scenarios.
Author contributions
YH: Conceptualization, Methodology, Investigation, Writing – original draft, Supervision. RC: Writing – original draft, Resources, Project administration. SX: Data curation, Writing – review & editing, Formal analysis. HC: Writing – review & editing. GL: Writing – review & editing. JY: Writing – review & editing. XZ: Writing – review & editing. HD: Methodology support, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research and/or publication of this article. This research was funded by the Natural Science Foundation of Fujian Province (No. 2022J01644), the Fujian Agriculture and Forestry University Science and Technology Innovation Fund (No. KFB24043), the Big Data in Agroforestry (Cross-Disciplinary) of Fujian Agriculture and Forestry University (No. 712023030), the Infrastructure Development Fund of Fujian Agriculture and Forestry University (No. KXNDM0001), and the Higher Education Scientific Research Planning Project (No. ZD202309).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Arad, B., Balendonck, J., Barth, R., Ben-Shahar, O., Edan, Y., Hellstrom¨, T., et al. (2020). Development of a sweet pepper harvesting robot. J. Field Robot. 37, 1027–1039. doi: 10.1002/rob.21937
Arefi, A., Motlagh, A. M., Mollazade, K., and Teimourlou, R. F. (2011). Recognition and localization of ripen tomato based on machine vision. Aust. J. Crop Sci. 5, 1144–1149. doi: 10.3316/informit.745798602538938
Aulia, L. H., Azhari, F., and Bimantara, M. D. (2023). Implementation of hsv imagery with k-nearest neighbor for classification of maturity levels in tomatoes. Bigint. Comput. J. 1, 62–69. doi: 10.55537/bigint.v1i2.779
Aygün, S. and Güneş, E. O. (2017). “A benchmarking: Feature extraction and classification of agricultural textures using lbp, glcm, rbo, neural networks, k-nn, and random forest,” in 2017 6th International Conference on Agro-Geoinformatics. 1–4 (Piscataway, NJ, IEEE). doi: 10.1109/Agro-Geoinformatics.2017.8047000
Babellahi, F., Paliwal, J., Erkinbaev, C., Amodio, M. L., Chaudhry, M. M. A., and Colelli, G. (2020). Early detection of chilling injury in green bell peppers by hyperspectral imaging and chemometrics. Postharvest. Biol. Technol. 162, 111100. doi: 10.1016/j.postharvbio.2019.111100
Baeten, J., Donné, K., Boedrij, S., Beckers, W., and Claesen, E. (2008). “Autonomous fruit picking machine: A robotic apple harvester,” in Field and service robotics: Results of the 6th international conference. 531–539 (Berlin, Heidelberg, Springer). doi: 10.1007/978-3-540-75404-6_51
Ban, Y., Liu, M., Wu, P., Yang, B., Liu, S., Yin, L., et al. (2022). Depth estimation method for monocular camera defocus images in microscopic scenes. Electronics 11, 2012. doi: 10.3390/electronics11132012
Besnassi, M., Neggaz, N., and Benyettou, A. (2020). Face detection based on evolutionary haar filter. Pattern Anal. Appl. 23, 309–330. doi: 10.1007/s10044-019-00784-5
Birrell, S., Hughes, J., Cai, J. Y., and Iida, F. (2020). A field-tested robotic harvesting system for iceberg lettuce. J. Field Robot. 37, 225–245. doi: 10.1002/rob.21888
Burusa, A. K., van Henten, E. J., and Kootstra, G. (2024). Attention-driven next-best-view planning for efficient reconstruction of plants and targeted plant parts. Biosyst. Eng. 246, 248–262. doi: 10.1016/j.biosystemseng.2024.08.002
Chao, D., Chen, Y., Junxi, Z., Guowei, S., and Chuyan, W. (2023). “Analysis of parallax characteristics of binocular vision based on no-similar imaging,” in AOPC 2022: optical sensing, imaging, and display technology, vol. 12557. (Bellingham, Washington, United States, SPIE), 576–583. doi: 10.1117/12.2669643
Chen, J., Ma, A., Huang, L., Li, H., Zhang, H., Huang, Y., et al. (2024). Efficient and lightweight grape and picking point synchronous detection model based on key point detection. Comput. Electron. Agric. 217, 108612. doi: 10.1016/j.compag.2024.108612
Chen, W., Rao, Y., Wang, F., Zhang, Y., Wang, T., Jin, X., et al. (2024). Mlp-based multimodal tomato detection in complex scenarios: Insights from task-specific analysis of feature fusion architectures. Comput. Electron. Agric. 221, 108951. doi: 10.1016/j.compag.2024.108951
Chen, M., Tang, Y., Zou, X., Huang, Z., Zhou, H., and Chen, S. (2021). 3d global mapping of large-scale unstructured orchard integrating eye-in-hand stereo vision and slam. Comput. Electron. Agric. 187, 106237. doi: 10.1016/j.compag.2021.106237
Chen, H., Wu, K., Lin, H., Zhou, H., Zhou, Z., Mai, Y., et al. (2025). A real-time vision guidance method for autonomous longan picking by the uav. Comput. Electron. Agric. 229, 109735. doi: 10.1016/j.compag.2024.109735
Cheng, Z., Zhang, Y., and Tang, C. (2021). Swin-depth: Using transformers and multi-scale fusion for monocularbased depth estimation. IEEE Sensors. J. 21, 26912–26920. doi: 10.1109/JSEN.2021.3120753
Chunjiang, Z., Beibei, F., Jin, L., and Qingchun, F. (2023). Agricultural robots: Technology progress, challenges and trends. Smart. Agric. 5, 1. doi: 10.12133/j.smartag.SA202312030
De Kleine, M. E. and Karkee, M. (2015). A semi-automated harvesting prototype for shaking fruit tree limbs. Trans. ASABE. 58, 1461–1470. doi: 10.13031/trans.58.11011
Gallego, G., Delbruck¨, T., Orchard, G., Bartolozzi, C., Taba, B., Censi, A., et al. (2021). Event-based vision: A survey. IEEE Transactions on Pattern Analysis and Machine Intelligence 44, 154–180. doi: 10.1109/TPAMI.2020.3008413
Gehrig, M. and Scaramuzza, D. (2023). “Recurrent vision transformers for object detection with event cameras,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Vancouver, Canada, IEEE 13884–13893.
Girshick, R. (2015). Fast r-cnn. arXiv. preprint. arXiv:1504.08083, 1440–48. doi: 10.1109/ICCV.2015.169
Guo, Q., Chen, Y., Tang, Y., Zhuang, J., He, Y., Hou, C., et al. (2019). Lychee fruit detection based on monocular machine vision in orchard environment. Sensors 19, 4091. doi: 10.3390/s19194091
Guo, J., Xiao, X., Miao, J., Tian, B., Zhao, J., and Lan, Y. (2023). Design and experiment of a visual detection system for zanthoxylum-harvesting robot based on improved yolov5 model. Agriculture 13, 821. doi: 10.3390/agriculture13040821
Guo, C., Zhu, C., Liu, Y., Huang, R., Cao, B., Zhu, Q., et al. (2024) End-to-End lightweight TransformerBased neural network for grasp detection towards fruit robotic handling. Computers and Electronics in Agriculture 221, 109014. doi: 10.1016/j.compag.2024.109014
Gurubelli, Y., Malmathanraj, R., and Palanisamy, P. (2020). “Texture and colour gradient features for grade analysis of pomegranate and mango fruits using kernel-svm classifiers,” in 2020 6th International Conference on Advanced Computing and Communication Systems (ICACCS). 122–126 (Coimbatore, India, IEEE). doi: 10.1109/ICACCS48705.2020.9074221
He, K., Gkioxari, G., Dollar´, P., and Girshick, R. (2017). “Mask r-cnn,” in Proceedings of the IEEE international conference on computer vision. Venice, Italy, IEEE, 2961–2969.
Hornung, A., Wurm, K. M., Bennewitz, M., Stachniss, C., and Burgard, W. (2013). Octomap: An efficient probabilistic 3d mapping framework based on octrees. Autonomous. Robots. 34, 189–206. doi: 10.1007/s10514-012-9321-0
Hou, C., Zhang, X., Tang, Y., Zhuang, J., Tan, Z., Huang, H., et al. (2022). Detection and localization of citrus fruit based on improved you only look once v5s and binocular vision in the orchard. Front. Plant Sci. 13, 972445. doi: 10.3389/fpls.2022.972445
Huang, T., Huang, L., You, S., Wang, F., Qian, C., and Xu, C. (2022). LightViT: towards light-weight convolution-free vision transformers. ArXiv. preprint. arXiv:2207.05557. doi: 10.48550/arXiv.2207.05557
Huang, Y., Zhong, Y., Zhong, D., Yang, C., Wei, L., Zou, Z., et al. (2024). Pepper-yolo: an lightweight model for green pepper detection and picking point localization in complex environments. Front. Plant Sci. 15, 1508258. doi: 10.3389/fpls.2024.1508258
Hussain, M. (2024). Yolov5, yolov8 and yolov10: The go-to detectors for real-time vision. arXiv. preprint. arXiv:2407.02988. doi: 10.48550/arXiv.2407.02988
Jia, W., Zhang, Y., Lian, J., Zheng, Y., Zhao, D., and Li, C. (2020). Apple harvesting robot under information technology: A review. Int. J. Adv. Robot. Syst. 17, 1729881420925310. doi: 10.1177/1729881420925310
Jun, J., Kim, J., Seol, J., Kim, J., and Son, H. I. (2021). Towards an efficient tomato harvesting robot: 3d perception, manipulation, and end-effector. IEEE Access 9, 17631–17640. doi: 10.1109/ACCESS.2021.3052240
Junge, K., Pires, C., and Hughes, J. (2023). Lab2field transfer of a robotic raspberry harvester enabled by a soft sensorized physical twin. Commun. Eng. 2, 40. doi: 10.1038/s44172-023-00089-w
Khan, F., Salahuddin, S., and Javidnia, H. (2020). Deep learning-based monocular depth estimation methods—a state-of-the-art review. Sensors 20, 2272. doi: 10.3390/s20082272
Khanam, R. and Hussain, M. (2024). Yolov11: An overview of the key architectural enhancements. ArXiv. preprint. arXiv:2410.17725 20, 2272. doi: 10.48550/arXiv.2410.17725
Lawal, M. O. (2021). Tomato detection based on modified yolov3 framework. Sci. Rep. 11, 1–11. doi: 10.1038/s41598-021-81216-5
Lawal, O. M. (2021). Development of tomato detection model for robotic platform using deep learning. Multimedia. Tools Appl. 80, 26751–26772. doi: 10.1007/s11042-021-10933-w
Lei, M., Li, S., Wu, Y., Hu, H., Zhou, Y., Zheng, X., et al. (2025). YOLOv13: real-time object detection with hypergraph-enhanced adaptive visual perception. ArXiv. preprint. arXiv:2506.17733. doi: 10.48550/arXiv.2506.17733
Li, H., Huang, J., Gu, Z., He, D., Huang, J., and Wang, C. (2024). Positioning of mango picking point using an improved yolov8 architecture with object detection and instance segmentation. Biosyst. Eng. 247, 202–220. doi: 10.1016/j.biosystemseng.2024.09.015
Li, C., Li, L., Jiang, H., Weng, K., Geng, Y., Li, L., et al. (2022). Yolov6: A single-stage object detection framework for industrial applications. arXiv. preprint. arXiv:2209.02976. doi: 10.48550/arXiv.2209.02976
Li, C., Ma, W., Liu, F., Fang, B., Lu, H., and Sun, Y. (2023). Recognition of citrus fruit and planning the robotic picking sequence in orchards. Signal. Image. Video. Process. 17, 4425–4434. doi: 10.1007/s11760-023-02676-y
Li, H., Shi, L., Fang, S., and Yin, F. (2023). Real-time detection of apple leaf diseases in natural scenes based on yolov5. Agriculture 13, 878. doi: 10.3390/agriculture13040878
Li, T., Xie, F., Qiu, Q., and Feng, Q. (2023a). “Multi-arm robot task planning for fruit harvesting using multi-agent reinforcement learning,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 4176–4183 (Detroit, MI, USA, IEEE). doi: 10.1109/IROS55552.2023.10341822
Li, T., Xie, F., Zhao, Z., Zhao, H., Guo, X., and Feng, Q. (2023b). A multi-arm robot system for efficient apple harvesting: Perception, task plan and control. Comput. Electron. Agric. 211, 107979. doi: 10.1016/j.compag.2023.107979
Li, Z., Yuan, X., and Wang, C. (2022). A review on structural development and recognition–localization methods for end-effector of fruit–vegetable picking robots. Int. J. Adv. Robot. Syst. 19, 17298806221104906. doi: 10.1177/17298806221104906
Lin, T. Y., Dollár, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017). “Feature pyramid networks for object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Honolulu, HI, USA, IEEE, 2117–2125.
Lin, G., Tang, Y., Zou, X., Cheng, J., and Xiong, J. (2020). Fruit detection in natural environment using partial shape matching and probabilistic hough transform. Precis. Agric. 21, 160–177. doi: 10.1007/s11119-019-09662-w
Ling, X., Zhao, Y., Gong, L., Liu, C., and Wang, T. (2019). Dual-arm cooperation and implementing for robotic harvesting tomato using binocular vision. Robot. Autonomous. Syst. 114, 134–143. doi: 10.1016/j.robot.2019.01.019
Liu, Z., Hu, H., Lin, Y., Yao, Z., Xie, Z., Wei, Y., et al. (2022). Swin transformer v2: Scaling up capacity and resolution. ArXiv. preprint. arXiv:2111.09883, 12009–19. doi: 10.1109/CVPR52688.2022.01170
Liu, X., Jing, X., Jiang, H., Younas, S., Wei, R., Dang, H., et al. (2024). Performance evaluation of newly released cameras for fruit detection and localization in complex kiwifruit orchard environments. J. Field Robot. 41, 881–894. doi: 10.1002/rob.22297
Liu, G., Nouaze, J. C., Touko Mbouembe, P. L., and Kim, J. H. (2020). Yolo-tomato: A robust algorithm for tomato detection based on yolov3. Sensors 20, 2145. doi: 10.3390/s20072145
Lopez-Barrioś, J. D., Escobedo Cabello, J. A., Gomez-Espinosa´, A., and Montoya-Cavero, L. E. (2023). Green sweet pepper fruit and peduncle detection using mask r-cnn in greenhouses. Appl. Sci. 13, 6296. doi: 10.3390/app13106296
Lu, A., Lv, X., Ma, L., Li, Q., and Ma, Q. (2024). An effective picking point localization method for multi-posture lotus pods based on three-view depth vision observation. Comput. Electron. Agric. 227, 109492. doi: 10.1016/j.compag.2024.109492
Luo, L., Tang, Y., Zou, X., Ye, M., Feng, W., and Li, G. (2016). Vision-based extraction of spatial information in grape clusters for harvesting robots. Biosyst. Eng. 151, 90–104. doi: 10.1016/j.biosystemseng.2016.08.026
Lv, J., Wang, F., and Ma, Z. (2016). “Peach fruit recognition method under natural environment,” in Eighth International Conference on Digital Image Processing (ICDIP 2016), Vol. 10033. 232–236 (Bellingham, SPIE). doi: 10.1117/12.2244945
Lv, J., Wang, F., Ma, Z., and Rong, H. (2015). “Yellow apple recognition method under natural environment,” in 2015 7th International Conference on Intelligent Human-Machine Systems and Cybernetics, Vol. 1. 46–49 (New York, IEEE). doi: 10.1109/IHMSC.2015.91
Maud, M., Kadim, Z., and Hon, H. (2023). Real-time 3d mapping and localization of palm oil tree for harvest data management system. J. Inf. Syst. And. Technol. Manage. (JISTM). 8, 91–101.
Menon, R., Zaenker, T., Dengler, N., and Bennewitz, M. (2023). “Nbv-sc: Next best view planning based on shape completion for fruit mapping and reconstruction,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 4197–4203 (New York, IEEE). doi: 10.1109/IROS55552.2023.10341855
Mia, M. S., Voban, A. A. B., Arnob, A. B. H., Naim, A., Ahmed, M. K., and Islam, M. S. (2023). “Danet: Enhancing small object detection through an efficient deformable attention network,” in 2023 International Conference on the Cognitive Computing and Complex Data (ICCD). 51–62 (New York, IEEE). doi: 10.1109/ICCD59681.2023.10420622
Miriti, E. (2016). . Classification of selected apple fruit varieties using Naive Bayes (University of Nairobi, Nairobi, Kenya).
Ning, Z., Luo, L., Ding, X., Dong, Z., Yang, B., Cai, J., et al. (2022). Recognition of sweet peppers and planning the robotic picking sequence in high-density orchards. Comput. Electron. Agric. 196, 106878. doi: 10.1016/j.compag.2022.106878
Oliveira, L. F., Moreira, A. P., and Silva, M. F. (2021). Advances in agriculture robotics: A state-of-the-art review and challenges ahead. Robotics 10, 52. doi: 10.3390/robotics10020052
Park, Y., Seol, J., Pak, J., Jo, Y., Kim, C., and Son, H. I. (2023). Human-centered approach for an efficient cucumber harvesting robot system: Harvest ordering, visual servoing, and end-effector. Comput. Electron. Agric. 212, 108116. doi: 10.1016/j.compag.2023.108116
Parvathi, S. and Selvi, S. T. (2021). Detection of maturity stages of coconuts in complex background using faster r-cnn model. Biosyst. Eng. 202, 119–132. doi: 10.1016/j.biosystemseng.2020.12.002
Pohle-Fröhlich, R., Gebler, C., and Bolten, T. (2024). Stereo-event-camera-technique for insect monitoring. VISIGRAPP 3, 375–384. doi: 10.5220/0012326500003660
Rabby, M. K. M., Chowdhury, B., and Kim, J. H. (2018). “A modified canny edge detection algorithm for fruit detection & Classification,” in 2018 10TH INTERNATIONAL CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (ICECE), New York. 237–240 (Dhaka, Bangladesh, IEEE). doi: 10.1109/ICECE.2018.8636811
Ramisetty, U. M., Gundavarapu, V. N. K., Rajender, R., Segovia Ramírez, I., and García Márquez, F. P. (2022). “Prediction analysis of crop and their futuristic yields using random forest regression,” in The International Conference on Industrial Engineering and Industrial Management. 280–285 (Cham, Springer). doi: 10.1007/978-3-031-27915-7_50
Rapado-Rincón, D., van Henten, E. J., and Kootstra, G. (2023). Development and evaluation of automated localisation and reconstruction of all fruits on tomato plants in a greenhouse based on multi-view perception and 3d multi-object tracking. Biosyst. Eng. 231, 78–91. doi: 10.1016/j.biosystemseng.2023.06.003
Rebecq, H., Gallego, G., Mueggler, E., and Scaramuzza, D. (2018). EMVS: event-based multi-view stereo—3D reconstruction with an event camera in real-time? International Journal of Computer Vision 126, 1394–1414. doi: 10.1007/s11263017-1050-6
Redmon, J. (2016). “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Las Vegas, NV, USA, IEEE 1–10.
Rehman, H. U. and Miura, J. (2021). “Viewpoint planning for automated fruit harvesting using deep learning,” in 2021 IEEE/SICE International Symposium on System Integration (SII). 409–414 (New York, IEEE). doi: 10.1109/IEEECONF49454.2021.9382628
Ren, S., He, K., Girshick, R., and Sun, J. (2016). Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39, 1137–1149. doi: 10.1109/TPAMI.2016.2577031
Sanders, K. (2005). Orange harvesting systems review. Biosyst. Eng. 90, 115–125. doi: 10.1016/j.biosystemseng.2004.10.006
Sapkota, R. and Karkee, M. (2025). Improved YOLOv12 with LLM-generated synthetic data for enhanced apple detection and benchmarking against YOLOv11 and YOLOv10? arXiv. doi: 10.48550/arXiv.2503.00057. ArXiv preprint arXiv:2503.00057.
Schertz, C. E. and Brown, G. (1968). Basic considerations in mechanizing citrus harvest. Trans. ASAE. 11, 343–0346. doi: 10.13031/2013.39405
Sola-Guirado, R. R., Sánchez-Cachinero, P., and Blanco-Roldán, G. (2023). Simultaneous trunk and branch shaking in an over-the-row olive harvester. Biosyst. Eng. 231, 92–103. doi: 10.1016/j.biosystemseng.2023.06.005
Sozzi, M., Cantalamessa, S., Cogato, A., Kayad, A., and Marinello, F. (2022). Automatic bunch detection in white grape varieties using yolov3, yolov4, and yolov5 deep learning algorithms. Agronomy 12, 319. doi: 10.3390/agronomy12020319
Suresh Kumar, M. and Mohan, S. (2023). Selective fruit harvesting: Research, trends and developments towards fruit detection and localization–a review. Proc. Inst. Mechanical. Eng. Part C: J. Mechanical. Eng. Sci. 237, 1405–1444. doi: 10.1177/09544062221128443
Tafuro, A., Adewumi, A., Parsa, S., Amir, G. E., and Debnath, B. (2022). “Strawberry picking point localization ripeness and weight estimation,” in 2022 International conference on robotics and automation (ICRA). 2295–2302 (Philadelphia, PA, USA, IEEE). doi: 10.1109/ICRA46639.2022.9812200
Tan, S. H., Lam, C. K., Kamarudin, K., Ismail, A. H., Rahim, N. A., Azmi, M. S. M., et al. (2021). Vision-based edge detection system for fruit recognition. J. Phys.: Conf. Ser. 2107, 012066.
Tian, K., Li, J., Zeng, J., Evans, A., and Zhang, L. (2019). Segmentation of tomato leaf images based on adaptive clustering number of k-means algorithm. Comput. Electron. Agric. 165, 104962. doi: 10.1016/j.compag.2019.104962
Tian, Y., Ye, Q., and Doermann, D. (2025). YOLOv12: attention-centric real-time object detectors. arXiv doi: 10.48550/arXiv.2502.12524
Urvina, R. P., Guevara, C. L., Vásconez, J. P., and Prado, A. J. (2024). An integrated route and path planning strategy for skid–steer mobile robots in assisted harvesting tasks with terrain traversability constraints. Agriculture 14, 1206. doi: 10.3390/agriculture14081206
Wan, S. and Goudos, S. (2020). Faster r-cnn for multi-class fruit detection using a robotic vision system. Comput. Networks 168, 107036. doi: 10.1016/j.comnet.2019.107036
Wan, H., Ou, Y., Guan, X., Jiang, R., Zhou, Z., and Luo, X. (2024). Review of the perception technologies for unmanned agricultural machinery operating environment. Transactions of the Chinese Society of Agricultural Engineering 40, 1–18. doi: 10.11975/j.issn.10026819.202402020
Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J., et al. (2024). Yolov10: Real-time end-to-end object detection. arXiv. preprint. arXiv:2405.14458 37, 107984–108011.
Wang, N., Jin, Z., Wang, T., Xiao, J., Zhang, Z., Wang, H., et al. (2025a). Hybrid path planning methods for complete coverage in harvesting operation scenarios. Comput. Electron. Agric. 231, 109946. doi: 10.1016/j.compag.2025.109946
Wang, N., Li, S., Xiao, J., Wang, T., Han, Y., Wang, H., et al. (2025b). A collaborative scheduling and planning method for multiple machines in harvesting and transportation operations-part: Harvester task allocation and sequence optimization. Comput. Electron. Agric. 232, 110060. doi: 10.1016/j.compag.2025.110060
Wang, G., Li, H., Zhang, S., Liu, Y., and Liu, H. (2024). Observe then act: Asynchronous active vision-action model for robotic manipulation. arXiv. preprint. arXiv:2409.14891.
Wang, J., Lin, X., Luo, L., Chen, M., Wei, H., Xu, L., et al. (2024). Cognition of grape cluster picking point based on visual knowledge distillation in complex vineyard environment. Comput. Electron. Agric. 225, 109216. doi: 10.1016/j.compag.2024.109216
Wang, P., Luo, W., Liu, J., Zhou, Y., Li, X., Zhao, S., et al. (2025). Real-time semantic slam-based 3d reconstruction robot for greenhouse vegetables. Comput. Electron. Agric. 237, 110582. doi: 10.1016/j.compag.2025.110582
Wang, Y., Ouyang, C., Peng, H., Deng, J., Yang, L., Chen, H., et al. (2025). YOLO-ALW: an enhanced high-precision model for chili maturity detection. Sensors 25, 1405. doi: 10.3390/s25051405
Wang, D., Song, H., Tie, Z., Zhang, W., and He, D. (2016). Recognition and localization of occluded apples using k-means clustering algorithm and convex hull theory: a comparison. Multimedia. Tools Appl. 75, 3177–3198. doi: 10.1007/s11042-014-2429-9
Wang, F., Sun, Z., Chen, Y., Zheng, H., and Jiang, J. (2022). Xiaomila green pepper target detection method under complex environment based on improved yolov5s. Agronomy 12, 1477. doi: 10.3390/agronomy12061477
Wang, L., Wang, Z., Liu, M., Ying, Z., Xu, N., and Meng, Q. (2022). Full coverage path planning methods of harvesting robot with multi-objective constraints. J. Intelligent. Robot. Syst. 106, 17. doi: 10.1007/s10846-022-01722-0
Wang, Y., Wu, H., Zhu, Z., Ye, Y., and Qian, M. (2023). Continuous picking of yellow peaches with recognition and collision-free path. Comput. Electron. Agric. 214, 108273. doi: 10.1016/j.compag.2023.108273
Wang, C. Y., Yeh, I. H., and Mark Liao, H. Y. M. (2024). “Yolov9: Learning what you want to learn using programmable gradient information,” in European conference on computer vision. 15089, 1–21 (Cham, Springer). doi: 10.1007/978-3-031-72751-1_1
Wang, Z., Jin, L., Wang, S., and Xu, H (2022) “Apple stem/calyx real-time recognition using YOLO-v5 algorithm for fruit automatic loading system”, Postharvest Biology and Technology, Vol. 185, p. 111808. doi: 10.1016/j.postharvbio.2021.111808
Wei, X., Jia, K., Lan, J., Li, Y., Zeng, Y., and Wang, C. (2014). Automatic method of fruit object extraction under complex agricultural background for vision system of fruit picking robot. Optik 125, 5684–5689. doi: 10.1016/j.ijleo.2014.07.001
Wen, J., Yin, Y., Zhang, Y., Pan, Z., and Fan, Y. (2022). Detection of wheat lodging by binocular cameras during harvesting operation. Agriculture 13, 120. doi: 10.3390/agriculture13010120
Wu, Y., Qiu, C., Liu, S., Zou, X., and Li, X. (2021). “Tomato harvesting robot system based on binocular vision,” in 2021 IEEE international conference on unmanned systems (ICUS). 757–761 (Beijing, China, IEEE). doi: 10.1109/ICUS52573.2021.9641260
Xia, C. (2024). Rapid strawberry ripeness detection and 3d localization of picking point based on improved yolo v8-pose with rgbcamera. J. Electric. Syst. 20, 2171–2187. doi: 10.52783/jes.1840
Xiao, X., Wang, Y., and Jiang, Y. (2024). Review of research advances in fruit and vegetable harvesting robots. J. Electric. Eng. Technol. 19, 773–789. doi: 10.1007/s42835-023-01596-8
Yamamoto, S., Hayashi, S., Yoshida, H., and Kobayashi, K. (2014). Development of a stationary robotic strawberry harvester with a picking mechanism that approaches the target fruit from below. Jpn. Agric. Res. Quarterly.: JARQ. 48, 261–269. doi: 10.6090/jarq.48.261
Yan, D., Luo, L., Zhang, P., Liu, W., Wang, J., Lu, Q., et al. (2023). Vibration analysis and experimental study of the effects of mechanised grape picking on the fruit–stem system. Biosyst. Eng. 227, 82–94. doi: 10.1016/j.biosystemseng.2023.01.019
Yang, N., Stumberg, Lv, Wang, R., and Cremers, D. (2020). “D3vo: Deep depth, deep pose and deep uncertainty for monocular visual odometry,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Seattle, WA, USA, IEEE 1281–1292.
Yi, T., Zhang, D., Luo, L., and Luo, J. (2024). View planning for grape harvesting based on active vision strategy under occlusion. IEEE Robot. Automat. Lett. 9, 2535–2542. doi: 10.1109/LRA.2024.3357397
Yin, R., Wu, H., Li, M., Cheng, Y., Song, Y., and Handroos, H. (2023). Monocular camera-based robotic pick-and-place in fusion applications. Appl. Sci. 13, 4487. doi: 10.3390/app13074487
Yogesh, Dubey, A. K., Arora, R. R., and Mathur, A. (2021). Fruit defect prediction model (fdpm) based on three-level validation. J. Nondestruct. Eval. 40, 45. doi: 10.1007/s10921-021-00778-6
Yoshida, T., Kawahara, T., and Fukao, T. (2022). Fruit recognition method for a harvesting robot with rgb-d cameras. ROBOMECH. J. 9, 15. doi: 10.1186/s40648-022-00230-y
Yu, Y., Zhang, K., Liu, H., Yang, L., and Zhang, D. (2020). Real-time visual localization of the picking points for a ridge-planting strawberry harvesting robot. IEEE Access 8, 116556–116568. doi: 10.1109/Access.6287639
Zaenker, T., Ruckin¨, J., Menon, R., Popovic´, M., and Bennewitz, M. (2023). “Graph-based view motion planning for fruit detection,” in 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 4219–4225 (Detroit, MI, USA, IEEE). doi: 10.1109/IROS55552.2023.10341874
Zaenker, T., Smitt, C., McCool, C., and Bennewitz, M. (2021). “Viewpoint planning for fruit size and position estimation,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 3271–3277 (IEEE).
Zhang, J., Kang, N., Qu, Q., Zhou, L., and Zhang, H. (2024). Automatic fruit picking technology: a comprehensive review of research advances. Artif. Intell. Rev. 57, 54. doi: 10.1007/s10462-023-10674-2
Zhang, K., Lammers, K., Chu, P., Dickinson, N., Li, Z., and Lu, R. (2022). “Algorithm design and integration for a robotic apple harvesting system,” in 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 9217–9224 (Kyoto, Japan, IEEE). doi: 10.1109/IROS47612.2022.9981417
Zhang, Y. and Wu, L. (2012). Classification of fruits using computer vision and a multiclass support vector machine. sensors 12, 12489–12505. doi: 10.3390/s120912489
Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., et al. (2024). “DETRs beat YOLOs on real-time object detection,” in 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). Seattle, WA, USA, IEEE 16965–74. doi: 10.1109/CVPR52733.2024.01605
Zheng, C., Chen, P., Pang, J., Yang, X., Chen, C., Tu, S., et al. (2021). A mango picking vision algorithm on instance segmentation and key point detection from rgb images in an open orchard. Biosyst. Eng. 206, 32–54. doi: 10.1016/j.biosystemseng.2021.03.012
Zhong, Z., Xiong, J., Zheng, Z., Liu, B., Liao, S., Huo, Z., et al. (2021). A method for litchi picking points calculation in natural environment based on main fruit bearing branch detection. Comput. Electron. Agric. 189, 106398. doi: 10.1016/j.compag.2021.106398
Zhou, T., Fan, D. P., Cheng, M. M., Shen, J., and Shao, L. (2021). Rgb-d salient object detection: A survey. Comput. Visual Media. 7, 37–69. doi: 10.1007/s41095-020-0199-z
Zhou, J., He, L., Karkee, M., and Zhang, Q. (2016). Analysis of shaking-induced cherry fruit motion and damage. Biosyst. Eng. 144, 105–114. doi: 10.1016/j.biosystemseng.2016.02.007
Zhou, H., Wang, X., Au, W., Kang, H., and Chen, C. (2022). Intelligent robots for fruit harvesting: Recent developments and future challenges. Precis. Agric. 23, 1856–1907. doi: 10.1007/s11119-022-09913-3
Zhou, H. and Yu, G. (2021). Research on pedestrian detection technology based on the svm classifier trained by hog and ltp features. Future Generat. Comput. Syst. 125, 604–615. doi: 10.1016/j.future.2021.06.016
Zhu, Y., Lu, W., Ren, G., Ying, Y., Vougioukas, S., and Peng, C. (2025). Optimal scheduling of a dual-arm robot for efficient strawberry harvesting in plant factories. arXiv. preprint. arXiv:2507.04240. doi: 10.48550/arXiv.2507.04240
Keywords: intelligent fruit harvesting robots, agricultural robotics, visual perception, object detection, visual servoing, V-SLAM
Citation: Huang Y, Xu S, Chen H, Li G, Dong H, Yu J, Zhang X and Chen R (2025) A review of visual perception technology for intelligent fruit harvesting robots. Front. Plant Sci. 16:1646871. doi: 10.3389/fpls.2025.1646871
Received: 19 June 2025; Accepted: 17 July 2025;
Published: 19 August 2025.
Edited by:
Aichen Wang, Jiangsu University, ChinaCopyright © 2025 Huang, Xu, Chen, Li, Dong, Yu, Zhang and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Shuyan Xu, eHN5QG1udXN0LmVkdS5jbg==; Riqing Chen, cmlxaW5nLmNoZW5AZmFmdS5lZHUuY24=