Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Sustainable and Intelligent Phytoprotection

TeaNeRF: An Integrated 3D Visual Perception Pipeline for Tea Bud Harvesting

Provisionally accepted
Weiheng  ChenWeiheng ChenXun  LiXun Li*Lei  RaoLei RaoXiang  XiaXiang Xia
  • Wuhan Institute of Technology, Wuhan, China

The final, formatted version of the article will be published soon.

Accurate perception of tea buds is a fundamental prerequisite for intelligent and precise tea harvesting planning. However, in real tea plantation environments, reliable harvesting-oriented perception at the planning level remains highly challenging due to the small size of tea buds, severe occlusion, complex background clutter, and the lack of accurate three-dimensional spatial information. To address these challenges, we propose TeaNeRF, an integrated three-dimensional visual perception pipeline designed for harvesting-oriented tea bud analysis. Instead of treating detection, segmentation, and spatial analysis as independent tasks, TeaNeRF integrates sequential two-dimensional recognition, monocular depth estimation, and neural radiance field reconstruction into a coherent perception pipeline, allowing accurate spatial understanding of tea buds in complex natural scenes. It should be noted that the proposed integration is conducted at the perception-output level, where multiple modular components are connected through fixed interfaces, rather than through joint optimization or an end-to-end trainable formulation. The proposed framework combines an enhanced YOLO-based detector, prompt-guided segmentation, and monocular depth priors to guide NeRF-based three-dimensional reconstruction. By incorporating depth supervision and semantic-aware neural fields, TeaNeRF generates dense and geometrically consistent point clouds with reliable semantic separation. Quantitative evaluations show consistent improvements in reconstruction fidelity, as reflected by increased PSNR and reduced LPIPS across multiple tea tree scenes. Based on the reconstructed semantic point cloud, a three-dimensional clustering and geometric fitting strategy is further developed to enable tea bud counting and harvesting-oriented candidate point estimation at the perception level. Experiments conducted on a real-world dataset of 4,700 tea plantation images demonstrate that TeaNeRF improves detection accuracy (mAP@50 = 91.7%), segmentation quality (IoU = 0.640), and overall three-dimensional perception performance. Case-level counting results on representative tea trees indicate that the proposed 3D semantic point cloud–based approach can provide feasible tea bud counting behavior and consistent spatial guidance cues for downstream harvesting planning. Chen et al. Running Title By providing structured three-dimensional spatial information, including tea bud locations, counts, and harvesting-oriented candidate points, TeaNeRF offers practical perception-level outputs for downstream planning in automated tea harvesting systems.

Keywords: 3D Reconstruction, Automated harvesting, Nerf, SAM2, Semantic segmentation, Tea bud detection, YOLOv11

Received: 10 Nov 2025; Accepted: 12 Feb 2026.

Copyright: © 2026 Chen, Li, Rao and Xia. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Xun Li

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.