Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Technical Advances in Plant Science

Tea bud pose estimation and grading detection network based on improved YOLOv7

Provisionally accepted
Yuchen  YaoYuchen Yao1,2Gui  ZhiyongGui Zhiyong2Haoyang  LiuHaoyang Liu2,3Zidong  YangZidong Yang1Lijian  YaoLijian Yao1Kai  LiKai Li1Zhenchuan  LinZhenchuan Lin4Yihu  MaoYihu Mao2Zhijun  JiaZhijun Jia1,2Yang  LiYang Li2,3*Rong  MaRong Ma1*
  • 1Zhejiang A and F University, Hangzhou, China
  • 2Tea Research Institute, Chinese Academy of Agricultural Sciences, Hangzhou, China
  • 3National Key Laboratory for Tea Plant Germplasm Innovation and Resource Utilization, Hangzhou, China
  • 4FUJIAN PIN PIN XIANG TEA INDUSTRY CO., LTD, Ningde, China

The final, formatted version of the article will be published soon.

Intelligent recognition and rapid grading of tea buds are significant for the development of tea-picking machinery. However, due to the complex background of tea plantations and the inconsistent growth of tea buds, traditional recognition algorithms have only ad-dressed the issue of identifying picking points, without fully considering the pose and grade of buds. As a result, the picking end-effector cannot adaptively harvest tea buds based on their pose and grade, thereby limiting the success rate and efficiency of picking. To address this issue, we propose YOLO-PC (You Only Look Once-Pose estimation and Classification), a deep neural network for tea bud pose estimation and classification. This model enhances shape feature extraction using a dynamic snake convolution (DSConv) module and improves spatial pooling with an ELA-SPPCSPC attention mechanism. Replacing CIoU with EIoU accelerates regression and boosts localization accuracy. Experiments show 91.5% accuracy for one-bud-one-leaf detection and 93.2% for one-bud-two-leaf detection, with an average keypoint detection accuracy (Pose_mAP) of 89.7%. The Normalized Mean Error (NME) achieved 0.047. The improved model compared to YOLOv7-pose increases mean average precision by 7.26% and pose accuracy by 9.65%, while reducing parameters by 14.99 M. Ablation assays confirm its superior performance in tea bud detection, providing practical support for intelligent harvesting.

Keywords: Deep neural network, Grading, Lightweight, Pose estimation, Tea bud

Received: 12 Jan 2026; Accepted: 12 Feb 2026.

Copyright: © 2026 Yao, Zhiyong, Liu, Yang, Yao, Li, Lin, Mao, Jia, Li and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence:
Yang Li
Rong Ma

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.