Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci.

Sec. Technical Advances in Plant Science

Volume 16 - 2025 | doi: 10.3389/fpls.2025.1664718

This article is part of the Research TopicPlant Phenotyping for AgricultureView all 11 articles

CGA-ASNet: An RGB-D Amodal Segmentation Network for Restoring Occluded Tomato Regions

Provisionally accepted
Zhaoyang  LiZhaoyang LiYong  YinYong YinZhihong  XingZhihong XingHanbing  DengHanbing Deng*
  • Shenyang Agricultural University, Shenyang, China

The final, formatted version of the article will be published soon.

Obtaining the complete morphology of the fruit under non-destructive conditions during tomato growth is crucial for tomato phenotype research. However, occlusions among tomato fruits can impede deep learning-based image segmentation methods from accurately capturing the true shape of the occluded areas during training. This issue results in the model being unable to accurately predict the complete morphology of the fruit, which in turn affects the model's ability to obtain accurate tomato phenotype data. To address this issue, we proposed CGA-ASNet, an RGB-D amodal segmentation network based on a Contextual and Global Attention (CGA) module. First of all, we constructed a synthetic tomato dataset (Tomato-sim) using NVIDIA Isaac Sim's Replicator Composer (ISRC). This dataset accurately simulated tomato morphology and greenhouse environments and had been trained by CGA-ASNet. To evaluate the generalization of model, we tested CGA-ASNet on both the synthetic and a separate real-world dataset. Although no explicit domain adaptation techniques were used, we adopted an implicit strategy by simulating diverse lighting conditions (e.g., strong, normal, and weak illumination) to narrow the domain gap.Additionally, a mean coordinate fusion algorithm was applied to improve annotation completeness in real-world occlusion scenarios. CGA-ASNet leveraged contextual information between feature input keys for self-attention learning, captured global information, and expanded the receptive field, thus enhancing the network's representation capacity and enabling more accurate semantic understanding and precise localization. The experimental results demonstrated that CGA-ASNet achieved an F@.75 score of 94.2 and a mIoU of 82.4% in amodal segmentation accuracy in greenhouse scenarios. The results demonstrate that training with well-designed synthetic data can effectively support accurate occlusion-aware segmentation in real environments that offers a practical solution for tomato phenotyping in greenhouse.

Keywords: Amodal, segmentation, Occlusion-aware Segmentation, RGB-D image segmentation, plant phenotyping, Tomato, smart agriculture

Received: 12 Jul 2025; Accepted: 29 Aug 2025.

Copyright: © 2025 Li, Yin, Xing and Deng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

* Correspondence: Hanbing Deng, Shenyang Agricultural University, Shenyang, China

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.