ORIGINAL RESEARCH article
Front. Plant Sci.
Sec. Technical Advances in Plant Science
Volume 16 - 2025 | doi: 10.3389/fpls.2025.1664718
This article is part of the Research TopicPlant Phenotyping for AgricultureView all 11 articles
CGA-ASNet: An RGB-D Amodal Segmentation Network for Restoring Occluded Tomato Regions
Provisionally accepted- Shenyang Agricultural University, Shenyang, China
Select one of your emails
You have multiple emails registered with Frontiers:
Notify me on publication
Please enter your email address:
If you already have an account, please login
You don't have a Frontiers account ? You can register here
Obtaining the complete morphology of the fruit under non-destructive conditions during tomato growth is crucial for tomato phenotype research. However, occlusions among tomato fruits can impede deep learning-based image segmentation methods from accurately capturing the true shape of the occluded areas during training. This issue results in the model being unable to accurately predict the complete morphology of the fruit, which in turn affects the model's ability to obtain accurate tomato phenotype data. To address this issue, we proposed CGA-ASNet, an RGB-D amodal segmentation network based on a Contextual and Global Attention (CGA) module. First of all, we constructed a synthetic tomato dataset (Tomato-sim) using NVIDIA Isaac Sim's Replicator Composer (ISRC). This dataset accurately simulated tomato morphology and greenhouse environments and had been trained by CGA-ASNet. To evaluate the generalization of model, we tested CGA-ASNet on both the synthetic and a separate real-world dataset. Although no explicit domain adaptation techniques were used, we adopted an implicit strategy by simulating diverse lighting conditions (e.g., strong, normal, and weak illumination) to narrow the domain gap.Additionally, a mean coordinate fusion algorithm was applied to improve annotation completeness in real-world occlusion scenarios. CGA-ASNet leveraged contextual information between feature input keys for self-attention learning, captured global information, and expanded the receptive field, thus enhancing the network's representation capacity and enabling more accurate semantic understanding and precise localization. The experimental results demonstrated that CGA-ASNet achieved an F@.75 score of 94.2 and a mIoU of 82.4% in amodal segmentation accuracy in greenhouse scenarios. The results demonstrate that training with well-designed synthetic data can effectively support accurate occlusion-aware segmentation in real environments that offers a practical solution for tomato phenotyping in greenhouse.
Keywords: Amodal, segmentation, Occlusion-aware Segmentation, RGB-D image segmentation, plant phenotyping, Tomato, smart agriculture
Received: 12 Jul 2025; Accepted: 29 Aug 2025.
Copyright: © 2025 Li, Yin, Xing and Deng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
* Correspondence: Hanbing Deng, Shenyang Agricultural University, Shenyang, China
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.