Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 19 December 2025

Sec. Technical Advances in Plant Science

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1698843

Structure-aware completion of plant 3D LiDAR point clouds via a multi-resolution GAN-inversion network

Zhiming Wei,,Zhiming Wei1,2,3Jianing LongJianing Long4Zhihong ZhangZhihong Zhang5Xinyu Xue*Xinyu Xue1*Yitian Sun,Yitian Sun2,3Qinglong Li,Qinglong Li2,3Wu LiuWu Liu6Jingxin Shen,Jingxin Shen2,3Zhikai ZhangZhikai Zhang7Xiaoju LiXiaoju Li7Zhengguo MaZhengguo Ma7
  • 1Nanjing Institute of Agricultural Mechanization, Ministry of Agriculture and Rural Affairs, Nanjing, China
  • 2Shandong Academy of Agricultural Machinery Sciences, Jinan, China
  • 3Huang Huai Hai Key Laboratory of Modern Agricultural Equipment, Ministry of Agriculture and Rural Affairs, Jinan, China
  • 4College of Information and Telecommunications, China Agricultural University, Beijing, China
  • 5Faculty of Intelligence Technology, Shanghai Institute of Technology, Shanghai, China
  • 6Bureau of Agriculture and Rural Affairs of Yanggu County, Liaocheng, China
  • 7College of Mechanical and Electrical Engineering, Qingdao Agricultural University, Qingdao, China

Introduction: Three-dimensional (3D) point clouds acquired by LiDAR are fundamental for applications such as autonomous navigation, mobile robotics, infrastructure inspection, and cultural-heritage documentation. However, environmental disturbances and sensor limitations often yield incomplete or noisy point clouds, degrading downstream performance. This study addresses robust, high-fidelity point cloud completion under such practical conditions.

Methods: We propose an unsupervised deep learning framework, Multi-Resolution Completion Net (MRC-Net), which builds on ShapeInversion by integrating a Generative Adversarial Network (GAN) inversion strategy with multi-resolution principles. The architecture comprises an encoder for feature extraction, a generator for completion, and a discriminator to assess geometric integrity and detail. Two key designs enable strong performance without supervision: (i) a multi-resolution degradation mechanism that guides reconstruction across coarse-to-fine scales, and (ii) a multi-scale discriminator that captures both global structure and local details.

Results: Extensive experiments on multiple datasets demonstrate that MRC-Net achieves accuracy comparable to leading supervised approaches. On virtual datasets (e.g., CRN), MRC-Net attains an average Chamfer Distance (CD) of 8.0 and an F1 score of 91.3. On a custom dataset targeting agricultural scenarios, the model preserves object integrity across varying complexity: for regular cartons, it achieves CD 3.3 and F1 97.3; for structurally complex simulated plants, it maintains overall shape while delivering average CD 8.6 and F1 88.1.

Discussion: These results indicate that MRC-Net advances unsupervised point cloud completion by balancing global shape consistency with fine-grained detail. The method provides a reliable data foundation for downstream tasks—including autonomous navigation, high-precision 3D modeling, and agricultural robotics—thereby contributing to improved data quality in precision-agriculture and related domains.

1 Introduction

The important role of 3D point cloud data in representing object shapes and environmental structures is well-recognized across a myriad of fields, including agricultural computer vision, robotics, and autonomous driving (Li et al., 2022; Tang et al., 2024; Huang et al., 2025; Li et al., 2025a; Qiao et al., 2025). However, real-world scanning processes, particularly those involving Light Detection and Ranging (LiDAR) technology, often yield incomplete or partially observed point clouds due to factors like occlusions, limited fields of view, and sensor noise (Yan et al., 2022; Khorsandi et al., 2023; Wang et al., 2025a, Wang et al., 2025b). This data incompleteness presents a significant bottleneck, as it can severely impair the effectiveness of downstream tasks such as object detection, segmentation, and recognition.

These challenges are particularly consequential within the realm of precision agriculture. Technologies such as variable-rate application depend on detailed 3D field maps to enable the precise delivery of inputs like pesticides (Zhu et al., 2017; Fessler et al., 2021; Wang et al., 2023), herbicides (Andújar et al., 2013), or fertilizers (Zhao et al., 2021; Cruz Ulloa et al., 2022). Point cloud data is invaluable for creating prescription maps by capturing critical metrics like plant height, density, and volume (Abbas et al., 2020; Gu et al., 2020; Zeng et al., 2020; Zhao et al., 2020; Wang et al., 2022). Yet, when the underlying point clouds are incomplete, the accuracy of the prescription maps is compromised, thereby limiting the precision of agricultural inputs (Zhang et al., 2023). Therefore, the field of point cloud completion—which focuses on estimating and integrating missing data—is of paramount importance.

Point cloud completion technologies predominantly fall under two methodologies, supervised and unsupervised learning. Given the inherent limitations of real-world scanning data, the potential for geometric structure loss, and the challenges that supervised learning faces in these circumstances, the exploration and advancement of unsupervised solutions become critical. This need underscores the importance of continuous research in this field. To elaborate, supervised and unsupervised learning, the two primary methodologies in point cloud completion tasks, each present unique advantages and limitations when applied to the process of completion. Supervised approaches (Dai et al., 2017b; Dai et al., 2018; Mo et al., 2019; Zhang et al., 2020), typically necessitate paired data sets of incomplete and complete point clouds for training. Eigen et al. (2014) pioneered the recovery of voxel-based depth maps using a single image. Choy et al. (2016) proposed the 3D-R2N2 Encoder3D LSTM-Decoder to model the mapping from 2D images to 3D voxels. Subsequently, the 3D-Encoder-Predictor Networks (3D-EPN) introduced by Dai, et al. (2017b) was used to predict the missing parts of point clouds. The PointNet presented by Qi et al. (2017a) addressed the issue of point cloud disorder, while Yang et al. (2017) enabled 3D reconstruction from a single image and direct output of point cloud coordinates. Achlioptas et al. (2018) proposed a L-GAN model for generating depth in point clouds, and PCN developed by Yuan et al. (2018) was pioneered to facilitate shape completion based on deep learning. The PF-Net presented by Huang et al. (2020) introduced the Generative Adversarial Network (GAN) to generate complete point clouds, though it was limited to filling only spherical missing parts. Techniques that progress from coarse to fine detail (Huang et al., 2020; Liu et al., 2020; Wang et al., 2020; Xie et al., 2020) have shown significant results in 3D virtual datasets. Despite their theoretical utility, these methodologies present practical challenges in real-world 3D completion applications, primarily due to their need for paired incomplete-complete datasets for training. Furthermore, their generalization capacity is often limited, constraining the applicability of these methods to diverse or unanticipated scenarios. These limitations necessitate further advancements in the field to increase robustness and versatility of point cloud completion techniques.

Considering the intrinsic constraints of LiDAR in obtaining exhaustive real-world scanning data—making it less compatible with conventional supervised learning methods—an unsupervised learning-based approach is proposed for 3D point cloud completion. This approach aims to mitigate the dependency on complete data sets and enhance the model’s capacity to generalize to diverse and complex real-world applications. For instance, Chen et al. (2019) proposed pcl2pcl that employs adversarial training to complete unpaired point clouds. Further, RL-GAN Net proposed by Sarmad et al. (2019) integrates Reinforcement Learning (RL) with Generative Adversarial Networks (GAN) to enhance robustness and accelerate prediction. Additionally, ShapeInversion proposed by Zhang et al. (2021) was the first to incorporate GAN inversion into the process of 3D shape completion.

More recently, unsupervised and generative‐prior frameworks have advanced label-free point cloud completion. Lee et al. (2024) cast unpaired completion as an unbalanced optimal transport problem (UOT-UPC), improving robustness under class imbalance and without paired supervision. Complementarily, diffusion-based priors enable strong generalization: Huang et al. (2024) complete partial shapes at test time using 2D diffusion guidance with no additional training; Wei et al. (2025) leverage multi-view diffusion priors with fusion and consolidation to recover both global symmetry and thin structures; and Zhang and Liu (2024) introduce an occupancy-diffusion scheme with a coarse-to-fine design. To further enhance cross-domain robustness, Li et al. (2025b) propose DAPoinTr, a domain-adaptive Point Transformer that narrows distribution gaps and stabilizes predictions across datasets. Notwithstanding, existing unsupervised point cloud completion methods often primarily focus on the overall shape completion, while detailed structural aspects tend to be overlooked. This discrepancy poses a significant challenge as it may lead to the loss of essential, intricate details in the completion process.

To address the aforementioned challenges, this study proposes the Multi-Resolution Completion Network (MRC-Net), an unsupervised framework for 3D point cloud completion that preserves global structure while recovering fine-grained details by coupling a multi-resolution degradation mechanism with multi-scale discriminators. The approach introduces a detail-aware completion architecture that jointly models global geometry and local structures to complete imperfect, LiDAR-like inputs without paired supervision; employs a principled training strategy that synthesizes realistic partial observations (e.g., occlusion, sparsity, and noise) and applies multi-scale adversarial guidance to enhance fidelity and robustness across resolutions; and is validated through comprehensive experiments on synthetic benchmarks (e.g., CRN) and real-world agricultural datasets, including comparisons to recent baselines and ablation studies, demonstrating strong generalization under diverse levels of incompleteness and noise.

2 Materials and methods

2.1 Construction of point cloud completion network model

2.1.1 The application of MRC-net for point cloud completion

In this investigation, a methodology was developed which broadens the foundational work of ShapeInversion, addressing the intricate challenges presented by 3D point cloud completion. The procedural flow, as illustrated in Figure 1, is detailed as follows. The procedure of feature extraction started by extracting features from both complete and incomplete point clouds. The generator was initially trained to produce coarse point cloud data, which was subsequently degraded to resemble the input point cloud at varying resolutions. Multiple discriminators, guided by a specific loss function, further instructed the generator to create more realistic point clouds. In pre-training and initial generation, the model was pre-trained on a set of complete point cloud models to facilitate the extraction of latent vectors from the latent space. These vectors, encompassing vital features, enabled the generator (G) to produce an approximate complete point cloud. In multi-resolution down sampling, the Iterative Farthest Point Sampling (IFPS) method is employed to down sample the roughly generated complete point cloud into three point clouds with different resolutions: Xa,  Xa',  and Xa'' (consisting of 2048, 1024, and 512 points respectively). In degradation using multi-resolution mechanism, a mechanism known as Mk-Mask, perceptible as a clustering task using the K-Nearest Neighbor (KNN) method, degraded these point clouds at various resolutions into shapes akin to the incomplete input Xin, resulting in Xb,  Xb',  and Xb'', respectively. In progressive optimization and evaluation, the generator was progressively fine-tuned through a multi-stage reconstruction loss and multi-scale feature matching loss. This optimization process involves the MRC-Net matching a set of latent codes in the pre-trained GAN’s latent space, from which a complete point cloud structure was reconstructed. The completed point cloud was then subdivided and degraded into incomplete structures via IFPS and KNN, concluding with the integration of multi-discriminators to enhance the generator. The performance of the generation network was evaluated in the final step. This methodological framework served as a robust analytical approach, extending existing point cloud completion models, and set the stage for the development and refinement of techniques tailored to the intricate nature of 3D point cloud completion.

Figure 1
Flowchart depicting a process involving various steps. The input enters a pre-trained TreeGAN generator, which feeds into an IFPS module. This is followed by clustering with the KNN algorithm. Multi-scale degradation branches into three paths, leading to discriminators D1, D2, and D3. Outputs are combined to produce G. Each step is represented with icons and arrows indicating direction.

Figure 1. Unsupervised MRC-net application for 3D point cloud completion.

In addressing incomplete point cloud data acquisition by LiDAR sensors, the MRC-Net was introduced in this study. MRC-Net, illustrated in Figure 2, predicts the comprehensive structure of point clouds from incomplete depictions, adapting the GAN inversion concept for 3D structure completion. The architecture primarily incorporates a multi-scale degradation network Mk-Mask and an optimization network. Initially, the generator, collaborating with IFPS, fabricates multi-resolution point clouds. These are then degraded to mirror the incomplete input Xin via the multi-scale degradation mechanism. The process culminates with the optimization of the generator using a progressive, multi-stage reconstruction loss in combination with a multi-scale feature matching loss, followed by an evaluation of the resultant output.

Figure 2
Diagram illustrating a computational model with branching and graph convolution processes. Trees with branching and graph convolution are shown, progressing through stages labeled \(p^1\), \(p^2\), etc. Inputs include \(y\) and \(\delta\), processed by function \(G(\delta)\). Subsequent steps involve IFPS processing with degraded parts. The output is visualized with color-coded data flows \(D1\), \(D2\), and \(D3\), showing the relationship between stages.

Figure 2. Illustration of the MRC-net’s strategy: seeking latent codes ‘y’ within the hidden space of a pre-trained GAN and incrementally updating the parameters ‘δ’ of the generator ‘G’ to enhance the quality of generation.

2.1.2 Adaptive GAN inversion for enhanced 3D point cloud completion and multi-resolution reconstruction

In 3D point cloud completion tasks, GAN networks typically employ a generator to capture latent vectors, which are further used to accomplish the generation task. Traditional GAN networks, despite having dynamically changing input parameters, usually employ a fixed generator architecture. A pre-trained GAN was used on a set of complete point clouds from a particular class to obtain a collection of effective prior knowledge containing rich 3D geometric structure and semantic information. The generator exploits the latent vectors from this prior knowledge to generate complete point cloud models. Furthermore, the generator is allowed to utilize a combination of latent vectors y and generator parameters δ for real-time fine-tuning and continuous optimization of the generator. Applying GAN inversion to point cloud completion tasks enables the generator G to identify the optimal latent vectors for reconstructing complete point cloud models.

During the pre-training phase, a generator parameterized by δ is trained on complete point cloud models. In this study, the pre-trained generator is used to generate shape XaFm×3 from latent vectors yFd. The incomplete input point cloud model Xin can be reconstructed into a complete point cloud model closely resembling the ground truth through G, as shown in (Equation 1):

y*=argminyFdE(G(y; δ), Xin), Xa*=G(y*; δ) (1)

An approach inspired by GAN inversion for image generation in 2D images (Pan et al., 2021) has been adopted to achieve improved completion results with the generator G. For each point cloud, the generator is adapted by updating the generator parameters δ, while the latent vectors y was concurrently optimized for the progressive fine-tuning of the generator. This was represented in (Equation 2).

δ, y*=argminy, δE(G(y; δ), Xin) (2)

The GAN inversion initialization began with an extensive random sampling of latent vectors. The latent vector associated with the smallest ϵ value was then chosen as the initial parameter for fine-tuning the generator in relation to y. In real-time, the parameters y and δ are updated using gradient descent as illustrated in (Equation 3). Subsequently, the generator integrated with the latent vectors to construct a complete Xa from the incomplete input Xin. Leveraging an approach inspired by FPN (Lin et al., 2017), the complete point cloud structure Xa was then merged with the multi-resolution degradation mechanism Mk-Mask. This process ultimately leads to the transformation of the complete 3D model into point cloud models of diverse resolutions, analogous to the input.

y*=argminyRdE(N(G(y; δ)), Xin) (3)

2.1.3 Efficient feature extraction from point cloud data using iterative farthest point sampling

This study extracted a limited number of key feature points from large quantities of point cloud data to efficiently represent the model’s structural information, thus enhancing the computational efficiency of deep learning networks. The IFPS method, as utilized in PointNet++ (Qi et al., 2017), was employed to extract the principal feature points from the point cloud data. This method, despite its increased computational complexity compared to uniform and random sampling, provided a more uniform distribution of sampled points across the entirety of the model’s point set. In Figure 3, point clouds at various resolutions were depicted. It became evident that even as the point quantity decreases in low-resolution point clouds, the fundamental structural features of the 3D point cloud were still adequately captured. The application of the IFPS method ensures computational efficiency while preserving essential geometric and semantic information. Notably, even in reduced-resolution representations, the structural integrity of the 3D point cloud was maintained. Through the use of IFPS, an equilibrium was struck between computational demands and the preservation of crucial geometric and semantic aspects.

Figure 3
Three scatter plots show progressively less dense arrangements of blue dots, resembling abstract leaf or tree-like patterns. Each pattern decreases in dot density from left to right.

Figure 3. Demonstration of point clouds at varied resolutions.

2.1.4 Multi-resolution degradation mechanism component in point cloud completion

The generator used latent vectors to create a complete point cloud structure, denoted as Xa = G(δ). Through the application of IFPS, initial point cloud shapes of varying resolutions were produced. The multi-resolution degradation mechanism, a variant of the recent degradation transformation module k-Mask (Zhang et al., 2021), referred to as Mk-Mask in this study, served as the initial comparator for the multi-resolution incomplete point cloud model. Given the irregular and unstructured nature of point clouds in space, variations in spatial position information could occur within the same point cloud. Consequently, a direct degradation of the complete point cloud into the desired shape was not feasible via the degradation function alone. The Mk-Mask mechanism addressed this challenge by adopting ideas from the KNN algorithm, used for representing spatial-local correlations in 2D classification tasks, and from the Point CNN (Dai et al., 2017b) feature extractor, used for 3D point cloud tasks. The Mk-Mask was not confined to the degradation of generator-produced point clouds at a single resolution but could simultaneously degrade point clouds across multiple resolutions. Specifically, it performed a clustering task to group generated points into different resolutions, preserving the structure and information of the point cloud. The KNN algorithm aided in capturing the local structure of the point cloud by considering each point’s nearest neighbors. In contrast, Point CNN extracted hierarchical features from point clouds. By integrating these concepts, the Mk-Mask degraded the generated point cloud structures at multiple resolutions to more closely approximate the incomplete input point cloud. This mechanism was pivotal in ensuring alignment between the generated point clouds and the real-world structures and semantics represented by the incomplete input data, contributing to the generation of more accurate and realistic point cloud completions.

The IFPS method was employed to sample the generator-produced point cloud into various resolutions, denoted as Fi, with sizes of 2048, 1024, and 512 for i = 1, 2, 3, respectively. This strategy enabled the multi-resolution degradation module, Mk-Mask, to concentrate on essential feature points. Subsequently, Mk-Mask degraded the multi-resolution complete point clouds into forms akin to the incomplete input, Xin. Specifically, the degradation through F3 to MF3 primarily ensured the effect on the backbone structures. The transition from F2 to MF2 predominantly safeguards the secondary overall structures, while MF1 chiefly focused on details. By adopting a multi-resolution degradation approach, it was possible to extract more features pertinent to both high-resolution structures and semantics and low-resolution structures and semantics compared to single-resolution degradation when calculating feature distance losses. This procedure facilitated the generator to produce the most effective point cloud completion by using the optimal solution. Mk-Mask’s utilization of multi-resolution degradation was crucial to capturing different levels of detail in the point cloud. The mechanism allowed for the preservation of a wide array of features, formed primary structural components to finer details.

Furthermore, a methodology inspired by K-nearest neighbors (KNN) was implemented in this investigation to enhance the precision and robustness of the degradation mechanism when dealing with incomplete point clouds, represented as Xin. This approach uses Euclidean distance as the metric for proximity assessment. Each point within Xin, designated as bi, is assigned a neighborhood defined by a parameter K and compared against the complete point cloud, Xa, produced by the generator. The collection of points within Xa located within the K-defined neighborhood relative to Xin are represented as Mkxa(bi). By executing this procedure for every point within the point cloud, the complete point cloud generated by the generator, Xb, is gradually degraded into a configuration that closely mirrors the incomplete input Xin. Therefore, as shown in (Equations 3, 4), Xb is composed of the aggregate of points encapsulated within these K-defined vicinities.

Xb=i=1n{UjMkXa(bi)biXin} (4)

2.1.5 Integrated loss function components of the MRC-net

The loss function of the MRC-Net integrated two distinct elements, the multi-stage reconstruction loss and the multi-scale feature distance loss. The former, the reconstruction loss, functioned to measure discrepancies between the generated point cloud and the known ground truth. Meanwhile, the latter, the feature distance loss, employed a discriminator, which was trained in tandem with the generator, for feature matching within the feature space, with the objective of maintaining semantic and structural consistency. The calculation of the feature distance loss was facilitated through the utilization of the L1 distance metric. Together, these two interlinked elements of the MRC-Net’s loss function guide the model effectively towards accurate point cloud completion.

2.1.5.1 Multi-stage reconstruction loss and the use of chamfer distance in point cloud completion

In the sphere of point cloud completion, two commonly utilized metrics for assessing the effectiveness of point cloud structure reconstruction were CD and Earth Mover’s Distance (EMD). Despite EMD’s heightened sensitivity towards capturing finer detail in point cloud generation, it exhibited limitations in our context since the training process entails degrading the complete shape, which was generated, to a form analogous to the input incomplete point cloud for comparison purposes. It was important to note that the number of points in the degraded shape may diverge from that in the input incomplete point cloud. Conversely, CD, being differentiable, supported more efficient computation. Hence, CD is chosen as the metric for reconstruction loss, as illustrated in (Equation 5).

ƐCD(Xb,Xin)=1|Xb|bXbmincXinbc22 +1|Xin|cXinminbXbcb22(5)

The CD was used to determine the mean shortest point distance between the degraded point cloud, Xa, and the incomplete input point cloud, Xin. Two versions, CD-P and CD-T (Dai et al., 2017a), support the calculation of the average closest point distance between the generated and real point clouds. CD-P was defined as CD-P=(LX,Y+LY,X)/2, while CD-T is expressed as CD-T=EX,Y+EY,X. In all training phase experiments conducted within this study, CD-T was utilized, which involves the computation of the L2 distance.

Utilizing the IFPS method, the incomplete input Xin is partitioned into two novel sets: Xin' and Xin'', serving as the incomplete input feature points of size N3. Further, after GAN inversion, the complete point cloud was sampled via IFPS into three point clouds at disparate resolutions, each of size N3, analogous to the incomplete input, labeled as Xb, Xb', and Xb''.

As a result, the multi-stage reconstruction loss is assembled from three components: ƐCD1, ƐCD2, and ƐCD3. These are weighed by the hyperparameters α and β, as per (Equations 6). ƐCD1, the first term, determines the squared distance between the refined points of the degraded point cloud Xb and the actual incomplete input Xin. The remaining terms, ƐCD2 and ƐCD3, calculate the squared distances between the primary structure points Xb' and secondary structure points Xb'', and their corresponding real incomplete inputs Xin' and Xin''. Thus, the multi-stage reconstruction loss is formulated as per (Equation 6).

TC=ƐCD1(Xb,Xin)+αƐCD2(Xb',Xin')+βƐCD3(Xb'',Xin'')(6)
2.1.5.2 Adopting feature matching loss and multi-scale discriminators for 3D point cloud tasks

Inspired by AniGAN’s (Li et al., 2021) use of feature matching loss for 2D image tasks to yield more natural data across various dimensions, and the proposed use of multiple GAN discriminators for a single image in 2D image tasks (Durugkar et al., 2016), this study adopted an analogous approach for 3D point cloud tasks. While the reconstruction loss function primarily emphasizes the overall structure, it failed short in finely detailing the completion process. To ameliorate this, L1 distance in the discriminator feature space was employed as a distance metric, aiming to optimize the generation of network details via discriminator’s distance metrics. Throughout the experimentation process, three discriminators, D1, D2, and D3, with identical network structures, were deployed. These discriminators operate on different point cloud resolutions, as determined by the multi-scale degradation mechanism. The choice of discriminators learned the method employed in 2D image tasks (Pan et al., 2021), where discriminators pretrained alongside the generator for the same task were used. As these discriminators exhibited high compatibility with the generator, they could finely tune the generator’s output distribution, thereby enhancing the quality of the results. This setup was reflected in (Equation 7).

TD=D1(Xb)-D1(Xin)1+D2(Xb')-D2(Xin')1+D3(Xb'')-D3(Xin'')1(7)

The overall loss function is as shown in (Equation 8):

T=hCTC+hDTD(8)

The total loss function, as applied in this study, had an important role in both the generation and reconstruction of the complete shape. This function consisted of two major components, the reconstruction loss, weighted by hC, and the feature matching loss, weighted by hD. A crucial aspect of these weight parameters was that their sum was constrained to equal one, ensuring a balanced influence of both components on the overall loss. This restriction ensured that the overall loss function accurately reflects the interplay between the reconstruction and feature matching aspects, thus effectively guiding the process of point cloud completion.

2.1.6 Comparison to baseline and design rationale

This subsection contrasts the proposed MRC-Net with the canonical baseline adopted in this study and clarifies the rationale behind each design choice. The baseline is taken to be a single-resolution encoder–decoder trained with reconstruction losses on synthetically partialed inputs. By comparison, MRC-Net is designed to address the principal failure modes encountered in real LiDAR scans—occlusion, sparsity/limited field of view, and measurement noise—through coordinated changes to input modeling, network architecture, and learning objectives. At the level of input modeling, MRC-Net employs a multi-resolution degradation mechanism that synthesizes LiDAR-like partial observations across several scales by stochastically composing occlusion masks, sparsity schedules, and noise injections. This broadens the training distribution to better match the statistics of real-world partial scans and thereby improves robustness without requiring paired supervision. In contrast, the baseline’s single-resolution partialing under-represents difficult cases such as large self-occlusions or highly anisotropic sparsity, which often leads to brittle behavior at test time. Architecturally, MRC-Net adopts a detail-aware global–local design that aggregates multi-resolution features to preserve coarse, topology-stabilizing structure while recovering high-frequency details. The coupling of global geometry with local structural cues helps retain symmetry and volumetric plausibility and, at the same time, reconstructs thin and highly curved parts. The baseline’s single-scale pathway tends to either over-smooth these fine structures or hallucinate disconnected fragments when incompleteness is severe. In terms of the learning objective, MRC-Net augments standard reconstruction losses with multi-scale adversarial guidance and geometric-consistency regularizers that align normal and regulate local curvature across resolutions. These additions encourage shape plausibility and suppress both over-smoothing and implausible hallucination under heavy missingness and noise. By comparison, reconstruction-only training in the baseline frequently trades detail fidelity for numerical stability, yielding overly conservative completions.

2.2 Dataset selection and preprocessing of point cloud completion based on MRC-net

2.2.1 Dataset composition, preprocessing, and configuration for model evaluation

This study employed both synthetic and real-world scan datasets to rigorously evaluate the performance of the model across various contexts. For synthetic data, the baseline dataset, CRN (Wang et al., 2020), was used, selecting and annotating models from eight categories: chair, car, cabinet, couch, lamp, plane, table and watercraft. These models were organized in a common directory. The dataset was split as follows: 60% for training, 20% for validation, and 20% for testing. The input point cloud data undergo common preprocessing procedures, such as normalization of coordinate scales to the range [-1, 1], and recentering them around the origin. Both complete and incomplete point clouds within the synthetic dataset comprise 2048 points, with the incomplete point clouds forming subsets of their complete counterparts. Real-world scan datasets mirror the ones employed in ShapeInversion (Zhang et al., 2021) and pcl2pcl (Chen et al., 2019), which includes (1) KITTI (cars) (Geiger et al., 2012), (2) ScanNet (chairs and tables) (Dai et al., 2017a), and (3) MatterPort3D (chairs and tables) (Chang et al., 2017).

2.2.2 Data acquisition, preprocessing, and augmentation for point cloud samples

The dataset for this experiment comprised point cloud data samples acquired through the 3D scanning process of a diverse range of objects, including cartons and simulated plants, as illustrated in Figure 4. The 3D scanning method effectively captures the point cloud data of these objects in various sizes.

Figure 4
A three-dimensional scanning setup featuring an artificial plant on a circular base, a three-dimensional scanner mounted on a tripod, and a laptop displaying a virtual point cloud dataset of the plant.

Figure 4. Three-dimensional scanner scanning of artificial plant physical model.

The collected 3D point cloud data undergone a processing step, utilizing (Equation 9) to calculate the offset in each direction (x, y, z). This offset was then subtracted from the entire dataset, resulting in the center of the point cloud data being realigned to the coordinate origin. Consequently, this process ensured the centralization of the complete dataset.

{Xm=Xmin+(Xmax-Xmin2)Ym=Ymin+(Ymax-Ymin2)Zm=Zmin+(Zmax-Zmin2)(9)

In the normalization process, the maximum value (Mmax) from the longest axis across the x, y, and z directions of the point cloud data was identified. This maximum value was then used to normalize all the points within the dataset, effectively bounding the data within the (-1, 1) range. To generate a robust dataset, the normalized data was subject to augmentation techniques, involving random rotations about the coordinate center across the x, y, and z axes. This allowed for the acquisition of point cloud models from various perspectives, enhancing the diversity of the dataset. In keeping with practical applications, the study incorporated 3D point cloud data collected. This data served as the incomplete input with a resolution of 2048 points within the network. Post-reconstruction and augmentation, this 3D point cloud data formed the complete dataset used in this study.

2.2.3 Evaluation metrics for shape completion performance

To evaluate shape completion performance for the paired complete and incomplete data, methodologies from ShapeInversion (Lin et al., 2017) were utilized, incorporating CD and F1 scores as metrics. The CD measures the average minimum distance between the generated point cloud and the ground truth, with larger values suggesting greater discrepancies and smaller values indicating better reconstruction accuracy. The F1 score, a harmonic mean of precision and recall, provided an overall measure of model performance. For real-world datasets that lack ground truth data and consist only of incomplete inputs, Unidirectional Chamfer Distance (UCD) and Unidirectional Hausdorff Distance (UHD) were adopted as evaluative metrics (Lin et al., 2017). Both metrics provided a measure of similarity between the reconstructed and incomplete point cloud datasets, with smaller distances suggesting better alignment with the original incomplete data.

2.2.4 Implementation and configuration details of the MRC-net network

The MRC-Net network developed in this study operated on a Linux server and is primarily implemented using PyTorch (version 1.2.0) and Python (version 3.7). Experiments were conducted on a server equipped with a Quadro RTX 5000 graphics card. During the pre-training phase, the ShapeNet dataset served as the primary dataset and the ADAM optimizer is employed with an initial learning rate of 0.0001. The batch size was configured to 64 and the number of epochs is set to 2000. This study focused on enhancing precision by training individual models specific to each object class. Within the generative network, the batch size was configured to 1, allowing for the output of completion results sequentially. For all cases, the final complete point cloud resolution output by the generative network was 2048. The configuration of these parameters ensures the reliable and efficient operation of the MRC-Net, yielding high-quality results.

3 Results and discussion

3.1 Visualization and efficacy evaluation of MRC-net across diverse object categories

Figure 5 displays the completion results of the proposed MRC-Net across eight varied categories: chair, car, cabinet, couch, lamp, plane, table, and watercraft. The representation emphasized the proficiency of MRC-Net in accurately reconstructing these diverse objects, capturing intricate details. By employing a consistent dataset for both training and testing phases, the evaluation of MRC-Net’s efficacy was ensured to be impartial and grounded. This dataset, supplied by CRN (Wang et al., 2020), established a level playing field for comparisons with multiple existing supervised and unsupervised techniques under identical virtual scanning conditions. The sequence demonstrated in the figure initiated with the input of fragmented point clouds for the mentioned categories, progresses through an intermediate degradation phase, and culminates in the generation of the finalized complete point clouds. This sequence vividly validated the ability of the proposed method to adeptly complete point clouds.

Figure 5
A grid showing a progression of point cloud models for objects like a chair, car, cabinet, couch, lamp, plane, table, and watercraft. Rows represent different stages: “Input” with faint shapes, “Degenerated Point Cloud” in red, “Complete” in blue, and “Ground Truth” in yellow, demonstrating refinement and completion of each model.

Figure 5. The performance of MRC-Net for eight different object types.

3.2 Comparative performance analysis of MRC-net in point cloud completion on virtual dataset

The performance of the proposed MRC-Net was evaluated using a virtual dataset. For a comprehensive understanding, its efficacy was benchmarked against several established point cloud completion techniques, encompassing both supervised and unsupervised approaches. The central goal of this study was to assess the quality of the completed point clouds, focusing on their detail, overall structure, and other crucial facets. Such evaluations are critical for determining the capability of MRC-Net in addressing missing sections. The CD and F1 scores served as key performance metrics, providing a quantitative measure of the completion outcomes. As presented in Table 1, the results of the supervised MRC-Net were described using the format [CD↓/F1↑], with CD values scaled by 104 for clarity. Lower CD values indicated superior reconstruction quality, whereas higher F1 scores signified improved accuracy and completeness of the reconstructed point cloud. MRC-Net’s average CD values were found to be comparable with other supervised techniques and significantly surpassed results from unsupervised pcl2pc methods. Its slight edge over the ShapeInversion technique could be ascribed to its multi-stage refinement mechanism.

Table 1
www.frontiersin.org

Table 1. Shape completion results from supervised network. Metrics are reported in the format [CD↓/F1↑] where CD values are multiplied by 104.

From Table 1, it was observed that the CRN method consistently outperformed other techniques, denoting its robustness in shape reconstruction across varied objects. Categories such as ‘plane’ were easily reconstructed by most methods, whereas ‘lamp’ and ‘cabinet’ proved challenging, as reflected by their lower F1 scores. Yet, most methodologies achieved an F1 score above 80%, emphasizing their effectiveness in point cloud completion tasks. Collectively, while multiple techniques demonstrated notable proficiency, CRN stood out across numerous categories. It was also evident that the reconstruction’s effectiveness was subject to the inherent intricacies of different objects. Table 2 detailed the outcomes of unsupervised methods. Performance was gauged using CD, where lower scores were optimal, and the F1 score, with higher values preferred. The Pcl2pcl, as proposed by Chen et al. (2019), generally lagged in precision compared to other methods. In contrast, ShapeInversion, developed by Zhang et al. (2021), showcased notable improvements over Pcl2pcl, especially in the ‘Cabinet’ and ‘Chair’ categories. Yet, MRC-Net outdid both, presenting reduced CD values and superior F1 scores. For categories like ‘Plane’ and ‘Lamp’, MRC-Net’s advancements were particularly evident. On average, its results, with a CD of 13.9 and an F1 of 84.4%, proved its dominance in unsupervised shape completion. The MRC-Net was validated as the preeminent method in this comparative assessment, manifesting the strength and adaptability of its foundational algorithm.

Table 2
www.frontiersin.org

Table 2. Effect of unsupervised network shape complementation.

It could be observed a clear indication of the efficacy of the CRN method in shape reconstruction across various objects. While its consistent performance across categories was desirable, it was essential to recognize the nuances and limitations associated with it, especially when compared to results focusing on unsupervised methods. MRC-Net, showcased exceptional results in unsupervised shape completion. This distinction was crucial given the industry’s increasing emphasis on unsupervised learning methodologies. Despite the superior performance of CRN, MRC was the recommended approach. There exist inherent challenges with supervised learning. The need for paired incomplete and complete datasets for training makes their application in real-world scenarios challenging. CRN, being more reliant on supervised techniques, might be less versatile in unanticipated scenarios due to its dependence on specific training data. However, the versatility unsupervised learning was greater than supervised one. The strength of MRC-Net lies in its foundational algorithm that adapts to unsupervised shape completion. In situations where obtaining paired datasets became challenging, MRC-Net offers a solution that is more flexible and adaptable. Moreover, in terms of generalization capacity, supervised methods often had limited generalization capabilities. This limitation implied that while CRN might excel in specific, known scenarios, it might not be as effective in diverse or unanticipated situations. In contrast, MRC-Net, with its unsupervised learning foundation, offered a broader applicability range, making it more suitable for dynamic real-world applications. In addition, the industry’s trajectory appears to be favoring advancements in unsupervised solutions due to the inherent limitations of real-world scanning data and potential for geometric structure loss. Recommending MRC aligned with this trajectory and emphasizes the importance of continuous research in unsupervised methodologies. Therefore, while CRN demonstrates superior results in specific controlled scenarios, MRC-Net’s adaptability, versatility, and alignment with industry trends make it a recommended choice for point cloud completion tasks.

In the findings presented, an enhanced method’s performance was rigorously assessed against the established ShapeInversion technique, specifically regarding the completion of structures featuring intricate characteristics. As discernibly illustrated in Figure 6, across the eight categories of the CRN dataset, this improved methodology was observed to markedly outperform ShapeInversion. A meticulous scrutiny, particularly into categories such as cabinets, chairs, and cars, brought to light its superior proficiency both in the distribution of points encompassing the overall structure and in the point density within more nuanced regions. Delving further into specific categories, such as couches and tables, revealed even more nuanced insights. When presented with input point clouds of atypically structured couches, the ShapeInversion technique was found to proficiently supplement the upper portions of the couches, yet neglecting the foundational support frames beneath. Conversely, the introduced method not only exhibited a comprehensive completion of the entire couch structure but also displayed a keen detection and reconstruction of the distinctive, brief protruding support frame at the base. This advantageous performance continuity was also identified within the table category, where the proposed approach was observed to equally emphasize the table’s surface as well as the often-overlooked crossbeam. Such meticulous attention ensures a harmonious equilibrium between the generation of intricate details and the creation of the overarching structure. The obtained results validated MRC-Net’s elevated competence in furnishing highly precise and detail-oriented structure completion, especially when confronted with multifaceted and detailed entities, when juxtaposed against alternative techniques.

Figure 6
Comparison of shape reconstruction for various objects: chair, car, cabinet, couch, lamp, plane, table, and watercraft. Rows show input, Shapeinversion, MRC-Net, and GT results. Colors denote varying reconstruction quality.

Figure 6. A visualization of the point cloud completion results attained by MRC-Net on the virtual scanning dataset.

3.3 Analysis of comparative experiments on real datasets

The quality of data obtained through real-world scanning was often affected by factors such as environment and equipment, resulting in excessive noise and significant missing information. Qualitative comparisons on real-world scans are shown in Figure 7, and the corresponding quantitative results—evaluated using UCD and UHD are summarized in Table 3. Therefore, this study further employed the method developed within it for real-world scan datasets ScanNet (Lin et al., 2017), MatterPort (Chang et al., 2017), KITTI (Geiger et al., 2012) and compares it with existing methods. It was notable that most of the existing methods primarily focus on generating the overall geometric structure, while the method proposed in this study places greater emphasis on the generation of geometric structural details. The comparison experiment results were quantitatively evaluated using UCD (Unidirectional Chamfer Distance) and UHD (Unidirectional Hausdorff Distance) (Choy et al., 2016; Zhang et al., 2021). The experimental results showed that the method proposed in this study was generally superior to other existing methods, particularly in terms of the reliability of the generated structures. Figure 7 displays the visual comparison of completion results on real-world scanned datasets. For the chair and table examples in MatterPort, it could be observed that the Shapeinversion method achieves fairly ideal completion results in terms of the overall structure of the chairs and tables. However, there were holes and missing parts of varying degrees, particularly in the chair’s backrest and the corners of the table. The method proposed in this study handled these types of issues quite effectively. In the case of the ScanNet dataset for chairs and tables, the ShapeInversion method appears to lack the connection between the incomplete chair input and the armrests when generating chairs. Additionally, there are excess parts generated at the upper left corner leg of the table. Compared to the method proposed in this study, the reliability and realism of the generated structures were superior to those of ShapeInversion. The method proposed in this study demonstrates better performance in terms of handling detailed structural components and generating more reliable and realistic structures, especially in datasets with real-world scans that often have noisy and incomplete data.

Figure 7
Comparison image showing 3D point cloud reconstructions of objects using different methods. Top row: gray input images from KITTI, MatterPort3D, and ScanNet datasets. Middle row: Shapeinversion results in light blue. Bottom row: Our method's results in darker blue, showing clearer and more detailed reconstructions for cars, chairs, and tables.

Figure 7. Comparative analysis of point cloud completion in real-world scanned datasets: ShapeInversion vs. proposed method.

Table 3
www.frontiersin.org

Table 3. Complementary results for the real dataset. The results were evaluated using [UCD↓/UHD↓] with UCD scaled by 104,UHD scaled by 102.

3.4 Performance assessment and practical application of the revised MRC-net algorithm

In the experiments undertaken, a revised algorithm was assessed against its predecessor, specifically focusing on detail generation. The results indicated that the enhanced algorithm exhibited superior performance. To validate the algorithm’s practical application, tests were conducted using datasets created from standard cartons and artificially simulated plants. As can be inferred from Figure 8; Table 4, the updated MRC-Net algorithm, when applied to datasets of standard objects, showed impressive results, evidenced by average CD and F1 values of 3.3 and 97.3, respectively. Upon a detailed analysis of the edge completions for simulated plants, it was noted that the algorithm performed desirable for plants with straightforward external contours. In contrast, for those with more complex exteriors, minor discrepancies in edge completion were detected, yet the integrity of the overall structure was retained. Additional data from Table 4 revealed the algorithm’s resilience and efficacy, even when applied to plants with intricate contours, demonstrated by average CD and F1 scores of 8.6 and 88.1, respectively. In essence, the findings affirmed the enhanced capability of the point cloud completion deep learning model, introduced in this research, to adeptly complete irregular structures.

Figure 8
Grid showing a series of images for five plants and a carton box across five stages: Input, Degenerated Point Cloud, Complete, Ground Truth, and Picture. Each column represents a plant or the carton, and each row shows a different stage of data visualization or rendering.

Figure 8. Comparative analysis of MRC-net algorithm performance on standard and irregular objects.

Table 4
www.frontiersin.org

Table 4. The shape completion effects of cartons and simulated plants on the dataset benchmark.

4 Conclusions

This research introduced and validated the MRC-Net, an enhanced unsupervised network model predicated on ShapeInversion. Building upon the principles of GAN inversion, MRC-Net innovatively integrates a multi-resolution degradation mechanism with multi-scale discriminators. This design enabled the model to effectively reconstruct both the global structure and intricate details of 3D objects from incomplete point clouds, addressing a significant limitation of prior models which often neglect fine-grained features.

The superiority of MRC-Net was substantiated through comprehensive experiments. On virtual datasets, the proposed network demonstrated performance comparable to existing supervised methods, achieving an average CD of 8.0 and an F1 score of 91.3. Furthermore, its practical efficacy was validated on custom datasets, where it exhibited exceptional proficiency in completing regular objects (cartons: CD 3.3, F1 97.3) and maintained high structural integrity for complex, irregular objects (simulated plants: CD 8.7, F1 88.1). These results confirm that MRC-Net offers a high-performance, stable, and generalizable solution for point cloud completion, providing a reliable data foundation for downstream applications.

Despite strong performance across synthetic and real-scan benchmarks, MRC-Net still exhibits minor artifacts along highly irregular or noisy boundaries. In addition, the multi-resolution degradation, multi-scale discrimination, and multi-stage objectives introduce non-trivial training and inference costs relative to minimalist baselines. A further methodological limitation is that, in our unsupervised GAN-inversion setting, these components are tightly coupled; destructive, module-removal ablations tend to destabilize optimization and conflate training failure with true marginal utility, yielding analyses that are less informative than holistic comparisons.

Future work will therefore proceed along three axes: efficiency—via model distillation, dynamic-resolution routing, and lighter-weight discriminators; robustness—through richer geometric priors, uncertainty-aware completion, and test-time adaptation to severely degraded scans; and principled component analysis—using larger-scale studies under evaluation protocols that avoid objective degeneracy, alongside expansion to more diverse real-world datasets. We expect these directions to reduce compute requirements, improve boundary fidelity, and broaden the practical applicability of MRC-Net.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author.

Author contributions

ZW: Writing – review & editing, Conceptualization, Validation. JL: Writing – review & editing, Writing – original draft. ZhihZ: Writing – original draft. XX: Writing – review & editing, Funding acquisition, Supervision. YS: Writing – review & editing, Investigation. QL: Writing – review & editing. WL: Writing – review & editing. JS: Writing – review & editing. ZhikZ: Data curation, Writing – review & editing. XL: Data curation, Writing – review & editing. ZM: Data curation, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This research was funded by the Key R&D Program of Shandong Province (Science and Technology Demonstration Project) (No. 2022SFGC0204), Shanghai Pujiang Program (23PJ1411900), Shandong Provincial Agricultural Machinery R&D, Manufacturing, Promotion, and Application Integration Pilot Project (No. NJYTHSD-202306). This work was also supported by the earmarked fund for SDARS-5.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbas, I., Liu, J., Faheem, M., Noor, R. S., Shaikh, S. A., Solangi, K. A., et al (2020). Different sensor based intelligent spraying systems in Agriculture. Sensors Actuators A: Phys 316, 112265. doi: 10.1016/j.sna.2020.112265

Crossref Full Text | Google Scholar

Achlioptas, P., Diamanti, O., Mitliagkas, I., and Guibas, L. (2018). “Learning representations and generative models for 3d point clouds,” in International conference on machine learning (Sweden: PMLR), 40–49.

Google Scholar

Andújar, D., Escolà, A., Rosell-Polo, J. R., Fernández-Quintanilla, C., and Dorado, J. (2013). Potential of a terrestrial LiDAR-based system to characterise weed vegetation in maize crops. Comput. Electron. Agric 92, 11–15. doi: 10.1016/j.compag.2012.12.012

Crossref Full Text | Google Scholar

Chang, A., Dai, A., Funkhouser, T., Halber, M., Niessner, M., Savva, M., et al (2017). Matterport3d: Learning from rgb-d data in indoor environments. arXiv preprint arXiv:1709.06158. doi: 10.48550/arXiv.1709.06158

Crossref Full Text | Google Scholar

Chen, X., Chen, B., and Mitra, N. J. (2019). Unpaired point cloud completion on real scans using adversarial training. arXiv preprint arXiv:1904.00069. doi: 10.48550/arXiv.1904.00069

Crossref Full Text | Google Scholar

Choy, C. B., Xu, D., Gwak, J., Chen, K., and Savarese, S. (2016). “3d-r2n2: A unified approach for single and multi-view 3d object reconstruction,” in Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part VIII 14 (Cham, Switzerland: Springer), 628–644.

Google Scholar

Cruz Ulloa, C., Krus, A., Barrientos, A., Del Cerro, J., and Valero, C. (2022). Trend technologies for robotic fertilization process in row crops. Front. Robotics AI 9, 808484. doi: 10.3389/frobt.2022.808484

PubMed Abstract | Crossref Full Text | Google Scholar

Dai, A., Chang, A. X., Savva, M., Halber, M., Funkhouser, T., and Nießner, M. (2017a). “Scannet: Richly-annotated 3d reconstructions of indoor scenes,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 5828–5839.

Google Scholar

Dai, A., Ruizhongtai Qi, C., and Nießner, M. (2017b). “Shape completion using 3d-encoder-predictor cnns and shape synthesis,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 5868–5877.

Google Scholar

Dai, A., Ritchie, D., Bokeloh, M., Reed, S., Sturm, J., and Nießner, M. (2018). “Scancomplete: Large-scale scene completion and semantic segmentation for 3d scans,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA, USA: IEEE Computer Society 4578–4587.

Google Scholar

Durugkar, I., Gemp, I., and Mahadevan, S. (2016). Generative multi-adversarial networks. arXiv preprint arXiv:1611.01673. doi: 10.48550/arXiv.1611.01673

Crossref Full Text | Google Scholar

Eigen, D., Puhrsch, C., and Fergus, R. (2014). Depth map prediction from a single image using a multi-scale deep network. Adv. Neural Inf. Process. Syst 27, 2366–2374. doi: 10.5555/2969033.2969091

Crossref Full Text | Google Scholar

Fessler, L., Fulcher, A., Schneider, L., Wright, W. C., and Zhu, H. (2021). Reducing the nursery pesticide footprint with laser-guided, variable-rate spray application technology. HortScience 56, 1572–1584. doi: 10.21273/HORTSCI16157-21

Crossref Full Text | Google Scholar

Geiger, A., Lenz, P., and Urtasun, R. (2012). “Are we ready for autonomous driving? the kitti vision benchmark suite,” in 2012 IEEE conference on computer vision and pattern recognition (Los Alamitos, CA, USA: IEEE), 3354–3361.

Google Scholar

Gu, C., Wang, X., Wang, X., Yang, F., and Zhai, C. (2020). Research progress on variable-rate spraying technology in orchards. Appl. Eng. Agric 36, 927–942. doi: 10.13031/aea.14201

Crossref Full Text | Google Scholar

Huang, Z., Lee, W. S., Zhang, P., Jeon, H., and Zhu, H. (2025). SASP: Segment any strawberry plant, an end-to-end strawberry canopy volume estimation. Smart Agric. Technol 11, 101017. doi: 10.1016/j.atech.2025.101017

Crossref Full Text | Google Scholar

Huang, T., Yan, Z., Zhao, Y., and Lee, G. H. (2024). ComPC: completing a 3D point cloud with 2D diffusion priors. arXiv preprint arXiv:2404.06814. doi: 10.48550/arXiv.2404.06814

Crossref Full Text | Google Scholar

Huang, Z., Yu, Y., Xu, J., Ni, F., and Le, X. (2020). “Pf-net: Point fractal network for 3d point cloud completion,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 7662–7670.

Google Scholar

Khorsandi, A., Tanino, K., and Noble, S. D. (2023). The effects of sampling and instrument orientation on LiDAR data from crop plots. Front. Plant Sci 14. doi: 10.3389/fpls.2023.1087239

PubMed Abstract | Crossref Full Text | Google Scholar

Lee, T., Choi, J., Choi, J., and Kang, M. (2024). Unsupervised point cloud completion through unbalanced optimal transport. arXiv preprint arXiv:2410.02671. doi: 10.48550/arXiv.2410.02671

Crossref Full Text | Google Scholar

Li, H., Huang, K., Sun, Y., Lei, X., Yuan, Q., Zhang, J., et al (2025a). An autonomous navigation method for orchard mobile robots based on octree 3D point cloud optimization. Front. Plant Sci 15. doi: 10.3389/fpls.2024.1510683

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Z., Xie, D., Liu, L., Wang, H., and Chen, L. (2022). Inter-row information recognition of maize in the middle and late stages via LiDAR supplementary vision. Front. Plant Sci 13. doi: 10.3389/fpls.2022.1024360

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Zhou, Q., Gong, J., Zhu, Y., Dazeley, R., Zhao, X., et al (2025b). “Dapointr: Domain adaptive point transformer for point cloud completion,” in Proceedings of the AAAI Conference on Artificial Intelligence. 5066–5074.

Google Scholar

Li, B., Zhu, Y., Wang, Y., Lin, C.-W., Ghanem, B., and Shen, L. (2021). Anigan: Style-guided generative adversarial networks for unsupervised anime face generation. IEEE Trans. Multimedia 24, 4077–4091. doi: 10.1109/TMM.2021.3113786

Crossref Full Text | Google Scholar

Lin, T.-Y., Dollár, P., Girshick, R., He, K., Hariharan, B., and Belongie, S. (2017). “Feature pyramid networks for object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 2117–2125.

Google Scholar

Liu, M., Sheng, L., Yang, S., Shao, J., and Hu, S.-M. (2020). “Morphing and sampling network for dense point cloud completion,” in Proceedings of the AAAI conference on artificial intelligence. Palo Alto, California, USA: AAAI Press. 11596–11603.

Google Scholar

Mo, K., Zhu, S., Chang, A. X., Yi, L., Tripathi, S., Guibas, L. J., et al (2019). “Partnet: A large-scale benchmark for fine-grained and hierarchical part-level 3d object understanding,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 909–918.

Google Scholar

Pan, X., Zhan, X., Dai, B., Lin, D., Loy, C. C., and Luo, P. (2021). Exploiting deep generative prior for versatile image restoration and manipulation. IEEE Trans. Pattern Anal. Mach. Intell 44, 7474–7489. doi: 10.1109/TPAMI.2021.3115428

PubMed Abstract | Crossref Full Text | Google Scholar

Qi, C. R., Su, H., Mo, K., and Guibas, L. J. (2017a). “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 652–660. doi: 10.1109/CVPR.2017.16

Crossref Full Text | Google Scholar

Qi, C. R., Yi, L., Su, H., and Guibas, L. J. (2017b). Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst 30, 652–660. doi: 10.1109/CVPR.2017.16

Crossref Full Text | Google Scholar

Qiao, G., Zhang, Z., Niu, B., Han, S., and Yang, E. (2025). Plant stem and leaf segmentation and phenotypic parameter extraction using neural radiance fields and lightweight point cloud segmentation networks. Front. Plant Sci 16. doi: 10.3389/fpls.2025.1491170

PubMed Abstract | Crossref Full Text | Google Scholar

Sarmad, M., Lee, H. J., and Kim, Y. M. (2019). “Rl-gan-net: A reinforcement learning agent controlled gan network for real-time point cloud shape completion,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA, USA: IEEE Computer Society. 5898–5907.

Google Scholar

Tang, F., Zhang, S., Zhu, B., and Sun, J. (2024). Outdoor large-scene 3D point cloud reconstruction based on transformer. Front. Phys 12. doi: 10.3389/fphy.2024.1474797

Crossref Full Text | Google Scholar

Tchapmi, L. P., Kosaraju, V., Rezatofighi, H., Reid, I., and Savarese, S. (2019). “Topnet: Structural point cloud decoder,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA, USA: IEEE Computer Society. 383–392.

Google Scholar

Wang, X., Ang, M. H., Jr., and Lee, G. H. (2020). “Cascaded refinement network for point cloud completion,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. Los Alamitos, CA, USA: IEEE Computer Society. 790–799.

Google Scholar

Wang, N., Cao, Y., Zhang, K., Xu, S., Dong, R., Kang, F., et al (2025a). BranchMatch: point cloud registration for individual apple trees with limited overlap based on local structure characteristics. Front. Plant Sci 16. doi: 10.3389/fpls.2025.1616611

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, M., Dou, H., Sun, H., Zhai, C., Zhang, Y., and Yuan, F. (2023). Calculation method of canopy dynamic meshing division volumes for precision pesticide application in orchards based on liDAR. Agronomy 13, 1077. doi: 10.3390/agronomy13041077

Crossref Full Text | Google Scholar

Wang, G., Huang, D., Zhou, D., Liu, H., Qu, M., and Ma, Z. (2022). Maize (Zea mays L.) seedling detection based on the fusion of a modified deep learning model and a novel Lidar points projecting strategy. Int. J. Agric. Biol. Eng 15, 172–180. doi: 10.25165/j.ijabe.20221505.7830

Crossref Full Text | Google Scholar

Wang, Z., Zhang, H., Bian, L., Zhou, L., and Ge, Y. (2025b). Accurate plant 3D reconstruction and phenotypic traits extraction via stereo imaging and multi-view point cloud alignment. Front. Plant Sci 16. doi: 10.3389/fpls.2025.1642388

PubMed Abstract | Crossref Full Text | Google Scholar

Wei, G., Feng, Y., Ma, L., Wang, C., Zhou, Y., and Li, C. (2025). “Pcdreamer: Point cloud completion through multi-view diffusion priors,” in Proceedings of the Computer Vision and Pattern Recognition Conference. Los Alamitos, CA, USA: IEEE Computer Society. 27243–27253.

Google Scholar

Xie, H., Yao, H., Zhou, S., Mao, J., Zhang, S., and Sun, W. (2020). “Grnet: Gridding residual network for dense point cloud completion,” in European Conference on Computer Vision (Cham, Switzerland: Springer), 365–381.

Google Scholar

Yan, X., Lin, L., Mitra, N. J., Lischinski, D., Cohen-Or, D., and Huang, H. (2022). “Shapeformer: Transformer-based shape completion via sparse representation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. Los Alamitos, CA, USA: IEEE Computer Society. 6239–6249.

Google Scholar

Yang, B., Wen, H., Wang, S., Clark, R., Markham, A., and Trigoni, N. (2017). “3d object reconstruction from a single depth view with adversarial learning,” in Proceedings of the IEEE International Conference on Computer Vision Workshops. 679–688.

Google Scholar

Yuan, W., Khot, T., Held, D., Mertz, C., and Hebert, M. (2018). “Pcn: Point completion network,” in 2018 international conference on 3D vision (3DV) (Los Alamitos, CA, USA: IEEE), 728–737.

Google Scholar

Zeng, L., Feng, J., and He, L. (2020). Semantic segmentation of sparse 3D point cloud based on geometrical features for trellis-structured apple orchard. Biosyst. Eng 196, 46–55. doi: 10.1016/j.biosystemseng.2020.05.015

Crossref Full Text | Google Scholar

Zhang, J., Chen, X., Cai, Z., Pan, L., Zhao, H., Yi, S., et al (2021). “Unsupervised 3d shape completion through gan inversion,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Los Alamitos, CA, USA. 1768–1777.

Google Scholar

Zhang, G. and Liu, J. (2024). Diffusion-occ: 3D point cloud completion via occupancy diffusion. arXiv preprint arXiv:2408.14846. doi: 10.48550/arXiv.2408.14846

Crossref Full Text | Google Scholar

Zhang, Z., Long, J., Lai, Q., Zhu, Q., He, H., Salcedo, R., et al (2023). Static and dynamic performance evaluation of a solid-state liDAR for 3D object detection in greenhouse spray applications. J. ASABE 66, 1367–1378. doi: 10.13031/ja.15285

Crossref Full Text | Google Scholar

Zhang, W., Yan, Q., and Xiao, C. (2020). “Detail preserved point cloud completion via separated feature aggregation,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXV 16 (Springer), 512–528.

Google Scholar

Zhao, X., Li, H., Zhu, Q., Huang, M., Guo, Y., and Qin, J. (2020). Automatic sweet pepper detection based on point cloud images using subtractive clustering. Int. J. Agric. Biol. Eng 13, 154–160. doi: 10.25165/j.ijabe.20201303.5460

Crossref Full Text | Google Scholar

Zhao, J., Wang, X., Tian, H., Lu, Y., Guo, C., and Liu, H. (2021). A fertilizer discharge detection system based on point cloud data and an efficient volume conversion algorithm. Comput. Electron. Agric 185, 106131. doi: 10.1016/j.compag.2021.106131

Crossref Full Text | Google Scholar

Zhu, H., Rosetta, R., Reding, M. E., Zondag, R. H., Ranger, C. M., Canas, L., et al (2017). Validation of a laser-guided variable-rate sprayer for managing insects in ornamental nurseries. Trans. ASABE 60, 337–345. doi: 10.13031/trans.12020

Crossref Full Text | Google Scholar

Keywords: plant canopy architecture, 3D plant modeling, point cloud completion, deep learning, multi-resolution, GAN inversion

Citation: Wei Z, Long J, Zhang Z, Xue X, Sun Y, Li Q, Liu W, Shen J, Zhang Z, Li X and Ma Z (2025) Structure-aware completion of plant 3D LiDAR point clouds via a multi-resolution GAN-inversion network. Front. Plant Sci. 16:1698843. doi: 10.3389/fpls.2025.1698843

Received: 04 September 2025; Accepted: 23 October 2025;
Published: 19 December 2025.

Edited by:

Ning Yang, Jiangsu University, China

Reviewed by:

Yatian Pang, National University of Singapore, Singapore
Tong Zheng, Beijing Technology and Business University, China

Copyright © 2025 Wei, Long, Zhang, Xue, Sun, Li, Liu, Shen, Zhang, Li and Ma. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xinyu Xue, NzM1MTc4MzEyQHFxLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.