Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 29 January 2026

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1705564

This article is part of the Research TopicAI-Driven Plant Intelligence: Bridging Multimodal Sensing, Adaptive Learning, and Ecological Sustainability in Precision Plant ProtectionView all 12 articles

PointNeXt-DBSCAN: a hybrid point cloud deep learning framework for multi-stage cotton leaf instance segmentation

Zeyu Lei,&#x;Zeyu Lei1,2†Debin Zeng,&#x;Debin Zeng1,3†Liangfang Zheng,*Liangfang Zheng1,3*
  • 1College of Information Engineering, Tarim University, Alaer, China
  • 2College of Big Data and Information Engineering, Xinjiang University of Technology, Hotan, China
  • 3Key Laboratory of Tarim Oasis Agriculture, Ministry of Education, Tarim University, Alaer, China

This study addresses the challenge of organ-level instance segmentation in cotton point clouds, which arises from significant morphological variations and leaf occlusion across growth stages. To achieve high-precision leaf extraction, a hybrid framework integrating PointNeXt and DBSCAN is proposed. A dataset containing 1,065 cotton plants from seedling to boll-opening stages was constructed via multi-view image reconstruction and augmented through random rotation and scaling. Methodologically, a two-stage pipeline was designed: semantic segmentation was first performed using the PointNeXt network, where its residual MLP blocks enhanced edge and local feature learning; instance segmentation was then conducted by applying density-adaptive DBSCAN clustering to the semantic results, effectively mitigating over-segmentation in emerging leaves. Experimental results indicate that the semantic segmentation achieved an mIoU of 0.9846, representing a 7.2% improvement over PointNet++. The subsequent instance segmentation attained an ARI of 0.983, reduced the over-segmentation rate by 63%, and maintained an error below 3% for leaves smaller than 5 cm2. The framework provides reliable technical support for the automated extraction of key phenotypic traits such as leaf area index and leaf inclination distribution.

1 Introduction

Cotton leaf phenotype, which refers to the physical, physiological, and biochemical traits exhibited by cotton leaves during their growth and development, serves as a critical indicator for evaluating cotton health and photosynthetic productivity. Precise leaf phenotypic analysis is of significant importance for optimizing cotton cultivation strategies and enhancing crop management. However, traditional two-dimensional image analysis methods present substantial limitations in processing plant leaves, particularly in addressing overlapping and occlusion challenges specific to cotton leaves. These limitations reduce the accuracy of analytical results, thereby compromising the quality and credibility of plant phenotypic data. To overcome these challenges, three-dimensional point cloud technology has been introduced into agricultural research (Ding et al., 2023; Sohail et al., 2024; Yu et al., 2024; Zhu et al., 2024). This technology utilizes three-dimensional spatial data to effectively resolve inter-leaf overlapping issues, enabling more accurate segmentation of plant organs. Recent studies have further advanced instance segmentation techniques for crops like rice and wheat, demonstrating the potential of deep learning-based approaches in complex agricultural environments (Wang et al., 2024; Chen et al., 2025).

The application of high-precision 3D point cloud technology in agriculture extends beyond basic organ segmentation to enable the quantitative extraction of key phenotypic traits that are crucial for breeding and precision farming (Lahoud et al., 2019; Han et al., 2020; Xuelei et al., 2024; Yu and Qiujie, 2024; Zhenyu et al., 2024). For instance, accurately measuring leaf area index (LAI), leaf inclination distribution, and individual leaf growth dynamics in 3D space provides invaluable insights into plant photosynthetic efficiency, lodging resistance, and overall crop vigor. Scaling this technology to field conditions is feasible through the integration of portable Light Detection and Ranging (LiDAR) scanners or Unmanned Aerial Vehicle (UAV)-based photogrammetry systems, which can rapidly capture 3D data across large plots (Sai and Pande, 2024; Xiao et al., 2024). This aligns with the emerging trend of high-throughput field phenotyping, supported by robust point cloud processing frameworks for unstructured environments (Li et al., 2025) and precise positioning systems such as LiDAR-inertial-ultrasonic simultaneous localization and mapping (SLAM) (Gong et al., 2024). Previous studies have highlighted the limitations of 2D imaging in such scenarios; for example, attempts to measure leaf area in dense canopies using 2D images often fail due to severe leaf occlusion and perspective distortion, leading to significant underestimation (Vinodkumar et al., 2024; Yasir and Ahn, 2024). In contrast, the 3D point cloud approach proposed in this study directly captures the spatial architecture of plants, inherently avoiding perspective errors and mitigating occlusion issues through multi-view reconstruction and advanced quality enhancement techniques (Wei et al., 2025). This allows for a more faithful representation of the true leaf area and spatial arrangement, thereby providing a robust foundation for reliable phenotypic analysis under real-world field conditions.

Current mainstream point cloud instance segmentation algorithms adopt semantic fusion approaches that integrate semantic segmentation into basic point cloud instance segmentation frameworks. By partitioning complex spatial point clouds into semantically homogeneous subspaces, this method reduces processing complexity. Within these subspaces, neural networks are tasked with handling lower-complexity operations, thereby improving processing efficiency and precision. This approach is particularly crucial for enhancing segmentation accuracy in cotton leaf point cloud data, especially during the instance segmentation of overlapping and adjacent leaves, where clustering methods can effectively differentiate and identify individual leaf instances. Recent advances in cross-modal alignment and fusion for 3D semantic occupancy prediction (Xu et al., 2025), along with specialized architectures for handling class imbalance and limited data (VSegNet–A variant segNet for improving segmentation accuracy in medical images with class imbalance and limited data), have shown promising results in improving segmentation precision through optimized feature integration.

Yan Yu et al (Xiao et al., 2024). proposed a real-time segmentation method for street tree target point clouds using Mobile Laser Scanning (MLS). This method processes MLS data via a First-In-First-Out (FIFO) buffer and optimizes segmentation performance by combining image instance segmentation with the K-Nearest Neighbors (KNN) algorithm. Ma Xuelei et al (Sai and Pande, 2024) developed a 3D point cloud feature point detection method based on hierarchical clustering. The algorithm adjusts normal vectors using minimum spanning trees and depth-first traversal, followed by feature point identification through hierarchical clustering. Their study demonstrated that this algorithm achieves effective feature point detection across diverse point cloud models, outperforming conventional methods. Wang Zhenyu et al (Yasir and Ahn, 2024). introduced a loop closure detection algorithm based on clustered edge points in 3D point clouds. The algorithm generates descriptors via clustering and stores them in a bag-of-words model, enabling precise location recognition and six-degree-of-freedom pose correction through hash tables and inverted indices. Tests on the Multi-Modal and Multi-Robot Ground Dataset (M2DGR) and Karlsruhe Institute of Technology and Toyota Technological Institute (KITTI) datasets, as well as real-world environments, confirmed its high accuracy and rotational invariance. Huang He et al (Vinodkumar et al., 2024). proposed a fuzzy C-means clustering (FCM) point cloud simplification method that integrates the Pelican Optimization Algorithm (DEAMPOA) with weighted entropy. This approach enhances clustering efficiency and data simplification accuracy through algorithmic optimization. Evaluations on University of California, Irvine (UCI) Machine Learning Repository and KITTI datasets revealed superior clustering performance, simplified point cloud quality, accelerated processing speed, and reduced errors compared to traditional methods. Wu Jianqing et al (Singh and Yadav, 2024). designed a Real-Time Point Cloud Clustering algorithm for Road LiDAR (RTPCC-RL), comprising three modules: online capture, background filtering, and voxel feature clustering, which enables rapid point cloud processing. Their C++ implementation significantly reduces data processing complexity and optimizes point cloud localization through voxel grid partitioning. Testing demonstrated that the algorithm achieves high-precision processing of 32-channel LiDAR data within 100 milliseconds, substantially improving clustering efficiency and precision. These foundational studies are complemented by recent advances in target detection for complex scenes (Wang et al., 2025), precise 3D circle identification in point clouds (Deng et al., 2025), and drone-based surveillance using neuro-fuzzy classifiers (Abbas et al., 2025), which further refine point cloud processing in challenging environments.

In recent years, advances in deep learning technologies have established plant point cloud organ segmentation as a viable frontier research direction (Arthur, 2007; Girardeau-Montaut, 2016; Han Dong et al., 2018; Rakotosaona et al., 2020). Current 3D point cloud segmentation methods are primarily categorized into three groups: multi-view-based approaches (Bono et al., 2024; Shen et al., 2024; Zhao et al., 2024), voxelization-based methods (Zhang et al., 2023), and point-based techniques (Li et al., 2023; Patel et al., 2023; Sun et al., 2023). These methods have been extended to tasks such as 3D reconstruction of transparent objects under limited constraints (Sha et al., 2025b), while weakly supervised learning approaches leveraging pixel-level noise mining offer potential for reducing annotation dependency (Pixel-level noise mining for weakly supervised salient object detection. IEEE transactions on neural networks and learning systems, 2025).

Despite these advancements, accurately segmenting individual leaves in cotton plants across multiple growth stages remains a significant challenge due to severe leaf occlusion and substantial morphological variations. To address this gap, this study aims to develop a robust and high-precision framework for cotton leaf instance segmentation. The primary objectives are: (1) To construct a comprehensive 3D point cloud dataset of cotton plants spanning key growth stages (from seedling to boll-opening), incorporating data augmentation to enhance model generalizability; (2) To propose a hybrid two-stage segmentation pipeline that leverages an improved deep learning network (PointNeXt) for precise semantic segmentation of leaves, followed by a density-based clustering algorithm (DBSCAN) for effective instance discrimination; (3) To rigorously evaluate the performance of the proposed framework against other state-of-the-art methods in both semantic and instance segmentation tasks, demonstrating its superiority in handling complex plant architectures; and (4) To investigate the parameter sensitivity of the clustering module to establish optimal settings for cotton leaf data, ensuring robust performance across different plant densities and growth conditions. This work seeks to provide a reliable technical solution for automated phenotypic trait extraction, facilitating advanced crop management and breeding programs.

2 Materials and methods

This study utilized cotton plants from a state-owned farm in Alar City as experimental subjects. The cotton plants were categorized into four growth stages: vegetative stage (with 2–4 leaves per plant), squaring stage (5–12 leaves), boll-forming stage (13–24 leaves), and opening stage (>24 leaves). Leaves exhibited variations in vertical positioning along stems and unequal sizes across developmental phases.

The complex vertical arrangement and size variation of leaves across these growth stages necessitate precise and consistent imaging conditions for reliable 3D reconstruction and subsequent analysis. Therefore, this study employs controlled, constant artificial lighting during image acquisition. This deliberate design establishes standardized visual conditions to isolate and validate the core geometric segmentation performance of the proposed framework, independent of variable lighting effects. Addressing the primary challenge of complex plant morphology and severe occlusion first under these controlled settings provides a clear performance benchmark, with the explicit research progression to then test model robustness under natural illumination in subsequent field applications.

To acquire high-precision 3D point clouds for segmentation, this study employed a multi-view photogrammetry approach based on a DSLR camera rather than direct LiDAR scanning. This choice was motivated by three principal considerations: (1) Cost-effectiveness and accessibility, making the pipeline more viable for broader research and potential agricultural applications; (2) Rich data capture, as RGB images retain detailed texture and color information beneficial for future phenotypic analysis beyond geometry; and (3) Research focus, which is on developing and validating the segmentation algorithm itself, rather than comparing sensor technologies. A dedicated image acquisition platform was therefore constructed to ensure systematic and consistent data collection. Photographic capture was performed using a digital single-lens reflex (DSLR) camera (Canon EOS 200D II) equipped with a 50mm macro lens. The camera was set to automatic mode, allowing it to automatically determine the aperture, shutter speed, and ISO sensitivity for optimal exposure in the given lighting conditions. The imaging was conducted under controlled artificial lighting conditions to ensure consistency. A uniform illumination setup was achieved using two 100W LED panel lights (color temperature: 5500 K, illuminance: approximately 1200 lux at the plant canopy level) positioned at 45-degree angles relative to the plant’s central axis to minimize shadows and specular reflections. The lighting intensity was calibrated using a digital lux meter prior to each acquisition session to maintain stable illumination across all samples.

The camera was mounted on a tripod, and the 0-degree reference position was defined as the camera lens being perpendicular to and directly facing the central stem of the cotton plant at a height level with the plant’s canopy. This frontal view established the primary axis for angular measurements.

Image acquisition for the cotton plant was performed from three distinct angular configurations (30°, 60°, and 90° relative to the plant's central axis). A goniometer fixed to the tripod head ensured precise angular positioning. Examples from different perspectives are shown in Figure 1. For the 30° and 60° setups, the camera maintained a fixed radial distance from the plant base and was moved along a circular path to capture 60-80 images, ensuring comprehensive circumferential coverage. The top-down (90°) view was crucial for capturing the planform structure of the canopy. The resulting 3D point cloud, generated using RealityCapture software, is presented in Figure 2a. The model of the cotton plant is visualized within a transparent cubic bounding box against a black background. The numerous white points surrounding the plant in the visualization represent the precise camera stations, clearly illustrating the consistent fixed distance and circular trajectory used during data acquisition. The apparent close proximity of the plant to the bounding box is a perspective effect of this 3D visualization environment and does not reflect the actual camera distance.All images were acquired with a uniform resolution of 6000 × 4000 pixels in JPG format.

Figure 1
Three images, labeled (a), (b), and (c), the images show a cotton plant from different angles. The plant is placed on a black background.

Figure 1. Multi-angle sample images of data acquisition; (a) Image acquired at 30° angular configuration relative to the plant canopy; (b) Image acquired at 60° angular configuration relative to the plant canopy; (c) Image acquired at 90° angular configuration relative to the plant canopy.

Figure 2
A series of five images illustrating a 3D plant modeling process. (a) Display a plant model surrounded by shots taken from different angles. (b) Displays a wireframe model of the plant. (c) Shows a segmented model with red and blue coloring. (d) Depicts isolated plant parts in gray. (e) Displays colored segments of plant parts, each in a different color, over a blue gradient background.

Figure 2. 3D Reconstruction of cotton plant and leaf annotation diagram (a) Multi-angle image alignment and 3D reconstruction with acquisition angles; (b) Raw point cloud data of unlabeled cotton plant structure; (c) Semantic-labeled point cloud with leaf/stem classification; (d) Unannotated leaf cluster point cloud prior to segmentation; (e) Instance-segmented leaves with individual identity labels.

2.1 Point cloud preprocessing

The multi-angle 2D images were imported into the 3D modeling software RealityCapture. Three-dimensional reconstruction was performed using the Structure from Motion (SFM) algorithm (Zhao et al., 2023) and Cluster Multi-view Stereo (CMVS) algorithm (Ni et al., 2023), as illustrated in Figure 2a, which also displays the acquisition angles and relative positions of each cotton plant image. To enhance visualization quality, the 3D models underwent completion and colorization processes to improve their realism and detail representation. Subsequently, non-plant-related point cloud data were removed to ensure the accuracy and efficiency of subsequent analyses.

The resulting dataset comprised 1, 065 cotton plant point cloud samples spanning four distinct growth stages, each characterized by specific morphological traits that influence segmentation complexity. At the vegetative stage (285 samples; 2–4 leaves per plant), leaves were typically small (5–15 cm² area) and predominantly distributed in the upper canopy, presenting minimal occlusion challenges but requiring precise segmentation of emerging structures. During the squaring stage (320 samples; 5–12 leaves), plants exhibited medium-sized leaves (15–35 cm²) with increased vertical distribution along the main stem, where moderate occlusion began appearing particularly in middle canopy layers. The boll-forming stage (280 samples; 13–24 leaves) was characterized by large leaves (30–60 cm²) and complex architecture with significant vertical span (40–80 cm height), where severe occlusion in lower canopy regions presented the greatest segmentation challenges. Finally, the opening stage (180 samples; >24 leaves) featured very large leaves (45–80 cm²) with extensive overlapping and curling edges, forming a highly complex canopy structure with multiple branching patterns and self-occlusion. This progressive increase in morphological complexity across stages directly influenced the difficulty of segmentation tasks.

This distribution ensured comprehensive coverage of morphological variations, with leaf size variability and vertical positioning complexity increasing progressively across stages. These characteristics directly influence segmentation difficulty, particularly regarding occlusion handling and instance discrimination accuracy.

Due to environmental factors and equipment limitations, the point cloud data inevitably contained noise. Point cloud noise is primarily categorized into two types: outliers and noise points (Comba et al., 2018). Outliers refer to isolated points distant from other points, potentially caused by acquisition errors or foreign objects. Noise points denote irregularly distributed points with significant local deviations, likely resulting from environmental interference or data acquisition errors. Noise predominantly clusters near leaf surfaces, primarily caused by leaf surface reflections and inter-leaf occlusion-induced feature matching errors.

To eliminate noise and smooth density irregularities in the point cloud data, statistical filtering was applied to the downsampled colored leaf point clouds, identifying and removing noise points through statistical methods. Additionally, voxel filtering was employed to downsample the plant point cloud data, converting the point cloud into voxel data and removing redundant voxels, thereby reducing the original point cloud density.

Normalization was subsequently applied to standardize the spatial coordinates of the point cloud data, addressing scale variations arising from differences in plant size and scanning distance. The normalization process utilized min-max scaling, which linearly transforms the coordinate values of each point cloud sample to a predefined range of [–1, 1]. This technique was implemented using custom Python scripts leveraging the NumPy library for efficient numerical computations.

The primary rationale for applying normalization was to mitigate the influence of absolute scale variations on subsequent neural network training and geometric processing algorithms. By transforming all point clouds into a consistent coordinate range, the process enhances model convergence stability and reduces algorithmic sensitivity to spatial disparities, thereby improving the generalizability of the segmentation model across growth stages with divergent morphological scales.

Finally, manual annotation of the point cloud data was performed using the segmentation tools in the 3D point cloud processing software CloudCompare (Zhao et al., 2021). Semantic labeling classified each point as either Leaf or Stem (Figures 2b, c), while instance labeling distinguished individual leaves as separate instances (Figures 2d, e) to facilitate subsequent segmentation and analysis. To ensure annotation quality and consistency, we implemented a rigorous quality control protocol involving dual independent annotation by two trained experts. The annotators first independently labeled a randomly selected subset of 100 point clouds using identical guidelines, achieving substantial agreement with a Cohen’s Kappa coefficient of 0.73. All discrepancies were resolved through consensus discussions with a third senior researcher before the remaining samples were divided between annotators for completion. Through these procedures, a rigorously annotated dataset containing 1, 065 individual plant samples was established with comprehensive coverage of growth stages and morphological variations. To facilitate model training and evaluation, the complete dataset was divided into training, validation, and test sets using a 7:2:1 ratio. This partition was conducted with two primary considerations. First, to avoid data leakage and ensure that model performance reflects generalization to unseen plants, all point cloud data originating from the same individual plant were kept entirely within one subset. Second, a stratified sampling strategy was employed to maintain a balanced distribution of the four growth stages across all three subsets, promoting unbiased learning and reliable evaluation (Arthur, 2007; Girardeau-Montaut, 2016; Schonberger and Frahm, 2016; Han Dong et al., 2018; Rakotosaona et al., 2020; Wu et al., 2024). Following the partition, standard data augmentation techniques—including rotation, scaling, and Gaussian noise addition—were applied to the training set to increase its variability and enhance the model’s robustness.

2.2 Research methodology

The research workflow is illustrated in Figure 3. Following the completion of multi-angle image acquisition for cotton plants, the collected plant data underwent 3D reconstruction and preprocessing. The preprocessed cotton plant point clouds were then annotated to establish a dedicated point cloud dataset. Subsequently, neural network models were applied to perform semantic segmentation on the preprocessed point clouds, aiming to isolate leaf point clouds from the plant structure. The model demonstrating optimal segmentation performance was selected as the core algorithm for the semantic segmentation module (Zhang et al., 2020; Zhao and Tao, 2020; Zhou et al., 2020; Huang et al., 2021). The leaf point clouds obtained from semantic segmentation were further subjected to clustering experiments to achieve instance segmentation of cotton leaves. Multiple clustering algorithms were evaluated, with the highest-performing method being designated as the primary algorithm for the instance segmentation module (Huang et al., 2024). Finally, parameter sensitivity analysis was conducted to evaluate the selected clustering algorithm’s segmentation effectiveness under different parameter configurations when applied to the cotton leaf point cloud dataset.

Figure 3
Flowchart detailing a process for cotton model analysis. It includes Data Collection: 2D image acquisition and 3D reconstruction; Data Processing: preprocessing, downsampling, normalization, normal vector reconstruction, annotation; Semantic Segmentation Algorithm: selecting from PointNet, PointNet++, DGCNN, SceneEncoder, PointNeXt; Cluster-Based Instance Segmentation Algorithm: selecting clustering algorithms like MeanShift, K-means, K-means++, Mini Batch K-Means, DBSCAN, with parameter optimization.

Figure 3. Research workflow diagram.

2.3 Network architecture

Current mainstream point cloud instance segmentation algorithms adopt semantic fusion strategies that integrate semantic segmentation into basic instance segmentation frameworks. By dividing complex spatial point clouds into semantically consistent subspaces, this approach reduces processing complexity (Wang et al., 2018; Wang et al., 2019; Hu et al., 2020; Fang et al., 2021; Qiu et al., 2022; Zhao et al., 2022; Guo et al., 2023; Ren et al., 2024). Within these subspaces, network models only need to handle lower-complexity tasks, thereby improving processing efficiency and accuracy. This is particularly critical for enhancing segmentation precision in cotton leaf point cloud data, especially during instance segmentation of overlapping and adjacent leaves where clustering methods can effectively distinguish and identify individual leaf instances.

The selection of PointNeXt for semantic segmentation and DBSCAN for instance clustering in this study is based on a thorough consideration of the unique characteristics of cotton leaf point clouds (Rakotosaona et al., 2020; Qian et al., 2022). The rationale for this design is twofold.

Firstly, PointNeXt was chosen over other semantic segmentation networks due to its superior capability in handling the intricate geometry of leaves.Compared to voxel-based or multi-view methods, point-based networks (such as PointNet++ (Qi et al., 2017a; Qi et al., 2017b)) maximize the preservation of original point cloud geometry and avoid quantization errors, which is crucial for accurately segmenting thin leaf edges. However, the classic PointNet++ has limitations in complex local feature extraction. PointNeXt addresses this by introducing inverted residual MLP modules and enhanced residual connections, significantly boosting its ability to model local geometric structures. This improvement allows PointNeXt to more effectively learn the subtle features of cotton leaves, such as complex vein textures and curled edges, leading to more precise semantic segmentation at the junctions between leaves and stems, and between adjacent leaves. Furthermore, compared to graph convolutional networks like DGCNN (Phan et al., 2018), PointNeXt achieves superior convergence and stability through architectural optimization while maintaining computational efficiency (as shown in Figure 4), a critical advantage when processing large, whole-plant point cloud models.

Figure 4
Line graph depicting mIoU values over epochs for five models: PointNet, PointNet++, SceneEncoder, DGCNN, and PointNeXt. The x-axis shows epochs from 0 to 200, and the y-axis shows mIoU values from 0.4 to 1.0. PointNeXt has the highest performance throughout, reaching above 0.9, while the others stabilize around 0.8.

Figure 4. Variation of mIoU during training process.

Secondly, DBSCAN was selected for instance clustering because its fundamental principles align perfectly with the challenges of segmenting individual cotton leaves.Traditional clustering algorithms require pre-defining the number of clusters (K), which is impractical for cotton plants where the leaf count varies dynamically across growth stages and is often unknown due to occlusion. DBSCAN inherently overcomes this limitation by automatically determining the number of instances based on the data’s density distribution, requiring no pre-set cluster count. Moreover, cotton leaves are irregularly shaped and vary greatly in size, contrary to the spherical clusters assumed by many algorithms. DBSCAN’s core strength lies in its ability to discover clusters of arbitrary shapes, making it ideally suited for segmenting natural leaf structures. An additional benefit is its robustness to noise; points in low-density regions can be identified as noise, enhancing the algorithm’s reliability in complex plant environments. In contrast, as evidenced by the results in Section 3.2, algorithms like K-means produce unreasonable fragmentation when applied to this data.

Therefore, the combination of PointNeXt and DBSCAN forms a complementary hybrid framework: the former is responsible for achieving high-precision “pixel-level” semantic recognition, separating leaves from the complex background; the latter then performs “instance-level” object discrimination based on the semantic results using its density-adaptive characteristics. This design directly addresses the core challenges of the cotton leaf segmentation task.

This paper proposes a PointNeXt-based semantic segmentation method and a DBSCAN-based point cloud clustering method for cotton leaf instance segmentation. PointNeXt, an improved version of PointNet++, enhances model performance through architectural optimization. It employs an encoder-decoder architecture with skip connections, primarily composed of three modules: Set Abstraction (SA) module, Feature Propagation (FP) module, and Inverted Residual MLP (InvResMLP) module. The PointNeXt structure is shown in Figure 5.

Figure 5
Diagram illustrating a neural network pipeline for processing point cloud data of a plant. The process begins with an MLP layer, followed by stages of set abstraction and Inverted Residual MLP layers, reducing dimensions from N by 32 to N divided by 256 by 512. Three phases of feature propagation increase dimensions back to N by 32. Intermediate processes like subsampling, grouping, reduction, and concatenation are depicted, with transformation details below each step. The image transitions from a black-and-white to a colorized point cloud representation of the plant.

Figure 5. PointNeXt network architecture diagram.

In the encoder section, multiple SA modules and InvResMLP modules progressively reduce point cloud resolution while extracting high-dimensional features. Each SA module contains a subsampling layer, grouping layer, and shared MLP layer responsible for local feature extraction and aggregation. The InvResMLP module following each SA module incorporates residual connections, depthwise-separated MLP, and inverted bottleneck designs to further enhance feature extraction capability and alleviate gradient vanishing issues.

The decoder consists of multiple FP modules that gradually restore point cloud resolution by fusing multi-level features. Each FP module contains an interpolation layer and MLP layer, integrating features from the encoder through skip connections to output high-dimensional features. The decoder’s high-dimensional features are processed through an MLP layer and divided into two branches: the semantic prediction branch outputs per-point semantic category predictions, while the instance embedding branch outputs per-point instance embedding vectors to facilitate subsequent instance segmentation tasks. PointNeXt achieves model width and depth scaling by adjusting channel number C in Stem MLP and quantity B of InvResMLP modules.

On the other hand, this study employs the DBSCAN algorithm for instance segmentation clustering. DBSCAN is a density-based unsupervised clustering algorithm that exhibits multiple significant advantages in point cloud clustering and instance segmentation.

For point cloud clustering, DBSCAN eliminates the need for predefining cluster numbers, enabling automatic partitioning of distinct clusters based on point cloud density distribution. This characteristic makes DBSCAN particularly effective when handling dynamic and variable datasets. Additionally, DBSCAN can effectively identify and exclude noise points and outliers from clustering results, thereby significantly enhancing the robustness of clustering outcomes. The algorithm also excels at recognizing arbitrarily shaped clusters without constraints on their geometry or size, demonstrating superior performance when processing complex point cloud data.

However, DBSCAN’s performance exhibits sensitivity to parameter selection, requiring adjustment based on specific datasets. For instance, the choice of neighborhood radius (eps) and minimum point threshold (minPts) directly influences clustering results. For any point x in the dataset, its ϵ-neighborhood Nϵ(x) can be expressed as Equation 1:

Nϵ(x)={yD|dist(x,y)ϵ}(1)

where D represents the dataset, and dist(x, y) denotes the distance between points x and y.

In point cloud instance segmentation, DBSCAN’s density-based clustering method partitions point clouds into distinct instances based on density distribution, achieving precise instance segmentation. DBSCAN also excels at handling overlapping instances by identifying them as separate clusters, thereby enhancing segmentation accuracy.

While DBSCAN offers significant advantages in point cloud clustering and instance segmentation—including automated processing, noise robustness, and adaptability to complex geometries—it also presents limitations. By integrating with PointNeXt’s feature extraction capabilities, DBSCAN enables more efficient point cloud clustering while maintaining segmentation precision, providing robust support for cotton leaf instance segmentation. This synergistic combination not only improves segmentation accuracy for cotton leaves but also enhances the model’s adaptability to environmental variations and biodiversity.

2.4 Parameter settings

All experiments in this study were conducted on a 64-bit server equipped with an AMD RYZEN 7 3700X 8-Core Processor CPU (base frequency: 4.20 GHz) and an NVIDIA GeForce RTX 2080TI GPU. The server operated on Ubuntu 22.04.4 LTS, with CUDA Toolkit 11.3 as the computing platform and PyTorch 1.10.1 as the deep learning framework. During training, the batch size was uniformly set to 16, with an initial learning rate of 0.001. The network model underwent 200 training epochs, with the learning rate reduced by 50% every 20 epochs. The optimization process employed momentum-based stochastic gradient descent (SGD) with a momentum value of 0.9 and a weight decay coefficient of 0.0005. For the DBSCAN algorithm, the neighborhood radius parameter (ϵ) was set to 0.04, and the minimum sample parameter (min_samples) was configured as 10.

2.5 Evaluation metrics

The methodology in this study comprises two independent branches performing distinct tasks: predicting leaf point clouds of cotton plants and conducting instance prediction on leaf point clouds. The proposed method thus requires two distinct sets of evaluation metrics (De De Vargas and Bedregal, 2013).

For the semantic segmentation task, the performance of the plant segmentation network was evaluated at the point level. Commonly used metrics for semantic categories were employed: Overall Accuracy (OA) and mean Intersection over Union (mIoU).

For the instance segmentation task, clustering metrics for point clouds were adopted: Adjusted Rand Index (ARI), Adjusted Mutual Information (AMI), Normalized Mutual Information (NMI), and Fowlkes-Mallows Index (FMI) served as evaluation criteria[70-71], Its computational formula is given in Equation 2.

ARI=RIExpected_RImax(RI_range)Expected_RI(2)

The Adjusted Mutual Information (AMI) ranges from 0 to 1. Its physical meaning is the amount of shared information between the clustering result and the ground truth classification, corrected for chance.It is based on information theory principles, quantifying how much knowing one cluster label reduces uncertainty about the other. Higher AMI values indicate greater alignment between clustering outcomes and ground truth categories, signifying improved clustering performance. Its computational formula is given in Equation 3 (Amelio and Pizzuti, 2015).

AMI=MIExpected_MIH(U)+H(V)2Exprcted_MI(3)

The Normalized Mutual Information (NMI) ranges from 0 to 1. Its physical meaning is a normalized measure of the mutual dependence between the clustering result and the ground truth, representing the ratio of the mutual information to the square root of the product of their individual entropies.Higher NMI values indicate better clustering performance by reflecting a stronger statistical relationship between the obtained clusters and the true labels. Its computational formula is given in Equation 4 (Amelio and Pizzuti, 2015).

NMI=2MIH(U)+H(V)(4)

The FMI ranges from 0 to 1, measuring similarity between clustering results and ground truth labels. Values closer to 1 indicate higher consistency between clustering outcomes and true classifications. Its computational formula is given in Equation 5 (Guo et al., 2019; Tongtong et al., 2023).

FMI=TPTP+FP*TPTP+FN(5)

3 Experimental results and analysis

3.1 Comparative analysis of semantic segmentation results for cotton point clouds

This study evaluated the performance of the PointNeXt model on semantic segmentation tasks using the constructed cotton plant point cloud dataset. We conducted comparative analyses between PointNeXt and several state-of-the-art models, including PointNet, PointNet++, DGCNN, and SceneEncoder (Xu et al., 2020). The performance metrics, along with statistical significance tests, are summarized in Table 1.

Table 1
www.frontiersin.org

Table 1. Performance comparison of semantic segmentation for different deep learning networks.

The results demonstrate that PointNeXt’s performance advantage is statistically significant. Our model achieved superior mIoU and OA scores of 0.954 and 0.985, respectively. As shown in Table 1, PointNeXt significantly outperformed all baseline models in semantic segmentation (all p-values< 0.05). The 95% confidence intervals for the differences in mIoU consistently exclude zero, providing strong evidence that the observed performance improvements are robust. Figure 6 presents visualization results across different growth stages, showing that although minor missegmentation persists at stem-leaf junctions, nearly all leaf point clouds are accurately segmented, confirming the model’s effectiveness in complex canopy environments.

Figure 6
Comparison of plant branch 3D reconstructions. Each row illustrates plants labeled as Ground Truth, PointNet, and PointNeXt. Ground Truth shows the original structure, while PointNet and PointNeXt present the segmentation results of different models. The legend section highlights the subtle differences at the branch connections. The four examples (a, b, c, d) demonstrate the differences in details and accuracy of samples at different growth stages.

Figure 6. Visualization examples of language segmentation for cotton plant point clouds. (a) Vegetative stage: Original data vs PointNet/PointNet++ on simple structure; (b) Squaring stage: Raw point cloud vs models during bud formation; (c) Boll-forming stage: Ground truth vs segmentation with dense bolls; (d) Opening stage: Original vs model results on mature cotton clusters.

Based on the comparative experimental data from Figure 4, the PointNext algorithm demonstrates three core advantages in point cloud segmentation tasks. The model exhibits significant convergence superiority during early training stages, achieving an mIoU value of 0.728 at 25 epochs — 25.9% and 5.6% faster than DGCNN (0.578) and PointNet++ (0.772), respectively — and breaks the 0.8 performance threshold first at 50 epochs with an mIoU of 0.813. As training progresses, PointNext’s mIoU curve shows a smooth monotonic upward trend, ultimately reaching a peak performance of 0.959, significantly surpassing DGCNN (0.865) and PointNet++ (0.892), while maintaining an mIoU standard deviation of only 0.003 throughout training — an 89.3% reduction compared to DGCNN (0.028). This performance enhancement stems from its innovative multi-scale residual feature propagation mechanism, which optimizes local geometric feature modeling through cross-layer gradient flow enhancement. Combined with an adaptive neighborhood sampling strategy that dynamically corrects point cloud density distribution bias, PointNext achieves dual breakthroughs in feature representation robustness and training stability. Experimental results confirm that PointNext sets new benchmarks in precision, efficiency, and generalization capabilities.

3.2 Comparative analysis of instance segmentation results using different clustering algorithms on cotton leaf point clouds

Following the completion of semantic segmentation on cotton point clouds, the leaf point cloud data were obtained. To validate DBSCAN’s superior suitability compared to other algorithms for cotton leaf instance segmentation, this study compared the clustering performance of DBSCAN (Ester et al., 1996) against MeanShift (Comaniciu and Meer, 2002), K-means (Lloyd, 1982), K-means++, and Mini Batch K-Means (Sculley, 2010). The comparative results are presented in Table 2.

Table 2
www.frontiersin.org

Table 2. Performance and time consumption comparison of clustering algorithms.

As demonstrated in the table, distinct differences in clustering performance emerge among various algorithms when applied to cotton leaf point cloud data. The DBSCAN algorithm significantly outperforms other methods across all evaluation metrics, highlighting its superior capability in clustering cotton leaf point clouds. From the cotton point cloud dataset, we visualized instance segmentation results of three test samples to compare five clustering algorithms: K-means, K-means++, Mini Batch K-Means, MeanShift, and DBSCAN. Figure 7 explicitly illustrates their segmentation effects on cotton leaf point clouds, where subfigures (a)-(e) respectively correspond to the outcomes of these five algorithms. The visualization reveals notable disparities in precision and detail processing: K-means exhibits fragmented segmentation at leaf margins (a), K-means++ shows improved consistency yet residual fragmentation (b), Mini Batch K-Means demonstrates accelerated computation at the cost of edge artifacts (c), while MeanShift displays adaptive density recognition with boundary ambiguities (d). Visual comparison confirms DBSCAN (e) achieves optimal instance segmentation, characterized by well-defined cluster boundaries, natural preservation of insect-damaged cavities, and precise delineation of leaf vein branching patterns through spatially compact yet distinct clusters.

Figure 7
A series of scientific visualizations labeled (a) to (e), each figure presents 3 data cases. The detailed parts are connected to the enlarged view within the red box via red dotted lines. Each panel highlights different parts within the cluster and demonstrates different clustering effects.

Figure 7. Visualization of clustering effectiveness for different algorithms on cotton point cloud dataset. (a) K-means: Fragmented edge segmentation in leaf margin regions; (b) K-means++: Residual fragmentation despite improved cluster consistency; (c) Mini Batch K-Means: Edge artifacts induced by accelerated computation; (d) MeanShift: Boundary ambiguity in adaptive density recognition; (e) DBSCAN: Optimal instance segmentation with natural cavity preservation and compact cluster boundaries.

3.3 Investigation of neighborhood radius parameter ϵ on cotton leaf instance segmentation

The neighborhood radius parameter ϵ in the DBSCAN algorithm critically influences clustering efficacy. This experiment investigates the clustering performance of DBSCAN on cotton leaf point clouds under varying ϵ values, aiming to understand how parameter adjustments affect clustering precision and instance segmentation quality. The optimal value of ϵ=0.04 was determined through a systematic empirical process. We performed a grid search by incrementally increasing ϵ from 0.01 to 0.07 with a step size of 0.01. The evaluation metrics (ARI, NMI, FMI) showed a consistent trend: performance improved gradually as ϵ increased from 0.01, reached an optimal plateau around ϵ=0.04, and then began to decline as ϵ increased further beyond 0.05. This pattern confirms the existence of a well-defined optimum and underscores the sensitivity of the algorithm to this parameter. The value ϵ=0.04 was selected as it yielded the best overall performance across our dataset, effectively balancing the trade-off between over-segmentation in dense regions and under-segmentation in sparse regions.

However, it is important to note that this value is dataset-specific. Its efficacy relies on the consistent point cloud density and spatial scale of the plants in our current study. Variations in plant size, growth stage, or data acquisition protocols that alter the point cloud’s inherent density distribution would necessitate a re-calibration of this parameter for optimal results. Such analysis provides fundamental guidance for parameter selection in cotton leaf point cloud processing to achieve enhanced leaf recognition and feature extraction. The instance segmentation results of DBSCAN under different ϵ configurations are detailed in Table 3:

Table 3
www.frontiersin.org

Table 3. Effectiveness comparison of DBSCAN instance segmentation with varying ϵ.

As evident from Table 3, the DBSCAN algorithm achieved optimal instance segmentation performance on cotton leaf point cloud data when the neighborhood radius parameter ϵ was set to 0.04, with all evaluation metrics approaching their peak values. As ϵ increased, the algorithm’s performance initially improved before marginally declining, demonstrating the significant impact and sensitivity of parameter selection on clustering outcomes.

Two test samples were selected from the cotton point cloud dataset to visualize segmentation results, as shown in Figures 8, 9.

Figure 8
Six panels labeled (a) to (f) display colorful 3D shapes of DBSCAN clustering under different parameters. Each panel groups several figures, with certain shapes highlighted in red bordered squares.

Figure 8. Parameter comparison for DBSCAN clustering (sparse leaf arrangement). (a) DBSCAN with the minimum neighborhood radius (ϵ=0.02); (b) DBSCAN with a reduced neighborhood radius (ϵ=0.03); (c) DBSCAN with the recommended neighborhood radius (ϵ=0.04); (d) DBSCAN with an increased neighborhood radius (ϵ=0.05); (e) DBSCAN with a high neighborhood radius (ϵ=0.06); (f) DBSCAN with the maximum neighborhood radius (ϵ=0.07).

In both figures, subfigures (a)-(f) correspond to neighborhood radius parameter settings ranging from 0.02 to 0.07. As shown in Figures 8, 9, under dense leaf arrangements, lower ϵ values yield better instance segmentation efficacy for cotton leaves. Conversely, with sparser leaf distributions, higher ϵ values achieve optimal segmentation performance. When ϵ was set to 0.04, the algorithm demonstrated superior generalizability by maintaining relatively high segmentation accuracy across both dense and sparse leaf arrangements.

Figure 9
Six panels (a-f) depict multicolored three-dimensional shapes with a specific purple shape highlighted by a red box in each panel. Red arrows indicate focus on the purple shape, with close-up insets shown above each. Panels are arranged with (a), (b), and (c) on the top row, and (d), (e), and (f) directly below. Shapes appear varied in color and form across the panels.

Figure 9. Parameter comparison for DBSCAN clustering (dense leaf arrangement). (a) DBSCAN with the smallest neighborhood radius (ϵ=0.02); (b) DBSCAN with a small neighborhood radius (ϵ=0.03); (c) DBSCAN with the optimal neighborhood radius (ϵ=0.04); (d) DBSCAN with a moderate neighborhood radius (ϵ=0.05); (e) DBSCAN with a large neighborhood radius (ϵ=0.06); (f) DBSCAN with the largest neighborhood radius (ϵ=0.07).

4 Discussion

4.1 Analysis of segmentation performance and research implications

This study successfully addresses the technical challenges of cotton plant point cloud leaf segmentation by introducing a two-stage segmentation method that integrates the PointNeXt deep learning framework with the DBSCAN density clustering algorithm. To clarify the contribution of this research, we note that similar deep learning approaches have been validated in phenotyping analyses of different crops. For instance, a recent study on apple trees also utilized PointNeXt to achieve high-precision semantic segmentation of branches and trunks (mIoU of 0.9481) (Jiang et al., 2025). That study demonstrated the effectiveness of the method in parsing the architecture of woody perennial fruit trees, while this research validates and extends this technical pathway to annual field crops, specifically targeting the more complex and variable organ structures of cotton leaves. This cross-crop comparison underscores the robustness and adaptability of point cloud deep learning in agricultural phenotyping.

When directly compared with contemporary research on cotton, our method demonstrates distinct advantages. For example, although Song et al (Jiazheng et al., 2024). employed Laplacian skeleton extraction combined with Quickshift++ for cotton plant organ segmentation, our approach achieved a higher mean intersection over union (mIoU) of 0.9846 and an accuracy of 0.9537 in the semantic segmentation task. This performance highlights the powerful capability of deep feature learning in parsing complex plant structures, particularly in distinguishing fine leaf veins and wrinkled edges—a challenge acknowledged in prior point cloud segmentation studies. For instance segmentation, the optimized DBSCAN algorithm (ϵ=0.04) exhibited excellent adaptability to non-uniform leaf distributions, achieving a high adjusted Rand index (ARI) of 0.983 and normalized mutual information (NMI) of 0.992. The end-to-end inference pipeline required only 0.88 seconds per sample, establishing a solid foundation for large-scale, high-throughput phenotyping applications and confirming its practical value for automated analysis.

4.2 Limitations and future perspectives

The primary application value of this research lies in its contribution to non-destructive, automated phenotypic analysis. Our framework enables the accurate extraction of key morphological parameters—such as leaf count, unfolding angle, and main stem height—replacing time-consuming and subjective manual measurements. This capability is crucial for precise growth monitoring and lays a foundation for future applications in stress resistance evaluation, for instance, by automating the calculation of metrics like wilted leaf area to support breeding decisions.

Despite these promising results, this study acknowledges certain limitations that present avenues for future work. First, the model was primarily trained and validated on data from controlled environments. Plant architecture in open fields is considerably more complex due to factors such as planting density, wind, pest/disease interactions, and varied lighting conditions. Therefore, the model’s performance may degrade under such unconstrained field conditions, representing a major constraint for large-scale practical deployment. Second, manual point cloud annotation remains a bottleneck for data production efficiency. Third, a critical technical limitation concerns the fixed neighborhood parameter in the DBSCAN algorithm, which caused approximately 7.1% over-segmentation in regions with extremely high point density (e.g., near the main stem). This issue arises because a static parameter cannot adapt to spatial density variations within the complex plant architecture.

To address these challenges and advance the methodology, future research will be guided by insights from several cutting-edge studies. To enhance field robustness, we will prioritize expanding the dataset to include point clouds of plants exposed to major agronomic stresses through controlled experiments and collaborative field collections. Such data would be instrumental in developing models capable of segmenting organs and quantifying stress-induced phenotypic changes. To overcome the limitation of fixed DBSCAN parameters, we will explore adaptive clustering approaches such as Hierarchical DBSCAN (HDBSCAN), which can dynamically adjust based on local density variations. Inspired by the homology constraint employed in HCMPE-Net for enforcing feature consistency in complex structures (Xiang et al., 2025), we plan to integrate constraints on leaf curvature and normal vector consistency to further mitigate over-segmentation. Furthermore, the hypergraph-based attention mechanism of the Hypergraph BiFormer (Jing et al., 2025) offers a promising path to better model the complex, non-pairwise spatial relationships between leaves in dense canopies. To specifically enhance the model’s robustness and adaptability in real-world field applications, the joint learning framework of SSC-Net (Sha et al., 2025a), which simultaneously handles segmentation and classification, provides a valuable blueprint for introducing auxiliary tasks such as stress state classification. Finally, to alleviate the burden of manual annotation and enhance data efficiency, techniques from small-sample learning models like ZHPO-LightXBoost (Sha et al., 2025c) will be explored to reduce reliance on massive labeled datasets, potentially combined with AI-assisted pre-labeling strategies.

5 Conclusions

This study developed and validated a novel two-stage segmentation method that effectively integrates the PointNeXt deep learning network and the DBSCAN clustering algorithm for cotton plant point cloud leaf segmentation. The framework demonstrated outstanding performance in both semantic and instance segmentation tasks, establishing a robust technical foundation for high-throughput phenotyping applications.

The main conclusions of this study are summarized as follows: First, the proposed method exhibits high accuracy and efficiency in segmenting cotton plant point clouds, with PointNeXt achieving an mIoU of 0.9846 in semantic segmentation and DBSCAN achieving an ARI of 0.983 in instance segmentation. Second, compared to existing methods for cotton organ segmentation, this study demonstrates superior performance in semantic segmentation, and its successful application to cotton extends the validation of point cloud deep learning from woody fruit trees to annual field crops. Third, this research provides a reliable technical basis for automating the extraction of key phenotypic traits, promoting the development of non-destructive and efficient crop phenotyping. Finally, the study also acknowledges limitations including reliance on manual annotation, over-segmentation issues, and primary validation in controlled environments.

Future research will focus on three key directions: First, Field Validation and Adaptation: Priority will be given to testing and optimizing the model on cotton plants grown in open fields to assess its robustness under real and complex agricultural conditions. Second, Advanced Synthetic Data Generation: Plans include developing a physics engine-based virtual point cloud generation system to create large-scale, accurately annotated training data that simulates field variability and various stress conditions. Third, Enhanced Segmentation Framework: Integrating geometric constraints with the deep learning framework to establish a semantics-geometry dual-driven adaptive segmentation framework, improving parsing accuracy in overlapping regions and under challenging field conditions.

This work pioneers a valuable technical pathway for 3D phenotypic analysis of cotton. The tools and insights provided can advance breeding programs and precision agriculture practices. By addressing the future directions outlined above, the potential of this method to create a tangible impact in real-world agricultural scenarios can be fully realized.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

ZL: Writing – original draft. DZ: Writing – original draft. LZ: Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported in part by the Science and Technology Key Project of the Xinjiang Production and Construction Corps (2023AB063) Development and application demonstration of green technology for online monitoring of fresh fruits and cold chain logistics in Xinjiang.

Conflict of interest

The authors declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbas, Y., Alarfaj, A. A., Alabdulqader, E. A., Algarni, A., Jalal, A., Liu, H., et al. (2025). Drone-based public surveillance using 3D point clouds and neuro-fuzzy classifier. Comput. Mater. Contin. 82. doi: 10.32604/cmc.2025.059224

Crossref Full Text | Google Scholar

Amelio, A. and Pizzuti, C. (2015). “Is normalized mutual information a fair measure for comparing community detection methods?,” in Proceedings of the 2015 IEEE/ACM international conference on advances in social networks analysis and mining, Vol. 2015. 1584–1585.

Google Scholar

Arthur, D. (2007). “Vassilvitskii S.K-means++: the advantages of careful seeding,” in Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, Louisiana, USA, January 7-9 (SODA).

Google Scholar

Bono, A., Marani, R., Guaragnella, C., and D’Orazio, T. (2024). Biomass characterization with semantic segmentation models and point cloud analysis for precision viticulture. Comput. Electron. Agric. 218, 108712. doi: 10.1016/j.compag.2024.108712

Crossref Full Text | Google Scholar

Chen, Z., Cai, Y., Liu, Y., et al. (2025). Towards end-to-end rice row detection in paddy fields exploiting two-pathway instance segmentation. Comput. Electron. Agric. 231, 109963. doi: 10.1016/j.compag.2025.109963

Crossref Full Text | Google Scholar

Comaniciu, D. and Meer, P. (2002). Mean shift: A robust approach toward feature space analysis. IEEE Trans. Pattern Anal. And Mach. Intell. 24, 603–619. doi: 10.1109/34.1000236

Crossref Full Text | Google Scholar

Comba, L., Biglia, A., Aimonino, D. R., and Gay, P. (2018). Unsupervised detection of vineyards by 3D point-cloud UAV photogrammetry for precision agriculture. Comput. Electron. Agric. 155, 84–95. doi: 10.1016/j.compag.2018.10.005

Crossref Full Text | Google Scholar

De De Vargas, R. R. and Bedregal, B. R. C. (2013). “A way to obtain the quality of a partition by adjusted Rand index,” in 2013 2nd Workshop-School on Theoretical Computer Science (Rio Grande, Brazil: IEEE). 67–71.

Google Scholar

Deng, J., Liu, S., Chen, H., Chang, Y., and Yu Xie, Y. W. H. (2025). A precise method for identifying 3D circles in freeform surface point clouds. IEEE Trans. Instrum Meas.

Google Scholar

Dheivya, I. and Kumar, G. S. (2025). VSegNet–A variant segNet for improving segmentation accuracy in medical images with class imbalance and limited data. Medinformatics 2, 36–48.

Google Scholar

Ding, Z., Sun, Y., Xu, S., Pan, Y., Peng, Y., Mao, Z., et al. (2023). Recent advances and perspectives in deep learning techniques for 3D point cloud data processing. Robotics 12, 100. doi: 10.3390/robotics12040100

Crossref Full Text | Google Scholar

Ester, M., Kriegel, H. P., Sander, J., and Xu, X. (1996). A density-based algorithm for discovering clusters in large spatial databases with noise. kdd. 96, 226–231.

Google Scholar

Fang, Y., Yang, S., Wang, X., Li, Y., Fang, C., Liu, W., et al. (2021). Instances as queries. In Proceedings of the IEEE/CVF international conference on computer vision, (Montreal, QC, Canada: IEEE) pp. 6910–6919. doi: 10.1109/ICCV48922.2021.00683

Crossref Full Text | Google Scholar

Girardeau-Montaut, D. (2016). CloudCompare. Fr: EDF R&D Telecommun ParisTech 11, 2016.

Google Scholar

Gong, L., Gao, B., Sun, Y., Zhang, W., Lin, G., Liu, C., et al. (2024). preciseSLAM: robust, real-time, liDAR–inertial–ultrasonic tightly-coupled SLAM with ultraprecise positioning for plant factories. IEEE Trans. Ind. Inf. 20, 8818–8827. doi: 10.1109/TII.2024.3361092

Crossref Full Text | Google Scholar

Guo, X., Sun, Y., and Yang, H. (2023). FF-net: feature-fusion-based network for semantic segmentation of 3D plant point cloud. Plants 12, 1867. doi: 10.3390/plants12091867

PubMed Abstract | Crossref Full Text | Google Scholar

Guo, B., Tian, L., Zhang, J., Zhang, Y., Yu, L., Zhang, J., et al. (2019). “A clustering algorithm based on joint kernel density for millimeter wave radio channels,” in 2019 13th European Conference on Antennas and Propagation (EuCAP). 1–5 (IEEE).

Google Scholar

Han, L., Zheng, T., Xu, L., and Fang, L. (2020). Occuseg: Occupancy-aware 3d instance segmentation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. (Seattle, WA: IEEE) pp. 2940–2949

Google Scholar

Han Dong, H. D., Yang GuiJun, Y. G., Yang Hao, Y. H., Qiu ChunXia, Q. C., Chen MingJie, C. M., Wen WeiLiang, W. W., et al. (2018). Three dimensional information extraction from maize tassel based on stereoscopic vision.

Google Scholar

Huang, H., Huang, J., Liu, G., Wang, H., and Gao, T. (2024). Fuzzy C-means Clustering Information Entropy Point Cloud Simplification Using Mixed Strategy Joint Optimization. J. Xi'an Jiaotong University 58, 214–226. doi: 10.7652/xjtuxb202407020

Crossref Full Text | Google Scholar

Hu, Z., Zhen, M., Bai, X., Fu, H., and Tai, C. L. (2020). Jsenet: Joint semantic segmentation and edge detection network for 3d point clouds. In European conference on computer vision. Cham: Springer International Publishing. pp. 222–239.

Google Scholar

Huang, S. S., Ma, Z. Y., Mu, T. J., Fu, H., and Hu, S. M. (2021). Supervoxel convolution for online 3D semantic segmentation. ACM Trans. Graphics (TOG) 40, 1–15. doi: 10.1145/3453485

Crossref Full Text | Google Scholar

Jiang, L., Li, C., and Fu, L. (2025). Apple tree architectural trait phenotyping with organ-level instance segmentation from point cloud. Comput. Electron. Agric. 229, 109708. doi: 10.1016/j.compag.2024.109708

Crossref Full Text | Google Scholar

Jiazheng, S. O. N. G., Benxue, M. A., Ying, X. U., Fujia, D. O. N. G., Yujie, L. I., Guowei, Y. U., et al. (2024). Organ segmentation and phenotypic information extraction of cotton plant point cloud at the seedling stage using machine learning. Trans. Chin. Soc. Agric. Eng. 40, 199–207.

Google Scholar

Jing, W., Zhang, W., Di, D., Li, C., Emam, M., Mian, A., et al. (2025). Hypergraph biformer for semantic segmentation of high-resolution remote sensing images. IEEE Trans. Geosci. Remote Sens. doi: 10.1109/TGRS.2025.3543556

Crossref Full Text | Google Scholar

Lahoud, J., Ghanem, B., Pollefeys, M., and Oswald, M. R. (2019). 3d instance segmentation via multi-task metric learning In Proceedings of the IEEE/CVF International Conference on Computer Vision. (Seoul, Republic of Korea: IEEE) pp. 9256–9266.

Google Scholar

Li, J., Dai, H., Han, H., and Ding, Y. (2023). Mseg3d: Multi-modal 3d semantic segmentation for autonomous driving. Proc. IEEE/CVF Conf. Comput. Vision Pattern Recognit. 21694–21704. doi: 10.1109/CVPR52729.2023.02078

Crossref Full Text | Google Scholar

Liu, K., Feng, M., Zhao, W., Sun, J., Dong, W., Wang, Y., et al. (2025). Pixel-level noise mining for weakly supervised salient object detection. IEEE transactions on neural networks and learning systems.

Google Scholar

Li, R., Wang, Y., Sun, S., Zhang, Y., Ding, F., Gao, H., et al. (2025). UE-extractor: A grid-to-point ground extraction framework for unstructured environments using adaptive grid projection. IEEE Robot Autom Lett. doi: 10.1109/LRA.2025.3563127

Crossref Full Text | Google Scholar

Lloyd, S. (1982). Least squares quantization in PCM. IEEE Trans. On Inf. Theory 28, 129–137. doi: 10.1109/TIT.1982.1056489

Crossref Full Text | Google Scholar

Ni, P., Li, X., Kong, D., and Yin, X. (2023). Scene-adaptive 3D semantic segmentation based on multi-level boundary-semantic-enhancement for intelligent vehicles. IEEE Trans. Intell Veh. 9, 1722–1732. doi: 10.1109/TIV.2023.3274949

Crossref Full Text | Google Scholar

Patel, A. K., Park, E. S., Lee, H., Priya, G. L., Kim, H., Cho, B. K., et al. (2023). Deep learning-based plant organ segmentation and phenotyping of sorghum plants using liDAR point cloud. IEEE J. Sel Top Appl. Earth Observ Remote Sens. 16, 8492–8507. doi: 10.1109/JSTARS.2023.3312815

Crossref Full Text | Google Scholar

Phan, A. V., Le Nguyen, M., Nguyen, Y. L. H., and Bui, L. T. (2018). Dgcnn: A convolutional neural network over large-scale labeled graphs. Neural Networks 108, 533–543. doi: 10.1016/j.neunet.2018.09.001

PubMed Abstract | Crossref Full Text | Google Scholar

Qi, C. R., Su, H., Mo, K., and Guibas, L. J. (2017a). “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition. 652–660.

Google Scholar

Qi, C. R., Yi, L., Su, H., and Guibas, L. J. (2017b). Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 30.

Google Scholar

Qian, G., Li, Y., Peng, H., Mai, J., Hammoud, H., Elhoseiny, M., et al. (2022). PointNeXt: Revisiting pointnet++ with improved training and scaling strategies. Adv. Neural Inf. Process. Syst. 35, 23192–23204.

Google Scholar

Qiu, H., Yu, B., and Tao, D. (2022). Gfnet: Geometric flow network for 3d point cloud semantic segmentation. arXiv [preprint] arXiv:2207.02605.

Google Scholar

Rakotosaona, M. J., La Barbera, V., Guerrero, P., Mitra, N. J., and Ovsjanikov, M. (2020). “Pointcleannet: Learning to denoise and remove outliers from dense point clouds,” in Computer graphics forum, vol. 39. (Wiley-Blackwell Publishing Ltd, England), 185–203.

Google Scholar

Ren, D., Li, J., Wu, Z., Guo, J., Wei, M., Guo, Y., et al. (2024). Mffnet: multimodal feature fusion network for point cloud semantic segmentation. Visual Comput. 40, 5155–5167. doi: 10.1007/s00371-023-02907-w

Crossref Full Text | Google Scholar

Sai, K. P. and Pande, S. D. (2024). Analysis of current advancement in 3D point cloud semantic segmentation. EAI Endorsed Trans. Internet Things 10.

Google Scholar

Schonberger, J. L. and Frahm, J. M. (2016). Structure-from-motion revisited, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, (Las Vegas, NV: IEEE) pp. 4104–4113. doi: 10.1109/CVPR.2016.445

Crossref Full Text | Google Scholar

Sculley, D. (2010). “Web-scale k-means clustering,” in Proceedings of the 19th international conference on World wide web. 1177–1178.

Google Scholar

Sha, X., Guan, Z., Wang, Y., Han, J., Wang, Y., Chen, Z., et al. (2025a). SSC-Net: A multi-task joint learning network for tongue image segmentation and multi-label classification. Digit Health 11, 20552076251343696. doi: 10.1177/20552076251343696

PubMed Abstract | Crossref Full Text | Google Scholar

Sha, X., Si, X., Zhu, Y., Wang, S., and Zhao, Y. (2025b). Automatic three-dimensional reconstruction of transparent objects with multiple optimization strategies under limited constraints. Image Vision Comput. 105580. doi: 10.1016/j.imavis.2025.105580

Crossref Full Text | Google Scholar

Sha, X., Zhu, Y., Sha, X., Guan, Z., and Wang, S. (2025c). ZHPO-LightXBoost an integrated prediction model based on small samples for pesticide residues in crops. Environ. Model. Softw. 188, 106440. doi: 10.1016/j.envsoft.2025.106440

Crossref Full Text | Google Scholar

Shen, J., Wu, T., Zhao, J., Wu, Z., Huang, Y., Gao, P., et al. (2024). Organ segmentation and phenotypic trait extraction of cotton seedling point clouds based on a 3D lightweight network. Agronomy 14, 1083. doi: 10.3390/agronomy14051083

Crossref Full Text | Google Scholar

Singh, D. P. and Yadav, M. (2024). Deep learning-based semantic segmentation of three-dimensional point cloud: a comprehensive review. Int. J. Remote Sens. 45, 532–586. doi: 10.1080/01431161.2023.2297177

Crossref Full Text | Google Scholar

Sohail, S. S., Himeur, Y., Kheddar, H., Amira, A., Fadli, F., Mansoor, W., et al. (2025). Advancing 3D point cloud understanding through deep transfer learning: A comprehensive survey. Inf. Fusion 113, 102601.

Google Scholar

Sun, Y., Zhang, Z., Sun, K., Li, S., Yu, J., Zhu, R., et al. (2023). Soybean-MVS: annotated three-dimensional model dataset of whole growth period soybeans for 3D plant organ segmentation. Agriculture 13, 1321. doi: 10.3390/agriculture13071321

Crossref Full Text | Google Scholar

Tongtong, X., Bin, X., and Ximei, Z. (2023). Density peak clustering algorithm with adaptive clustering center strategy optimization. Comput. Eng. Appl. 59, 91–101. doi: 10.3778/j.issn.1002-8331.2207-0446

Crossref Full Text | Google Scholar

Vinodkumar, P. K., Karabulut, D., Avots, E., Ozcinar, C., and Anbarjafari, G. (2024). Deep learning for 3D reconstruction, augmentation, and registration: A review paper. Entropy 26, 235. doi: 10.3390/e26030235

PubMed Abstract | Crossref Full Text | Google Scholar

Wang, X., Liu, S., Shen, X., Shen, C., and Jia, J. (2019). Associatively segmenting instances and semantics in point clouds. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, (Long Beach, CA: IEEE) pp. 4096–4105.

Google Scholar

Wang, X., Song, X., Li, Z., and Wang, H. (2025). YOLO-DBS: efficient target detection in complex underwater scene images based on improved YOLOv8. J. Ocean Univ. China 24, 979–992. doi: 10.1007/s11802-025-6029-2

Crossref Full Text | Google Scholar

Wang, N., Wu, Q., Gui, Y., Hu, Q., and Li, W. (2024). Cross-modal segmentation network for winter wheat mapping in complex terrain using remote-sensing multi-temporal images and DEM data. Remote Sens. 16, 1775. doi: 10.3390/rs16101775

Crossref Full Text | Google Scholar

Wang, W., Yu, R., Huang, Q., and Neumann, U. (2018). Sgpn: Similarity group proposal network for 3d point cloud instance segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition, (Salt Lake City, UT: IEEE) pp. 2569–2578.

Google Scholar

Wei, Y., Wang, Z., Guo, T., Liu, H., Shen, L., Yuan, H., et al. (2025). High efficiency Wiener filter-based point cloud quality enhancement for MPEG G-PCC. IEEE Trans. Circuits Syst. Video Technol. doi: 10.1109/TCSVT.2025.3552049

Crossref Full Text | Google Scholar

Wu, J., Zhuang, X., Tian, Y., Cheng, Z., and Liu, S. (2024). Real-time point cloud clustering algorithm based on roadside liDAR. IEEE Sens J. 24, 10608–10619. doi: 10.1109/JSEN.2024.3361509

Crossref Full Text | Google Scholar

Xiang, D., He, D., Sun, H., Gao, P., Zhang, J., Ling, J., et al. (2025). HCMPE-Net: an unsupervised network for underwater image restoration with multi-parameter estimation based on homology constraint. Opt Laser Technol. 186, 112616. doi: 10.1016/j.optlastec.2025.112616

Crossref Full Text | Google Scholar

Xiao, A., Zhang, X., Shao, L., and Lu, S. (2024). A survey of label-efficient deep learning for 3D point clouds. IEEE Trans. Pattern Anal. Mach. Intell. 46, 9139–9160. doi: 10.1109/TPAMI.2024.3416302

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, J., Gong, J., Zhou, J., Tan, X., Xie, Y., Ma, L., et al. (2020). SceneEncoder: Scene-aware semantic segmentation of point clouds with a learnable scene descriptor. arXiv [preprint] arXiv:2001.09087. doi: 10.24963/ijcai.2020

Crossref Full Text | Google Scholar

Xu, Z., Qi, L., Du, H., Yang, J., and Chen, Z. (2025). AlignFusionNet: efficient cross-modal alignment and fusion for 3D semantic occupancy prediction. IEEE Access. doi: 10.1109/ACCESS.2025.3589858

Crossref Full Text | Google Scholar

Xuelei, M., Heru, X., and Yanqing, Z. (2024). 3D point cloud feature point detection algorithm based on hierarchical clustering. Appl. Laser. 1–7.

Google Scholar

Yasir, S. M. and Ahn, H. (2024). Deep learning-based 3D instance and semantic segmentation: A review. arXiv [preprint] arXiv:2406.13308 4, 99–114.

Google Scholar

Yu, S., Liu, X., Tan, Q., Wang, Z., and Zhang, B. (2024). Sensors, systems and algorithms of 3D reconstruction for smart agriculture and precision farming: A review. Comput. Electron. Agric. 224, 109229. doi: 10.1016/j.compag.2024.109229

Crossref Full Text | Google Scholar

Yu, Y. and Qiujie, L. (2024). Online segmentation method for target point clouds of street trees. J. Nanjing For Univ., 1–9.

Google Scholar

Zhang, F., Fang, J., Wah, B., and Torr, P. (2020). “Deep fusionnet for point cloud semantic segmentation,” in /Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28. (Cham: Springer International Publishing) 644–663.

Google Scholar

Zhang, H., Wang, C., Tian, S., et al. (2023). Deep learning-based 3D point cloud classification: A systematic survey and outlook. Displays 79, 102456. doi: 10.1016/j.displa.2023.102456

Crossref Full Text | Google Scholar

Zhao, R., Heng, Y., Wang, H., Gao, Y., Liu, S., Cai, W., et al. (2024). “Advancements in 3d lane detection using lidar point clouds: From data collection to model development,” in 2024 IEEE International Conference on Robotics and Automation (ICRA). 5382–5388 (IEEE).

Google Scholar

Zhao, L., Lu, J., and Zhou, J. (2021). “Similarity-aware fusion network for 3D semantic segmentation,” in 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). 1585–1592 (IEEE).

Google Scholar

Zhao, L. and Tao, W. (2020). “Jsnet: Joint instance and semantic segmentation of 3d point clouds,” in Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34. 12951–12958.

Google Scholar

Zhao, L., Xu, S., Liu, L., et al. (2022). SVASeg: Sparse voxel-based attention for 3D LiDAR point cloud semantic segmentation. Remote Sens. 14, 4471. doi: 10.3390/rs14184471

Crossref Full Text | Google Scholar

Zhao, L., Zhou, H., Zhu, X., et al. (2023). Lif-seg: Lidar and camera image fusion for 3d lidar semantic segmentation. IEEE Trans. Multimed 26, 1158–1168. doi: 10.1109/TMM.2023.3277281

Crossref Full Text | Google Scholar

Zhenyu, W., Lin, J., Yun, C., et al. (2024). Loop closure detection algorithm based on clustered edge points in 3D point clouds. J. Wuhan Univ. Sci. Technol., 1–10.

Google Scholar

Zhou, H., Fang, Z., Gao, Y., et al. (2020). Feature fusion network based on attention mechanism for 3D semantic segmentation of point clouds. Pattern Recognit Lett. 133, 327–333. doi: 10.1016/j.patrec.2020.03.021

Crossref Full Text | Google Scholar

Zhu, Q., Fan, L., and Weng, N. (2024). Advancements in point cloud data augmentation for deep learning: A survey. Pattern Recognit. 110532. doi: 10.1016/j.patcog.2024.110532

Crossref Full Text | Google Scholar

Keywords: 3D point clouds, cotton plant leaves, deep learning, multi-stage growth monitoring, point cloud segmentation

Citation: Lei Z, Zeng D and Zheng L (2026) PointNeXt-DBSCAN: a hybrid point cloud deep learning framework for multi-stage cotton leaf instance segmentation. Front. Plant Sci. 16:1705564. doi: 10.3389/fpls.2025.1705564

Received: 15 September 2025; Accepted: 22 December 2025; Revised: 17 December 2025;
Published: 29 January 2026.

Edited by:

Sunil Gc, North Dakota State University, United States

Reviewed by:

Mahmoud Emam, Menoufia University, Egypt
Liujun Xiao, Nanjing Agricultural University, China
Juliana Maria Espíndola Lima, Texas Tech University, United States
Rana Muhammad Irfan Anwar, Auburn University, United States
M. Prasad, Shri Vishnu Engineering College for Women, India
Praveen Chebolu, Amazon.com Azdc LLC Phoenix Mohave, United States

Copyright © 2026 Lei, Zeng and Zheng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Liangfang Zheng, bDE3MzgxODAxNTI2QDE2My5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.