Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 06 February 2026

Sec. Technical Advances in Plant Science

Volume 17 - 2026 | https://doi.org/10.3389/fpls.2026.1765604

ELGCot3D: a lightweight 3D cotton point cloud segmentation model based on EdgeConv-Local Attention-GCN and semantic feature enhancement

Hao QiuHao Qiu1Xiaoyan Meng,,Xiaoyan Meng1,2,3Yongke Li,,Yongke Li1,2,3Yunjie Zhao,,*Yunjie Zhao1,2,3*Xiaoyu LiXiaoyu Li1Shuai YinShuai Yin1Haoyuan NiuHaoyuan Niu1
  • 1School of Computer and Information Engineering, Xinjiang Agricultural University, Urumqi, China
  • 2Ministry of Education Engineering Research Center for Intelligent Agriculture, Urumqi, China
  • 3Xinjiang Agricultural Informatization Engineering, Technology Research Center, Urumqi, China

Efficient and non-destructive cotton organ extraction is crucial for automatic cotton phenotyping. However, limited by leaf occlusion, large model parameters, and inefficient manual observation, it fails to meet current high-throughput phenotyping demands. To address these challenges, this paper propose ELGCot3D, a lightweight 3D point cloud-based cotton organ segmentation method, enabling high-precision segmentation in resource-constrained environments. First, a new module called ELG3D replaces traditional Set Abstraction structures, enhancing local cotton data learning capability via multi-mechanism feature fusion and boosting segmentation accuracy. Second, a cotton-specific feature enhancement module is proposed to secondary optimize the features output from the Feature Propagation layer. This module significantly increases feature discriminability while substantially reducing redundant and high consumption network layers, achieving a balance between performance and efficiency. Finally, a cotton point cloud-adapted training strategy improves model training stability and prediction accuracy. Experimental results on the Crops3D dataset show ELGCot3D achieves 76.7% mIoU and 86.1% OA for cotton segmentation, Meanwhile, the number of parameters and computational complexity are reduced by 50.1% and 50.7%, respectively, demonstrating the model’s lightweight characteristics. Notably, it performs well in segmenting other Crops3D crops and exhibits strong generalization on the other cotton point cloud datasets. The proposed method offers a reliable approach for cotton phenotyping and precision agriculture. Future work will extend its high-throughput extraction capability for individual plant organs in large cotton fields, providing breeders with accurate data to support efficient breeding and new variety development.

1 Introduction

Cotton, as a globally important cash crop (Zhu et al., 2025), its fiber products support the core supply chain of the textile industry, while its seed oil (Younas et al., 2025) and by-products also have irreplaceable value in fields such as food (Lv et al., 2025) and chemical engineering (Yan et al., 2025). According to statistics from the Food and Agriculture Organization of the United Nations (FAO) (OECD/FAO, 2024), the global annual planting area of cotton exceeds 30 million hectares, directly related to the livelihood of hundreds of millions of people and the agricultural economic structure of dozens of countries. In modern agriculture, the core demand of the cotton industry has shifted from mere yield increase to a production model emphasizing both high-quality and efficient precision management (Xiao et al., 2024). The realization of this goal is highly dependent on the accurate quantification of cotton phenotypic traits. Studying cotton phenotypes not only enables a more precise understanding of plant physiological status, adaptability, and environmental responses (Hao et al., 2024) but also serves as a key basis for selecting high-yield and high-quality genotypes in crop breeding (Liu et al., 2025b), as well as core data support for formulating precision cultivation measures such as field water and fertilizer regulation (Jia et al., 2025) and pest early warning (Li et al., 2024).

Traditional crop phenotyping relies on manual observation and counting, which is time-consuming, labor-intensive, and susceptible to the observer’s subjective experience, making it difficult to meet the needs of large-scale breeding trials or dynamic field monitoring (Gupta et al., 2005). With the development of relevant demand-driven technologies, current crop phenotyping analysis technologies focus on 2D images and 3D point clouds (Gené-Mola et al., 2020; Sun et al., 2020; Zhang et al., 2023a). Among them, 2D images emerged as the core of early research and applications for convenient data acquisition. The technical route has evolved from traditional algorithms like threshold segmentation (Fu et al., 2019) and edge detection (Wang et al., 2018) to breakthroughs driven by deep learning. For instance, Sadeghi-Tehran et al. (Sadeghi-Tehran et al., 2019) realized wheat ear semantic segmentation by combining simple linear iterative clustering with CNN, and Yi et al. (Yi et al., 2023) modified the UNet network for accurate segmentation of birch stems and leaves. However, 2D images have inherent limitations: lack of 3D spatial information, easy key information loss caused by organ occlusion, and image quality interference from light and shadows. These issues are particularly prominent in complex field or fine phenotyping scenarios (Wei et al., 2025). With the development of computer vision and 3D perception technologies, 3D point cloud technology has gradually become a core tool in phenotyping research (Gu et al., 2024). Compared with 2D images that only capture planar texture information, it can fully reconstruct the three-dimensional morphology of cotton plants, accurately restore the spatial positional relationship between organs, effectively solve the occlusion problem, and exhibit stronger robustness to environmental interference, providing a more reliable data source for organ-level segmentation and parameter extraction (Minervini et al., 2015; Gibbs et al., 2016). This advantage has been fully verified in field crop point cloud segmentation research. For instance, Shen et al. (Shen et al., 2024) realized organ segmentation and phenotypic trait extraction of cotton seedling point clouds based on a 3D lightweight network. Liu et al (Liu et al., 2025a). proposed the TPointNetPlus model integrated with a Transformer attention module to enhance the feature discrimination ability of occluded regions. Lin et al. (Lin et al., 2022) employed 3D laser point cloud technology to develop a 3D information interpretation model for organs using field maize point cloud data. Yao et al. (Yao et al., 2025) leveraged unmanned aerial vehicle (UAV) 3D point cloud coupled with the improved PointNet++ technology to achieve automated determination of tobacco phenotypic traits, and Dong et al. (Dong et al., 2025) combined PointNeXt with the Quickshift++ algorithm to realize instance segmentation of multiple crop organs.

Despite the significant advantages of 3D point cloud technology (Qi et al., 2017b; Zhao et al., 2021), research on cotton still faces multiple challenges. Cotton plants have multiple leaf layers, dense interlaced branches and leaves, and similar morphologies between bolls and leaves, leading to severe organ occlusion. Traditional point cloud segmentation models struggle to effectively distinguish the boundaries of overlapping organs. Beyond these technical challenges inherent in cotton point cloud analysis itself, practical application scenarios pose additional hurdles. Implementing cotton phenotyping analysis in field deployment faces multiple challenges. A major algorithmic bottleneck lies in the fact that existing high-precision 3D segmentation models [e.g., PointMLP (Ma et al., 2022), CurveNet (Muzahid et al., 2020)] are often parameter-heavy and computationally complex, making it difficult to meet the requirements for low-power, real-time inference on resource-constrained platforms like drones. Therefore, improving computational efficiency is the primary objective of this study. Furthermore, transitioning from the lab to field application involves overcoming more front-end challenges, such as robustly isolating individual plants from dense crop point clouds (i.e., instance segmentation). Currently, this study focuses on achieving high-precision, lightweight organ segmentation on already isolated individual plant point clouds, which is a crucial core component for building a complete field-based phenotyping analysis pipeline. Beyond these practical implementation challenges, a deeper underlying factor that exacerbates such difficulties lies in the difference in core goals between crop phenotypic segmentation and general point cloud segmentation. The former needs to adapt to the structural characteristics of crops such as dense overlapping leaves and similar organ morphologies, as well as the deployment constraints of agricultural scenarios. Its core difficulties focus on fine organ-level separation and field data noise interference (Zhu and Yu, 2025), while the latter only pursues universal accuracy. Although general point cloud segmentation models perform excellently in conventional 3D object recognition tasks, they are difficult to directly adapt to crop scenarios. Existing crop-specific research mainly focuses on crop adaptation optimization and lightweight deployment of general architectures, and there are still obvious shortcomings in dedicated solutions (Chen et al., 2025). Turgut et al. (Turgut et al., 2022) adopted the deep learning network RoseSegNet based on 3D point clouds to successfully segment organs such as leaves, stems, and flowers from 3D point cloud data of rose bushes. Song et al. (Song et al., 2025) specifically realized cotton organ segmentation, including stems, petioles, and leaves, through the deep learning network CotSegNet. Ghahremani et al. (Ghahremani et al., 2021) utilized the deep learning model Pattern-Net to realize semantic segmentation of wheat ears and non-ear wheat organs for 3D point cloud data of wheat. Yang et al. (Yang et al., 2025) employed the deep learning model PlaneSegNet to achieve accurate segmentation of plant and non-plant targets, and validated its efficacy on the Crop3D dataset (Zhu et al., 2024). Yet this model is plagued by an excessively large number of parameters. Notably, this drawback is not exclusive to PlaneSegNet. In fact, it epitomizes a prevalent dilemma across similar deep learning-based 3D point cloud segmentation methods. Although such approaches deliver outstanding segmentation accuracy, they impose stringent demands on hardware resources, and their robustness in complex field environments remains to be enhanced. Addressing these issues is intrinsically tied to the core technical bottlenecks of 3D point cloud-based crop semantic segmentation.

Modern 3D point cloud crop semantic segmentation is mainly based on deep learning (Zhang et al., 2019), with the core challenge lying in handling the disorder, sparsity, and complex spatial correlations of point clouds (Zhang et al., 2023b). The early representative work PointNet (Qi et al., 2017a) pioneered the use of symmetric functions to address the disorder of point clouds, utilizing MLP (Taud and Mas, 2017) to directly map individual point features for end-to-end classification and segmentation. However, it can only capture global features and lacks the ability to model local neighborhood structures, making it difficult to handle occluded or dense point cloud scenarios. To overcome this limitation, progressive feature aggregation of “downsampling-local grouping-feature extraction” is achieved through the Set Abstraction (SA) module, which significantly improves the ability to capture local geometric structures. In the SA module, Farthest Point Sampling selects key center points (Li et al., 2022), grouping operations construct local neighborhoods, and then a mini-PointNet extracts neighborhood features, enabling the model to gradually learn hierarchical features from local to global. This architecture has become the foundation for subsequent point cloud segmentation models, but it still suffers from insufficient local feature discrimination when processing highly overlapping or morphologically similar objects. With in-depth research, researchers have optimized point cloud feature extraction mechanisms from various perspectives. Graph convolution (Bhatti et al., 2023) based models [e.g., Point-GNN (Shi and Rajkumar, 2020), DGCNN (Phan et al., 2018)] regard point clouds as topological graphs, model point-to-point relationships through adjacency matrices, and transmit spatial correlation information using graph convolution operators, enhancing adaptability to non-Euclidean structures. CurveNet (Muzahid et al., 2020) aggregates and synthesizes curve sequences in point clouds through a guided walking strategy, and aggregates curve features back to point features to enhance geometric description ability, significantly improving the recognition accuracy of complex surfaces and fine components. PointCNN (Li et al., 2018) is a deep learning algorithm that directly processes unstructured point clouds. It learns local neighborhood features of each point through X-Conv operators, enabling core tasks such as point cloud classification and segmentation without converting point clouds into structured forms such as voxels or images. PointMLP abandons complex geometric extractors and relies on a residual MLP architecture and geometric affine modules to dynamically adjust local features, achieving efficient local feature aggregation.

However, existing architectures still face unique challenges in practical agricultural scenarios (Saeed and Li, 2021; Guo et al., 2023). On the one hand, the dense occlusion and morphological convergence of crop organs place higher requirements on the model’s local feature discrimination ability. On the other hand, the hardware resource constraints of field deployment further strengthen the demand for model lightweighting. Most existing models either focus on accuracy improvement, leading to a surge in parameter size and computational cost, or simplify the network structure to pursue inference efficiency, thereby sacrificing segmentation performance, making it difficult to achieve an effective balance between accuracy and efficiency. At the same time, relevant phenotypic research on structurally complex crops such as cotton remains scarce. To address the core issues of complex and indistinguishable plant structures (Shen et al., 2024) and large model parameter sizes in current intelligent cotton organ extraction, this paper proposes ELGCot3D, a lightweight 3D point cloud cotton organ segmentation method. Unlike mainstream models such as the traditional PointNet++ and PointMLP, which are typically plagued by excessively large parameter volumes and limited to general-purpose segmentation tasks, this method takes the synergistic optimization of high precision and lightweighting as its core design objective. It thereby enables efficient, non-destructive extraction of cotton organ point clouds even in resource-constrained environments, through the deployment of improved network modules and tailored training strategies. The main work and contributions of this paper are as follows:

1. The traditional Set Abstraction module is reconstructed using the EdgeConv-Local Attention-GCN Hybrid 3D Cotton Phenotype Segmentation Module (ELG3D), and EdgeConv, local attention mechanism, and graph convolution are adopted to further enhance the ability to capture local features of overlapping cotton organs and alleviate the problem of low segmentation accuracy in occluded regions.

2. A feature enhancement module suitable for cotton plant structures is designed to perform secondary optimization on the high-level features output by the Feature Propagation layer. While improving feature discriminability, it achieves a balance between model performance and computational efficiency by streamlining redundant network layers and introducing efficient convolution operators.

3. An innovative training strategy adapted to the distribution of cotton point cloud data is proposed. By combining the proposed MuSGD hybrid optimizer, cosine annealing learning rate scheduler, and targeted data augmentation, it significantly enhances the training effectiveness of the lightweight model, improves the model’s adaptability to cotton phenotypes with different morphological structures, and further stabilizes the segmentation performance.

2 Methods

2.1 ELGCot3D architecture

To address the technical bottlenecks in 3D phenotypic segmentation of cotton, this paper proposes ELGCot3D, a lightweight and high-fidelity segmentation architecture. Focusing on resource-constrained scenarios, this architecture aims to maximize simultaneous improvements in accuracy and computational efficiency. Specifically, it greatly enhances the ability to capture local features of overlapping cotton organs by reconstructing the traditional Set Abstraction module into the ELG3D module. Second, a cotton-specific feature enhancement module is specially designed to streamline model parameters while strengthening feature discriminability, ultimately achieving an efficient balance between performance and computational efficiency. The overall architecture of ELGCot3D is illustrated in Figure 1.

Figure 1
Diagram of a point cloud processing pipeline. It shows an input tree point cloud progressing through stages: Hierarchical Point Set Feature Learning, Feature Propagation, and Feature Enhancement. ELG3D Modules and Set Abstraction involve multi-scale grouping and feature fusion. Local Feature Fusion uses parallel Local Attention and GCN Block. Feature Propagation includes dimension permutation, interpolation, MLP, and feature concatenation. The bottom section illustrates feature stages using colored panels, ending with a tree output point cloud.

Figure 1. The overall architecture of the ELGCot3D model. This architecture redesigns the traditional SetAbstraction module into an ELG3D module with enhanced local point cloud capture capability, and adds a feature enhancement component, significantly improving feature discriminability for cotton phenotypic segmentation.

2.2 ELG3D module

The Set Abstraction (SA) module of the original PointNet++ (Qi et al., 2017b) is the core of low-level feature extraction, with a logic of sampling-grouping-MLP aggregation. It performs multi-scale grouping after dimensionality reduction via Farthest Point Sampling (FPS) and aggregates local features relying on Multi-Layer Perceptron (MLP). However, considering the characteristics of cotton point clouds, such as severe organ occlusion, dense foliage, and the morphological similarity between bolls and leaves, the SA module design has limitations: relying solely on MLP mapping fails to distinguish the edges of occluded regions and is difficult to focus on key features, while the lack of modeling for organ topological correlations results in inadequate feature representation of complex local structures, ultimately compromising segmentation accuracy.

To address the above issues, this paper reconstructs the SA module into the EdgeConv-Local Attention-GCN Hybrid 3D Cotton Phenotype Segmentation Module (ELG3D), retaining the advantage of multi-scale grouping while introducing three core operators to enhance feature learning. First, the EdgeConv (Wang et al., 2019) operator is introduced (Figure 2) to accurately capture local geometric structures of cotton organs, such as leaf serrations and spherical cotton boll contours. Second, a Local Attention (Harley et al., 2017) operator is designed to enhance the model’s focus on key segmentation-related features (e.g., boll-petiole connections, leaf edge inflection points) and suppress redundant information in background or overlapping regions, improving feature discriminability under complex morphologies. Third, a Graph Convolution (GCN) (Kipf, 2016) operator is embedded (Figure 3) to strengthen topological correlation modeling, enabling accurate capture of inter-organ connections (e.g., leaf-main stem attachment, boll-fruiting branch connection) and enriching semantic dependence of local features, thus enhancing segmentation performance in complex scenarios like interlaced branches.

Figure 2
Diagram depicting EdgeConv operation. On the left, a yellow central node labeled X sub i connects to five blue nodes labeled X sub j1 to X sub j5 with red lines. On the right, a green central node labeled X prime sub i connects to the same blue nodes, now with pink arrows labeled e sub ji1 to e sub ji5, indicating directed relationships.

Figure 2. Visualization of the EdgeConv operation. The output of EdgeConv is calculated by aggregating the edge features of all edges associated with each connected vertex.

Figure 3
Flowchart illustrating a neural network architecture with input features branching into three processes: Center Feature Extraction, Attention Weight Calculation, and Neighbor Feature Convolution. Outputs from these are aggregated in Attention Aggregation, followed by Output Enhancement. The final stage is the Final Output. Each process specifies dimensions like [B, D, K, S].

Figure 3. Schematic diagram of the graph convolutional network (GCN) operator’s processing flow and topological correlation modeling. This figure illustrates the core workflow of the GCN operator in the ELG3D module, emphasizing its advantage in modeling the topological correlations of cotton organs. The direction of arrows in the figure indicates the feature flow direction.

The three operators process multi-scale grouped neighborhood features in parallel, with EdgeConv capturing geometric details, Local Attention emphasizing key region weights, and GCN modeling organ topological correlations to generate three differentiated feature types. First, residual fusion is applied to features of the same scale to retain core information without attenuation, followed by integration of cross-scale fused features. Finally, 1×1 convolution unifies dimensions and strengthens cross-scale correlations, forming the complete ELG3D module. Compared with the original SA module’s single feature output, the module’s enhanced local features integrate geometric, weight, and topological information with multi-scale expression capability, significantly boosting feature discriminability and laying a core foundation for subsequent high-level feature enhancement and high-precision cotton organ segmentation.

2.3 Cotton-specific feature enhancement module

Cotton point cloud semantic segmentation faces three core challenges that drive the design of the Cotton-Specific Feature Enhancement Module. First, there is significant organ scale heterogeneity, organs such as cotton bolls, leaves, and main stems vary greatly in size and morphology, making traditional single-scale feature extraction difficult to simultaneously adapt to the representation needs of different organs. Second, severe feature-background confusion exists. In scenarios with dense branches and leaves, point clouds of target organs are easily submerged by redundant backgrounds, which makes it hard to accurately capture key semantic information such as the connection between cotton bolls and petioles and leaf edge contours. Third, balancing lightweighting and representativeness is challenging. Simply stacking network layers to improve feature expression leads to a sharp increase in parameters, failing to meet the efficiency requirements of crop phenotyping. Meanwhile, lightweight designs tend to lose fine-grained structural information of point clouds.

The proposed Cotton-Specific Feature Enhancement Module is based on multi-scale feature fusion, centered on attention mechanism, and supported by efficient convolution, with its overall architecture illustrated in Figure 4. It is particularly important to note that there are significant differences in both action levels and targets between the attention mechanism of this module and the Local Attention in the ELG3D module (Section 2.2). The Local Attention in ELG3D focuses on fine-grained feature selection between points in local neighborhoods, while the dual-channel attention of this module acts at the global feature level. Specifically, the module first performs multi-branch parallel processing on the point cloud features after Feature Propagation. It extracts receptive field features of multiple scales through branches with different convolution kernel sizes to cover the scale differences of cotton organs, and forms a full-scale feature base after feature concatenation and integration. On this basis, the introduced dual-channel attention mechanism conducts global feature selection. Channel attention enhances the discriminability of different semantic categories (such as cotton bolls, stems, and leaves) in the channel dimension by modeling the dependencies between feature channels. Spatial attention focuses on the distribution pattern of target organs in the entire point cloud space, suppressing feature interference from redundant background points. The synergy of the two mechanisms achieves accurate enhancement of key information in both global semantic and spatial dimensions. Finally, to balance computational efficiency and representational capability, the module adopts depthwise separable convolution instead of traditional convolution operations. It first performs independent convolution on each channel through depthwise convolution to retain fine-grained structures, then completes information interaction between channels via pointwise convolution. This reduces parameters and computational complexity by more than 50% while ensuring the discriminability of enhanced features. It forms a hierarchical complementarity with the ELG3D module, from local point-level feature purification to global feature optimization, providing efficient and accurate feature input for subsequent segmentation heads.

Figure 4
Flowchart of a neural network architecture for point cloud segmentation. Boxes represent stages: Propagated Point Cloud Features, Multi-scale Feature Fusion, Attention Enhancement, Efficient Convolution, and Final Segmentation Head. Steps include Multi-branch Receptive Field Extraction, Feature Concatenation, Channel and Spatial Attention, Depthwise and Pointwise Convolution. Stages are linked by arrows indicating data flow.

Figure 4. Architecture of the cotton-specific feature enhancement module. The direction of arrows in the figure indicates the feature flow direction, and dashed boxes mark the core innovations of the module. It forms a hierarchical complementarity with the ELG3D module (Section 2.2), ranging from local purification to global enhancement.

2.4 Innovative training strategy for cotton 3D point cloud distribution

The ELGCot3D model faces two core bottlenecks in training for cotton 3D point cloud segmentation tasks, limiting its segmentation accuracy and robustness. On one hand, training stability is insufficient. Traditional schemes struggle to adapt to the fine-grained feature learning needs of cotton point clouds. A single optimizer cannot balance convergence speed and gradient stability, making it prone to gradient explosion or vanishing. The learning rate scheduling with fixed step decay (StepLR) causes abrupt learning rate changes, leading to rough parameter adjustments in the late training stage and oscillating segmentation errors. On the other hand, generalization ability in complex scenarios is limited. Practical cotton point cloud collection involves issues such as rotational posture differences, point cloud missing, and occlusion. However, traditional data augmentation is not designed for the characteristics of 3D point clouds, resulting in insufficient data distribution diversity. This makes the model poorly adaptable to complex scenarios and unable to stably extract edge details and local features.

To address the aforementioned issues, a coordinated improvement plan is designed from three aspects: optimizer, learning rate scheduling, and data augmentation. In the optimizer part, this paper adopts a custom MuSGD optimizer, which combines the momentum update of Momentum-SGD with the exponential moving average of squared gradients from RMSProp to construct a hybrid optimization strategy with adaptive scaling capability, with the corresponding formulas shown in Equations 15. Ideologically, this method shares conceptual similarities with Muon’s “update-norm consistency” but removes its expensive orthonormalization constraint, thus maintaining a lightweight computational overhead, and relying on a momentum buffer and a squared-gradient accumulator, MuSGD can effectively smooth gradient changes and mitigate gradient explosion or oscillation, thereby enhancing the stability of the training process.

gt=pLt+λpt(1)
mt=βmt1+gt(2)
vt=μvt1+gt2(3)
dt=vt+ϵ(4)
pt+1=ptηmtdt(5)

During the MuSGD optimization process, η denotes the learning rate, which is used to adjust the update magnitude of model parameters in each iteration; β is the momentum coefficient, which performs exponential moving average on the first-order gradient term to construct a momentum buffer and accelerate the consistency of the descent direction; µ is the exponential decay factor of the squared gradient accumulation term, which determines the smoothness of adaptive scaling (RMSProp-like scaling); λ is the weight decay coefficient, i.e., the parameter of the L2 regularization term, which is used to constrain the parameter norm and reduce the risk of overfitting; ϵ=108 is the numerical stability term, which is used in the adaptive denominator vt+ε to avoid division by zero and improve the stability of the training process.

For learning rate scheduling, a cosine annealing strategy is employed to allow the learning rate to smoothly decrease along a cosine curve—ensuring fast parameter updates in the early stage and achieving fine-tuning in the later stage to reduce error oscillation. In terms of data augmentation, three types of targeted augmentation operations are designed: First, introducing random Z-axis rotation to enhance the model’s adaptability to different postures of cotton 3D point clouds by randomly adjusting the rotation angle of point clouds in the plane. Second, implementing a random point dropout strategy to randomly eliminate part of the point cloud data with a maximum proportion of 10%, simulating point cloud missing scenarios that may occur in actual collection processes and improving the model’s anti-interference generalization ability. Third, adopting a mirror flipping operation to perform mirror transformation along the X-axis or Y-axis with a 50% probability, further enriching the spatial distribution characteristics of training data; the three work synergistically to improve the model’s convergence efficiency, adaptability to complex scenarios, and segmentation accuracy.

3 Materials

3.1 Point cloud dataset

To comprehensively and multi-dimensionally verify the performance and adaptability of the proposed model, this paper adopts a dual verification strategy combining general public datasets and cotton-specific 3D point cloud datasets, aiming to cover the evaluation requirements of cross-crop generalization and adaptability to cotton-specific scenarios.

First, systematic experiments are conducted on the public multi-crop mixed Crops3D dataset (Zhu et al., 2024), which contains abundant multi-crop point cloud samples. Partial sample examples are shown in Figure 5. Among them, cotton, potato, rice, and rapeseed point clouds are collected via Terrestrial Laser Scanning (TLS), with 176 individual plants, 118 individual plants, 84 individual plants, and 150 individual plants respectively. Tomato point clouds are collected through Structure from Motion (SfM) technology, totaling 83 individual plants.

Figure 5
Illustration depicting five plants: Cotton, Potato, Rice, Tomato, and Rapeseed. Each plant part is color-coded: Cotton (cotton balls, branches, soil), Potato (leaves, stem), Rice (pot, shoot system), Tomato (fruits, leaves, stem), and Rapeseed (soil, leaves, stem, pot). A table includes X, Y, Z coordinates, RGB values, and labels, representing data attributes for the plants.

Figure 5. Examples of the public Crops3D multi-crop mixed dataset.

Second, ablation experiments are conducted on the cotton subset of the Crops3D dataset, while generalization experiments use the Cotton data cotton-specific point cloud dataset (https://github.com/UGA-BSAIL/Cotton_plants_with_foliage). Cotton data contains 80 individual plant point clouds, divided into two categories: Cotton Ball and Cotton Foliage. It focuses on covering typical challenging scenarios in cotton segmentation such as dense foliage occlusion and organ scale heterogeneity, which can specifically verify the model’s segmentation accuracy and environmental robustness. The dataset addresses the limitation of general datasets in covering cotton-specific phenotypes and scenarios, with partial sample examples shown in Figure 6.

Figure 6
Five 3D models of cotton plants are shown, with red indicating foliage and blue for cotton balls. To the right, there are XYZ coordinates, normal vector coordinates, and labels indicating data points associated with each model. A legend explains the color coding.

Figure 6. Examples of the cotton data cotton-specific point cloud dataset.

3.2 Experimental details

To ensure the fairness of experimental comparisons, credibility of results, and reproducibility, all comparative models are trained from scratch without pre-trained weights, with experiments conducted on a server equipped with an NVIDIA GeForce RTX 4090 GPU (Santa Clara, CA, USA) and implemented based on the PyTorch v2.2.2 deep learning framework. At the input data level, considering that the Crops3D dataset contains both XYZ coordinates and RGB color parameters, while the specialized cotton point cloud dataset Cotton data includes both XYZ coordinates and Normal vector coordinates, all models are uniformly trained with only XYZ coordinates as input features to eliminate interference caused by input data differences. Regarding training configurations, key hyperparameters are summarized in Table 1, and unless otherwise specified, all experimental datasets are fixedly split into training, validation, and test sets at a 7:2:1 ratio to ensure different models are evaluated under consistent data distribution conditions.

Table 1
www.frontiersin.org

Table 1. Key hyperparameters used for training all models.

3.3 Evaluation metrics

This paper employs two types of evaluation metrics, namely segmentation accuracy and model efficiency, to assess the model. For comprehensive evaluation of segmentation accuracy, mean Intersection over Union (mIoU) and Overall Accuracy (OA) are adopted as specific metrics, with their calculation formulas shown in Equations 6, 7.

mIoU=1Mi=1MTPiTPi+FPi+FNi(6)
OA=TP+TNTP+TN+FP+FN(7)

In the formula, M in the formulas denotes the total number of categories in the segmentation task. True Positive (TP) refers to the number of samples that actually belong to the positive class and are correctly classified as the positive class. True Negative (TN) is the number of samples that actually belong to the negative class and are correctly classified as the negative class. False Positive (FP) represents the number of samples that actually belong to the negative class but are incorrectly classified as the positive class. False Negative (FN) denotes the number of samples that actually belong to the positive class but are incorrectly classified as the negative class.

Meanwhile, to measure the model efficiency, we report the total number of trainable parameters in the model (Params) and Giga Floating-Point Operations Per Second (GFLOPs). Params refers to the total number of trainable parameters in the model, measured in millions (M), and reflects the model’s size and memory footprint. GFLOPs measures the computational complexity of a single forward pass and serves as a key indicator of inference speed.

4 Results

4.1 Comparison with state-of-the-art methods

On the Crops3D validation set, we conducted comparative evaluations of ELGCot3D against various classic and cutting-edge target detectors, with results shown in Table 2. The model achieved 76.7% mIoU and 85.8% OA, increasing by 5.1% and 4.2% respectively compared with the PointNet++ baseline model. It has only 0.86 M parameters, a reduction of 50.1% compared with PointNet++, and a computational complexity of 2.41 GFLOPs, a reduction of 50.7% compared with PointNet++. All performance metrics outperform high-performance models such as PointMLP. This result is attributed to the rationality of the task-aware design approach in this paper. Instead of relying on computational power stacking, the model optimizes the local feature extraction module to specifically adapt to core segmentation challenges in cotton such as dense foliage occlusion and organ scale heterogeneity. While streamlining redundant parameters, it effectively retains key cotton phenotypic information, ultimately achieving the coordinated improvement of accuracy and efficiency under a lightweight architecture. This provides a practical and deployment-feasible solution for cotton segmentation tasks.

Table 2
www.frontiersin.org

Table 2. Performance comparison with state-of-the-art models on crops3D validation set.

4.2 Ablation study

To verify the effectiveness of each proposed core component, this paper conducts systematic ablation experiments on the cotton subset of the Crops3D dataset. Using PointNet++ as the baseline model, the performance of the model under different component combinations is compared among the proposed ELG3D module, specialized feature enhancement module, and innovative training strategy to quantify the value of each component. As shown in Table 3, the experimental results confirm that each proposed component not only can independently bring significant gains to the model’s segmentation accuracy but also can further improve the overall performance through synergistic adaptation between components, fully verifying the necessity and synergy of each component’s design.

Table 3
www.frontiersin.org

Table 3. Ablation performance on Crops3D validation set.

Among them, the ELG3D module achieves the optimal performance improvement, with mIoU increased by 3.7% and OA by 2.8%, verifying our hypothesis that the module can more efficiently capture fine-grained features and spatial correlation information of cotton organs by reconstructing the traditional Set Abstraction module and optimizing local feature fusion logic, laying a core foundation for improved segmentation accuracy. Meanwhile, integrating only the specialized feature enhancement module increases mIoU by 2.5% while further reducing the number of parameters compared with the baseline model, highlighting its strong adaptability to the semantic characteristics of cotton 3D point clouds. It strengthens organ-level information of cotton bolls and foliage through multi-scale feature fusion and precisely focuses on core semantic regions with a dual-channel attention mechanism, effectively suppressing background redundant interference and reducing unnecessary computational redundancy to balance accuracy improvement and lightweight requirements. Additionally, adopting the innovative training strategy increases mIoU by 1.1%, underscoring its precise adaptability to the data characteristics of cotton 3D point clouds; compared with general training schemes, it optimizes the model’s learning focus on complex scenarios by matching the point cloud distribution law of cotton organs, effectively reducing segmentation deviations caused by mismatched data characteristics and providing reliable training-level support for the stable improvement of model performance.

The aforementioned three core components are synergistically integrated to finally construct the complete ELGCot3D 3D cotton point cloud segmentation model. Experimental results show that compared with the baseline model, the total mIoU gain of this model reaches 5.1%, while the number of parameters is reduced by 50.1% compared with the baseline model, ultimately only 0.86M. The above results fully verify that the contributions of each core component have significant complementarity, jointly establishing an efficient and lightweight 3D point cloud segmentation framework that balances segmentation accuracy and computational efficiency, which can effectively adapt to the practical application needs of cotton point cloud segmentation.

4.3 Qualitative analysis

To intuitively understand the advantages of ELGCot3D, we conducted a qualitative comparison. Regarding the segmentation results, Figure 7 shows the segmentation outputs of ELGCot3D on point cloud samples of five crops (Cotton, Potato, Rice, Tomato, Rapeseed) covered by the Crops3D dataset, and the model exhibits outstanding performance across different crop scenarios. Specifically, it achieves more accurate segmentation of error-prone fine structures at the junctions between cotton foliage and bolls, and can clearly define organ boundaries to avoid structural adhesion issues. Meanwhile, it recognizes key phenotypic organs of crops more completely, showing significant improvement over the small target missing problem of the PointNet++ baseline model. In addition, facing occlusion scenarios such as dense overlapping of tomato leaves and intertwined canopy branches and leaves of rapeseed, the model has higher robustness—it can effectively distinguish the contours of occluded organs to reduce false negative annotations. Moreover, in areas such as junctions between rice stems and soil, and areas where rapeseed plants are mixed with the environmental background, it can completely suppress background redundant point clouds and reduce false positive annotations, resulting in higher consistency between segmentation results and the ground truth (GT). These visual results correspond with quantitative improvements, intuitively confirming that ELGCot3D can significantly reduce false negative and false positive annotations, and fully verifying its good adaptability to multi-crop point cloud segmentation scenarios.

Figure 7
Five rows of plant images labeled from A to E, each row showing a different plant: Cotton, Potato, Rice, Tomato, and Rapeseed. Labeling indicates parts of each plant using different colors: red for significant features like cotton balls and fruits, blue for branches or shoot systems, green for leaves, and additional colors for items like soil and pots. Each row presents both original and color-coded versions, illustrating the plant's structure.

Figure 7. Visual comparison results of multi-crop segmentation performance on the Crops3D dataset. Compared with the PointNet model and PointNet++ baseline model, ELGCot3D achieves more accurate segmentation results and more complete detail restoration—especially at fine structures such as crop foliage junctions, with clearer segmentation boundaries and lower misclassification rates. (A) Point Cloud. (B) GT. (C) PointNet. (D) PointNet++ (baseline). (E) ELGCot3D.

4.4 Generalization analysis

For comprehensive validation of ELGCot3D’s generalization capabilities and preventing its performance from being confined to a single dataset, this paper further conducts experiments on the Cotton data specialized cotton point cloud dataset. It specifically evaluates the model’s generalization performance in cotton-specific scenarios to verify its cross-dataset adaptability and task adaptability.

On the Cotton_data dataset covering different growth stages and morphological differences, ELGCot3D still exhibits segmentation performance exceeding all comparative baseline models. As shown in Table 4, its mIoU and OA reach 80.4% and 94.2%, respectively, significantly outperforming baseline models such as PointNet++. This result fully proves that the model has excellent environmental adaptability, it can stably adapt to the appearance, morphology, and complex background scenarios of different cotton plants, effectively reducing false positive and false negative errors during segmentation. Meanwhile, both the number of parameters and computational complexity of ELGCot3D are controlled within 50% of those of the baseline model, achieving ultra-lightweight while ensuring high-precision segmentation. This dual advantage of high precision and low overhead provides key support for the reliable deployment of the model in practical scenarios such as field portable devices and real-time phenotypic monitoring, fully meeting the actual application needs of cotton point cloud segmentation tasks.

Table 4
www.frontiersin.org

Table 4. Performance of cotton data specialized cotton point cloud dataset.

The high precision and low consumption advantages demonstrated by the above quantitative results can be more intuitively verified in the visual segmentation results. Figure 8 shows the visual segmentation comparison between ELGCot3D and the PointNet, PointNet++ baseline models on the Cotton data specialized cotton point cloud dataset. By intuitively presenting the segmentation effects of different models on the junction areas of cotton bolls and foliage, the detail processing capability of ELGCot3D in complex scenarios can be clearly observed, further confirming its performance superiority in cotton-specific segmentation tasks.

Figure 8
Comparison of cotton plants in five columns labeled A to E. In column A, the images are in grayscale. Columns B to E show colored diagrams with red representing cotton foliage and blue representing cotton bolls. Circles in columns D and E highlight specific areas on the plants.

Figure 8. Visual comparison results of segmentation performance on the cotton data specialized cotton point cloud dataset. Compared with the PointNet model and PointNet++ baseline model, ELGCot3D has minor misjudgments only in a few complex areas with dense foliage occlusion or local point cloud missing, but the overall accuracy and detail restoration of its segmentation results are significantly superior. (A) Point Cloud. (B) GT. (C) PointNet. (D) PointNet++ (baseline). (E) ELGCot3D.

From the perspective of model interpretability, ELGCot3D’s precise segmentation capability for cotton’s core organs fundamentally derives from its attention focus on two key semantic categories: branches and leaves, and bolls. Figure 9 presents the attention heatmap comparison between ELGCot3D and the PointNet++ baseline model for these core organs, which intuitively visualizes the distribution of model attention weights on each point cloud to reveal semantic capture differences between the models. As shown in the results, ELGCot3D’s attention weights are significantly denser in the effective structural regions of foliage and the surface of cotton bolls, with a higher matching degree to ground truth labels. In contrast, the PointNet++ baseline model exhibits relatively scattered attention distribution and insufficient weight intensity in target organ regions. This difference clearly confirms the effectiveness of the dual-channel attention mechanism in the specialized feature enhancement module—by precisely allocating channel semantic weights and spatial position weights for foliage and cotton bolls, it guides the model to actively focus on core organs and suppress non-target interference, providing direct mechanism-level support for improved segmentation accuracy from the perspective of decision-making logic visualization.

Figure 9
Four panels labeled A, B, C, and D, display abstract visualizations of plant structures. Panels A and B depict colored branches with leaves, showing variations in color intensity. Panels C and D present isolated colored clusters, resembling leaf outlines, in similar color schemes.

Figure 9. Attention heatmap comparison of “Cotton Foliage - Cotton Ball” on the cotton data specialized cotton point cloud dataset. In the heatmap, a brighter yellow color indicates a higher attention weight of the model on the point cloud, while a darker yellow color indicates a lower weight. (A) PointNet++ (Cotton Foliage). (B) ELGCot3D (Cotton Foliage). (C) PointNet++ (Cotton Ball). (D) ELGCot3D (Cotton Ball).

Finally, Figure 10 presents the point cloud confusion matrices of ELGCot3D and the baseline model PointNet++ on the Cotton data test set. These matrices intuitively reveal the differences in classification performance between the two methods, clearly showing the correct segmentation and misclassification distribution characteristics of cotton components. Key evaluation metrics including precision, recall, and F1-score are labeled below the matrices. Comprehensive comparison indicates that ELGCot3D exhibits superior segmentation performance.

Figure 10
Two confusion matrices labeled A and B compare classification results for “Cotton Ball” and “Cotton Foliage.” Matrix A shows 137,475 true positives and 104,749 false positives for Cotton Ball, and 805,452 true negatives and 15,166 false negatives for Cotton Foliage. Precision and recall are lower for Cotton Ball. Matrix B shows improved metrics with 130,105 true positives and 37,017 false positives for Cotton Ball, and 873,184 true negatives and 22,536 false negatives for Cotton Foliage, with higher precision and recall for both categories.

Figure 10. Confusion matrix comparison on cotton data. Comparison of F1-scores shows that ELGCot3D is significantly superior to PointNet++. (A) PointNet++ (baseline). (B) ELGCot3D.

5 Discussion

The ELGCot3D model designed in this paper achieves an effective balance between segmentation accuracy and computational efficiency in 3D cotton point cloud segmentation tasks. It exhibits superior performance over traditional baseline models on both the Crops3D multi-crop dataset and the Cotton data specialized cotton point cloud dataset. Through a multi-operator fusion mechanism, it precisely focuses on effective foliage structures and cotton boll regions, providing technical support for the deployment of field portable devices. However, the model still has limitations: it currently only conducts semantic segmentation for individual cotton plants, and has not yet realized instance segmentation for distinguishing individual organs within the same category. Additionally, the model has not yet been validated with large-scale field point cloud data of densely planted cotton populations, and lacks targeted research on the processing and analysis pipeline for large-area field scenarios, which restricts its direct scalability to practical large-scale cotton phenotyping applications. In the future, the research direction will be further expanded: on the one hand, an instance segmentation branch will be introduced to distinguish individual organs; on the other hand, the model will be extended to the verification and adaptation of large-scale field dense planting scenarios, and a set of efficient point cloud down-sampling and batch processing strategies will be established to promote the practical application of the model in large-area cotton fields. This will provide a more comprehensive technical solution for cotton growth monitoring and yield estimation.

6 Conclusion

To address the core pain points of insufficient segmentation accuracy and low deployment efficiency in 3D cotton point cloud segmentation, this paper develops an efficient and lightweight ELGCot3D model, achieving technical breakthroughs through three key improvements. Specifically, it reconstructs the traditional Set Abstraction module into the ELG3D module to enhance the ability to capture fine-grained features of cotton organs. A cotton-specific feature enhancement module integrating multi-scale features and dual-channel attention is designed to precisely focus on core semantic regions, suppress background redundant interference, and reduce computational consumption. Meanwhile, an innovative training strategy adapted to the distribution of cotton point cloud data is proposed to ensure more stable model training. Experimental verification shows that the model performs excellently on both the Crops3D multi-crop dataset and the Cotton data specialized cotton point cloud dataset. Particularly on the Cotton data specialized cotton point cloud dataset, it achieves an mIoU of 80.4% and an OA of 94.2%, with only 0.86M parameters and computational complexity controlled below 50% of the baseline model, realizing a balance between high precision and lightweight. With its lightweight characteristics and efficient inference capability, the model can directly provide technical support for real-time monitoring of cotton phenotypes. Nevertheless, this study has certain limitations that warrant further exploration: the model’s performance is constrained when handling extreme leaf occlusion scenarios where overlapping organs are almost indistinguishable, and its inference efficiency and segmentation stability may decline when processing ultra-large-scale field point cloud data due to the current architecture’s limited scalability for massive data. In the future, relying on the point cloud segmentation model, we will further conduct research on high-throughput phenotypic extraction technology for large-area field cotton plants, optimize the model architecture to address extreme occlusion and large-scale data processing issues, and aim to provide more accurate phenotypic data support for breeders, thereby assisting in optimizing the efficient cotton breeding process and cultivating high-quality new varieties.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Author contributions

HQ: Visualization, Conceptualization, Writing – review & editing, Writing – original draft, Methodology. XM: Supervision, Writing – original draft, Funding acquisition. YL: Writing – original draft, Supervision, Funding acquisition. YZ: Conceptualization, Writing – review & editing, Funding acquisition, Writing – original draft. XL: Writing – review & editing, Visualization, Validation. SY: Data curation, Project administration, Writing – review & editing. HN: Visualization, Investigation, Writing – review & editing.

Funding

The author(s) declared that financial support was received for this work and/or its publication. This work was supported by the fund of the National Natural Science Foundation of China (32560419), the open research projects of the Xinjiang Agricultural Informatization Engineering Technology Research Center (XJAIEC2025K011), and the Provincial Key S&T Program of Xinjiang (2022A02011).

Acknowledgments

We would like to express our sincere gratitude to all the experts who provided various support for this research, as well as the reviewers for their valuable comments and suggestions.

Conflict of interest

The author(s) declared that this work was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declared that generative AI was not used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpls.2026.1765604/full#supplementary-material

References

Bhatti, U. A., Tang, H., Wu, G., Marjan, S., and Hussain, A. (2023). Deep learning with graph convolutional networks: An overview and latest applications in computational intelligence. Int. J. Intell. Syst. 2023, 8342104. doi: 10.1155/2023/8342104

Crossref Full Text | Google Scholar

Chen, X., Wen, S., Zhang, L., Lan, Y., Ge, Y., Hu, Y., et al. (2025). A calculation method for cotton phenotypic traits based on unmanned aerial vehicle lidar combined with a three-dimensional deep neural network. Comput. Electron. Agric. 230, 109857. doi: 10.1016/j.compag.2024.109857

Crossref Full Text | Google Scholar

Dong, S., Fan, X., Li, X., Liang, Y., Zhang, M., Yao, W., et al. (2025). Automatic 3d plant organ instance segmentation method based on pointnext and quickshift++. Plant Phenom., 100065. doi: 10.1016/j.plaphe.2025.100065

PubMed Abstract | Crossref Full Text | Google Scholar

Fu, L., Tola, E., Al-Mallahi, A., Li, R., and Cui, Y. (2019). A novel image processing algorithm to separate linearly clustered kiwifruits. Biosyst. Eng. 183, 184–195. doi: 10.1016/j.biosystemseng.2019.04.024

Crossref Full Text | Google Scholar

Gené-Mola, J., Sanz-Cortiella, R., Rosell-Polo, J. R., Morros, J.-R., Ruiz-Hidalgo, J., Vilaplana, V., et al. (2020). Fruit detection and 3d location using instance segmentation neural networks and structure-from motion photogrammetry. Comput. Electron. Agric. 169, 105165. doi: 10.1016/j.compag.2019.105165

Crossref Full Text | Google Scholar

Ghahremani, M., Williams, K., Corke, F. M., Tiddeman, B., Liu, Y., and Doonan, J. H. (2021). Deep segmentation of point clouds of wheat. Front. Plant Sci. 12, 608732. doi: 10.3389/fpls.2021.608732

PubMed Abstract | Crossref Full Text | Google Scholar

Gibbs, J. A., Pound, M., French, A. P., Wells, D. M., Murchie, E., and Pridmore, T. (2016). Approaches to three-dimensional reconstruction of plant shoot topology and geometry. Funct. Plant Biol. 44, 62–75. doi: 10.1071/FP16167

PubMed Abstract | Crossref Full Text | Google Scholar

Gu, Y., Wang, Y., Wu, Y., Warner, T. A., Guo, T., Ai, H., et al. (2024). Novel 3d photosynthetic traits derived from the fusion of uav lidar point cloud and multispectral imagery in wheat. Remote Sens. Environ. 311, 114244. doi: 10.1016/j.rse.2024.114244

Crossref Full Text | Google Scholar

Guo, R., Xie, J., Zhu, J., Cheng, R., Zhang, Y., Zhang, X., et al. (2023). Improved 3d point cloud segmentation for accurate phenotypic analysis of cabbage plants using deep learning and clustering algorithms. Comput. Electron. Agric. 211, 108014. doi: 10.1016/j.compag.2023.108014

Crossref Full Text | Google Scholar

Gupta, P. K., Rustgi, S., and Kulwal, P. L. (2005). Linkage disequilibrium and association studies in higher plants: present status and future prospects. Plant Mol. Biol. 57, 461–485. doi: 10.1007/s11103-005-0257-z

PubMed Abstract | Crossref Full Text | Google Scholar

Hao, H., Wu, S., Li, Y., Wen, W., Zhang, Y., Zhuang, L., et al. (2024). Automatic acquisition, analysis and wilting measurement of cotton 3d phenotype based on point cloud. Biosyst. Eng. 239, 173–189. doi: 10.1016/j.biosystemseng.2024.02.010

Crossref Full Text | Google Scholar

Harley, A. W., Derpanis, K. G., and Kokkinos, I. (2017). “Segmentation-aware convolutional networks using local attention masks,” in Proceedings of the IEEE International Conference on Computer Vision. (Venice, Italy: IEEE) 5038–5047.

Google Scholar

Jia, Y., Li, Y., He, J., Biswas, A., Siddique, K. H., Hou, Z., et al. (2025). Enhancing precision nitrogen management for cotton cultivation in arid environments using remote sensing techniques. Field Crops Res. 321, 109689. doi: 10.1016/j.fcr.2024.109689

Crossref Full Text | Google Scholar

Kipf, T. (2016). Semi-supervised classification with graph convolutional networks. arXiv. arXiv, 1609.02907.

Google Scholar

Li, J., Zhou, J., Xiong, Y., Chen, X., and Chakrabarti, C. (2022). “An adjustable farthest point sampling method for approximately-sorted point cloud data,” in 2022 IEEE workshop on signal processing systems (SiPS). 1–6 (Rennes, France: IEEE).

Google Scholar

Li, R., He, Y., Li, Y., Qin, W., Abbas, A., Ji, R., et al. (2024). Identification of cotton pest and disease based on cfnet-vov-gcsp-lsknet-yolov8s: a new era of precision agriculture. Front. Plant Sci. 15, 1348402. doi: 10.3389/fpls.2024.1348402

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Bu, R., Sun, M., Wu, W., Di, X., and Chen, B. (2018). Pointcnn: Convolution on x-transformed points. Adv. Neural Inf. Process. Syst. 31, 828–838.

Google Scholar

Lin, C., Hu, F., Peng, J., Wang, J., and Zhai, R. (2022). Segmentation and stratification methods of field maize terrestrial lidar point cloud. Agriculture 12, 1450. doi: 10.3390/agriculture12091450

Crossref Full Text | Google Scholar

Liu, F.-Y., Geng, H., Shang, L.-Y., Si, C.-J., and Shen, S.-Q. (2025a). A cotton organ segmentation method with phenotypic measurements from a point cloud using a transformer. Plant Methods 21, 37. doi: 10.1186/s13007-025-01357-w

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, Z., Fan, L., Shu, S., Qanmber, G., Chen, E., Huang, J., et al. (2025b). Cotton metabolism regulatory network: Unraveling key genes and pathways in fiber development and growth regulation. Plant Commun. 6, 101221. doi: 10.1016/j.xplc.2024.101221

PubMed Abstract | Crossref Full Text | Google Scholar

Lv, L., Lv, Q., Yang, Y., Xiong, F., Pei, S., He, S., et al. (2025). Identification of novel antioxidant peptides from cottonseed meal co-fermented with lactobacillus mucosae llk-xr1 and acid proteases: In silico screening, molecular simulation, and in vitro functional analysis. Food Chem. 483, 144285. doi: 10.1016/j.foodchem.2025.144285

PubMed Abstract | Crossref Full Text | Google Scholar

Ma, X., Qin, C., You, H., Ran, H., and Fu, Y. (2022). Rethinking network design and local geometry in point cloud: A simple residual mlp framework. arXiv. arXiv, 2202.07123.

Google Scholar

Minervini, M., Scharr, H., and Tsaftaris, S. A. (2015). Image analysis: the new bottleneck in plant phenotyping [applications corner. IEEE Signal Process. Mag. 32, 126–131. doi: 10.1109/MSP.2015.2405111

Crossref Full Text | Google Scholar

Muzahid, A., Wan, W., Sohel, F., Wu, L., and Hou, L. (2020). Curvenet: Curvature-based multitask learning deep networks for 3d object recognition. IEEE/CAA. J. Automatica. Sin. 8, 1177–1187. doi: 10.1109/JAS.2020.1003324

Crossref Full Text | Google Scholar

OECD/FAO (2024). Oecd-fao agricultural outlook 2024–2033 (Paris, France: OECD). Available online: https://www.fao.org/documents/card/en/c/cc9016en. (Accessed September 25, 2025).

Google Scholar

Phan, A. V., Le Nguyen, M., Nguyen, Y. L. H., and Bui, L. T. (2018). Dgcnn: A convolutional neural network over large-scale labeled graphs. Neural Networks 108, 533–543. doi: 10.1016/j.neunet.2018.09.001

PubMed Abstract | Crossref Full Text | Google Scholar

Qi, C. R., Su, H., Mo, K., and Guibas, L. J. (2017a). “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition. (New York, NY, USA: IEEE) 652–660.

Google Scholar

Qi, C. R., Yi, L., Su, H., and Guibas, L. J. (2017b). Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 30, 4–9.

Google Scholar

Sadeghi-Tehran, P., Virlet, N., Ampe, E. M., Reyns, P., and Hawkesford, M. J. (2019). Deepcount: in-field automatic quantification of wheat spikes using simple linear iterative clustering and deep convolutional neural networks. Front. Plant Sci. 10, 1176. doi: 10.3389/fpls.2019.01176

PubMed Abstract | Crossref Full Text | Google Scholar

Saeed, F. and Li, C. (2021). “Plant organ segmentation from point clouds using point-voxel cnn,” in 2021 ASABE annual international virtual meeting, Vol. 1 (St. Joseph, Michigan: American Society of Agricultural and Biological Engineers).

Google Scholar

Shen, J., Wu, T., Zhao, J., Wu, Z., Huang, Y., Gao, P., et al. (2024). Organ segmentation and phenotypic trait extraction of cotton seedling point clouds based on a 3d lightweight network. Agronomy 14, 1083. doi: 10.3390/agronomy14051083

Crossref Full Text | Google Scholar

Shi, W. and Rajkumar, R. (2020). “Point-gnn: Graph neural network for 3d object detection in a point cloud,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. (Seattle, WA, USA: IEEE) 1711–1719.

Google Scholar

Song, J., Ma, B., Xu, Y., Yu, G., and Xiong, Y. (2025). Organ segmentation and phenotypic information extraction of cotton point clouds based on the cotsegnet network and machine learning. Comput. Electron. Agric. 236, 110466. doi: 10.1016/j.compag.2025.110466

Crossref Full Text | Google Scholar

Sun, S., Li, C., Chee, P. W., Paterson, A. H., Jiang, Y., Xu, R., et al. (2020). Three-dimensional photogrammetric mapping of cotton bolls in situ based on point cloud segmentation and clustering. ISPRS. J. Photogramm. Remote Sens. 160, 195–207. doi: 10.1016/j.isprsjprs.2019.12.011

Crossref Full Text | Google Scholar

Taud, H. and Mas, J.-F. (2017). “Multilayer perceptron (mlp),” in Geomatic approaches for modeling land change scenarios (Berlin/Heidelberg, Germany: Springer), 451–455.

Google Scholar

Turgut, K., Dutagaci, H., and Rousseau, D. (2022). Rosesegnet: An attention-based deep learning architecture for organ segmentation of plants. Biosyst. Eng. 221, 138–153. doi: 10.1016/j.biosystemseng.2022.06.016

Crossref Full Text | Google Scholar

Wang, Y., Sun, Y., Liu, Z., Sarma, S. E., Bronstein, M. M., and Solomon, J. M. (2019). Dynamic graph cnn for learning on point clouds. ACM Trans. Graphics (Tog). 38, 1–12. doi: 10.1145/3326362

Crossref Full Text | Google Scholar

Wang, Z., Wang, K., Yang, F., Pan, S., and Han, Y. (2018). Image segmentation of overlapping leaves based on chan–vese model and sobel operator. Inf. Process. Agric. 5, 1–10. doi: 10.1016/j.inpa.2017.09.005

Crossref Full Text | Google Scholar

Wei, X., Wang, Q., Li, K., and Zhang, W. (2025). A method for sesame (sesamum indicum l.) organ segmentation and phenotypic parameter extraction based on cavf-pointnet++. Plants 14, 2898. doi: 10.3390/plants14182898

PubMed Abstract | Crossref Full Text | Google Scholar

Xiao, S., Fei, S., Ye, Y., Xu, D., Xie, Z., Bi, K., et al. (2024). 3d reconstruction and characterization of cotton bolls in situ based on uav technology. ISPRS. J. Photogramm. Remote Sens. 209, 101–116. doi: 10.1016/j.isprsjprs.2024.01.027

Crossref Full Text | Google Scholar

Yan, Z., Li, T., Zou, G., Zhang, X., Qu, L., and Wei, Y. (2025). Probiotic fermentation of defatted cottonseed meal for sustainable foods and non-food applications. Microorganisms 13, 1020. doi: 10.3390/microorganisms13051020

PubMed Abstract | Crossref Full Text | Google Scholar

Yang, X., Xu, C., Wang, Y., Feng, R., Yu, J., Su, Z., et al. (2025). Planesegnet: A deep learning network with plane attention for plant point cloud segmentation in agricultural environments. Artif. Intell. Agric. 16, 284–299. doi: 10.1016/j.aiia.2025.10.015594

Crossref Full Text | Google Scholar

Yao, J., Wang, W., Fu, H., Deng, Z., Cui, G., Wang, S., et al. (2025). Automated measurement of field crop phenotypic traits using uav 3d point clouds and an improved pointnet++. Front. Plant Sci. 16, 1654232. doi: 10.3389/fpls.2025.1654232

PubMed Abstract | Crossref Full Text | Google Scholar

Yi, X., Wang, J., Wu, P., Wang, G., Mo, L., Lou, X., et al. (2023). Ac-unet: an improved unet-based method for stem and leaf segmentation in betula luminifera. Front. Plant Sci. 14, 1268098. doi: 10.3389/fpls.2023.1268098

PubMed Abstract | Crossref Full Text | Google Scholar

Younas, A. Y., Parveen, A., and Waqar, S. (2025). Versatile applications, challenges, and future prospects of cottonseed oil. BinBee–Arı. Ve. Doğal. Ürünler. Dergisi. 5, 24–33.

Google Scholar

Zhang, H., Wang, L., Jin, X., Bian, L., and Ge, Y. (2023a). High-throughput phenotyping of plant leaf morphological, physiological, and biochemical traits on multiple scales using optical sensing. Crop J. 11, 1303–1318. doi: 10.1016/j.cj.2023.04.014

Crossref Full Text | Google Scholar

Zhang, R., Wu, Y., Jin, W., and Meng, X. (2023b). Deep-learning-based point cloud semantic segmentation: A survey. Electronics 12, 3642. doi: 10.3390/electronics12173642

Crossref Full Text | Google Scholar

Zhang, J., Zhao, X., Chen, Z., and Lu, Z. (2019). A review of deep learning-based semantic segmentation for point cloud. IEEE Access 7, 179118–179133. doi: 10.1109/ACCESS.2019.2958671

Crossref Full Text | Google Scholar

Zhao, H., Jiang, L., Jia, J., Torr, P. H., and Koltun, V. (2021). “Point transformer,” in Proceedings of the IEEE/CVF international conference on computer vision. (Nashville, TN, USA: IEEE) 16259–16268.

Google Scholar

Zhu, Y., Wang, G., Du, H., Liu, J., and Yang, Q. (2025). The effect of agricultural mechanization services on the technical efficiency of cotton production. Agriculture 15, 1233. doi: 10.3390/agriculture15111233

Crossref Full Text | Google Scholar

Zhu, Q. and Yu, M. (2025). A corn point cloud stem-leaf segmentation method based on octree voxelization and region growing. Agronomy 15, 740. doi: 10.3390/agronomy15030740

Crossref Full Text | Google Scholar

Zhu, J., Zhai, R., Ren, H., Xie, K., Du, A., He, X., et al. (2024). Crops3d: a diverse 3d crop dataset for perception and segmentation toward agricultural applications. Sci. Data 11, 1438. doi: 10.1038/s41597-024-04290-0

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: 3D point cloud, cotton organ segmentation, ELG3D, ELGCot3D, feature enhancement, lightweight

Citation: Qiu H, Meng X, Li Y, Zhao Y, Li X, Yin S and Niu H (2026) ELGCot3D: a lightweight 3D cotton point cloud segmentation model based on EdgeConv-Local Attention-GCN and semantic feature enhancement. Front. Plant Sci. 17:1765604. doi: 10.3389/fpls.2026.1765604

Received: 11 December 2025; Accepted: 19 January 2026; Revised: 12 January 2026;
Published: 06 February 2026.

Edited by:

Zhenghong Yu, Guangdong Polytechnic of Science and Technology, China

Reviewed by:

Kai Huang, Jiangsu Academy of Agricultural Sciences, China
Wei She, Hunan Agricultural University, China

Copyright © 2026 Qiu, Meng, Li, Zhao, Li, Yin and Niu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yunjie Zhao, enlqQHhqYXUuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.