Your new experience awaits. Try the new design now and help us make it even better

ORIGINAL RESEARCH article

Front. Plant Sci., 12 September 2025

Sec. Sustainable and Intelligent Phytoprotection

Volume 16 - 2025 | https://doi.org/10.3389/fpls.2025.1654232

This article is part of the Research TopicPlant Phenotyping for AgricultureView all 12 articles

Automated measurement of field crop phenotypic traits using UAV 3D point clouds and an improved PointNet++

Jiatong YaoJiatong Yao1Wei WangWei Wang2Hongyu FuHongyu Fu3Zhehong DengZhehong Deng1Guoxian CuiGuoxian Cui2Shuaibin WangShuaibin Wang4Dong WangDong Wang4Wei She*Wei She2*Xiaolan Cao*Xiaolan Cao1*
  • 1College of Information and Intelligence, Hunan Agricultural University, Changsha, China
  • 2College of Agriculture, Hunan Agricultural University, Changsha, China
  • 3Hunan Cultivated Land and Agricultural Eco-Environment Institute, Changsha, China
  • 4Technology Center, China Tobacco Hunan Industrial Co., Ltd, Changsha, China

Accurate acquisition of tobacco phenotypic traits is crucial for growth monitoring, cultivar selection, and other scientific management practices. Traditional manual measurements are time-consuming and labor-intensive, making them unsuitable for large-scale, high-throughput field phenotyping. The integration of 3D reconstruction and stem–leaf segmentation techniques offers an effective approach for crop phenotypic data acquisition. In this study, we propose a tobacco phenotyping method that combines unmanned aerial vehicle (UAV) remote sensing with an improved PointNet++ model. First, a 3D point-cloud dataset of field-grown tobacco plants was generated using multi-view UAV imagery. Next, the PointNet++ architecture was enhanced by incorporating a Local Spatial Encoding (LSE) module and a Density-Aware Pooling (DAP) module to improve the accuracy of stem and leaf segmentation. Finally, based on the segmentation results, an automated pipeline was developed to compute key phenotypic traits, including plant height, leaf length, leaf width, leaf number, and internode length. Experimental results demonstrated that the improved PointNet++ model achieved an overall accuracy (OA) of 95.25% and a mean intersection over union (mIoU) of 93.97% for tobacco plant segmentation—improvements of 5.12% and 5.55%, respectively, over the original PointNet++ model. Moreover, using the segmentation results from the improved PointNet++ model, the predicted phenotypic values exhibited strong agreement with ground-truth measurements, with coefficients of determination (R²) ranging from 0.86 to 0.95 and root mean square errors (RMSE) between 0.31 and 2.27 cm. This study provides a technical foundation for high-throughput phenotyping of tobacco and presents a transferable framework for phenotypic analysis in other crops.

1 Introduction

Tobacco, as one of the world’s major economic crops, requires accurate plant phenotypic parameters for cultivar improvement and optimization (Koh et al., 2021). Traditional tobacco phenotyping data collection primarily relies on manual measurements, which are time-consuming, costly, prone to human error, and may damage plant morphology, making them unsuitable for large-scale phenotypic analysis tasks (Kaiser et al., 2018). UAV remote sensing technology addresses the limitations of traditional phenotyping methods by offering advantages such as rapid, non-destructive, and flexible data collection, becoming a promising tool for acquiring field crop phenotypic information. Using multi-view remote sensing imagery captured by UAVs, combined with 3D reconstruction technology, it is possible to obtain the complete geometric structure of plants in a non-contact manner, enabling refined phenotypic parameter extraction through plant segmentation.

In UAV-based plant phenotyping studies, stem-leaf segmentation remains a challenging task (Chen et al., 2025). Traditional segmentation methods primarily rely on techniques such as region growing, attribute clustering, and skeleton extraction. For instance, Li et al. proposed a leaf segmentation method for dense plant point clouds based on small planar region growing, which involves three steps: point cloud preprocessing, over-segmentation of planar regions, and region growing, to achieve individual leaf segmentation in greenhouse ornamental plants (Li et al., 2018). Ferrara et al. employed the DBSCAN density-based clustering algorithm to analyze plant point clouds collected by LiDAR, enabling the automatic separation of tree leaves and trunks (Ferrara et al., 2018). Sun et al. constructed a dual-threshold segmentation approach using the Otsu algorithm, based on plant height and reflectance intensity in different parts of rice panicles, and optimized point cloud processing through super pixel and mean-shift clustering methods (Sun et al., 2021). Although these methods have achieved certain success in plant stem-leaf segmentation, they heavily depend on manually defined rules, involve high computational complexity, and are difficult to scale for high-throughput phenotyping. With the advancement of deep learning, its end-to-end feature extraction and adaptive optimization capabilities offer a more robust and computationally efficient solution for plant stem-leaf segmentation.

Deep learning-based segmentation methods are generally categorized into four types: projection-based, voxel-based, graph-based, and point-based approaches (Sarker et al., 2024). Among these, projection, voxelization, and graph-based methods require the transformation of point cloud data into regular grid or graph structures (Shi et al., 2019; Yang et al., 2021; Saeed et al., 2023). However, these transformations often lead to the loss of three-dimensional information and fine details during processing (Das Choudhury et al., 2020; Zięba-Kulawik et al., 2021). And may also introduce instability in adjacency construction and feature propagation (Phan et al., 2018; Mirande et al., 2022).

In contrast, point-based methods operate directly on raw point cloud data without requiring spatial transformations, allowing for better preservation of original geometric information. This advantage is particularly important in point cloud segmentation tasks, where maintaining fine-grained local geometric features is essential. Point-based methods also effectively avoid the detail loss issues commonly associated with voxelization and graph construction. In addition, these methods typically require fewer computational resources and are well-suited for handling complex and irregular point cloud structures, making them increasingly popular in plant phenotyping applications (Turgut et al., 2022; Yang et al., 2024).

In the field of tobacco phenotypic trait extraction, the application of these methods remains relatively limited. Existing approaches primarily rely on depth cameras or LiDAR to obtain point cloud data of individual tobacco plants, followed by segmentation using clustering algorithms. However, such methods often struggle with insufficient feature extraction capability, low computational efficiency, and uneven point cloud density when processing the detailed structures of tobacco organs (Briechle et al., 2020; Zhu et al., 2023). Furthermore, variations in plant growth stages introduce additional noise and structural changes in the point cloud data, further affecting segmentation accuracy. To address these challenges, this study proposes a deep learning-based method for tobacco organ segmentation and phenotypic trait computation. The main contributions are as follows:

1. In this study, we constructed a tobacco point cloud dataset to enhance model robustness. A 3D point cloud model of tobacco plants was generated using multi-view UAV image acquisition, followed by point cloud preprocessing and manual annotation. This dataset improves the uniformity of point cloud density and contributes to more accurate phenotypic analysis.

2. We proposed an improved PointNet++ model. Building upon the original PointNet++ architecture, the model adopts a multi-layer feature extraction strategy and integrates a Local Spatial Encoding (LSE) module and a Density-Aware Pooling (DAP) module. First, a multilayer perceptron (MLP) is used to learn basic point cloud features. Then, the LSE module captures local spatial relationships to enhance the representation of complex plant structures. Finally, the DAP module adaptively selects pooling strategies based on neighborhood point cloud density, improving the fusion of salient local features and addressing the limitations of traditional methods in feature extraction.

3. We integrated stem-leaf segmentation with phenotypic parameter computation to improve the efficiency of phenotypic data acquisition. This method extracts phenotypic parameters, including plant height, leaf length, leaf width, leaf number, and internode length. In the field of tobacco, a phenotypic parameter calculation method has been proposed, which successfully resolves the contradiction between measurement accuracy and efficiency in phenotypic parameter determination.

2 Materials and methods

This study is divided into the following four main parts (Figure 1):

1. Tobacco point cloud dataset construction: Multi-view images of tobacco plants were collected using a UAV, and a 3D point cloud model was reconstructed. Individual plant extraction and point cloud preprocessing were performed to provide a data foundation for model training.

2. Tobacco stem-leaf segmentation network: An improved network based on PointNet++ was proposed. Coordinate normalization was first applied to the input point cloud to standardize spatial information, followed by enhancement of the feature extraction module to improve the segmentation performance of the model for tobacco stem and leaf structures. The model was then trained accordingly.

3. Phenotypic parameter calculation: Based on the segmentation results, coordinate de-normalization was performed to restore the actual scale, and phenotypic traits such as plant height, leaf length, leaf width, leaf number, and internode length were calculated.

4. Result evaluation: Evaluation was conducted from two perspectives: segmentation performance and phenotypic accuracy. The segmentation results were validated through comparisons with mainstream models and ablation experiments, while the accuracy of phenotypic parameter estimation was verified against measured data.

Figure 1
Flowchart illustrating the process of tobacco plant analysis. It consists of four main steps: 1) Tobacco Point Cloud Dataset Construction, featuring data acquisition and processing stages. 2) Tobacco Stem-Leaf Segmentation Network, detailing feature extraction and model training. 3) Phenotypic Parameter Calculation, involving segmentation results and calculations of plant characteristics. 4) Result Evaluation, highlighting comparison experiments and result visualization. Images and diagrams within each step depict drones, point cloud models, and graphical data representations.

Figure 1. Overall workflow for calculating tobacco phenotypic parameters.

2.1 Dataset and collection

2.1.1 Experimental area

The experimental area is located at the Guandu Experimental Base of the Hunan Tobacco Technology Center in Liuyang, Hunan Province (Figure 2). This region has a subtropical monsoon humid climate with abundant rainfall and favorable light and heat conditions. A total of 400 tobacco breeding materials were planted in the experimental area. Tobacco seedlings were transplanted in March 2024 with a row spacing of 120 cm and a plant spacing of 50 cm.

Figure 2
Map and aerial imagery of Liuyang, Hunan. Panel (a) shows the location of Liuyang within Hunan. Panel (b) presents an aerial view of agricultural fields captured by a drone. Panel (c) displays six close-up images of crops arranged in rows.

Figure 2. Tobacco image data collection. (A) schematic of the experimental base; (B) experimental data collection area; (C) UAV captured multi-view tobacco images during two different periods.

2.1.2 Data collection

Data collection for this study was conducted at two critical developmental stages of tobacco: the peak growth stage (June 23, 2024) and the post-topping stage (July 13, 2024) (Figure 2). The data types included UAV imagery and detailed ground-based phenotypic measurements, all collected on the same day for each stage.

(1) UAV imagery data

The image data were acquired using a DJI Inspire 2 unmanned aerial vehicle (UAV) equipped with a Zenmuse X5s camera (35 mm focal length). The camera has an effective pixel count of 20.8 million and a maximum resolution of 5280×3956. The UAV flew at a height of 5 meters and captured images of the tobacco plants at 30°, 60°, and 90° angles. The flight speed was set to 2 meters per second, resulting in a total of 2220 images collected (Figure 2).

(2) Ground-based tobacco phenotypic data

Phenotypic data were obtained through manual measurements, including the true values of tobacco plant height, leaf length, leaf width, leaf number, and internode length. During measurement, plant height was defined as the straight-line distance from the ground to the highest leaf; leaf length was the straight-line distance from the point where the leaf connects to the stem to the leaf tip; leaf width was measured as the maximum width perpendicular to the main vein direction; leaf number was determined by manual counting; and internode length was defined as the vertical distance between the base of the upper and middle leaves.

2.2 Tobacco point cloud model reconstruction and processing

2.2.1 Tobacco point cloud model reconstruction

This study employs a multi-view image-based reconstruction technique, using Structure From Motion (SFM) and Multi-View Stereo (MVS) methods to construct the tobacco point cloud model (Luo et al., 2024). The reconstruction process includes the generation of sparse point clouds and the reconstruction of dense point clouds (Gonçalves et al., 2021). The entire reconstruction process was done using the structure-from-light measurement software Agisoft Metashape (version 2.1.2).

The SFM-MVS workflow includes the following steps (Gao et al., 2022) (Figure 3): (1) Feature Extraction and Matching: Key feature points are extracted from the input multi-view tobacco images, and feature matching relationships are established between the images. (2) Camera Pose Estimation: Based on the feature matching results, the camera’s position and orientation are estimated for each capture location. (3) Sparse Point Cloud Generation: Using the camera poses and matched feature points, the 3D coordinates of the feature points are recovered through triangulation, generating the sparse point cloud. (4) Dense Point Cloud Reconstruction: The sparse point cloud is refined by increasing point density, fusing multi-view image information to generate a detailed dense point cloud.

Figure 3
Diagram illustrating two processes: “Tobacco Point Cloud Model Reconstruction” and “Ground Point Removal.” The top section shows steps from feature extraction to dense point cloud reconstruction. The bottom section depicts a process from a tobacco field model to ground point removal and tobacco point preservation using region segmentation and RANSAC clustering.

Figure 3. Overall workflow of tobacco model reconstruction and preprocessing. (A) Tobacco Point Cloud Model Reconstruction; (B) Ground Point Removal.

2.2.2 Ground point removal

Due to the large space occupied by ground points in the model and their frequent mixing with tobacco point clouds, it is necessary to remove the ground points. First, the entire tobacco model is segmented to reduce computational load (Ghahremani et al., 2021). Then, the Random Sample Consensus (RANSAC) algorithm is used to remove ground points (distance threshold: 0.2; number of points for fitting the plane: 3; number of iterations: 500) for ground point cloud removal (Figure 3).

The RANSAC algorithm iteratively selects random point sets to fit a plane model, and based on the distance from points to the plane, it identifies and segments ground and non-ground points (Harintaka and Wijaya, 2023). Specifically, three points are randomly selected to fit the plane model (Equation 1), the distance from each point to the plane is calculated (Equation 2), and ground points are identified based on a pre-set distance threshold. If the distance from a point to the plane is smaller than the threshold, it is classified as a ground point. Through multiple iterations, the plane with the most inliers is selected as the final ground model.

ax+ by+cz+d=0(1)
d(pi)=|axi+byi+czi+d|a2+b2+c2(2)

where (a,b,c) represents the normal vector of the plane; (x,y,z) represents the coordinates of any point in the tobacco model; pi​ represents the i-th point; and d(pi) represents the distance from the point to the plane.

2.2.3 Point cloud denoising, downsampling, and labeling

After removing ground points, individual tobacco plants were extracted, and statistical filtering (neighboring points: 50; outlier threshold: 1.0) was applied for denoising to remove scattered points and improve data quality (Zhao et al., 2021). Uniform downsampling (sampling rate: 2) was performed to reduce the number of points and increase computational efficiency (Al-Rawabdeh et al., 2020). Additionally, based on the structural features of tobacco, point cloud data for the stem and leaves were manually labeled to clearly define the spatial position and geometric shape of the tobacco stem, which was used for model prediction comparison (Figure 4). Finally, region segmentation, individual plant extraction, and point cloud labeling were all conducted using CloudCompare software (version 2.14).

Figure 4
Four-panel sequence showing a 3D point cloud of a tree. Panel (a) depicts the initial point cloud. Panel (b) and (c) show processing stages with reduced complexity. Panel (d) displays the final segmented tree with distinct colored sections. Arrows indicate progression between stages.

Figure 4. Individual tobacco plant extraction process. (A) individual tobacco plant; (B) statistical filtering; (C) uniform downsampling; (D) point cloud labeling.

A total of 122 tobacco plants were labeled to construct the tobacco point cloud dataset. The training set included 102 plants, and the test set included 20 plants. Table 1 presents the statistics for each category.

Table 1
www.frontiersin.org

Table 1. Category statistics for training and test sets.

2.3 Tobacco stem-leaf segmentation network based on improved PointNet+

Although PointNet++ is capable of capturing local features through its hierarchical structure, its feature extraction accuracy remains limited when processing crop point clouds, especially for crops like tobacco with complex morphological structures. To enhance the segmentation performance of tobacco stems and leaves, this study introduces two key improvements to the original PointNet++: the integration of the LSE module and the incorporation of the DAP module.

The improved PointNet++ model adopts a multi-level feature extraction strategy. After extracting fundamental features using an MLP, the network enhances its feature representation at two hierarchical levels, as described below:

For enhancing local spatial information, traditional ball query grouping relies on a fixed radius, which may result in overly large or small local regions and thus fails to accurately capture fine-grained details of target structures in non-uniform point clouds. To address this limitation, this study adopts a K-Nearest Neighbors (KNN) grouping strategy and integrates the LSE module (Cheng et al., 2021). The LSE encodes local spatial relationships by computing the relative coordinates between each center point and its neighboring points. The encoded features are then concatenated with the original features and passed through an MLP to adjust the output dimension, thereby enhancing the representation of local spatial information (Hu et al., 2020; Chen et al., 2024).

For local feature aggregation, to enhance the model’s ability to extract features from non-uniform point clouds, this study incorporates the DAP module. It first computes the neighborhood density of each center point and then applies a relative density-based strategy to distinguish between high- and low-density regions. In high-density areas, max pooling is used to extract the most prominent local features, while in low-density regions, attention pooling is applied to perform weighted aggregation, emphasizing salient feature representation (Wu et al., 2022). Finally, the features from both areas are concatenated to obtain enhanced point cloud features, thereby improving the model’s segmentation performance in tobacco point cloud tasks (Deng et al., 2023).

2.3.1 PointNet++ network architecture

PointNet++ is a deep learning network designed for point cloud processing (Figure 5). As an improved version of PointNet, it enables hierarchical feature learning from point cloud data (Qi et al., 2017a).

Figure 5
Diagram illustrating a hierarchical point set feature learning process for point cloud segmentation. It shows two main sections: “Sampling & Grouping” with “PointNet” in green and yellow blocks labeled SA1 and SA2, and “Interpolation” with “Unit PointNet” in orange and yellow blocks labeled FP1 and FP2. The diagram includes arrows indicating the workflow and “Skip link concatenation” connections.

Figure 5. PointNet++ network architecture.

The network architecture follows an encoder–decoder structure composed of multiple Set Abstraction (SA) modules and Feature Propagation (FP) modules. Each SA module consists of a sampling layer, a grouping layer, and a PointNet layer. The sampling layer adopts Farthest Point Sampling (FPS), the grouping layer constructs local regions using ball query, and the PointNet layer employs an MLP to extract features, followed by max pooling to aggregate global features. In the decoding stage, the FP module upsamples features through distance-weighted interpolation to recover spatial resolution and utilizes a Unit PointNet to further extract global features for each point, ultimately completing the point cloud segmentation (Qi et al., 2017b).

2.3.2 Improved PointNet++ network architecture

To enhance the spatial alignment capability of the model, the input point cloud data were first normalized to a unified coordinate scale. The overall architecture of the improved PointNet++ consists of an encoder and a decoder (Figure 6).

Figure 6
Diagram illustrating an encoder-decoder architecture for plant image data. The encoder has multiple layers with skip link concatenation, showing stages SA1 to SA4 with specific parameters. The decoder mirrors this with modules labeled FP1 to FP4. The process begins with a plant image and results in a segmented output, demonstrating data flow through modules like Sampling, MLP, LSE, and DAP.

Figure 6. Improved PointNet++ network architecture.

The encoder is composed of multiple stacked SA modules. Each module incorporates a sampling layer, a grouping layer, an MLP layer, an LSE module, and a DAP module. These elements are employed to extract point cloud features layer by layer, encode spatial positional relationships, and conduct local feature aggregation.

Specifically, in the encoder process, the input is tobacco point cloud data in the (N, 3) format, where N indicates the number of points and 3 represents the feature dimension. The network utilizes the FPS strategy to determine the number of central points. To balance the point cloud information between shallow and deep layers, each layer retains half of the central-point count from the previous layer using this strategy. Subsequently, each central point selects its nearest neighboring points to construct local regions. Next, feature dimension elevation is carried out via the MLP to extract deep-level point cloud information. Finally, point position information is enhanced through the LSE and DAP modules, thereby obtaining key features of the entire tobacco plant.

The decoder adopts the original FP module to perform step-by-step interpolation recovery on the downsampled features, reconstructing a high-resolution point cloud structure. Ultimately, accurate segmentation of the stem and leaf structures in the tobacco point cloud is achieved.

2.3.3 Local spatial encoding module

This study introduces the LSE module (Figure 7), which explicitly encodes the local spatial relationships between center points and neighboring points to enhance point cloud positional information (Hu et al., 2020; Cheng et al., 2021).

Figure 7
Flowchart illustrating a neural network process. It starts with an input labeled \(N, 3, d\). The input undergoes operations including KNN grouping and feature concatenation. These processes result in intermediate outputs with dimensions \((K, 3)\), \((K, d)\), and \((K, 2d)\), leading to the final output. Components involve MLP and relative position encoding, with arrows indicating the data flow.

Figure 7. Structure of the LSE module.

The specific implementation steps of the module are as follows:

First, the input point cloud is represented in the format of (N, 3, d), where N is the number of points, each containing 3D coordinates (x,y,z) and d-dimensional features.

Next, after constructing local regions using KNN grouping, the Euclidean distances between each center point and its K nearest neighbors are calculated. The absolute coordinates of the center and neighboring points, their relative coordinates, and the corresponding distances are concatenated to form the relative position encoding. This encoding is then passed through an MLP for nonlinear mapping, with the output dimension adjusted to match the original feature size (Equation 3).

Finally, the encoded features are concatenated with the original features to form the position-enhanced feature vector (Equation 4). This encoding method, by fusing positional information with original features, provides an enhanced representation for each neighboring point, thereby clearly expressing the local structure of the center point and improving the network’s spatial perception capabilities.

rik=MLP((pipik)(pipik)d(pi,pik))(3)
tik=rikqik(4)

where rik​ represents the encoded feature; pi represents the feature of the center point;​ pik represents the feature of the neighboring point; tik​ represents the position-enhanced feature; qik​ represents the original feature; represents feature concatenation; (pipik)​ represents the concatenation of the absolute coordinates of the center and neighboring points; (pipik)​ represents their relative coordinates; and d(pi,pik) represents the Euclidean distance between them.

2.3.4 Density-aware pooling module

After enhancing the local spatial information of the point cloud with the LSE module, this study introduces a Density-Aware Dynamic Pooling (DAP) module (Figure 8) to further extract key feature information. Traditional max pooling effectively extracts prominent features but ignores the relationships between features. Attention pooling can adaptively focus on key features but suffers from redundancy in high-density regions.

Figure 8
Diagram illustrating a neural network flow for density-based region classification. It shows input processing through neighborhood density computation and attention pooling, using components like MLP, softmax function, dot product, and max pooling, culminating in an output. Blocks are colored and labeled to indicate different stages, with annotations for dimensions and operations.

Figure 8. Structure of the DAP module.

To combine the advantages of both, this study uses KNN to compute the neighborhood density of the center point after position enhancement (Equation 5), and normalizes the density values of all points (Equation 6), mapping them to a unified range. Then, a relative density-based strategy is introduced: if the normalized density of the center point is greater than the average density of the points in the region, it is classified as a high-density region (Equation 7); otherwise, it is classified as a low-density region (Equation 8). This strategy dynamically selects the pooling method based on the density values, adaptively processing features from different density regions, thereby improving the model’s robustness to density variations.

Di=1Kk=1K1d(pi,pik)+(5)
D˜i=DiDminDmaxDmin(6)
iHigh :D˜i>1Kk=1KD˜ik(7)
iLow :D˜i1Kk=1KD˜ik(8)

where Di  represents the neighborhood density of the center point; K represents the number of neighboring points; represents a small constant added to prevent division by zero; D˜i​ represents the normalized density of the point; Dmin and Dmax  represent the minimum and maximum densities in the entire point cloud, respectively; 1Kk=1KD˜ik  represents the average normalized density of the neighboring points.

Different pooling strategies are employed for local feature extraction in regions with varying densities:

Low-density regions: An attention pooling strategy is applied. First, the features of neighboring points are fed into a shared function composed of an MLP and softmax to compute attention scores. These scores are then used to weight the local features through element-wise multiplication (dot product), and the weighted features are summed to obtain the aggregated local representation (Equation 9). This strategy adaptively emphasizes key features in sparse areas and improves the feature representation in low-density regions.

High-density regions: A max pooling strategy is adopted. In dense regions such as stems or leaves, max pooling effectively suppresses redundant information and extracts the most prominent structural features within the local area (Equation 10).

Finally, the features extracted from both regions are concatenated along the channel dimension and passed through an MLP for feature fusion and dimensionality reduction, yielding the enhanced point cloud representation (Equation 11). This module fully leverages the structural characteristics of regions with different densities and achieves fine-grained local feature extraction.

Fatt=k=1KSoftmax(g(aik;W))·aik(9)
Fmax=maxk=1K(mik)(10)
Fi=MLP(FattFmax)(11)

where Di  represents the neighborhood density of the center point; K represents the number of neighboring points; represents a small constant added to prevent division by zero; D˜i represents the normalized density of the point; Dmin and Dmax  represent the minimum and maximum densities in the entire point cloud, respectively; 1Kk=1KD˜ik  represents the average normalized density of the neighboring points.

2.4 Tobacco phenotypic parameter calculation

2.4.1 Coordinate de-normalization

To ensure consistent data scaling, coordinate de-normalization was applied to the normalized point cloud data during the computation process, restoring it to the original scale to ensure the accuracy of the results. The core of de-normalization involves using the minimum value and range recorded during normalization to fix the point cloud coordinates to the actual physical space. Specifically, all normalized point cloud coordinates (x, y, z) are de-normalized to align with the measured data (Equation 12).

xorig=xnorm×Rx+minx(12)

where xorig represents the original coordinate value; xnorm represents the normalized coordinate value within the range [0,1]; Rx represents the value range for the corresponding dimension; and minx represents the minimum value of that dimension.

2.4.2 Phenotypic feature extraction

Extraction process of tobacco phenotypic features (Patel et al., 2023) (Figure 9).

Figure 9
Diagram showcasing the process of tobacco phenotypic feature extraction and visualization. The top section illustrates steps such as plant height, leaf number, PCA analysis, projection, and MBB calculation. The bottom section displays 3D visualizations of plant height, leaf number, internode length, and a chart for leaf blade dimensions, highlighting various phenotypic calculation methods.

Figure 9. Visualization of phenotypic feature calculation and methods. (a) Tobacco Phenotypic Feature Extraction Process; (b) Visualization of Phenotypic Calculation Methods.

Plant Height: The height difference between the highest and lowest points in the segmentation results is used as the predicted value for plant height (Equation 13).

Hpred=ZmaxZmin(13)

where Hpred represents the predicted plant height; Zmax  and Zmin represent the z-coordinates of the highest and lowest points, respectively.

1. Leaf number: Based on the label grouping of the segmentation results, the point cloud data is processed for region identification, with each leaf treated as an independent region and numbered. Different colors correspond to different labels.

2. Leaf length and leaf width: These are calculated using Principal Component Analysis (PCA) combined with the Minimum Bounding Box (MBB) method. First, based on the segmentation results, PCA is applied to compute the main direction of the leaf point cloud, and the point cloud is projected onto the XY plane. Then, the minimum bounding rectangle on this plane is calculated, with the longest and shortest sides defined as the approximate values for leaf length and leaf width, respectively.

3. Internode length: The vertical distance between the labels of the upper and middle leaves. The predicted value is calculated from the vertical distance between the corresponding leaf label regions in the segmentation results.

2.4.3 Visualization of phenotypic calculation methods

This section employs visualization techniques to present the phenotypic features of tobacco plants Figure 9.

Figure A visually illustrates the distribution of plant height using a color gradient; Figure B shows the relationship between leaf length and width; Figure C distinguishes the leaf number of different plants using color differences; and Figure d reveals the internode length of the plants, displaying the vertical arrangement density of the leaves.

2.5 Network training and evaluation metrics

2.5.1 Experimental environment and parameter settings

To achieve accurate point cloud segmentation of tobacco plants, this study reconstructed the experimental framework from two perspectives: experimental environment configuration and model training strategy design. Table 2 presents the hardware and software configurations used in the experiment. Table 3 presents the key hyperparameters optimized for the segmentation experiments.

Table 2
www.frontiersin.org

Table 2. Experimental environment parameters.

Table 3
www.frontiersin.org

Table 3. Hyperparameters for the segmentation experiments.

2.5.2 Segmentation evaluation metrics

In this study, the segmentation results of the tobacco point cloud dataset were quantitatively evaluated using four performance metrics: Overall Accuracy (OA), Intersection over Union (IoU), and mean Intersection over Union (mIoU).

OA assesses the overall classification performance of the model by calculating the ratio of correctly classified points to the total number of points (Equation 14).

IoU is used to measure the segmentation performance for each class, defined as the ratio of the intersection to the union between the predicted region and the ground-truth region of a specific class (Equation 15).

The mIoU reflects the overall segmentation performance across all classes and is calculated as the average IoU over all categories (Equation 16).

OA=i=0kPiii=0kj=0kPij(14)
IoU=Piii=0kPij+j=0kPjiPii(15)
mIoU=1k+1i=0kPiii=0kPij+j=0kPjiPii(16)

where k represents the total number of classes; Pii represents the number of points correctly predicted as belonging to class i ; Pij represents the number of points from class  i incorrectly predicted as belonging to class j; i=0kPij represents the total number of points in class i (including both correctly and incorrectly classified points).

2.5.3 Phenotypic evaluation metrics

This study adopts the coefficient of determination (R²) and the root mean square error (RMSE) to evaluate model performance.

R² assesses the degree of fit between the predicted and measured values (Equation 17), where a value closer to 1 indicates a better fit.

RMSE quantifies the average deviation between the predicted and measured values and reflects the magnitude of prediction error, with units consistent with the original data (Equation 18).

R2=1i=1n(yiy^i)2i=1n(yiy¯)2(17)
RMSE=1ni=1n(yiy^i)2(18)

where yi  represents the measured value, y^i represents the predicted value, y¯  represents the mean of the measured values, and n represents the total number of data samples.

3 Results

3.1 Accuracy analysis of the improved PointNet++ tobacco stem-leaf segmentation model

To validate the effectiveness of the proposed method, under consistent experimental conditions, the segmentation accuracy of the improved PointNet++ model was compared not only with that of existing mainstream point cloud segmentation models but also with that of the classic DBSCAN clustering method for tobacco stem - leaf segmentation, as shown in Table 4.

Table 4
www.frontiersin.org

Table 4. Comparison of results for different models on the tobacco dataset (%).

The experimental results show that the improved PointNet++ model performs best in segmentation accuracy, with mIoU and OA reaching 93.97%, 95.25%, respectively. Compared to PointNet++ (Qi et al., 2017b), these two metrics improved by 5.55%, 5.12%, respectively, demonstrating its advantages in fine-grained feature extraction and structural recognition.

In terms of class segmentation performance, the Improved PointNet++ model achieved Stem IoU of 92.60% and Leaf IoU of 95.33%, respectively, significantly outperforming other comparison models. Specifically, the DBSCAN (Gaonkar and Sawant, 2013) clustering exhibits certain performance in stem segmentation but performs poorly in terms of Leaf IoU and OA metrics. The DGCNN (Phan et al., 2018) and PointNet (Qi et al., 2017a) models are limited in local feature extraction, resulting in relatively poor overall segmentation performance; RandLA-Net (Hu et al., 2020) exhibits some instability in leaf segmentation; although GrowSP (Zhang et al., 2023) outperforms the first three models (in comparison) in terms of stem and leaf IoU, there is still significant room for improvement. The results of PointNeXt (Qian et al., 2022) and PointNet++ are relatively close, but the Leaf IoU of PointNet++ is higher than that of PointNext.

Therefore, the incorporation of the LSE and DAP modules has enhanced the accuracy and robustness of the model in tobacco stem-leaf segmentation, particularly in category differentiation and feature aggregation.

Meanwhile, Table 5 presents the results of ablation experiments conducted to further assess the contribution of the LSE and DAP modules to the overall performance of the improved model. Compared to the baseline model, the inclusion of the LSE module led to increases of 3.88% and 3.34% in mIoU and OA, respectively, indicating that local spatial representation was effectively enhanced. This improvement alleviates the issue of incomplete spatial structure expression caused by sparsely distributed point clouds.

Table 5
www.frontiersin.org

Table 5. Comparison of results for different modules on the tobacco dataset (%).

Subsequently, the incorporation of the DAP module resulted in additional improvements of 3.32% and 2.04% across the two evaluation metrics, respectively. These results suggest that density-aware pooling effectively optimizes local feature extraction across regions with varying point densities, addressing the limited adaptability of conventional pooling strategies in both high- and low-density areas.

Ultimately, the improved PointNet++ model, incorporating both LSE and DAP modules, achieved the best overall performance, with a mIoU of 93.97% and OA of 95.25%. In particular, the IoU values for stem and leaf segmentation improved by 4.90% and 6.19%, respectively, confirming that the synergy between the LSE and DAP modules significantly strengthens the model’s ability to extract structural features and improves the accuracy of complex plant structure recognition.

3.2 Visualization of tobacco stem-leaf segmentation results

To visually demonstrate the phenotypic characteristics of tobacco in the point cloud segmentation task, three tobacco plants with varying levels of structural complexity were selected. The visualized results of five segmentation models are compared (Figure 10), clearly reflecting the differences between the ground truth and the predicted results.

Figure 10
Comparative image showing segmentation results using different methods on a 3D structure. The top row is the ground truth followed by results from DBSCAN, DGCNN, PointNet, and RandLA-Net. The second row shows GrowSP, PointNeXt, PointNet++, and Improved PointNet++. Each method visualizes segmented regions in distinct colors for three different instances.

Figure 10. Visualization comparison of segmentation results.

For plants with simple structures, the DBSCAN algorithm exhibits leaf segmentation errors. The prediction results of the other models are relatively close to the ground truth labels, but there are still a small number of mis-segmentations in the details.

As the complexity of plant structures increases, the DBSCAN algorithm shows limitations in leaf segmentation, with relatively severe mis-segmentation situations. the segmentation results displayed by the DGCNN and PointNet models are relatively coarse, with some regions not matching the ground truth, and they also have limitations in capturing different structural boundaries. RandLA-Net exhibits some inconsistencies, especially in the leaf regions, where there are mismatches between the color-coded segments and the ground truth. The GrowSP model shows improvements compared to the first three models, but there are still deviations between the segmentation results in some regions and the ground truth, indicating room for improvement. PointNeXt demonstrates more accurate segmentation, approaching the ground truth in many regions. PointNet++ further enhances the segmentation accuracy, with better alignment between the segmented parts and the ground truth.

In contrast, the improved PointNet++ achieves better segmentation performance, outperforming other models in small structure regions. It demonstrates superior segmentation ability in dense leaf regions, highlighting its advantages in point cloud segmentation tasks for complex structures.

3.3 Extraction results of tobacco phenotypic parameters based on the improved model

This study predicts five key phenotypic parameters of tobacco (Figure 11).

Figure 11
Scatter plots compare predicted versus measured values for five plant traits: plant height, leaf length, leaf width, leaf number, and internode length. Each plot includes a trend line with corresponding equations and statistical metrics such as R-squared and RMSE values, indicating model accuracy. Data points are in varying colors to differentiate traits.

Figure 11. Comparison of predicted values and measured values. (a) computed results from the PointNet++ model; (b) computed results from the improved PointNet++ model.

The results show that the model performs well in predicting plant height, leaf length, leaf width, and leaf number, with high accuracy. Specifically, the R² value for the measured and predicted plant height is 0.95, with an RMSE of 2.03 cm, indicating a small deviation between predicted and measured values and accurately reflecting the trend in plant height variation. The prediction accuracies for leaf length and leaf width are 0.91 and 0.89, with RMSE values of 2.27 cm and 2.26 cm, respectively, with leaf length showing slightly better fitting performance than leaf width. Overall, the model exhibits low prediction errors for leaf morphological features and can consistently estimate leaf growth parameters. For leaf number estimation, the model achieves a fitting accuracy of 0.94 and an RMSE of 0.31 cm, while the internode length prediction has an R² of 0.86 and an RMSE of 1.16 cm.

As demonstrated in Table 6, the improved PointNet++ model exhibits superior fitting performance compared to the PointNet++ model across all phenotypic prediction tasks. The R2 analysis reveals consistent improvements in the five tobacco phenotypic traits, with enhancements of 0.06, 0.04, 0.06, 0.05, and 0.02, respectively. The modified model also demonstrates significant advantages in prediction accuracy, showing error reductions of 1.07 cm in plant height, 0.46 cm in leaf length, 0.64 cm in leaf width, 0.13 in leaf number, and 0.08 cm in internode length. These results substantiate the improved model’s enhanced capability in handling complex overlapping leaf scenarios. The substantial reduction in prediction errors across all phenotypic parameters validates the effectiveness of both the LSA module for feature extraction and the DAP module for structural recognition.

Table 6
www.frontiersin.org

Table 6. Calculated results for tobacco phenotypic indicators.

4 Discussion

4.1 3D reconstruction accuracy improvement with multi-view imaging and point cloud preprocessing

Due to self-shadowing effects, leaf overlap, and a lack of depth information, traditional 2D imaging techniques based on UAV remote sensing struggle to effectively extract low-noise and evenly distributed plant point cloud features from 2D images. Multi-view imaging was employed in the UAV flight mission to address this issue. As shown in Figure 3, images were captured from multiple viewpoints and densely reconstructed, effectively capturing the complex morphological details of tobacco plants (Hu et al., 2020). Furthermore, ground points and outliers in the reconstructed tobacco point cloud interfere with canopy segmentation and subsequent phenotypic extraction, leading to significant errors in phenotypic extraction. To tackle the issue of ground point interference, the RANSAC algorithm was employed to remove ground points, as illustrated in Figure 3. Through multiple iterations, the RANSAC algorithm randomly selects point sets to fit a plane model and determines ground and non-ground points based on their distance from the plane, enabling ground point identification and segmentation (Ghahremani et al., 2021; Harintaka and Wijaya, 2023). This effectively mitigates ground interference and noise in field point clouds, improving the completeness of individual tobacco plant point clouds.

Figure 4 illustrates the overall workflow of individual tobacco plant point cloud preprocessing. The statistical filtering method reduces redundant information while preserving the overall structural features. On this basis, a uniform downsampling method is applied to reduce point cloud density and improve subsequent processing efficiency. Additionally, the preprocessed point cloud is manually annotated to provide precise semantic labels for the supervised learning model, laying the foundation for the subsequent 3D segmentation task. The results indicate that this method effectively resolves ground interference and noise issues in field point clouds, enhancing the completeness of individual tobacco plant point clouds.

4.2 Application of the improved PointNet++ in tobacco stem-leaf segmentation

Traditional point cloud processing methods, such as clustering analysis and skeleton extraction, often rely on manually set thresholds and multi-stage processing pipelines for tobacco stem-leaf segmentation. However, these methods struggle to achieve stable segmentation, particularly in the presence of complex plant structures and overlapping leaves (Miao et al., 2021). In contrast, deep learning models offer end-to-end feature learning capabilities, enabling automatic extraction and differentiation of stem and leaf shape and texture features. This reduces reliance on manual expertise and enhances the model’s generalization ability in complex environments.

In this study, the LSE and DAP modules were introduced to the original PointNet++ model. The LSE module enhances the model’s ability to perceive local positional information, while the DAP module allows the model to extract stable and representative local features even in sparse or overlapping regions. The comparison experiment results in Table 4 show that OA and mIoU improved by 5.12% and 5.55%, respectively, compared to the original PointNet++. This improvement demonstrates that the introduced modules effectively compensate for PointNet++’s shortcomings in fine-grained stem-leaf feature extraction, enhancing the model’s segmentation ability and robustness in complex agricultural point cloud structures.

To further validate the individual contributions and synergistic effects of the introduced modules, an ablation experiment was conducted. The results in Table 5 show that the model’s OA and mIoU improved compared to the original architecture, regardless of whether the LSE or DAP module was introduced alone, confirming the effectiveness of each module in its respective function.

4.3 Phenotypic parameter analysis

Building on the implementation of stem-leaf segmentation, this study further calculated the phenotypic parameters of tobacco plants. As shown in Table 6, the study achieved the automated extraction of key phenotypic traits, including plant height, leaf length, leaf width, leaf number, and internode length. The phenotypic traits obtained by this method show a strong correlation with manual measurements (R² > 0.86). Compared to traditional manual phenotyping, this method demonstrates improvements in both efficiency and stability. Manual measurements typically rely on manual recording, which is time-consuming, inefficient, and prone to human error, making it difficult to meet the high-throughput demands of large-scale field phenotyping research (Li et al., 2022). In contrast, this study leverages a 3D point cloud model combined with semantic segmentation results to automate the extraction of phenotypic parameters, significantly enhancing data acquisition speed.

In addition, the method considers spatial scale consistency during parameter extraction by performing coordinate de-normalization on the segmentation results, ensuring the comparability of the calculated phenotypic parameters in real-world dimensions. Compared to the common point selection errors and reading fluctuations in manual measurements, this method performs automatic identification and measurement based on structural logic, enhancing the stability of parameter computation.

4.4 Future work

The model still faces some limitations when dealing with complex plant structures. When processing tobacco plants at different growth stages or with intricate morphologies, the segmentation accuracy of the model may be affected, particularly in predicting fine structures such as overlapping leaf regions and internode length, where some errors still exist.

Future research will focus on optimizing local feature extraction and phenotypic parameter calculation, particularly improving the segmentation accuracy of complex tobacco stem-leaf structures and the precision of leaf length, leaf width, and internode length measurements. This will further enhance the model’s generalization ability across different growth stages and tobacco varieties. Additionally, the tobacco point cloud dataset will be expanded to include various growth environments and stages, thereby strengthening the model’s robustness and adaptability while optimizing data annotation strategies. Furthermore, to validate the model’s generalizability, future studies will explore its application to other economic crops, assessing its suitability for point cloud segmentation and phenotypic calculation tasks in different crops, thereby enhancing its potential for agricultural intelligence and precision phenotypic analysis.

5 Conclusion

This study presents a tobacco three-dimensional phenotypic analysis method tailored for field applications, covering key aspects such as data collection, 3D reconstruction, stem-leaf segmentation, and phenotypic parameter extraction. By integrating multi-view imaging with SFM-MVS technology, tobacco plant models were successfully reconstructed, providing a reliable approach for non-contact data acquisition in field environments.

Building upon this, an improved PointNet++ segmentation model was proposed, which enhanced the accuracy of tobacco stem-leaf recognition. Based on the segmentation results, the automated extraction of phenotypic parameters, including plant height, leaf length, leaf width, leaf number, and internode length, was achieved. The predicted values showed good correlation with the measured values (R² > 0.86), demonstrating the method’s potential for practical applications.

Additionally, this method is non-destructive and highly adaptable, making it suitable for phenotypic monitoring and variety evaluation in large-scale planting areas, thus providing technical support for the application of intelligent tobacco phenotyping in agricultural production.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

JY: Writing – original draft, Methodology, Writing – review & editing, Conceptualization. WW: Data curation, Supervision, Writing – original draft. HF: Writing – original draft, Supervision, Data curation. ZD: Validation, Writing – review & editing, Software, Visualization. GC: Writing – review & editing, Project administration, Funding acquisition, Resources. SW: Formal analysis, Funding acquisition, Investigation, Writing – review & editing. DW: Funding acquisition, Investigation, Formal analysis, Writing – review & editing. WS: Funding acquisition, Writing – review & editing, Resources, Project administration. XC: Writing – review & editing, Conceptualization, Methodology, Writing – original draft.

Funding

The author(s) declare financial support was received for the research and/or publication of this article. This research was supported by key funding from the China National Tobacco Corporation (CNTC) (No. 110202101003(JY-03)), the China Tobacco Hunan Industrial Co., Ltd. Research Project (KY2024YC0015), the Ministry of Finance and Ministry of Agriculture and Rural Affairs: National Modern Agricultural Industry Technology System (CARS-16-E11), the Scientific Research Fund of Hunan Provincial Education Department (23A0178), the Yuelushan Laboratory Talent Project (2024RC2097), and the Graduate Student Scientific Research and Innovation Project at Hunan Agricultural University (2024XKC059).

Acknowledgments

We sincerely thank Hunan Agricultural University for their invaluable technical assistance throughout this research. We also gratefully acknowledge the Technology Center of China Tobacco Hunan Industrial Co., Ltd. for providing materials and experimental resources. Additionally, we deeply appreciate the collaborative efforts of all team members.

Conflict of interest

Authors SW and DW were employed by China Tobacco Hunan Industrial Co., Ltd.

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Generative AI was used in the creation of this manuscript.

Any alternative text (alt text) provided alongside figures in this article has been generated by Frontiers with the support of artificial intelligence and reasonable efforts have been made to ensure accuracy, including review by the authors wherever possible. If you identify any issues, please contact us.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Al-Rawabdeh, A., He, F., and Habib, A. (2020). Automated feature-based down-sampling approaches for fine registration of irregular point clouds. Remote Sensing. 12. doi: 10.3390/rs12071224

Crossref Full Text | Google Scholar

Briechle, S., Krzystek, P., and Vosselman, G. (2020). Classification of tree species and standing dead trees by fusing Uav-based lidar data and multispectral imagery in the 3d deep neural network pointnet++. ISPRS Ann. Photogrammetry Remote Sens. Spatial Inf. Sci. V-2-2020, 203–210. doi: 10.5194/isprs-annals-V-2-2020-203-2020

Crossref Full Text | Google Scholar

Chen, J., Liu, F., Ren, D., Guo, L., and Liu, Z. (2024). Point cloud completion via relative point position encoding and regional attention. IEEE Trans. Emerging Topics Comput. Intelligence. 8, 3807–3820. doi: 10.1109/tetci.2024.3375614

Crossref Full Text | Google Scholar

Chen, X., Wen, S., Zhang, L., Lan, Y., Ge, Y., Hu, Y., et al. (2025). A calculation method for cotton phenotypic traits based on unmanned aerial vehicle LiDAR combined with a three-dimensional deep neural network. Comput. Electron. Agriculture. 230. doi: 10.1016/j.compag.2024.109857

Crossref Full Text | Google Scholar

Cheng, H., Zhao, B., Zhang, Z., and Liu, X. (2021). Research on classification method of 3D point cloud model based on KE-pointVNet. J. Physics: Conf. Series. 1966. doi: 10.1088/1742-6596/1966/1/012023

Crossref Full Text | Google Scholar

Das Choudhury, S., Maturu, S., Samal, A., Stoerger, V., and Awada, T. (2020). Leveraging image analysis to compute 3D plant phenotypes based on voxel-grid plant reconstruction. Front. Plant Sci. 11. doi: 10.3389/fpls.2020.521431

PubMed Abstract | Crossref Full Text | Google Scholar

Deng, C., Peng, Z., Chen, Z., and Chen, R. (2023). Point cloud deep learning network based on balanced sampling and hybrid pooling. Sensors (Basel). 23. doi: 10.3390/s23020981

PubMed Abstract | Crossref Full Text | Google Scholar

Ferrara, R., Virdis, S. G. P., Ventura, A., Ghisu, T., Duce, P., and Pellizzaro, G. (2018). An automated approach for wood-leaf separation from terrestrial LIDAR point clouds using the density based clustering algorithm DBSCAN. Agric. For. Meteorology. 262, 434–444. doi: 10.1016/j.agrformet.2018.04.008

Crossref Full Text | Google Scholar

Gao, L., Zhao, Y., Han, J., and Liu, H. (2022). Research on multi-view 3D reconstruction technology based on SFM. Sensors (Basel). 22. doi: 10.3390/s22124366

PubMed Abstract | Crossref Full Text | Google Scholar

Gaonkar, M. N. and Sawant, K. (2013). AutoEpsDBSCAN: DBSCAN with Eps automatic for large dataset. Int. J. Advanced Comput. Theory Eng. 2, 11–16.

Google Scholar

Ghahremani, M., Williams, K., Corke, F., Tiddeman, B., Liu, Y., Wang, X., et al. (2021). Direct and accurate feature extraction from 3D point clouds of plants using RANSAC. Comput. Electron. Agriculture. 187. doi: 10.1016/j.compag.2021.106240

Crossref Full Text | Google Scholar

Gonçalves, G., Gonçalves, D., Gómez-Gutiérrez, Á., Andriolo, U., and Pérez-Alvárez, J. A. (2021). 3D reconstruction of coastal cliffs from fixed-wing and multi-rotor UAS: impact of SfM-MVS processing parameters, image redundancy and acquisition geometry. Remote Sensing. 13. doi: 10.3390/rs13061222

Crossref Full Text | Google Scholar

Harintaka, H. and Wijaya, C. (2023). Automatic point cloud segmentation using RANSAC and DBSCAN algorithm for indoor model. TELKOMNIKA (Telecommunication Computing Electron. Control). 21. doi: 10.12928/telkomnika.v21i6.25299

Crossref Full Text | Google Scholar

Hu, Q., Yang, B., Xie, L., Rosa, S., Guo, Y., Wang, Z., et al. (2020). “RandLA-net: efficient semantic segmentation of large-scale point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Seattle, WA, United States). doi: 10.48550/arXiv.1911.11236

Crossref Full Text | Google Scholar

Kaiser, S., Dias, J. C., Ardila, J. A., Soares, F. L. F., Marcelo, M. C. A., Porte, L. M. F., et al. (2018). High-throughput simultaneous quantitation of multi-analytes in tobacco by flow injection coupled to high-resolution mass spectrometry. Talanta 190, 363–374. doi: 10.1016/j.talanta.2018.08.007

PubMed Abstract | Crossref Full Text | Google Scholar

Koh, J. C. O., Spangenberg, G., and Kant, S. (2021). Automated machine learning for high-throughput image-based plant phenotyping. Remote Sensing. 13. doi: 10.3390/rs13050858

Crossref Full Text | Google Scholar

Li, D., Cao, Y., Tang, X. S., Yan, S., and Cai, X. (2018). Leaf segmentation on dense plant point clouds with facet region growing. Sensors (Basel). 18. doi: 10.3390/s18113625

PubMed Abstract | Crossref Full Text | Google Scholar

Li, Y., Liu, J., Zhang, B., Wang, Y., Yao, J., Zhang, X., et al. (2022). Three-dimensional reconstruction and phenotype measurement of maize seedlings based on multi-view image sequences. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.974339

PubMed Abstract | Crossref Full Text | Google Scholar

Luo, H., Zhang, J., Liu, X., Zhang, L., and Liu, J. (2024). Large-scale 3D reconstruction from multi-view imagery: A comprehensive review. Remote Sensing. 16. doi: 10.3390/rs16050773

Crossref Full Text | Google Scholar

Miao, T., Zhu, C., Xu, T., Yang, T., Li, N., Zhou, Y., et al. (2021). Automatic stem-leaf segmentation of maize shoots using three-dimensional point cloud. Comput. Electron. Agriculture. 187, 106310. doi: 10.1016/j.compag.2021.106310

Crossref Full Text | Google Scholar

Mirande, K., Godin, C., Tisserand, M., Charlaix, J., Besnard, F., and Hetroy-Wheeler, F. (2022). A graph-based approach for simultaneous semantic and instance segmentation of plant 3D point clouds. Front. Plant Sci. 13. doi: 10.3389/fpls.2022.1012669

PubMed Abstract | Crossref Full Text | Google Scholar

Patel, A. K., Park, E.-S., Lee, H., Priya, G. G. L., Kim, H., Joshi, R., et al. (2023). Deep learning-based plant organ segmentation and phenotyping of sorghum plants using liDAR point cloud. IEEE J. Selected Topics Appl. Earth Observations Remote Sensing. 16, 8492–8507. doi: 10.1109/jstars.2023.3312815

Crossref Full Text | Google Scholar

Phan, A. V., Nguyen, M. L., Nguyen, Y. L. H., and Bui, L. T. (2018). DGCNN: A convolutional neural network over large-scale labeled graphs. Neural Netw. 108, 533–543. doi: 10.1016/j.neunet.2018.09.001

PubMed Abstract | Crossref Full Text | Google Scholar

Qi, C. R., Su, H., Mo, K., and Guibas, L. J. (2017a). “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of the IEEE conference on computer vision and pattern recognition (Honolulu, HI, United States). doi: 10.1109/CVPR.2017.16

Crossref Full Text | Google Scholar

Qi, C. R., Yi, L., Su, H., and Guibas, L. (2017b). Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Adv. Neural Inf. Process. Syst. 30. doi: 10.48550/arXiv.1706.02413

Crossref Full Text | Google Scholar

Qian, G., Li, Y., Peng, H., Mai, J., Hammoud, H., Elhoseiny, M., et al. (2022). Pointnext: Revisiting pointnet++ with improved training and scaling strategies. Adv. Neural Inf. Process. systems. 35, 23192–23204. doi: 10.48550/arXiv.2206.04670

Crossref Full Text | Google Scholar

Saeed, F., Sun, S., Rodriguez-Sanchez, J., Snider, J., Liu, T., and Li, C. (2023). Cotton plant part 3D segmentation and architectural trait extraction using point voxel convolutional neural networks. Plant Methods 19, 33. doi: 10.1186/s13007-023-00996-1

PubMed Abstract | Crossref Full Text | Google Scholar

Sarker, S., Sarker, P., Stone, G., Gorman, R., Tavakkoli, A., Bebis, G., et al. (2024). A comprehensive overview of deep learning techniques for 3D point cloud classification and semantic segmentation. Mach. Vision Applications. 35. doi: 10.1007/s00138-024-01543-1

Crossref Full Text | Google Scholar

Shi, W., van de Zedde, R., Jiang, H., and Kootstra, G. (2019). Plant-part segmentation using deep learning and multi-view vision. Biosyst. Engineering. 187, 81–95. doi: 10.1016/j.biosystemseng.2019.08.014

Crossref Full Text | Google Scholar

Sun, Y., Luo, Y., Chai, X., Zhang, P., Zhang, Q., Xu, L., et al. (2021). Double-threshold segmentation of panicle and clustering adaptive density estimation for mature rice plants based on 3D point cloud. Electronics 10. doi: 10.3390/electronics10070872

Crossref Full Text | Google Scholar

Turgut, K., Dutagaci, H., Galopin, G., and Rousseau, D. (2022). Segmentation of structural parts of rosebush plants with 3D point-based deep learning methods. Plant Methods 18, 20. doi: 10.1186/s13007-022-00857-3

PubMed Abstract | Crossref Full Text | Google Scholar

Wu, X., Lao, Y., Jiang, L., Liu, X., and Zhao, H. (2022). Point transformer v2: Grouped vector attention and partition-based pooling. Adv. Neural Inf. Process. Systems. 35, 33330–33342. doi: 10.48550/arXiv.2210.05666

Crossref Full Text | Google Scholar

Yang, X., Miao, T., Tian, X., Wang, D., Zhao, J., Lin, L., et al. (2024). Maize stem–leaf segmentation framework based on deformable point clouds. ISPRS J. Photogrammetry Remote Sensing. 211, 49–66. doi: 10.1016/j.isprsjprs.2024.03.025

Crossref Full Text | Google Scholar

Yang, Y., Zhang, J., Wu, K., Zhang, X., Sun, J., Peng, S., et al. (2021). 3D point cloud on semantic information for wheat reconstruction. Agriculture 11. doi: 10.3390/agriculture11050450

Crossref Full Text | Google Scholar

Zhang, Z., Yang, B., Wang, B., and Li, B. (2023). “Growsp: Unsupervised semantic segmentation of 3d point clouds,” in Proceedings of the IEEE/CVF Conference on computer vision and pattern recognition (Vancouver, BC, Canada). doi: 10.48550/arXiv.2305.16404

Crossref Full Text | Google Scholar

Zhao, Q., Gao, X., Li, J., Luo, L., and Liguori, R. (2021). Optimization algorithm for point cloud quality enhancement based on statistical filtering. J. Sensors. 2021. doi: 10.1155/2021/7325600

Crossref Full Text | Google Scholar

Zhu, B., Zhang, Y., Sun, Y., Shi, Y., Ma, Y., and Guo, Y. (2023). Quantitative estimation of organ-scale phenotypic parameters of field crops through 3D modeling using extremely low altitude UAV images. Comput. Electron. Agriculture. 210. doi: 10.1016/j.compag.2023.107910

Crossref Full Text | Google Scholar

Zięba-Kulawik, K., Skoczylas, K., Wężyk, P., Teller, J., Mustafa, A., and Omrani, H. (2021). Monitoring of urban forests using 3D spatial indices based on LiDAR point clouds and voxel approach. Urban Forestry Urban Greening 65. doi: 10.1016/j.ufug.2021.127324

Crossref Full Text | Google Scholar

Keywords: UAV remote sensing, 3D point cloud, deep learning, phenotypic trait extraction, stem-leaf segmentation

Citation: Yao J, Wang W, Fu H, Deng Z, Cui G, Wang S, Wang D, She W and Cao X (2025) Automated measurement of field crop phenotypic traits using UAV 3D point clouds and an improved PointNet++. Front. Plant Sci. 16:1654232. doi: 10.3389/fpls.2025.1654232

Received: 26 June 2025; Accepted: 25 August 2025;
Published: 12 September 2025.

Edited by:

Alejandro Isabel Luna-Maldonado, Autonomous University of Nuevo León, Mexico

Reviewed by:

Ajay Kumar Patel, Saint Louis University, United States
Lei Lei, Chang’an University, China

Copyright © 2025 Yao, Wang, Fu, Deng, Cui, Wang, Wang, She and Cao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Wei She, d2Vpc2hlQGh1bmF1LmVkdS5jbg==; Xiaolan Cao, Y3hsQGh1bmF1Lm5ldA==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.